Abstract
The notion of entropy is used to compare the complexity associated with 12 common cancers based on metastatic tumor distribution autopsy data. We characterize powerlaw distributions, entropy and KullbackLiebler divergence associated with each primary cancer as compared with data for all cancer types aggregated. We then correlate entropy values with other measures of complexity associated with Markov chain dynamical systems models of progression. The Markov transition matrix associated with each cancer is associated with a directed graph model where nodes are anatomical locations where a metastatic tumor could develop and edge weightings are transition probabilities of progression from site to site. The steadystate distribution corresponds to the autopsy data distribution. Entropy correlates well with the overall complexity of the reduced directed graph structure for each cancer and with a measure of systemic interconnectedness of the graph, called graph conductance. The models suggest that grouping cancers according to their entropy values, with skin, breast, kidney and lung cancers being prototypical high entropy cancers, stomach, uterine, pancreatic and ovarian being midlevel entropy cancers and colorectal, cervical, bladder and prostate cancers being prototypical low entropy cancers, provides a potentially useful framework for viewing metastatic cancer in terms of predictability, complexity and metastatic potential.
Introduction
Metastatic cancer is a dynamic disease of relentlessly increasing entropy. From an initial primary tumor located at a single anatomical site (a zeroentropy state), the metastatic cascade leads to a proliferation of tumors at other sites on a timescale of months, or years in most cases, if left untreated^{1,2,3,4}. Entropy is a quantity deeply connected with notions of complexity and predictability^{5,6} used primarily in the fields of information theory^{7,8,9} and statistical thermodynamics^{10}. It is used to quantify the level of disorder associated with a stochastic dynamical process that has a number of sites that it can occupy^{11,12}. Systems that can visit these sites with relatively equal probability have higher entropy (they are considered more disordered and less predictable) than systems that can only occupy a few sites with very different probabilities (considered less disordered and more predictable). In this paper we demonstrate how entropy, relative entropy^{6,13} and graph conductance^{30} can be used in the context of metastatic spread to quantify, compare and cogroup the 12 most prevalent cancer types worldwide. The view of cancer that we describe is based on its dynamical characteristics, which offers a more nuanced view than the static view of classifying only according to site of origin. To put it differently, we characterize cancers based not just on their initial conditions, but on a collection of features that are associated with their dynamical predictability throughout the course of disease progression.
To fix ideas further, suppose each anatomical site where a primary or metastatic tumor could appear is indexed by ‘i’, (i = 1,…N). Let σ_{i} represent the probability that site ‘i’ is occupied (i.e. has a metastatic tumor) and let represent a probability mass distribution over a collection of potentially occupied sites, so that , with 0 ≤ σ_{j} ≤ 1. The level of disorder associated with the distribution is captured by a scalar quantity H_{N}_{′}called the entropy of the state, a quantity that is a function both of N and the way the probabilities are distributed among the N sites. The lowest entropy state, corresponding to the one of least disorder, would be represented by a distribution such as , in which case H_{N} = 0. In this distribution, state i = 3 is occupied with probability 1, making it predictively certain. Typically, this site would be the anatomical location of the primary tumor in a Stage I patient whose disease has not yet progressed. The highest entropy state, corresponding to the one of most disorder, would be represented by the uniform distribution . For this uniform distribution, each site is occupied with equal probability. This distribution, which constitutes a mathematical upper bound on the entropy, represents a state of maximal disorder. It corresponds to the least predictable state. In the case where only two states are possible (N = 2), a familiar example of a maximum entropy state would be the probability of heads (H) and tails (T) when flipping a fair coin, P(H) = ½, P(T) = ½, . The point we want to emphasize is that associated with any specific probabilistic distribution of occupied sites (typically falling between the above two extremes, as shown, for example in Figure 1 for all cancers grouped together), is a quantitative notion of disorder, which in turn is related to the system’s predictability and complexity^{5,6,11,12}. Since each cancer type has a different empirical metastatic tumor distribution, each will have a different entropy value and these entropy values can be thought of as convenient and simple surrogates that represent metastatic tumor complexity and disorder associated with ensemble populations of patients with a given primary cancer type.
We show in this paper how entropy is a useful metric of metastatic complexity that correlates well with features associated with the anatomical pathways of disease progression, as well as the graph conductance measuring the density of the network diagram, which in turn is associated with rates of convergence to the steadystate tumor distribution of the disease in a population of patients. We note that notions of entropy have been used fruitfully to characterize other aspects of metastatic cancer both at the genotypic and phenotypic levels^{24,25,26,27,29,33,34,35,36} but none have related it to large scale progression patterns with the goal of quantifying complexity and predictability.
Results
Distribution of metastatic tumors from autopsy data
Figure 1 shows the tumor distribution for all cancers collected from 3827 untreated patients, with a total of 9484 metastatic tumors distributed over 30 distinct anatomical sites^{14}. On the left we show the histograms, normalized so that the total area under the bars is one, hence the distribution represents the probability mass function associated with ‘all’ cancers. On the right we show the same data plotted on a loglog plot to more clearly bring out the fact that there is a powerlaw scaling range, where the distribution follows the form p(x)~x^{−α}, with α = 1.46, obtained using maximum likelihood estimators and a goodnessoffit criterion for the optimal range over which the powerlaw holds^{15}. We note that powerlaw distributions arise in many other contexts, most relevant might be the distribution of edges from nodes on the WorldWideWeb^{16}. The analogy of websurfing from sitetosite and modeling cancer progression as a random walk process from sitetosite has been used fruitfully in^{17,18} and is the basis for the Markov diagrams described later.
The panels shown in Figure 2al break the data of Figure 1 into 12 groupings associated with 12 major primary cancer types (a. Skin; b. Breast; c. Kidney; d. Lung; e. Stomach; f. Uterine; g. Pancreatic; h. Ovarian; i. Colorectal; j. Cervical; k. Bladder; l. Prostate) and the ensemble metastatic distributions associated with each. Each of the empirical distributions shows a clear powerlaw range (details are described in Figure caption), each with a distinct powerlaw exponent and approximate range of validity.
As detailed in^{17,18} for our lung cancer model, these metastatic tumor distributions are used to construct Markov chain dynamical system models of cancer progression. Markov transition matrices^{28} are constructed conditioned on (i) these distributions being steadystates and (ii) an initial ‘trial’ matrix which is used to produce converged Markov matrices for each cancer type using a stochastic iteration scheme, where the entries of the trial matrix are obtained from the data. Details are described in the Materials and Methods section and specifically for the lung cancer model in^{17,18}.
Metastatic entropy for 12 major cancer types
Because of well known difficulties inherent with pinning down precise values for powerlaw exponents^{15}, we do not use their value for comparative purposes. For the purposes of quantifying the complexity associated with each primary cancer type, we calculate the entropy associated with each, given by the formula , where σ_{i} represents the proportion of metastatic tumors found at anatomical site ‘i’, for a given primary cancer type. The constraints are given by 0 ≤ σ_{j} ≤ 1, (i = 1, …, N), . Entropy has been used in various contexts related to cancer in the literature, see for example^{33,34,35,36}. It should be intuitively clear that an increase in complexity is associated with two distinct features associated with each of the distributions: (i) the total number of sites, N, at which metastatic tumors are found and (ii) relatively flat distributions, meaning that the probabilities of spreading to each site are more equally probable than what a steep distribution would show. Both of these factors play an important role in the entropy values. Table 1 shows the value of the metastatic entropy for each of the 12 cancer types, as well as the ‘All Cancer’ aggregated data. The first column lists the primary cancer type, the second column lists the number of sites, N, over which the metastatic tumors are distributed, while the third column lists the metastatic entropy associated with the empirical distributions shown in Figures 1 and 2. We list the sites according to the descending values of the entropy shown in Table 1, column 3, thus skin (2.9945), breast (2.7798), kidney (2.7554) and lung (2.7454) all have entropy values higher than the value for all cancers combined (2.7136), which we use as a benchmark for comparisons. The cancer type with the lowest entropy value is prostate (2.0960), consistent with the relatively small number of sites to which it distributes (N = 21) and the relatively sharp drop in the empirical distribution shown in Figure 2l. It is useful to compare this distribution with skin, shown in Figure 2a, which has more sites to which it distributes itself (N = 30) and has a distinctly flatter distribution to those sites. For ovarian cancer, whose entropy is relatively low (2.5193), we have grouped large intestine, small intestine, diaphragm, ovary, omentum and peritoneum all as one site which we call ‘peritoneal cavity’, due to the fact that metastases in each of these regions likely represent random spread of disease within an anatomically connected region, as opposed to hematogenously seeded metastases. To get an idea of the robustness of these ‘Autopsy Entropy’ values, we show in column 4 ‘Sample Entropy’ values, with +/− standard deviations. These are computed using the Markov transition matrices for each cancer type to produce artificial sample populations (100 sample populations for each cancer type, each with the same number of patients as in the autopsy data) with the same statistical characteristics as the autopsy populations (i.e. same total number of patients, same number of metastatic tumors and tumor distributions). We use these sample populations to obtain standard deviation values.
Relativeentropy between each primary cancer type and the aggregate entropy associated with all cancers
Columns 5 and 6 in Table 1 show the KullbackLiebler divergence^{8} between each cancer type and the all cancer category. We use Q as the all cancer distribution, while P is the distribution associated with each specific cancer type. While the value of entropy shown in column 3 is independent of the ordering in which the sites are listed, the KL divergence is not. In column 6 we calculate this quantity using the P distribution and the Q distribution arranged in decreasing order in each case. As Table 1 column 5 indicates, the KL divergence between prostate and ‘All’ is the highest (0.1620), indicating that its shape is most different from the all cancer category. By contrast, stomach cancer has the smallest KL divergence from the all cancer group (0.0213), making it in this sense, the most similar to the aggregate.
Column 6 in Table 1 shows the KL computations between each of the cancer types and ‘All’ on a site specific basis, as shown in Figure 3. Here, we list the sites in decreasing order according to the all cancer category, meaning that the comparative histogram heights for each of the specific primary cancers generally are not arranged in strictly decreasing order. Thus, on this sitespecific way of computing the KL divergence, ovarian cancer (0.7995) and prostate cancer (0.2750) have the largest values, making them the most distinct from the ‘all’ cancer aggregate on a sitebysite comparison. By contrast, breast cancer (0.0759) and cervical cancer (0.0979) have the smallest values of site specific KL divergence, meaning these are the most similar to the all cancer aggregate. In Figure 4, we show the same histograms as in Figure 3, but we arrange the sites in order of decreasing size. This way of comparing the distributions focuses on the shape of the distribution, i.e. the rate at which it drops to zero, rather than the actual sites to which the disease spreads.
Markov diagrams, spreaders and sponges and graph conductance
The notion of entropy of metastatic tumor distributions is closely tied to systemic complexity of the disease, which in turn is tied to the fact that the metastatic process, both on the molecular level^{24,25,26,27,29} and across anatomical scales, is on average, an entropy increasing (or at least nondecreasing) dynamical process. The dynamics of progression from one anatomical site to the next can be captured reasonably well by modeling it as a Markovian process^{28,30} as it spreads from site to site. The statevector, at discrete timestep k has entries that represent the distributed probabilities of a tumor developing at anatomical site ‘i’, 1 ≤ i ≤ N. The transition matrix which is made up of probabilities of tumor spread from site ‘i’ to site ‘j’, propagates the statevector forward in time via the Markov equations . By using patient population data (this could be longitudinal or autopsy data), we obtain estimates of the model parameters, which are the transition probabilities that fill out the transition matrix A. See Materials and Methods section for details. Figure 5 shows the entropy for each cancer type, as the discrete model timestep, k, advances forward. For each of the cancers, the entropy values start at zero (at step k = 0, only the primary site has a tumor), then increase to their maximum value (corresponding to the steadystate) as the disease progresses, confirming that metastatic cancer is an entropy increasing process.
The conductance is a measure of how fast a random walk converges to its steadystate distribution, which is tied to the Markov mixing time and convergence rate^{30}. From the conductance values listed in Table 2, we can see a clear (although not perfectly onetoone) correlation of entropy value and graph conductance value. The highest entropy and conductance cancers are skin, breast, kidney and lung, the lowest in both are cervical, bladder and prostate cancers. Table 2 summarizes the network based data for the 12 cancers. Note the overall decrease in graph conductance values listed in the 6^{th} column of the table, correlating roughly with decreasing entropy values for each. The correlation is more clearly shown in Figure 6 across all 12 cancer types.
The reduced Markov diagrams for each of the 12 cancers, listed in the same order of decreasing entropy, are shown in Figure 7. These diagrams are based on retaining only the top 30 twostep pathways from primary site, to the first metastatic site, to the second metastatic site^{17,18}. The % listed under the main primary circle represents the % that these 30 paths capture out of the total, hence is a separate measure of complexity of the cancer type. The top 30 paths associated with skin cancer (Figure 7a), for example, capture only 23.8% of the total, indicating that it has a more diversified set of alternative pathways available to it than, say, prostate cancer (Figure 7l), in which the top 30 pathways captures over 80% of the total. In order of decreasing entropy of each of the cancers whose reduced Markov diagrams are laid out from Figure 7a (highest entropy) down to Figure 7l (lowest entropy), the % of total pathways captured by the top 30 paths clearly increases (although, not exactly onetoone), indicating that high entropy cancers have many alternative pathways and high graph conductance value (fast convergence rates to steadystate), whereas low entropy cancers have few pathways available and low graph conductance (slower convergence rates to steadystate). Table 3 summarizes these metrics for the 12 cancer types, listed in order of decreasing entropy. Column 2 shows clearly how the percentages covered by the top 30 pathways increase with entropy, column 3 shows how the number of paths associated with a fixed % (in the case 35%) decreases with decreasing entropy.
Also summarized in this table are the main spreader and sponge metastatic sites^{18} associated with each tumor type. This notion, developed and used for lung cancer models in^{18} is based on a calculation of the probability out (P_{out}) of each node in the directed graph as compared with the probability in (P_{in}), obtained by adding up the edge weights of all the outward directed edges from a site compared with the inward directed edges. A node in which P_{out} > P_{in} is called a spreader site, whereas a node in which P_{out} < P_{in} is called a sponge site. From our lung cancer models and data, we know that the combined characteristics of the primary tumor and the first metastatic site to which it spreads is an important determining factor of the future course of disease progression, particularly if the first metastatic site is a spreader site associated with that cancer. At this point, for cancers other than lung, this is a graph based metric only that would need further clinical and biological correlates.
Discussion
Grouping cancers according to their dynamical and probabilistic characteristics offers an alternative point of view from the classical approach of classifying according to site of origin. In our models, site of origin does play an important role in determining which transition matrix governs disease progression and therefore a crucial role in determining which metastatic sites are spreaders and which are sponges and pinpointing the entropy values of each cancer. But grouping cancers according to metrics that are associated with key dynamical features such as entropy and graph conductance is a potentially enlightening way to think about similarities and differences between cancers based on their comparative metastatic potential. It provides a quantitative framework that could help guide clinical strategies whose end goals could be restated in language highlighting, for example, entropy reduction strategies and strategies that decrease mixing times, with clear ways of measuring and optimizing these quantities. Similarly, identifying therapeutic strategies that target metastatic spreaders, particularly in the oligometastatic setting^{19,20,32}, might prove to be an area where these mathematical models could be of particular clinical value.
Methods
Summary of autopsy data set
We used the DiSibio and French^{14} data set of metastatic tumor distributions based on autopsy studies collected for 3827 untreated cadavers from 5 different cancer facilities in New England between 19141943. The data reflect 9484 distinct metastatic tumors distributed over 30 anatomical sites for all of the major tissue cancers. The data represent natural disease progression, which is useful, but we caution that brain metastases are underrepresented in the data since examination of the intracranial contents at that time was not routinely performed. See also studies such as^{21} which focus on anatomical progression patterns. The data has been used in^{17,18,31} to develop a Markov chain model for lung cancer progression, where the autopsy data is used as the Markov chain steadystate, from which transition probabilities are calculated. In this paper, we directly characterize the data, shown in Figure 1 (All cancers) and Figure 2 (12 different primary cancers) in terms of their empirical distributions, which predominantly follow powerlaw form^{15}. Other related work focusing on the development of dynamical models based on metastatic progression patterns includes references^{18,21,22,23}. While notions of entropy have been used previously in the context of gene expression profiles and epidemiology^{24,25,26,27,29}, we know of no previous work that uses these notions to characterize the complexity of largescale progression patterns.
Definition of entropy
The notion of entropy we use is from the field of information theory and statistical mechanics^{7,8,9,10}. Given a probabilistic distribution of states spread over N sites, the entropy associated with the distribution is given by the quantity where 0 ≤ H_{N} ≤ ln N. [To be clear, we are using the natural logarithm to define the entropy, hence the unit of measurement is commonly denoted as ‘nat’^{8}. If base 2 logarithms were used, the units would be ‘bits’. One nat corresponds to 1.44 bits.] There are two factors that lead to increased entropy: (i) the larger the number N of sites over which the disease is distributed, the larger the entropy; (ii) the more even the probabilities are distributed among those sites, the larger the entropy. Thus, the lowest entropy state, given by H_{N} = 0, corresponds to the distribution σ_{k} = 1, σ_{i} = 0 (i ≠ k)Since the probability of site ‘k’ being occupied is 1 and the probability of sites i ≠ k being occupied is 0, this state is associated with predictive certainty. In the language of statistical thermodynamics^{10}, this would be called a completely ordered state. By contrast, the highest entropy state corresponds to the uniform distribution in which each site is equally probable, hence This uniform distribution gives rise to a maximal entropy value of H_{N} = ln(N) For this distribution, since each site is equally likely to occur with probability 1/N, the predictive certainty associated with this distribution is minimal, yielding the highest possible entropy value. We note that the entropy value is independent of the ordering of the sites. Thus, higher values of entropy are intimately tied to notions of disorder and complexity and have been used productively across a wide range of disciplines^{6,7,8,9,10,11,12,13}.
Definition of relativeentropy
The concept of relative entropy, or KullbackLiebler distance, is used to measure the distance between two distributions of random variables^{8,13}. One way to think of the relative entropy D(PQ) between two random variables P and Q is to view D(PQ) as a measure of inefficiency associated with assuming that the distribution is Q, when in fact the true distribution is P^{8,24}.
It is defined as . In our comparisons, we use the symbol Q to represent the ‘All’ cancer empirical distribution, whereas P will represent a specific primary cancer type. Thus, the notion of relative entropy quantifies the relative inefficiency of using the generic ‘All’ cancer distribution instead of the more targeted and informative primary cancer type.
Definition of graph conductance
Graph conductance is a measure of graph density, or how ‘wellknit’ the full graph is. The conductance is a measure of how fast a random walk converges to its steadystate distribution, which is tied to the Markov mixing time and convergence rates^{30}. It is calculated for a network, A, by first partitioning A into two distinct sets (where comprises all of the nodes in the network). The conductance associated with that particular partition is computed as the sum of the transition probabilities from all the sites in S to all those in , normalized by dividing by the sum of transition probabilities from sites in S or sites in to ALL sites in the network , whichever of those two numbers is smaller. Then, the graph conductance is the minimum conductance achieved by calculating the conductance associated with all possible partitions of the network. More formally, it is defined as:
with
where α_{ij} are the transition probabilities. Note that it is a quantity which uses the edge weightings of the directed graph, not just the adjacency matrix values.
Construction of Markov transition matrices
The Markov transition matrices used in this paper are calculated using the same conditional random search algorithm used to construct the lung transition matrices described in^{17,18}. Briefly, they are calculated using the following procedures:

1
The ‘target’ steadystate distribution for a given cancer type, denoted , is defined as the right eigenvector = 0, of transition matrix , corresponding to unit eigenvalue. To construct the transition matrix for a given cancer type, we take the target steadystate distribution to correspond to the probability mass function associated with that cancer type from the data shown in Figure 2.

2
To construct the ‘final’ transition matrix for a given cancer type, A_{f}, we construct a sequence of increasingly accurate approximations to A_{f}, denoted (A_{0},A_{1},…,A_{i},…), with corresponding steadystate vectors so that and , where is called a residual vector, where and is the identity matrix. We condition our sequence based on an initial matrix A_{0}, chosen so that the row associated with the primary tumor corresponds to the empirical distribution for that cancer type from the data, shown in Figure 2. All other rows are scaled so that they correspond to the empirical distributions associated with the appropriate cancer type, as shown in Figure 2. Note that this initial assumption cannot be the final converged transition matrix because it treats each metastatic tumor as a primary tumor. The iteration scheme then corrects for this quite efficiently in all 12 cases we considered and finds the converged transition matrix which is nearest to this initial matrix. More details of this scheme and issues associated with uniqueness, robustness and convergence are described for the case of lung cancer in Ref. 17.

3
To construct A_{i}_{+1} from A_{i} (the iteration procedure), starting with A_{0}, we randomly perturb the entries of A_{i}, in sequence from upper left to bottom right (see^{17} for more details), each time computing the corresponding residual . For each perturbation, if , we keep the perturbation, if not, we discard it and perturb the next entry. The size of the perturbations are chosen to scale with the size of the residual vector, i.e. so that as we get closer and closer to converging to A_{f}, the perturbations get smaller.

4
When , for a given convergence threshold value of ϵ (typically taken as O(10^{−5}))), we stop the iteration and take the corresponding matrix A_{i} as our converged transition matrix for that cancer type. We document the robustness and stability of the scheme by plotting the full set of singular values associated with an ensemble of Markov matrices produced this way, as well the one produced by averaging the ensemble – the method is robust, stable and produces a unique Markov matrix with the correct steadystate for all 12 cancers studied in this paper.
References
Weiss, L. Metastasis of cancer: a conceptual history from antiquity to the 1990's. Cancer Metastasis Rev. 19, 193–204 (2000).
Fidler, I. J. Timeline: The pathogenesis of cancer metastasis: the ‘seed and soil' hypothesis revisited. Nat. Rev. Cancer 3, 453–458 (2003).
Chambers, A. F., Groom, A. C. & MacDonald, I. C. Dissemination and growth of cancer cells in metastatic sites. Nat. Rev. Cancer 2, 563–573 (2002).
Weinberg, R. A. The Biology of Cancer (Garland Science, New York, 2006).
Haven, K., Majda, A. J. & Abramov, R. Quantifying predictability through information theory: small sample estimation in a nonGaussian framework. J. Comp. Phys. 206, 334–362 (2005).
Kleeman, R. Information theory and dynamical system predictability. Entropy 13, 612–649 (2011).
Shannon, C. E. A mathematical theory of communication. Bell. Syst. Tech. J. 27, 379–423 623–656 (1948).
Cover, T. & Thomas, J. Elements of Information Theory 2^{nd} Ed. (WileyInterscience, New York, 2006).
Jaynes, E. T. Information Theory and Statistical Mechanics [Ford, K. (ed.)] (W.A. Benjamin Inc., New York, 1963).
Kinchin, A. I. Mathematical Foundations of Statistical Mechanics (Dover Publications, New York, 1949).
Pincus, S. M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 88, 2297–2301 (1991).
Crutchfield, J. P. & Young, K. Inferring statistical complexity. Phys. Rev. Lett. 63, 10 (1989).
Majda, A. J., Kleeman, R. & Cai, D. A mathematical framework for quantifying predictability through relative entropy. Meth. Appl. Anal. 9, 425–444 (2002).
DiSibio, G. & French, S. W. Metastatic patterns of cancers: Results from a large autopsy study. Arch. Pathol. Lab. Med. 132, 931–939 (2008).
Newman, M. E. J. Power laws, Pareto distributions and Zipf’s law. Cont. Phys. 46, 323–351 (2005).
Barabasi, A. L. & Albert, R. Emergence of scaling in random networks. Science 286, 509–511 (1999).
Newton, P. K. et al. A stochastic Markov chain model to describe lung cancer growth and metastasis. PLoS ONE 7, e34637 (2012).
Newton, P. K. et al. Spreaders and sponges define metastasis in lung cancer: A Markov chain mathematical model. Cancer Res. 73, 2760–2769 (2013).
Kumar, V. et al. Radiomics: the process and the challenges. Mag. Res. Imaging 30, 1234–1248 (2012).
Corbin, K. S., Hellman, S. & Weichselbaum, R. R. Extracranial oligometastases: A subset of metastases curable with stereotactic radiotherapy. J. Clin. Onc. 31, 1384–1390 (2013).
Chen, L. L., Blumm, N., Christakis, N. A., Barabasi, A. L. & Deisboeck, T. S. Cancer metastasis networks and the prediction of progression patterns. Brit. J. of Cancer 101, 749–758 (2009).
Iwata, K., Kawasaki, K. & Shigesada, N. A dynamical model for the growth and size distribution of multiple metastatic tumors. J. Theor. Biol. 203, 177–186 (2000).
Haustein, V. & Schumacher, U. A dynamic model for tumor growth and metastasis formation. J. Clinical Bio. 2, 10.1186/20439113211, 1–11 (2012).
Rajni, B. & Agrawal, R. K. Mutual information and cross entropy framework to determine relevant gene subset for cancer classification. Informatica 35, 375–382 (2011).
Tritchler, D. L., Sucheston, L., Chanda, P. & Ramanathan, M. Information metrics in genetic epidemiology. Stat. Apps. In Genetics and Mol. Bio. 10, 1–20 (2011).
Ritchie, W., Granjeaud, S., Puthier, D. & Gautheret, D. Entropy measures quantify global splicing disorders in cancer. PLoS Comp. Bio. E1000011, 4, 1–9 (2008).
Molnar, J. et al. Thermodynamic aspects of cancer: possible role of negative entropy in tumor growth, its relation to kinetic and genetic resistance. Lett. in Drug Design and Rec. 2, 429–438 (2005).
Norris, J. R. Markov Chains (Cambridge Series in Statistical and Probabilistic Mathematics, Cambridge University Press, 1997).
Almendro, V. et al. Genetic and phenotypic diversity in breast tumor metastases. Cancer Res. 74, 1338–48 (2014).
Levin, D. A., Peres, Y. & Wilmer, E. L. Markov Chains and Mixing Times (American Mathematical Society, Providence, 2009).
Bazhenova, L. et al. Adrenal metastases in lung cancer: Clinical implications of a mathematical model. J. Thor. Onc. 9, 442–446 (2014).
Tait, C. R., Waterworth, A., Loncaster, J., Horgan, K. & Dodwell, D. The oligometastatic state in breast cancer: hypothesis or reality. Breast 14, 87–93 (2005).
Banerji, C. R. S. et al. Cellular network entropy as the energy potential in Waddington’s differentiation landscape. Nat. Sci. Rep. 3, 3039 (2013).
Gulati, S., Cheng, T. M. K. & Bates, P. Cancer networks and beyond: Interpreting mutations using the human interactome and protein structure. Sem. in Cancer Bio. 23, 219–226 (2013).
West, J., Bianconi, G., Severini, S. & Teschendorff, A. E. Differential network entropy reveals cancer system hallmarks. Nat. Sci. Rep. 2, 802 (2012).
Teschendorff, A. E. & Severini, S. Increased entropy of signal transduction in the cancer metastasis phenotype. BMC Sys. Bio. 4, 104 (2010).
Acknowledgements
The project described was supported by Award Number U54CA143906 from the National Cancer Institute and the Bill and Melinda Gates Foundation through the Gates Millennium Fellowship Program. The content is solely the responsibility of the authors and does not necessarily represent the official view of the National Cancer Institute or the National Institutes of Health.
Author information
Authors and Affiliations
Contributions
P.N. and J.M. wrote the main manuscript text. J.M. prepared figures 3–7 and Tables 1–3. B.H. and J.M. prepared figures 1, 2. K.B., L.B., J.N. and P.K. reviewed and edited the manuscript.
Ethics declarations
Competing interests
The authors declare no competing financial interests.
Rights and permissions
This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article's Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder in order to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
About this article
Cite this article
Newton, P., Mason, J., Hurt, B. et al. Entropy, complexity and Markov diagrams for random walk cancer models. Sci Rep 4, 7558 (2014). https://doi.org/10.1038/srep07558
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/srep07558
This article is cited by

Modeling cancer immunoediting in tumor microenvironment with system characterization through the isingmodel Hamiltonian
BMC Bioinformatics (2022)

TimeSeries Analysis of Tumorigenesis in a Murine Skin Carcinogenesis Model
Scientific Reports (2018)

Investigating cellular network heterogeneity and modularity in cancer: a network entropy and unbalanced motif approach
BMC Systems Biology (2016)

Spatiotemporal progression of metastatic breast cancer: a Markov chain model highlighting the role of early metastatic sites
npj Breast Cancer (2015)

Celltocell communication in cancer: workshop report
npj Breast Cancer (2015)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.