Abstract
Biological organisms adapt to changes by processing informations from different sources, most notably from their ancestors and from their environment. We review an approach to quantify these informations by analyzing mathematical models of evolutionary dynamics and show how explicit results are obtained for a solvable subclass of these models. In several limits, the results coincide with those obtained in studies of information processing for communication, gambling or thermodynamics. In the most general case, however, information processing by biological populations shows unique features that motivate the analysis of specific models.
Similar content being viewed by others
References
Bialek, W.: Biophysics: Searching for Principles. Princeton University Press, Princeton (2013)
Nemenman, I.: Quantitative Biology: From Molecular to Cellular Systems. Information Theory and Adaptation, vol. 4. CRC Press, Boca Raton (2012)
Tkačik, G., Walczak, A.M.: Information transmission in genetic regulatory networks: a review. J. Phys. 23, 153102 (2011)
Brennan, M.D., Cheong, R., Levchenko, A.: How information theory handles cell signaling and uncertainty. Science 338, 334–335 (2012)
Bowsher, C.G., Swain, P.S.: Environmental sensing, information transfer, and cellular decision-making. Curr. Opin. Biotechnol. 28, 149–155 (2014)
Cover, T.M., Thomas, J.A.: Elements of Information Theory. Wiley Interscience, Hoboken (1991)
Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 27, 379 (1948)
Rivoire, O., Leibler, S.: The value of information for populations in varying environments. J. Stat. Phys. 142, 1124–1166 (2011)
Kelly, J.L.: A new interpretation of information rate. Inf. Theor. 2, 185–189 (1956)
Barron, A.R., Cover, T.M.: A bound on the financial value of information. IEEE Trans. Inf. Theor. 34, 1097–1100 (1988)
Bergstrom, C.T., Lachmann, M: Shannon information and biological fitness. IEEE Proceedings of Information Theory, pp. 50–54 (2004)
Kussell, E., Leibler, S.: Phenotypic diversity, population growth, and information in fluctuating environments. Science 309, 2075–2078 (2005)
Taylor, S.F., Tishby, N., Bialek, W.: Information and fitness. arXiv:0712.4382 (2007)
Donaldson-Matasci, M.C., Bergstrom, C.T., Lachmann, M.: The fitness value of information. Oikos 119, 219–230 (2010)
Permuter, H.H., Kim, Y.-H., Weissman, T.: Interpretations of directed information in portfolio theory, data compression, and hypothesis testing. IEEE Trans. Inf. Theor. 57, 3248–3259 (2011)
Cheong, R., Rhee, A., Wang, C.J., Nemenman, I., Levchenko, A.: Information transduction capacity of noisy biochemical signaling networks. Science 334, 354–358 (2011)
Rivoire, O., Leibler, S.: A model for the generation and transmission of variations in evolution. Proc. Nat. Acad. Sci. 111, E1940–E1949 (2014)
Ziv, E., Nemenman, I., Wiggins, C.H.: Optimal signal processing in small stochastic biochemical networks. PLoS One 2, e1077 (2007)
Tkačik, G., Callan, C.G., Bialek, W.: Information flow and optimization in transcriptional regulation. Proc. Nat. Acad. Sci. 105, 12265–12270 (2008)
Tostevin, F., ten Wolde, P.R.: Mutual information between input and output trajectories of biochemical networks. Phys. Rev. Lett. 102, 218101 (2009)
Shannon, C.E.: Communication in the presence of noise. Proc. IRE 37, 10–21 (1949)
Lynch, M., Walsh, B.: Genetics and Analysis of Quantitative Traits. Sinauer, Sunderland (1998)
Kalman, R.: A new approach to linear filtering and prediction problems. J. Basic Eng. 82, 35–45 (1960)
Vinkler, D.A., Permuter, H.H., Merhav, N.: Analogy between gambling and measurement-based work extraction. IEEE International Symposium on Information Theory (ISIT), pp. 1111–1115 (2014)
Parrondo, J.M.R., Horowitz, J.M., Sagawa, T.: Thermodynamics of information. Nat. Phys. 11, 131–139 (2015)
Haccou, P., Iwasa, Y.: Optimal mixed strategies in stochastic environments. Theor. Popul. Biol. 47, 212–243 (1995)
Donaldson-Matasci, M.C., Lachmann, M., Bergstrom, C.T.: Phenotypic diversity as an adaptation to environmental uncertainty. Evol. Ecol. Res. 10, 493–515 (2008)
Swain, P.S., Elowitz, M.B., Siggia, E.D.: Intrinsic and extrinsic contributions to stochasticity in gene expression. Proc. Nat. Acad. Sci. 99, 12795–12800 (2002)
Massey, J.: Causality, feedback and directed information. Proceedings of International Symposium on Information Theory Application (ISITA-90), pp. 303–305 (1990)
Schreiber, T.: Measuring information transfer. Phys. Rev. Lett. 85, 461 (2000)
Amblard, P.-O., Michel, O.: The relation between granger causality and directed information theory: a review. Entropy 15, 113–143 (2013)
Leff, H., Rex, A.F.: Maxwell’s Demon 2 Entropy, Classical and Quantum Information, Computing. CRC Press, Boca Raton (2002)
Sagawa, T., Ueda, M.: Second law of thermodynamics with discrete quantum feedback control. Phys. Rev. Lett. 100, 080403 (2008)
Sagawa, T., Ueda, M.: Nonequilibrium thermodynamics of feedback control. Phys. Rev. E 85, 021104 (2012)
Esposito, M., Van den Broeck, C.: Second law and Landauer principle far from equilibrium. Europhys. Lett. 95, 40004 (2011)
Horowitz, J.M., Parrondo, J.M.R.: Designing optimal discrete-feedback thermodynamic engines. New J. Phys. 13, 123019 (2011)
Abreu, D., Seifert, U.: Extracting work from a single heat bath through feedback. Europhys. Lett. 94, 10001 (2011)
Bauer, M., Abreu, D., Seifert, U.: Efficiency of a Brownian information machine. J. Phys. A 45, 162001 (2012)
Csiszár, I.: Axiomatic characterizations of information measures. Entropy 10, 261–273 (2008)
Horowitz, J.M., Sandberg, H.: Second-law-like inequalities with information and their interpretations. New J. Phys. 16, 125007 (2014)
Cover, T.M., Ordentlich, E.: Universal portfolios with side information. IEEE Trans. Inf. Theor. 42, 348–363 (1996)
Kobayashi, T.J., Sughiyama, Y.: Fluctuation relations of fitness and information in population dynamics. arXiv:1505.02201 (2015)
Hirono, Y., Hidaka, Y.: Jarzynski-type equalities in gambling: role of information in capital growth. arXiv:1505.06216 (2015)
Iwasa, Y.: Free fitness that always increases in evolution. J. Theor. Biol. 135, 265–281 (1988)
Barton, N.H., Coe, J.B.: On the application of statistical physics to evolutionary biology. J. Theor. Biol. 259, 317–324 (2009)
Mustonen, V., Lassig, M.: Fitness flux and ubiquity of adaptive evolution. Proc. Nat. Acad. Sci. 107, 4248–4253 (2010)
Lan, G., Sartori, P., Neumann, S., Sourjik, V., Tu, Y.: The energy-speed-accuracy trade-off in sensory adaptation. Nat. Phys. 8, 422–428 (2012)
Lang, A.H., Fisher, C.K., Mora, T., Mehta, P.: Thermodynamics of statistical inference by cells. Phys. Rev. Lett. 113, 148103 (2014)
England, J.L.: Statistical physics of self-replication. J. Chem. Phys. 139, 121923 (2013)
François, P.: Evolving phenotypic networks in silico. Semin. Cell Dev. Biol. 35, 90–97 (2014)
Hemery, M., Rivoire, O.: Evolution of sparsity and modularity in a model of protein allostery. Phys. Rev. E 91, 042704 (2015)
Acknowledgments
I thank B. Houchmandzadeh and M. Ribezzi for helpful comments.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix 1: Mapping from Eqs. (6) to (2)
The model described by Eq. (6) is mapped to the model described by Eq. (2) by defining
and
where \(\tilde{\phi }^k_t\) corresponds to the k-th component of \(\tilde{\phi }_t\), i.e., \(\tilde{\phi }_t=(\tilde{\phi }_t^1,\tilde{\phi }_t^2)\). Note that \(\tilde{S}\) is of the form \(\tilde{S}=\tilde{K}\tilde{\Delta }\) as in Eq. (1) if S is itself of the form \(S=K\Delta \).
Appendix 2: Decomposition of the Growth Rate
We detail here the decomposition of the growth rate given in Eq. (14) for the discrete model in absence of inheritance. The idea is to write
and to recognize that \(\mathbb {E}_{X,Y}[\ln P_X(x)]=-H(X)\), \(\mathbb {E}_{X,Y}[\ln P_{X|Y}(X|Y)/P_X(X)]=I(X;Y)\) and
Appendix 3: Gaussian Random Variables
A Gaussian random variable X is characterized by its mean \(x_0=\mathbb {E}[X]\) and its variance \(\sigma ^2_x=\mathbb {E}[X^2]-\mathbb {E}[X]^2\), and its probability density is \(P_X(x)=G_{\sigma ^2_x}(x-x_0)\), where \(G_{\sigma ^2_x}(x)=(2\pi \sigma ^2_x)^{-1/2}\exp (-x^2/2\sigma ^2_x)\).
Its differential entropy \(h(X)=-\int \mathrm {d}xP_X(x)\ln P_X(x)\) is
The mutual information \(I(X;Y)=h(X)-h(X|Y)\) between X and another Gaussian random variable Y whose conditional probability given x is \(P_{Y|X}(y|x)=G_{\sigma ^2_{y|x}}(y-x)\) is
The relative entropy between two Gaussian probability densities is
Finally, given \(P_{X_1|X_0}(x_1|x_0)=G_{\sigma ^2_{x_1|x_0}}(x_1-ax_0)\) and \(P_{Y_1|X_1}(y_1|x_1)=G_{\sigma ^2_{y_1|x_1}}(y_1-x_1)\), the conditional probability \(P_{X_1|Y_1,X_0}\), which by Bayes’ rule is proportional to \(P_{Y_1|X_1}P_{X_1|X_0}\), is also Gaussian and given by
with
or, equivalently, \(\sigma ^{-2}_{x_1|y_1,x_0}=\sigma _{x_1|x_0}^{-2}+\sigma _{y_1|x_1}^{-2}\).
Appendix 4: Growth Rate of the Gaussian Model
Equation (30) for the growth rate \(\Lambda \) of the Gaussian model is obtained by considering
with \(n_t(\phi _t)\) of the form \(n_t(\phi _t)=G_{\sigma ^2_t}(\phi _t-m_t)\), which leads to
The variance \(\sigma ^2_t\) has a fixed point \(\sigma ^2_\infty \) in terms of which the growth rate can be rewritten as
where \(\Lambda ^*=\mathbb {E}_X[\ln K(X)]\),
and
Given that \(x_{t+1}=ax_t+b_t\) and \(y_{t+1}=x_{t+1}+b'_{t+1}\) with \(b_t\sim \mathcal {N}(0,\sigma ^2_{x_1|x_0})\) and \(b'_{t+1}\sim \mathcal {N}(0,\sigma ^2_{y_1|x_1})\), we have
Using \(\sum _{k=0}^t\alpha ^{t-k}x_k=\sum _{k=0}^t(\alpha ^{t-k}-a^{t-k})/(\alpha -a)b_k\), we obtain
and, since the \(b_k\) and \(b'_k\) are all independent, with variances \(\mathbb {E}[b_k^2]=\sigma ^2_{x_1|x_0}\) and \(\mathbb {E}[b_k'^2]=\sigma ^2_{y_1|x_1}\),
Plugged into Eq. (72), it leads to Eq. (30).
Appendix 5: Decomposition of the Growth Rate of the Gaussian Model
Since the Gaussian model can be obtained as a continuous limit of the discrete model, Eqs. (32)–(37) directly result from Eqs. (14)–(22) by taking the same limit. The decomposition can also be derived directly from the general formula of Eq. (30) as we illustrate it here in the simplest case where the two limits are taken.
The first limit, of perfect selectivity, corresponds to \(\sigma ^2_s\rightarrow 0\), such that Eq. (30) becomes
The second limit, of no inheritance, simply corresponds to setting \(\lambda = 0\) in this equation, so that
where \(\sigma ^2_{x_1}=\sigma ^2_{x_1|x_0}/(1-a^2)\) represents the stationary variance of the environmental process, \(\sigma ^2_{x_1}=\mathbb {E}[X_1^2]\). The optimal strategy \(\hat{\pi }\) is obtained by optimizing \(\Lambda \) over \(\kappa \) and \(\sigma ^2_\pi \), which leads to
\(\hat{\kappa }\) can also be written \(\hat{\kappa }=\sigma ^2_{x_1}/\sigma ^2_{y_1}\) where \(\sigma ^2_{y_1}=\sigma ^2_{x_1}+\sigma ^2_{y_1|x_1}\) represents the stationary variance of \(y_t\). As expected from the analysis of the discrete model, we verify that the optimal strategy implements a Bayesian estimation, i.e., \(\hat{\pi }=P_{X|Y}\) [see Appendix 3]. We also verify that the optimal optimal growth rate,
is equivalently written
where
More generally, by introducing \(\sigma _{x_1|y_1}^{-2}=\sigma _{y_1|x_1}^{-2}+\sigma _{x_1}^{-2}\), so that \((1-\kappa )^2\sigma ^2_{x_1}+\kappa ^2\sigma ^2_{y_1|x_1}=\sigma ^2_{x_1|y_1}+(\hat{\kappa }-\kappa )^2\sigma ^2_{y_1}\), we verify that Eq. (80) is equivalent to
as expected from Eq. (13).
Appendix 6: Gaussian Model with Individual Sensors
Using the formulae of Appendix 3, the term to maximize in Eq. (42) can be written
When \(Y\rightarrow X\), \(\kappa _0\rightarrow 1\), \(\sigma ^2_y\rightarrow 0\), \(\sigma ^2_{y|x}\rightarrow 0\) and \(\sigma ^2_{x|y}\rightarrow 0\) but \(\sigma ^2_{x|y}/\sigma ^2_{y|x}\rightarrow 1\) and it simplifies to
The maximum over \(\sigma ^2_\pi \) is reached for \(\sigma ^2_\pi =0\) and taking the derivative with respect to \(\kappa \) leads to
whose solution is given in Eq. (43).
Appendix 7: The Gaussian Model as a Limit of the General Model of Ref. [17]
A general model is defined by Eqs. [S1]–[S2] in the Supporting Information of [17], which we repeat here with only slightly modified notations:
Without loss of generality it can be assumed that \(\sigma ^2_s=1\). The formula for the growth rate of this general model is given with an error in Eq. [S3] of [17]. The correct formula is
where \(\sigma ^2_x\equiv \sigma ^2_{x_1|x_0}/(1-a^2)\), where \(\eta \) and \(\upsilon \) given by
and
with
These formulae reduce to Eq. (30) when taking \(\theta _0=\lambda \), \(\rho _0=\kappa \), \(\omega _0=1\), \(\sigma ^2_D=\sigma ^2_\pi \), \(\lambda _0=\kappa _0=\sigma ^2_H=\sigma ^2_{z|x}=0\) and \(r_\mathrm{max}=\ln K-(1/2)\ln (2\pi \sigma ^2_s)\).
Appendix 8: Feedback Control Out of Equilibrium
The state \(x_t\) of a system in contact with a heat bath is measured as \(y_t\) at regular intervals of time \(\tau \), upon which the potential in which the system evolves is changed from \(V_{t-1}(x)\) to \(V_t(x)\). This change is done without knowing the current state \(x_t\), but may depend on the history of past measurements \(y^t=(y_1,\dots ,y_t)\) as well as on the history of past states at the time of these measurements, \(x^{t-1}=(x_1,\dots ,x_{t-1})\). If we assume that the potential is controllable by one or several parameters \(\ell \), we therefore consider, in the more general case, that \(\ell _t=\ell (y^t,x^{t-1})\) [in more constrained cases, \(\ell _t\) may depend only on some of variables, e.g., \(\ell _t=\ell (y^t)\) when only the present and past measurements are available]. In-between two measurements, the system relaxes in a constant potential \(V_t(x)\) but may not reach equilibrium; its dynamics is generally stochastic, due to the interaction with the heat bath, and may for instance be described by a Master equation with rates satisfying detailed balance. When changing the potential from \(V_{t-1}(x)\) to \(V_t(x)\), a demon extracts a work \(\mathcal {W}_t=V_{t-1}(x_t)-V_{t}(x_t)\). The goal of the demon is either to optimize the total extracted work \(\mathcal {W}_\mathrm{tot}=\mathbb {E}[\sum _t\mathcal {W}_t]\) or, if \(\tau \) itself is controllable, to optimize the power \(\mathcal {W}_\mathrm{tot}/\tau \).
To formalize this problem, we denote by \(p^{\tau }_{t-1}(x_t)\) the probability of the system to be in state \(x_t\) at the time of the t-th measurement: this probability depends explicitly only on \(\ell _{t-1}\) and \(x_{t-1}\), which characterize, respectively, the potential \(V_{t-1}(x)\) and the state of the system when this potential is switched on. Introducing \(F_t=-\beta ^{-1}\ln \sum _x e^{-\beta V_t(x)}\) and \(p^{(\infty )}_t(x)=e^{\beta [F_t-V_t(x)]}\) (also denoted \(\pi \) in the main text), the extracted work may be decomposed as
We now consider the past history \((x^{t-1},y^{t-1})\) as given and average over \((X_t,Y_t)\) to define
Since \(P_{X_t|X^{t-1},Y^{t-1}}(x_t|x^{t-1},y^{t-1})=p^{\tau }_{t-1}(x_t)\), we have
where
and
The total work \(\mathcal {W}_\mathrm{tot}\) is obtained as \(\mathcal {W}_\mathrm{tot}=\sum _t\mathbb {E}_{X^{t-1},Y^{t-1}}[\mathbb {E}_t[\mathcal {W}_t]]\). When \(\tau \rightarrow \infty \), the third term on the right-hand side of Eq. (101) vanishes and we recover the equilibrium result, Eq. (51).
This formalism can be applied to a Brownian particle in a controllable harmonic potential. For simplicity, we assume that only the location of the potential can be controlled, and its stiffness k is fixed to \(k=1\). We also set \(\beta =1\). The potential \(V_t(x)=(x-\ell _t)^2/2\) is characterized by the location \(\ell _t\) of its minimum, and \(F_t=F_{t-1}\) for all t. We take the relaxation dynamics between measurements to be described by a Fokker-Planck equation,
with the initial condition \(p_t^0(x)=\delta (x-x_t)\). This equation is easily solved as its solution is Gaussian at all time: \(p^\tau _t(x)=G_{\varsigma ^2_\tau }(x-\mu _t^\tau )\) with
so that
When \(\tau \rightarrow \infty \), \(p^\tau _t(x)\) converges to the equilibrium distribution \(p_t^\infty (x)=G_{1}(x-\ell _t)\). Using \(P_{Y|X}(y_t|x_t)=G_{\sigma ^2_{y|x}}(y_t-x_t)\) and applying Eq. (66), \(q_t^\tau \) is found to be
The first term in Eq. (101) is therefore
The second term is
with
The third term is
with
Given \((x^{t-1},y^{t-1})\), the only term depending on \(y_t\) is \(\mathbb {E}[z_t^2]\) in Eq. (111). It is optimized by choosing \(\ell _t\) so as to have \(z_t=0\):
By taking \(\ell _{t-1}=\hat{\ell }_{t-1}\), this defines recursively a series of optimal translations \(\hat{\ell }^t\).
To express the optimal work, it remains to evaluate \(\mathbb {E}[z_t'^2]\) for \(\ell ^t=\hat{\ell }^t\). Since \(x_t-\hat{\ell }_t=(1-\kappa )(x_t-\mu _{t-1}^\tau )-\kappa (y_t-x_t)\) where \(x_t-\mu _{t-1}^\tau \) and \(y_t-x_t\) are statistically independent, we have
and therefore \(\mathbb {E}[z_t'^2]=e^{-2\tau }\sigma ^2_{x|y}\). All together, we obtain
which, given that \(\varsigma ^2_\tau =1-e^{-2\tau }\) and \(\sigma ^2_{x|y}=(\varsigma ^{-2}_\tau +\sigma ^{-2}_{y|x})^{-1}\), simplifies to \(\max _{\ell ^t}\mathbb {E}[\mathcal {W}_t]=\varsigma ^2_\tau (1-\sigma ^2_{x|y})/2\), or, in terms of \(\tau \) and \(\sigma ^2_{y|x}\) only,
When \(\tau \rightarrow \infty \), we recover the equilibrium result, \(\mathbb {E}[\mathcal {W}_t]\le I(X;Y)-\min _\phi \mathbb {E}_Y[D(P_{X|Y}(.-Y)\Vert G_1(.-\phi (Y)))]\), with \(I(X;Y)=[\ln (1+1/\sigma ^2_{y|x})]/2\) and \(\min _\phi \mathbb {E}_Y[D(P_{X|Y}(.-Y)\Vert G_1(.-\phi (Y)))]=D(G_{\sigma ^2_{x|y}}\Vert G_1)=(\sigma ^2_{x|y}-\ln \sigma ^2_{x|y}-1)/2\).
Rights and permissions
About this article
Cite this article
Rivoire, O. Informations in Models of Evolutionary Dynamics. J Stat Phys 162, 1324–1352 (2016). https://doi.org/10.1007/s10955-015-1381-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10955-015-1381-z