This list contains references from the content that can be linked to their source. For a full set of references and notes please see the PDF or HTML where available.
P. H. Algoet and T. M. Cover (1988) A sandwich proof of the Shannon–McMillan–Breiman theorem. Annals of Probability 16 899–909.
J. Avery (2003) Information theory and evolution, World Scientific.
R. Badii and A. Politi (1997) Complexity. Hierarchical structures and scaling in physics, Cambridge University Press.
R. Balian (2005) Information in statistical physics. Studies in History and Philosophy of Modern Physics 36 323–353.
L. Breiman (1957) The individual ergodic theorem of information theory. Annals of Mathematical Statistics 28 809–811. (Correction: (1957) 31 809–810.)
L. Brillouin (1951a) Maxwell's demon cannot operate: Information and entropy. Journal of Applied Physics 22 334–337.
L. Brillouin (1951b) Physical entropy and information. Journal of Applied Physics 22 338–343.
L. Brillouin (1953) Negentropy principle of information. Journal of Applied Physics 24 1152–1163.
M. Brin and A. Katok (1983) On local entropy. In: J. Palis (ed.) Geometric dynamics. Springer-Verlag Lecture Notes in Mathematics 1007 30–38.
P. Castiglione , M. Falcioni , A. Lesne and A. Vulpiani (2008) Chaos and coarse-graining in statistical mechanics, Cambridge University Press.
C. Cercignani (1988) The Boltzmann equation and its applications, Springer-Verlag.
G. J. Chaitin (1966) On the length of programs for computing finite binary sequences. Journal of the ACM 13 547–569.
E. G. D. Cohen and G. Gallavotti (1999) Note on two theorems of nonequilibrium statistical mechanics. Journal of Statistical Physics 96 1343–1349.
R. T. Cox (1946) Probability, frequency, and reasonable expectation. American Journal of Physics 14 1–13.
I. Csiszár (1975) I-divergence geometry of probability distributions and minimization problems. Annals of Probability 3 146–158.
I. Csiszár (1998) The Method of types. IEEE Transactions on Information Theory 44 2505–2523.
B. Durand and A. Zvonkine (2007) Kolmogorov complexity. In: E. Charpentier , A. Lesne and N. Nikolski (eds.) Kolmogorov's Heritage in Mathematics, Springer-Verlag281–300.
R. S. Ellis (1985) Entropy, large deviations and statistical mechanics, Springer-Verlag.
D. J. Evans and D. J. Searles (2002) The fluctuation theorem. Advances in Physics 51 1529–1585.
M. Falcioni , V. Loreto and A. Vulpiani (2003) Kolmogorov's legacy about entropy, chaos and complexity. In: A. Vulpiani and R. Livi (eds.) The Kolmogorov Legacy in Physics, Springer-Verlag85–108.
D. P. Feldman and J. P. Crutchfield (1998) Measures of statistical complexity: Why? Physics Letters A 238 244–252.
K. Ford (2007) From Kolmogorov's theorem on empirical distribution to number theory. In: E. Charpentier , A. Lesne and N. Nikolski (eds.) Kolmogorov's heritage in mathematics, Springer-Verlag97–108.
S. A. Frank (2009) The common patterns of nature. Journal of Evolutionary Biology 22 1563–1585.
G. Gallavotti (1998) Chaotic dynamics, fluctuations, nonequilibrium ensembles. Chaos 8 384–393.
G. Gallavotti (2006) Entropy, thermostats and the chaotic hypothesis. Chaos 16 043114.
P. Gaspard (2004) Time-reversed dynamical entropy and irreversibility in Markovian random processes. Journal of Statistical Physics 117 599–615.
M. Gell-Mann and S. Lloyd (1996) Information measures, effective complexity, and total information. Complexity 2 44–52.
E. Glasner (2003) Ergodic theory via joinings, American Mathematical Society.
A. N. Gorban (2007) Order-disorder separation: Geometric revision. Physica A 374 85–102.
P. Grassberger (1986) Toward a quantitative theory of self-generated complexity. International Journal of Theoretical Physics 25 907–938.
C. Gruber , S. Pache and A. Lesne (2004) On the second law of thermodynamics and the piston problem. Journal of Statistical Physics 117 739–772.
B. Haegeman and R. S. Etienne (2010) Entropy maximization and the spatial distribution of species. American Naturalist 175 E74–E90.
J. Honerkamp (1998) Statistical physics, Springer-Verlag.
S. Ihara (1993) Information theory for continuous systems, World Scientific.
E. T. Jaynes (1957a) Information theory and statistical mechanics Part I. Physical Review 106 620–630.
E. T. Jaynes (1957b) Information theory and statistical mechanics Part II. Physical Review 108 171–190.
E. T. Jaynes (1973) The well-posed problem. Foundations of Physics 3 477–493.
E. T. Jaynes (1980) The minimum entropy production principle. Annual Review of Physical Chemistry 31 579–601.
E. T. Jaynes (1982) On the rationale of maximum entropy methods. Proceedings of the IEEE 70 939–952.
S. Karlin and H. M. Taylor (1975) A first course in stochastic processes, Academic Press.
W. Krieger (1970) On entropy and generators of measure-preserving transformations. Transactions of the American Mathematical Society 149 453–464.
S. Kullback and R. Leibler (1951) On information and sufficiency. Annals of Mathematical Statistics 22 79–86.
R. Landauer (1961) Irreversibility and heat generation in the computing process. IBM Journal of Research and Development 5 183–191.
J. L. Lebowitz (1993a) Boltzmann's Entropy and Time's Arrow. Physics Today 46 32–38.
J. L. Lebowitz (1993b) Macroscopic laws, microscopic dynamics, time's arrow and Boltzmann's entropy. Physica A 194 1–27.
F. Ledrappier and J. M. Strelcyn (1982) A proof of the estimation from below in Pesin's entropy formula. Ergodic Theory and Dynamical Systems 2 203–219.
A. Lempel and J. Ziv (1976) On the complexity of finite sequences. IEEE Transactions on Information Theory 22 75–81.
A. Lesne and A. Benecke (2008) Feature context-dependency and complexity reduction in probability landscapes for integrative genomics. Theoretical Biology and Medical Modelling 5 21.
A. Lesne , J. L. Blanc and L. Pezard (2009) Entropy estimation of very short symbolic sequences. Physical Review E 79 046208.
M. Li and P. Vitanyi (1997) An Introduction to Kolmogorov complexity and its applications, Springer.
H. Mahara and T. Yamaguchi (2010) Entropy balance in distributed reversible Gray-Scott model. Physica D 239 729–734.
P. Martin-Löf (1966) The definition of random sequence. Information and Control 9 602–619.
B. McMillan (1953) The basic theorems of information theory. Annals of Mathematical Statistics 24 196–219.
G. Nicolis and P. Gaspard (1994) Toward a probabilistic approach to complex systems. Chaos, Solitons and Fractals 4 41–57.
G. Parisi (2003) Complexity and intelligence. In: A. Vulpiani and R. Livi (eds.) The Kolmogorov Legacy in Physics, Springer-Verlag109–122.
Y. Pesin (1997) Dimension theory in dynamical systems. Contemporary views and applications, University of Chicago Press.
S. J. Phillips and M. Dudík (2008) Modeling of species distributions with Maxent: new extensions and a comprehensive evaluation. Ecography 31 161–175.
S. J. Phillips , R. P. Anderson and R. E. Schapire (2006) Maximum entropy modeling of species geographic distribution. Ecological Modelling 190 231–259.
Z. Rached , F. Alajaji and L. Campbell (2001) Rényi's divergence and entropy rates for finite alphabet Markov sources. IEEE Transactions on Information Theory 47 1553–1562.
C. Robert (1990) An entropy concentration theorem: applications in artificial intelligence and descriptive statistics. Journal of Applied Probability 27 303–313.
D. P. Ruelle (2003) Extending the definition of entropy to nonequilibrium steady states. Proceedings of the National Academy of Sciences of the United States of America 100 3054–3058.
I. Samengo (2002) Estimating probabilities from experimental frequencies. Physical Review E 65 046124.
T. Sagawa and M. Ueda (2009) Minimal energy cost for thermodynamic information processing: measurement and information erasure. Physical Review Letters 102 250602.
L. S. Schulman (2010) We know why coffee cools. Physica E 42 269–272.
C. Shannon (1948) A mathematical theory of communication. Bell System Technical Journal 27 379–423.
J. S. Shinner , M. Davison and J. T. Landsberg (1999) Simple measure for complexity. Physical Review E 59 1459–1464.
A. D. Sokal (1997) Monte Carlo methods in statistical mechanics: Foundations and new algorithms. In: C. C. De Witt-Morette and A. Folacci (eds.) Functional integration: basics and applications (1996 Cargèse summer school), Plenum Press.
A. D. Sokal and L. E. Thomas (1989). Exponential convergence to equilibrium for a class of random-walk models. Journal of Statistical Physics 54 797–828.
R. Solomonoff (1978). Complexity-based induction systems: comparisons and convergence theorems. IEEE Transactions on Information Theory 24 422–432.
L. Szilard (1929) Uber die Entropieverminderung in einem thermodynamischen System bei Eingriffen intelligenter Wesen. (On the lessening of entropy in a thermodynamic system by interference of an intelligent being). Zeitschrift für Physik 53 840–856.
H. Touchette (2009) The large deviation approach to statistical mechanics. Physics Reports 478 1–69.
M. Tribus and E. C. McIrvine (1971) Energy and information. Scientific American 225 179–188.
J. M. Van Campenhout and T. M. Cover (1981) Maximum entropy and conditional entropy. IEEE Transactions on Information Theory 27 483–489.
V. Vovk and G. Shafer (2003) Kolmogorov's contributions to the foundations of probability. Problems of Information Transmission 39 21–31.
H. White (1993) Algorithmic complexity of points in dynamical systems. Ergodic Theory and Dynamical Systems 13 807–830.
A. D. Wyner and J. Ziv (1989) Some asymptotic properties of the entropy of a stationary ergodic data source with applications to data compression. IEEE Transactions on Information Theory 35 1250–1258.
J. Ziv and A. Lempel (1977) A universal algorithm for sequential data compression. IEEE Transactions on Information Theory 23 337–343.
J. Ziv and A. Lempel (1978) Compression of individual sequences by variable rate coding. IEEE Transactions on Information Theory 24 530–536.
O. Zuk , I. Kanter and E. Domany (2005) The entropy of a binary hidden Markov process. Journal of Statistical Physics 121 343–360. (Conference version: Aymptotics of the entropy rate for a hidden Markov process. Proceedings DCC'05 173–182.)