Maximum information entropy principle and the interpretation of probabilities in statistical mechanics − a short review

Topical Review

DOI: 10.1140/epjb/e2016-70175-6

Cite this article as:
Kuić, D. Eur. Phys. J. B (2016) 89: 124. doi:10.1140/epjb/e2016-70175-6


In this paper an alternative approach to statistical mechanics based on the maximum information entropy principle (MaxEnt) is examined, specifically its close relation with the Gibbs method of ensembles. It is shown that the MaxEnt formalism is the logical extension of the Gibbs formalism of equilibrium statistical mechanics that is entirely independent of the frequentist interpretation of probabilities only as factual (i.e. experimentally verifiable) properties of the real world. Furthermore, we show that, consistently with the law of large numbers, the relative frequencies of the ensemble of systems prepared under identical conditions (i.e. identical constraints) actually correspond to the MaxEnt probabilites in the limit of a large number of systems in the ensemble. This result implies that the probabilities in statistical mechanics can be interpreted, independently of the frequency interpretation, on the basis of the maximum information entropy principle.


Statistical and Nonlinear Physics 

Copyright information

© EDP Sciences, SIF, Springer-Verlag Berlin Heidelberg 2016

Authors and Affiliations

  1. 1.University of Split, Faculty of ScienceSplitCroatia

Personalised recommendations