# Newtonian Dynamics from the Principle of Maximum Caliber

- First Online:

- Received:
- Accepted:

DOI: 10.1007/s10701-014-9819-8

- Cite this article as:
- González, D., Davis, S. & Gutiérrez, G. Found Phys (2014) 44: 923. doi:10.1007/s10701-014-9819-8

- 7 Citations
- 214 Downloads

## Abstract

The foundations of Statistical Mechanics can be recovered almost in their entirety from the principle of maximum entropy. In this work we show that its non-equilibrium generalization, the principle of maximum caliber (Jaynes, Phys Rev 106:620–630, 1957), when applied to the unknown trajectory followed by a particle, leads to Newton’s second law under two quite intuitive assumptions (both the expected square displacement in one step and the spatial probability distribution of the particle are known at all times). Our derivation explicitly highlights the role of mass as an emergent measure of the fluctuations in velocity (inertia) and the origin of potential energy as a manifestation of spatial correlations. According to our findings, the application of Newton’s equations is not limited to mechanical systems, and therefore could be used in modelling ecological, financial and biological systems, among others.

### Keywords

Newtonian mechanics Maximum caliber## 1 Introduction

Newtonian mechanics is a cornerstone of our civilization. From its simpler formulation, \(F=ma\), to Lagrangian and Hamiltonian mechanics it has evolved an elegant mathematical apparatus around it. However, we still have to assume the principle of minimum action as a postulate (that is, if we do not consider Newtonian mechanics as a limiting consequence of quantum mechanics).

There have been attempts to derive Newtonian dynamics from different principles. Particularly related to this work, it has been previously derived from information-geometric arguments leading to the idea of entropic dynamics [1]. This idea is based on the assumption of an irreducible uncertainty in the position of a particle, implying an information metric for space from which Newton’s second law naturally emerges. Caticha and Cafaro’s derivation is founded on Jaynes’ maximum entropy principle [2] (MaxEnt for short), a generic rule for the construction of probabilistic models.

In 1957, Edwin T. Jaynes [2] postulated that Statistical Mechanics has to be understood, not as a physical theory in the same footing as, say, classical mechanics or electromagnetism, but as an application of statistical inference to a system with macroscopically large numbers of degrees of freedom. The question was reversed from “given the microscopic evolution of the system, what is the probability distribution for the macroscopic quantities?” to “given a few known macroscopic properties, what are the possible microstates compatible with said knowledge?”. The answer, as initially proposed by Gibbs, was the probability distribution with maximum entropy \(S=-\sum _i P_i\ln P_i\) subjected to constraints reflecting the known macroscopic properties. Jaynes, after the work of Shannon in information theory, realized that this procedure (maximization of \(S\) constrained only by the known information) is not limited to Statistical Mechanics but a valid principle in any problem of statistical inference. This is the principle of Maximum Entropy. Due to the uniqueness of Shannon’s entropy in characterizing uncertainty, it is the most unbiased procedure for the construction of statistical models. Later, it has been axiomatically derived [3, 4] from requirements of internal consistency, so its validity is independent of any meaning assigned to the quantity \(S\).

A full understanding of the maximum entropy principle requires the adoption of the Bayesian viewpoint on probability: An “absolute” probability distribution \(P(x)\) does not exist, only conditional probability distributions \(P(x|H)\) where \(H\) represents a given state of knowledge. For a given \(H\), MaxEnt gives us the most unbiased model we can propose with our (possibly limited) information, therefore we cannot guarantee beforehand that it is the correct model. As we acquire more and more information, \(P(x|H)\) becomes increasingly more independent of \(H\), and for all practical purposes the distribution becomes objective (independent of the observer).

The principle of Maximum Caliber [5, 6] generalizes this idea to dynamical systems, including time explicitly. It has been applied recently to discrete dynamics [7] and earlier to derive the Fokker–Planck equations [8] and the Markov process formalism [9].

In this work we show that if we use the maximum caliber principle to find the unknown trajectory of a particle, there are two general conditions that automatically lead to Newton’s second law, namely that (a) the expected square displacement per step is known at all times, and (b) that the time-independent probability of finding the particle at any coordinate is also known. Knowledge of both (a) and (b) leads to Newton’s second law in expectation over trajectories, and what is perhaps more interesting, any dynamical system not following Newton’s second law has to violate at least one of these assumptions.

## 2 The Maximum Entropy and Maximum Caliber Formalism

*Bayesian state of knowledge*, that we will denote by \(H\). Maximization of the Shannon entropy

*emerge*naturally from simpler constraints in a Maximum Caliber problem?

## 3 Derivation of Newton’s Second Law

Consider a single particle following an unknown trajectory \(x(t)\) in one spatial dimension. This can be easily generalized to many particles in arbitrary dimensions, at the cost of overcomplicated notation.

If we discretize time, such that the continuous variable \(t\) is sampled at intervals \(\Delta t\) and becomes \(t_k=k\Delta t\) (with \(k=0,\ldots ,n-1\)), the positions \(x(t)\) have to be replaced with \(x_k=x(t_k)\). Now the trajectory \(x(t)\) itself becomes a vector \(\vec {x}=(x_0, \ldots , x_{n-1})\), and the Lagrange function \(\lambda (t)\) also becomes a vector \(\vec \lambda =(\lambda _0, \ldots , \lambda _{n-1})\). This is the method of finite differences [12]. We can see that, using this discretized version, we recover Eqs. 2 and 3, therefore we can also employ the identity given by Eq. 4 for this kind of discretized Maximum Caliber problem.

Here we must remark that, as we are considering the Bayesian view of probability, the last constraint can always be fulfilled: \(P(x_i=X|H)\) is a model comprising our knowledge of \(x_i\)*under the state H*, not a physical property of the phenomenon. Different states of information \(H\) may assign different probability distributions \(P(x_i=X|H)\), and the “quality” of this information will be reflected in the resulting potential energy function emerging from this formalism (as will be seen later).

In appendix A we explore the validity of some aspects of the canonical formalism, namely the Poisson bracket, for the expectation over trajectories.

## 4 Concluding Remarks

We have found that two simple constraints are sufficient to recover Newton’s second law in expectation for the probable trajectories of a particle. The first constraint, on the step size as a function of time, leads to the existence of an inertial mass \(m(t)\) proportional to the Lagrange multiplier \(\lambda (t)\). To understand the meaning of this, remember that for any variational problem solved using Lagrange multipliers, the larger the value of the multiplier, the more restrictive (and therefore more relevant) the constraint. An irrelevant constraint has always a vanishing multiplier. As Jaynes [13] (p. 945) clearly states, “The Lagrange multipliers \(\lambda _k\) in the MAXENT formalism have therefore a deep meaning: \(\lambda _k\) is the ’potential’ of the datum \(R'_k\), that measures how important a constraint it represents.”

Now we motivate the following principle: constraints related to conserved quantities are always more relevant. For instance, this explains the fact that the canonical ensemble in equilibrium statistical mechanics is correctly derived just from a single constraint, the energy or expectation of the Hamiltonian, which is an integral of motion. Another illustration is the following: suppose we are trying to recover the trajectory of a particle from information about the distance to a particular point. If this distance is a constant, this is enough to isolate a unique trajectory, the circle. If we only know that the distance varies between \(r_1\) and \(r_2\), the number of compatible trajectories will increase with \(\Delta r=r_2-r_1\), thus the strength of the constraint will correspondingly decrease with increasing \(\Delta r\).

Given the earlier discussion, the closer \(d_i^2\) is to be a conserved quantity, the more relevant the first constraint is. In this case, \(\lambda (t)\) is large and therefore, \(m(t)\) is also large. Conversely, if the value of \(m\) is small, this means \(\lambda (t)\) is small and therefore \(d_i^2\) has larger fluctuations. In the continuous limit it is the instantaneous speed that fluctuates (there is a non-zero acceleration). This embodies the idea of inertia, and is reminiscent of the ideas of Smolin [14] and of Nelson [15] about inertia being inversely proportional to the size of quantum fluctuations.

## Acknowledgments

GG and SD thank Jorge Zanelli for useful conversations at the beginning of this work. DG gratefully acknowledges the access to resources provided by Grupo de Nano Materiales (Departamento de Física, Facultad de Ciencias, Universidad de Chile). SD acknowledges funding from FONDECYT 1140514.