Thermal state from entropy maximization

Maximum-entropy derivation of the canonical (and generalized Gibbs) distribution from macroscopic constraints.
Thermal state from entropy maximization

A central construction of equilibrium statistical mechanics is: given only partial macroscopic information, choose the least biased microscopic distribution consistent with that information. This principle is implemented by maximizing the .

Problem (choose a distribution from constraints)

Let xx denote a in , and let H(x)H(x) be the . A statistical state is a probability density ρ(x)0\rho(x)\ge 0 normalized with respect to the dΓd\Gamma:

ρ(x)dΓ(x)  =  1. \int \rho(x)\,d\Gamma(x) \;=\; 1.

The (dimensionless) entropy functional is

s[ρ]  =  ρ(x)logρ(x)dΓ(x), s[\rho] \;=\; -\int \rho(x)\,\log\rho(x)\,d\Gamma(x),

and the physical entropy is S[ρ]=kBs[ρ]S[\rho]=k_B\,s[\rho] with kBk_B the .

Assume we only know the mean energy (a macroscopic constraint)

ρ(x)H(x)dΓ(x)  =  U. \int \rho(x)\,H(x)\,d\Gamma(x) \;=\; U.

Solution (canonical Gibbs distribution)

Maximizing s[ρ]s[\rho] subject to normalization and fixed mean energy gives, via Lagrange multipliers, a unique maximizer of the form

ρβ(x)  =  eβH(x)Z(β),Z(β)  =  eβH(x)dΓ(x). \rho_\beta(x) \;=\; \frac{e^{-\beta H(x)}}{Z(\beta)}, \qquad Z(\beta) \;=\; \int e^{-\beta H(x)}\,d\Gamma(x).

The normalizing factor Z(β)Z(\beta) is the , and the resulting state is exactly the . The multiplier β\beta is identified with the .

This provides a variational (information-theoretic) characterization of thermal equilibrium: among all distributions with the same mean energy, ρβ\rho_\beta has maximal entropy.

Generalization (multiple constraints)

If, in addition to the mean energy, one constrains the mean values of other quantities Qi(x)Q_i(x) (often conserved charges),

ρ(x)Qi(x)dΓ(x)  =  qi, \int \rho(x)\,Q_i(x)\,d\Gamma(x) \;=\; q_i,

the same entropy-maximization procedure yields

ρ(x)exp ⁣(λ0iλiQi(x)), \rho(x) \propto \exp\!\Big(-\lambda_0 - \sum_i \lambda_i Q_i(x)\Big),

which is the basis of the .

Interpretation via relative entropy

Entropy maximization can also be phrased as an optimization over information distance: the maximizer is the distribution that is “closest” to a chosen reference measure while satisfying the constraints, using the . The uniqueness/optimality is underwritten by the .

Physical meaning

  • The constraints encode what is known macroscopically (here, the mean energy).
  • The maximizer is the least structured microscopic state compatible with those constraints.
  • When a small system is weakly coupled to a large heat bath, this construction matches the canonical state obtained from .