Maximum entropy principle
A maximum entropy principle is the prescription: given a nonempty set of candidate probability distributions, select a distribution satisfying
where is Shannon entropy for discrete laws (and typically differential entropy for absolutely continuous laws). In applications, is usually defined by constraints such as normalization, support restrictions, or moment/expectation constraints of the form for a random variable and given functions (using expectation ).
The guiding idea is to choose the least-committal distribution consistent with the stated information. In many common settings, maximum-entropy problems can be reformulated in terms of minimizing KL divergence to a reference distribution, with Gibbs' inequality guaranteeing nonnegativity of the objective.
Examples:
- On a finite set of outcomes, if the only constraint is that the distribution is supported on those outcomes, the maximizer of Shannon entropy is the uniform distribution.
- Among all distributions on with fixed mean and variance, the maximizer of differential entropy is the normal distribution with those parameters.