Simple Principle of Maximum Entropy, without Lagrange Multipliers.
6.050J / 2.110J, April 28, 2004.
- The set-up phase:
- Identify the states (assume a finite number of them), Ai
- Identify the energy of each state, Ei
- Identify the expected value of energy E which must be between the lowest
and highest Ei (or possibly equal to one of them)
- Phase of assigning probability distribution p(Ai):
- We want the p(Ai) with the highest uncertainty
S = Σi p(Ai) log2 (1/p(Ai) )
consistent with the right value for energy
E = Σi p(Ai)Ei
- Consider the candidate probability distributions of the form
pc(Ai) = e-α e-βEi
where α is chosen so as to make the sum of all pc(Ai) = 1, i.e.,
α = log2 [Σi e-βEi]
- These candidate distributions are functions of β, a real number
- These candidate distributions have an uncertainty
Sc = Σi pc(Ai) log2 (1/pc(Ai) )
- These candidate distributions have an expected value of energy
Ec = Σi pc(Ai)Ei
which may or may not be the right energy E
- Assertion: If p(Ai) is any other probability distribution with the same
expected value of energy Ec, then its uncertainty
S = Σi p(Ai) log2 (1/p(Ai) ) is no greater than Sc
- Therefore, if β is chosen to make the energy Ec equal to the desired
expected value E, the resulting
candidate distribution is the one which has the maximum entropy
Proof of the assertion:
- The Gibbs inequality (6.4) is:
Σi p(Ai) log2 (1/p(Ai) ) ≤
Σi p(Ai) log2 (1/pc(Ai) )
- Just substitute the form for pc(Ai) in the Gibbs
inequality, and use the fact that
Σi p(Ai)Ei =
Σi pc(Ai)Ei
At this point all the general properties of Chapter 11 can be used for the
candidate distribution. In particular, a value of β can be found for
which the energy matches what is needed, as a sketch of Ec(β) vs.
β reveals.
Click here for information on MIT Accessibility