I was going through my lecture notes and I found something I could not quite understand. First, it starts by deriving an expression for the information entropy (as used in physics?):
Let pi be the probability of finding the system in microstate i. With the total number of accessible Microstates N(A), this can be written pi=1N for all microstates compatible with the macrostate A. We can write lnN(A)=−1⋅lnpi=−(∑N(A)i=1pi)lnpi due to the normalization of the pi. [...]
Therefore, we can write for the information entropy of a macrostate A: S(A)=−kBN(A)∑i=1pilnpi
Later, it tries to derive the Boltzmann distribution for the ideal gas:
We will do so by finding the extremum of ϕ=S(A)−λ1(∑ipi−1)−λ2(∑ipiEi−Eavg)
using the method of Lagrange multipliers. With S(A)=−kB∑N(A)i=1pilnpi.
It goes on to find the correct formula.
My question is, why this expression for the entropy S(A) can be used, even though for the second example the pi are obviously not constant and equal to 1N?
Answer
You can actually derive the Gibbs entropy from purely mathematical concerns and the properties of probability. The properties we require entropy to have are:
- Extensivity - the entropy of two independent systems, considered as a whole, should be the sum of the entropies of the individual systems S(A∩B)=S(A)+S(B).
- Continuity - the entropy should be a smoothly differentiable function of the probabilities assigned to each state.
- Minimum - the entropy should be zero if and only if the system is in a single state with probability 1.
- Maximum - the entropy should be maximized when every state is equally probable.
It follows from probability theory that, when A and B are independent, then P(A∩B)=P(A)P(B).
In physics we choose the constant of proportionality to be kB, Boltzmann's constant, and assign it units JK−1, Joules per Kelvin, in order to match Clausius's formula for classical entropy. When all of the pi are equally probable, the formula reduces to the Boltzmann entropy.
You get the classical canonical ensembles and their corresponding distributions when you maximize the entropy of a system that is interacting with a 'bath' in a way that constrains the average value of a parameter (e.g. energy, volume, particle number) without specifying the value that parameter takes. The MB distribution comes, as the questioner saw, when the average energy is constrained but the total energy is allowed to vary; total energy would be fixed by adding a Lagrange multiplier of the form λE(Ei−Etot), producing the microcanonical ensemble.
No comments:
Post a Comment