14

It is known that for a system at thermal equilibrium described by the canonical ensemble, the probability of being in a state of energy $E$ at temperature $T$ is given by the Boltzmann distribution: $$ \tag{1} P(\text{state with energy } E) \propto e^{-E/kT}. $$

I have no problems with the mathematical derivation of this result, but I do wonder if there is a way to understand what (1) means intuitively.

In other words: Is there a way to "guess" that the correct distribution for such a system is given by (1), using solely physical arguments?

Qmechanic
  • 220,844
glS
  • 15,488
  • 5
  • 42
  • 114

4 Answers4

9

What is most intuitive to me is to look at what we're asking. What is the probability that we find the system in a state with total energy $\hat{E}$? It is just the fraction of all possible states that have total energy $\hat{E}$, i.e.

$$p(E) \equiv \frac{\int \delta\left(E(\Omega)- \hat{E}\right)d\Omega}{\int d\Omega}\sim N_\hat{E}$$

But $N_\hat{E}$ is just $$e^{\ln N_\hat{E}}\equiv e^{S_S(\hat{E})/k_B}$$ where $e^{S_S(\hat{E})}$ is the number of states in which our system has total energy $\hat{E}$, which is the same as the number of states in which the rest of the universe has dropped to the energy $E_{tot}-\hat{E}$ by giving $\hat{E}$ to our system, i.e. $S_S(\hat{E}) = S_U(E-\hat{E}) \approx S_U(E) - \hat{E}\frac{\partial S_U(E)}{\partial E}$.

To bring it together, I would do the usual demonstration that at equilibrium, it must be the case that in a simple system that exchanges heat $\frac{\partial S}{\partial E}$ must be the same value in every zone of the system.

Furthermore, if one zone of the system has a smaller value of this than the others, the system will tend to bring more energy to that zone.

Therefore $1/\frac{\partial S}{\partial E}$ is naturally identified with our notion of temperature, i.e. $1/\frac{\partial S}{\partial E} \equiv T$.

Then we clearly have $$p(\hat{E})\sim N_\hat{E} \equiv e^{S_S(\hat{E})/k_B} \propto e^{-\hat{E}/k_BT}$$ This always makes it clear to me that we're just taking sub-volumes of the total set of microstates and asking how big it is next to the collection of all microstates.

Also, picking the sub-volumes out with the $\delta$-function naturally generalizes to all other scenarios, that are usually classified as different ensembles. Here it is just a different constraint relation inside the $\delta$-picker.

If there's another thing we want to constrain, maybe the number of particles, then the $\delta$ installs a factor of $\frac{\partial{S}}{\partial N}$ upstairs, along with the $\frac{\partial{S}}{\partial E}$.

This view has been the most useful to me as it unifies all the descriptions, and makes explicit what is actually being done.

2

Ok. Let's put in this way:

  1. We know the average energy: $\langle E\rangle=\sum P_k E_k $
  2. $P_K$ is a probability distribution: $\sum P_k=1$
  3. We don't know nothing more about the system.

The question is, what probability distribution $P_k$ capture the previous assertions in an unbiased way? The answer is, the distribution that maximize the Shannon's entropy $$ H(P_1...P_k)=-k_b\sum P_kln(P_k) $$ and respect the constraints (1) and (2).

To do this we use the Lagrange multipliers. The multiplier correspondent to the constraint (1) is the temperature. The constraint (2) gave us the partition function.

The boltzmann constant have the task of coupling the entropy to temperature. We use a euler basis for the logarithms and the exponentials. If we change that we need to change the boltzmann constant.

Here, temperature appear as average energy tax associated to an variation of entropy. If the maximum of $H(P_1...P_k)$ for some $\langle E \rangle$ is $S$ , then $$ d \langle E\rangle=TdS $$

Note that we can make entropy an adimensional quantity, and uses the Boltzmann constant as conversor of kelvin to joules.

How we can interpret the result?

We have the probability $P(E)=e^{-E/kT}$. This mean that

$$\frac{k_bT}{P(E)}\frac{dP(E)}{dE}= -1$$.

This equation tells us that for each temperature we have a fixed scale of energy. We can see this by doing $k_bT=1$. In the scale defined by the temperature we can see that the probability of energy $E$ have a maximum in $E=0$ (This zero energy is defined by the normalization,i.e. partition function of the system). We can think that this exponential represent the thermal fluctuation of the energy in some zero energy defined by the system.

Nogueira
  • 5,892
2

I find this the simplest argument. It does not explicitly involve entropy.

  • the multiplicity of two systems is the product of their multiplicities
  • in thermodynamic equilibrium the multiplicity is at its maximum
  • the fractional change of multiplicity with internal energies ($\beta = \frac{1}{\Omega} \frac{d\Omega}{dE}$) is equal for both systems.
  • identify $\beta = 1/kT$

So far this is standard. I suspect what follows is a bit too dubious, because I have not seen it anywhere else.

  • regard a small system with non-degenerate energy levels exchanging energy with a large reservoir
  • write the expression above as a differential equation for $\Omega(E)$ and adjust the sign for the energy of the small system : $\frac{d\Omega}{dE} = -\beta \Omega$
  • solution for the multiplicity of the combined system in terms of large-system $\beta$ and the small-system energy $E$: $\Omega \propto e^{-\beta E}$

I can see problems with this derivation, but I do not see why it would be so much worse than many other arguments in statistical physics.

A numerical example: systems not too far from equilibrium with our environment have a $\beta$ of about 4 % per meV. This value can be computed from the barometric formula. Now contemplate a harmonic oscillator with $\hbar \omega = 1$ meV with a reservoir at $\beta = 0.04$/meV. So for the $n$th excited state, the reservoir will have its multiplicity reduced by a factor of $0.96^n$, the negative exponential which gives the probability of the Boltzmann factor.

1

When deriving a result (whether physically or mathematically), it can be helpful to first find out the limits of the result. Then if your alleged explanation "explains" the result in regions where it doesn't hold, you know your explanation is faulty.

So, the Boltzmann factor for the canonical ensemble. We have to assume that the particle number isn't changing, and we have to assume the volume isn't changing. Having both of these is actually quite restrictive. For instance you could have a gas of hydrogen, but to preserve particle number they need to be cool enough to avoid any possibility of fusion (even through tunnelling, which in the sun is how it happens since even in the sun it is cold relative to fusion being classically achievable just from KE overcoming PE). But to preserve volume you have to have room for all your hydrogen, which means they all have to be near the ground state, because Rydberg atoms (highly excited but not quite ionized hydrogen) can get very very very large if the principle quantum number is insanely large. If you want this to never ever ever happen you'd have to have so little energy available that even if every atom was low in kinetic energy, that there isn't enough energy left over to use as internal energy within a hydrogen to ionize even one single hydrogen. So technically we'd only expect Boltzmann to hold for very very cold hydrogen.

So looking at $TdS=dU+PdV-\mu dN$, we can see that in those situations, $dV=0$ because no change in volume and $dN=0$ because no change in particle number, so $dS=dU/T$, now we are most of the way there.

So since we don't expect it to hold exactly, the real question is why does it work well at all. It must be that the effects of the violations are small, transitory, or their net effects cancel. So for instance if the hydrogen atoms only overlap a little bit (in space and time) we might be able to ignore it. If there is an equilibrium amount of free electrons, ionized hydrogen, and neutral hydrogen then we might be able to ignore some of them if the effects are small. If we have a population of hydrogen at different states, and even though some are large on average there is plenty of room, we can get that.

One way to see this physically is to imagine a whole bunch of roughly identical regions each with some gas, enough that there are a very large number of hydrogen atoms in the many regions all put together. Then we can try to find an equilibrium for how many electrons, ions, and hydrogen, how many are large on average. Then we can try to take those typical values and see how realistic it is to ignore certain abilities.

In the end, taking no change in particle number, no change in volume, and a relationship between entropy and probability leads to a Boltzmann factor. So if you want to see it physically, focus on the relationship between entropy and internal energy in the absence of change in volume or particle number.

Timaeus
  • 26,055
  • 1
  • 36
  • 71