Statistical Mechanics

Ensembles, partition functions, quantum statistics, phase transitions, critical phenomena, and the renormalization group.


Statistical mechanics is the bridge between the microscopic world of atoms and molecules and the macroscopic world of thermodynamics. Where thermodynamics treats energy, entropy, and temperature as given macroscopic quantities, statistical mechanics derives them from first principles — starting with the laws governing individual particles and using probability theory to predict the collective behavior of 102310^{23} of them. Founded by Ludwig Boltzmann and Josiah Willard Gibbs in the late nineteenth century, statistical mechanics provides the microscopic foundation for all of thermodynamics and extends far beyond it, explaining phenomena from the heat capacity of solids to the critical behavior of magnets.

Ensembles and the Boltzmann Distribution

The fundamental idea of statistical mechanics is to replace the impossible task of tracking every particle with a probability distribution over microstates — the complete specification of all microscopic degrees of freedom. An ensemble is a conceptual collection of copies of the system, each in a different microstate but all sharing the same macroscopic constraints.

The microcanonical ensemble describes an isolated system with fixed energy EE, volume VV, and particle number NN. The fundamental postulate (the equal a priori probability hypothesis) states that all accessible microstates with energy between EE and E+δEE + \delta E are equally probable. The entropy is then:

S(E,V,N)=kBlnΩ(E,V,N)S(E, V, N) = k_B \ln \Omega(E, V, N)

where Ω\Omega is the number of such microstates. Temperature emerges as a derived quantity: 1/T=S/E1/T = \partial S / \partial E. This definition guarantees that when two systems are brought into thermal contact, energy flows from the one with higher TT to the one with lower TT — reproducing the zeroth law of thermodynamics from microscopic assumptions.

The canonical ensemble describes a system in thermal contact with a heat bath at temperature TT, with fixed VV and NN. The probability of finding the system in a microstate with energy EiE_i is given by the Boltzmann distribution:

Pi=eEi/kBTZ,Z=ieEi/kBTP_i = \frac{e^{-E_i / k_B T}}{Z}, \qquad Z = \sum_i e^{-E_i / k_B T}

The normalization factor ZZ is the partition function, and it is the master key of statistical mechanics. All thermodynamic quantities follow from ZZ: the Helmholtz free energy is F=kBTlnZF = -k_B T \ln Z, the average energy is E=lnZ/β\langle E \rangle = -\partial \ln Z / \partial \beta where β=1/kBT\beta = 1/k_BT, and the entropy is S=F/TS = -\partial F / \partial T. The partition function factorizes for independent subsystems — Ztotal=Z1Z2Z_{\text{total}} = Z_1 Z_2 — making it tractable for ideal gases and weakly interacting systems.

The grand canonical ensemble further allows particle exchange with a reservoir at chemical potential μ\mu. Its grand partition function Z=N=0eβμNZN\mathcal{Z} = \sum_{N=0}^{\infty} e^{\beta \mu N} Z_N controls the pressure via PV=kBTlnZPV = k_B T \ln \mathcal{Z} and is essential for treating systems where particle number fluctuates, such as adsorbed gases, quantum gases, and chemical reactions.

The Partition Function in Action

The power of the partition function is best seen through examples. For a classical ideal gas of NN indistinguishable particles in volume VV, the canonical partition function is:

Z=1N!(Vλ3)N,λ=2π2mkBTZ = \frac{1}{N!}\left(\frac{V}{\lambda^3}\right)^N, \qquad \lambda = \sqrt{\frac{2\pi\hbar^2}{mk_BT}}

where λ\lambda is the thermal de Broglie wavelength and the 1/N!1/N! factor resolves the Gibbs paradox — the observation that entropy must be extensive, requiring identical particles to be treated as indistinguishable even in classical mechanics. From this partition function, one recovers the ideal gas law PV=NkBTPV = Nk_BT, the energy E=32NkBTE = \frac{3}{2}Nk_BT, and the Sackur-Tetrode equation for the entropy.

The equipartition theorem states that each quadratic term in the energy contributes 12kBT\frac{1}{2}k_BT to the average energy. For a monatomic ideal gas, three translational degrees of freedom give E=32NkBTE = \frac{3}{2}Nk_BT; a diatomic gas adds two rotational degrees of freedom to give E=52NkBTE = \frac{5}{2}Nk_BT at room temperature. But equipartition fails at low temperatures, where quantum effects freeze out vibrational and rotational modes — a failure that was one of the earliest hints of quantum mechanics.

For a harmonic solid (the Einstein model), each atom oscillates independently at frequency ω\omega. The partition function for a single oscillator is Z1=[2sinh(ω/2kBT)]1Z_1 = [2\sinh(\hbar\omega / 2k_BT)]^{-1}, giving a heat capacity that vanishes exponentially at low TT — in dramatic contrast to the classical prediction of 3NkB3Nk_B (the Dulong-Petit law). The Debye model improves on Einstein’s by treating the solid as a collection of coupled oscillators with a realistic density of states g(ω)ω2g(\omega) \propto \omega^2 up to a cutoff frequency ωD\omega_D. Debye’s result, CVT3C_V \propto T^3 at low temperatures, matches experiment beautifully and was confirmed shortly after its prediction in 1912.

Quantum Statistics: Fermions and Bosons

At sufficiently low temperatures or high densities, the quantum nature of particles becomes essential. Identical particles fall into two classes: fermions (half-integer spin, obeying the Pauli exclusion principle) and bosons (integer spin, with no occupancy restriction). Their statistics are fundamentally different.

For fermions, the average occupation of a single-particle state with energy ϵ\epsilon is given by the Fermi-Dirac distribution:

nϵ=1e(ϵμ)/kBT+1\langle n_\epsilon \rangle = \frac{1}{e^{(\epsilon - \mu)/k_BT} + 1}

At T=0T = 0, all states below the Fermi energy ϵF\epsilon_F are occupied and all above are empty — a sharp step function. The Fermi energy of conduction electrons in metals is typically several electron-volts, corresponding to tens of thousands of kelvin; hence electrons in metals are in a degenerate Fermi gas regime even at room temperature. This explains why metals have much smaller electronic heat capacities than the classical prediction: only electrons within kBT\sim k_BT of ϵF\epsilon_F can be thermally excited, giving CVTC_V \propto T rather than the classical 32NkB\frac{3}{2}Nk_B.

For bosons, the Bose-Einstein distribution is:

nϵ=1e(ϵμ)/kBT1\langle n_\epsilon \rangle = \frac{1}{e^{(\epsilon - \mu)/k_BT} - 1}

Below a critical temperature TcT_c, a macroscopic fraction of bosons condenses into the ground state — Bose-Einstein condensation (BEC). Predicted by Einstein in 1924 based on Bose’s work on photon statistics, BEC was first observed in dilute atomic gases by Eric Cornell and Carl Wieman in 1995. The photon gas (blackbody radiation) and the phonon gas in solids are also bosonic systems; the Planck distribution for blackbody radiation, which launched quantum mechanics in 1900, is a special case of Bose-Einstein statistics with μ=0\mu = 0.

Phase Transitions and Critical Phenomena

Statistical mechanics provides the microscopic theory of phase transitions — abrupt changes in the macroscopic state of matter. The partition function, though a sum of smooth exponentials, can develop singularities in the thermodynamic limit (NN \to \infty, VV \to \infty, N/VN/V fixed), and these singularities correspond to phase transitions.

The Ising model — a lattice of spins σi=±1\sigma_i = \pm 1 interacting via H=JijσiσjH = -J\sum_{\langle ij \rangle} \sigma_i \sigma_j, where J>0J > 0 favors alignment — is the simplest model exhibiting a phase transition. In one dimension, Ernst Ising showed in 1925 that no spontaneous magnetization occurs at any T>0T > 0. But in two dimensions, Lars Onsager found the exact solution in 1944, proving that a sharp transition occurs at Tc=2J/(kBln(1+2))T_c = 2J / (k_B \ln(1 + \sqrt{2})). Below TcT_c, the system spontaneously magnetizes; above TcT_c, thermal fluctuations destroy long-range order. Onsager’s solution was a tour de force that confirmed the statistical-mechanical origin of phase transitions.

Near a continuous phase transition, physical quantities diverge with characteristic critical exponents: the specific heat as CTTcαC \sim |T - T_c|^{-\alpha}, the order parameter as mTTcβm \sim |T - T_c|^{\beta}, the susceptibility as χTTcγ\chi \sim |T - T_c|^{-\gamma}, and the correlation length as ξTTcν\xi \sim |T - T_c|^{-\nu}. The remarkable discovery of universality is that these exponents depend only on a few features — the spatial dimensionality dd and the symmetry of the order parameter — and not on microscopic details. Liquid-gas transitions, ferromagnetic transitions, and superfluid transitions in the same universality class share the same exponents.

The Renormalization Group

The explanation of universality came from the renormalization group (RG), developed by Kenneth Wilson in the early 1970s (Nobel Prize, 1982). The core idea is coarse-graining: systematically integrating out short-wavelength fluctuations and rescaling the remaining degrees of freedom. Under this transformation, the effective Hamiltonian flows through a space of couplings. A fixed point of the RG flow corresponds to a scale-invariant system — precisely the behavior at a critical point.

Near the fixed point, perturbations are classified as relevant (growing under RG flow, driving the system away from criticality), irrelevant (shrinking, and therefore unobservable at long distances), or marginal (requiring higher-order analysis). The critical exponents are determined entirely by the relevant perturbations at the fixed point. Since many different microscopic Hamiltonians flow to the same fixed point, they share the same critical behavior — this is the origin of universality.

The RG approach is not limited to equilibrium phase transitions. It provides a unified framework for understanding phenomena across physics: quantum field theory uses renormalization to handle ultraviolet divergences (the RG was, in fact, borrowed from particle physics), polymer physics uses RG to predict the scaling of polymer sizes, and non-equilibrium systems are increasingly analyzed with RG techniques. Wilson’s work showed that the physics of phase transitions is fundamentally a question about scale and symmetry — a perspective that has transformed theoretical physics.

Non-Equilibrium Statistical Mechanics

Most systems in nature are not in equilibrium. Non-equilibrium statistical mechanics extends the formalism to systems driven by external forces, relaxing toward equilibrium, or maintained in steady states by continuous energy input.

The simplest non-equilibrium framework is the Boltzmann equation for the single-particle distribution function f(r,v,t)f(\mathbf{r}, \mathbf{v}, t):

ft+vf+Fmvf=(ft)coll\frac{\partial f}{\partial t} + \mathbf{v} \cdot \nabla f + \frac{\mathbf{F}}{m} \cdot \nabla_v f = \left(\frac{\partial f}{\partial t}\right)_{\text{coll}}

The left side describes free streaming and external forces; the right side is the collision integral, encoding molecular interactions. Boltzmann’s H-theorem shows that a quantity H=flnfd3vH = \int f \ln f\, d^3v decreases monotonically in time, reaching its minimum at the Maxwell-Boltzmann distribution — providing a microscopic derivation of the approach to equilibrium and the second law.

Linear response theory, developed by Ryogo Kubo in the 1950s, relates the response of a system to a small external perturbation to equilibrium fluctuations via the fluctuation-dissipation theorem. For example, the electrical conductivity of a material is determined by the equilibrium current-current correlation function — the Green-Kubo relation. More recently, fluctuation theorems (Jarzynski equality, Crooks relation) have extended the second law to individual microscopic trajectories, showing that the free energy difference between two states can be extracted from non-equilibrium work measurements — even when the process is far from reversible.