RemNote Community
Community

Introduction to Statistical Mechanics

Understand how microscopic particle behavior yields macroscopic thermodynamic properties, the role of ensembles and partition functions, and key applications such as ideal gases and phase transitions.
Summary
Read Summary
Flashcards
Save Flashcards
Quiz
Take Quiz

Quick Practice

What does statistical mechanics provide for the laws of thermodynamics?
1 of 24

Summary

Fundamentals of Statistical Mechanics What is Statistical Mechanics? Statistical mechanics is the bridge between the microscopic world of atoms and molecules and the macroscopic world we observe. It answers a fundamental question: How do the properties of matter that we measure in the lab—temperature, pressure, entropy—arise from the behavior of countless individual particles? Classical thermodynamics describes these bulk properties through empirical laws, but it doesn't explain why those laws work. Statistical mechanics fills this gap. By counting the vast numbers of possible particle configurations and using probability, it shows why the laws of thermodynamics must hold. In essence, statistical mechanics provides the microscopic foundation for everything thermodynamics describes phenomenologically. Microstates and Macrostates: Two Views of a System To understand statistical mechanics, you need to grasp the distinction between two complementary descriptions of a system. A microstate is a complete specification of every detail of the system: the exact position and momentum of every single particle. For a macroscopic system with $10^{23}$ particles, describing a microstate requires specifying roughly $10^{24}$ numbers. Clearly, this is impossibly detailed information. A macrostate is what we actually measure and observe: aggregate properties like total energy $E$, volume $V$, temperature $T$, and particle number $N$. A macrostate is coarse-grained—it ignores microscopic details and specifies only a few observable quantities. Here's the crucial insight: A single macrostate corresponds to an enormous number of different microstates. For example, two arrangements of gas molecules in a container might have the same total energy but different individual particle positions—these are different microstates of the same macrostate. Because we cannot know the exact microstate of a real system (and don't need to), statistical mechanics assigns probabilities to different microstates. The fundamental assumption is that, in equilibrium, all microstates consistent with the constraints of the system are equally probable—or more generally, that the probability of a microstate depends in a specific way on its energy and other factors. The key prediction of statistical mechanics: the most probable macrostate is the one we observe. The system naturally evolves toward the macrostate with the greatest number of compatible microstates. Entropy and Statistical Weight The connection between microscopic counting and the macroscopic quantity entropy is one of the triumphs of statistical mechanics. Define the statistical weight $\Omega$ of a macrostate as the number of distinct microstates that produce that macrostate. A macrostate with more accessible microstates has a larger statistical weight. The fundamental relationship is: $$S = kB \ln \Omega$$ where $S$ is the entropy, $kB$ is Boltzmann's constant ($1.38 \times 10^{-23}$ J/K), and $\Omega$ is the statistical weight. This equation shows that entropy is literally a measure of the number of microstates available to the system. A system with more possible microstates has higher entropy. Why does the system increase its entropy? Because states with more microstates are more probable. If you have a million-to-one ratio of microstates, the more probable macrostate is $10^6$ times more likely to occur. At macroscopic scales, this makes the probability of observing the high-entropy state essentially 100%. This gives a microscopic explanation for the second law of thermodynamics: systems naturally evolve toward states with higher entropy because those states have more accessible microstates and are therefore more probable. Ensembles: Accounting for Different Constraints Not all systems have the same constraints. A statistical ensemble is a collection of all possible microstates of a system subject to certain constraints. Different physical situations call for different ensembles, each with its own probability distribution. The Microcanonical Ensemble In the microcanonical ensemble, the system is isolated: its total energy $E$ is fixed and cannot change. No energy can be exchanged with the surroundings. The defining assumption: all microstates with the specified energy are equally probable. This makes intuitive sense—if energy is conserved and we have no other information, we shouldn't prefer one energy-conserving arrangement over another. This ensemble is useful for understanding fundamental principles, but it's rarely encountered in practice (true isolation is hard to achieve). The Canonical Ensemble In the canonical ensemble, the system is in thermal contact with a much larger heat reservoir at temperature $T$. The system can exchange energy with the reservoir, but its volume and particle number are fixed. Because energy can flow to or from the reservoir, individual microstates no longer have equal probability. Instead, lower-energy microstates are more probable than higher-energy ones. The probability of finding the system in a microstate with energy $Ei$ is: $$P(Ei) \propto e^{-Ei/(kB T)}$$ This is the Boltzmann distribution. Notice that at high temperature, the exponent is small (less negative), so the probabilities become more uniform—thermal energy dominates, and the system explores many different energy states. At low temperature, $e^{-Ei/(kB T)}$ strongly suppresses high-energy states, so the system concentrates its probability on low-energy microstates. The canonical ensemble is the most commonly used in statistical mechanics because it mirrors the real world: systems are typically held at constant temperature by their environment. The Grand-Canonical Ensemble In the grand-canonical ensemble, both energy and particle number can fluctuate. The system is in thermal contact with a reservoir at temperature $T$ and can exchange both energy and particles with it. The chemical potential $\mu$ (a measure of how many particles the system "wants" to have) is held constant. The probability of a microstate with energy $Ei$ and particle number $Ni$ is: $$P(Ei, Ni) \propto e^{-(Ei - \mu Ni)/(kB T)}$$ This ensemble is useful when studying systems where the number of particles varies, such as adsorption of gases on surfaces or chemical reactions. Why These Probability Distributions? Each probability distribution reflects the physical constraints on the system: Microcanonical: energy fixed → all compatible microstates equally probable Canonical: temperature fixed → energy varies → $\propto e^{-E/(kB T)}$ Grand-canonical: temperature and chemical potential fixed → $\propto e^{-(E-\mu N)/(kB T)}$ These aren't arbitrary assumptions—they follow from the principle that, given constraints, we should maximize entropy subject to those constraints. This is a theorem: the probability distribution that maximizes entropy under the given constraints is exactly the one listed above for each ensemble. The Partition Function: The Central Object The partition function $Z$ is the most important quantity in statistical mechanics. It's a sum over all microstates that encodes the statistical properties of a system. For the canonical ensemble: $$Z = \sumi e^{-Ei/(kB T)}$$ where the sum runs over all microstates $i$, and $Ei$ is the energy of microstate $i$. The word "partition" refers to how this function partitions the total probability among all microstates. The factor $e^{-Ei/(kB T)}$ is called the Boltzmann factor. The partition function is "magic" because once you know $Z$, you can calculate every thermodynamic quantity of the system through differentiation. This is why statistical mechanics problems often reduce to calculating $Z$. Deriving Thermodynamic Quantities from the Partition Function The power of the partition function lies in its ability to generate thermodynamic quantities through mathematical operations. Here are the key relationships: Internal Energy: $$U = -\frac{\partial \ln Z}{\partial \beta}$$ where $\beta = 1/(kB T)$. Intuitively, the internal energy is the weighted average energy of all microstates, weighted by their Boltzmann factors. Helmholtz Free Energy: $$F = -kB T \ln Z$$ Free energy is the thermodynamic potential that governs which processes are spontaneous at constant temperature and volume. Pressure (Equation of State): $$P = kB T \frac{\partial \ln Z}{\partial V}$$ This shows how pressure arises from the partition function. By calculating this derivative for a specific system (like an ideal gas), we can derive the equation of state. Heat Capacity at Constant Volume: $$CV = \frac{\partial U}{\partial T}$$ This can be expressed as derivatives of $Z$. Heat capacity measures how much the internal energy changes with temperature. The general pattern: thermodynamic quantities = derivatives of the partition function. This is the essence of how statistical mechanics connects microscopic details to macroscopic measurements. Application 1: The Ideal Gas and the Molecular Basis of Temperature One of the clearest applications of statistical mechanics is deriving the ideal gas law from first principles. For an ideal gas, statistical mechanics uses the Maxwell-Boltzmann distribution to count how gas molecules occupy different velocities. The calculation proceeds by: Computing the partition function for non-interacting particles in a box Finding the probability distribution for particle speeds Calculating the average kinetic energy The remarkable result: the average kinetic energy per molecule is $$\langle E{\text{kinetic}} \rangle = \frac{3}{2} kB T$$ This equation reveals what temperature actually is at the microscopic level: temperature is a measure of average molecular kinetic energy. Higher temperature means faster-moving particles. From this average kinetic energy and basic mechanics, the ideal gas law immediately follows: $$PV = NkB T$$ where $P$ is pressure, $V$ is volume, and $N$ is the number of molecules. This is remarkable: we've derived this empirical equation from the statistical behavior of molecules, showing that it's a statistical necessity, not an assumption. Application 2: Phase Transitions Phase transitions—like ice melting or water boiling—are qualitative changes in the macroscopic properties of matter. Statistical mechanics explains these through the lens of competing entropy and energy. As temperature changes, the number of accessible microstates (statistical weight) changes dramatically for different phases. A liquid has vastly more microstates than a solid (molecules can rearrange freely vs. being locked in a lattice). At low temperature, energy dominates, so the system prefers the low-energy solid state. At high temperature, entropy dominates, so the system prefers the high-entropy liquid state. Statistical mechanics analyzes phase transitions by tracking how the partition function changes with temperature and pressure. Discontinuous changes in thermodynamic quantities (like density or entropy) at the transition point emerge naturally from the partition function calculation. <extrainfo> Application 3: Magnetic Ordering Materials like iron become magnetic when cooled below a critical temperature. This spontaneous magnetic order arises from competition between energy (magnetic moments prefer to align) and entropy (misaligned states have more microstates). Statistical models like the Ising model count spin configurations and their energies. Analysis shows that below a critical temperature, aligned spins become statistically dominant—the system "chooses" the ordered, magnetized state. Above this temperature, thermal entropy wins, and the spins are disordered. </extrainfo> Summary: Statistical mechanics reveals that the thermodynamic laws we observe macroscopically emerge inevitably from counting microstates. Entropy is literally the number of accessible microstates. Different ensembles account for different physical constraints. The partition function encodes all thermodynamic information through a single mathematical object. These principles explain everything from why gases follow $PV=NkT$ to why materials undergo phase transitions—not through assumptions, but through the inexorable logic of probability and counting.
Flashcards
What does statistical mechanics provide for the laws of thermodynamics?
Microscopic justification.
How does statistical mechanics differ from classical thermodynamics in explaining laws?
By counting particle configurations rather than describing bulk properties phenomenologically.
What is the definition of a microstate?
A complete specification of the positions and momenta of every particle in a system.
Why are probabilities assigned to different microstates in a real system?
Because the exact microstate cannot be known.
What defines a macrostate in statistical mechanics?
A few observable quantities such as energy, volume, and particle number.
Which macrostate dominates the observed behavior of a system?
The most probable macrostate.
What is the statistical weight of a macrostate?
The number of microstates it contains.
What is the formula relating entropy to statistical weight?
$S = k{B} \ln \Omega$ (where $S$ is entropy, $k{B}$ is Boltzmann’s constant, and $\Omega$ is the number of microstates).
How are average values of physical quantities calculated using probability distributions?
By weighted sums over microstates.
What property is fixed in a microcanonical ensemble?
Total energy.
What probability is assigned to microstates within a microcanonical ensemble?
Equal probability for all microstates with the specified energy.
With what does a system in a canonical ensemble exchange energy?
A heat reservoir at temperature $T$.
What is the probability of a microstate with energy $E{i}$ in a canonical ensemble?
Proportional to $e^{-E{i}/(k{B}T)}$.
What is the formula for the partition function $Z$ in a canonical ensemble?
$Z = \sum{i} e^{-E{i}/(k{B}T)}$.
Which quantities can fluctuate in a grand-canonical ensemble?
Energy and particle number.
What is the probability of a microstate with energy $E{i}$ and particle number $N{i}$ in a grand-canonical ensemble?
Proportional to $e^{-(E{i}-\mu N{i})/(k{B}T)}$ (where $\mu$ is the chemical potential).
What is the general definition of the partition function $Z$?
The sum of the Boltzmann factors over all microstates of a system.
What is the formula for internal energy $U$ in terms of the partition function?
$U = -\frac{\partial \ln Z}{\partial \beta}$ (where $\beta = 1/(k{B}T)$).
What is the formula for Helmholtz free energy $F$ using the partition function?
$F = -k{B}T \ln Z$.
What is the formula for pressure $P$ derived from the partition function?
$P = k{B}T \frac{\partial \ln Z}{\partial V}$ (where $V$ is volume).
What is the formula for heat capacity at constant volume $C{V}$ in thermodynamics?
$C{V} = \frac{\partial U}{\partial T}$.
What is the average kinetic energy of a molecule in an ideal gas according to statistical mechanics?
$\frac{3}{2}k{B}T$.
What is the ideal-gas law formula expressed with Boltzmann's constant?
$PV = N k{B}T$ (where $P$ is pressure, $V$ is volume, and $N$ is the number of molecules).
What specific constraints are reflected by the different probability distributions in statistical ensembles?
Fixed energy Fixed temperature Fixed temperature and chemical potential

Quiz

What is the relationship between a macrostate and microstates?
1 of 14
Key Concepts
Fundamentals of Statistical Mechanics
Statistical mechanics
Microstate
Macrostate
Entropy
Statistical Ensembles
Canonical ensemble
Grand canonical ensemble
Partition function
Applications and Phenomena
Ideal gas
Phase transition
Magnetic ordering