Graduate Statistical Mechanics
Leon Hostetler
December 16, 2018 Version 0.5 Contents
Preface v
1 The Statistical Basis of Thermodynamics1 1.1 Counting States...... 1 1.2 Connecting to Thermodynamics...... 4 1.3 Intensive and Extensive Parameters ...... 7 1.4 Averages...... 8 1.5 Summary: The Statistical Basis of Thermodynamics...... 9
2 Ensembles and Classical Phase Space 13 2.1 Classical Phase Space ...... 13 2.2 Liouville’s Theorem ...... 14 2.3 Microcanonical Ensemble ...... 16 2.4 Canonical Ensemble ...... 20 2.5 Grand Canonical Ensemble ...... 29 2.6 Summary: Ensembles and Classical Phase Space ...... 32
3 Thermodynamic Relations 37 3.1 Thermodynamic Potentials ...... 37 3.2 Fluctuations...... 44 3.3 Thermodynamic Response Functions...... 48 3.4 Minimizing and Maximizing Thermodynamic Potentials...... 49 3.5 Summary: Thermodynamic Relations ...... 51
4 Quantum Statistics 55 4.1 The Density Matrix ...... 55 4.2 Indistinguishable Particles...... 57 4.3 Thermodynamic Properties ...... 59 4.4 Chemical Potential...... 63 4.5 Summary: Quantum Statistics ...... 65
5 Boltzmann Gases 67 5.1 Kinetic Theory...... 68 5.2 Equipartition and Virial Theorems...... 70 5.3 Internal Degrees of Freedom...... 71 5.4 Chemical Equilibrium ...... 74 5.5 Summary: Boltzmann Gases ...... 76
6 Ideal Bose Gases 79 6.1 General Bose Gases...... 79 6.2 Bose-Einstein Condensation...... 80 6.3 Photons and Black-body Radiation...... 84 6.4 Phonons...... 85 Contents iii
6.5 Summary: Ideal Bose Gases...... 88
7 Ideal Fermi Gases 91 7.1 General Fermi Gases...... 91 7.2 Pauli Spin Susceptibility...... 93 7.3 Sommerfeld Expansion...... 95 7.4 Summary: Ideal Fermi Gases ...... 98
8 Interacting Gases 101 8.1 Correlation Functions ...... 101 8.2 Virial Expansion ...... 104 8.3 Summary: Interacting Gases ...... 111
9 Phase Transitions 113 9.1 Liquid-Gas Phase Transition ...... 116 9.2 Ising Model...... 119 9.3 Critical Exponents...... 124 9.4 Summary: Phase Transitions ...... 128
10 Landau-Ginzburg Field Theory 133 10.1 Introduction...... 133 10.2 Phase Transitions...... 135 10.3 Fluctuations and Correlations...... 137 10.4 Symmetry Breaking ...... 139 10.5 The Ginzburg Criterion ...... 143 10.6 Scaling Hypothesis...... 145 10.7 Summary: Landau-Ginzburg Field Theory...... 149
11 Non-Equilibrium Statistical Mechanics 153 11.1 BBGKY...... 153 11.2 Hydrodynamics...... 156 11.3 Summary: Non-Equilibrium Statistical Mechanics ...... 159
12 Review 161 12.1 Entropy...... 161 12.2 Ensembles...... 161 12.3 Classical Limit ...... 162 12.4 Thermodynamic Relations...... 164 12.5 Quantum Statistics...... 165 12.6 Interacting Gases...... 167 12.7 Phase Transitions...... 167 12.8 Landau-Ginzburg Field Theory...... 167 Preface
About These Notes
These are my class notes from the graduate statistical mechanics class I took at Michi- gan State University with Professor Luke Roberts. The primary textbook we used was Statistical Mechanics by Pathria and Beale. My class notes can be found at www.leonhostetler.com/classnotes Please bear in mind that these notes will contain errors. Any errors are certainly my own. If you find one, please email me at [email protected] with the name of the class notes, the page on which the error is found, and the nature of the error. If you include your name, I will probably list your name on the thank you page if I decide to compile and sell my notes. This work is currently licensed under a Creative Commons Attribution-NonCommercial- NoDerivatives 4.0 International License. That means you are free to copy and distribute this document in whole for noncommercial use, but you are not allowed to distribute derivatives of this document or to copy and distribute it for commercial reasons.
Updates
Last Updated: December 16, 2018 Version 0.5: (Dec. 16, 2018) First upload. Chapter 1
The Statistical Basis of Thermodynamics
1.1 Counting States Tip In general, we often want to know how to go from the Hamiltonian H of a system to the physical, macroscopic properties of the system. These properties typically include Often we will work in units where kB = 1. • N: The number of particles the system is confined to. This is typically a large number ∼ 1023 • V : The volume of the system • E: The energy of the system
46 The classical Hamiltonian H({qi, pi}) will have a huge number of variables, e.g. a 10 - dimensional space. Generally, we don’t know all the coordinates qi and momenta pi of the particles in our system, but we may know the overall volume V of the system and its energy E. There are typically many microstates of a system each with the same overall energy E. In fact, the quantity Ω(N,V,E), denotes the number of microstates that correspond to the single macrostate with particle number N, volume V , and energy E. The number Ω tells us something about our uncertainty of the microscopic state of the system. For example, if Ω = 1, i.e. there is only a single microstate with N,V,E, then there is no uncertainty because we know exactly which microstate the system must be in. Typically, however, Ω ∼ 1023, which means we are very uncertain about exactly which microstate the system is in. Another useful quantity is the Boltzmann entropy
S = kB ln Ω(N,V,E).
The entropy is also a measure of the uncertainty of the system. Example 1.1.1: Paramagnet
Suppose we have N spin-1/2 particles locked on a lattice so there are no translational degrees of freedom. Assume there are no interactions between the different particles. We will assume a magnetic field of the form
B~ = Bzˆ. 2 The Statistical Basis of Thermodynamics
The energy of a single particle is
εi = −2µBSz,iB,
where µB is the Bohr magneton, and Sz,i = ±1/2 is the spin of the particle in the z-direction. Our notation can be made more compact by defining the constant b = gµbB/2, and then the total energy of the system is
N N X X E = εi = −b σz,i, i=1 i=1
where σz,i = ±1. For our system, there is a spectrum of energies
E = {−Nb, −(N − 2)b, . . . , (N − 2)b, Nb} ,
where, −Nb corresponds to all spins being down and Nb corresponds to all spins being up. There are a total of 2N states. However, there are only N + 1 different energy levels. The question then becomes, how many states are there per energy level? For example, there is only a single state with energy −Nb, and that is the state with all spins being down ↓↓↓ · · · . Likewise, there is a single state with energy Nb, and that is the state with all spins being up ↑↑↑ · · · . Since 2N is generally much larger than N + 1, clearly, the states are not distributed uniformly among the energy levels. For example, the energy level −(N − 2)b is the state with all spins but one being down. There are N ways to arrange N spins if only one spin is up. For a given energy E, how many microstates are there with this energy? We know that E = −b (N↑ − N↓) ,
where, for example, N↑ is the number of spin-up particles. How can we enumerate the number of microstates with this energy? That is, how many states are there with N↑ up spins and N↓ down spins? In general, if you have N = N1 + N2 + ··· total objects, consisting of N1 objects of one kind, N2 objects of another kind, and so on, then the total number of permutations is N! . N1!N2! ··· In our case, N! N! Ω(N↑) = = . N↑!N↓! N↑!(N − N↑)! Then N! Ω(E) = N E N E . 2 + 2b ! 2 − 2b ! Often, we write N N N = + s, N = − s, ↑ 2 ↓ 2 then the total energy is
E = −b (N↑ − N↓) = −b(2s),
where 2s is called the spin excess. 1.1. Counting States 3
The entropy of the paramagnet is S kT S(E) = k ln Ω = k [ln N! − ln N↑! − ln N↓!] .
In the large N limit, we can apply the lowest order Stirling approximation x ln N! = N ln N − N.
If we define Figure 1.1: A plot of the “net N − N N magnetization” x versus inverse x ≡ ↑ ↓ ,N ≡ (1 ± x), N ↑/↓ 2 temperature for a paramagnet. then E = −bNx, Tip and the entropy can be written as When counting mi- 1 + x 1 + x 1 − x 1 − x crostates, make sure S = −kN ln + ln . 2 2 2 2 you are counting the states and not just the energy If we plot S/kN as a function of x, we see that it is positive in [−1, 1], symmetric levels. In general, because about x = 0, and it peaks at x = 0 with a value of ln 2. That is of degeneracy, the number of states is greater than the S(x = 0) number of energy levels. = ln 2. kN This implies that S(x = 0) = k ln 2N , but we know that S = k ln Ω, so Ω(x = 0) = 2N . This is the total number of microstates of all energies. This is due to our approximation using Stirling’s formula, but it illustrates the fact that the probabilities are dominated by a single term.
When counting states, it is often useful to use Stirling’s approximation
√ 1 1 N! ' 2πN N N exp −N + + O . 12N N 2
Often this is more useful in the logarithmic form (and then just taking the exponential in the end)
1 1 1 1 ln N! ' ln (2π) + N + ln N − N + + O . 2 2 12N N 2
The lowest order Stirling approximation can be derived as follows
N! = N(N − 1)(N − 2) ··· 2 · 1 ln N! = ln N + ln(N − 1) + ln(N − 2) + ··· + ln 2 + ln 1 N X = ln n n=1 N ' ln n dn ˆ0 = N ln N − N.
One approach to the continuum limit is to make N! continuous via the gamma func- tion. The gamma function Γ(N), where N is a continuous variable in the real numbers, 4 The Statistical Basis of Thermodynamics
is the function with the properties
N! = Γ(N + 1) Γ(N + 1) = NΓ(N) 0! = Γ(1) = 1
To satisfy these properties, the gamma function is defined as
∞ Γ(z) = dz xz−1e−x. ˆ0
The gamma function is the continuous version of the factorial function. Useful values to know, in addition to all the non-negative integer values, include 1 √ Γ = π 2 3 1√ Γ = π. 2 2 The gamma function will come up repeatedly in statistical mechanics. We can also define the digamma function ∂ Ψ(z) = ln Γ(z). ∂z The digamma function has a nice asymptotic expansion. For large z,
Ψ(z + 1) ' ln z.
Example 1.1.2
Using the Gamma function, we can write the entropy of a paramagnet
S(E) = k [ln N! − ln N↑! − ln N↓!] ,
as S(E) = k [ln Γ(N + 1) − ln Γ(N↑ + 1) − ln Γ(N↓ + 1)] . Then we can write ∂S = −k [Ψ(N + 1) − Ψ(N + 1)] , ∂x ↑ ↓
where x ≡ (N↑ − N↓)/N. Applying the asymptotic expansion of the digamma function gives us ∂S 1 1 + x = − kN ln . ∂x 2 1 − x
1.2 Connecting to Thermodynamics
Previously, we defined the Boltzmann entropy. We want to connect this with Clausius’ thermodynamic entropy. Recall the laws of thermodynamics: Zeroth law: If systems A and B are individually in equilibrium with C, then A and B are in equilibrium with each other 1.2. Connecting to Thermodynamics 5
First law: Energy is conserved Tip dE = δQ − dW. That is, the energy increase in a system is due to the heat added to the system minus Thermal contact between the work done by the system. Keep in mind that the work done by the system could two systems means that en- be a more complicated expression like dW = −P dV + µ dN + B~ · dS~ + ··· . ergy can be exchanged be- tween the two systems. Second law: Entropy never decreases. That is, dS ≥ 0. In general,
δQ C dS ≥ = V dT, T T Tip which implies that A general principle is that Tf CV equilibrium occurs when S − S ≥ dT. f 0 the accessible states Ω is ˆT0 T maximized. Third law: The entropy of a system approaches a constant value as the temperature of the system approaches absolute zero (T = 0).
Suppose we have two systems in equilibrium. The first system has the macroscopic properties N1,V1,E1, and the second has N2,V2,E2. We put the two systems in thermal contact with each other such that energy can be exchanged between the two systems but not particle number N or volume V . The energy of the combined system is
0 E = E1 + E2 = const.
The number of microstates of the system is multiplicative. For each individual system, the number of microstates depends only on the energy and not the particle number or volume, so for the combined system, the number of microstates is
0 Ω = Ω1Ω2, or more explicitly,
0 0 Ω(E ,E1) = Ω1(E1)Ω2(E2) = Ω1(E1)Ω2(E − E1).
Equilibrium in the combined system is achieved when Ω0 is maximized. This makes sense statistically because an equilibrium system is by definition in a highly probable mi- crostate, and the maximum of Ω0 corresponds to the macrostate with the largest number of microstates. An equilibrium system is then most likely in a microstate associated with 0 0 the maximum of Ω . To maximize Ω , we differentiate with respect to E1 and set it equal to zero
0 ∂Ω ∂Ω1 ∂Ω2 ∂Ω1 ∂Ω2 ∂E2 ∂Ω1 ∂Ω2 0 = = Ω2 + Ω1 = Ω2 + Ω1 = Ω2 − Ω1 . ∂E1 ∂E1 ∂E1 ∂E1 ∂E2 ∂E1 ∂E1 ∂E2 We can write this result in the form ∂ ln Ω ∂ ln Ω 1 = 2 ≡ β, ∂E ∂E 1 V,N,E1=E1 2 V,N,E2=E2 where the subscript on the first term means V and N are held constant, and the result is evaluated at E1 equal to the equilibrium energy E1. In general, then
∂ ln Ω ≡ β. (1.1) ∂E V,N,E=E 6 The Statistical Basis of Thermodynamics
Previously, we defined the Boltzmann entropy as
SB = kB ln Ω. This implies that ∂ ln Ω 1 ∂S = B . ∂E kB ∂E Recall the thermodynamic identity
dE = T dST − P dV + µ dN.
The entropy, ST used here is the thermodynamic entropy. This implies that ∂S 1 T = . (1.2) ∂E V,N T Comparing Eq. (1.1) and (1.2), we conclude that dS 1 T = = const. d ln Ω βT
−1 This implies that β ∝ T . In order to have ST = SB, we need
1 β = . kT This essentially defines the Boltzmann constant k. Instead of exchanging E, we can have our two systems in contact exchange N or V to get other relations ∂ ln Ω ∂ ln Ω 1 = 2 ≡ η ∂V ∂V 1 N,E,V1=V 1 2 N,E,V2=V 2 ∂ ln Ω ∂ ln Ω 1 = 2 ≡ ζ. ∂N ∂N 1 V,E,N1=N 1 2 V,E,N2=N 2 These relations define pressure ∂ ln Ω P = , ∂V kT and chemical potential ∂ ln Ω µ = − . ∂N kT
This implies that in equilibrium, T1 = T2, P1 = P2, and µ1 = µ2 if the systems are allowed to exchange energy, volume, and particles. In general, if we have two systems A and B in contact and they are capable of exchanging quantities X,Y,..., then the number of microstates will depend on these quantities Ω(X,Y,...) = ΩA(XA,YA,...)ΩB(XB,YB,...),
where X = XA + XB, Y = YA + YB, and so on. Then,
∂Ω ∂ΩA ∂ΩB ∂ΩA ∂ΩB = ΩB + ΩA = ΩB − ΩA ∂XA ∂XA ∂XA ∂XA ∂XB ∂Ω ∂ΩA ∂ΩB ∂ΩA ∂ΩB = ΩB + ΩA = ΩB − ΩA ∂YA ∂YA ∂YA ∂YA ∂YB . . 1.3. Intensive and Extensive Parameters 7
The equilibrium of the combined system occurs when Ω is maximized with respect to each of these quantities. I.e. equilibrium occurs when the above equations are equal to zero, and then
∂ΩA ∂ΩB ΩB = ΩA ∂XA ∂XB ∂ΩA ∂ΩB ΩB = ΩA ∂YA ∂YB . .
1.3 Intensive and Extensive Parameters
Remember, an extensive parameter depends on the size (or extent) of the system. The way in which it depends on the size of the system is additive. In general, the parameters N,V,E are extensive. Suppose you chop a system with parameters E,N,V into two pieces—one with E1,N1,V1 and the other with E2,N2,V2, then E = E1 +E2, V = V1 +V2, and N = N1 + N2. This is what it means for these quantities to be extensive. In contrast, intensive (or inEXtensive) variables do not depend on the size (i.e. extent) of the system. Examples of intensive parameters include T , P , µ, and ratios of extensive quantities like n = N/V . The intensive quantities in terms of the extensive derivatives of the entropy are
∂S 1 ∂S P ∂S µ = , = , = − . ∂E N,V T ∂V N,E T ∂N V,E T
These can be derived by looking at the derivatives when a pair of systems in thermal contact is allowed to exchange E, V , or N. They can be remembered by comparing the differential ∂S ∂S ∂S dS = dE + dV + dN, ∂E ∂V ∂N with the thermodynamic identity
dE = T dS − P dV + µ, dN, which can be rearranged as 1 P µ dS = dE + dV − dN. T T T The intensive quantities in terms of the derivatives of the energy are
∂E ∂E ∂E = T, = −P, = µ. ∂S N,V ∂V N,S ∂N V,S
For the extensive quantities, we can write down scaling variables λ, as in,
λS(E,N,V ) = S(λE, λN, λV ).
This is essentially a definition of a system whose variables are extensive. For an extensive system, E = TS + µN − PV. Taking the total derivative, gives us the Gibbs-Duhem relation
S dT = V dP − N dµ. 8 The Statistical Basis of Thermodynamics
If we want to find the intensive entropy s = S/V , we let λ = 1/V
x S E N s = = S , , 1 /V = s(, n). V V V
1/T All of this is only true if our system experiences no surface effects. Example 1.3.1: Paramagnet Figure 1.2: A plot of the “net magnetization” x versus inverse We now continue the paramagnet example. temperature for a paramagnet. For a paramagnet, we know that
∂S kN 1 + x = − ln . ∂x 2 1 − x
From E = −bNx, we get
∂S 1 k 1 + x = = − ln . ∂E T 2b 1 − x
We can invert to get b x = tanh . kT Note that x is kind of a net magnetization. See Fig. (1.2). We cannot define a pressure P for the paramagnet because the system has no volume dependence.
1.4 Averages
If the probability to have state i is Pi and the value of a variable A in the state i is Ai, then the expectation or mean of A is X hAi = AiPi. i
In the continuous case, if you have the probability p(x) of a particular value of x, then
∞ hAi = A p(x) dx. ˆ−∞ 1.5. Summary: The Statistical Basis of Thermodynamics 9
1.5 Summary: The Statistical Basis of Thermodynamics
Skills to Master • Be able to count states and calculate the number of microstates Ω accessible to a system • Calculate the average value of quantities • Calculate the Boltzmann entropy • Understand the paramagnet model system • Know the laws of thermodynamics • Derive relations for two systems in equilibrium which are exchanging some quantity • Understand extensive vs. intensive variables
Counting States Equilibrium Exchanges The Boltzmann entropy is Given two systems individually in equilibrium with N1,V1,E1 and N2,V2,E2, we are often interested in S(N,V,E) = kB ln Ω(N,V,E), relations between quantities when the two systems are where Ω(N,V,E) is the number of microstates which placed in some kind of contact (allowing the exchange correspond to the single macrostate with properties of something, e.g. energy) and the combined system N,V,E. reaches equilibrium. The energy of the combined sys- When counting microstates, make sure you are tem is just the sum counting the states and not just the energy levels. In general, because of degeneracy, the number of states is E = E1 + E2 = const. greater than the number of energy levels. The number of microstates of the combined system is When counting states, it is often useful to use Stir- multiplicative. Given that Ω (E ) microstates of sys- ling’s approximation. To lowest order, it is 1 1 tem 1 correspond to E1, and so on, the number of ln N! = N ln N − N. microstates of the total system is Ω(E,E ) = Ω (E )Ω (E ) = Ω (E )Ω (E − E ). Laws of Thermodynamics 1 1 1 2 2 1 1 2 1 Zeroth law: If systems A and B are individually in Equilibrium of the combined system occurs when Ω is equilibrium with C, then A and B are in equilib- maximized. To find this value, we partially differenti- rium with each other. ate Ω with respect to E1 and set it equal to zero. The result can be written in terms of the partial derivatives First law: Energy is conserved of ln Ω1 and ln Ω2 equal to each other and constant. In general, if we have two systems A and B in dE = δQ − dW. contact and they are capable of exchanging extensive I.e., the energy increase in a system is due to the quantities X,Y,..., then the number of microstates will heat added to the system minus the work done depend on those quantities by the system. Ω(X,Y,...) = ΩA(XA,YA,...)ΩB(XB,YB,...), Second law: Entropy never decreases. That is, dS ≥ 0. In general, where X ≡ XA + XB, Y ≡ YA + YB, and so on. Then,
δQ CV ∂Ω ∂ΩA ∂ΩB ∂ΩA ∂ΩB dS ≥ = dT, = ΩB + ΩA = ΩB − ΩA T T ∂XA ∂XA ∂XA ∂XA ∂XB which implies that ∂Ω ∂ΩA ∂ΩB ∂ΩA ∂ΩB = Ω + Ω = Ω − Ω ∂Y B ∂Y A ∂Y B ∂Y A ∂Y Tf A A A A B CV Sf − S0 ≥ dT. . . ˆT0 T . . Third law: The entropy of a system approaches a This is just the partial derivative with the product rule constant value as the temperature of the system followed by simplification. The equilibrium of the com- approaches absolute zero (T = 0). bined system occurs when Ω is maximized with respect 10 The Statistical Basis of Thermodynamics to each of these quantities. I.e. equilibrium occurs These can be derived by looking at the derivatives when when the above equations are equal to zero, and then a pair of systems in thermal contact is allowed to ex- change E, V , or N. They can be remembered by com- ∂ΩA ∂ΩB ΩB = ΩA paring the differential ∂XA ∂XB ∂ΩA ∂ΩB ∂S ∂S ∂S ΩB = ΩA dS = dE + dV + dN, ∂YA ∂YB ∂E ∂V ∂N . . with the fundamental thermodynamic relation . . dE = T dS − P dV + µdN, A general principle is that equilibrium occurs when the accessible states Ω is maximized. For the extensive quantities, we can write down Using this approach in combination with the ther- scaling variables λ, as in, modynamic differential identity gives us the following λS(E,N,V ) = S(λE, λN, λV ). relations This is essentially a definition of a system whose vari- ∂ ln Ω 1 ≡ β = . ables are extensive. For an extensive system, ∂E kT V,N,E=E E = TS + µN − PV. These relations define pressure Taking the total derivative, gives us the Gibbs-Duhem ∂ ln Ω P ≡ , relation ∂V N,E,V =V kT S dT = V dP − N dµ. and chemical potential Model: Paramagnet ∂ ln Ω µ ≡ − . For a paramagnet, we have N non-interacting spin-1/2 ∂N V,E,N=N kT particles on a lattice in an external magnetic field B~ = This implies that in equilibrium, T1 = T2, P1 = P2, Bzˆ. The energy of a single particle is εi = −2µBSz,iB, and µ1 = µ2 if the systems are allowed to exchange where µB is the Bohr magneton, and Sz,i = ±1/2 is energy, volume, and particles. the spin of the particle in the z-direction. The total energy of the system is
Intensive and Extensive Parameters N N X X An extensive parameter depends on the size (or ex- E = εi = −b σz,i = −b (N↑ − N↓) , tent) of the system. The way in which it depends on i=1 i=1 the size of the system is additive. In general, the pa- where σz,i = ±1, and b = gµbB/2. rameters N,V,E are extensive. Suppose you chop a There are a total of 2N states. The energy of the system with parameters E,N,V into two pieces—one lattice depends only on the number of up spins ver- with E1,N1,V1 and the other with E2,N2,V2, then sus the number of down spins. But given a specified E = E1 + E2, V = V1 + V2, and N = N1 + N2. This is number of up and down spins, there are many different what it means for these quantities to be extensive. possible arrangements of those spins. The number of In contrast, intensive (or inEXtensive) variables arrangements of spins which add up to the energy E is do not depend on the size (i.e. extent) of the system. N! N! Examples of intensive parameters include T , P , µ, and Ω(E) = = , N !N ! N N ratios of extensive quantities like n = N/V . ↑ ↓ 2 + s ! 2 − s ! The intensive quantities in terms of the extensive where N↑ = N/2 + s, N↓ = N/2 − s, and the quantity derivatives of the entropy are 2s is called the spin excess. Now the total energy can ∂S 1 be written as = E = −b(2s). ∂E N,V T ∂S P The entropy of the paramagnet is S(E) = k ln Ω. = Applying the lowest order Stirling approximation, we ∂V T N,E get that in the large N limit, ∂S µ = − . 1 + x 1 + x 1 − x 1 − x ∂N T S = −kN ln + ln , V,E 2 2 2 2 1.5. Summary: The Statistical Basis of Thermodynamics 11 where The expectation value of a quantity A is N − N x ≡ ↑ ↓ , N X hAi = AiPi, is kind of a net magnetization. Now, i
E = −bNx. where Pi is the probability of getting Ai. In the con- tinuous case, if your probability distribution is p(x), The temperature of the paramagnet is given by then ∞ 1 ∂S kB 1 + x hAi = Ap(x) dx. = = − ln . ˆ−∞ T ∂E 2b 1 − x The gamma function Inverting gives us ∞ z−1 −x b Γ(z) = dz x e , x = tanh . ˆ0 kT is the continuous version of the factorial, and its prop- Miscellaneous erties include
In general, if you have N = N1 + N2 + ··· total ob- N! = Γ(N + 1) = NΓ(N). jects, consisting of N1 objects of one kind, N2 objects of another kind, and so on, then the total number of Useful values to know include permutations is N! 1 √ . Γ = π N !N ! ··· 2 1 2 √ The higher order Stirling approximation is 3 π Γ = . 2 2 √ 1 1 N! ' 2πN N N exp −N + + O . 12N N 2 The digamma function is
In logarithmic form it is ∂ Ψ(z) = ln Γ(z). ∂z 1 1 1 ln N! ' ln (2π) + N + ln N − N + + ··· 2 2 12N It has a nice asymptotic expansion. For large z,
The general Gaussian integral is Ψ(z + 1) ' ln z. ∞ r 2k −ax2 (2k − 1)!! π A useful approximation if x << 1 is x a dx = k+1 k , ˆ0 2 a a 1 ln(1 + x) ' x − x2. provided that k ∈ Z, and a > 0. 2 Chapter 2
Ensembles and Classical Phase Space
2.1 Classical Phase Space
Consider an N particle system with coordinates and momenta q1, . . . , q3N , p1, . . . , p3N . Using the Hamiltonian H({qi, pi}), we can write down Hamilton’s equations
∂H ∂H q˙i = , p˙i = − . ∂p ∂qi
Our system corresponds to a single point in a 6N-dimensional phase space. As our system evolves, its corresponding point in phase space traces out some trajectory. Given some function of time f({qi(t), pi(t)}), we want to know how this function evolves in time. The total derivative is
df ∂f X ∂f ∂f = + q˙i +p ˙ dt ∂t ∂qi ∂pi ∂f X ∂f ∂H ∂f ∂H = + − ∂t ∂qi ∂pi ∂pi ∂qi ∂f = + {f, H} , ∂t This can be written as df ∂f = + {f, H} , dt ∂t where X ∂f ∂H ∂f ∂H {f, H} = − , ∂qi ∂pi ∂pi ∂qi
df is called a Poisson bracket. The total derivative dt is a convective time derivative since it tells us how the system is evolving in phase space. Now instead of a single system corresponding to a single point in phase space, let’s think about an entire ensemble of systems (with the same H) corresponding to an entire ensemble of points moving through our phase space. An ensemble is simply many different copies of our system. The phase space density is a function of the form
ρ = ρ (qi, . . . , q3N , pi, . . . , p3N ) .
For equilibrium,
∂tρ = 0. 2.2. Liouville’s Theorem 13
Phase-space averaged quantities are calculated as
d3N q d3N p f ({q , p }) ρ hf(t)i = i i . ´ d3N q d3N p ρ ´ The question then becomes, how does ρ evolve through phase space?
2.2 Liouville’s Theorem
We can define some generalized phase space velocity
~vps = ({q˙i}, {p˙i}) .
To get the total number of systems in a volume V , we integrate ρ over the volume
NV = dω ρ. ˆV The flux of points through a surface is given by
F~ = ρ~vps.
Then ∂ ρ = − ρ~v · d~s. ∂t ˆV ˆ So the rate of change of the number of systems in the volume equals the flux through the surface. This implies a continuity equation in the phase space
∂tρ + ∇ · (ρ ~vps) = 0.
This tells us how the phase space evolves if we know the phase space velocities. Note, the divergence here has many dimensions. We can expand the divergence as
∇ · (ρ ~v) = ρ ∇ · ~v + ~v · ∇ρ.
The first term on the right can be written as
2 2 X ∂q˙i ∂p˙i X ∂ H ∂ H ρ ∇ · ~v = + = − = 0. ∂q ∂p ∂q ∂p ∂p ∂q i i i i i i i i To simplify the above result, we plugged in Hamilton’s equations
∂H ∂H q˙i = , p˙i = − . ∂pi ∂qi 14 Ensembles and Classical Phase Space
So our divergence simplifies to
X ∂ρ ∂ρ ∇ · (ρ ~v) = ~v · ∇ρ = q˙ +p ˙ . i ∂q i ∂p i i i Again, we apply Hamilton’s equations to get
X ∂ρ ∂H ∂ρ ∂H ∇ · (ρ ~v) = − = {ρ, H} . ∂q ∂p ∂p ∂q i i i i i So, we have that
∂tρ + {ρ, H} = 0. This is Liouville’s theorem. Recall the convective derivative df ∂f = {f, H} + . dt ∂t The convective derivative of ρ is dρ = 0. dt This has to be true for any ensemble of systems that is described by Hamilton’s equations. For a stationary ensemble,
∂tρst = 0 =⇒ {ρst, H} = 0. This is satisfied if ρ = ρ[H], i.e. if ρ is a functional of H. One possibility (a microcanonical ensemble) is when
ρst = const. This is consistent with the Boltzmann hypothesis. I.e. the Boltzmann hypothesis is plausible. Another example (a canonical ensemble) is when
ρ ∝ e−βH.
The Master Equation is another approach—a discrete approach to justify all this. We have an ensemble of N systems that can take on some number of states. There are Ni states with energy Ei in the overall ensemble of N systems. The rate of transition from state i to state j, denoted ωij is
dNi X = (ω N − ω N ) . dt ji j ij i j
dNi We want to find an equilibrium state where dt = 0. The most obvious way to find such a state is to say that ωjiNj = ωijNi. Note that this is a particular solution—not a general solution. If we assume time-reversal invariance, then ωij = ωji.
This implies that Ni = Nj, which implies that N i = P = P . N i j That is, we are justified to assume that the probabilities are equal. 2.3. Microcanonical Ensemble 15
2.3 Microcanonical Ensemble
In the microcanonical ensemble, the macrostate is defined by a fixed number of particles 1 1 N, a fixed volume V , and a fixed energy E or a fixed energy range E − 2 ∆,E + 2 ∆ . Then the chief problem is to calculate the number of microstates Ω(N,V,E) that are accessible to the system. Unfortunately, for most physical systems, calculating Ω is very difficult. Furthermore, we typically do not know the energy E for a physical system. Because of such problems, later we will look at other approaches, e.g. the canonical ensemble. What does our ensemble tells us about real systems? For the microcanonical ensemble, the phase density is constant ( const if E − 1 ∆ ≤ H(q, p) ≤ E + 1 ∆ ρ = 2 2 . 0 otherwise.
In other words, the phase volume of the microcanonical ensemble is the volume of the “shell” defined by 1 1 E − ∆ ≤ H(q, p) ≤ E + ∆. (2.1) 2 2 The volume enclosed by this shell is
ω = dω = d3N q d3N p, ˆ ˆ where the integral goes over the region of phase space for which Eq. (2.1) is true. The ensemble average of a function f is defined as dω fρ hfi = . ´ dω ρ ´ The time average is defined as
N 1 X f = f ({q(t ), p(t )}) . N j j j=1 We want to assume that hfi = f. This is essentially the ergodic hypothesis. Note, there are systems which do not satisfy this. A system that satisfies it must trace out all of its available phase space over some time. We define the classical weight function as integrating over the shell in phase space d3q ··· d3q d3p ··· d3p 1 1 Ω(E) = 1 N 1 N θ H − E − ∆E θ H − E + ∆E , ˆ ω0 2 2 (2.2) where ( 1 if x > 0 θ(x) = . 0 if x < 0. is the Heaviside step function. All that the product of Heaviside functions is doing is selecting the shell. It multiplies the integral by 1 if the condition in Eq. (2.1) is true, and otherwise, it multiplies the integral by 0. In the limit ∆E → 0, the Heaviside functions become a delta function dω Ω(E) = ∆E δ(E − H). ˆ ω0 16 Ensembles and Classical Phase Space
We will later derive that 3N ω0 = h , where h is Planck’s constant, and 3N are the degrees of freedom for N particles in 3D space. The quantity ω0, called the fiducial element of phase space, can be thought of as the infinitesimal phase space volume corresponding to a single microstate. For discrete states, as in our paramagnet example, we would have X Ω(E) = δEi,E. i
Example 2.3.1: Classical Ideal Gas
We will now apply the microcanonical machinery to an example—the classical (i.e. non-quantum) non-relativistic ideal gas. For the ideal gas, we have N particles in a volume V = L3 for a cubic box. The total Hamiltonian is N 2 X ~pj H = . 2m j=1 We can split Eq. (2.2) into two pieces. We start by calculating the volume of phase space with energy less than E, denoting it Σ(E) and then we can find the density of states around E as ∂Σ(E) Ω(E) = ∆E, ∂E where ∆E << E. For the number of states with energy less than E, we get
N N N 2 1 Y 1 Y X ~pj Σ(E) = d3q d3p θ (E − H) = d3q d3p θ E − . ω ˆ i i ω ˆ i i 2m 0 i 0 i j=1
This is the volume of phase space below energy E. This calculation assumes that the particles are distinguishable. In reality, our particles are indistinguishable, and so we are overcounting with this formula. To remove the overcounting, we need to divide by N!
N N 2 1 Y X ~pj Σ(E) = d3q d3p θ E − . ω N! ˆ i i 2m 0 i j=1
We can write this as N ∞ N 2 V 1 X ~pj Σ (E) = d3p ··· d3p θ E − , CL N! ω ˆ 1 N 2m 0 −∞ j=1
where the subscript “CL” indicates that this is for the classical gas. Note, the √integral here is the volume of a 3N-dimensional hypersphere with radius R = 2mE. What about for the quantum-mechanical case? For a QM ideal gas, the wave function of the system is
23N/2 Y ψ(~x ,... ~x ) = sin (kxx ) sin (kyy ) sin (kzz ) . 1 N V N/2 i i i i i i i 2.3. Microcanonical Ensemble 17
The energy eigenvalues are N 2 X 2 E = ~ ~k . α 2m j j where π ~k = nx + ny + nz , nl = 1, 2,... j L j j j j l The subscript α simply takes the place of the quantum numbers {nj} that the system depends on. Note that this is simply the wave function and energy for N particles in an infinite square well. Then the number of states with energy less than E is ∞ N 1 X 2 X 2 Σ (E) = θ E − ~ ~k . QM N! 2m j ~n1,...,~nN j Schematically, we expect our energy to have the form
2/3 1/3 Eα = Nε + N c1 + N c2 + ··· ,
where ε is the average energy per particle. The first term is the bulk energy, the second term is a correction due to surface effects, and the third term is a correction due to edge effects. In the large N limit, only the bulk (first) term survives. For large ~k, we can approximate the sum over one of the quantum numbers as a momentum integral
∞ X L ∞ L ∞ L ∞ → dk = dk = dp. π ˆ 2π ˆ 2π ˆ n=1 0 −∞ ~ −∞
We can do this since the functions depend only on k2. In the first equality, we double the integration region and divide by two. In the second, we use the fact that pi = ~ki. Then we end up with N ∞ N 2 1 V X ~pj Σ (E) = d3p ··· d3p θ E − . QM N! (2π )3N ˆ 1 N 2m ~ −∞ j
This is the same as the classical result provided that we define
3N ω0 = h .
Generically, this is the value of the phase space normalization factor ω0 and it can be shown using a variety of approaches and a variety of systems. It is also the smallest resolvable volume that is predicted by the Heisenberg uncertainty principle ∆x∆p ≥ ~. 2 So far we have not made any assumptions about the form of the single particle energies, so we expect the classical ideal gas law
PV = NkT
to hold true. We can write this relation as ∂S P ∂k ln Ω kN = = = , ∂V E,N T ∂V E,N V 18 Ensembles and Classical Phase Space
which implies that ln Ω = N ln V + ln Ωp(E,N). It can be shown that the volume of a 3N-dimensional hypersphere of radius R is 3N/2 π 3N 3N Λ3N (R) = 3N R = C3N R . Γ 2 + 1 √ In our case, R = 2mE, so we have that
3N/2 Λ3N (R) = C3N (2mE) .
Then ∂Σ(E) V N 3N 2mE 3N/2 ∆E Ω(E) = ∆E = C . ∂E N! 3N 2 h2 E Now we just have to take the log and apply Stirling’s approximation to get the entropy using S/k = ln Ω
S 3N 3N 2mE ∆E = N ln V − ln N! + ln C + ln + ln + ln . k 3N 2 2 h2 E
Using Stirling’s approximation,
N ln N! ' N ln N − N = N ln N − N ln e = N ln , e
and it turns out that 3N 2πe ln C ' ln . 3N 2 3N With these approximations, we get " # S V e emE 3/2 3N ∆E ' N ln + N ln + ln + ln . k N 3π~2N 2 E
The last two terms go to zero as N → ∞, so we get
" # S V mE 3/2 5 = N ln + N. k N 3π~2N 2
This is the Sackur-Tetrode equation for the entropy of an ideal gas. Now that we have the entropy, we can extract some thermodynamic quanti- ties. For example, ∂S 1 = , ∂E V,N T gives us 3 E = NkT, 2 and ∂S µ = − , ∂N E,V T 2.4. Canonical Ensemble 19
gives us Tip " 3/2# V mE An “adiabatic process” is µ = T ln 2 . N 3π~ N often taken to mean a pro- cess that occurs at constant For adiabatic (i.e. constant entropy) changes, we have entropy.
VE3/2 = const VT 3/2 = const PV 5/3 = const.
We imposed no symmetry requirements on the wave function of the system. This is okay to do if our gas is a very low density. So our results here are only really valid in the low density N/V limit. The main purpose of this example was to illustrate that calculations done in the microcanonical ensemble are long and tiresome. This motivates our develop- ment of more useful ensembles such as the canonical ensemble.
The Gibb’s Paradox If the particles are distinguishable, then we have to drop the 1/N! factor from Σ(E). Then the entropy gets an extra contribution of N ln(N/e), which would make it non-extensive. Suppose we have two boxes with the same volume, same number of particles, and with entropies S1 and S2. Suppose these two boxes are next to each other and we remove the barrier between them. The total entropy of the combined system is now " # 2V m2E 3/2 2N Stot = k2N ln + 5kN + 2N ln . 2N 3π~2 e
We can write this as
Stot = S1 + S2 + Smix, where S1 and S2 are just the Sackur-Tetrode entropies of the individual boxes, and
2E S = 2N ln , mix e is the additional entropy term that appears when the two boxes are mixed together. Apparently, removing the barrier and allowing distinguishable particles to mix creates entropy. Such a process is irreversible.
2.4 Canonical Ensemble
In the microcanonical ensemble, we had N, V , and E as fixed quantities. Then we count states to calculate the entropy, and from that we calculate thermodynamic quantities. In the canonical ensemble, the macrostate is defined by the parameters N, V , and T . The energy E of the system can vary (in principle from 0 to ∞). In the canonical ensemble, we split a very large system into two pieces which can exchange energy. We assume that one of the systems is very large (i.e. N ∼ 1046), and we call it the reservoir and the other system called the subsystem is just large (i.e. N ∼ 1023). 20 Ensembles and Classical Phase Space
Considering the total system, we can neglect the multiplicity of the subsystem and write the multiplicity of the total system as the multiplicity of the reservoir
Ωt = Ωr.
Then, assuming the total system has energy E0, the probability of finding the subsystem in a state j with energy Ej is
P (Ej) ∝ Ωr(E0 − Ej),
or in terms of entropy Sr(E0 − Ej) P (Ej) ∝ exp . kB Expanding this gives us " # S (E ) ∂S E E r 0 r j j −βEj P (Ej) ≈ exp − ≈ exp − = e . kB ∂E V,N kB kBT