Theoretical Statistical Physics: Prof. Dr. Christof Wetterich Institute For Theoretical Physics Heidelberg University
Theoretical Statistical Physics: Prof. Dr. Christof Wetterich Institute For Theoretical Physics Heidelberg University
Theoretical Statistical Physics: Prof. Dr. Christof Wetterich Institute For Theoretical Physics Heidelberg University
Contents
I.
1. Introduction
1.1. From Microphysics to Macrophysics . . . . . . . . . . . . . . . . . . . . . . . .
1.2. Macroscopic laws and simplicity . . . . . . . . . . . . . . . . . . . . . . . . . .
1.3. Fundamental laws are of probabilistic nature . . . . . . . . . . . . . . . . . . .
11
11
12
12
15
15
15
15
15
16
16
16
17
17
18
18
20
23
23
24
24
25
25
26
26
29
29
30
31
33
33
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
35
35
36
36
38
38
38
39
39
40
41
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
43
43
43
43
45
45
46
46
47
49
50
51
53
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
55
55
55
56
56
57
61
61
63
63
65
65
66
68
68
70
70
71
4. Equilibrium Ensembles
4.1. Microcanonical ensemble . . . . . . . . . . . . . . . . . . . . . .
4.1.1. Fundamental statistical postulate . . . . . . . . . . . . .
4.1.2. Microcanonical partition function . . . . . . . . . . . . .
4.1.3. Entropy S . . . . . . . . . . . . . . . . . . . . . . . . . .
4.1.4. Partition function for ideal classical gas . . . . . . . . .
4.1.5. Entropy of an ideal classical gas . . . . . . . . . . . . . .
4.1.6. Thermodynamic limit . . . . . . . . . . . . . . . . . . . .
4.1.7. The limit E E 0 . . . . . . . . . . . . . . . . . . . . . .
4.1.8. Derivatives of the entropy . . . . . . . . . . . . . . . . .
4.2. Canonical ensemble . . . . . . . . . . . . . . . . . . . . . . . . .
4.2.1. System in a heat bath . . . . . . . . . . . . . . . . . . . .
4.2.2. Canonical partition function and Boltzmann factor . . .
4.2.3. Relation between temperature and energy . . . . . . . .
4.2.4. Energy distribution . . . . . . . . . . . . . . . . . . . . .
4.2.5. Equivalence of microcanonical and canonical ensemble .
4.2.6. Temperature . . . . . . . . . . . . . . . . . . . . . . . . .
4.2.7. Temperature and energy distribution in subsystems . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
5. Thermodynamics
5.1. Thermodynamic potentials . . . . . . . . . . . . . . . . . .
5.1.1. State variables, equations of state . . . . . . . . . .
5.1.2. Thermodynamic potentials and partition functions
5.1.3. Dierential relations . . . . . . . . . . . . . . . . .
5.1.4. Legendre transform . . . . . . . . . . . . . . . . . .
5.1.5. Properties of entropy . . . . . . . . . . . . . . . . .
5.1.6. Zero of absolute temperature . . . . . . . . . . . .
5.1.7. 3rd law of thermodynamics . . . . . . . . . . . . .
5.2. Reversible and irreversible processes . . . . . . . . . . . .
5.2.1. Subsystems and constraints . . . . . . . . . . . . .
5.2.2. Removing the constraint . . . . . . . . . . . . . . .
5.2.3. Increase of entropy by removing the constrains . .
5.2.4. 2nd law of thermodynamics . . . . . . . . . . . . .
5.2.5. Reversible processes for isolated systems . . . . . .
5.2.6. Reversible processes for open systems . . . . . . .
5.2.7. Thermodynamics and non-equilibrium states . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
71
72
73
73
74
74
75
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
77
77
77
79
80
81
83
84
85
85
85
86
87
88
89
89
90
91
93
93
94
95
95
96
97
97
98
99
99
101
103
104
105
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
105
105
106
106
107
107
108
109
110
110
111
112
113
114
114
115
115
115
116
116
116
117
118
118
119
119
119
120
121
.
.
.
.
.
.
.
.
.
.
.
.
123
123
123
124
124
125
125
126
127
127
128
128
129
129
8. Magnetic systems
8.1. Magnetization . . . . . . . . . . . . . . . . . . . . . . .
8.1.1. Degrees of freedom . . . . . . . . . . . . . . . .
8.1.2. Hamiltonian . . . . . . . . . . . . . . . . . . . .
8.1.3. Magnetization . . . . . . . . . . . . . . . . . . .
8.1.4. Thermodynamic potential . . . . . . . . . . . .
8.1.5. Eective potential . . . . . . . . . . . . . . . .
8.1.6. Minimal principle . . . . . . . . . . . . . . . . .
8.2. Spontaneous symmetry breaking . . . . . . . . . . . .
8.2.1. Symmetries . . . . . . . . . . . . . . . . . . . .
8.2.2. Vanishing magnetic eld . . . . . . . . . . . . .
8.2.3. Domains . . . . . . . . . . . . . . . . . . . . . .
8.2.4. General form of U . . . . . . . . . . . . . . . .
8.3. Landau theory and phase transition . . . . . . . . . . .
8.3.1. Quartic expansion of U . . . . . . . . . . . . . .
8.3.2. Phase transition for j = 0 . . . . . . . . . . . .
8.3.3. Discontinuity in magnetization for non-zero B
8.4. Magnetic susceptibility . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
133
133
133
134
134
135
136
137
137
137
138
138
139
139
139
140
142
143
9. Phase transitions
9.1. Order of phase transition . . . . . . .
9.1.1. First order phase transition .
9.1.2. Second order phase transition
9.1.3. Probabilistic interpretation . .
9.2. Liquid-gas phase transition . . . . . .
9.2.1. Eective potential . . . . . .
9.2.2. First order transition . . . . .
9.2.3. Compressibility . . . . . . . .
9.2.4. Chemical potential . . . . . .
9.3. Phase diagram . . . . . . . . . . . . .
9.3.1. Vapor pressure curve . . . . .
9.3.2. p T diagram . . . . . . . . .
9.3.3. Clausius-Clapeyron relation .
9.3.4. Latent heat . . . . . . . . . .
9.3.5. Superheating, supercooling .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
147
147
147
147
149
150
150
151
152
152
153
153
153
154
155
155
.
.
.
.
.
.
.
.
.
.
157
157
157
157
157
158
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
158
160
161
161
162
163
164
164
164
164
165
167
170
173
Part I.
1. Introduction
1.1. From Microphysics to Macrophysics
Fundamental laws for basic interactions do not explain everything, e. g. the description of
macroscopic objects, consisting of many particles (solids, uids, gases).
A typical number of particles is Avogadro's number: N a ' 6 1023 . It denotes the number of
molecules per mol, e. g. the number of water molecules in 18 g water.
Typical questions are:
connections between pressure (p), volume (V ), and temperature (T ) of gases (equation of
state)
specic heat capacity of a material
conductivity of a solid, colour of a uid
connection between magnetization and magnetic eld
melting of ice
How does a refrigerator work?
How does a supernova explode?
energy budget of the Earth's atmosphere
diusion through a membrane in a plant
particle beam in a high energy collider
Typical area of application:
systems with many degrees of freedom macroscopic objects
For example a water droplet:
Consider an isolated system a droplet of water with volume V and number of particles N .
A classical mechanical description needs 6N initial conditions 6N degrees of freedom.
The distance between two trajectories, which initially are close to each other, grows
exponentially with time. Initial state can never be prepared (known) with arbitrary
precision: Limit for predictability (weather forecast)
11
1. Introduction
Try deterministic classical mechanics as a description of a water droplet for one second.
One needs more information than can be saved in the entire universe, even if one stores
one bit in every volume with Planck size (1034 m) 3
The goal of the statistical physics is to develop laws for the macrostates by deriving them from
the microphysics with the help of statistical methods.
1
N
predictive power only for systems with many degrees of freedom; uncertainty about the
behaviour of single atoms (quantum physics), but precise predictions about the whole
system
thermodynamics
material properties
non-equilibrium processes
Scales for macroscopic systems:
L mean distance between particles in uids and solids ' 107 cm, N 1
t timescale of the microscopic motions:
motion of an atom: 1013 s
motion of an electron (in solids): 1015 s
12
Abstraction
Understanding
of the complexity
Fundamental laws
(1)
(2)
To calculate the predictions, which follow from the fundamental theory, one uses methods of
statistical physics just like in this lecture.
Basic laws of physics are statistical in nature!
They make statements about:
the probability that, when measuring an observable A once, one gets the value Ai
the conditional probability for two measurements: if one measures a value Ai for an observable A, what is the probability that one measures the value B j for a second observable
B?
Remarks
1. From microphysics to microphysics and from Planck's scale (1019 GeV) to the energy scale
of LHC (1000 GeV) 16 orders of magnitude.
2. Universality: one does not consider all the details of the microscopic laws predictive
power of the QFT!
3. Conceptional revolution due to the statistical physics.
Modern physics: Plank, Boltzmann, Einstein.
(((( probabilistic theory predictive power
4. (
Determinism
(
13
(2.1)
p 0
positivity,
(2.2)
p = 1
normalization.
(2.3)
0 p 1 .
(2.4)
The probability distribution is {p } = {p1 ,p2 ,p3 , . . . }, which is an ordered sequence of numbers.
2.1.3. Observables
Capital roman letters are used for observables, e. g. A. A classical observable A has a xed value
A in every state :
A Spec A ,
(2.5)
where the spectrum Spec A of an observable is the set of all possible measurement values { A },
A R.
The probability to measure a certain value A is given by:
X
pA =
p ,
(2.6)
A = A
where one sums over all states for which A = A . A linear combination of two classical
observables A and B is again a classical observable:
C = f AA + f B B
C = f A A + f B B .
with
f A,B R ,
(2.7)
(2.8)
15
A(2)
A(3)
0
0
1
0
0
1
0
0
f (A)
= f (A ) .
(2.9)
p A .
(2.10)
p (A hAi) 2
(2.11)
(2.12)
p A2 hAi2
(2.13)
2.1.5. Dispersion
The dispersion is given by
A2 =
= hA2 i hAi2 .
(2.14)
The observables A( ) are called projectors. They can be considered as a basis for observables.
All classical observables can be expressed as linear combinations of projector observables. An
association of A( ) with possible events (i. e. that state is realized) leads to the Kolmogorov
axioms for probabilities.
16
~
A.
2
(2.16)
(2.17)
= p1 (1 p1 )
(2.18)
= 2p1 1 ,
(2.19)
hN i = h 12 (A + 1)i
(2.20)
= p1 N 1 + p2 N 2
(2.21)
= p1 .
(2.22)
(2.23)
= p1 + (1 p1 )
(2.24)
= 1,
(2.25)
(2.26)
2
= 4p1 (1 p1 ) ,
(2.27)
(2.28)
and therefore
A2 = 0
for p1 = 0
or p1 = 1 .
(2.29)
(2.30)
N = p1 (1 p1 ) .
(2.31)
17
11
10
01
00
|1i
|2i
|3i
|4i
|1i
A
B
C
2
1
1
|2i
0
1
1
|3i
0
1
1
|4i
2
1
1
total spin
spin 1
spin 2
(2.32)
hA2 i = 2 ,
(2.33)
B = C = 1 ,
A = 2 .
(2.34)
18
same B and arbitrary C . As a consequence, the system turns into an eective two-state
system.
Here, = | 1 i groups together these microscopic states = |1i, = |2i, in which B
has the same values B = 1. Similarly, = | 2 i groups together = |3i and = |4i with
B = 1. The states of this reduced ensemble are illustrated in Table 2.4. The probabilities
of the subsystem states are
p1 = p1 + p2 ,
p2 = p3 + p4 .
(2.35)
(2.36)
Using B1 = 1 and B2 = 1, which are the classical observalbles in the subsystem, the
expectation value of B can be computed from the subsystem:
hBi = p1 B1 + p2 B2
= p1 p2
(2.37)
(2.38)
= p1 + p2 p3 p4 .
(2.39)
As it should be, this is the same as computed from the underlying microscopic system:
hBi =
4
X
B p .
(2.40)
=1
In this case, B is a system observable, which means that the information in the system
is enough to compute the distribution of the measurement values. In contrast, A and C
are environment observables.
2. Consider the case when A is the system observable of the subsystem. Thus, the system
becomes an eective three-state system with
A1 = 2 ,
A2 = 0 ,
A3 = 2 .
(2.41)
p2 = p2 + p3 ,
p3 = p4 .
(2.42)
State | 2 i of the subsystem groups together the states |2i and |3i of the microsystem that
both have A = 0. The states | 1 i and | 3 i of the subsystem are identical to the microstates
|1i and |4i with A = 2 and A = 2. We will continue examining this example in
section 2.4.
19
| 1 i
| 2 i
(2.44)
n =
dn n w (n)
(2.45)
dn w (n) = 1.
The Motivation of these concepts is, to characterize microstates by density and state
of water. So, we have microstates (n, ) with probability distribution P (n, ), obeying
X
dn P (n, ) = 1 .
(2.46)
In this setting, n is a classical observable with xed value n in each state (n, ). It is
hni = dn P (n, ) n ,
(2.47)
hn2 i = dn P (n, ) n 2 .
(2.48)
20
dn P (n, ) ,
(2.49)
P (n, )
.
p
(2.50)
dn w (n) = 1 .
(2.51)
Then,
P (n, ) = p w (n) ,
and we can rewrite (2.47) and (2.48) as
X
hni =
dn w (n) n
hn i =
2
p n ,
(2.53)
p (n2 ) .
(2.54)
p
(2.52)
dn w (n) n2 =
(2.55)
For hni the specication of n for every state is sucient (at this point no distinction from
classical observable), but hn2 i needs additional information about distribution of values n
in each state .
n is a probabilistic observable on the level of states .
2. Subsystem of states with total spin:
Probabilistic observables typically occur for subsystems. For example we can identify
the two-spin system with the microstates (n, ) of water. Then, the three states
characterized by total spin A in subsection 2.2.2:
A1 = 2 ,
A2 = 0 ,
A3 = 2 ,
(2.56)
are analogous to the states ice, water, vapor in the rst example.
21
to nd B = 1 ,
(2.57)
to nd B = 1 .
(2.58)
(2.59)
hBi = p1 B2 + p2 B2 + p3 B3
p2 p3
= p1 + (p2 + p3 )
p4
p +p
| {z } |2{z }3
(2.60)
p2
(2.61)
B 2
= p1 + p2 p3 p4
q. e. d.
(2.62)
However,
(B2 )2 = 1 ,
(2.63)
3
X
p B2 .
(2.64)
=1
additional information
B system observable.
(2.65)
We have listed all the observables that we just have computed in Table 2.5.
3. Quantum mechanics:
Consider two-state quantum mechanics:
=1:
=2:
where Sk = ~2 Sk .
22
Sz = 1 ,
Sz = 1 ,
(2.66)
(2.67)
| 1 i
p
A
N
B
B2
p1
2
2
1
1
| 2 i
| 3 i
p2 + p3 p4
0
2
1
0
p2 p3
1
p2 +p3
1
1
The additional observable Sx has no xed value in the states , only probabilities to nd
Sx = 1:
w 1 (1) = a ,
w 1 (1) = 1 a ,
(2.68)
w 2 (1) = b ,
(Sx )1 = 2a 1 ,
w 2 (1) = 1 b ,
(Sx )2 = 2b 1 .
(2.69)
(2.70)
(2.71)
(2.72)
So,
1 hSx i 1 ,
(2.73)
(2.74)
23
(2.76)
= A
(2.77)
= A
.
= tr (A)
(2.78)
(2.79)
2.5.2. Dispersion
Accordingly, the dispersion is given by
hA2 i = tr (A2 ) ,
(2.80)
A = hA i hAi .
2
(2.81)
(2.82)
(2.83)
hAi = tr (A)
X
=
A
(2.84)
(2.85)
A .
(2.86)
So, one only needs the diagonal elements of the density matrix. They can be associated with
probabilities p = , where 0. This way, we recover classical statistics:
hf (A)i =
24
p f (A ) .
(2.87)
Example
A two-state system:
11
Let A = A01 A02 and = 12
12
22 ,
then:
!
!
!
12
A1 0
A1 11 A2 12
=
,
22
0 A2
A1 12
A2 22
A = 11
12
= A1 11 + A2 22 ,
tr (A)
X
X
hAi =
A =
p A .
n
(2.88)
(2.89)
(2.90)
n = 1 and n : 0 n 1
Example
Quantum mechanical two-state system:
In this case, the density matrix is a complex 2 2 matrix of the form
!
11
12
=
,
12 1 11
with 11 R, 0 11 1. Consider the observable
!
1 0
A =
.
0 1
(2.91)
(2.92)
12
1 + 11
2
~
= (2 11 1) .
(2.95)
2
From the classical statistics we have hAi = p1A1 + p2A2 . If we set p1 = 11 , p2 = 1 11 ,
A1 = ~2 , A2 = ~2 , then we will get the same result (2.95).
25
11 (1 11 )
1
4
11
1
2
(2.96)
= + 11 (1 11 ) f ,
(2.97)
where
f = | 12 | 2 0 .
(2.98)
(2.99)
We nd
1,2 =
1
2
1 4[ 11 (1 11 ) f ] .
(2.100)
11 (1 11 ) 0 .
(2.101)
and therefore,
The full positivity condition for the two-state matrix is
f = | 12 | 2 11 (1 11 ) .
(2.102)
In Figure 2.2 we have plotted 11 (1 11 ). From the positivity condition it follows that | 12 | 2
only can take values on or below this curve.
26
hSz i
hSx i
2
hS x i, hSz2 i
|1i
|2i
12 + 12
1
12 + 12
1
Table 2.6.: Mean values of spin observables in the states |1i and |2i
Example
Spin in arbitrary direction Si = ~2 Si :
The spin components Sz and Sx are dierent types of observables:
Sz : Classical observable
Sz = 1 ,
Sz = 1 ,
1
2
Sz2 = 1 ,
Sz2 = 1 .
1
(2.103)
(2.104)
Sx : Probabilistic observable
There is a probability distribution for nding the values +1 or 1 in the states |1i
and |2i:
hSx i1 = hSx i2 = 12 + 12
.
(2.105)
But in both states the square has a xed value:
Sx2 = Sx2 = 1 .
1
(2.106)
w 1 (1) = 12 (1 + 12 + 12
),
w 1 (1) =
1
2 (1
12
12
),
(2.107)
(2.108)
such that
w 1 (1) + w 1 (1) = 1 .
(2.109)
(2.110)
Both, w 1 (1) and w 1 (1), have to be positive, which can be inferred from the condition
| 12 + 12
| 1,
(2.111)
or equivalently
| Re 12 |
1
2
(2.112)
27
| 12 | 2 | Im 12 | 2
(2.113)
11 (1 11 ) | Im 12 | 2
(2.114)
11 (1 11 )
(2.115)
1
2
(2.116)
Classical:
= 11 ( 12 +
= 12 +
12
)
(2.117)
+ (1 11 )( 12 +
12
12
)
(2.118)
(2.119)
= 2 Re 12
(2.120)
"
Quantum:
!
!#
0
1
11
12
hSx i = tr
12
22
1 0
!
11
= tr 12
22 12
(2.121)
(2.122)
= 12 + 12
(2.123)
= 2 Re 12 .
(2.124)
One can see that (2.120) and (2.124) give exactly the same result.
Actually, this could be generalized for any hermitian quantum operator
!
A11 A12
A=A =
.
A12 A22
(2.125)
The probabilistic observable has mean values in the states |1i and |2i, given by hAi1 and hAi2 :
hAi1 = A11 + 12
A12 + 12 A12
hAi2 = A22 +
12
A12
12 A12
(2.126)
.
(2.127)
It turns out that quantum mechanics can be interpreted as classical statistics, but with additional
information for probabilistic observables. As long as we are dealing with diagonal observables,
there is no dierence between both approaches. In this case, non-diagonal elements of are of
no importance. Actually, the last statement is only true for a given point in time or for a static .
Non-diagonal elements of are important for the time evolution of the probability distribution:
{p (t )} { (t )} ,
where the sum over repeated indices is implied.
28
(2.128)
hAi = tr (A)
= A
= A
(2.129)
(2.130)
(2.131)
i,
= h |A|
(2.132)
(2.133)
As already mentioned, only is needed for diagonal observables. So, for diagonal observables
there is no dierence between pure and mixed states.
We can state the condition for pure states as
2 = ,
(2.134)
(2.135)
because
(2.136)
= .
(2.137)
(2.138)
)
= tr (U U U AU
)
= tr (U AU
(2.139)
= tr (U U A )
= tr (A ) .
(2.141)
(2.140)
(2.142)
29
lz
pure states
1
*.
.0
,0
density matrix
probability
-1
1
*. +/
.0/
,00
0
0
0
*. +/
.1/
,00
1
0
0
*. +/
.0/
,10
0
0
0
+
0//
0-
0
*.
.0
,0
p1
0
+
0//
0-
p2
0
*.
.0
,0
0
+
0//
1-
p3
02 = 0 .
(2.143)
= UU
(2.144)
(2.145)
(2.146)
= .
0
(2.147)
From this it follows that n2 = n , leaving only the possible values n = 0, 1. This means that
0 = and therefore,
0 = ei ,
(2.148)
= ei ,
(2.149)
(2.150)
This is a mixed state as long as there are pi , 0, 1. One could check and see that in this case
p2 0 0
*. 1 2
+
= . 0 p2 0 // , 0 .
2
, 0 0 p3 2
30
(2.151)
2.6. Microstates
Example
Let us examine a system characterized by the following density matrix:
1
0 0
*. 3 1
+
= . 0 3 0 // .
(2.152)
1
,0 0 3 Here, all the states lz = 1, 0, 1 have equal probability, so hlz i = 0. If we compare this
with a pure state:
1
1 * +
= ..1//
(2.153)
3 1
, leading to the density matrix
1 1 1
+
1 *.
= .1 1 1// .
(2.154)
3
1
1
1
,
-
we can see that both, (2.152) and (2.154), have the same diagonal elements. Consequently,
hlz i = 0, again. They dier, though, in the mean values of the probabilistic observables,
because of the dierent o-diagonal elements.
2.6. Microstates
Back to the quantum-mechanical point of view: provided that | i is a complete orthonormal
basis, then are microstates.
Examples
1. Hydrogen atom:
Here, = (n,l,m,s)
E
.
n 2
(2.155)
n = 1, l = 0, m = 0, s = 1; E = E 0 ,
E0
n = 2, l = 0, 1, m, s; E = ,
4
(2.156)
(2.157)
where N = 2 + 8 = 10. The energy operator is given by E = E 0 diag 1, 1, 14 , . . . , 14 . Let
1
1
the distribution be uniform: = diag 10
, . . . , 10
. Then, the expectation value of the
energy is
1
hEi = 10
2 + 84 (E 0 ) = 0.4 E0 .
(2.158)
31
cos
Lx
mx =
1
2
sin
2
Lx
mx = 1
x
cos
3
Lx
mx =
3
2
2. A particle in a box:
The particle is described by the wave equation
~
= E .
2M
(2.159)
Consider a cubic box with an edge length L. Then, the solution is found to be
=e
=e
ipx x
~
ip x
~
ipy y
~
ipz z
~
0 ,
(2.160)
(2.161)
where
1 2
p .
(2.162)
2M
The periodical boundary conditions give L2 ,y,z = L2 ,y,z (and the same for y and
z). Then px,y,z have discrete values:
E=
p=~
2
m,
L
(2.163)
with
m = (m x ,my ,mz ) ,
m x,y,z Z .
Three examples of these periodical wave functions are illustrated in Figure 2.3.
32
(2.164)
(2.165)
2 ~2 n 2
< E + E .
2M L2
(2.166)
so,
E<
If E E, we can approximate
leading to
1 E
E + E E 1 +
2 E
1 E
= E+ ,
2 E
2ME
L<n<
~
2ME
1 2M
L+
~
2~ E
(2.167)
(2.168)
! 1/2
LE .
(2.169)
! 1/2
E ,
(2.170)
dx p(x ) = 1
(2.171)
dx p(x ) A(x ) = hAi .
dx
2 ,x
dx
2
): mean probability in I (x ),
p(x
(x ): number of states belonging to I (x ),
33
) (x ) ,
p = p(x
(2.172)
I (x )
(2.173)
dx p(x ) A(x )
(2.174)
X
x
! P
I (x ) p A
p
P
I (x ) p
I (x )
| {z }
X
(2.175)
dx p (x )
X X
(2.176)
p A
x I (x )
p A
(2.177)
q. e. d.
P
Note that hA2 i = dx p(x ) A2 (x ) = p A holds only, if the dispersion of A vanishes in the
continuum limit dx 0.
We shall see that p(x ) depends on the choice of the variables x. For that purpose, consider
the explicit coordinate transformation
x = f (x 0 ) ,
df
dx = 0 dx 0 .
dx
(2.178)
(2.179)
dx p(x ) A(x ) =
dx 0
df
p[f (x 0 )] A[f (x 0 )] .
0
dx
| {z } | {z }
p 0 (x 0 )
(2.180)
A0 (x 0 )
34
dx 1 dx n p(x 1 , . . . ,x n ) A(x 1 , . . . ,x n )
(2.181)
..
.
d f
.
0 = det .. ...
.
dx
df 1
, dx n0
(2.182)
(2.183)
df n
dx 10
..
.
+/
// .
/
df n
dx n0
(2.184)
p(E) dE = p(E)
via
(E)
dE ,
E
(2.185)
,
E
E
E
(2.186)
leading us to
p(E) dE = p(E)
dE .
E
(2.187)
Then,
V
2~
p j (E) =
0
V 2M
2~ E
1
2M /2 1
Z
E
!1/2
E .
(2.188)
(2.189)
!1/2
(2.190)
Here,
Z is a constant factor, which could be determined by the normalization condition
dE p(E) = 1:
!1/2
Emax
V
2M
Z=
dE
.
(2.191)
2~ 0
E
35
p = Z 1 e kbT = Z 1 e T .
(2.192)
We have set k b = 1, which means that we measure the temperature in energy units. E is the
energy of state . In Figure 2.4 the Boltzmann distribution is compared to an equipotential one.
We nd
!1/2
E
V 2M
pb (E) =
e T Z 1 ,
(2.193)
2~ E
Emax
hEi =
dE p(E) E ,
(2.194)
0
Emax
2
hE i =
dE p(E) E 2 .
(2.195)
0
It should be marked that p(E) is the macroscopic probability distribution and p is the microscopic one.
L
px .
2~
(2.196)
2~
n < px + px .
L
(2.197)
V
For an arbitrary dimension d we have (p) = (2~)
d p 1 pd , which in the most usual case
of d = 3 is
3
V
=
p1 p2 p3 .
(2.198)
px py pz
(2~) 3
36
=
E
dd p
!
p2
d
E ,
p1 pd
2M
V
=
E
(2~)d
(2.199)
!
p2
d p
E .
2M
d
(2.200)
For d = 1 it is
!
p2
1
E = 2M
d /2 ( E)
2M
0
!1/2
2M
=
,
E
dp
with =
p2
2M .
(2.202)
2ME and
dp = 2
(2.203)
d|p|
=2
1 1
2M /2 .
2
So, we arrive at
L 2M
=
E
2~ E
For d = 3 we have
(2.201)
(2.204)
!1/2
.
(2.205)
d p = 4
d|p| |p| 2 ,
(2.206)
with =
|p | 2
2M .
So, we nd
r
!
2
p
M 1/2
dd p
E = 4
d ( E)
2M
2M
2
0
| {z
} |{z}
|p |
= 4 2 M /2 E /2 ,
3
(2.207)
|p | 2
(2.208)
V
3
1
=
4 2 M /2 E /2 .
3
E
(2~)
(2.209)
37
E 1
e TZ ,
E
where Z is the normalization factor. To nd it, we use
dE p(E) = 1 .
p(E) =
So, we get
Z=
dE
0
where
= F E 1/2 . Therefore,
Z =F
and
E
e T ,
E
1
dE E /2 eE/T = F Z
p(E) = Z 1 E /2 e T
1
(2.210)
(2.211)
(2.212)
(2.213)
(2.214)
So, it is possible to nd an expression for this probability, where ~ does not longer appear.
p( B , A ) B A
(2.216)
( B , A )
XX
p( B | A ) p( A ) B A .
(2.217)
B A
38
(2.218)
(2.219)
1
2,
X
(2.220)
pA A =
hBi =
A
X
hBi hAi =
hB Aic =
1
4
1
2
1
2
pB B =
|{z}
(2.221)
1
2
(2.222)
1
2
(2.223)
1
4
1
4
(2.224)
(2.225)
(2.226)
39
1
2
1
2
1
2
1
2
A
1
2
1
2
B
(a) Two possibilities
But we must keep in mind that this is not true in general! In this case, the correlation function
can be written as
XX
hB Ai =
p( B ) p( A ) B A
(2.227)
B A
p( B ) B
! X
!
p( A ) A
(2.228)
= hBi hAi .
(2.229)
Then,
hB Aic = 0 .
(2.230)
(2.231)
{z
hBAi
} |
(2.232)
{z
hBi
}|
{z
hAi
40
(2.233)
(2.234)
It is A 0 = A and B 0 = B .
3. The result is:
p( B , A ) =
A = A
B = B
p .
One should note that after the measurement of A the relative probabilities
p
p
(2.235)
p 0
p0
(2.236)
In this lecture we will be mostly concerned with classical products of two observables, i. e.
BA = AB, and with classical correlation functions, which are special correlation functions.
However, if the relative probabilities change after the measurement of A, then we have dierent
conditional probabilities: B A , BA. Note that every consistent choice of conditional probabilities p( B | A ) leads to the denition of a product of two observables B A. Many dierent
products and correlation functions are possible. In general, the correct choice depends on the
way how measurements are performed. This choice is important for the statistical description of
systems with few states like atoms (with bounded energy), where the measurement of a Fermi
observable has a strong inuence on the possible outcome of the measurement of a second
observable.
!
AB + BA
,
2
(2.237)
(2.238)
The quantum product coincides with the classical one, in this case. However, if A and B do
not commute, this is no longer correct, because in that case A and B cannot simultaneously be
diagonal.
41
Lattice sites are independent and we say that the probability of a site being occupied is q. So,
1 q is the probability of it being unoccupied. Then
p{s j } =
N f
Y
j=1
1
2
g
+ q 12 s j
= q N (1 q) N N .
(3.2)
(3.3)
N!
N ! (N N )!
(3.4)
(3.5)
43
N sites
Particle 2
(N 1) sites
Particle 3
(N 2) sites
(1) = N
N (N 1)
(2) =
2
N (N 1)(N 2)
(3) =
23
..
.
N!
.
N ! (N N )!
(3.6)
N
X
N!
q N r N N ,
N
!
(N
N
)!
N =0
(3.7)
p(N ) = 1
(3.8)
N =0
by setting r = 1 q.
We conclude that there are two dierent levels at which one can characterize this system.
One can consider the 2 N microstates micro = {s j } or the N + 1 macrostates macro = N .
44
N
X
p(N ) N
(3.9)
N =0
N
X
N!
N q N (1 q) N N
N
!
(N
N
)!
N =0
N
X
N!
q q N r N N
N
!
(N
N
)!
q
N =0
N
X
N!
=q
q N r N N
q N =0 N ! (N N )!
= q (q + r ) N
q
(3.10)
(3.11)
(3.12)
r =1q
r =1q
(3.13)
r =1q
= N (q + r ) N 1q
r =1q
= Nq .
(3.14)
(3.15)
To arrive at this result, we used the following trick: We expressed N via a q-derivative acting
on q N . Thereto we considered r to be independent of q. Only in the end, after using (3.7) and
performing the q-derivative, we substituted r = 1 q again.
3.1.4. Dispersion
The dispersion can be computed using the same trick as above:
hN 2 i =
N
X
p(N ) N 2
(3.16)
N =0
!2
= q
(q + r ) N
q
r =1q
=q
Nq (q + r ) N 1
q
r =1q
f
g
= qN (q + r ) N 1 + (N 1) q (q + r ) N 2
r =1q
2
= qN + q N (N 1)
= q 2 N 2 + N (q q 2 ) .
(3.17)
(3.18)
(3.19)
(3.20)
(3.21)
45
From the expectation value and the dispersion we can calculate the variance
N 2 = hN 2 i hN i2
(3.22)
= Nq (1 q) .
(3.23)
N
=
hN i
(3.24)
(3.25)
(3.26)
1
N
N
1010 !
hN i
(3.27)
This is one of the main reasons why statistical physics has such a predictive power. The
statistical uncertainty is so small that one virtually can calculate deterministically
by using mean values.
46
N
X
1
2 (s j
+ 1)
j=1
1
1
2S + 2 N
(3.29)
(3.30)
and therefore
S = 2N N
(3.31)
= N (2q 1) .
(3.32)
Spin expectation value, dispersion, variance and relative standard deviation are then found to
be
hSi = 2 hN i N
(3.33)
= N (2q 1) ,
(3.34)
hS 2 i = 4 hN 2 i 4 N hN i + N 2 ,
(3.35)
2
S 2 = 4 hN 2 i hN i
(3.36)
= 4 N 2
(3.37)
= 4 Nq (1 q) ,
(3.38)
S
1
,
hSi
N
(3.39)
N +S
2
(1 q)
N S
2
N!
N +S
N S
!
!
2
2
(3.40)
(3.41)
N f
Y
j=1
N
Y
1
2
g
+ q 12 s j
P (s j ) .
(3.42)
(3.43)
j=1
The latter is a product of independent probabilities for each spin s j . That is because we assumed
uncorrelated spins, i. e. the spin s j does not know anything about its neighbours. In the
following this property is used to compute correlation functions.
(3.44)
{s j }
47
Using (3.43) and the fact that the sum over all micro-states can be written as
X Y X
=
j
{s j }
X X
(3.45)
s j =1
s 1 =1 s 2 =1
(3.46)
s N =1
we nd
hsl sk i =
Y X Y
P (s j ) sl sk
j
Y X
j
(3.47)
s j =1
P (s j ) sl sk
(3.48)
s j =1
Y X
P (s j )
X X
j,l,k s j =1
P (sl )P (sk ) sl sk .
j,l,k s j =1
(3.50)
s j =1
to compute
Y X
(3.49)
sl =1 s k =1
P (s j ) = P
sl =1
s k =1 P (sl )P (s k )
(3.51)
sl =1
s k =1 P (sl )P (s k ) sl s k
sl =1
s k =1 P (sl )P (s k )
(3.52)
Notice, that this is the same expression as for a system consisting of only two spins! We conclude
that in the case of uncorrelated spins the spins s j , j , l,k, do not matter for hsl sk i.
By a similar calculation we also get an expression for the expectation value:
P
s =1 P (s k ) s k
hsk i = Pk
.
(3.53)
s k =1 P (s k )
So, obviously
hsl sk i = hsl i hsk i ,
(3.54)
which leads to the observation that the connected correlation function, dened in (2.219) as
hsl sk ic = hsl sk i hsl i hsk i ,
(3.55)
48
(3.56)
This resembles the fact that we are dealing with uncorrelated spins here.
In contrast to that, an example for a system with correlated spins is the Ising model. Its
microscopic probability is given by
P
p{s j } = e
j (s j s j+1 )
(3.57)
In this case there are next neighbour interactions, so the spins know about each other. We will
come back to this later.
(3.58)
If the probability for one spin is P (sk ), the probability for a given sequence is
Y
p{s j } =
P (s j ) ,
(3.59)
hsi =
(3.60)
(3.61)
Of course, there always is some statistical error for a nite sequence, i. e. the value of the total
spin S uctuates between dierent sequences. The dispersion is computed as
S 2 = hS 2 i hSi2 ,
X
hS 2 i =
p{s j } S 2 ,
where
(3.62)
(3.63)
{s j }
S2
and
X
2
sm
(3.64)
q
N
q (1 q) ,
(3.65)
N
1
.
N
s =
(3.66)
(3.67)
(3.68)
49
(3.69)
2
= 4 q (1 q) .
(3.70)
(3.71)
So,
s j = 2
q (1 q) ,
q (1 q)
= s .
(3.72)
50
(x x ) 2
2 2
(3.73)
where x = hxi is the mean value of the macroscopic variable x, = x = x 2 is the variance
and c = c ( ) is a normalization constant which is determined by demanding
1=
dx p(x )
= c ( )
(3.74)
2 2 .
(3.75)
1 1
c ( ) =
.
2
The generalization to several variables x i is given by:
(3.76)
We nd
p = c e 2 Ai j
1
(x i xi )(x j x j )
(3.77)
where Ai j are the components of a positive denite symmetric Matrix A, i. e. a matrix with only
positive eigenvalues. As an example we can think about fermions on a 1-dimensional chain.
We will show that it is the correct large N limit of the binomial distribution
p(N ) =
N!
q N r N N ,
N ! (N N )!
r = 1q.
(3.78)
(3.79)
we get
(N ) = ln p(N )
= N + 12 ln N N + 12 ln (2)
N + 12 ln N + N 12 ln (2)
N N + 12 ln (N N ) + N N 12 ln (2)
(3.80)
(3.81)
+ N ln q + (N N ) ln (1 q)
= N + 12 ln N N + 12 ln N N N + 12 ln (N N )
+ N ln q + (N N ) ln (1 q) 12 ln (2) .
(3.82)
51
(N N 2 )
,
2 N 2
(3.83)
where
N = Nq ,
(3.84)
N = Nq (1 q) ,
2
(3.85)
N + 12
N N +
+ ln (N N ) +
N
N N
1q
N N
= ln
ln
+ O( N 1 ) .
q
N
= ln N
So, up to N 1 corrections we nd
1
2
+ ln q ln (1 q)
(3.87)
(3.88)
1q
N N
=
,
q
N
(3.89)
N = Nq .
(3.90)
and therefore
Moreover, from the second derivative we can infer the variance:
1
2
=
N 2 N 2
(3.91)
N =N
1
1 N
1
1
=
2
+
2
N N N
N (N N ) (N N ) 2 N
N
=
+ O( N 2 ) .
(N N ) N
!
(3.92)
(3.93)
1
N
=
,
N 2
(N Nq) Nq
(3.94)
and therefore
N = Nq (1 q) .
(3.95)
3
N2
=
N 3 N =N
(N N ) 2 N 2
(3.96)
All in all, this leads to the central limit theorem of statistics which states that any uncorrelated
ensemble can be described by a Gaussian probability distribution in the limit N .
52
N 2 = 2N = 2 1020 ,
(3.97)
(3.98)
(N N ) 2
p(N )
= e 2 N 2
p( N )
(3.99)
1010 1040
4 1020
(3.100)
= e
=e
14 1010
10
109
(3.101)
.
(3.102)
N = N (1 + 1010 )
(3.103)
1
p(N )
= e 4 = O(1) .
p( N )
(3.104)
N 5 N . N . N + 5 N
(3.105)
53
4. Equilibrium Ensembles
Equilibrium states are described by statistical ensembles with a static probability
distribution.
If the probability distribution {p } is time independent, then all expectation values of the observables are time independent, too. One example would be an isolated system after suciently
long time or, respectively, just parts of the isolated system. Another one would be an open
system being immersed in a heat bath.
Basic postulate for equilibrium states
An isolated system approaches an equilibrium state. After suciently long time an
isolated system can be described by an equilibrium ensemble.
This conjecture was made by Boltzmann, but was never proven. Actually, it most probably is not
true, as there are two problems with it. First, the necessary time scale could be extremely large.
For example think of systems like water, honey or glass. And second, the microscopic equations
are invariant under time reversal, but the time direction always points towards equilibrium.
But even though the basic postulate may not be strictly true, and despite the existence of
important practical systems which do not equilibrate, the basic postulate is useful for many
applications and it often provides very good approximations.
(4.1)
Or equivalently: An isolated system in equilibrium can be found with equal probability in each
of its allowed states.
55
4. Equilibrium Ensembles
Microcanonical ensemble
The number of microstates with given E and N is denoted by (E, N ). Then the
probability of the system being in one these is given by
p =
1
.
(E, N )
(4.2)
This raises several arguments in favour of the basic postulate for equilibrium states. To begin
with, no state is preferred to another state 0. In addition to that, this so-called equipartition
of states (i. e. equal p ) is constant in time. But one has to be careful. The existence of an
asymptotic state for t is only valid for large enough N . Moreover there is an obstruction
due to conserved correlations, i. e. a system whose initial correlations dier from the ones of
the microcanonical ensemble cannot reach the latter asymptotically.
Note that for a microcanonical ensemble all expectation values and classical correlations for
classical observables are uniquely xed.
1
Z mic
(4.3)
(4.4)
Z mic (E, N ,V ) depends on the thermodynamical variables E, N ,V . Recall that (E) (E) is
the number of states with energy in between E and E + E.
4.1.3. Entropy S
For a microcanonical ensemble the entropy is given by:
S = k b ln ,
(4.5)
where k b is the Boltzmann constant. Setting it to 1 simply means to measure the temperature in
energy units, e. g. eV. Then we have
S (E, N ,V ) = ln (E, N ,V ) .
(4.6)
This equation is valid for isolated systems in equilibrium. The entropy S is an extensive
thermodynamic potential, i. e. it is additive when uniting subsystems. One can conclude that,
in the limit of large N , the entropy is proportional to N :
SN
56
for
N .
(4.7)
Examples
1. Consider N uncorrelated particles where each of them can be in F dierent states. Furthermore, say the energy is given by E = cN , independent of the particles states. Then
there are F N possible microstates.
= FN
(4.8)
ln = N ln F
(4.9)
S = k b N ln F
(4.10)
(4.11)
(4.12)
(4.13)
one nds
S = [N ln N N ] [N ln N N ] [(N N ) ln (N N ) (N N )]
N N
N
= N ln
+ N ln
N !
N N !
N
N
= N ln
1 N ln 1
N
N
"
!
!#
bV
bV
N
= N ln
1
ln 1
.
N
N
bV
(4.14)
(4.15)
(4.16)
(4.17)
Now it is obvious that S is extensive, since one can write it as N times a function of the
particle density n = VN :
S = N f (n) .
(4.18)
For small n we can approximate
"
!
#
b
S = N ln
1 +1 .
n
(4.19)
57
4. Equilibrium Ensembles
(4.20)
(4.21)
E=
n(p)E(p) ,
(4.22)
with E(p) =
p2
.
2M
(4.23)
2M
!
p2
3
= dp 3
E E
p
2M
!
p2
V
3
=
d
p
E
E
2M
(2~) 3
V
3
1
=
4
2 M /2 E /2 E .
3
(2~)
3
=E
(4.24)
(4.25)
(4.26)
(4.27)
E=
n(p)E(p) .
(4.28)
We should discriminate between fermions, for which n(p) {0, 1}, and bosons, for which n(p)
can take arbitrary values. In any case, we can split up the partition sum into two parts:
= 1 + 2 .
(4.29)
Here, 1 is the number of states with n(p 1 ) = 1, n(p 2 ) = 1 for some p 1 , p 2 , and 2 is the
number of states with n(p) = 2 for some p. These two contributions are given by
1 =
1X X
2 p p ,p
1
E=
p 21 +p 22
2M
2 = (b) P
2 = p p 2
E= M
58
(4.30)
for fermions,
for bosons.
(4.31)
The factor 12 in (4.30) is due to the fact that we are dealing with indistinguishable quantum
mechanical particles. So, swapping both particles does not count as a dierent state.
The full partition sum then takes the form
!
1 X X
1 (b)
bosons
=
.
(4.32)
fermions
2 p p E= p 21 +p 22 2 2
2M
1
2
Note that in (4.32) the p 2 sum runs over all possible momentum values, whereas in (4.30) it
excludes p 2 = p 1 .
If we neglect the 12 2(b) term in (4.32) we arrive at the approximate description of a classical
gas. Then there is no distinction between fermions and bosons any more. Performing the two
momentum sums analogously to the one particle case leads to
!2
!
p 21 + p 22
1
V
3
3
=
d p1 d p2
E E .
(4.33)
2
2M
(2~) 3
3) Quantum states for N particles
The extension to the general N particle case is straight-forward:
X
N =
n(p) ,
(4.34)
E=
n(p)E(p) ,
(4.35)
1
=
N!
3
d p 1 d pn
V
(2~) 3
!N
N p2
X
j
j=1
2M
!
E E .
(4.36)
As in the two particle case, equation (4.36) neglects the particular behaviour when two or more
momenta commute. Again, this is the classical approximation where no distinction between
fermions and bosons is made.
4) Partition function
We dene
=
.
(4.37)
E
E
To nd , we have to evaluate a 3N -dimensional integral. For this purpose, we will perform a
change of variables:
p 2j
59
4. Equilibrium Ensembles
Then
d p1 d p N = (2ME)
3
3N/2
d3 x 1 d3 x N ,
(4.40)
X p 2j
2M
X
j
x 2j E .
(4.41)
For our purposes, we also introduce the 3N -component vector y = (x 1,1 ,x 1,2 ,x 1,3 ,x 2,1 , . . . ).
Now, nally, we can calculate :
1
V
=
N ! (2~) 3
!N
(2ME)
1
E
dy1 dy3N
1 N ME
=
V
N!
22 ~2
The factor
3N/2
! 3N/2
E
(4.42)
(4.43)
F=
yi2 E
1
F .
E
X
dy1 dy3N 1
(X ) and
1
E
3N
X
i=1
!
yi2
(4.44)
is the volume of the (3N 1)-dimensional unit sphere S 3N , which is equal to the surface of a
3N -dimensional unit ball. It can be calculated from the general formula
F=
=
3N
2
3N/2
2
3N
2
1 !
(4.45)
3N/2
(4.46)
2
1
2!
/2
3
(4.47)
4 3
= /2
= 4
q. e. d.
Here we have used 12 ! =
partition sum is given by
60
3
2
1
2
2
N!
3N
2
(4.48)
(4.49)
M
2
1 ! 2~
! 3N/2
E
3N/2
VN
E
.
E
(4.50)
Finally, using
E
E
1
3N
, we arrive at
Z mic = =
N!
1
3N
2
M
2
2~
!
! 3N/2
E
3N/2
VN .
(4.51)
N is typically of the order 6 1026 . Therefore, Z mic increases extremely rapidly with E and V !
ln
N
!
ln
!
(4.52)
2
2~2
2
2
(
)
1
M
1
1
1
1 1 3N 1
= k b 3N
ln
+ ln V + ln E ln N + ln
+
.
(4.53)
2 2~2 3
2
3
3 2
2
2
So,
(
S = k b 3N
)
1 V
1 2E 1
M
5
ln + ln
+ ln
+
.
3 N 2 3 N 2 2~2 6
(4.54)
and therefore
S = c kb f
(4.55)
(E) E cf ,
(4.56)
where c depends on the system. So, the partition sum increases rapidly with increasing energy.
This form of the entropy and the partition sum applies too many other systems, too.
It is important to note, that the factor N1 ! in (E) was necessary for the entropy to be an
extensive quantity. This factor is absent in the classical description! It was rst introduced by
Max Planck in 1900 and was heavily discussed back then.
N ,
E .
(4.57)
(particle density) ,
(4.58)
(energy density) ,
(4.59)
61
4. Equilibrium Ensembles
Extensive
Intensive
p (pressure)
n
s = VS (entropy density)
V
N
E
S
as these stay constant in the limiting process. So, mathematically more rigorously we dene
the limiting process via an auxiliary parameter a:
V aV ,
N aN ,
E aE
(4.60)
(4.61)
(4.62)
(4.63)
Some extensive and intensive quantities are listed in table Table 4.1.
Again note, that the entropy is an extensive quantity:
S = 3k bnV c (n, ) ,
(4.64)
where
1 2 1
1
M
5
ln
ln n + ln
+ .
(4.65)
2
2
3n
3
2 2~
6
On the other hand, ratios of extensive quantities, such as n, and s, are intensive quantities.
By setting k b = 1 and ~ = 1 we nd the entropy density to be given by
c (n, ) =
s=
S
V
(4.66)
1/2
1
1
= 3n ln n /3 + ln
+ ln M /2 + const. .
n
(4.67)
So
r
s = 3n ln
!
M
+ const. .
n5/3
(4.68)
In general we have to be careful with units, as we can only take logarithms of dimensionless
numbers. But a quick dimensional analysis conrms this here:
M M1 ,
(4.69)
M ,
(4.70)
nM .
(4.71)
4
3
62
(4.72)
(4.73)
lim S (E) = 0 .
(4.74)
EE 0
and as a consequence
EE 0
EE 0
(4.75)
So, often the assumption is made that S vanishes in this limit, as ln q is tiny compared to the
usually huge f . Then
S
0
(4.76)
V
in the thermodynamic limit.
(4.77)
S (E, N ,V ) = k b ln (E, N ,V ) .
(4.78)
Now we take the derivative with respect to E at xed particle number and volume:
1
S
=
.
T
E N ,V
(4.79)
This denes the temperature T = T (E, N ,V ). It can be used as a new thermodynamic state
variable. So, we can make a variable substitution in order to express the entropy not as a
function of E but of T: S = S (T , N ,V ). Moreover, we see that the temperature is an intensive
quantity, as S and E are extensive and so their ratio must be intensive.
63
4. Equilibrium Ensembles
Pressure
Now we take the derivative with respect to V at xed particle number and energy:
p
S
=
.
T
V E,N
(4.80)
This denes the pressure p. And it can immediately be concluded that the pressure is an intensive
quantity.
Let us know compute this for our ideal gas of particles:
!
S
V
= kb
N ln + f (E, N )
(4.81)
V
V
N
N
= kb
(4.82)
V
p
= .
(4.83)
T
From this we can infer the equation of state for an ideal classical gas:
pV = k b NT .
(4.84)
In thermodynamics T and p are determined by the change of the number of available states
when changing E and V .
Chemical potential
Now we take the derivative with respect to N at xed volume and energy:
S
= T
N
(4.85)
E,V
64
S
V
1
T
S (E,N ,V )
E
N ,V
Reservoir
System
R
S
E
(4.86)
Nr Ns 1 ,
(4.87)
i. e. the reservoir R is much larger than the system S concerning volume as well as particle
number.
The total system can be described by a microcanonical ensemble with total energy
E g = E s + E r = const.
(4.88)
However, E r and E s are not xed individually. More precisely, the total system is described by a
Hamiltonian
H = H s + H r + H int ,
(4.89)
where H int describes the interaction between S and R. We assume this coupling to be weak:
| hH int i | hH s i , hH r i .
(4.90)
Now we denote by |n i the quantum states of S and by | i the quantum states of R. They form
a complete basis, i. e. a quantum state (microstate) of the full system can always be expressed as
= n ,
(4.91)
65
4. Equilibrium Ensembles
E sn = n H s n ,
E r = H r ,
(4.92)
(4.93)
E gn = E sn + E r = E g .
(4.94)
(4.95)
We now consider operators which only depend on n, i. e. they are independent of the properties
of the reservoir:
XX
hAi =
pn An
(4.96)
=
n
X
p n An ,
(4.97)
where
pn =
pn .
(4.98)
In other words, we look at the reduced system S. This restriction to leads to the concept of a
canonical ensemble. Our task is now to compute the reduced probability distribution pn . These
En
considerations will lead us to the famous Boltzmann factor e kbT .
(4.99)
This means that the sum only runs over such which full the constraint (4.94). So, the sum
is just the number of states in R. Therefore pn will depend on E sn and the properties of the
reservoir (Tr ) and we conclude that
pn (E s ) =
r (E r )
.
g (E g ) Er =Eg Es
(4.100)
ln r (E r ) ln r (E g ) + (E r E g )
ln r (E g ) .
(4.101)
E g
| {z } |
{z
}
E s
For
66
Er
Es
We have dened
ln r (E)
,
E
E=Er
(4.102)
where the evaluation at E r instead of E g is ne, as this dierence is a higher order eect in the
Taylor expansion. Note that 0 for monotonously increasing r (E).
So, just depends on the density of states of the reservoir:
=
1 S r
k b E E=Er
(4.103)
1
,
k bTr
(4.104)
(4.105)
Here,
Z can =
g (E g )
r (E g )
(4.106)
(4.107)
is the canonical partition function. It only depends on properties of the reservoir. It can be
computed from the normalisation of pn :
X
pn = 1 .
(4.108)
n
So,
e En .
(4.109)
(E) e E
(4.110)
Z can =
X
E
=
dE (E) e E .
(4.111)
So all in all,
Z can = Z can (, N ,V ) ,
(4.112)
where N and V are the particle number and the volume of S. So, Z can only depends on R via .
67
4. Equilibrium Ensembles
e En
(4.114)
1
1
=
.
T Tr
(4.115)
Then we nd
E = hEi =
(4.116)
pn E n
P En
En
ne
= P
E n
e
n
X
= ln
e En .
(4.117)
(4.118)
E = ln Z can .
(4.119)
c s f s E
(4.120)
,
(4.121)
where f s , the number of degrees of freedom, is usually a huge number. Then, this distribution
describes an extremely sharp peak, as sketched in Figure 4.2.
Say, that w (E) is given by a Gaussian distribution:
E
( E ) 2
w (E) = const. e 2 E 2 .
(4.122)
ln w =
68
(4.123)
(4.124)
E
Figure 4.2.: Energy distribution
cs fs
=
.
E E
E
(4.125)
(4.126)
and therefore
T =
E
,
cs fs
(4.127)
so the temperature can be interpreted as something like the energy per degree of freedom.
The energy dispersion can be inferred from the second derivative:
cs fs
1
2
= 2 .
=
2
2
E
E E
E
(4.128)
So,
E = p
(4.129)
E
= p
.
E
cs fs
(4.130)
cs fs
1
In the thermodynamic limit the relative energy uctuations are negligible. So, we can assume
E = E and associate statistical quantities with actual numbers, this way. This means that we
can actually speak of the energy in a canonical ensemble. Then, microcanonical and canonical
ensemble are equivalent!
69
4. Equilibrium Ensembles
Microcanonical
Canonical
E mic
Z mic (E) = (E)
Z can ( )
E = ln Z can ( ) ,
(4.131)
mic =
ln Z mic (E) .
(4.132)
E
= E can , can we also con conclude mic = can , i. e. Ts = Tr ?
4.2.6. Temperature
Microcanonical ensemble (T = Ts )
mic =
ln
1
=
E N ,V k bT
(4.133)
Here, T is an absolute temperature. So, its units are Kelvin or eV, if we set k b = 1. For example
1
300 K 40
eV. Recall that T is an intensive quantity and that it can be interpreted as an energy
per degree of freedom:
Es
T = 1 =
.
(4.134)
cs fs
In the case of an ideal gas we can compute T from the entropy
S=
via
3N
2E
ln
+ f (V , N )
2
3N
(4.135)
1
S
3N
=
=
.
T
E N ,V
2E
(4.136)
E
E 1
= = T,
3N
f
2
(4.137)
70
(4.138)
Canonical ensemble (T = Tr )
T is the characteristic quantity for the canonical ensemble, i. e. it is a property of the total
system. So to speak, we can forget the concept of the reservoir. T can be measured by
measuring the mean energy hEi in the system.
As a next step, we will show that Tr = Ts .
(4.139)
It is
As said before, w (E) is a sharply peaked distribution, dominated by E max = E.
ln w
E
E=E
=
ln 1 (E)
+
ln 2 (E g E)
E
E=E E
E=E
0=
= 1 2
1
1
=
,
T1 T2
(4.140)
(4.141)
(4.142)
(4.143)
where we used the denition of in the third equation. The minus sign in front of 2 arises
from the derivative of the argument E g E of 2 with respect to E. So, we nd
T1 = T2 .
(4.144)
We conclude, that we can dene the temperature of any ensemble with small uctuations of
E around E via
1
=
ln (E)
.
(4.145)
T
E
E=E
But note, that in general the temperature is determined by the energy dependency of the
partition sum.
71
4. Equilibrium Ensembles
Reservoir
System
R
S
E, N
Figure 4.3.: System connected to a reservoir with energy and particle exchange
N ,E
Canonical:
N ,T
Grand canonical:
,T
The grand canonical ensemble describes an open system S in equilibrium that is able to exchange
energy and particles with a reservoir R, as illustrated in Figure 4.3. For example one can consider
a part of a gas volume.
We will treat it in the same manner as we treated the canonical ensemble. The probability for
a certain state is now given by
1 ( E n + N n )
pn (En , Nn ) = Z gc
e
,
(4.146)
where
ln r (E g , N g ) ,
E
=
ln r (E g , N g ) .
N
We dene the chemical potential as
= .
(4.147)
(4.148)
(4.149)
=
72
dE
dN (E, N ,V ) e (E N ) ,
(4.150)
(4.151)
(4.152)
(4.153)
The averaged quantities can again be extracted via derivatives of the logarithm of the partition
function:
E N = ln Z gc (, ,V ) ,
N =
ln Z gc (, ,V ) .
(4.154)
(4.155)
e E
= (E)
dE e 2 (EE ) + . . .
|
{z
}
q
(4.156)
(4.157)
2
a
where
cf
.
E2
(4.158)
E
2
e E
= (E)
E
cf
| {z }
(4.159)
a=
Then the partition function takes the form
Z can
O(1)
(4.160)
S = ln Z can + E = 1
ln Z can .
(4.161)
Note that, as said before, the dierence between the xed energy E for the microcanonical
and E = hEi = E for the canonical ensemble is negligible. It vanishes in the thermodynamic
limit.
73
4. Equilibrium Ensembles
(4.162)
S = ln Z gc + (E N ) = 1
ln Z gc .
(4.163)
(4.164)
where we had n in the previous cases. The validity of this expression can be checked easily:
In the microcanonical case it is
1
p = ,
(4.165)
so we have
X 1
1
S=
ln
(4.166)
!
1 X
= ln
1
(4.167)
|{z}
= ln
q. e. d.
(4.168)
e E ,
Z can
ln p = ln Z can E ,
p =
(4.169)
(4.170)
so we have
S=
(4.171)
p ln p
p ln Z can + E
(4.172)
= ln Z can + E
q. e. d.
74
(4.173)
(4.174)
This can also be considered as a general denition of the entropy, as it coincides with S = ln (E)
for microcanonical, canonical and grand canonical ensembles.
Most importantly, this expression is valid in any basis. Therefore it also holds when diagonalizing , which proves (4.164) more generally:
X
S=
ln
(4.175)
=
p ln p
q. e. d.
(4.176)
75
5. Thermodynamics
5.1. Thermodynamic potentials
5.1.1. State variables, equations of state
Classical thermodynamics describes equilibrium states and transitions from one equilibrium
state to another. Historically it was developed before statistical physics, but it can be derived as
an eective macroscopic theory from the microscopic statistical theory. Without any doubt it
has an immense importance regarding technical aspects. It is mostly axiomatic, based on the
laws of thermodynamics and their consequences. Classical thermodynamics only uses a few
macroscopic variables (both extensive and intensive) such as E, T , S, N , , V , p. We distinguish
state variables or independent thermodynamic variables, which characterise the ensemble, and
dependent thermodynamic variables. For the dierent types of ensembles (systems) this is
illustrated in Table 5.1
The thermodynamic limit provides a functional relationship between the state variables.
Usually, no average values are used, because hEi E, etc.
There are several practical question, which we will try to answer in this chapter:
How much energy should be deposited in a system, so that its temperature rises by a
certain amount, provided the volume stays constant?
How much will the pressure increase during this process?
How is heat transformed into mechanical energy? What is the eciency of machines?
When is a process reversible and when not?
The dependence of the dependent variables on the state variables is called equation of state.
Examples
1. Equation of state for an ideal gas:
pV = RT n M = k bT N
(5.1)
n M : number of moles
R: gas constant (R = 8.31 molJ K )
This follows from
p
S
= .
V E,N T
(5.2)
77
5. Thermodynamics
State variables
Dependent variables
ln Z mic (E, N ,V )
ln Z can (T , N ,V )
ln Z gc (T , ,V )
S,T , ,p
S,E, ,p
S,E, N ,p
Table 5.1.: Dierent kinds of ensembles (systems) and their corresponding state and dependent variables
(5.3)
nT
1 bn
(5.4)
(5.5)
E = V f E (T , )
(5.6)
N = V f N (T , )
(5.7)
p = fp (T , )
(5.8)
Therefore, (p,T , ) cannot be used as three independent variables at the same time!
If the particle number is not conserved (e. g. photon), there are only two state variables, as
= 0. Then,
p = fp (T ) .
(5.9)
So, there is a relation between pressure and temperature. Dening =
E
V,
we nd
p = fp ( ) ,
(5.10)
p = 13 .
(5.11)
78
k b ln Z mic (E, N ,V )
(entropy)
(5.12)
F = F (T , N ,V ) =k bT ln Z can (T , N ,V )
(free energy)
(5.13)
J = J (T , ,V ) =k bT ln Z gc (T , ,V )
(Gibbs potential)
(5.14)
S
E
S
N
S
V
N ,V
E,V
E,N
1
T
(5.15)
(5.16)
p
T
(5.17)
At this point, these are just denitions. Their physical meaning will be explained later. Note that,
traditionally, in thermodynamics one denotes the xed quantities explicitly. These 3 relations
together with S (E, N ,V ) lead to 4 relations for the 7 variables in total.
We now use
S = ln Z can + E
and =
1
T
(5.18)
to nd
F
E
= ln Z can = S ,
T
T
(5.19)
and therefore
F = E TS .
(5.20)
S = ln Z gc + (E N ) ,
(5.21)
E N
J
= ln Z gc = S
,
T
T
(5.22)
we have
and therefore
J = E N T S
(5.23)
J = F N .
(5.24)
or
79
5. Thermodynamics
S
,
=
T
N E,V
p
S
=
.
T
V
(5.25)
(5.26)
(5.27)
E,N
With these, one arrives at the following expressions for the dierentials of S:
dS =
p
1
dE dN + dV ,
T
T
T
(5.28)
and of E:
dE = T dS +
|{z}
heat
Q
dN
|{z}
added particles
EN
pdV .
|{z}
(5.29)
work
A
for xed V , N .
(5.30)
(5.31)
Again, note that this no property of a system. Neither is it a thermodynamic variable. But it
can be used as a thermodynamic denition of the entropy:
Q = T dS .
80
(5.32)
ds
Fp
Mechanical work
A denition of mechanical work can be given via a slow change of volume at xed S, N (compare
Figure 5.1):
A = Fp ds
(5.33)
= |Fp | ds
(5.34)
= p a ds
(5.35)
= p dV ,
(5.36)
(5.37)
A > 0 :
(5.38)
In general, A describes the change in energy by work at xed S, N . For example, it also
describes the change of an external magnetic eld instead of the change of volume.
We note that the statistical denition of pressure coincides with the mechanical denition
force per area, but the statistical one is more general!
F ( ) = F ( ) .
(5.39)
S
=.
E
(5.40)
(5.41)
81
5. Thermodynamics
and therefore
F = E TS .
(5.42)
F
E S
=E+
E
S E
=E+
|{z}
E
(5.43)
(5.44)
=E,
the original variable, as it should be.
On the other hand, we nd
(5.45)
F
= S ,
T N ,V
(5.46)
dF = SdT + dN pdV .
(5.47)
dJ = SdT N d pdV ,
(5.48)
ln Z gc = N ,
J = ln Z gc ,
so,
J
= N ,
(5.49)
(5.50)
(5.51)
as well as
!
S = 1
ln Z gc
= 1
( J )
J
= J + J + 2
J
=
.
T
Now we have everything at hand to construct thermodynamics:
82
(5.52)
(5.53)
(5.54)
(5.55)
p
1
dE dN + dV
T
T
T
(5.56)
3. Computation of the partition function for the given system S (E, N ,V ), or equivalently
F (T , N ,V ) or J (T , ,V )
(aim of statistical physics, learn counting)
Note that no other than these 3 axioms are needed for thermodynamics! And all fundamental
laws of thermodynamics can be derived from these! We already have shown the latter for the
0th and the 1st law.
(5.57)
(5.58)
(5.59)
(5.60)
We dene E 1 = E and want to nd the condition for the maximum of the probability distribution
W (E), which describes the equilibrium state:
w (E) =
1 (E 1 ) 2 (E 2 )
.
g (E g )
E1 +E2 =Eg
(5.61)
83
5. Thermodynamics
Another property of the entropy can be inferred from this. Let us start with two isolated
systems in equilibrium, with arbitrary energies E 1(0) , E 2(0) and E g = E 1(0) + E 2(0) . They have
entropies S 1(0) , S 2(0) . Then we bring them into thermal contact (i. e. we allow energy exchange).
This leads to an energy ow E 1(0) E 1 , E 2(0) E 2 until S 1 + S 2 is maximal for a given E g . This
means that
S 1 + S 2 S 1(0) + S 2(0) .
(5.62)
Entropy can only increase!
This is a key ingredient of the 2nd law of thermodynamics!
(or S S 0 ,
Using
S0
V
S 0
(5.64)
ln
1
=
,
E E0 0 E
(5.65)
0).
we nd
(5.66)
EE0
,
E
(5.67)
E E f ,
(5.68)
(5.69)
as, choosing E 0 0,
e f .
This means that in the thermodynamic limit the ground state temperature approaches 0:
f
T f ef 0 ,
(5.70)
T (E = E 0 ) = 0 .
(5.71)
84
ln
E
E0
Figure 5.2.: Behaviour of the partition sum at zero temperature
S0
0
0+
85
5. Thermodynamics
Gas 1
(E 1 ,V1 , N 1 )
Vacuum
Gas
(V1 )
Gas 2
Wall
(a) Example 1 (xed volume V1 )
Examples
1. Fixed volume V1 (Figure 5.4a):
V1 : volume occupied by the gas
2. Fixed energy of a subsystem E 1 (Figure 5.4b):
E 1 : energy of gas 1
V1 : volume of gas 1
N 1 : particle number of gas 1
This example can be considered as two isolated subsystems:
g (E,E 1 ,V ,V1 , N , N 1 ) = (E 1 ,V1 , N 1 ) (E E 1 ,V V1 , N N 1 ) .
(5.73)
(5.74)
f i .
(5.75)
(5.76)
(The nal and initial states are equilibrium states. It is not important what happens in between.)
86
Examples
1. Removing the wall:
i V1N
(5.77)
f V N
(5.78)
f i
for V > V1
(5.79)
(5.80)
m 3N1/2
1
E 1
3N
3N
3N
i1
V1 1/2 E 1 1/2
(diluted gas at high energy)
2
E1
N 1 ! 2 1 ! 2~
f i
(i w (E 1 ), f w (E1 ))
E i Ei
f i for E 1 , E 1max = E1 (recall Figure 4.2)
(5.83)
f i
(5.84)
for T1 = T2
(5.81)
(5.82)
1
2
2. Probability that the energy of gas 1 remains E 1 after the wall becomes thermally conductive:
w i = w (E 1 ) E
(5.88)
For the interval [E max 12 E,E max + 12 E] with
E
E
1
N
E 1 .
w (E max ) E = w (E)
it is
(5.89)
87
5. Thermodynamics
Let us look at the eect of removing the constraint on a general variable y (e. g. y = E 1 , the
energy in the subvolume V1 ). Beforehand, it has a xed initial value yi . Afterwards, it satises a
probability distribution characterized by
w (y)y (y) ,
where, in general, y , yi , with y being the position of the maximum of w (y),
generalization to multiple variables y1 , . . . ,yk is straightforward.)
Now we consider the limit
N .
(5.90)
w
y y
= 0. (The
(5.91)
This will lead to a sharp value yf = y = ymax (again, generally yf , yi ). yf is determined by the
entropy becoming maximal ((ymax )):
S (yf ) maximal.
(5.92)
Without constraints the parameters will take such values that the entropy becomes maximal (at
given boundary conditions like V , N , E of the system).
Example
y is the energy in a subvolume:
Beforehand, S (E, N ,V ; y) is calculated with given constraints. Afterwards, we have to
satisfy yS
= 0 to nd ymax .
E,N ,V
wi =
(5.93)
(5.94)
So, from the equilibrium state point of view approaching the equilibrium state corresponds to
the transition from an unlikely to a likely state (content of the rst postulate).
88
Kelvin:
It is impossible to produce work continuously by lowering the temperature of a
system, without other changes in the environment.
In other words, the latter formulation states the impossibility of perpetuum motion machines.
Short summary of the laws of thermodynamics
0th law: T , T1 = T2
1st law: E, conservation of E
2nd law: S, S cannot decrease
3rd law: S min = 0
89
5. Thermodynamics
(a)
(d)
(b)
(c)
V
Figure 5.5.: Carnot cycle
90
Part II.
Statistical Systems
91
n = 0, 1
(6.1)
bosons:
n = 0, 1, 2, . . . ,
(6.2)
(6.3)
This can be used to express the sum over all microstates via the one-particle states:
X X
(6.4)
{n }
X
X
n1
n2
YX
(6.5)
nk
(6.6)
93
This way, every possible sequence appears exactly once! For the dierent types of particles this
yields:
1 !
X Y X
fermions:
=
(6.7)
{n }
bosons:
{n }
n =0
!
X
n =0
(6.8)
As said before, we can relate a sequence of occupation numbers {n } to a function n(p). This
means that we also can identify the sum over all possible sequences with a sum over all possible
functions n(p):
X
D n(p) .
(6.9)
{n }
e (
E n
n )
(6.11)
{n }
n (E )
(6.12)
e n (E )
(6.13)
{n }
XY
{n }
Y X! Y
0
!
(6.14)
n0
Y X
n (E )
n (E )
!
.
(6.15)
The last step may seem confusing, but by explicitly writing it out, it is easy to see that every
summand in (6.14) appears exactly once and with the same weight in (6.15), too.
The logarithm of the partition function is given by
X X
ln Z gc =
ln
en (E ) .
(6.16)
94
ln 1 + e (E ) ,
(6.17)
ln
"X
(E )
j#
(6.18)
j=0
1
,
1x
(6.19)
(6.20)
we arrive at
(b)
ln Z gc
=
ln
e (E )
Now we only have to carry out sums over one-particle states! This is an enormous simplicaP
P
P
tion as it is much easier to perform rather than = {n } ! But note that this just works
for free particles without any interactions.
(6.21)
N p
X N e (E N )
Z gc
1
ln Z gc .
(6.22)
(6.23)
X" 1
#
(E )
ln 1 e
X 1 e (E )
=
1 e (E )
X e (E )
=
,
(E )
1
e
|
{z
}
(6.24)
(6.25)
(6.26)
(b)
and therefore
N =
n(b) ,
(6.27)
95
1
e (E )
(Bose-Einstein statistics)
e (E n n )
n1
and
n =
(6.28)
(6.29)
1
ln Z
.
=
(6.30)
ln
1
1
e (E )
(6.31)
6.3.2. Fermions
As for bosons, the mean occupation number of fermions is given by the sum of the mean
one-particle occupation numbers:
X
N =
n(f) ,
(6.32)
1
(E )
=
ln 1 + e
e (E )
=
.
1 + e (E )
(6.33)
(6.34)
And therefore
n(f) =
1
e (E )
+1
(Femi-Dirac-statistics)
(6.35)
From this it is easy to see that n(f) only can take values in the following range:
0 n(f) 1 .
(6.36)
96
n(f) 0
for E T ,
n(f) 1
for E < 0
(6.37)
and
|E | T .
(6.38)
E=
+
ln Z gc
= ln Z gc + N .
(6.39)
(6.40)
For bosons it is
X (E )e (E )
+ N
1 e (E )
X
X
=
(E )n(b) +
n(b)
E =
(6.41)
(6.42)
E n(b) .
(6.43)
Performing the same calculation for fermions, we nd that both are given by
X
E =
E n .
(6.44)
Similar formulas can be found for other macroscopic quantities, too! Note that n itself already
is a macroscopic quantity!
1
N =
ln Z gc (, = N (, ) ,
(6.45)
= ( N ,T ) .
(6.46)
so we conclude
In the thermodynamic limit the grand canonical and the canonical ensemble are equivalent,
which leads to
= (N ,T ) .
(6.47)
Note that, if there are several conserved quantum numbers Ni , there also are several independent
i .
97
(6.48)
P
But there is no restriction of the summands in {n } , anyway, as N is no conserved quantum
number. Here, the mean occupation numbers characterize the canonical ensemble.
(6.49)
(6.50)
(6.51)
which holds for bosons and fermions, equally. This recovery of the classical description in
the large energy limit is an example of the correspondence principle between QM and classical
mechanics.
We also can write this as
n = C e
E
T
ET
= C e
E
(6.52)
(6.53)
98
Hole
(6.56)
E = E (r ) eit ,
(6.57)
Assuming
it takes the form
2 E (r ) +
2
E = 0,
c2
(6.58)
99
which is solved by
E(r ) = A eik r .
(6.59)
The exact form of boundary conditions is not important, so we choose periodic ones:
k=
2
n
L
with n x,y,z Z .
(6.60)
2
,
c2
= c |k | .
(6.61)
Now, using the fact that the diversion of the electric eld vanishes (no sources):
E (r ) = 0 ,
(6.62)
(6.63)
This means that there are 2 degrees of freedom for every k (massless particle with spin 1 and
helicity H = 1).
Quantisation
For an electromagnetic wave, energy and momentum are given by
E = ~ ,
(6.64)
p = ~k .
(6.65)
2~c
,
L
(6.66)
with n x,y,z Z , H = 1 .
(6.67)
(6.68)
t (n x ,ny ,nz )
(6.69)
i H
(6.70)
into
Note that the translational degrees of freedom are the same for all kinds of particles, e. g. atoms,
molecules.
100
2V
d3p
,
(2~) 3
(6.71)
d3 k
,
(2~) 3
(6.72)
where we dened the volume V = L3 . To arrive at (6.72), we used p = ~k and the fact that E
is independent of H . So, the factor of 2 arises due to the two possible values of the helicity
H = 1.
2
e ~
d3 k
.
1 (2) 3
(6.73)
(6.74)
(6.75)
d =
8
2
d .
(2c) 3 e ~ 1
(6.76)
(k bT ) 4 3
d ,
2 (~c) 3 e 1
(6.78)
101
10
~
k bT
u (,T ) d = 162 ~c
e
2~c
kbT
d
.
5
1
2c
:
(6.79)
Let us look at an example: the spectrum of sunlight. Its maximum is at max 500 nm
which is just what one would expect from a black body radiation spectrum at the sun's surface
temperature T 5800 K. This maximum lies in the regime of green visible light which, as an
interesting side note, is also the maximum of the sensitivity of the human eye!
Furthermore, we can consider the low frequency limit ~ k bT , where the frequency
spectrum can be approximated by
~ 3
d
2c 3 ~
k bT
= 2 3 2 d ,
c
u (,T ) d =
(6.80)
(6.81)
which recovers classical electrodynamics (Einstein 1905). In this limit ~ does not appear, any
more.
The extrapolation to large frequencies ~ k bT , on the other hand, leads to a divergence:
d u (,T ) .
(6.82)
This demonstrates the incompleteness of the theory! So, classical electrodynamics has to
be modied. A solution is given by quantum statistics of photons, which accounts for the
particle-wave duality (for large occupation number we are in the regime of classical elds,
whereas for small occupation numbers we are in the regime of classical particles).
102
d u (,T )
(6.83)
d u (,T )
3
(k bT ) 4
= 2 3 3
d
.
c ~ 0
e 1
|
{z
}
(6.84)
0
(6.85)
4
15
2 (k bT ) 4
.
15 (c~) 3
(Stefan-Boltzmann law)
(6.86)
(6.87)
(6.88)
(6.89)
1 eV = 1.783 10
33
g,
(6.90)
(6.91)
(6.92)
In Table 6.1 some physical quantities and constants and their respective mass (energy) dimensions in natural units are listed.
103
Quantity
Mass dimension
E
T
L,t
V
N
M
M
M 1
M 3
M0
M
M4
M0
M
M0
M0
M0
M0
Z mic,can,gc
F
S
e
Q
1
.
1 e E
(6.93)
1
d 2 ln
1 e
{z
}
(6.94)
ln
this yields
ln Z can
V k bT
= 2
~c
!3
|0
2 3
V ,
45
4
45
(6.95)
where we used natural units in the last step again. Let us check, if this gives the correct result
for the energy density:
E=
=
ln Z can
2 4
V
15
(6.96)
(6.97)
and therefore
E
2
= T4
V
15
which agrees with (6.86), using natural units.
=
104
(6.98)
Remarks
1. Not only E itself is a macroscopic observable, but the energy distribution as a function of
the frequency is one, too.
2. From Z can other quantities, e. g. the pressure of the photon gas, easily can be attained.
thermodynamic potentials classical thermodynamics
= VT 4 ,
3
where we dened
=
or, using natural units,
k b4 2
15c 3 ~3
2
,
15
respectively. Then, the entropy, energy and pressure are given by
!
F
4
S=
= VT 3 ,
T V 3
=
E = F + T S = VT 4 ,
!
F
p=
= T4 .
T V 3
(6.99)
(6.100)
(6.101)
(6.102)
(6.103)
(6.104)
(6.105)
Combining the expressions for E and p and using = VE , we arrive at the equation of state for a
photon gas:
1
p = ,
(6.106)
3
where, as a reminder, the energy density is given by
=
2 4
T .
15
(6.107)
1
.
e (E ) + 1
(6.108)
105
n
1
(6.109)
Thereby, the zero temperature limit of the mean occupation number depends on the relative
size of E and :
k bT 0
n 0
k bT 0
n 1
for E > ,
(6.110)
for E < .
(6.111)
(6.112)
and the Fermi momentum pf or k f , respectively, via the usual energy momentum relation:
f =
pf2
~2k f2
=
,
2M
2M
(6.113)
k 2 <k f2
kf
V
d|k | |k | 2
2 0
V
= 2 k f3 .
3
106
(6.115)
(6.116)
So, using n =
N
V
f =
(6.117)
1
2
(32n) /3 .
2M
(6.118)
(6.119)
For example the Fermi temperature of electrons in copper is approximately 80 000 K. So, we see
that in this case (and also in general) k bT f is a very good approximation.
6.5.4. Thermodynamics
for our purposes the energy E(T ,V , N ) is the most suitable thermodynamic potential. To
calculate it, we perform an expansion in TTf . In the lowest order approximation the energy is
given by the ground state energy:
E = E0 .
(6.120)
=2
V
(2) 3
(6.121)
d3 k
k2
2M
(6.122)
k 2 <k f2
kf
V 1
= 2
d|k | |k | 4
2M 0
V
=
k5
102 M f
3 k f2 N
=
,
10 M
(6.123)
(6.124)
(6.125)
where we used (6.116) in the last step. Also using (6.113), leads us to
3
E0 = N f .
5
(6.126)
(6.127)
107
E
N
= f =
(6.128)
An additional particle always has the energy f ! (not the mean energy of the already present
particles)
So far, we have found the zero temperature equation of state for our Fermi gas:
E(T = 0, N ,V ) =
3
2
(32n) /3 N .
10M
(6.129)
(6.130)
which also considers the interaction with the lattice. Note that denotes the one-particle energy
here, not the energy density VE . It is given by
=
~2
|k | 2
2M
(6.131)
(6.132)
for a relativistic Fermi gas. In principle, this does not change anything regarding the computation
of k f , now we just allow for a general function f (k f ). But nevertheless,
N = i
still holds, as it is a general result.
So, the energy is now computed as
E = i
= i
V
(2) 3
V
(2) 3
V 3
k
62 f
(6.133)
d3k (k ) n(k
)
k 2 <k f2
(6.134)
d
d3k [ (k )] n()
,
(6.135)
and therefore
E=N
d 1 ( ) n(
) ,
(6.136)
d3 k
[ (k )] ,
(2) 3
(6.137)
108
the number of states per particle for given or rather , as sketched in Figure 6.4. (It corresponds
to the quantity (E) for statistics of one particle in the rst part of this lecture, as it has mass
dimension M 1 .) It is an important microscopic property of a specic Fermi gas, as it inuences
the conductivity of metals, for example.
The result is
1 ( ) =
3 3/2 1/2
( ) ,
2 f
(6.139)
(6.140)
(6.141)
(6.142)
109
The method, used for this, is to compute how observables depend on 1 () by leaving 1 () a
free function.
E(T , ) = N
d 1 ( )
e ( )
+1
(6.143)
which we obtain by inserting (6.108) into (6.136). It holds for arbitrary T . So, in principle, this
yields a general equation of state for NE , but only if we know the form of (T ). To compute it,
we start with the general expression for the number of particles:
X
N =
n
(6.144)
=
d n()
i V
d3 k
[ (k )]
(2) 3
(6.145)
=N
d 1 () n()
.
(6.146)
(6.147)
E0 = N
d 1 ()
(6.148)
(6.149)
For T Tf the main contribution to this deviation comes from values close to f . Note that
f = f (n), which follows from the calculation at T = 0, whereas = (T ), as , f for T , 0.
A remark towards nomenclature: If electrons have an energy larger than f , we call them
(excited) particles. If they have an energy less than f , we can call them holes.
110
n
1
f
Figure 6.5.: Smooth Fermi surface
"
d 1 ()
1
e ( ) + 1
( f ) = 0 .
(6.151)
To calculate it, we will perform the so-called Sommerfeld approximation: We perform a Taylor
expansion of 1 () around f :
1 () = 1 ( f ) + 10 ( f ) ( f ) + . . .
(6.152)
= f + ,
(6.153)
x = ( f ) ,
(6.154)
( ) = ( f )
(6.155)
and dene
as well as as the new variable
such that
= x .
(6.156)
This leads to
0 ( f ) x
0=
dx 1 ( f ) + 1
"
#"
1
ex e + 1
(x ) + . . .
(6.157)
1
(1 ) + 1
+1
1
(ex + 1) 1
(6.158)
ex
ex
ex +1
1
ex
+
,
(ex + 1)
(ex + 1) 2
(6.159)
(6.160)
111
yields
0 ( f ) x
0=
dx 1 ( f ) + 1
"
#"
#
ex
f (x ) + x
,
(e + 1) 2
(6.161)
where we dened
f (x ) =
ex
1
(x ) ,
+1
(6.162)
Now, (6.161) is just a linear expression in , but the x-integrations still have to be performed.
f (x ) =
(6.163)
(6.164)
(6.165)
(6.166)
(6.167)
This means that at T , 0 the occupation number for the particles is enhanced in the same extent
as it is lowered for the holes, when compared to the occupation number at T = 0. It is easy to
this by looking at the plot of f (x ) in Figure 6.6.
We also can immediately conclude that
dx f (x ) = 0 .
112
(6.168)
ex
dx x
= 1,
(e + 1) 2
ex
dx x x
= 0,
(e + 1) 2
ex
(ex +1) 2
(6.169)
(6.170)
(6.171)
being even.
10 ( f ) 2
.
(6.172)
2 10 ( f )
(k bT ) 2 .
6 1 ( f )
(6.173)
Let us compute this explicitly for the example of free non-relativistic particles:
3 3/2 1/2
2 f
3 3 1
10 = f /2 /2
4
1 =
Therefore it is
3 1
,
2 f
3
10 ( f ) = f2 .
4
1 ( f ) =
10 ( f ) 1 1
= ,
1 ( f ) 2 f
(6.174)
(6.175)
(6.176)
!2
!
+... .
2
f 1 ( f ) + (k bT ) 2 [1 ( f ) + f 10 ( f )] .
6
(6.177)
(6.178)
(6.179)
2
(k bT ) 2 1 ( f ) N .
6
(6.180)
113
3 He
Tf [K]
T [K]
0.006
12
37 000
1011
4 1012
300
3
300
104
3
,
21 ( f )
(6.181)
which also reproduces the result k bTf = f for free non-relativistic particles, we can bring this
into the form
2 (k bT ) 2
E = E0 +
N.
(6.182)
4 k bTf
This way, one also can consider Tf as a measure for 1 ( f ).
With an expression for the energy at hand we now can calculate the specic heat
E
.
T V ,N
(6.183)
2
T
kb N
,
2
Tf
(6.184)
cv =
It is given by
cv =
which is linear in T ! Indeed, this is in good agreement with experimental results for metals at
low temperatures! (the lattice contribution is c v T 3 ) Moreover, the measurement of Tf allows
us to determine the density of states at the Fermi edge 1 ( f ). This as an example of getting
microphysical information out of the macroscopic thermodynamic behaviour of a system! In
Table 6.2 we have listed the Fermi temperature in comparison to the characteristic temperature
of some materials.
114
1
.
e (E ) 1
(6.185)
We demand n 0. It immediately follows that < E 0 , if E 0 is the lowest energy. Without loss
of generality we set E 0 = 0, i. e. < 0 (otherwise rescale E and by some additive shift, as
p2
only their dierence E matters, which is shift invariant), for example E = 2M .
(6.186)
1
1 1
1
= ,
(6.187)
(6.188)
no condensate,
(6.189)
Bose-Einstein condensate,
(6.190)
) for T 0
6.6.3. Qualitative behaviour of n(E
For smaller T at xed N there are fewer particles at high E due to stronger Boltzmann suppression. On the other hand, of course, this has to be compensated by more particles at small E .
We have plotted this in Figure 6.7. It follows that the low temperature limit of is
T 0
(T , N ) 0 .
(6.191)
J = T ln Z gc = pV .
(6.192)
(6.193)
= T i V
ln
e (E )
d3 k
T1
ln
1
e
(2) 3
~2 k 2
2M
(6.194)
115
n
T2 > T1
T2
T1
E
E0
Figure 6.7.: Occupation number for dierent temperatures at xed N
(6.195)
~2 k 2
T .
2M
(6.196)
or, inserting E ,
1
e (E )
e (E )
(6.197)
for bosons and fermions, equally. So, the classical approximation is only valid for n 1.
d3k T1
e
(2) 3
~2 k 2
2M
1
2 T
d|k | |k | e
~2 |k | 2
2M
(6.198)
,
(6.199)
which, of course, is again the same for bosons and fermions. So,
J = T i
116
V
e T I3 ,
22
(6.200)
where
I3 =
0
dx x 2 ebx
2
dx ebx
0
r !
1
2 b
=
b
1 1 3
= /2b /2 ,
4
=
with
b=
(6.201)
(6.202)
(6.203)
(6.204)
~2
.
2Mk bT
(6.205)
! 1/2
,
(6.206)
! 1/2
.
(6.207)
(6.208)
(6.209)
(6.210)
and therefore
b /2 = 3 (4) /2 .
3
(6.211)
This way, we can express the factor I 3 in the Gibbs potential (6.200) in terms of the thermal de
Broglie wavelength:
I 3 = 22 3 .
(6.212)
117
6.7.4. Thermodynamics
The Gibbs potential is now given by
J = T i V 3 e T ,
(6.213)
J = pV .
(6.214)
as well as
So, the number of particles is computed as
J
N =
1
= J
T
pV
=
,
T
!
(6.215)
T ,V
(6.216)
(6.217)
(6.218)
(6.219)
1.
T
(6.220)
pv
T
J
=
T
= i V 3 e T ,
(6.221)
We know that
N =
(6.222)
(6.223)
leading to
eT =
N 3
,
iV
(6.224)
and therefore
V
= ln i + ln
.
T
N 3
From this we can read o that the approximation is valid if
V
3 (T ) .
N
118
(6.225)
(6.226)
Therefore, recalling that T 1/2 , the classical approximation is valid for a diluted gas at
suciently high temperature:
n 3 (T ) ,
(6.227)
6.8.1. n()
The density is given by
N
n=
=
V
Using
d3 k
(2) 3
1
e
k2
2M
(6.228)
d3 k
1
=
d|k | |k | 2
(2) 3 22
p
1
= 2
d|k | 2 |k | 2
4
(6.229)
(6.230)
and dening
|k | 2
2MT
(T ) 2 |k | 2
=
,
4
z=
this becomes
1 4
n= 2
4 2
! 3/2
and therefore
2
n3 =
dz
0
dz
(6.231)
(6.232)
1
ez
1
.
ez 1
(6.233)
(6.234)
2
1
3
(n )c =
dz z z
.
e 1
|
{z
}
3
2
(6.235)
3
2
119
Using
3
2
1
2
, we arrive at
(n3 )c =
3
2
(6.236)
1/3
3
2
nc /3 ,
1
(6.237)
which, using
2
,
MTc
leads us to an expression for the critical temperature:
Tc =
Using
3
2
2c =
(6.238)
2 3 2/3 2/3
2
nc .
M
(6.239)
1 2/3
n .
(6.240)
M c
So, if we lower T until it reaches Tc for a given n, reaches 0. But what happens for T < Tc ? Is
there no solution?
Tc 3.31
(6.242)
(6.243)
we have a problem, as this diverges in the thermodynamic limit. The macroscopic number of
particles in the ground state is
n =0 N
=
= n0 .
(6.244)
V
V
So, the atoms in the ground
d3 k state are not properly counted in the transition from the discrete
P
sum to the integral (2)
3 . The total number number of atoms is given by
n = n 0 (T ) + n(T
)
(6.245)
where n(T
) is the number of non-condensed thermal atoms. If n 0 , 0 we speak of Bose-Einstein
condensation (BEC).
120
n0
T
Tc
Figure 6.8.: Number of condensed atoms
(6.246)
(6.247)
The qualitative behaviour of this function is plotted in Figure 6.8. At the critical temperature it
is
n 0 (Tc ) = 0
(6.248)
and it occurs a second order phase transition! For n 0 , 0 we also observe superuidity!
Experimental images of a BEC are shown in Figure 6.9.
We can express the number of condensed atoms in terms of the critical temperature. To do
so, we use
!3/2
n(T
)
3 (Tc )
T
= 3
=
,
(6.249)
n(T
c)
(T )
Tc
121
T
n(T
)=
Tc
n,
and therefore
(6.250)
!3/2 #
"
T
n 0 (T ) = 1
Tc
n.
(6.251)
This has consequences for the total energy, too, as only particles with E , E 0 = 0 contribute
if T < Tc :
i M 3/2 T 5/2 V
z 3/2
E=
dz z
.
(6.252)
1
2 2 ~3 |0 {ze }
5
5
2 2
So, using
5
5
2 2 =
3
4
5
2
we arrive at
E
T
0.77
NT
Tc
For comparison, for a classical gas we had
122
E
NT
= 32 .
1.341 ,
(6.253)
!3/2
.
(6.254)
R0
R
!12
R0
2
R
!6 #
,
(7.1)
with R being the distance between two particles. This potential is used to describe Van-der-Waals
forces at long distances, as well as the eect of Pauli exclusion at short distances. We have
plotted it in Figure 7.1a. A simpler approximation would be
V (R) =
V R0 6
0
R
for R < R 0 ,
for R < R 0 .
(7.2)
This describes hard spheres with an attractive interaction. It is sketched in Figure 7.1b.
123
R0
R0
Virial expansion
An important technique, used to describe real gases, is the virial expansion. It is an expansion
of the equation of states in terms of the density n = VN :
pV
= 1 + B 2 (T ) n + B 3 (T ) n 2 + . . . ,
N k bT
(7.3)
where the Bi are the so-called virial coecients. Using the grand canonical partition function,
we nd
p
1
ln Z gc =
V
k bT
= f (n,T )
(7.4)
(7.5)
= n (1 + B 2 (T ) n + B 3 (T ) n + . . . ) .
2
(7.6)
X p12
i
124
2M
X
ij
V (r i r j ,
(7.7)
P
where i j means the sum over pairs of particles. The problem here is that in quantum mechanics
pi and r i do not commute. So, it is not easy to diagonalize H ! Therefore, we will consider the
classical limit for xed N . In this limit, the sum over microstates becomes
X
1 Y
N ! i=1
d3pi
.
(2~) 3
d3 r i
(7.8)
The factor N1 ! takes care of the fact that we are dealing with identical particles. It can be shown
that this is the correct limit by using the Wigner representation of the quantum density matrix.
If we had free particles, then the spatial integration would just lead to a volume factor for
each particle:
N
X
d3pi
1 NY
=
V
.
(7.9)
N!
(2~) 3
i=1
This is an exact result!
Here, Ueff is an average potential seen by one particle, generated by the interaction with all
other particles.
With this replacement the canonical partition function factorizes:
Z can
f
g
2
1 Y d3pi d3r i Pj p2Mj +Pjk V (r j r k )
=
e
N! i
(2~) 3
f 2
g
i +U (r )
1 Y d3pi d3r i p2M
eff i
=
e
.
N! i
(2~) 3
(7.11)
(7.12)
f 2
g#N
p
d3p d3r 2M
+Ueff (r )
e
(2~) 3
"
#N
N
3 Ueff (r )]
= Z ideal V
dre
1
N!
"
(1) N
= Z ideal Z ww
,
(7.13)
(7.14)
(7.15)
125
with Z ideal being the canonical partition function for an ideal gas. Using this, the free energy
becomes
F = k bT ln Z can
(7.16)
= k bT ln Z ideal N k bT
= F ideal N k bT
(1)
ln Z ww
(7.17)
(1)
ln Z ww
,
(7.18)
F ideal
!3/2 # !
,
(7.19)
N 3
mk bT
N k bT ln
.
V
2
2~2
(7.20)
(1)
7.3.4. Computation of Z ww
(1)
To compute this deviation Z ww
from the ideal gas partition function, we have to consider two
eects:
1. excluded volume,
2. attractive interaction.
It is
(1)
Z ww
1
=
V
d3r e Ueff (r ) .
(7.21)
1
V Vex
d3r e Ueff (r ) .
(7.22)
V Vex
126
V Vex W
e
.
V
(7.23)
1
N (N 1)
2 {z }
|
4 3
a ,
3
|{z}
(7.24)
with a being the minimal distance between two particles, i. e. a = R 0 in Figure 7.1b. So, we nd
1 4 3
N
a
2
3
2 3
=
a N
3
= b0 N ,
Vex
(7.25)
(7.26)
(7.27)
2 3
a .
3
(7.28)
1
N (N 1) u ,
2
(7.29)
(7.30)
and therefore
1
W = N u
2
2N
=
dR R 2 W (R) .
V Vex a
(7.31)
(7.32)
(7.33)
Note that W n!
127
Altogether, we arrive at
V b0 N kWT
e b ,
V
V b0 N
W
= ln
V
k bT
V b0 N
N
= ln
+
b1 .
V
V k bT
(1)
Z ww
=
(1)
ln Z ww
(7.34)
(7.35)
(7.36)
(7.37)
(7.38)
So, we arrive at
F = N k bT + N k bT ln
N
3
mk bT
N2
N k bT ln
b
.
1
V b0 N 2
2~2
V
(7.39)
F
,
N N ,T
(7.40)
!2
.
(7.41)
Note that here only the structural form is important. The computation of b0 and b1 can be
improved by a self-consistent computation of Ueff (r .
Often, it is more useful to express these equations just in terms of the density n = VN and T .
So, instead of F we look at VF :
F
n
3
mk bT
= n k bT + n k bT ln
n k bT ln
b 1n 2 ,
V
1 b 0n 2
2~2
(7.42)
nk bT
b 1n 2 .
1 b 0n
(7.43)
This is the van der Waals equation. One nds it to describe the phase transition from gas to
liquid qualitatively correct!
128
S
T T ,N
2 F
= T 2
T
cV = T
(7.44)
(7.45)
V ,N
S
,
=V
lnT V n
with
S
V
(7.46)
(7.47)
n
3
mk bT 3
+ k bn ln
+ k bn .
1 b 0n 2
2~2 2
(7.48)
So,
!
S
5
n
3 mk bT
= k bn
ln
+ ln
,
V
2
1 b 0n 2
2~2
which is independent of b1 .
This leads to
cV
s
=
V
lnT V
!
= 3 k n ,
b
2
n
(7.49)
(7.50)
3
kb N .
2
(7.51)
So, regarding the specic heat there is no dierence between a van der Waals gas and an ideal
gas!
3
kb N ,
2
cV = 3 k b N .
cV =
(7.52)
(7.53)
In general, it is
cV =
1
kb N f1 ,
2
(7.54)
129
with f 1 being the number of excited degrees of freedom per particle. The total number f excited
degrees of freedom is f = N f 1 . So, for example one nds:
Monatomic gas:
f1 = 3
(translations),
(7.55)
Phonons:
f1 = 6
(7.56)
Diatomic molecule:
f 1 = 3 + 2 + (2)
Regarding the diatomic molecule: note that one (longitudinal) vibration corresponds to f 1 = 2.
Alternatively, one can split up its total f 1 = 7 into 6 velocity components and the distance
between the atoms.
The relation (7.54) is always true, if the quadratic approximation for the energy holds:
X
E = E0 +
b q 2 .
(7.58)
Here, the q are the excited degrees of freedom of the system and the b are suitable prefactors.
P
labels the excited degrees of freedom and therefore = N f 1 . We denote the non-excited
degrees of freedom by q.
So, for a general degree of freedom qr one nds qr {q , q}.
P
2
P
Dqr (E 0 + 0 b 0q 2 0 ) e E0 b q
,
P
2
Dqr e E0 b q
(7.59)
where
Dqr =
dq 1 dqr dq N
(7.60)
Dq Dq .
(7.61)
It is easy to see that the Dq integration and the factor e E0 in enumerator and denominator
cancel. Moreover, we can pull the E 0 out of the remaining Dq integration. We end up with
b q 2
dq e
hEi =
Q
130
!
P
2
0 b 0q 0
b q 2
dq e
(7.62)
b 0 q 2
dq 0 b 0q 2 0 e 0
hEi E 0 =
b 0 q 2
dq 0 e 0
0
!
X
b 0 q 2
=
ln dq 0 e 0
0
!
r
X
ln
b 0
0
| {z }
X
(7.63)
(7.64)
(7.65)
12 ln + const.
X 1
=
.
2
0
(7.66)
And therefore,
1
k bT N f 1 .
(7.67)
2
Note that we implicitly assumed a classical treatment, i. e. we neglected the discreteness
of energy levels. This is only valid if the dierence between the ground state energy and the
energy levels in the -channel is small compared to k bT :
hEi = E 0 +
k bT E 1, E 0 .
(7.68)
131
8. Magnetic systems
8.1. Magnetization
8.1.1. Degrees of freedom
We consider a lattice with points x. In this case it is a cubic lattice:
x = na ,
(8.1)
n = (n 1 ,n 2 ,n 3 )
with nk Z ,
(8.2)
where a is the lattice spacing. There are classical spins a (x ) at each lattice site x which satisfy
3
X
a (x ) a (x ) = 1 .
(8.3)
a=1
The microstates of this system are given by the possible spin congurations a (x ), the
so-called lattice elds. Moreover, we have a Hamiltonian
H = H [a (x )] ,
(8.4)
with a = 1, . . . , N ,
a (x ) a (x ) = 1 .
(8.5)
(8.6)
a=1
One example of such a system is the Ising model, which is characterized by N = 1 and
(x ) = 1 (discrete number of microstates). So, its microstates are given by
: { (x )} {s j }
with s j = 1
(8.7)
e E
e H
=
,
Z
Z
(8.8)
as we are describing it as a canonical ensemble. Note that we are not dealing with uncorrelated
probabilities as in chapter 3, any more!
133
8. Magnetic systems
e2
e1
Figure 8.1.: Unit lattice vectors
8.1.2. Hamiltonian
We are considering the following Hamiltonian:
XXX
XX
H =
a (x ) a (x + e ) c B
a (x ) Ba .
x
{z
(8.9)
H int
(8.10)
e is the unit lattice vector in -direction, as shown in Figure 8.1. Of course, the sum over
runs from 1 to the dimension d of the lattice (d = 3 for a cubic lattice):
X
d
X
(8.11)
=1
So, the properties of H int are that it only considers next-neighbour interactions and that it tries
to align the spins. On the other hand, B is a magnetic eld which tries to align the spins in the
same direction as B.
The ground state of the system is characterized by the limit . In this limit the spin
expectation values will be proportional to the magnetic eld:
ha (x )i Ba .
(8.12)
8.1.3. Magnetization
The magnetization is dened as
Ma =
ha (x )i .
(8.13)
Since Ba is an intensive quantity, Ma is an extensive one. Now, the question is: How does the
magnetization depend on the magnetic eld?
M (B) = ?
134
(8.14)
8.1. Magnetization
(8.15)
From (8.9) we can read o that it includes a source term, which is linear in the eld (x ).
Compenentwise, the corresponding source eld is given by
ja = c B Ba .
(8.16)
So, instead of considering the partition function as B-dependent, we also can consider it as
j-dependent:
Z (T ,B) = Z (T , j) ,
(8.17)
which is a mere change of notation. But we will stick to the notation using j, as it is the same
notation as in QFT. Note that the volume V and the particle number N do not play a role here
they are no thermodynamic variables!
Again, we can perform the transition from a sum over states to a functional integral:
! X
#
X Y" Y
da (x )
a (x ) a (x ) 1 .
(8.18)
On the other hand, for the discrete Ising model (N = 1) we just have
X Y X
=
.
(8.19)
(x )=1
We dene
W (T , j) = ln Z (T , j) ,
(8.20)
(8.21)
This way, the basic relation for the magnetization takes the form
W
= Ma
ja
(8.22)
W
1 Z
=
ja
Z ja
1 X Hint +Px
=
e
Z ja
*X
+
=
a (x )
(8.23)
a (x ) j a (x )
(8.24)
(8.25)
= Ma .
(8.26)
135
8. Magnetic systems
Therefore, one also can compute the magnetization from the free energy:
M
F
,
B
(8.27)
(8.28)
MN,
(8.29)
(Gibbs potential).
(8.30)
W (T , j)
as thermodynamic potential,
(8.31)
T,j
as thermodynamic variables.
(8.32)
F J
So, in the following we will use
g
1 f
W (j,T ) + Ma ja ,
N
(8.34)
W (j,T )
+ a ja ,
N
(8.35)
U
= ja .
a
(8.37)
This is the basic relation between magnetization and magnetic eld, which can be proven by
direct computation:
!
U
jb
W (j,T )
= ja +
b
.
(8.38)
a
a
jb N
| {z }
b
136
(8.39)
U
= 0.
j
(8.40)
Ma = N 0,a .
(8.41)
N U = U jM
(8.42)
= W
F
= ,
T
(8.43)
(8.44)
where F (T , j; M ) is the free energy with the constraint that M takes the given value (by
inverting M (j), one gets the free energy F (T , j) in standard form). This is similar to the entropy
of a system with constraint y (S (y) maximal F (y) minimal).
What is the probability to nd a given M? To answer that, we need the probability distribution
w (M ):
w (M ) =
X
P
e H [a (x )]
x
(M ) e E (M )
S (M ) T1
=e
)
F (M
T
=e
(8.45)
a (x )=M a
E (M )
(8.46)
(8.47)
(8.48)
where we used F = E T S in the last step. From this we conclude that the maximum of w (M )
corresponds to the minimum of F (M ) and therefore to the minimum of U (). For j = 0 this
then corresponds to a minimum of U :
U
= 0.
a
(8.49)
137
8. Magnetic systems
O (N )-symmetry
a (x ) and ja (x ) are N -component vectors. So, H int is
invariant under N -dimensional rotations, i. e. SO (N )-invariant,
invariant under a (x ) a (x ) (internal reection).
From a group theoretical point of view the symmetry group SO (N ) together with reections
P
yields the symmetry group O (N ), which describes all linear transformations that leave a (a a )
invariant. Therefore, H int is O (N )-invariant.
But the source term breaks this O (N )-symmetry for xed j , 0. So, only simultaneous rotations
of a (x ) and ja (x ) leave H invariant. From this we can conclude that W (j,T ) is invariant under
O (N ) rotations of ja and that, as a also is a vector under rotations, U ( a ) is invariant under
O (N ) rotations of a .
This symmetry of U (N ) leads to a simplication of the computations. As already said, U is
invariant with respect to O (N )-symmetry and the explicit breaking by a magnetic eld appears
only in the source term. Therefore, the solution of
U
= ja
a
(8.50)
is not O (N )-symmetric for a given ja , 0. This means, that the magnetic eld singles out a
particular direction.
(x ) = 1
for all x, = 1 ,
(8.51)
or
(x ) = 1
for all x, = 1 .
(8.52)
8.2.3. Domains
A domain is a continuous region of spins pointing in the same direction. In Figure 8.2 we
sketched three neighbouring domains. At the boundary between two domains there is a positive
138
domain
+ domain
But what wins? p or the number of contributions? We expect, that for low T the system will
have near ground state order with , 0 (SSB), and that for high T the system will be completely
disordered with = 0 (SYM). There will be a phase transition between these two cases at the
critical temperature Tc .
Note that in practice domains are visible as Weiss domains in ferromagnets.
(8.55)
where U0 is irrelevant for the magnetization, as it drops out when performing the derivative:
U
= ja .
a
(8.56)
U
U
U
=
=
a .
a
a
(8.57)
139
8. Magnetic systems
U
= 2 a + a
a
= ( 2 + 12 b b ) a .
(8.58)
(8.59)
2
,
0 0 =
2 2
.
(8.60)
( a a ) 2 .
8
(8.61)
Because | 0 |( 2 ) is a continuous function, as one can see in Figure 8.6, we speak of a second
order phase transition.
Critical temperature
2 = 2 (T ) ,
(8.62)
= (T )
(8.63)
< 0,
(8.64)
(T large) > 0 .
(8.65)
2 (T = Tc ) = 0 ,
(8.66)
i. e. it is the temperature where the phase transition occurs. The resulting phase diagram is
sketched in Figure 8.7.
140
U
2 > 0
U
2 < 0
U
2 = 0
141
8. Magnetic systems
| 0 |
SSB
SYM
T
Tc
Figure 8.7.: Phase diagram of Landau theory
(8.67)
which we have plotted in Figure 8.8 for dierent cases of j So, we are looking for the solutions
of
U
0=
= 2 + 12 3 j .
(8.68)
142
U = U j
j>0
j=0
j<0
0
j
,
B T
(8.69)
.
j
(8.70)
(8.71)
By inverting it:
1 =
143
8. Magnetic systems
2
Figure 8.10.: Magnetic susceptibility for j = 0
and using j =
, we nd (j = 0)
1 =
2U
.
2 =0
(8.72)
1 = 2
1 = 2 + 32 2
=0
2
= + 3 0 .
1
,
2
(8.73)
(8.74)
(8.75)
1
1
= 2.
2 0
2
(8.76)
So, by combining (8.73) and (8.76), we can plot the magnetic susceptibility for j = 0 over the
complete 2 range in Figure 8.10. We see that it diverges at the phase transition ( 2 0, T Tc ).
For j , 0 we would nd a nite susceptibility at Tc .
N >1
In this case, we have a susceptibility tensor:
a
.
jb
(8.77)
2U
,
a b
(8.78)
ab =
Dening the mass matrix
2
mab
=
144
(8.79)
a U
=
jb c b
a jb
=
jb c
a
=
c
= ac .
!
(8.80)
(8.81)
(8.82)
(8.83)
2U
,
a b 0
(8.84)
with
U = 2 + 12 2 .
(8.85)
2
mab
= 2 ab ,
(8.86)
(8.87)
Take
0,1 = 0 ,
0,a,1 = 0
0 = 12 02 ,
(8.88)
for a , 1, b , 1,
(8.89)
which leads to
m211 = 2 0 ,
1
mab
=0
and therefore
2 0 0 0
*.
+
0
0
0//
.
2
m = .. .
. . .. // .
. ./
. ..
0 0, 0
(8.90)
(8.91)
(8.92)
This means that in the SSB- phase the susceptibility tensor has N 1 diverging eigenvalues for
j 0, 0 const. , 0 (N diverging eigenvalues for j 0, 0 0).
145
8. Magnetic systems
( 1 + i 2 ) ,
(8.93)
with
= = = 12 ( 12 + 22 ) .
(8.94)
In this case the system experiences superuidity in the SSB-phase, which is related to the
Goldstone mode.
The case of N = 4 corresponds to two complex elds. Its relativistic version is the Higgs
mechanism (Goldstone bosons become longitudinal components of massive vector bosons in
presence of local gauge symmetry).
146
9. Phase transitions
9.1. Order of phase transition
We speak of phase transitions when we observe non-analytic behaviour of thermodynamic
potentials or their derivatives, while varying thermodynamic variables.
(9.1)
Thereby, 2 (T ) increases monotonically and 2 (Tc ) = 0. For temperatures T near the critical
temperature Tc it behaves like
2 = a (T Tc ) + . . . ,
(9.2)
with some constant a. From this the behaviours of 0 and 0 follow directly:
2 (T )
a (Tc T )
0 (T ) =
=
,
q
p
0 (T ) = 2 0 (T ) = c Tc T ,
(9.3)
(9.4)
147
9. Phase transitions
where c =
2a
.
In general, one nds 0 to behave like a power law near the critical temperature:
0 (Tc T ) ,
(9.5)
with being a so-called critical exponent. In this case = 12 (mean eld value).
For our example the derivative of 0 with respect to T is given by
0
1
.
T
Tc T
(9.6)
(9.7)
U = 18 ( a a ) 2 ja a .
(9.8)
and therefore
What is the form of 0 (j) in this case? To compute it, we take ja = j a1 , leading to 0,a = 0 a1 .
Then, we nd
U
j=
= 03 ,
(9.9)
1 2
and therefore
0 j /3 ,
1
M B /3 .
1
(9.10)
(9.11)
with being another critical exponent. In this case = 3 (mean eld value).
The derivative of M with respect to B is given by
M
2
B /3 ,
(9.12)
148
w
T < Tc
w
T = Tc
w
T > Tc
(9.13)
149
9. Phase transitions
Magnetic systems
Real gas
j
U
U
v
p
f
!2
,
(9.14)
with b0 , b1 being the so-called virial coecients. The free energy F is related to the pressure via
F
= p ,
V
(9.15)
N
3 mk bT
1 ln
V b0 N
2
2~2
V
N
b1
1
n
N2
.
V
(9.16)
f
= p ,
v
(9.18)
(v,T
,p) = f (v,T ) + pv .
(9.19)
and
plays the role of U for magnetic systems, i. e. the maximal probability w (v) corresponds to
the minimum of (v):
= 0.
(9.20)
v T ,p
In Table 9.1 we have listed the corresponding quantities of magnetic systems and real gases.
But note that there is one important dierence: there is no O (N )-symmetry for the real gas.
150
Liquid
Gas
Liquid
Gas
v
(a) T < Tc
(b) T > Tc
Figure 9.2.: Eective potential 1 above and below the critical temperature
n
Liquid
Gas
T
Tc
Figure 9.3.: Jump in density during liquid-gas transition
into a v-dependent
and a non-v-dependent part:
(v)
= 0 (T ) + 1 (v,T ,p) ,
(9.21)
with (k b = ~ = 1)
b1
+ pv ,
v
(9.22)
1
T
b1
=
+ 2 +p .
v
v b0 v
(9.23)
1 = T ln (v b0 )
leading to
0=
We nd that the equation of state has three solutions for given T , p (2 minima, 1 maximum),
as sketched in Figure 9.2. For T = Tc both local minima have equal 1 . Moreover, there is a jump
in density at T = Tc , as shown in Figure 9.3.
151
9. Phase transitions
9.2.3. Compressibility
The isothermal compressibility T is dened as
p
1
= V
.
T
V T ,N
(9.24)
Note that V < 0, as larger pressure corresponds to smaller volume. Therefore T > 0 (if T < 0,
this would be an unstable situation).
So, in terms of intensive variables,
p
1
= v
,
T
v
and therefore
T
b1
=
+ 2
v T v v b0 v
(9.25)
!
.
T
(9.26)
!
2
T
b1
=
,
+
v 2 T ,p v v b0 v 2 T
(9.27)
2
1
= v 2 .
T
v v0
(9.28)
By comparing this to
we nd
(9.29)
= f + pv 0
1
= (F + pV )
N
1
= (F JGibbs )
N
1
= (N )
N
=.
(9.30)
JGibbs = F N = pV
(9.35)
(9.31)
(9.32)
(9.33)
(9.34)
Recall that
152
(9.36)
We are considering chemical reactions where N atoms of a given species can be bound either
in molecule A or molecule B. The corresponding variable is the concentration, e. g.
cb =
#molecules B
.
#molecules A + #molecules B
(9.37)
N (c b )
T
v
v2
= 0, from which we get the equation of state and the solution for the two minima v 1 ,
as a function of p,
2. (v
1 ) = (v
2 ), which is equivalent to l = g (liquid-gas coexistence condition) and from
which we get the vapor pressure curve pc (T ).
So, there is only one free parameter (for xed N , V ), e. g. T or p:
given p
Tc ,
(9.38)
given T
pc .
(9.39)
9.3.2. p T diagram
The free enthalpies of the two phases are
G l = N l l (p,T ) ,
G g = N g g (p,T ) .
(9.40)
(9.41)
l > g :
(9.42)
l = g :
transition line (N l , N g , 0, N l + N g = N ).
(9.43)
The transition line is described by p = pc (T ), which we have sketched in Figure 9.4. So,
coexistence in equilibrium is possible. Note that this is a universal curve, as it is independent of
Nl , Ng .
153
9. Phase transitions
p
l = g
Liquid
l < g
Gas
l > g
T
Figure 9.4.: p T transition line between liquid and gas phase
(9.44)
1 (T + dT ,p + dp) = 2 (T + dT ,p + dp) ,
(9.45)
d 1 = d 2 ,
(9.46)
to arrive at
with
d i =
i
i
dT +
dp .
T
p
|{z}
|{z}
S
N
(9.47)
V
N
(9.48)
When expressing it in terms of intensive quantities and combining it with (9.46) we arrive at
v 1 dp s1 dT = v 2 dp s2 dT ,
(9.49)
(s s ) dT = (v 2 v 1 ) dpc ,
| 2{z 1}
| {z }
(9.50)
(9.51)
It tells us that the slope of pc (T ) is given by the ratio of entropy dierence and volume dierence.
154
dT
T V .
dpc
(9.53)
The latent heat is necessary to separate the water molecules during the transition from the
liquid to the gas phase or to break up the crystal structure during the transition from the ice
(solid) to the liquid phase, respectively.
(9.55)
(9.56)
In this case, the system has not enough time for the transition B A, but instead stays in the
local equilibrium B even beyond the critical point. This phenomenon is called supercooling
(one speaks of superheating if during the transition A B the systems stays in A beyond the
critical point). Eventually, the local instability will become extremal when B becomes a saddle
point and nally a maximum. In practice one can observe the coexistence of water and vapor
not only at Tc .
For magnets the same eect appears in the form of hysteresis. Indeed, this is a characteristic
eect for rst order transitions in general.
155
9. Phase transitions
w
T > Tc
T < Tc
A
Liquid
B
Gas
G
T < Tc
156
(10.1)
with c reg being the simplest approximation expression for the specic heat, being a critical
exponent, and A being some constant. We have plotted (10.1) in Figure 10.1 for dierent cases
of .
(10.2)
for T < Tc (compare (9.5)) and M = 0 for T > Tc . Here, is another critical exponent.
10.1.3. Susceptibility
The susceptibility in magnetic systems corresponds to the compressibility T in uids. At
B = 0 / p = p both behave the same close to the critical temperature:
|T Tc | ,
(10.3)
(10.4)
T |T Tc |
This hold for T > Tc as well as for T < Tc . They both share the same critical exponent .
157
(a) < 0
(b) > 0
(10.5)
(10.6)
with some critical exponent . For magnetic systems we already found this relation in (9.11).
A (T Tc ) 2 + 4 + U0 ,
2
8
(10.7)
= A (T Tc ) + 3 ,
2
2
U
3
= A (T Tc ) + 2 .
2
(10.8)
(10.9)
The rst derivative also directly gives us the position of the minimum for SSB:
1 2 A (Tc T )
=
,
2 0
158
(10.10)
1)
The specic heat is dened as
c = T
2 F
,
T 2
(10.11)
where
F = T N U ( 0 ) ,
(10.12)
T U 0
F T U ( 0 ) =
T U 0
1 A 2T
2
(Tc T ) 2
for SYM,
for SSB.
(10.13)
c reg
c
c reg
A 2T
with
c reg T 2
(3T 2Tc )
for SYM,
for SSB,
2U0
U0
+T
.
T 2
T
(10.14)
(10.15)
There appears no T Tc in the expression for c, so = 0. Or, in other words, c has no pole at
T = Tc . But there is a jump in specic heat at the phase transition:
A2Tc
c sym c ssb =
.
Tc
(10.16)
0 (Tc T ) /2
(10.17)
2)
From
1
we can read o = 12 .
3)
By combining (8.72) and (10.9) we nd
2U
!1
(T Tc ) 1 .
phi 0
(10.18)
So, it follows = 1.
159
4)
At T = Tc the eective potential is given by
U =
Moreover, we know that
4
.
8
U
( 0 ) = j .
(10.19)
(10.20)
=
,
(10.21)
1
2 = + 2 + .
(10.22)
The concepts which are necessary to understand where the scaling relations come from and to
understand this universal behaviour of systems with equal symmetries, in general, have been
developed by Kadano, Wilson and Wegner. We will introduce these concepts in the following
sections.
160
Theoretical results:
Mean eld
0.5
Ising model in d = 2
0.125
Renormalization group in d = 3
n=1
n=3
0.11
0.1159
1
1.75
3
15
0.3249
0.3647
1.2371(4)
1.3866
4.817
4.802
Experimental results:
Liquid-gas / uniaxial magnet (n = 1)
0.1
0.34
1.2
4.4
Isotropic magnet (n = 3)
0.15
0.36
1.25
4.3
(10.25)
with r = r 1 r 2 . Often, the correlation function shows a very typical behaviour for large r :
r
G (r ) e
(10.26)
M = 0,
|M | > 0
(10.27)
(10.28)
161
in magnetic systems at B = 0. To make this transition possible, the spins have to know of
each other, i. e. they have to be correlated. But not only that, the correlation length has to be a
function of temperature, too, to explain these dierent behaviours above and below the critical
temperature.
(10.29)
where j (r ) are articial local sources. They resemble an arbitrary inhomogeneous magnetic
eld. Moreover, H [(r )] and Z [j (r )] are functionals, i. e. functions of functions. Note that they
are not just functions of some constant or j.
Now, the so-called generating functional is dened as
W = ln Z .
(10.30)
P
1 X
+
(r ) (r 0 ) e H + j
Z
D
E D
ED
E
= (r ) (r 0 ) (r ) (r )
Df
D
Eg f
D
Eg E
= (r ) (r )
(r 0 ) (r 0 )
=G (r ,r 0 ) .
(10.31)
(10.32)
(10.33)
(10.34)
(10.35)
(10.36)
(10.37)
G (r ,r ) = G (r r ) .
0
(10.38)
162
(10.39)
j,
(10.40)
with j being independent of r . Then, the rst two derivatives of W with respect to j yield
E
W X D
=
(r )
j
r
= N ,
(10.41)
(10.42)
g f
gE
2W X Df
(r 0 )
=
(r
)
j 2
r ,r 0
X
=
G (r ,r 0 )
(10.43)
(10.44)
r ,r 0
=N
G (r , 0)
(10.45)
G (r ) ,
(10.46)
=N
X
r
where we used translational and rotational symmetry in the last two steps.
On the other hand, the susceptibility is dened as
j
1 W
=
j N j
1 2W
=
.
N j 2
(10.47)
(10.48)
(10.49)
So, the susceptibility and the correlation function are related via
=
G (r ) .
(10.50)
While we have considered a magnetic system, so far, these calculations can be performed
completely analogously for uids. In that case, one ends up with a relation between the
compressibility and the correlation function:
1
T = 2
nT
d3r G (r ) .
(10.51)
163
So, using the relations between susceptibility / compressibility and the correlation function,
we can conclude that and T are nite as long as remains nite. But as we have seen in
section 8.4, the susceptibility diverges at the critical temperature (e. g. recall Figure 8.10). The
same holds for the compressibility. We can immediately infer that the correlation length has to
diverge at T = Tc , too:
T Tc
,T ,
T Tc
(10.53)
(10.54)
(10.55)
In the mean eld approximation one nds = 12 . Moreover, the large r limit of G at T = Tc can,
in general, be expressed in the following way:
lim G (r ) r (d 2+) ,
(10.56)
with the so-called anomalous dimension being yet another critical exponent. Its mean eld
value is = 0.
10.3. Universality
10.3.1. Macroscopic laws for second order phase transition
At second order phase transitions the correlation length diverges, leading to long-range phenomena. This is the reason for universality: the long-range behaviour does not depend on details of
the microscopic interactions, any more, except for very few properties, like the dimension, the
symmetries and the number of relevant degrees of freedom. These properties determine the
universality class of the system.
164
10.3. Universality
We have seen that the mean led theory is not capable of describing this long-range behaviour
properly. But nevertheless, exact statements are possible! One nds laws which are only loosely
connected to the microphysics and mainly arise from the statistics of many body systems. These
eective laws strongly depend on the relevant distance scale. For example, particle and atomic
physics are suitable to describe systems at very small scales, whereas these systems are better
described by the laws of chemistry and biology at way larger, macroscopic scales.
So, what we are looking for, is a framework to explain this qualitatively dierent behaviour on
dierent scales to describe the transition from microscopic laws on small scales to macroscopic
laws on large scales. This framework is the so-called renormalization group. It describes the
distance scale dependence of the eective laws and makes it possible to cast the qualitative
description of these central ideas into a mathematically rigorous form.
In the following subsections, we will introduce this concept via the block spin description,
developed by Kadano.
(10.57)
eK
i =1 s i s i +1
(10.58)
eKsi si +1 .
(10.59)
PN
{s i }
N
XY
{s i } i=1
Now, we introduce the concept of a block lattice. Therefore, we relabel the spins in the following
way:
i +1
for i odd (spins on the block lattice, block-spins),
(10.60)
2
i
s j0 : j =
for i even (intermediate spins).
(10.61)
2
It is illustrated in Figure 10.2.
The idea now is to integrate out the intermediate spins, to obtain a new Ising model for the
block spins sj . We achieve this by always grouping two subsequent factors of the product in
(10.59) into one block factor. This way, we arrive at a product of blocks:
sj :
j=
Z=
N/2
XY
eK sj s j eK sj+1 s j
0
(10.62)
{s i } j=1
XY
{s i }
(10.63)
(10.64)
XY
{s i }
eK (sj s j +s j+1 s j )
0
165
2
1
j for s
j for s 0
Next, we perform the inner sum (integrate out the intermediate spins):
X Y
Z=
eK (sj +s j+1 ) + eK (sj +s j+1 ) .
{sj }
(10.65)
s j0 =1
(10.66)
This way, we arrive at a new Ising model for the block spins:
XY 0
Z=
eK sj sj +1 +C ,
{sj }
(10.67)
(s j , sj+1 ) :
++ :
e2K + e2K
+ :
=e
e2K + e2K
(10.69)
,
(10.70)
K 0 +C
(10.71)
=e
(10.68)
K 0 +C
=e
+ :
K 0 +C
=e
K 0 +C
= eK C ,
0
(10.72)
(10.73)
(10.74)
166
1
2
ln
f
1
2
e2K + e2K
g
1
2
ln cosh 2K .
(10.75)
10.3. Universality
The additive constant C is irrelevant and can be neglected. With that, (10.67) has exactly the
same form as (10.59). So, if the original lattice had lattice spacing a, the new partition function
is describing a lattice with lattice spacing 2a and a coupling K 0. But, of course, both expressions
of the partition function are the same we are still describing the same model! In other words,
(10.67) is describing the original model, just on a larger length scale. And on this length scale it
behaves like a model on a lattice with doubled lattice spacing and eective coupling K 0:
Renormalization: K K 0 ,
Lattice spacing:
(10.76)
a 2a .
K0
K (2a)
K 00
K (4a)
K 000
(10.77)
K (8a)
(10.78)
This procedure describes the renormalization group. It is a group, whose elements are all
possible sequences of renormalization steps. Actually, if you want to be mathematically more
rigorous, it is just a semigroup, as it does not contain the inverses of its elements.
Now, let us look at the dierent cases for K. If K > 0, then
so,
e2K =
1
2
(10.79)
(10.80)
e2K + e2K < e2K ,
and therefore,
0 < K0 < K ,
(10.81)
i. e. the coupling decreases as the length scale increases. So, the interaction goes to zero, if one
is going to larger and larger length scales:
l
K (l ) 0 .
(10.82)
167
Figure 10.3.: Complicated interactions in two-dimensional block lattice. The symbol denotes block
spins in the rst renormalization step. Links denote eective interactions. In addition to
next-neighbour interactions one now also nds diagonal interactions. A second renormalization step leads to the block spins . This induces even more complicated interactions, as
demonstrated by the dashed link.
2a
Block spins
Let us generalize the concept of block spins, that we have introduced in the one-dimensional
case. A block will now consist out of 2d spins, as illustrated in Figure 10.4 for d = 2. We dene
the spin of the k-th block as
X
si
sk =
(10.83)
,
2d
i block(k )
where we have chosen an arbitrary normalisation. Now, these block spins represent our new
degrees of freedom, i. e. we have reduced the number of degrees of freedom.
The partition function is given by
Z=
e J
<i j > s i s j
(10.84)
{s i =1}
X
{s i =1}
168
e H0 [si ] .
(10.85)
10.3. Universality
X
d
dsk sk 2 /2
si = 1 .
(10.86)
block
For each block we can now insert such a 1 into the partition function:
Z=
dsk
k block lattice
X
d
sk 2 /2
si e H0 [si ]
{s i =1}
original lattice
|
=
Y
(10.87)
block
{z
e H1 [sk ]
dsk e H1 [sk ] .
(10.88)
H1
H2
H3
(10.89)
This way, we end up with a so-called renormalization ow J (L) of the coupling, where L = 2n a.
Moreover, new couplings are generated by this procedure.
New variables
P
The sum of the spins in a block k, i block(k ) si , can take values from 2d to 2d . For our
normalisation (10.83) the spacing between the values sk amounts to 2d/2 , instead of 1 for the
original spins. The range of sk is given by
sk [2 /2 , 2 /2 ] ,
d
(10.90)
instead of si [1, 1] for the original spins. Thus, both range and spacing of the block variable
is dierent from the original variable.
After many renormalization steps we can replace our block spin by a continuous function :
sk
(x ) ,
(10.91)
169
Potential
The interactions between the spins of one given block spin are described by the potential U (sk ).
They are called ultra local or on-site interactions. Due to the symmetry under sk sk it
can only depend on the square of the spin:
U (sk ) = U (sk2 ) .
(10.92)
For the microscopic Ising spins there is only one possible value si2 = 1, so there are no on-site
interactions; i. e. microscopic spins si do not interact with themselves. For block spins, on the
other hand, U (sk2 ) is not trivial any more, as sk2 can take values between 0 and 2d . And in the
continuum limit, U [(x )] is a function of a continuous variable .
Let us denote the potential at step n of the blocking procedure, i. e. at eective lattice spacing
l = 2n a, by U [; l]. So, the potential changes as a function of length scale. Now, the Hamiltonian
at length scale l can be expressed as a potential for the on-site interactions plus additional terms
H :
X
H [(x ); l] =
U [(x ); l] + H .
(10.93)
x
k .
(10.95)
l
So, uctuations with momenta q 2 > k 2 are integrated out.
170
10.3. Universality
(a) l = 100 a
(b) l = 109 a
In momentum space, Hamiltonian and potential are denoted by H [; k] and U [; k]. The
microscopic limit of the Hamiltonian is given by
H [; k = /a ] = H 0 [] ,
(10.96)
(10.97)
In the latter limit, all uctuations are integrated out. The potential in this limit is called the
eective potential:
U [] = U [; k 0] .
(10.98)
Flow equation
The central equation of the renormalization group is the ow equation:
U [; k]
= ... ,
ln k
(10.99)
where the r. h. s. still remains to be computed. This equation describes the change of the potential
U when varying the momentum scale. To extract the free energy for constant (magnetisation),
it is sucient to solve this equation in the k 0 limit. Equivalently, we can consider this
equation in real space:
U [; l]
= ... ,
(10.100)
l
and solve it in the l limit.
As a simplication, we restrict the potential to a quartic polynomial at every scale l:
U [; l] = 12 l ( 0,l ) 2 ,
(10.101)
= 12 2 .
(10.102)
with
171
3c 1 1
0 = 2 2 ,
l
2 l
9c 2
= 2 2 ,
l
4
(10.103)
(10.104)
with some constants c 1 , c 2 of order O(1). (This result requires a computation not covered by this
lecture.) We can directly infer that 0 and decrease for increasing l. The qualitative change of
U due to that has been sketched in Figure 10.6.
Solution of flow equation
Let us rst consider the running minimum. The equation we have to solve, is
3c 1 dl
22 l 2
3c 1
= 2 d(l 1 ) .
2
d 0 =
(10.105)
(10.106)
,
2 l l 0
(10.107)
0 (l ) 0
3c 1
= 0 0,c .
22l 0
172
(10.108)
10.3. Universality
0 < 0,c : In this case, 0 (l ) reaches 0 before l . So, it describes the disordered phase,
where the symmetry is unbroken (SYM).
0 = 0,c : At this point, the phase transition occurs! The corresponding critical temperature
can be deduced from 0 (Tc ) = 0,c .
Now, let us consider the running coupling. Its dierential equation is given by
d
9c 2
= d(1 ) = 2 dl .
2
(10.109)
1
1
9c 2
= 2 (l l 0 ) ,
0 4
where 0 = (l 0 ). Solving this for gives us
!1
1
9c 2
=
+
(l l 0 )
0 42
!1
9c 2 0
= 0 1 +
(l l 0 ) .
42
(10.110)
(10.111)
(10.112)
42 1
.
9c 2 l
(10.113)
Note that this is independent of the initial value! This is an example of how the memory of the
microscopic physics is lost at macroscopic scales!
(10.114)
= 0l ,
(10.115)
l
.
l0
(10.116)
(10.117)
(10.118)
173
(b) Flow of
(a) Flow of
3c 1
.
22
(10.119)
(10.120)
9c 2 2
.
42
(10.121)
Here, we adopted the usual QFT notation of -functions. Note that no mass or length scale
appears in these ow equations!
We have sketched the ow of in Figure 10.7a. Thereby, the arrows indicate the ow towards
larger l, i. e. smaller t. We see that, independent of the starting point, ows to the xed point
for l . From
( ) = 0
(10.122)
it follows
4
=
.
9c 2
(10.123)
So, we can conrm the large length scale behaviour found in (10.113):
l
4
=
.
l
9c 2l
(10.124)
This ow of the eective couplings towards an infrared xed point is related to the observed
loss of memory of the microscopic physics. The details of the microphysics determine the starting
point of the ow of the coupling. But as we have seen, for larger (macroscopic) lenght scales the
174
10.3. Universality
coupling ows towards this xed point independent of the specic starting point! This is also
the deep origin of the universality of the critical exponents. They are related to the behaviour
of the -functions around the xed point. This behaviour is determined by the form of the
potential U . And this, in turn, is determined by the dimension and the symmetries of our system.
Universality follows from the ow!
Contrary to the coupling, the ow of , as sketched in Figure 10.7b, shows the existence of an
ultraviolet xed point . Every other starting point than = leads to a divergent ow for
l . So, the critical point is described by = :
=
T = Tc .
(10.125)
That is because the critical point is characterized by 0 = l 0 for l . And this is only
possible for = .
So, we see that at the critical point all quantities are on a xed point. Therefore, our system
is scale-invariant at the critical point! This behaviour is characteristic for second order phase
transitions. But note that this needs some tuning for , in contrast to the infrared xed point
for .
175