CHAPTER2
CHAPTER2
1
B. Zheng
Zhejiang Institute of Modern Physics, Zhejiang University,
Hangzhou 310027, P.R. China
1
e-mail: zheng@zimp.zju.edu.cn; tel: 0086-571-87952753; Fax: 0086-571-87952754
1
Chapter 1
Thermodynamics
2
Chapter 2
2.1 Postulates
Historically, thermodynamics arose before the atomic
nature of matter was understood. The idea that heat is
a form of energy was first suggested around the end of
eighteenth century, and gained acceptance in the next
half century. The kinetic theory of dilute gases and more
3
general statistical mechanics were developed around the
second half of the nineteenth century and the beginning
of the twentieth century.
Statistical mechanics
• the equilibrium state
• non-equilibrium states
4
The thermodynamic limit
N → ∞, V →∞
but
V
→ v (a non-zero constant)
N
Here V is measured in a microscopic unit,e.g.,
5
• Quasi-fundamental level
Do not trace the motion of each molecule, and con-
sider only the probability distribution ρ(pi , qi , t) with
∏
ρ(pi , qi , t) d3 pi d3 qi
i
being the probability of the system lying within a vol-
∏
ume i d3 pi d3 qi of the coordinate qi and momentum pi .
Equations of motion:
∗ Liouville’s equation
∗ Boltzmann equation
These equations can be solved only in some simple
cases, such as dilute gases.
The time t and space coordinate qi are mesoscopic.
Otherwise there is no meaning for the probability dis-
tribution.
• Statistical mechanics
Do not solve any equations of motion, but assume a
form of ρ(pi , qi , t) in the equilibrium state, i.e., ρ(pi , qi , ∞) ≡
ρ(p, q),
* it can be tested by experiments
* it can be derived from equations of motion in
some special cases.
6
Γ space: the phase space spanned by (p, q), each point
in Γ space represents a microscopic state of the system.
An ensemble: A collection of systems, identical in
composition and macroscopic conditions, existing in (dif-
ferent or the same) microscopic states, without inter-
actions each other. A system in the ensemble can be
represented by a point in Γ space, and these systems
obey the probability distribution ρ(p, q), i.e.,
∏
ρ(p, q)dp dq ≡ ρ(pi , qi ) dpi dqi
i
7
where E is the energy, ∆ ≪ E, H(p, q) is the Hamilto-
nian, and the ensemble described by this distribution is
the so-called microcanonical ensemble.
Here it is important that the Hamiltonian is assumed
to be known, which contains the microstructure and
interactions of the system.
• why is ∆ introduced?
Theoretically, it is convenient
Practically, isolation is not strict
8
of
9
P
f0 f
Figure 2.1:
Figure 2.2:
10
the relative fluctuation goes to zero. This assumption is im-
portant in statistical mechanics.
11
internal energy. So, it is crucial how to define the en-
tropy and temperature.
Let us denote the volume in Γ space of the micro-
scopic ensemble
∫
Γ(E) ≡ dp dq
E<H(p,q)<E+∆
12
with fixed N1 , N2 , V1 and V2 , which satisfy
N = N1 + N2
V = V1 + V2
and
H(p, q) = H1 (p1 , q1 ) + H2 (p2 , q2 )
Here (p1 , q1 ) are the coordinates of particles in the sys-
tem 1, and (p2 , q2 ) are the coordinates of particles in the
system 2.
Especially it is assumed that the interaction between
two subsystems is negligibly small, although it must
exit to allow the exchange of energy between two sub-
systems. For example, if the intermolecular potential
is finite-range, and the surface-to-volume ratio of each
subsystem is negligibly small.
If we define
and
S(E, V ) = k log Γ(E, 2∆)
we should show in the thermodynamic limit
13
tively, the volume in Γ space of the whole system is
∫
Γ(E) = dp1 dp2 dq1 dq2
∫ ∫
E1 <H1 <E1 +∆, E2 <H2 <E2 +∆
and
E ∝ N1 + N2 = N
In addition, both N1 and N2 should be the order of
N , otherwise it may not be meaningful. Then
15
sense. At least, one should not mix this small parameter
with the thermodynamic limit.
Understanding:
Alternatively, we may consider Ē1 and Ē2 are the av-
eraged energy of the two subsystems. In the thermody-
namic limit, the fluctuation δ of E1 around Ē1 should
be much smaller than Ē1 . Let us take δ < ∆, then only
Γ1 (Ē1 )Γ2 (Ē2 ) is dominating
∑
E/∆
Γ(E) = Γ1 (Ei )Γ2 (E − Ei )
i=1
Therefore
16
In fact, Γ1 (E1 )Γ2 (E − E1 ) is usually a very sharp func-
tion E1 , and such an understanding is reasonable. For
example, suppose Γ1 (E1 ) = E1N1 , Γ2 (E − E1 ) = (E − E1 )N2
and N1 = N2 , then
Γ1 (E1 )Γ2 (E − E1 ) = [E1 (E − E1 )]N1
This function has a maximum at E1 = E/2, and it is very
sharp for a large N1 .
Reading materials:
Why
∑
E/∆
Γ(E) = Γ1 (Ei )Γ2 (E − Ei )?
i=1
e.g. N =2
1 2 1 2 2
H = p + ω q
2m 2
1 2 2 1 2
H1 = ω q , H 2 = p
2 2m
∫
Γ(E) = dq dp
∫ E<H<E+2∆
Γ1 (E1 ) = dq
∫ E1 <H1 <E1 +∆
Γ2 (E − E1 ) = dp
E−E1 <H2 <E−E1 +∆
√ 1 )Γ2 (E − E1 )
Γ(E) is just the area of a circular ring. Γ1 (E
is an element of the circular ring at q = E1 . If E1 is
fixed, one may write
Γ(E) = Γ1 (E1 )Γ2 (E − E1 )
17
If not, one needs to integrate over E1
∑
E/∆
Γ(E) = Γ1 (Ei )Γ2 (E − Ei )
i=1
since E1 takes value from 0 to E, corresponding to dif-
ferent states.
End reading materials
18
2.3 Thermodynamics
Let us define
∑ ∫
(E) = dp dq
H(p,q)<E
∑
∂ (E)
ω(E) =
∂E
Then
Γ(E) = ω(E)∆
∑ ∑
Γ(E) = (E + ∆) − (E)
It can be proved that up to an additive constant of the
order O(log N ), the following definitions are equivalent
S = k log Γ(E)
S = k log ω(E)
∑
S = k log (E)
Question: why?
∑
For example, suppose (E) ∝ E N , then Γ(E) ∝ N E N −1 .
∑
Or, more generally,
∑ Γ(E) < (E) < Γ(E) · E/∆. The
behaviors of log (E) and log Γ(E) look similar, up to
the order of O(N ).
19
and E can not change, V can not decrease. Therefore,
the second law here is simply stated as that S∑ is a non-
decreasing function of V . This is obvious, for (E) is a
non-decreasing function of V by its definition.
20
Question: Is it reasonable to define
( )
∂S
P ≡T
∂V E
Hints:
• P is intensive
• Exercise : prove that P defined above satisfies
( )
∂E
P =−
∂V S
21
2.4 Equipartition Theorem
The Hamiltonian of the system is written as
H = H(p, q) = H(pj , qj ), j = 1, · · · 3N
e.g., for an ideal gas,
1 ∑ 2
H= pi
2m
or, a set of harmonic oscillators
1 ∑ 2 1 2∑ 2
H= pi + ω qi
2m 2
Here we note that negligibly small interactions between
independent degrees of freedom have been omitted.
We will derive an equipartition theorem for a general
Hamiltonian, and then apply it to a simple one.
Let xi be either pi or qi
⟨ ⟩ ∫
∂H 1 ∂H
xi = dp dq xi
∂xj Γ(E) E<H<E+∆ ∂xj
(∫ ∫ )
∆ 1 ∂H
= dp dq − dp dq xi
Γ(E) ∆ ∂xj
∫ H<E+∆ H<E
∆ ∂ ∂H
= dp dq xi
Γ(E) ∂E H<E ∂xj
The second equality is based on the definition of the
integration.
∫ ∫
∂H ∂
dp dq xi = dp dq xi (H − E)
∂xj ∂xj
∫ H<E H<E
∫
∂
= dp dq [xi (H − E)] − δij dp dq (H − E)
H<E ∂xj H<E
22
The second equality is obtained through the integration
by parts. The first term in the last line reduces to a
surface integral over the boundary of the region defined
by H < E, i.e.,
∫
the first term = dp dq xi (H − E) = 0
H=E
⟨ ⟩ ∫
∂H δij ∂
∴ xi = dp dp(E − H)
∂xj ω(E) ∂E H<E
(∫ ∫ )
δij 1
= dp dq + dp dq(E − H)
ω(E) H<E ∆ E<H<E+∆
(the second term is negligible,
at least when compared to the first term)
( ∑ )−1
δij ∑ ∂ log (E)
= (E) = δij
ω(E) ∂E
( )−1
∂S
= δij k = δij kT
∂E
If i=j ⟨ ⟩
∂H
xi = kT
∂xi
If ∑ ∑
H= Ai Pi2 + Bi Q2i
i i
23
Pi , Qi are canonical conjugate variables
Then one may easily show
∑ ( ∂H ∂H
)
Pi + Qi = 2H
i
∂P i ∂Qi
1 ∑ 2
N
H = p
2m i=1 i
Then one may compute
∑ ∫
(E) = d3 p1 · · · d3 pN d3 q1 · · · d3 qN
∫
H<E
= VN d3 p1 · · · d3 pN
H<E
Let
√
R = 2mE
∑
(E) = V N Ω3N (R)
24
Ωn (R) is the volume of an n-sphere of radius R
Ωn (R) = Cn Rn
2π n/2
Cn =
Γ(n/2 + 1)
Γ(z) is the gamma function
n n n n
log Cn −−−−→ log π − log +
n→∞ 2 2 2 2
hence
∑ [ ]N
3/2
(E) = C3N V (2mE)
∑
S(E, V ) = k log (E)
[ ( )3/2 ]
4πmE 3
≃ N k log V + Nk
3N 2
( )
3N 2S
U (S, V ) ≡ E = exp −1
4πmV 2/3 3N k
( )
∂U 2U 3
T = = , U = N kT
∂S V 3N k 2
3
CV = Nk
2( )
∂U N kT
P = − =
∂V S V
From the above calculation, Ωn (R) ∝ Rn is essential,
and Cn is not much relevant.
Gibbs paradox
25
Consider the mixture of two ideal gases at the same
temperature and the same density, with the volumes V1
and V2 , and particle numbers N1 and N2 respectively.
Initially, two gases are separated by a partition wall,
then the wall is removed and two gases mix.
If two gases are different, the entropy increases, since
both gases expand. If two gases are identical, nothing
is changed in thermodynamics, but the entropy seems
increasing in statistical mechanics as described above,
since both gases expand respectively.
To solve this problem, one ∑may introduce a Gibbs
factor in the computation of (E), i.e.,
∑ ∫
1
(E) = dN p dN q
N ! H(p,q)<E
With this correction factor, and log N ! ≈ N log N − N ,
one obtains
[ ( )3/2 ]
V 4πmE 5
S(E, V ) = N k log + Nk
N 3N 2
Thus the entropy remains unchanged after the mixture
of two identical gases. In other words, S ∼ N k log V is
unreasonable, since V is extensive.
The Gibbs factor corrects it to S ∼ N k log(V /N ). The
Gibbs factor is needed, whenever there exists a change
of particles in different volumes.
In fact, it seems that the factor one needs is N N rather
than N !. Why does one not directly introduce N N ? In
quantum mechanics, an exchange of two identical parti-
cles does not yield a new quantum state, and N ! is the
total number of such exchanges for N particles.
26
Alternatively, when V is decomposed into V1 and V2 ,
∫ ∫ ∫
1 ∑
N
1 N!
N
d q= N1
d q dN −N1 q
N! V N! N1 !(N − N1 )! V1 V2
N1 =0
H(p, q) = E
27
µ space : the phase space spanned by a single-molecule
coordinates (p, q).
A microscopic state of a single molecule can be rep-
resented by a point in µ space. A microscopic state of
the system is described by a set of the points.
Since the energy of a molecule is bounded by E, the
points are confined to a finite region of µ space. We
divide the region into K elements of volume ω = d3 p d3 q,
and denote the number of the molecules in the l-th el-
ement by nl (need to draw a figure), then
∑
K ∑
K
nl = N ϵl nl = E.
l=1 l=1
28
However, it is difficult to perform this ensemble av-
erage.
N! ∏
K
Ω ({nl }) ∝ ∏K gknk
l=1 nl ! k=1
29
Understanding:
There are N ! ways of distributing N distinguishable
molecules to N positions. However, N positions form
K groups with the distribution {nl }. Inside a group,
there are nl ! ways of distributing nl molecules, which do
not lead to new states in Γ space. In other words, we
consider that molecules in a same element are in a same
state, since their contributions to the physical quantity
f (p, q) are the same.
For a large nl ,
log nl ! ≃ nl (log nl − 1) (i.e., nl ! ≃ nnl l )
∑
K
log Ω ({nl }) = N (log N − 1) − nl (log nl − 1)
l=1
∑
K
+ nl log gl + const
l=1
∑K
Now ∑ we vary {nl } under the condition of l=1 nl =
K
N and l=1 ϵl nl = E, to find the most probable {nl }.
We introduce the Lagrange multipliers α and β, and
calculate
( K )
∑ ∑
K
δ [log Ω ({nl })] − δ α nl + β ϵl nl = 0
l=1 l=1
30
∴ log n̄l = log gl − α − βϵl
n̄l = gl e−α−βϵl
Note that
* n̄l is only the function of ϵl = ϵ(p, q), does not gen-
erally depends on (p, q). This is an important feature of
the equilibrium state.
* The molecules tends to gather in the lower energy
states.
∑ ∑
N = n̄l = e−α−βϵl
l l
∑ ∑
E = ϵl n̄l = ϵl e−α−βϵl
l l
Let us assume
ϵl = ϵ(p, q, y)
The subscript l just labels p and q in µ space, and y =
{yk } represent macroscopic external parameters such as the
31
volume, and the spatial coordinates relevant to external
fields.
Define the partition function of a single particle
∑ ∫
−βϵl
Z(β, y) = e = dpdqe−βϵ(p,q,y)
l ϵ≤E
then
Z(β, y)
N = e−α Z(β, y) or α = log
N
∂ log Z(β, y)
E = −N
∂β
Suppose the system is changed very slowly.
∑ ∑
dE = n̄l dϵl + ϵl dn̄l
l l
Here we assume that the upper and lower bounds of the sum-
mation are fixed. The first term represents the macro-
scopic interaction with the external environment through
the parameters {yk }.
∑ ∑ ∂ϵl
n̄l dϵl = n̄l dyk
∂yk
l l,k
( )
∑ ∑ ∂ϵl
= n̄l dyk
∂yk
k l
Question: why not differentiate over p and q?
Answer: p and q are integrating variables.
Here
∑ ∂ϵl N ∂ log Z(β, y)
Yk ≡ − n̄l =
∂yk β ∂yk
l
32
is the force acting on the environment by the system. This
definition is similar to the pressure P = −(∂E/∂V ∑ )S in
the ensemble theory. Therefore, the term − l n̄l dϵl just
represents the work done by the system.
Proof: since
N
e−α =
Z(β, y)
therefore
N ∂ log Z(β, y) N ∂Z(β, y)
=
β ∂yk βZ(β, y) ∂yk
e−α ∂ ∑ −βϵl
= e
β ∂yk
l
∑ ∂ϵl
= − e−α−βϵl
∂yk
l
∑ ∂ϵl
= − n̄l
∂yk
l
33
one may perform the transformation
q → qV 1/3 , p → pV −1/3 .
The parameter V is then transformed into ϵ(p, q). For a
free particle, for example, ϵ(p, q) = p2 /2m is transformed
into ϵ(p, q)V −2/3 .
Since
∂ log Z(β, y) ∑ ∂ log Z(β, y)
d log Z(β, y) = dβ + dyk
∂β ∂yk
k
hence
( )
N ∂ log Z(β, y)
dQ = d log Z(β, y) − β
β ∂β
Define the temperature by
1
β=
kT
34
and the entropy by
( )
∂ log Z(β, y)
dS = N kd log Z(β, y) − β
∂β
( )
∂ log Z(β, y)
S = N k log Z(β, y) − β + const
∂β
Then, dQ = T dS, i.e., assuming the first law, we may
prove the second law. In fact, one may rewrite
35
and n! ≈ nn ,
∑
log Ω ({n̄l }) = − n̄l log n̄l
∑l
= (α + βϵl )e−α−βϵl
l
∑ ∑
= α e−α−βϵl + β ϵl e−α−βϵl
l l
Z(β, y) ∂ log Z(β, y)
= N log − Nβ
N ∂β
( )
Z(β, y) ∂ log Z(β, y)
∴ S = N k log −β
N ∂β
The difference comes from the Gibbs factor.
Fix y,
∂ log Z(β, y) ∂ log Z(β, y)
dS = N k dβ − N k dβ + kβdE
∂β ∂β
= kβdE
( )
1 ∂S
∴ = = kβ
T ∂E y
2.8 Summary
Ergodicity
→ Postulate of equal a prior probability for the mi-
crocanonical ensemble
A special case: the most probable distribution of
quasi-independent particles.
→ T, S, P and thermodynamics
36
The drawback is that the calculation of observables
is clumsy because of the restriction of the energy.
Exercise :
• Assuming ( ) ( )
∂S ∂S 1
P ≡T , = ,
∂V E ∂E V T
prove ( )
∂E
P =− .
∂V S
• Discussion
Negative temperature
37