0% found this document useful (0 votes)
32 views

8.06 Spring 2016 Lecture Notes 2. Time-Dependent Approximation Methods

The document summarizes lecture notes on time-dependent approximation methods. It covers topics like time-dependent perturbation theory, using a rotating frame to eliminate natural time evolution from the large static Hamiltonian H0. Perturbation expansions are used to approximate solutions when the perturbation δH(t) is small. Specific examples discussed include NMR, periodic perturbations, light and atoms interactions, adiabatic evolution, and scattering.

Uploaded by

OttoHinder
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
0% found this document useful (0 votes)
32 views

8.06 Spring 2016 Lecture Notes 2. Time-Dependent Approximation Methods

The document summarizes lecture notes on time-dependent approximation methods. It covers topics like time-dependent perturbation theory, using a rotating frame to eliminate natural time evolution from the large static Hamiltonian H0. Perturbation expansions are used to approximate solutions when the perturbation δH(t) is small. Specific examples discussed include NMR, periodic perturbations, light and atoms interactions, adiabatic evolution, and scattering.

Uploaded by

OttoHinder
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
You are on page 1/ 37

8.

06 Spring 2016 Lecture Notes


2. Time-dependent approximation methods

Aram Harrow

Last updated: March 12, 2016

Contents
1 Time-dependent perturbation theory 1
1.1 Rotating frame . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Perturbation expansion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3 NMR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.4 Periodic perturbations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

2 Light and atoms 8


2.1 Incoherent light . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.2 Spontaneous emission . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.3 The photoelectric effect . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

3 Adiabatic evolution 15
3.1 The adiabatic approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.2 Berry phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.3 Neutrino oscillations and the MSW effect . . . . . . . . . . . . . . . . . . . . . . . . 21
3.4 Born-Oppenheimer approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

4 Scattering 24
4.1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.2 Born Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
4.3 Partial Waves . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

1 Time-dependent perturbation theory


Perturbation theory can also be used to analyze the case when we have a large static Hamiltonian
H0 and a small, possibly time-dependent, perturbation δH(t). In other words

H(t) = H0 + δH(t). (1)

However, the more important difference from time-independent perturbation theory is in our goals:
we will seek to analyze the dynamics of the wavefunction (i.e. find |ψ(t)i as a function of t) rather
than computing the spectrum of H. In fact, when we use a basis, we will work in the eigenbasis of
H0 . For example, one common situation that we will analyze is that we start in an eigenstate of
H0 , temporarily turn on a perturbation δH(t) and then measure in the eigenbasis of H0 . This is a
bit abstract, so here is a more concrete version of the example. H0 is the natural Hamiltonian of

1
the hydrogen atom and δH(t) comes from electric and/or magnetic fields that we temporarily turn
on. If we start in the 1s state, then what is the probability that after some time we will be in the
2p state? (Note that the very definition of the states depends on H0 and not the perturbation.)
Time-dependent perturbation theory will equip us to answer these questions.

1.1 Rotating frame


We want to solve the time-dependent Schrödinger equation i~∂t |ψ(t)i = H(t)|ψi. We will assume
that the dynamics of H0 are simple to compute and that the computational difficulty comes from
δH(t). At the same time, if H0 is much larger than δH(t) then most of the change in the state will
come from H0 . In classical dynamics when an object is undergoing two different types of motion,
it is often useful to perform a change of coordinates to eliminate one of them. We will do the same
thing here. Define the state
iH0 t
˜
|ψ(t)i = e ~ |ψ(t)i. (2)
˜
We say that |ψ(t)i is in the rotating frame or alternatively the interaction picture. Multiplying by
iH0 t
e ~ cancels out the natural time evolution of H0 . In particular, if δH(t) = 0 then we would have
˜
|ψ̃(t)i = |ψ̃(0)i = |ψ(0)i. Thus, any change in |ψ(t)i must come from δH(t).

Aside: comparison to Schrödinger and Heisenberg pictures. In 8.05 we saw the Schrödinger
picture and the Heisenberg picture. In the former, states evolve according to H and operators re-
main the same; in the latter, states stay the same and operators evolve according to H. The
interaction picture can be thought of as intermediate between these two. We pick a frame rotating
with H0 , which means that the operators evolve according to H0 and the states evolve with the
remaining piece of the Hamiltonian, namely δH. As we will see below, to calculate this evolution
correctly we need δH to rotate with H0 , just like all other operators. This is a little vague but
below we will perform an exact calculation to demonstrate what happens.
Now let’s compute the time evolution of |ψ̃(t)i.

d d  iH0 t 
i~ |ψ̃(t)i = i~ e ~ |ψ(t)i
dt dt
iH0 t iH0 t
= −H0 e ~ |ψ(t)i + e ~ (H0 + δH(t)) |ψ(t)i
iH0 t iH0 t
=e ~ δH(t)|ψ(t)i since H0 and e ~ commute
iH0 t iH t
− ~0 ˜
=e ~ δH (t)e |ψ(t)i
| {z }
δH(t)
f

Thus we obtain an effective Schrödinger equation in the rotating frame

d ˜ ˜
i~ |ψ(t)i = δf
H(t)|ψ(t)i (3)
dt

where we have defined iH0 t iH0 t


δH(t)
f =e ~ δH(t)e− ~ .
This has a simple interpretation as a matrix. Suppose that the eigenvalues and eigenvectors of H0
(reminder: we work with the eigenbasis of H0 and not H(t)) are given by

H0 |ni = En |ni.

2
Define δHmn (t) = hm|δH(t)|ni. Then
iH0 t iH0 t i(Em −En )t
f mn (t) = hm|e
δH ~ δH(t)e− ~ |ni = e ~ δHmn (t) ≡ eiωmn t δHmn ,
where we have defined ωmn = Em −E~
n
. If we define cn (t) according to
X X iEn t
|ψ̃(t)i = cn (t)|ni =⇒ |ψ(t)i = e− ~ cn (t)|ni
n n

then we obtain the following coupled differential equations for the {cn }.
X X
i~ċm (t) = δf
H mn (t)cn (t) = eiωmn t δHmn cn (t).
n n

1.2 Perturbation expansion


So far everything has been exact, although sometimes this is already enough to solve interesting
problems. But often we will need approximate solutions. So assume that δH(t) = O(λ) and expand
the wavefunction in powers of λ, i.e.
(0) (1) (2)
cm (t) = cm (t) + cm (t) + cm (t) +...
O(1) O(λ) O(λ2 )
˜
|ψ(t)i = |ψ˜(0) (t)i + |ψ˜(1) (t)i + |ψ˜(2) (t)i + . . .

We can solve these order by order. Applying (3) we obtain

i~∂ |ψ̃ (0) (t)i + i~∂t |ψ̃ (1) (t)i + i~∂t |ψ˜(2) (t)i + . . . = δfH(t)|ψ˜(0) (t)i + δf
H(t)|ψ˜(1) (t)i + . . . (4)
| t {z } | {z } | {z } | {z } | {z }
O(1) O(λ) O(λ2 ) O(λ) O(λ2 )

The solution is much simpler than in the time-dependent case. There is no zeroth order term on
the RHS, so the zeroth order approximation is simply that nothing happens:

|ψ̃ (0) (t)i = |ψ̃ (0) (0)i = |ψ(0)i (5)

The first-order terms yield


H(t)|ψ˜(0) (t)i = δf
i~∂t |ψ̃ (1) (t)i = δf H(t)|ψ(0)i. (6)
Integrating, we find
Z t f t0 )
δH(
|ψ̃ (1)
(t)i = dt0 |ψ(0)i . (7)
0 i~
This leads to one very useful formula. If we start in state |ni, turn on H(t) for times 0 ≤ t ≤ T
and then measure in the energy eigenbasis of H0 , then we find that the probability of ending in
state |mi is
t δH (t0 )eiωmn t0 2
Z
0 mn
Pn→m = dt

i~

0
We can also continue to higher orders. The second-order solution is
t0
t
H(t0 ) δf
H(t00 )
Z Z
0 δf
|ψ̃ (2)
(t)i = dt dt00 |ψ(0)i . (8)
0 0 i~ i~

3
1.3 NMR
In some cases the rotating frame already helps us solve nontrivial problems exactly without going
to perturbation theory. Suppose we have a single spin-1/2 particle in a magnetic field pointing in
the ẑ direction. This field corresponds to a Hamiltonian
~
H0 = ω0 Sz = ω0 σz .
2
If the particle is a proton (i.e. hydrogen nucleus) and the field is typical for NMR, then ω0 might
be around 500 MHz.

Static magnetic field Now let’s add a perturbation consisting of a magnetic field in the x̂
direction. First we will consider the static perturbation

δH(t) = ΩSx ,

where we will assume Ω  ω0 , e.g. Ω might be on the order of 20 KHz. (Why are we considering a
time-independent Hamiltonian with time-dependent perturbation theory? Because really it is the
time-dependence of the state and not the Hamiltonian that we are after.)
We can solve this problem exactly without using any tools of perturbation theory, but it will be
instructive to compare the exact answer with the approximate one. The exact evolution is given
by precession about the
ω ẑ + Ωx̂
p0
ω02 + Ω2
p
axis at an angular frequency of ω02 + Ω2 . If Ω  ω0 then this is very close to precession around
the ẑ axis.
Now let’s look at this problem using first-order perturbation theory.
ω t
0 ω t 0
δfH(t) = ei 2 σz ΩSx e−i 2 σz = Ω (cos(ω0 t)Sx − sin(ω0 t)Sy )
Z t f 0)
δH(t
(1)
|ψ̃ (t)i = dt0 |ψ(0)i
0 i~
Z t

dt0 cos(ω0 t0 )Sx − sin(ω0 t0 )Sy |ψ(0)i

=
0 i~
1 Ω
= (sin(ω0 t)Sx + (cos(ω0 t) − 1)Sy ) |ψ(0)i
i~ ω0
We see that the total change is proportional to Ω/ω0 , which is  1. Since this is the difference
between pure rotating around the ẑ axis, this is consistent with the exact answer we obtained.
The result of this calculation is that if we have a strong ẑ field, then adding a weak static x̂
field doesn’t do very much. If we want to have a significant effect on the state, we will need to do
something else. The rotating-frame picture suggests the answer: the perturbation should rotate
along with the frame, so that in the rotating frame it appears to be static.

Rotating magnetic field Now suppose we apply the perturbation

δH(t) = Ω (cos(ω0 t)Sx + sin(ω0 t)Sy ) .

4
We have already computed S˜x above. In the rotating frame we have

S̃x = (cos(ω0 t)Sx − sin(ω0 t)Sy )


S̃y = (cos(ω0 t)Sy + sin(ω0 t)Sx )

Thus
δf
H(t) = ΩSx .
The rotating-frame solution is now very simple:
iΩt
|ψ̃(t)i = e− 2
σx
|ψ(0)i.

This can be easily translated back into the stationary frame to obtain
iω0 t iΩt
|ψ(t)i = e− 2
σz
e− 2
σx
|ψ(0)i.

1.4 Periodic perturbations


The NMR example suggests that transitions between eigenstates of H0 happens most effectively
when the perturbation rotates at the frequency ωmn . We will show that this holds more generally
in first-order perturbation theory. Suppose that

δH(t) = V cos(ωt),

for some time-independent operator V . If our system starts in state |ni then at time t we can
calculate Z t f mn (t0 )
δH
(1) ˜(1)
cm (t) = hm|ψ (t)i = dt0
0 i~
Z t 0
δHmn (t ) iωmn t0
= dt0 e
0 i~
Z t
Vmn 0
= dt0 cos(ωt)eiωmn t
0 i~
Vmn t 0  i(ωmn +ω)t0
Z 
0
= dt e + ei(ωmn −ω)t
2i~ 0
" #
Vmn ei(ωmn +ω)t − 1 ei(ωmn −ω)t − 1
= +
2i~ ωmn + ω ωmn − ω
The ωmn ± ω terms in the denominator mean that we will get the largest contribution when
ω ≈ |ωmn |. (A word about signs. By convention we have ω > 0, but ωmn is a difference of energies
and so can have either sign.) For concreteness, let’s suppose that ω ≈ ωmn ; the ω ≈ −ωmn case is
similar. Then we have
Vmn ei(ωmn −ω)t − 1
c(1)
m (t) ≈ .
2i~ ωmn − ω
If we now measure, then the probability of obtaining outcome m is
 
2 (ωmn −ω)t
sin 2 sin2 αt

| V | 2 2 |V |
mn mn
Pn→m (t) ≈ |c(1) 2
m (t)| = = 2
,
~2 (ωmn − ω)2 ~2 α2

5
where we have defined the detuning α ≡ ωmn − ω. The t, α dependence is rather subtle, so we
examine it separately. Define
sin2 αt

2
f (t, α) = (9)
α2
For fixed α, f (t, α) is periodic in t.
It is more interesting to consider the case of fixed t, for which f has a sinc-like appearance (see
Fig. 1).

Figure 1: The function f (α, t) from (9), representing the amount of amplitude transfered at a fixed
time t as we vary the detuning α ≡ ωmn − ω.

It has zeroes as α = 2π
t n for integers n 6= 0. Since the closest zeros to the origin are at ±2π/t,

we call the region α ∈ [−2π/t, 2π/t] the “peak” and the rest of the real line (i.e. R ∞ |α| ≥2 t ) the
2
“tails.” For α → ∞, f (t, α) ≤ 1/α . Thus, the tail has total area bounded by 2 2π/t 1/α = O(t).
For the peak, as α → 0, f (t, α) → t2 /4. On the other hand, sin is concave, so for 0 ≤ θ ≤ π/2
t2
we have sin(θ) ≥ sin(π/2) 2 π
π/2 θ = π θ. Thus for |α| ≤ t we have f (α, t) ≥ π . While these crude bounds
do not determine the precise multiplicative constants, this does show that there is a region of width
∼ 1/t and height ∼ t2R, and so the peak also has area O(t).

We conclude that −∞ dαf (t, α) ∼ t. Dividing by t, we obtain
Z ∞
f (t, α)
dα ∼ 1.
−∞ t

On the other hand, f (t,α)


t 6 0. So as t → ∞ we see that f (t,α)
→ 0 as t → ∞ for all α = t is always
nonnegative, always has total mass roughly independent of t, but approaches zero for all nonzero
α. This means that it approaches a delta function. A more detailed calculation (omitted, but it
uses complex analysis) shows that

f (t, α) π
lim = δ(α).
t→∞ t 2

6
The reason to divide by t is that this identifies the rate of transitions per unit time. Define
Rn→m = Pn→m
t . Then the above arguments imply that

π |Vmn |2
Rn→m ≈ δ(|ωmn | − ω) for large t. (10)
2 ~2

Linewidth In practice the frequency dependence is not a delta function. The term “linewidth”
refers to the width of the region of ω that drives a transition; more concretely, FWHM stands
for “full-width half-maximum” and denotes the width of the region that achieves ≥ 1/2 the peak
transition rate. The above derivation already suggests some reasons for nonzero linewidth.

1. Finite lifetime. If we apply the perturbation for a limited amount of time, or if the state we
are driving to/from has finite lifetime, then this will contribute linewidth on the order of 1/t.

2. Power broadening. If |Vmn | is large, then we will still see transitions for larger values of |α|.
For this to prevent us from seeing the precise location of a peak, we need also the phenomenon
of saturation in which transition rates all look the same above some threshold. (For example,
we might observe the fraction of a beam that is absorbed by some sample, and by definition
this cannot go above 1.)

There are many other sources of linewidth. In general we can think of both the driving frequency
ω and the gap frequency ωmn as being distributions rather than δ functions. The driving frequency
might come from a thermal distribution or a laser, both of which output a distribution of frequen-
cies. The linewidth of a laser is much lower but still nonzero. The energy difference ~ωmn seems
like a universal constant, can also be replaced by a distribution by phenomena such as Doppler
broadening, in which the thermal motion of an atom will redshift or blueshift the incident light.
This is just one example of a more general phenomenon in which interactions with other degrees of
freedom can add to the linewidth; e.g. consider the hyperfine splitting, which measures the small
shifts in an electron’s energy from its interaction with the nuclear spin. This can be thought of as
adding to linewidth in two different, roughly equivalent, ways. We might think of the nuclear spins
as random and thus the interaction adds a random term to the electon’s Hamiltonian. Alterna-
tively, we might view the interaction with the nuclear spin as a source of decoherence and thus as
contributing to the finite lifetime of the electron’s excited state. We will not explore those issues
further here.
The other contribution to the rate is the matrix element |Vmn |. This depends not only on the
strength of the perturbation, but also expresses the important point that we only see transitions
from n → m if Vmn 6= 0. This is called a selection rule. In Griffiths it is proved that transitions from
electric fields (see the next section) from Hydrogen state |n, l, mi to |n0 , l0 , m0 i are only possible
when |l − l0 | = 1 and |m − m0 | ≤ 1 (among other restrictions). Technically these constraints hold
only for first-order perturbation theory, but still selection rules are important, since they tell us
when we need to go to higher-order perturbation theory to see transitions (known as “forbidden
transitions”). In those cases transition rates are much lower. One dramatic example is that 2p → 1s
transition in hydrogen takes 1.6ns because it occurs at first order while the 2s → 1s transition takes
0.12 seconds. For this reason states such as the 2s states are called “metastable.”
We now consider the most important special case, which gets its own top-level section, despite
being an example of a periodic perturbation, which itself is an example of first-order perturbation
theory.

7
2 Light and atoms
~ and B
Light consists of oscillating E ~ fields. The effects of the B
~ fields are weaker by a factor
~
O(v/c) ∼ α, so we will focus on the E fields. Let
~ r) = E0 ẑ cos(ωt − kx).
E(~

However, optical wavelengths are 4000-8000Å, while the Bohr radius is ≈ 0.5Å, so to leading order
we can neglect the x dependence. Thus we approximate

δH(t) = eE0 z cos(ωt). (11)

We now can apply the results on transition rates from the last section with Vmn = eE0 hm|z|ni.
(This term is responsible for selection rules and for the role of polarization.) Thus the rate of
transitions is
π e2 E02
Rn→m = |hm|z |ni|2 δ(|ωmn | − ω). (12)
2 ~2
We get contributions at ωmn = ±ω corresponding to both absorption and stimulated emission.

Aside: quantizing light What about spontaneous emission? This does not appear in the
semiclassical treatment we’ve described here. Nor do the photons. “Absorption” means jumping
from a low-energy state to a higher-energy state, and “stimulated emission” means jumping from
high energy to low energy. In the former case, we reason from energy conservation that a photon
must have been absorbed, and in the latter, a photon must have been emitted. However, these
arguments are rather indirect. A much more direct explanation of what happens to the photon
comes from a more fully quantum treatment. This also yields the phenomenon of spontaneous
emission. Recall from 8.05 that oscillating electromagnetic fields can be quantized as follows:
r r
† 2π~ω ~ω
E0 = E0 (â + â ) E0 = (Gaussian units) = (SI units)
V 0 V
Using δH = eE0 z, we obtain
δH = eE0 z ⊗ (â + ↠).
 
0 α
If we look at the action of z in the {1s, 2pz } basis, then it has the form   with α = h1s|z |2pz i.
α 0
We then obtain the form of the Hamiltonian examined on pset 3.
This perspective also can be used to give a partial derivation of the Lamb shift, which can be
thought of as the interaction of the electron with fluctuations in the electric field of the vacuum.
In the vacuum (i.e. ground state of the photon field) we have hE0 i ∼ hâ + ↠i = 0 but hE02 i ∼
h(â + ↠)2 i > 0. These vacuum fluctuations lead to a small separation in energy between the 2s
and 2p levels of hydrogen.

Dipole moment In the Stark effect we looked at the interaction of the hydrogen atom with an
electric field. This was a special case of the interaction between a E ~ field and the dipole moment
of a collection of particles. Here we discuss the more general case.
Suppose that we have charges q1 , . . . , qN at positions ~x(1) , . . . , ~x(N ) , and we apply an electric
~ x). The energy is determined by the scalar potential φ(~x) which is related to the electric
field E(~

8
~ = −∇
field by E ~ φ. If E(~
~ x) = E ~ (i.e. independent of position ~x) then one possible solution is
~
φ(~x) = −~x · E. In this case the Hamiltonian will be
N
X N
X
qi φ ~x(i) = − ~ = −d~ · E
qi ~x(i) · E ~

H=
i=1 i=1
PN
where we have defined the dipole moment d~ = i=1 qi ~x(i) . Our choice of φ was not unique, and
we could have chosen φ(~x) = C − ~x · E ~ for any constant C. However, this would only have added
an overall constant to the Hamiltonian, which would have no physical effect.
What if the electric field is spatially varying? If this spatial variation is small and we are near
the origin, we use the first few terms of the Taylor expansion to approximate the field:
3
~ x) = E(0)
~
X ∂Ei
E(~ + êi xj + . . . .
∂xj
i,j =1

This corresponds to a scalar potential of the form


3 3
X 1 X ∂Ei
φ= xi Ei (0) + xi xj xi xj + . . .
2 ∂xj
i=1 i,j=1

For the quadratic termsP


we see that the field couples not to the dipole moment, but to the quadrupole
moment, defined to be N x(i) ⊗ ~x(i) . This is related to emission lines such as 1s → 3d in which
i=1 qi ~
` may change by up to ±2. Of course higher moments such as octupole moments can be also be
considered. We will not explore these topics further in 8.06.

2.1 Incoherent light


While we have so far discussed monochromatic light with a definite polarization, it is easier to
produce light with a wide range of frequencies and with random polarization. To analyze the rate
of transitions this causes we will average (12) over frequencies and polarizations.
Begin with polarizations. Instead of the field being E0 ẑ, let the electric field be E0 Pˆ for some
random unit vector Pˆ . We then replace V with −E0 Pˆ · d. ~ The only operator here is the dipole
moment d~ = (d1 , d2 , d3 ), so the matrix elements of V are given by
3
X
Vmn = −E0 Pˆ · d~mn = −E0 Pi hm|di |ni.
i=1

Since the transition rate depends on |Vmn |2 , we will average this quantity over the choice of polar-

9
ization. Denote the average over all unit vectors Pˆ by h·iPˆ .
D E
|Vmn |2 ˆ = E02 |Pˆ · d~mn |2


P P̂
3
X
= E02


hm|Pi di |ni hn|Pj dj |mi P̂
i,j =1
3
X
= E02


Pi Pj Pˆ
hm|di |ni hn|dj |mi
i,j=1
3
X δij
= E02 hm|di |ni hn|dj |mi explained below
3
i,j =1
E02 X
= |hm|di |ni|2
3
i
E02
≡ |d~mn |2
3
How did we calculate hPi Pj iPˆ ? This can be done by explicit calculation, but it is easier to use sym-
metry. First, observe that the uniform distribution over unit vectors is invariant under reflection.
Thus, if i =6 j, then hPi Pj i = h(−Pi )Pj i = 0. On the other hand rotation symmetry means that
hPi i should be independent of i. Since P12 + P23 + P32 = 1, we also have hP12 + P23 + P32 i = 1 and
2

thus hPi2 i = 1/3. Putting this together we obtain

δij
hPi Pj iPˆ = . (13)
3
Next, we would like to average over different frequencies. The energy density of an electric field
E02
is U =
R 8π (using Gaussian units). Define U (ω) to be the energy density at frequency ω, so that
U = U (ω) dω. If we consider light with this power spectrum, then we should integrate the rate
times this distribution over U (ω) to obtain

4π 2
Z
Rn→m = dω U (ω) 2 |d~mn |2 δ(ω − |ωmn |)
3~
4π 2 ~ 2
= |dmn | U (|ωmn |)
3~2

This last expression is known as Fermi’s Golden Rule. (It was discovered by Dirac, but Fermi called
it “Goldren Rule #2”.)

2.2 Spontaneous emission


The modern description of spontaneous emission requires QED, but the first derivation of it predates
even modern quantum mechanics! In a simple and elegant argument, Einstein:

(a) derived an exact relation between rates of spontaneous emission, stimulated emission and
absorption; and

(b) proposed the phenomenon of stimulated emission, which was not observed until 1960.

10
He did this in 1917, more than a decade before even the Schrödinger equation!
Here we will reproduce that argument. It assumes a collection of atoms that can be in either
state a or state b. Suppose that there are Na atoms in state a and Nb atoms in state b, and that the
states have energies Ea , Eb with Eb > Ea . Define ωba = Eb −E~
a
and β = 1/kB T . Assume further
that the atoms are in contact with a bath of photons and that the entire system is in thermal
equilibrium with temperature T . From this we can deduce three facts:
Fact 1. Equilibrium means no change: N˙ a = N˙ b = 0.
Nb e−βEb
Fact 2. At thermal equilibrium we have Na = e−βEa
= e−β ~ωba .

Fact 3. At thermal equilibrium the black-body radiation spectrum is


~ ω3
U (ω) = . (14)
π 2 c3 eβ~ω − 1

We would like to understand the following processes:

Process Explanation Rate

Absorption A photon of frequency ωba is absorbed and an Bab Na U (ωba )


atom changes from state a to state b.
Spontaneous A photon of frequency ωba is emitted and an atom ANb
emission changes from state b to state a.
Stimulated A photon of frequency ωba is emitted and an atom Bba Nb U (ωba )
emission changes from state b to state a.

These processes depend on the Einstein coefficients A, Bab and Bba for spontaneous emission,
absorption and stimulated emission respectively. They also depend on the populations of atoms
and/or photons that they involve; e.g. absorption requires an atom in state a and a photon of
frequency ωba , so its rate is proportional to Na U (ωba ). Here it is safe to posit the existence of
stimulated emission because we have not assumed that Bba is nonzero.
Having set up the problem, we are now almost done! Adding these processes up, we get

Ṅb = −Nb A − Nb Bba U (ωba ) + Na Bab U (ωba ). (15)

From Fact 1, N˙ b = 0 and so we can rearrange (15) to obtain

A Fact 2 A Fact 3 ~ω 3 1
U (ωba ) = Na
= β~ω
= 2 3 β~ω
Nb Bab − Bba
e ba Bab − Bba π c e ba −1

Since this relation should hold for all values of β, we can equate coefficients and find

Bab = Bba (16a)


~ωba3
A= Bab (16b)
π 2 c3
We see that these three processes are tightly related! All from a simple thought experiment, and
not even the one that Einstein is most famous for.
Today we can understand this as the fact that the electric field enters into the Hamiltonian as
a Hermitian operator proportional to â + ↠, and so the photon-destroying processes containing â

11
are inevitably accompanied by photon-creating processes containing ↠. Additionally the relation
between spontaneous and stimulated emission can be seen in the fact that both involve an a†
operator acting on the photon field. If there are no photons, then the field is in state |0i and we
get the term a† |0i = |1i, corresponding to√spontaneous emission. If there are already n photon in
the mode, then we get the term a† |ni = n + 1|n + 1i. Since the probabilities are the square of
the amplitudes, this means that we see photons emitted at n + 1 times the spontaneous emission
rate. In Einstein’s terminology, the n here is from stimulated emission and the +1 is from the
spontaneous emission which always occurs independent of the number of photons present.
Returning to (16), we plug in Fermi’s Golden Rule and obtain the rates

4π 2 ~ 2
Bab = Bba = |dab | and
3~2
3
4ωba
A= |d~ab |2 .
3~c3

2.3 The photoelectric effect


So far we have considered transitions between individual pairs of states. Ionization (aka the photo-
electric effect) involves a transition from a bound state of some atom to one in which the electron is
in a free state. This presents a few new challenges. First, we are used to treating unbound states as
~
unnormalized, e.g. ψ(~x) = eik·~x . Second, to calculate the ionization rate, we need to sum over final
states, since the quantity of physical interest is the total rate of electrons being dislodged from an
atom, and not the rate at which they transition to any specific final state. (A more refined analysis
might look at the angular distribution of the final state.)
Suppose our initial state is our old friend, the ground state of the hydrogen atom, and the
final state is a plane wave. If the final state is unnormalized, then matrix elements such as
hψfinal |V |ψinitial i become less meaningful. One way to fix this is to put the system in a box of
~2
size L × L × L with L  a0 ≡ me 2 and to impose periodic boundary conditions. The resulting
plane-wave states are now
~ · ~x)
exp(ik
ψ~k (~x) = ,
L3/2
where |ki = 2π ~n and ~n = (n1 , n2 , n3 ) is a vector of integers. (We use ~k instead of p~ = ~k
L
~ to keep
the notation more compact.) We will assume that L  a0 and also that the final energy of the
electron is  13.6 eV. This means that we can approximate the final state as a free electron and
can ignore the interaction with the proton.
Apply an oscillating electric field to obtain the time-dependent potential

δH = eE0 x3 cos(ωt) ≡ V cos(ωt).

The rate of the transitions will be governed by the matrix element


Z  
eE 0 r eE0
h~k|V |1, 0, 0i = p 3 d x x3 exp − − i~k · ~x ≡ p 3 A.
3
πa0 L3 a0 πa0 L3
| {z }
A

We have defined A to equal the difficult-to-evaluate integral. The factor of x3 can be removed by

12
writing A = i ∂k∂ 3 B, where
Z  
3 r ~
B = d x exp − − ik · ~x
a0
Z ∞ Z 1 
r
 ~k · ~x
= 2π r 2
dµ exp − − ikrµ defining µ = ~ r = |~x|
, k = |k|,
0 −1 a0 kr
Z 1 Z ∞
1 n!
= 4π dµ  3 using rn e−αr =
−1 1 0 αn+1
a0 + ikµ
Z 1
4πi 1
= 3 dµ  3
k −1 µ + kai 0
  2   2 
1 1
2πi  1 1  2πi  1 + ika0 − 1− ika0
= − 3  2 −  2  = 3 

2
k k
 
1 + kai 0 1 − kai 0 1 + k21a2
0
1
8πi ika0 8π 8π
= 2 = 2 = 2
k3 + a− 2
 
1+ 1
k 4 a0 1 + 1 a0 k2 0
k2 a20 k2 a20

∂ 2
To compute A, we use the fact that ∂k3 k = 2k3 . Thus

∂ −32πik3
A=i B= 3 .
∂k3 a0 k 2 + a−2
0

We can simplify this expression using our assumption that the photon energy (and therefore also
2 k2
the final state energy) is much larger than the binding energy. The final energy is ~2m and the
~2
binding energy is 2ma2 . Thus
0

~2 k 2 ~2
 =⇒ ka0  1.
2m 2ma20

We can use this to simplify A by dropping the a−2


0 term in the dominator:

−32πik3 −32πi cos(θ)


A≈ 6
= ,
a0 k a0 k 5

where θ is the angle between ~k and the z-axis.


We can now compute the squared matrix element to be (canceling a factor of π from numerator
and denominator)
e2 E 2 1024π cos2 (θ)
|h~k|V |1, 0, 0i|2 = 3 03 .
a0 L a20 k 10
1 f (t,α)
To compute the average rate over some time t, we will multiply this by ~2 t , where f (t, α) ≡
sin2 (αt/2)
α2
and ~α is the difference in energy between the initial and final states. If we neglect the
energy of the initial state, we obtain
 
~k2
2 2 2
1024πe E0 cos (θ) f t, 2m − ω
R1,0,0→k~ = 2 5 10 3
.
~ a0 k L t

13
We can simplify this a bit by averaging over all the polarizations of the light. (In fact, the angular
dependance of the free electron can often carry useful information, but here it will help simplify
some calculations.) The average of cos2 (θ) over the sphere is 1/3 (by the same arguments we used
in the derivation of Fermi’s golden rule), so we obtain
 
~k2
2
1024πe E 2 f t, 2m − ω
hR1,0,0→k~ i = 2 5 10 03 .
3~ a0 k L t

Let’s pause for a minute to look at what we’ve derived. One strange feature is the 1/L3 term,
because the rate of ionization should not depend on how much empty space surrounds the atom.
Another strange thing appears to happen when we take t large, so that f (t, α)/t will approach
π 2
2 δ(α). This would cause the transition rates to be nonzero only when 2mω/~ exactly equals k for
some valid vector ~k (i.e. of the form 2π L~n). We do not generally expect physical systems to have
such sensitive dependence on their parameters.
As often happens when two things look wrong, these difficulties can be made to “cancel each
other out.” Let us take t to be large but finite. It will turn out that t needs to be large only relative
2
to L~2 m , which is not very demanding when L is large. In this case, we can approximate f (t, α)
with a step function: (
π 2
t if 0 ≤ α ≤ 1t
f (t, α) ≈ f˜(t, α) ≡ 2
0 otherwise

In what sense is this a good approximation? We argue that for large t, f˜(t, α)/t ≈ π2 δ(α), just like
f (t, α)/t. Suppose that g(α) is a function satisfying |g 0 (α)| ≤ C for all α. Then
Z ! Z !
∞ f˜(t, α) 1/t
dα − δ(α) g(α) = t dα g(α) − g(0)

π
2t

−∞ 0
Z
1/t
= t dα (g(α) − g(0))

0
Z
1/t Z α
= t dα dβ g 0 (β)

0 0
Z 1/t Z α
dβ g 0 (β)

≤t dα triangle inequality
0 0
1 C
≤ t 2C =
2t 2t
This tends to 0 as t → ∞. (This is an example of a more general principle that the “shape” of a δ
function doesn’t matter. For example, the limit of a Gaussian distribution with σ 2 → 0 would also
work.)
Now using f˜(t, α), we get a nonzero contribution for ~k satisfying
~ 1
0≤2
2m − ω ≤
k  t 
2mω 2 2mω 1
⇔ ≤k ≤ 1+
~ ~ tω
r s   r  
2mω 2mω 1 2mω 1
⇔ ≤k≤ 1+ ≈ + (17)
~ ~ tω ~ 2tω

14
How many ~k satisfy (17)? Valid ~k live on a cubic lattice with spacing 2π/L, and thus have density
(L/2π)3 . Thus we can estimate the number of ~k satisfying (17) by (L/2π)3 times theqvolume of
2mω
k-space satisfying (17). This in turn corresponds to a spherical shell of inner radius ~ and
q
thickness 2mω 1
~ 2tω . Thus we have

3 3/2 r
L3 mL3 mk
 
L 2mω 1
2mω
# valid ~k = 4π = 2
= .
2π ~ 2tω~ 2π 2 ~t
2π ~t
q
In the last step we use the fact that spherical shell is thin to approximate k ≈ 2mω ~ . Thus, when
˜ ~
we sum f (t, α)/t over k we obtain
X f˜(t, α) π L3 mk
= t · # valid ~k = .
t 2 4π ~
~k

We have obtained our factor of L3 that removes the unphysical dependence on the boundary
conditions. Putting everything together we get
256me2 E02
R1,0,0→all ~k = .
3~3 a50 k 9

3 Adiabatic evolution
3.1 The adiabatic approximation
We now turn to a different kind of approximation, in which we consider slowly varying Hamiltonians.
We will consider a time-dependent Hamiltonian H(t). Let |ψn (t)i and En (t) be the “instantaneous”
eigenbases and eigenenergies, defined by

H(t)|ψn (t)i = En (t)|ψn (t)i E1 (t) ≤ E2 (t) ≤ . . . (18)

We also define |Ψ(t)i to be the solution of Schrödinger’s equation: i.e.



i~ |Ψ(t)i = H (t)|Ψ(t)i. (19)
∂t
Beware that (18) and (19) are not the same. You might think of (18) as a naive attempt to solve
(19). If the system starts in |ψn (0)i at time 0, there is of course no reason in general to expect that
|ψn (t)i will be the correct solution for later t. And yet, the adiabatic theorem states that in some
cases this is exactly what happens.
Theorem 1 (Adiabatic theorem). Suppose at t = 0, |Ψ(0)i = |ψn (0)i for some n. Then if H is
changed slowly for 0 ≤ t ≤ T , then at time T we will have |Ψ(T )i ≈ |ψn (T )i.
This theorem is stated in somewhat vague terms, e.g. what does “changed slowly” mean? H˙
should be small, but relative to what? One clue is the reference to the nth eigenstate |ψn (t)i.
This is only well defined if En (t) is unique, so clearly the theorem fails in the case of degenerate
eigenvalues. And since the theorem should not behave discontinuously with respect to H(t), it
should also fail for “nearly” degenerate eigenvalues. This gives us another energy scale to compare
with H˙ (which has units of energy/time, or energy squared once we multiply by ~). We will see
later the sense in which this can be shown to be the right comparison.

15
Example. Suppose we have a spin-1/2 particle in a magnetic field B(t). ~ Then the Hamiltonian
S~ ~ ~ both
is H(t) = ge µB ~ · B(t). The adiabatic theorem says that if we start with the spin and B
~ to point in the x̂ direction, then the spin will
pointing in the +ẑ direction and gradually rotate B
follow the magnetic field and also point in the x̂ direction. Given that the Schrödinger equation
prescribes instead that the spin precess around the magnetic field, this behavior appears at first
somewhat strange.

Derivation We will not rigorously prove the adiabatic theorem, but will describe most of the
derivation. Begin by writing X
|Ψ(t)i = cn (t)|ψn (t)i.
n

Taking derivatives of both sides we obtain


d X X
i~ |Ψ(t)i = i~ ċn (t)|ψn (t)i + cn (t)|ψ˙ n (t)i = cn (t)En (t)|ψn (t)i.
dt n n

Multiply both sides by hψk (t)| and we obtain


X
i~ċk = Ek ck − i~ hψk |ψ˙ n icn . (20)
n

Now we need a way to evaluate hψk |ψ˙ n i in terms of more familiar quantities.

Start with H|ψn i = En |ψn i


d ˙ n i + H |ψ˙ n i = E˙ n |ψn i + En |ψ̇n i
Apply H|ψ
dt
Apply hψk | hψk |Ḣ|ψn i + Ek hψk |ψ˙ n i = E˙ n δkn + En hψk |ψ˙ n i

This equation has two interesting cases: k = n and k = 6 n. The former will not be helpful in
estimating hψk |ψ̇n i, but does give us a useful result, called the Hellmann-Feynman theorem.

k=n E˙ n = hψn |Ḣ|ψn i


hψk |Ḣ|ψn i Ḣkn
6 n hψk |ψ˙ n i =
k= ≡
En − Ek En − Ek

In the last step, we used H˙ kn to refer to the matrix elements of H˙ in the {|ψn i} basis.
Plugging this into (20) we find
X Ḣkn
i~ċk = (Ek − i~hψk |ψ̇k i)ck − i~ cn . (21)
| {z } En − Ek
6
n=k
adiabatic approximation | {z }
error term

If the part of the equation denoted “error term” did not exist, then |ck | would be independent of
time, which would confirm the adiabatic theorem. Furthermore, the error term is suppressed by a
factor of 1/∆nk , where ∆nk ≡ En − Ek is the energy gap. So naively it seems that if H˙ is small
relative to ∆nk then the error term should be small. On the other hand, these two quantities do
not even have the same units, so we will have to be careful.

16
Phases Before we analyze the error term, let’s look at the phases we get if the error term were
not there. i.e. suppose that i~ċk = (Ek − i~hψk |ψ˙ k i)ck . The solution of this differential equation is

ck (t) = ck (0)eiθk (t) eiγk (t) (22a)


1 t t
Z Z
θk (t) ≡ − Ek (t0 )dt0 γk (t) ≡ νk (t0 )dt0 νk (t) ≡ ihψk |ψ˙ k i (22b)
~ 0 0

The θk (t) term is called the “dynamical” phase and corresponds to exactly what you’d expect from
a Hamiltonian that’s always on; namely the phase of state k rotates at rate −Ek /~. The γk (t) is
called the “geometric phase” or “Berry phase” and will be discussed further in the next lecture.
At this point, observe only that it is independent of ~ and that νk (t) can be seen to be real by
applying d/dt to the equation hψk |ψk i = 1.

Validity of the adiabatic approximation Let’s estimate the magnitude of the error term in a
toy model. Suppose that H(t) = H0 + Tt V , where H0 , V are time-independent and T is a constant
that sets the timescale on which V is turned on. Then H˙ = V /T . An important prediction about
the adiabatic theorem is that if the more slowly H changes from H0 to H0 + V , the lower the
probability of transition should be; i.e. increasing T should reduce the error term, even if we
integrate over time from 0 to T .
Let’s see how this works. If the gap is always & ∆, then we can upper-bound the transition
rate by some matrix element of TV∆ . This decreases as T and ∆ increase, which is good. But if we
add up this rate of transitions over time T , then the total transition amplitude can be as large as
∼ V /∆. Thus, going more slowly appears not to reduce the total probability of transition!
What went wrong? Well, we assumed that amplitude from state n simply added up in state
k. But if the states have different energies, then over time the terms we add will have different
phases, and may cancel out. This can be understood in terms of time-dependent perturbation
theory. Define c̃k (t) = e−iθk (t) ck (t). Observe that
d
i~ c̃k (t) = ~θ˙k (t)e−iθk (t) ck (t) + i~e−iθk (t) ċk (t)
dt
X Ḣkn
= −Ek (t)e−iθk (t) ck (t) + e−iθk (t) (Ek (t) − ~νk (t))ck (t) − i~ e−iθk (t) cn (t)
En − Ek
6
n=k
X Ḣkn
= −~νk (t)c̃k (t) − i~ ei(θn (t)−θk (t)) c̃n (t)
En − Ek
6
n=k

In the last step we have used cn (t) = eiθn (t) c̃n (t). Let’s ignore the νk (t) geometric phase term (since
our analysis here is somewhat heuristic). We see that the error term is the same as in (21) but
with an extra phase of ei(θn (t)−θk (t)) . Analyzing this in general is tricky, but let’s suppose that the
energy levels are roughly constant, so we can replace it with e−iωnk t , where ωnk = (En − Ek )/~.
Now when we integrate the contribution of this term from t = 0 to t = T we get

Ḣkn e−iωnk t V e−iωnk T − 1


Z T
V ~V
dt ∼ 2 ∼ 2 ∼ 2
0 En − Ek T ~ωnk T ~ωnk ∆ T

Finally we obtain that the probability of transition decreases with T . This can be thought of as a
rough justification of the adiabatic theorem, but it of course made many simplifying assumptions
and in general it will be only qualitatively correct.

17
This was focused on a specific transition. In general adiabatic transitions between levels m and
n are suppressed if
~|Ḣmn |  ∆2 = min(Em (t) − En (t))2 . (23)
t

Landau-Zener transitions One example that can be solved exactly is a two-level system with
a linearly changing Hamiltonian. Suppose a spin-1/2 particle experiences a magnetic field resulting
in the Hamiltonian
vt
H(t) = ∆σx + σz ,
T
p
for some constants ∆, v, T . The eigenvalues are ± ∆2 + (vt/T )2 . Assuming v > 0, then when
t = −∞ the top eigenstate is |−i and the bottom eigenstate is |+i. When t = ∞ these are reversed;
|+i is the top eigenstate and |−i is the bottom eigenstate. When t = 0, the eigenstates are |+i±|−i

2
.
See diagram on black-board for energy levels.
Suppose that ∆ = 0 and we start in the |−i at t = −∞. Then at t = ∞ we will still be in
the |−i state, with only the phase having changed. But if ∆ > 0 and we move slowly enough then
the adiabatic approximation says we will remain in the top eigenstate, which for t = ∞ will be
|+i. Thus, the presence of a very small transverse field can completely change the state if we move
slowly enough through it.
In this case, the error term in the adiabatic approximation can be calculated rather precisely
and is given by the Landau-Zener formula (proof omitted):

2π 2 ∆2 T
 
Pr[transition] ≈ exp − .
~v

Observe that it has all the qualitative features that we expect in terms of dependence on ∆, v, T ,
but that it corresponds to a rate of transitions exponentially smaller than our above estimate from
first-order perturbation theory. Note that here “transition” refers to transitions between energy
level. Thus starting in |−i and ending in |+i corresponds to “no transition” while ending in |−i
would correspond to “transition,” since it means starting in the higher energy level and ending in
the lower energy level.

3.2 Berry phase


Recall that the adiabatic theorem states that if we start in state |ψn (0)i and change the Hamiltonian
slowly, then we will end in approximately the state

eiθn (t) eiγn (t) |ψn (t)i (24a)


1 t t
Z Z
θn (t) ≡ − En (t0 )dt0 γn (t) ≡ νn (t0 )dt0 νn (t) ≡ ihψn |ψ˙ n i (24b)
~ 0 0

The phase γn (t) is called the geometric phase, or the Berry phase, after Michael Berry’s 1984
explanation of it.
Do the phases in the adiabatic approximation matter? This is a somewhat subtle question. Of
course an overall phase cannot be observed, but a relative phase can lead to observable interference
effects. The phases in (22) depend on the eigenstate label n, and so in principle interference is
possible. But solutions to the equation H(t)|ψn (t)i = En (t)|ψn (t)i are not uniquely defined, and
we can in general redefine |ψn (t)i by multiplying by a phase that can depend on both n and t.

18
To see how this works, let us consider the example of a spin-1/2 particle in a spatially varying
magnetic field. If the particle moves slowly, we can think of the position ~r(t) as a classical variable
causing the spin to experience the Hamiltonian H(~r(t)). This suggests that we might write the
state as a function of ~r(t), as |ψn (~r(t))i or even |ψn (~r)i. If the particle’s position is a classical
function of time, then we need only consider interference between states with the same value of ~r,
and so we can safely change |ψn (~r)i by any phase that is a function of n and ~r.
In fact, even if the particle were in a superposition of positions as in the two-slit experiment,
then we could still only see interference effects between branches of the wavefunction with the same
value of ~r. Thus, again we can define an arbitrary (n, ~r)-dependent phase.
More generally, suppose that H depends on some set of coordinates R(t) ~ = (R1 (t), . . . , RN (t)).
~ ~ ~ ~
The eigenvalue equation (18) becomes H(R)|ψn (R)i = En (R)|ψn (R)i where we leave the time-
dependence of R ~ implicit. This allows us to compute even in situations where R ~ is in a superposition
of coordinates at a given time t.
~
To express γn (t) in terms of |ψn (R)i, we compute
N ~ i dRi ~
d ~ =
X d|ψn (R) ~ i · dR
~ ~ |ψn (R)
|ψn (R)i =∇ R
dt dRi dt dt
i=1
Z t ~ ~
Z R(t)
γn (t) = i hψn |∇ ~ ~ |ψn i · dR dt = ~ ~ |ψn i · dR
ihψn |∇ ~
R dt ~ (0) R
0 R

The answer is in terms of a line integral, which depends only on the path and not on time (unlike
the dynamical phase).
How does this change if we reparameterize |ψn (R ~ )i? Suppose we replace |ψn (R) ~ i with |ψ˜n (R)
~ i=
~ ~
e−iβ(R) |ψn (R)i. Then the Berry phase becomes
~
Z R(t) ~
Z R(t)
~ ~ ~
γ̃n (t) = i hψ˜n (R) ~ ~ |ψ˜n (R)
~ |∇
R
~ i · dR
~ =i ~ |eiβ(R)
hψn (R) ∇ ~ e−iβ (R) |ψn (R)
R
~ i · dR
~
~ (0)
R ~
R(0)
~
= γn (t) + β(R(t)) ~
− β(R(0))
Changing β only changes phases as a function of the endpoints of the path. Thus, we can eliminate
~
the phase for any fixed path with R(t) ~
6= R(0), but not simultaneously for all paths. In particular,
if a particle takes two different paths to the same point, the difference in their phases cannot be
redefined away. More simply, suppose the path is a loop, so that R(0) ~ ~
= R(t). Then regardless of
β we will have γn = γ̃n . This suggests an important point about the Berry phase, which is that it
is uniquely defined on closed paths, but not necessarily open ones.
~
Suppose that R(t) follows a closed curve C. Then we can write
I I
~ ~
γn [C] = ihψn |∇R~ |ψn i ·dR = A ~ n (R)
~ · dR,
~
| {z }
~n (R)
A ~

where we have defined the Berry connection A ~ n (R)


~ = ihψn |∇ ~ ~ |ψn i. Note that it is real for the
R
same reason that νn (t) is real.
In some cases, we can simplify γn [C]. If N = 1 then the integral is always zero, since the line
integral of a closed curve in 1-d is always zero. In 2-d or 3-d we can use Green’s theorem or Stokes’s
theorem respectively to simplify the computation of γn [C]. Let’s focus on 3-d, because it contains
2-d as a special case. Then if S denotes the surface enclosed by curve C, we have
I ZZ ZZ
~ n (R)
A ~ · dR
~ = ~ ~ ×A
(∇ ~ n ) · d~a ≡ D~ n · d~a.
R
C S S

19
Here we define the Berry curvature D ~n = ∇ ~ ~ ×A
~ n and the infinitesimal unit of area d~a. We can
R
~
write Dn in a more symmetric way as follows:
 
X d d X dhψn | d|ψn i d d
(Dn )i = i ijk hψn | |ψn i = i ijk + hψn | |ψn i .
dRj dRk dRj dRk dRj dRk
j,k j,k

d d
Because ijk is antisymmetric in j, k and dR j dRk
is symmetric, the second term vanishes and we
are left with
D~ n = i(∇
~ r hψn |) × (∇
~ r |ψn i). (25)

Example: electron spin in a magnetic field. The Hamiltonian is

~ e~
H = µ~σ · B µ= .
mc

Suppose that B~ = B~r where B is fixed and we slowly trace out a closed path in the unit sphere
with ~r. Suppose that we start in the state
 
  sin(θ) cos(φ)
cos(θ/2)  
|~r; +i = |~ri =  with ~r =  sin(θ) sin(φ) 
  

e sin(θ/2)  
cos(θ)

Then the adiabatic theorem states that we will remain in the state |~ri at later points, up to an
overall phase. To compute the geometric phase observe that

~ = d r̂ + 1 d θˆ + 1

d ˆ
φ.
dr r dθ r sin θ dφ
d
Since dr |~
ri = 0 we have
   
− sin(θ/2) 0
~ ~ri = 1 
∇|  θˆ + 1  ˆ
 φ.
2r eiφ cos(θ/2) r sin θ ieiφ sin(θ/2)
| {z }
|−~
ri

This first term will not contribute to the Berry connection, and so we obtain
2
~ ri = − 1 sin (θ/2) φ.
~ + (~r) = ih~r|∇|~
A ˆ
r sin(θ)

Finally the Berry curvature is

~+ = ∇
~ ×A
~+ = 1 d r̂ d 1
D (sin θA+,φ ) r̂ = − 2 sin2 (θ/2) = − 2 r̂.
r sin θ dθ r sin θ dθ 2r
d d 1−cos θ
For this last computation, observe that dθ sin2 (θ/2) = dθ 2 = sin θ. We can now compute the
Berry phase as ZZ
γ+ [C] = ~ + · d~a = − 1 Ω.
D
S
|{z} 2
r2 dΩr̂

20
Here dΩ is a unit of solid angle, and Ω is the solid angle contained by C.
What if we used a different parameterization for |~ri? An equally valid choice is
 
e−iφ cos(θ/2)
|~ri =  . (26)
sin(θ/2)

If we carry through the same computation we find that now


2
~ + = 1 cos (θ/2) φˆ
A and ~ + = − 1 r̂.
D
r sin(θ) 2r2
d d
We see that the dθ sin2 (θ/2) was replaced by a dθ (− cos2 (θ/2)) which gives the same answer. This
is an example of the general principle that the Berry connection is sensitive to our choice of phase
convention but the Berry curvature is not. Accordingly the Berry curvature can be observed in
experiments.
What if we started instead with the state |~r; −i = | − ~ri? Then a similar calculation would find
that
1
γ− [C] = Ω.
2
Since the two states pick up different phases, this can be seen experimentally if we start in a
superposition of |~r; +i and |~r; −i.
More generally, if we have a spin-s particle, then its z component of angular momentum can be
anything in the range −s ≤ m ≤ s and one can show that

γm [C] = −mΩ.

There is much more that can be said about Berry’s phase. An excellent treatment is found in the
1989 book Geometric phases in physics by Wilczek and Shapere. There is a classical analogue called
Hannay’s phase. Berry’s phase also has applications to molecular dynamics and to understanding
electrical and magnetic properties of Bloch states. We will see Berry’s phase again when we discuss
the Aharonov-Bohm effect in a few weeks.

3.3 Neutrino oscillations and the MSW effect


In this section we will discuss the application of the adiabatic theorem to a phenomenon involving
solar neutrinos. The name neutrino means “little neutral one” and neutrinos are spin-1/2, electri-
cally neutral, almost massless and very weakly interacting particles. Neutrinos were first proposed
by Pauli in 1930 to explain the apparent violation of energy, momentum and angular momentum
conservation in beta decay. (Since beta decay involves the decay of a neutron into a proton, an
electron and an electron antineutrino, but only the proton and electron could be readily detected,
there was an apparent anomaly.)
Their almost complete lack of interaction with matter (it takes 100 lightyears of lead to ab-
sorb 50% of a beam of neutrinos) has made many properties of neutrinos remain mysterious.
Corresponding to the charged leptons e− , e+ (electron/position), µ− /µ+ (muon/antimuon) and
τ − /τ + (tau/antitau), neutrinos (aka neutral leptons) also exist in three flavors: νe , νµ , ντ , with
antineutrinos denoted ν̄e , ν̄µ , ν̄τ . Most, but not all, interactions preserve lepton number, defined
to be the number of leptons minus the number of antileptons. Indeed, most interactions preserve
#e− + #νe − #e+ − #ν̄e (electronic number) and similarly for muons and taus. However, these
quantities are not conserved by neutrino oscillations. Even the total lepton number is violated by
a phenomenon known as the chiral anomaly.

21
Solar neutrinos Solar neutrinos are produced via the p-p chain reaction, which converts (via a
series of reactions)
4 1H = 4p+ + 4e− 7→ 2p+ + 2n + 2e− +2νe .
| {z }
4 He

The resulting neutrinos are produced with energies in the range 0.5-20MeV. Almost all of neutrinos
produced in the sun are electron neutrinos.

Detection Neutrinos can be detected via inverse beta decay, corresponding to the reaction

A + νe 7→ A0 + e− ,

where A, A0 are different atomic nuclei. For solar neutrinos this will only happen for electron
neutrinos because the reaction A + νµ 7→ A0 + µ− will only happen for mu neutrinos carrying
at least 108 MeV of kinetic energy. So it is easiest to observe electron neutrinos. However other
flavors of neutrinos can also be detected via more complicated processes, such as neutrino-mediated
disassociation of deuterium.

Observations of solar neutrinos The first experiment to detect cosmic neutrinos was the 1968
Homestake experiment, led by Ray Davis, which used 100,000 gallons of dry-cleaning fluid (C2 Cl4 )
to detect neutrinos via the process 37 Cl + νe 7→37 Ar + e− . However, this only found about 1/3 as
many neutrinos as standard solar models predicted.
In 2002, the Sudbery Neutrino Observatory (SNO) measured the total neutrino flux and found
that once mu- and tau-neutrinos were accounted for, the total number of neutrinos was correct.
Thus, somehow electron neutrinos in the sun had become mu and tau neutrinos by the time they
reached the Earth.

Neutrino oscillations The first high-confidence observations of neutrino oscillations were by


the Super Kamiokande experiment in 1998, which could distinguish electron neutrinos from muon
neutrinos. Since neutrinos oscillate, they must have energy, which means they must have mass
(if we wish to exclude more speculative theories, such as violations of the principle of relativity).
This means that a neutrino, in its rest frame, has a Hamiltonian with eigenstates |ν1 i, |ν2 i, |ν3 i
that in general will be different from the flavor eigenstates |νe i, |νµ i, |ντ i that participate in weak-
interaction processes such as beta decay.
We will treat this in a simplified way by neglecting |ν3 i and |ντ i. So the Hamiltonian can be
modeled as (in the |νe i, |νµ i basis)  
Ee ∆
H= , (27)
∆ Eµ
where Ee , Eµ are the energies (possibly equal) of the electron and muon neutrinos and ∆ represents
a mixing term. Unfortunately, plugging in known parameter estimates for the terms in (27) would
predict that roughly a 0.57 fraction of solar neutrinos would end up in the |νe i state, so this still
cannot fully explain our observations.

The MSW effect It turns out that this puzzle can be resolved by a clever use of the adiabatic
theorem. Electron neutrinos scatter off of electrons and thus the Hamiltonian in (27) should be

22
modified to add a term proportional to the local density of electrons. Thus after some additional
rearranging, we obtain
   
−∆0 cos(2θ) ∆0 sin(2θ) CNe 0
H = E0 +  + , (28)
∆0 sin(2θ) ∆0 cos(2θ) 0 0

where ∆0 , θ come from (27) (θ ≈ π/6 is the “mixing angle” that measures how far the flavor states
are from being eigenstates), C is a constant and Ne = Ne (~r) is the local electron density. If the
neutrino is traveling at speed ≈ c in direction x̂, then ~r ≈ ctx̂. Thus we can think of Ne as
time-dependent. We then can rewrite H as
 
CNe (t)
const · I + − ∆0 cos(2θ) σz + ∆0 sin(2θ)σx . (29)
2

This looks like the adiabatic Landau-Zener transition we studied in the last lecture, although here
the σz term is no longer being swept from −∞ to +∞. Instead, near the center of the sun, Ne (0)
is large and the eigenstates are roughly |νe i, |νµ i. For large t, the neutrinos are in vacuum, where
their eigenstates are |ν1 i, |ν2 i.
If the conditions of the adiabatic theorem are met, then neutrinos that start in state |νe i (in
the center of the sun) will emerge in state |ν2 i (at the surface of the sun). They will then remain
in this state as they propagate to the Earth. It turns out that this holds for neutrinos of energies
' 2M eV . In this case, the probability of observing the neutrino on Earth in the |νe i state (thinking
of neutrino detectors as making measurements in the flavor basis) is sin2 (θ), which gives more or
less the observed value of 0.31.

3.4 Born-Oppenheimer approximation


Consider a system with N nuclei and n electrons. Write the Hamiltonian as
N
X ~2 ~ 2
H=− ∇ + Hel (R). (30)
2Mj R~ j
j=1

Here R = (R ~ 1, . . . , R
~ N ) denotes the positions of the N nuclei and Hel (R) includes all the other
terms, i.e. kinetic energy of the electrons as well as the potential energy terms which include
electron-electron, nuclei-nuclei and electron-nuclei interactions. Let r denote all of the coordinates
of the electrons. While (30) may be too hard to solve exactly, we can use a version of the adiabatic
theorem to derive an approximate solution.
We will consider a product ansatz:

Ψ(R, r) = γ(R)ΦR (r), (31)

where the many-electron wavefunction is an eigenstate of the reduced Hamiltonian:

Hel (R)ΦR (r) = Eel (R)ΦR (r). (32)

(Typically this eigenstate wil be simply the ground state.) This is plausible because of the adiabatic
theorem. If the nuclei move slowly then as this happens the electrons can rapidly adjust to remain
in their ground states. Then once we have solved (32) we might imagine that we can substitute

23
back to solve for the nuclear eigenstates. We might guess that they are solutions to the following
eigenvalue equation  
N 2
−
X ~ ~ 2 + Eel (R) γ(R) = Eγ(R).
∇ (33)
2Mj R~ j
j=1

~ ~ to (31) we obtain
However, this is not quite right. If we apply ∇ Rj

~ ~ Ψ(R, r) = (∇
∇ ~ ~ γ(R))ΦR (r) + γ(R)∇
~ ~ ΦR (r). (34)
Rj Rj Rj

Using the adiabatic approximation we neglect the overlap of ∇ ~ ~ Ψ(R, r) with all states to |ΦR i.
Rj
Equivalently we can multiply on the left by hΦR |. This results in
Z Z
d rΦR (r) ∇R~ j Ψ(R, r) = ∇R~ j γ(R) + γ(R) d3n rΦR (r)∗ ∇
3n ∗~ ~ ~ ~ ΦR (r)
Rj

~ ~ − iA
= (∇ ~ j )γ(R),
Rj

~ j is the familiar Berry connection


where A
~ j = ihΦR |∇
A ~ ~ |ΦR i. (35)
Rj

We conclude that the effective Hamiltonian actual experienced by the nuclei should be
N
X ~2 ~ ~ j )2 + Eel (R).
Heff = (∇ ~ − iA (36)
2Mj Rj
j=1

We will see these A ~ j terms again when we discuss electomagnetism later in the semester. In
systems of nuclei and atoms we need at least three nuclei before the A ~ j terms can have an effect,
for the same reason that we do not see a Berry phase unless we trace out a loop in a parameter
space of dimension ≥ 2.
The Born-Oppenheimer applies not just to nuclei and electrons but whenever we can divide a
system into fast and slow-moving degrees of freedom; e.g. we can treat a proton as a single particle
and ignore (or “integrate out”) the motion of the quarks within the proton. This is an important
principle that we often take for granted. Some more general versions of Born-Oppenheimer are
called “effective field theory” or the renormalization group.

4 Scattering
4.1 Preliminaries
One of the most important types of experiments in quantum mechanics is scattering. A beam of
particles is sent into a potential and scatters off it in various directions. The angular distribution
of scattered particles is then measured. In 8.04 we studied scattering in 1-d, and here we will
study scattering in 3-d. This is an enormous field, and we will barely scratch the surface of it. In
particular, we will focus on the following special case:
• Elastic scattering. The outgoing particle has the same energy as the incoming particle. This
means we can model the particles being scattered off semi-classically, as a static potential
V (~r). The other types of scattering are inelastic scattering, which can involve transformation
of the particles involved or creation of new particles, and absorption, in which there is no
outgoing particle.

24
• Non-relativistic scattering. This is by contrast with modern accelerators such as the LHC.
However, non-relativistic scattering is still relevant to many cutting-edge experiments, such
as modern search for cosmic dark matter (which is believed to be traveling at non-relativistic
speeds).
Even this special case can teach us a lot of interesting physics. For example, Rutherford scattering1
showed that atoms have nuclei, thereby refuting the earlier “plum pudding” model of atoms. This
led to a model of atoms in which electrons orbit nuclei like planets, and resolving the problems of
this model in turn was one of the early successes of quantum mechanics.

Scattering cross section: In scattering problems it is important to think about which physical
quantities can be observed. The incoming particles have a flux that is measured in terms of number
2N
of particles per unit area per unit time, i.e. ddAdt
in
. If we just count the total number of scattered
particles, then this is measured in terms of particles per time: dNdtscat . The ratio of these quantities
has units of area and is called the scattering cross section:
dNscat
dt
d2 Nin
= σ. (37)
dAdt

To get a sense of why these are the right units, consider scattering of classical particles off of a
classical hard sphere of radius a. If a particle hits the sphere it will scatter, and if it does not hit
the sphere it will not scatter. Assume that the beam of particles
p is much wider than the target,
i.e. each particle has trajectory ~r = (x0 , y0 , z0 + vt) with x20 + y02 given by a distribution
p with
standard deviation that is  a. The particles that scatter will be the ones with x0 + y0 ≤ a 2 2

which corresponds to a region with area πa2 , which is precisely the cross-sectional area of the
2
sphere. Since we have dNdtscat = ddANdt
in
πa2 , it follows that σ = πa2 . This simple example is good to
keep in mind to have intuition about the meaning of scattering cross sections.

Differential cross-section: We can get more information out of an experiment by measuring


the angular dependence of the scattered particles. The number of scattered particles can then be
2N
measured in terms of a rate per solid angle, i.e. ddΩ dσ
dt . The resulting differential cross-section dΩ
scat

is defined to be
d2 Nscat

(θ, φ) ≡ ddΩ 2N
dt
(38)
dΩ in
dAdt
Here the spherical coordinates (θ, φ) denote the direction of the outgoing particles. It is conventional
to define the axes so that the incoming particles have momentum in the ẑ direction, so θ is the angle
between the scattered particle and the incoming beam (i.e. θ = 0 means no change in direction
and θ = π means backwards scattering) while φ is the azimuthal angle. Integrating over all angles
gives us the full cross-section, i.e. Z

σ = dΩ . (39)
dΩ

Quantum mechanical scattering: Assume that the incoming particle states are wavepackets
that are large relative to the target. This allows us to approximate the incoming particles as plane
wave, i.e.
iEt
ψin ∝ eikz− ~ , (40)
1
Rutherford scattering is named after Ernest Rutherford for his 1911 explanation of the 1909 experiment which
was carried out by Geiger and Marsden.

25
2 2
k
where E = ~2m . Here we need to assume that the potential V (~r) → 0 as r → ∞ so that plane
waves are solutions to the Schrödinger equation for large r. For the scattered wave, we should seek
solutions satisfying
~2 ~ 2
− ∇ ψscat = Eψscat as r → ∞ (41)
2m 
1 d2

1 ˆ2
− 2
r + 2L ψscat = k 2 ψscat in spherical coordinates (42)
r dr r
A general solution can be written as a superposition of separable solutions. Separable solutions to
(42) can in turn be written as
rψ(r, θ, φ) = u(r)f (θ, φ), (43)
in terms of some functions u(r), f (θ, φ). In terms of these (42) becomes
1 ˆ2
u00 f + 2
uL f +k 2 uf = 0. (44)
r
| {z }
→0 as r→∞

Thus, for large r, we can cancel the f from each side and simply have u00 = −k 2 u, which has
solutions e±ikr . The eikr solution corresponds to outgoing waves and the e−ikr solution to incoming
waves. A scattered wave should be entirely outgoing, and so we obtain
r→∞ f (θ, φ) ikr− iEt
ψscat = e ~ (45)
r
or more precisely  
f (θ, φ) ikr− iEt 1
ψscat = e ~ + O . (46)
r r2
Because the scattering is elastic, the k and E here are the same as for the incoming wave.

Time-independent formulation: As with 1-d scattering problems, the true scattering process
is of course time-dependent, but the quantities of interest (transmission/reflection in 1-d, differential
cross section in 3-d) can be extracted by solving the time-independent Schrödinger equation with
suitable boundary conditions. In the true process, the incoming wave should really be a wavepacket
with well-defined momentum ≈ (0, 0, k) and therefore delocalized position. The outgoing wave will
be a combination of an un-scattered part, which looks like the original wave packet continuing
forward in the ẑ direction, and a scattered part, which is a spherical outgoing wavepacket with a
f (θ, φ) angular dependence. However, we can treat the incoming wave instead as the static plane
wave eikz and the scattered wave instead as the static outgoing wave f (θ,φ) r e
ikr . (Both of these

are when r → ∞.) Thus we can formulate the entire scattering problem as a time-independent
boundary-value problem. The high-level strategy is then to solve the Schrödinger equation subject
to the boundary conditions
r→∞ f (θ, φ) ikr
ψ(~r) = eikz + e . (47)
r
This is analogous to what we did in 1-D scattering, where the boundary conditions were that ψ(x)
should approach eikx + Re−ikx for x → −∞ and should approach T eikx for x → ∞. As in the 1-D
case, we have to remember that this equation is an approximation for a time-dependent problem.
As a result when calculating observable quantities we have to remember not to include interference
terms between the incoming and reflected waves, since these never both exist at the same point in
time.

26
Relation to observables. In the 1-D case, the probabilities of transmission and reflection are
|T |2 and |R|2 respectively. In the 3-d case, the observable quantities are the differential cross

sections dΩ . To compute these, first evaluate the incoming flux

~ ikz = ~k ẑ = vẑ
 
~in = ~ Im ψin
S ∗ ~
∇ψin =
~
Im e−ikz ∇e (48)
m m m
In the last step we have used v = ~k/m. The units
√here are off because technically the wavefunction
ikz ikz
should be not e but something more like e / V , where V has units of volume. But neglecting
this factor in both the numerator and denominator of (38) will cause this to cancel out. Keeping
that in mind, we calculate the denominator to be

d2 Nin ~in | = v.
= |S (49)
dAdt
~ = ∂ 1 ∂ ˆ 1 ∂ ˆ
Similarly the outgoing flux is (using ∇ ∂r r̂ + r ∂θ θ + r sin θ ∂φ φ)

e−ikr ∗ eikr
  
~scat = ~ Im 1 r̂
S f ikr̂ f +O =v |f |2 + O(1/r3 ). (50)
m r r r3 r2

To relate this to the flux per solid angle we use d~a = r2 dΩr̂ to obtain

d2 Nscat ~scat · d~a = (v r̂ |f |2 + O(r−3 )) · r2 dΩ r̂ = |f |2 vdΩ + O(1/r).


=S (51)
dΩ dt r2
We can neglect the O(1/r) term as r → ∞ (and on a side note, we see now why the leading-order
~scat was O(1/r2 )) and obtain the simple formula
term in S


= |f (θ, φ)|2 . (52)
dΩ

The Optical Theorem. (52) is valid everywhere except at θ = 0. There we have to also
consider interference between the scattered and the unscattered wave. (Unlike the incoming wave,
the outgoing unscattered does coexist with the scattered wave.) The resulting flux is

f ∗ −ikr
  
~ ~ −ikz ikz ikr̂ ikr 2
Sout = Im e + e ik ẑe + f e + O(1/r ) (53a)
m r r
 ∗ 
r̂ ẑf ik(z−r) r̂f ik(r−z)
vẑ + v 2 |f |2 + v Re
= |{z} e + e (53b)
r {z } r r
~unscat
S
| | {z }
~scat
S ~interference
S

This last term can be thought of as the peffects of interference. We will evaluate it for large r and
ρ2
for θ ≈ 0. Here r̂ ≈ ẑ and we define ρ = x2 + y 2 so that (to leading order) r = z + 2z . Then
Z Z 2π Z ∞  
~inter · d~a = v zˆ 2
∗ −ik ρ2z
2
ik ρ2z
S dφ ρdρ f e + fe (54a)
0 0 z
Z ∞ 2
v dρ ikρ2
= 4π Re f e 2z (54b)
z 2
Z0 ∞
2πv iky 4πv
= Re dye 2z f (0) = − Im f (0) (54c)
z 0 k

27
Since the outgoing flux should equal the incoming flux, we can define A to be the beam area and
find Z
4πv
Av = Av + v dΩ|f (θ, φ)|2 − Im f (0). (55)
k
Thus we obtain the following identity, known as the optical theorem:
Z

dΩ|f (θ, φ)|2 = Im f (0) (56)
k
All this is well and good, but we have made no progress at all in computing f (θ, φ). We will
discuss two approximation methods: the partial wave method (which is exact, but yields nice
approximations when k is very small) and the Born approximation, which is a good approximation
when we are scattering off a weak potential. This is analogous to approximations we have seen
before (see Table 1). We will discuss the Born approximation in Section 4.2 and the partial-wave

perturbation type time-independent time-dependent scattering

small TIPT TDPT Born

slow WKB adiabatic partial wave

Table 1: Summary of approximation techniques for scattering problems.

technique in Section 4.3.

4.2 Born Approximation


Zooming out, we want to solve the following eigenvalue equation:

~ 2 + k 2 )|ψi = U |ψi 2m
(∇ where U≡ V. (57)
~2
~ 2 + k 2 ) to obtain
This looks like a basic linear algebra question. Can we solve it by inverting (∇
? ~2
ψ = (∇ + k 2 )−1 U |ψi? (58)

To answer this, we first review some basic linear algebra. Suppose we want to solve the equation

A~x = ~b (59)

for some normal matrix A. We can write ~b = A−1 ~x only if A is invertible. Otherwise the solution
will not be uniquely defined. More generally, suppose that our vectors live on a space V . Then we
can divide up V as

V = Im A ⊕ ker A where ker A = {~x0 : A~x0 = 0}. (60)

If we restrict A to the subspace Im A then it is indeed invertible. The solutions to (59) are then
given by
~x = (A|Im A )−1~b + ~x0 where ~x0 ∈ ker A. (61)
Returning now to the quantum case, the operator (∇ ~ 2 + k 2 ) is certainly not invertible. States
~ 2 + k 2 )|ψ0 i = 0 exist, and are plane waves with momentum ~k. But if we restrict
satisfying (∇

28
~ 2 + k 2 ) to the subspace of states with momentum =
(∇ 6 ~k then it is invertible. Define the Green’s
~ 2 2 −1
operator G to be (∇ + k )|p=~k 6 . The calculation of G is rather subtle and the details can be
found in Griffiths. However, on general principles we can make a fair amount of progress. Since
~ 2 + k 2 ) is diagonal in the momentum basis, then G should be as well. Thus G should be written
(∇
as an integral over |p~ihp~| times some function of p~. By Fourier transforming this function we can
equivalently write G in terms of translation operators as
Z
r ·p
~ ~
G = d3~rG(~r)T~r where T~r ≡ e−i ~ . (62)

R 3 go through this more concretely. In the momentum basis we have the completeness relation
Let’s
d p~|p~ihp~| = I which implies
Z
~
∇ + k = d3 p~(−~2 p2 + k 2 )|p~ihp~|.
2 2
(63)

R0 R0
To invert this we might naively write G = d3 p~(−~2 p2 + k 2 )−1 |p~ihp~| where denotes the integral
over all p~ with p 6= ~k. To handle the diverging denominator, one method is to write
Z
G = lim d3 p~(−~2 p2 + k 2 + i)−1 |p~ihp~|. (64)
→0

Finally we can write this in the position basis according to (62) and obtain the position-space
Green’s function G(~r) by Fourier-transforming (−~2 p2 + k 2 + i)−1 . In Griffiths this integral is
carried out obtaining the answer:
eikr
G(~r) = − . (65)
4πr
This function G(~r) is called a Green’s function. We can thus write

−eikr
Z
G= d3~r T~r . (66)
4πr

Having computed G, we can now solve (57) and obtain

|ψi = |ψ0 i + GU |ψi (67)

for some free-particle solution |ψ0 i. Indeed for a scattering problem, we should have ψ0 (~r) = eikz .
(67) is exact, but not very useful because |ψi appears on both the LHS and RHS. However, it will
let us expand |ψi in powers of U .
The first Born approximation consists of replacing the |ψi on the RHS of (67) by |ψ0 i, thus
yielding
|ψi = |ψ0 i + GU |ψ0 i. (68)
The second Born approximation consists of using (68) to approximate |ψi in the RHS of (67), which
yields
|ψi = |ψ0 i + GU (|ψ0 i + GU |ψ0 i) = |ψ0 i + GU |ψ0 i + GU GU |ψ0 i. (69)
Of course we could also rewrite (67) as |ψi = (I − GU )−1 |ψ0 i = n≥0 (GU )n |ψ0 i (since (I − GU )
P
is formally invertible) and truncate this sum at some finite value of n.

29
These results so far have been rather abstract. Plugging in (65) and ψ0 (~r) = eikz we find that
the first Born approximation is
Z
ψ(~r) = ψ0 (~r) + d3~r0 ψ0 (~r0 )G(~r − ~r0 )U (~r0 ) (70a)
Z ~ r0 |
ik|r−~
3 0 ikz 0 e
ikz
= e − d ~r e U (~r0 ) (70b)
4π|~r − ~r0 |

If we assume that the potential is short range and we evaluate this quantity for r far outside the
range of U , then we will have r  r0 for all the points where the integral has a nonzero contribution.
In this case |~r − ~r0 | ≈ r − r̂ · ~r0 . Let us further define
~k = kr̂ and ~k 0 = kẑ, (71)

corresponding to the outgoing and incoming wavevectors respectively. Then we have (still in the
first Born approximation)
0
eik|r−~
~ r|
Z
0
ψscat (~r) = − d3~r0 eikz U (~r0 ) (72a)
4π|~r − ~r0 |
Z ~0
ikr−ikr̂·r
0e
≈ − d3~r0 eikz U (~r0 ) (72b)
4πr
r0 ) eikr
 Z 
3 0 i(k ~ r0 U (~
~ 0 −k)·~
= − d ~r e (72c)
4π r

The quantity in parentheses is then f (θ, φ). If we define V˜ (~q) to be the Fourier transform of V (~r)
then we obtain
m ˜ ~0 ~
f1 (θ, φ) = − V (k − k), (73)
2π ~2
where f1 refers to the first Born approximation. One very simple example is when V (~r) = V0 δ(~r).
mV0
Then we simply have f1 = − 2π~ 2.
A further simplification occurs in the case when V (~r) is centrally symmetric. Then
Z Z ∞ Z 1 Z ∞

Ṽ (~q) ≡ d3~rV (r)eiq~·~r = 2π dr dµr2 V (r)eiqrµ = dr rV (r) sin(qr). (74)
0 −1 q 0

Finally the momentum transfer ~q = ~k 0 − ~k satisfies q = 2k sin(θ/2).


One application of this (see Griffiths for details) is the Yukawa potential: V (r) = −βe−µr /r.
The first Born approximation yields

2mβ 2
f1Yukawa (θ) = − .
~2 (µ2 + q 2 )

Taking β = −eQ and µ = 0 recovers Rutherford scattering, with


2meQ meQ
f1Rutherford (θ) = 2 2
= 2 2 2 .
~ q 2~ k sin (θ/2)

A good exercise is to rederive the Born approximation by using Fermi’s Golden Rule and
counting the number of outgoing states in the vicinity of a given ~k. See section 7.11 of Sakurai for
details. Another version is in Merzbacher, section 20.1.

30
Rigorous derivation of Green’s functions. The above derivation of G was somewhat informal.
However, once the form of G(~r) is derived informally or even guessed, it can be verified that
~ 2 + k 2 )G acts as the identity on all states with no component in the null space of (∇
(∇ ~ 2 + k 2 ).
This is the content of Griffiths problem 11.8. Implicit is that we are working in a Schwartz space
which rules out exponentially growing wavefunctions, and in turn implies that ∇ ~ 2 has only real
eigenvalues and is in fact Hermitian. A more rigorous derivation of the Born approximation can
also be obtained by using time-dependent perturbation theory as described by Sakurai section 7.11
and Merzbacher section 20.1. I particularly recommend the discussion in Merzbacher.

4.3 Partial Waves


In this section assume we have a central potential, i.e. V (~r) = V (r). Since our initial conditions
are invariant under rotation about the ẑ axis, our scattering solutions will also be independent of
φ (but not θ).
Assume further that
lim r2 V (r) = 0. (75)
r→∞
We will see the relevance of this condition shortly.
A wavefunction with no φ dependence can be written in terms of Legendre polynomials2 (cor-
responding to the m = 0 spherical harmonics)as

X
ψ(r, θ) = Rl (r)Pl (cos θ). (76)
l=0

If we define ul (r) = rRl (r), then the eigenvalue equation becomes (for each l)
2mV l(l + 1)
− u00l + Veff ul = k 2 ul where Veff ≡ 2
+ . (77)
~ r2
If (75) holds, then for sufficiently large r, we can approximate Veff ≈ l(l + 1)/r2 . In this region
the solutions to (77) are given by the spherical Bessel functions. Redefining x ≡ kr and using the
assumption Veff ≈ l(l + 1)/r2 , (77) becomes
l(l + 1)
u00l − ul = −ul . (78)
x2
This has two linearly independent solutions: xjl (x) and xnl (x) where jl (x), nl (x) are the spherical
Bessel functions of the first and second kind respectively, and are defined as
1 d l sin(x) 1 d l cos(x)
   
l l
jl (x) = (−x) and nl (x) = −(−x) (79)
x dx x x dx x
These can be thought of as “sin-like” and “cos-like” respectively. For l = 0, 1, (79) becomes
sin(x) sin(x) cos(x)
j0 (x) = j1 (x) = − (80a)
x x2 x
cos(x) cos(x) sin(x)
n0 (x) = − n1 (x) = − − (80b)
x x2 x
2
R1
What are Legendre polynomials? One definition starts with the orthogonality condition −1 Pm (x)Pn (x)dx =
2 2
δ . (The δmn is the important term here; 2n+1
2n+1 mn
is a somewhat arbitrary convention.) Then if we apply the
Gram-Schmidt procedure to 1, x, x , . . . , we obtain the Legendre polynomials P0 = 1, P1 = 1, P2 = 12 (3x2 − 1), . . ..
2

Thus any degree-n polynomial can be written as a linear combination of P0 , . . . , Pn and vice-versa.
The reason for (76) is that if ψ(r, θ) is independent of φ then we can write it as a power series in r and z, or
equivalently r and zr = cos(θ). These power series can always be written in the form of (76).

31
One can check (by evaluating the derivatives in (79) repeatedly and keeping only the lowest power
of 1/x) that as x → ∞ we have the asymptotic behavior
1 1
jl (x) → sin(x − lπ/2) and nl (x) → − cos(x − lπ/2). (81)
x x
On the other hand, in the x → 0 limit we can keep track of only the lowest power of x to find that
as x → 0 we have
xl (2l − 1)!!
jl (x) → and nl (x) → − , (82)
(2l + 1)!! xl+1
where (2l + 1)!! ≡ (2l + 1)(2l − 1) · · · 3 · 1.
If jl and nl are sin-like and cos-like, it will be convenient to define functions that resemble
incoming and outgoing waves. These are the spherical Hankel functions of the first and second
kind:
(1) (2)
hl = jl + inl and hl = jl − inl (83)
(1) ikr (1)
For large r, hl (kr) → (−i)l+1 ekr , and so our scattered wave should be proportional to hl . More
(1)
precisely, Rl (r) (i.e. the angular-momentum-l component) should be proportional to hl (kr).
Putting this together we get
when V ≈0 ikz
X (1)
ψ(r, θ) = e + cl hl (kr)Pl (cos(θ)), (84)
l≥0
| {z }
ψscat

for some coefficients cl . It will be convenient to write the cl in terms of new coefficients al as
cl = kil+1 (2l + 1)al , so that we have
when V ≈0 (1)
X
ψscat (r, θ) = k il+1 (2l + 1)al hl (kr)Pl (cos(θ)) (85)
l≥0

when r→∞
X eikr
= (2l + 1)al Pl (cos(θ)) (86)
r
l≥0
| {z }
f (θ)

We can then compute the differential cross-section in terms of the al


2
dσ 2
X
= |f (θ)| = (2l + 1)al Pl (cos(θ)) . (87)

dΩ
l

Recall that the Legendre polynomials satify the orthogonality relation


Z 1
2
dzPl (z)Pl0 (z) = δl,l0 . (88)
−1 2l + 1

Thus we can calculate Z


dσ X
σ= dΩ = 4π (2l + 1)|al |2 . (89)
dΩ
l
While the differential cross section involves interference between different values of l, the total cross
section is simply given by an incoherent sum over l. Intuitively this is because we can think of l and
θ as conjugate observables, analogous to momentum and position. If we measure the probability

32
of observing something at a particular position, we will see interference effects between different
momenta, but if we integrate over all positions, these will go away.
The beauty of the partial-wave approach is that it reduces to a series of 1-d scattering problems,
one for each value of l. We have written down the form of the scattered wave. For the incoming
wave, we need to express
P eikz in the form of (76). From the arguments above we can see that eikz
should have form l (Al jl (kr) + Bl nl (kr))Pl (cos θ). The specific solution is given by Rayleigh’s
formula which we state without proof:
X
eikz = eikr cos θ = il (2l + 1)jl (kr)Pl (cos θ) (90)
l≥0

Plugging this in, we find that in the V ≈ 0 region we have


 
when V ≈0 (1)
X
ψ(r, θ) = il (2l + 1)Pl (cos θ)  jl (kr) + ikal hl (kr) (91)
 
| {z } | {z }
l≥0
plane wave scattered
 
1X  (1) (2)
= il (2l + 1)Pl (cos θ) hl (kr)(1 + 2ikal ) + hl (kr) (92)

2 | {z } | {z }
l≥0
outgoing incoming

Now we have really expressed this as a series of 1-d scattering problems. Here comes the crucial
simplifying move. Because we are assuming that the collision is elastic, probability and angular-
momentum conservation means that ”what goes in must come out”; i.e.
|1 + 2ikal | = 1. (93)
The outgoing wave’s amplitude must have the same absolute value as the incoming wave’s ampli-
tude. We can rewrite (93) by introducing the phase shift δl defined by
1 + 2ikal = e2iδl . (94)
The factor of 2 is conventional. We can rewrite (94) to solve for al as
2eiδl − 1 eiδl sin(δl ) 1 1
al = = = . (95)
2ik k k cot(δl ) − i
Many equivalent expressions are also possible. In terms of the phase shifts, we can write
1X
f (θ) = (2l + 1)Pl (cos θ)eiδl sin(δl ) (96)
k
l
4π X
σ= 2 (2l + 1) sin2 (δl ). (97)
k
l

As an application we can verify that this satisfies the optical theorem. Observe that Pl (1) = 1 for
all l. Thus
4π 4π 1 X
Im f (0) = (2l + 1)Pl (1) sin2 (δl ) = σ. (98)
k k k
l
Another easy application is partial-wave unitarity, which bounds the total amount of scattered
wave with angular momentum l. Define σl = 4π 2
P
k2
(2l + 1) sin (δ l ), so that σ = l l Then using
σ .
sin2 (δl ) ≤ 1 we have

σl ≤ 2 (2l + 1). (99)
k

33
This bound is called “partial-wave unitarity.”
How to compute phase shifts. Let us look again at the r → ∞ solution. We can use the fact
that overall phase and normalization don’t matter to obtain
(1) (2)
Rl (r) = hl (kr)e2iδl + hl (kr) (100a)
2iδl 2iδl
= (1 + e )jl (kr) + i(e − 1)nl (kr) (100b)
∝ cos(δl )jl (kr) − sin(δl )nl (kr) (100c)
∝ jl (kr) − tan(δl )nl (kr) (100d)

Suppose that we know the interior solution Rl (r) for r ≤ b and that V (r) ≈ 0 for r > b. Then we
R0 (r)
can compute the phase shift by matching Rll (r) for r = b ± .
Here is a simple example. Consider a hard sphere of radius b. Then Rl (r) = 0 for r ≤ b and we
have  
−1 jl (kb)
jl (kb) − tan(δl )nl (kb) = 0 =⇒ δl = tan . (101)
nl (kb)
One particularly simple case is when l = 0. Then
  sin(kb)
!
j0 (kb)
δ0 = tan−1 = tan−1 kb
= − tan−1 (tan(kb)) = −kb. (102)
n0 (kb) − cos(kb)
kb

It turns out that in general repulsive potentials yield negative phase shifts.
What about larger values of l? Suppose that kb  1. Then using the x → 0 approximations
for jl (x), nl (x), we obtain

(kb)2l+1 (kb)2l+1
 
k→0 −1
δl −−−→ tan ≈ ∼ (kb)2l+1 . (103)
(2l + 1)!!(2l − 1)!! 2l + 1!!(2l − 1)!!

Thus the l = 0 scattering dominates. In terms of cross-sections, we have


4π 4π
σl = (2l + 1) sin2 (δl ) ≈ (kb)4l b2 . (104)
k2 (2l + 1)((2l − 1)!!)4

For l = 0 this is 4πb2 which four times the classical value of πb2 , and for higher value of l this
drops exponentially (assuming kb  1). Even if kb > 1 this drops exponentially once l  kb, which
confirms our intuition that the angular momentum should be on the order of ~kb.
Another way to think about this reason to favor low values of l is because the l(l + 1)/r2
term in Veff forms an angular momentum “barrier” that prevents low-energy incoming waves from
penetrating to small enough r to see the potential V (r).
The high-energy limit. The partial wave approximation is easiest to use in the low-k limit
because then we can restrict our attention to a few values of l, or even just l = 0. But for the hard
sphere we can also evaluate the kb  1 limit. In this case we expect to find angular momenta up
to lmax ≡ kb. Thus we approximate the total cross section by
lmax
4π X
σ≈ (2l + 1) sin2 (δl ). (105)
k2
l=0

34
The phases δl will vary over the entire range from 0 to 2π so we simply approximate sin2 (δl ) by its
average value of 1/2. Thus we obtain
kb
2π X
σ≈ (2l + 1) = 2πb2 . (106)
k2
l=0

This is now twice the classical result. Even though the particles are moving quickly they still
diffract like waves. One surprising consequence is that even though a hard sphere leaves a shadow,
there is a small bright spot at the center of the shadow. Indeed the optical theorem predicts that
k
Im f (0) = 4π σ ≈ kb2 /2. Thus |f (0)|2 ≥ (kb)2 b2 /4. For a further discussion of this bright spot and
the role it played in early 19th-century debates about whether light is a particle and/or a wave,
look up “Arago spot” on wikipedia.
Phase shifts. As we have seen, scattering can be understood in terms of phase shifts. Now we
describe a simple physical way of seeing this. If V = 0 then a plane wave has u0 (r) = sin(kr),
due to the u0 (0) = 0 boundary condition. When there is scattering, the phase shift δ0 will become
nonzero and we will have
u0 (r) = sin(kr + δ0 ).
If the potential is attractive then the phase will oscillate more rapidly in the scattering region and
so we will have δ0 > 0 while if it is repulsive then the phase will oscillate more slowly and we will
have δ0 < 0. See Fig. 2 for an illustration.

Figure 2: The phase shift δ0 is positive for attractive potentials and negative for repulsive potentials.

Scattering length. In the regime of low k it turns out that many potentials behave qualitatively
like what we have seen with the hard sphere, with a characteristic length scale called a “scattering
length.” To derive this suppose there is some b such that V (r) ≈ 0 for r ≥ b. In this region we
have u0 (r) ≈ sin(kb + δ0 ) (neglecting normalization). In the vicinity of b we have

u0 (r) ≈ u0 (b) + u00 (b)(r − b).

35
If we extrapolate to smaller values of r our approximation hits 0 at r = a where

u0 (b) tan(kb + δ0 )
a=b− 0 =b− .
u0 (b) k

Using the tan addition formula and taking the limit k → 0 we find

1 tan(kb) + tan(δ0 ) kb + tan(δ0 ) tan(δ0 )


a=b− ≈= b − =− . (107)
k 1 − tan(kb) tan(δ0 ) k k

Rearranging we have tan(δ0 ) = −ka, and in the ka  1 limit this yields



σ0 = sin2 (δ0 ) ≈ 4πa2 ,
k2
which is again the hard sphere result. Similar results hold for larger value of l. Thus the scattering
length can be thought of as an effective size of a scattering target.

36
MIT OpenCourseWare
http://ocw.mit.edu

8.06 Quantum Physics III


Spring 2016

For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy