8.06 Spring 2016 Lecture Notes 2. Time-Dependent Approximation Methods
8.06 Spring 2016 Lecture Notes 2. Time-Dependent Approximation Methods
Aram Harrow
Contents
1 Time-dependent perturbation theory 1
1.1 Rotating frame . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Perturbation expansion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3 NMR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.4 Periodic perturbations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
3 Adiabatic evolution 15
3.1 The adiabatic approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.2 Berry phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.3 Neutrino oscillations and the MSW effect . . . . . . . . . . . . . . . . . . . . . . . . 21
3.4 Born-Oppenheimer approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
4 Scattering 24
4.1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.2 Born Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
4.3 Partial Waves . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
However, the more important difference from time-independent perturbation theory is in our goals:
we will seek to analyze the dynamics of the wavefunction (i.e. find |ψ(t)i as a function of t) rather
than computing the spectrum of H. In fact, when we use a basis, we will work in the eigenbasis of
H0 . For example, one common situation that we will analyze is that we start in an eigenstate of
H0 , temporarily turn on a perturbation δH(t) and then measure in the eigenbasis of H0 . This is a
bit abstract, so here is a more concrete version of the example. H0 is the natural Hamiltonian of
1
the hydrogen atom and δH(t) comes from electric and/or magnetic fields that we temporarily turn
on. If we start in the 1s state, then what is the probability that after some time we will be in the
2p state? (Note that the very definition of the states depends on H0 and not the perturbation.)
Time-dependent perturbation theory will equip us to answer these questions.
Aside: comparison to Schrödinger and Heisenberg pictures. In 8.05 we saw the Schrödinger
picture and the Heisenberg picture. In the former, states evolve according to H and operators re-
main the same; in the latter, states stay the same and operators evolve according to H. The
interaction picture can be thought of as intermediate between these two. We pick a frame rotating
with H0 , which means that the operators evolve according to H0 and the states evolve with the
remaining piece of the Hamiltonian, namely δH. As we will see below, to calculate this evolution
correctly we need δH to rotate with H0 , just like all other operators. This is a little vague but
below we will perform an exact calculation to demonstrate what happens.
Now let’s compute the time evolution of |ψ̃(t)i.
d d iH0 t
i~ |ψ̃(t)i = i~ e ~ |ψ(t)i
dt dt
iH0 t iH0 t
= −H0 e ~ |ψ(t)i + e ~ (H0 + δH(t)) |ψ(t)i
iH0 t iH0 t
=e ~ δH(t)|ψ(t)i since H0 and e ~ commute
iH0 t iH t
− ~0 ˜
=e ~ δH (t)e |ψ(t)i
| {z }
δH(t)
f
d ˜ ˜
i~ |ψ(t)i = δf
H(t)|ψ(t)i (3)
dt
H0 |ni = En |ni.
2
Define δHmn (t) = hm|δH(t)|ni. Then
iH0 t iH0 t i(Em −En )t
f mn (t) = hm|e
δH ~ δH(t)e− ~ |ni = e ~ δHmn (t) ≡ eiωmn t δHmn ,
where we have defined ωmn = Em −E~
n
. If we define cn (t) according to
X X iEn t
|ψ̃(t)i = cn (t)|ni =⇒ |ψ(t)i = e− ~ cn (t)|ni
n n
then we obtain the following coupled differential equations for the {cn }.
X X
i~ċm (t) = δf
H mn (t)cn (t) = eiωmn t δHmn cn (t).
n n
i~∂ |ψ̃ (0) (t)i + i~∂t |ψ̃ (1) (t)i + i~∂t |ψ˜(2) (t)i + . . . = δfH(t)|ψ˜(0) (t)i + δf
H(t)|ψ˜(1) (t)i + . . . (4)
| t {z } | {z } | {z } | {z } | {z }
O(1) O(λ) O(λ2 ) O(λ) O(λ2 )
The solution is much simpler than in the time-dependent case. There is no zeroth order term on
the RHS, so the zeroth order approximation is simply that nothing happens:
3
1.3 NMR
In some cases the rotating frame already helps us solve nontrivial problems exactly without going
to perturbation theory. Suppose we have a single spin-1/2 particle in a magnetic field pointing in
the ẑ direction. This field corresponds to a Hamiltonian
~
H0 = ω0 Sz = ω0 σz .
2
If the particle is a proton (i.e. hydrogen nucleus) and the field is typical for NMR, then ω0 might
be around 500 MHz.
Static magnetic field Now let’s add a perturbation consisting of a magnetic field in the x̂
direction. First we will consider the static perturbation
δH(t) = ΩSx ,
where we will assume Ω ω0 , e.g. Ω might be on the order of 20 KHz. (Why are we considering a
time-independent Hamiltonian with time-dependent perturbation theory? Because really it is the
time-dependence of the state and not the Hamiltonian that we are after.)
We can solve this problem exactly without using any tools of perturbation theory, but it will be
instructive to compare the exact answer with the approximate one. The exact evolution is given
by precession about the
ω ẑ + Ωx̂
p0
ω02 + Ω2
p
axis at an angular frequency of ω02 + Ω2 . If Ω ω0 then this is very close to precession around
the ẑ axis.
Now let’s look at this problem using first-order perturbation theory.
ω t
0 ω t 0
δfH(t) = ei 2 σz ΩSx e−i 2 σz = Ω (cos(ω0 t)Sx − sin(ω0 t)Sy )
Z t f 0)
δH(t
(1)
|ψ̃ (t)i = dt0 |ψ(0)i
0 i~
Z t
Ω
dt0 cos(ω0 t0 )Sx − sin(ω0 t0 )Sy |ψ(0)i
=
0 i~
1 Ω
= (sin(ω0 t)Sx + (cos(ω0 t) − 1)Sy ) |ψ(0)i
i~ ω0
We see that the total change is proportional to Ω/ω0 , which is 1. Since this is the difference
between pure rotating around the ẑ axis, this is consistent with the exact answer we obtained.
The result of this calculation is that if we have a strong ẑ field, then adding a weak static x̂
field doesn’t do very much. If we want to have a significant effect on the state, we will need to do
something else. The rotating-frame picture suggests the answer: the perturbation should rotate
along with the frame, so that in the rotating frame it appears to be static.
4
We have already computed S˜x above. In the rotating frame we have
Thus
δf
H(t) = ΩSx .
The rotating-frame solution is now very simple:
iΩt
|ψ̃(t)i = e− 2
σx
|ψ(0)i.
This can be easily translated back into the stationary frame to obtain
iω0 t iΩt
|ψ(t)i = e− 2
σz
e− 2
σx
|ψ(0)i.
δH(t) = V cos(ωt),
for some time-independent operator V . If our system starts in state |ni then at time t we can
calculate Z t f mn (t0 )
δH
(1) ˜(1)
cm (t) = hm|ψ (t)i = dt0
0 i~
Z t 0
δHmn (t ) iωmn t0
= dt0 e
0 i~
Z t
Vmn 0
= dt0 cos(ωt)eiωmn t
0 i~
Vmn t 0 i(ωmn +ω)t0
Z
0
= dt e + ei(ωmn −ω)t
2i~ 0
" #
Vmn ei(ωmn +ω)t − 1 ei(ωmn −ω)t − 1
= +
2i~ ωmn + ω ωmn − ω
The ωmn ± ω terms in the denominator mean that we will get the largest contribution when
ω ≈ |ωmn |. (A word about signs. By convention we have ω > 0, but ωmn is a difference of energies
and so can have either sign.) For concreteness, let’s suppose that ω ≈ ωmn ; the ω ≈ −ωmn case is
similar. Then we have
Vmn ei(ωmn −ω)t − 1
c(1)
m (t) ≈ .
2i~ ωmn − ω
If we now measure, then the probability of obtaining outcome m is
2 (ωmn −ω)t
sin 2 sin2 αt
| V | 2 2 |V |
mn mn
Pn→m (t) ≈ |c(1) 2
m (t)| = = 2
,
~2 (ωmn − ω)2 ~2 α2
5
where we have defined the detuning α ≡ ωmn − ω. The t, α dependence is rather subtle, so we
examine it separately. Define
sin2 αt
2
f (t, α) = (9)
α2
For fixed α, f (t, α) is periodic in t.
It is more interesting to consider the case of fixed t, for which f has a sinc-like appearance (see
Fig. 1).
Figure 1: The function f (α, t) from (9), representing the amount of amplitude transfered at a fixed
time t as we vary the detuning α ≡ ωmn − ω.
It has zeroes as α = 2π
t n for integers n 6= 0. Since the closest zeros to the origin are at ±2π/t,
2π
we call the region α ∈ [−2π/t, 2π/t] the “peak” and the rest of the real line (i.e. R ∞ |α| ≥2 t ) the
2
“tails.” For α → ∞, f (t, α) ≤ 1/α . Thus, the tail has total area bounded by 2 2π/t 1/α = O(t).
For the peak, as α → 0, f (t, α) → t2 /4. On the other hand, sin is concave, so for 0 ≤ θ ≤ π/2
t2
we have sin(θ) ≥ sin(π/2) 2 π
π/2 θ = π θ. Thus for |α| ≤ t we have f (α, t) ≥ π . While these crude bounds
do not determine the precise multiplicative constants, this does show that there is a region of width
∼ 1/t and height ∼ t2R, and so the peak also has area O(t).
∞
We conclude that −∞ dαf (t, α) ∼ t. Dividing by t, we obtain
Z ∞
f (t, α)
dα ∼ 1.
−∞ t
f (t, α) π
lim = δ(α).
t→∞ t 2
6
The reason to divide by t is that this identifies the rate of transitions per unit time. Define
Rn→m = Pn→m
t . Then the above arguments imply that
π |Vmn |2
Rn→m ≈ δ(|ωmn | − ω) for large t. (10)
2 ~2
Linewidth In practice the frequency dependence is not a delta function. The term “linewidth”
refers to the width of the region of ω that drives a transition; more concretely, FWHM stands
for “full-width half-maximum” and denotes the width of the region that achieves ≥ 1/2 the peak
transition rate. The above derivation already suggests some reasons for nonzero linewidth.
1. Finite lifetime. If we apply the perturbation for a limited amount of time, or if the state we
are driving to/from has finite lifetime, then this will contribute linewidth on the order of 1/t.
2. Power broadening. If |Vmn | is large, then we will still see transitions for larger values of |α|.
For this to prevent us from seeing the precise location of a peak, we need also the phenomenon
of saturation in which transition rates all look the same above some threshold. (For example,
we might observe the fraction of a beam that is absorbed by some sample, and by definition
this cannot go above 1.)
There are many other sources of linewidth. In general we can think of both the driving frequency
ω and the gap frequency ωmn as being distributions rather than δ functions. The driving frequency
might come from a thermal distribution or a laser, both of which output a distribution of frequen-
cies. The linewidth of a laser is much lower but still nonzero. The energy difference ~ωmn seems
like a universal constant, can also be replaced by a distribution by phenomena such as Doppler
broadening, in which the thermal motion of an atom will redshift or blueshift the incident light.
This is just one example of a more general phenomenon in which interactions with other degrees of
freedom can add to the linewidth; e.g. consider the hyperfine splitting, which measures the small
shifts in an electron’s energy from its interaction with the nuclear spin. This can be thought of as
adding to linewidth in two different, roughly equivalent, ways. We might think of the nuclear spins
as random and thus the interaction adds a random term to the electon’s Hamiltonian. Alterna-
tively, we might view the interaction with the nuclear spin as a source of decoherence and thus as
contributing to the finite lifetime of the electron’s excited state. We will not explore those issues
further here.
The other contribution to the rate is the matrix element |Vmn |. This depends not only on the
strength of the perturbation, but also expresses the important point that we only see transitions
from n → m if Vmn 6= 0. This is called a selection rule. In Griffiths it is proved that transitions from
electric fields (see the next section) from Hydrogen state |n, l, mi to |n0 , l0 , m0 i are only possible
when |l − l0 | = 1 and |m − m0 | ≤ 1 (among other restrictions). Technically these constraints hold
only for first-order perturbation theory, but still selection rules are important, since they tell us
when we need to go to higher-order perturbation theory to see transitions (known as “forbidden
transitions”). In those cases transition rates are much lower. One dramatic example is that 2p → 1s
transition in hydrogen takes 1.6ns because it occurs at first order while the 2s → 1s transition takes
0.12 seconds. For this reason states such as the 2s states are called “metastable.”
We now consider the most important special case, which gets its own top-level section, despite
being an example of a periodic perturbation, which itself is an example of first-order perturbation
theory.
7
2 Light and atoms
~ and B
Light consists of oscillating E ~ fields. The effects of the B
~ fields are weaker by a factor
~
O(v/c) ∼ α, so we will focus on the E fields. Let
~ r) = E0 ẑ cos(ωt − kx).
E(~
However, optical wavelengths are 4000-8000Å, while the Bohr radius is ≈ 0.5Å, so to leading order
we can neglect the x dependence. Thus we approximate
We now can apply the results on transition rates from the last section with Vmn = eE0 hm|z|ni.
(This term is responsible for selection rules and for the role of polarization.) Thus the rate of
transitions is
π e2 E02
Rn→m = |hm|z |ni|2 δ(|ωmn | − ω). (12)
2 ~2
We get contributions at ωmn = ±ω corresponding to both absorption and stimulated emission.
Aside: quantizing light What about spontaneous emission? This does not appear in the
semiclassical treatment we’ve described here. Nor do the photons. “Absorption” means jumping
from a low-energy state to a higher-energy state, and “stimulated emission” means jumping from
high energy to low energy. In the former case, we reason from energy conservation that a photon
must have been absorbed, and in the latter, a photon must have been emitted. However, these
arguments are rather indirect. A much more direct explanation of what happens to the photon
comes from a more fully quantum treatment. This also yields the phenomenon of spontaneous
emission. Recall from 8.05 that oscillating electromagnetic fields can be quantized as follows:
r r
† 2π~ω ~ω
E0 = E0 (â + â ) E0 = (Gaussian units) = (SI units)
V 0 V
Using δH = eE0 z, we obtain
δH = eE0 z ⊗ (â + ↠).
0 α
If we look at the action of z in the {1s, 2pz } basis, then it has the form with α = h1s|z |2pz i.
α 0
We then obtain the form of the Hamiltonian examined on pset 3.
This perspective also can be used to give a partial derivation of the Lamb shift, which can be
thought of as the interaction of the electron with fluctuations in the electric field of the vacuum.
In the vacuum (i.e. ground state of the photon field) we have hE0 i ∼ hâ + ↠i = 0 but hE02 i ∼
h(â + ↠)2 i > 0. These vacuum fluctuations lead to a small separation in energy between the 2s
and 2p levels of hydrogen.
Dipole moment In the Stark effect we looked at the interaction of the hydrogen atom with an
electric field. This was a special case of the interaction between a E ~ field and the dipole moment
of a collection of particles. Here we discuss the more general case.
Suppose that we have charges q1 , . . . , qN at positions ~x(1) , . . . , ~x(N ) , and we apply an electric
~ x). The energy is determined by the scalar potential φ(~x) which is related to the electric
field E(~
8
~ = −∇
field by E ~ φ. If E(~
~ x) = E ~ (i.e. independent of position ~x) then one possible solution is
~
φ(~x) = −~x · E. In this case the Hamiltonian will be
N
X N
X
qi φ ~x(i) = − ~ = −d~ · E
qi ~x(i) · E ~
H=
i=1 i=1
PN
where we have defined the dipole moment d~ = i=1 qi ~x(i) . Our choice of φ was not unique, and
we could have chosen φ(~x) = C − ~x · E ~ for any constant C. However, this would only have added
an overall constant to the Hamiltonian, which would have no physical effect.
What if the electric field is spatially varying? If this spatial variation is small and we are near
the origin, we use the first few terms of the Taylor expansion to approximate the field:
3
~ x) = E(0)
~
X ∂Ei
E(~ + êi xj + . . . .
∂xj
i,j =1
Since the transition rate depends on |Vmn |2 , we will average this quantity over the choice of polar-
9
ization. Denote the average over all unit vectors Pˆ by h·iPˆ .
D E
|Vmn |2 ˆ = E02 |Pˆ · d~mn |2
P P̂
3
X
= E02
hm|Pi di |ni hn|Pj dj |mi P̂
i,j =1
3
X
= E02
Pi Pj Pˆ
hm|di |ni hn|dj |mi
i,j=1
3
X δij
= E02 hm|di |ni hn|dj |mi explained below
3
i,j =1
E02 X
= |hm|di |ni|2
3
i
E02
≡ |d~mn |2
3
How did we calculate hPi Pj iPˆ ? This can be done by explicit calculation, but it is easier to use sym-
metry. First, observe that the uniform distribution over unit vectors is invariant under reflection.
Thus, if i =6 j, then hPi Pj i = h(−Pi )Pj i = 0. On the other hand rotation symmetry means that
hPi i should be independent of i. Since P12 + P23 + P32 = 1, we also have hP12 + P23 + P32 i = 1 and
2
δij
hPi Pj iPˆ = . (13)
3
Next, we would like to average over different frequencies. The energy density of an electric field
E02
is U =
R 8π (using Gaussian units). Define U (ω) to be the energy density at frequency ω, so that
U = U (ω) dω. If we consider light with this power spectrum, then we should integrate the rate
times this distribution over U (ω) to obtain
4π 2
Z
Rn→m = dω U (ω) 2 |d~mn |2 δ(ω − |ωmn |)
3~
4π 2 ~ 2
= |dmn | U (|ωmn |)
3~2
This last expression is known as Fermi’s Golden Rule. (It was discovered by Dirac, but Fermi called
it “Goldren Rule #2”.)
(a) derived an exact relation between rates of spontaneous emission, stimulated emission and
absorption; and
(b) proposed the phenomenon of stimulated emission, which was not observed until 1960.
10
He did this in 1917, more than a decade before even the Schrödinger equation!
Here we will reproduce that argument. It assumes a collection of atoms that can be in either
state a or state b. Suppose that there are Na atoms in state a and Nb atoms in state b, and that the
states have energies Ea , Eb with Eb > Ea . Define ωba = Eb −E~
a
and β = 1/kB T . Assume further
that the atoms are in contact with a bath of photons and that the entire system is in thermal
equilibrium with temperature T . From this we can deduce three facts:
Fact 1. Equilibrium means no change: N˙ a = N˙ b = 0.
Nb e−βEb
Fact 2. At thermal equilibrium we have Na = e−βEa
= e−β ~ωba .
These processes depend on the Einstein coefficients A, Bab and Bba for spontaneous emission,
absorption and stimulated emission respectively. They also depend on the populations of atoms
and/or photons that they involve; e.g. absorption requires an atom in state a and a photon of
frequency ωba , so its rate is proportional to Na U (ωba ). Here it is safe to posit the existence of
stimulated emission because we have not assumed that Bba is nonzero.
Having set up the problem, we are now almost done! Adding these processes up, we get
A Fact 2 A Fact 3 ~ω 3 1
U (ωba ) = Na
= β~ω
= 2 3 β~ω
Nb Bab − Bba
e ba Bab − Bba π c e ba −1
Since this relation should hold for all values of β, we can equate coefficients and find
11
are inevitably accompanied by photon-creating processes containing ↠. Additionally the relation
between spontaneous and stimulated emission can be seen in the fact that both involve an a†
operator acting on the photon field. If there are no photons, then the field is in state |0i and we
get the term a† |0i = |1i, corresponding to√spontaneous emission. If there are already n photon in
the mode, then we get the term a† |ni = n + 1|n + 1i. Since the probabilities are the square of
the amplitudes, this means that we see photons emitted at n + 1 times the spontaneous emission
rate. In Einstein’s terminology, the n here is from stimulated emission and the +1 is from the
spontaneous emission which always occurs independent of the number of photons present.
Returning to (16), we plug in Fermi’s Golden Rule and obtain the rates
4π 2 ~ 2
Bab = Bba = |dab | and
3~2
3
4ωba
A= |d~ab |2 .
3~c3
We have defined A to equal the difficult-to-evaluate integral. The factor of x3 can be removed by
12
writing A = i ∂k∂ 3 B, where
Z
3 r ~
B = d x exp − − ik · ~x
a0
Z ∞ Z 1
r
~k · ~x
= 2π r 2
dµ exp − − ikrµ defining µ = ~ r = |~x|
, k = |k|,
0 −1 a0 kr
Z 1 Z ∞
1 n!
= 4π dµ 3 using rn e−αr =
−1 1 0 αn+1
a0 + ikµ
Z 1
4πi 1
= 3 dµ 3
k −1 µ + kai 0
2 2
1 1
2πi 1 1 2πi 1 + ika0 − 1− ika0
= − 3 2 − 2 = 3
2
k k
1 + kai 0 1 − kai 0 1 + k21a2
0
1
8πi ika0 8π 8π
= 2 = 2 = 2
k3 + a− 2
1+ 1
k 4 a0 1 + 1 a0 k2 0
k2 a20 k2 a20
∂ 2
To compute A, we use the fact that ∂k3 k = 2k3 . Thus
∂ −32πik3
A=i B= 3 .
∂k3 a0 k 2 + a−2
0
We can simplify this expression using our assumption that the photon energy (and therefore also
2 k2
the final state energy) is much larger than the binding energy. The final energy is ~2m and the
~2
binding energy is 2ma2 . Thus
0
~2 k 2 ~2
=⇒ ka0 1.
2m 2ma20
13
We can simplify this a bit by averaging over all the polarizations of the light. (In fact, the angular
dependance of the free electron can often carry useful information, but here it will help simplify
some calculations.) The average of cos2 (θ) over the sphere is 1/3 (by the same arguments we used
in the derivation of Fermi’s golden rule), so we obtain
~k2
2
1024πe E 2 f t, 2m − ω
hR1,0,0→k~ i = 2 5 10 03 .
3~ a0 k L t
Let’s pause for a minute to look at what we’ve derived. One strange feature is the 1/L3 term,
because the rate of ionization should not depend on how much empty space surrounds the atom.
Another strange thing appears to happen when we take t large, so that f (t, α)/t will approach
π 2
2 δ(α). This would cause the transition rates to be nonzero only when 2mω/~ exactly equals k for
some valid vector ~k (i.e. of the form 2π L~n). We do not generally expect physical systems to have
such sensitive dependence on their parameters.
As often happens when two things look wrong, these difficulties can be made to “cancel each
other out.” Let us take t to be large but finite. It will turn out that t needs to be large only relative
2
to L~2 m , which is not very demanding when L is large. In this case, we can approximate f (t, α)
with a step function: (
π 2
t if 0 ≤ α ≤ 1t
f (t, α) ≈ f˜(t, α) ≡ 2
0 otherwise
In what sense is this a good approximation? We argue that for large t, f˜(t, α)/t ≈ π2 δ(α), just like
f (t, α)/t. Suppose that g(α) is a function satisfying |g 0 (α)| ≤ C for all α. Then
Z ! Z !
∞ f˜(t, α) 1/t
dα − δ(α) g(α) = t dα g(α) − g(0)
π
2t
−∞ 0
Z
1/t
= t dα (g(α) − g(0))
0
Z
1/t Z α
= t dα dβ g 0 (β)
0 0
Z 1/t Z α
dβ g 0 (β)
≤t dα triangle inequality
0 0
1 C
≤ t 2C =
2t 2t
This tends to 0 as t → ∞. (This is an example of a more general principle that the “shape” of a δ
function doesn’t matter. For example, the limit of a Gaussian distribution with σ 2 → 0 would also
work.)
Now using f˜(t, α), we get a nonzero contribution for ~k satisfying
~ 1
0≤2
2m − ω ≤
k t
2mω 2 2mω 1
⇔ ≤k ≤ 1+
~ ~ tω
r s r
2mω 2mω 1 2mω 1
⇔ ≤k≤ 1+ ≈ + (17)
~ ~ tω ~ 2tω
14
How many ~k satisfy (17)? Valid ~k live on a cubic lattice with spacing 2π/L, and thus have density
(L/2π)3 . Thus we can estimate the number of ~k satisfying (17) by (L/2π)3 times theqvolume of
2mω
k-space satisfying (17). This in turn corresponds to a spherical shell of inner radius ~ and
q
thickness 2mω 1
~ 2tω . Thus we have
3 3/2 r
L3 mL3 mk
L 2mω 1
2mω
# valid ~k = 4π = 2
= .
2π ~ 2tω~ 2π 2 ~t
2π ~t
q
In the last step we use the fact that spherical shell is thin to approximate k ≈ 2mω ~ . Thus, when
˜ ~
we sum f (t, α)/t over k we obtain
X f˜(t, α) π L3 mk
= t · # valid ~k = .
t 2 4π ~
~k
We have obtained our factor of L3 that removes the unphysical dependence on the boundary
conditions. Putting everything together we get
256me2 E02
R1,0,0→all ~k = .
3~3 a50 k 9
3 Adiabatic evolution
3.1 The adiabatic approximation
We now turn to a different kind of approximation, in which we consider slowly varying Hamiltonians.
We will consider a time-dependent Hamiltonian H(t). Let |ψn (t)i and En (t) be the “instantaneous”
eigenbases and eigenenergies, defined by
15
Example. Suppose we have a spin-1/2 particle in a magnetic field B(t). ~ Then the Hamiltonian
S~ ~ ~ both
is H(t) = ge µB ~ · B(t). The adiabatic theorem says that if we start with the spin and B
~ to point in the x̂ direction, then the spin will
pointing in the +ẑ direction and gradually rotate B
follow the magnetic field and also point in the x̂ direction. Given that the Schrödinger equation
prescribes instead that the spin precess around the magnetic field, this behavior appears at first
somewhat strange.
Derivation We will not rigorously prove the adiabatic theorem, but will describe most of the
derivation. Begin by writing X
|Ψ(t)i = cn (t)|ψn (t)i.
n
Now we need a way to evaluate hψk |ψ˙ n i in terms of more familiar quantities.
This equation has two interesting cases: k = n and k = 6 n. The former will not be helpful in
estimating hψk |ψ̇n i, but does give us a useful result, called the Hellmann-Feynman theorem.
In the last step, we used H˙ kn to refer to the matrix elements of H˙ in the {|ψn i} basis.
Plugging this into (20) we find
X Ḣkn
i~ċk = (Ek − i~hψk |ψ̇k i)ck − i~ cn . (21)
| {z } En − Ek
6
n=k
adiabatic approximation | {z }
error term
If the part of the equation denoted “error term” did not exist, then |ck | would be independent of
time, which would confirm the adiabatic theorem. Furthermore, the error term is suppressed by a
factor of 1/∆nk , where ∆nk ≡ En − Ek is the energy gap. So naively it seems that if H˙ is small
relative to ∆nk then the error term should be small. On the other hand, these two quantities do
not even have the same units, so we will have to be careful.
16
Phases Before we analyze the error term, let’s look at the phases we get if the error term were
not there. i.e. suppose that i~ċk = (Ek − i~hψk |ψ˙ k i)ck . The solution of this differential equation is
The θk (t) term is called the “dynamical” phase and corresponds to exactly what you’d expect from
a Hamiltonian that’s always on; namely the phase of state k rotates at rate −Ek /~. The γk (t) is
called the “geometric phase” or “Berry phase” and will be discussed further in the next lecture.
At this point, observe only that it is independent of ~ and that νk (t) can be seen to be real by
applying d/dt to the equation hψk |ψk i = 1.
Validity of the adiabatic approximation Let’s estimate the magnitude of the error term in a
toy model. Suppose that H(t) = H0 + Tt V , where H0 , V are time-independent and T is a constant
that sets the timescale on which V is turned on. Then H˙ = V /T . An important prediction about
the adiabatic theorem is that if the more slowly H changes from H0 to H0 + V , the lower the
probability of transition should be; i.e. increasing T should reduce the error term, even if we
integrate over time from 0 to T .
Let’s see how this works. If the gap is always & ∆, then we can upper-bound the transition
rate by some matrix element of TV∆ . This decreases as T and ∆ increase, which is good. But if we
add up this rate of transitions over time T , then the total transition amplitude can be as large as
∼ V /∆. Thus, going more slowly appears not to reduce the total probability of transition!
What went wrong? Well, we assumed that amplitude from state n simply added up in state
k. But if the states have different energies, then over time the terms we add will have different
phases, and may cancel out. This can be understood in terms of time-dependent perturbation
theory. Define c̃k (t) = e−iθk (t) ck (t). Observe that
d
i~ c̃k (t) = ~θ˙k (t)e−iθk (t) ck (t) + i~e−iθk (t) ċk (t)
dt
X Ḣkn
= −Ek (t)e−iθk (t) ck (t) + e−iθk (t) (Ek (t) − ~νk (t))ck (t) − i~ e−iθk (t) cn (t)
En − Ek
6
n=k
X Ḣkn
= −~νk (t)c̃k (t) − i~ ei(θn (t)−θk (t)) c̃n (t)
En − Ek
6
n=k
In the last step we have used cn (t) = eiθn (t) c̃n (t). Let’s ignore the νk (t) geometric phase term (since
our analysis here is somewhat heuristic). We see that the error term is the same as in (21) but
with an extra phase of ei(θn (t)−θk (t)) . Analyzing this in general is tricky, but let’s suppose that the
energy levels are roughly constant, so we can replace it with e−iωnk t , where ωnk = (En − Ek )/~.
Now when we integrate the contribution of this term from t = 0 to t = T we get
Finally we obtain that the probability of transition decreases with T . This can be thought of as a
rough justification of the adiabatic theorem, but it of course made many simplifying assumptions
and in general it will be only qualitatively correct.
17
This was focused on a specific transition. In general adiabatic transitions between levels m and
n are suppressed if
~|Ḣmn | ∆2 = min(Em (t) − En (t))2 . (23)
t
Landau-Zener transitions One example that can be solved exactly is a two-level system with
a linearly changing Hamiltonian. Suppose a spin-1/2 particle experiences a magnetic field resulting
in the Hamiltonian
vt
H(t) = ∆σx + σz ,
T
p
for some constants ∆, v, T . The eigenvalues are ± ∆2 + (vt/T )2 . Assuming v > 0, then when
t = −∞ the top eigenstate is |−i and the bottom eigenstate is |+i. When t = ∞ these are reversed;
|+i is the top eigenstate and |−i is the bottom eigenstate. When t = 0, the eigenstates are |+i±|−i
√
2
.
See diagram on black-board for energy levels.
Suppose that ∆ = 0 and we start in the |−i at t = −∞. Then at t = ∞ we will still be in
the |−i state, with only the phase having changed. But if ∆ > 0 and we move slowly enough then
the adiabatic approximation says we will remain in the top eigenstate, which for t = ∞ will be
|+i. Thus, the presence of a very small transverse field can completely change the state if we move
slowly enough through it.
In this case, the error term in the adiabatic approximation can be calculated rather precisely
and is given by the Landau-Zener formula (proof omitted):
2π 2 ∆2 T
Pr[transition] ≈ exp − .
~v
Observe that it has all the qualitative features that we expect in terms of dependence on ∆, v, T ,
but that it corresponds to a rate of transitions exponentially smaller than our above estimate from
first-order perturbation theory. Note that here “transition” refers to transitions between energy
level. Thus starting in |−i and ending in |+i corresponds to “no transition” while ending in |−i
would correspond to “transition,” since it means starting in the higher energy level and ending in
the lower energy level.
The phase γn (t) is called the geometric phase, or the Berry phase, after Michael Berry’s 1984
explanation of it.
Do the phases in the adiabatic approximation matter? This is a somewhat subtle question. Of
course an overall phase cannot be observed, but a relative phase can lead to observable interference
effects. The phases in (22) depend on the eigenstate label n, and so in principle interference is
possible. But solutions to the equation H(t)|ψn (t)i = En (t)|ψn (t)i are not uniquely defined, and
we can in general redefine |ψn (t)i by multiplying by a phase that can depend on both n and t.
18
To see how this works, let us consider the example of a spin-1/2 particle in a spatially varying
magnetic field. If the particle moves slowly, we can think of the position ~r(t) as a classical variable
causing the spin to experience the Hamiltonian H(~r(t)). This suggests that we might write the
state as a function of ~r(t), as |ψn (~r(t))i or even |ψn (~r)i. If the particle’s position is a classical
function of time, then we need only consider interference between states with the same value of ~r,
and so we can safely change |ψn (~r)i by any phase that is a function of n and ~r.
In fact, even if the particle were in a superposition of positions as in the two-slit experiment,
then we could still only see interference effects between branches of the wavefunction with the same
value of ~r. Thus, again we can define an arbitrary (n, ~r)-dependent phase.
More generally, suppose that H depends on some set of coordinates R(t) ~ = (R1 (t), . . . , RN (t)).
~ ~ ~ ~
The eigenvalue equation (18) becomes H(R)|ψn (R)i = En (R)|ψn (R)i where we leave the time-
dependence of R ~ implicit. This allows us to compute even in situations where R ~ is in a superposition
of coordinates at a given time t.
~
To express γn (t) in terms of |ψn (R)i, we compute
N ~ i dRi ~
d ~ =
X d|ψn (R) ~ i · dR
~ ~ |ψn (R)
|ψn (R)i =∇ R
dt dRi dt dt
i=1
Z t ~ ~
Z R(t)
γn (t) = i hψn |∇ ~ ~ |ψn i · dR dt = ~ ~ |ψn i · dR
ihψn |∇ ~
R dt ~ (0) R
0 R
The answer is in terms of a line integral, which depends only on the path and not on time (unlike
the dynamical phase).
How does this change if we reparameterize |ψn (R ~ )i? Suppose we replace |ψn (R) ~ i with |ψ˜n (R)
~ i=
~ ~
e−iβ(R) |ψn (R)i. Then the Berry phase becomes
~
Z R(t) ~
Z R(t)
~ ~ ~
γ̃n (t) = i hψ˜n (R) ~ ~ |ψ˜n (R)
~ |∇
R
~ i · dR
~ =i ~ |eiβ(R)
hψn (R) ∇ ~ e−iβ (R) |ψn (R)
R
~ i · dR
~
~ (0)
R ~
R(0)
~
= γn (t) + β(R(t)) ~
− β(R(0))
Changing β only changes phases as a function of the endpoints of the path. Thus, we can eliminate
~
the phase for any fixed path with R(t) ~
6= R(0), but not simultaneously for all paths. In particular,
if a particle takes two different paths to the same point, the difference in their phases cannot be
redefined away. More simply, suppose the path is a loop, so that R(0) ~ ~
= R(t). Then regardless of
β we will have γn = γ̃n . This suggests an important point about the Berry phase, which is that it
is uniquely defined on closed paths, but not necessarily open ones.
~
Suppose that R(t) follows a closed curve C. Then we can write
I I
~ ~
γn [C] = ihψn |∇R~ |ψn i ·dR = A ~ n (R)
~ · dR,
~
| {z }
~n (R)
A ~
19
Here we define the Berry curvature D ~n = ∇ ~ ~ ×A
~ n and the infinitesimal unit of area d~a. We can
R
~
write Dn in a more symmetric way as follows:
X d d X dhψn | d|ψn i d d
(Dn )i = i ijk hψn | |ψn i = i ijk + hψn | |ψn i .
dRj dRk dRj dRk dRj dRk
j,k j,k
d d
Because ijk is antisymmetric in j, k and dR j dRk
is symmetric, the second term vanishes and we
are left with
D~ n = i(∇
~ r hψn |) × (∇
~ r |ψn i). (25)
~ e~
H = µ~σ · B µ= .
mc
Suppose that B~ = B~r where B is fixed and we slowly trace out a closed path in the unit sphere
with ~r. Suppose that we start in the state
sin(θ) cos(φ)
cos(θ/2)
|~r; +i = |~ri = with ~r = sin(θ) sin(φ)
iφ
e sin(θ/2)
cos(θ)
Then the adiabatic theorem states that we will remain in the state |~ri at later points, up to an
overall phase. To compute the geometric phase observe that
~ = d r̂ + 1 d θˆ + 1
∇
d ˆ
φ.
dr r dθ r sin θ dφ
d
Since dr |~
ri = 0 we have
− sin(θ/2) 0
~ ~ri = 1
∇| θˆ + 1 ˆ
φ.
2r eiφ cos(θ/2) r sin θ ieiφ sin(θ/2)
| {z }
|−~
ri
This first term will not contribute to the Berry connection, and so we obtain
2
~ ri = − 1 sin (θ/2) φ.
~ + (~r) = ih~r|∇|~
A ˆ
r sin(θ)
~+ = ∇
~ ×A
~+ = 1 d r̂ d 1
D (sin θA+,φ ) r̂ = − 2 sin2 (θ/2) = − 2 r̂.
r sin θ dθ r sin θ dθ 2r
d d 1−cos θ
For this last computation, observe that dθ sin2 (θ/2) = dθ 2 = sin θ. We can now compute the
Berry phase as ZZ
γ+ [C] = ~ + · d~a = − 1 Ω.
D
S
|{z} 2
r2 dΩr̂
20
Here dΩ is a unit of solid angle, and Ω is the solid angle contained by C.
What if we used a different parameterization for |~ri? An equally valid choice is
e−iφ cos(θ/2)
|~ri = . (26)
sin(θ/2)
γm [C] = −mΩ.
There is much more that can be said about Berry’s phase. An excellent treatment is found in the
1989 book Geometric phases in physics by Wilczek and Shapere. There is a classical analogue called
Hannay’s phase. Berry’s phase also has applications to molecular dynamics and to understanding
electrical and magnetic properties of Bloch states. We will see Berry’s phase again when we discuss
the Aharonov-Bohm effect in a few weeks.
21
Solar neutrinos Solar neutrinos are produced via the p-p chain reaction, which converts (via a
series of reactions)
4 1H = 4p+ + 4e− 7→ 2p+ + 2n + 2e− +2νe .
| {z }
4 He
The resulting neutrinos are produced with energies in the range 0.5-20MeV. Almost all of neutrinos
produced in the sun are electron neutrinos.
Detection Neutrinos can be detected via inverse beta decay, corresponding to the reaction
A + νe 7→ A0 + e− ,
where A, A0 are different atomic nuclei. For solar neutrinos this will only happen for electron
neutrinos because the reaction A + νµ 7→ A0 + µ− will only happen for mu neutrinos carrying
at least 108 MeV of kinetic energy. So it is easiest to observe electron neutrinos. However other
flavors of neutrinos can also be detected via more complicated processes, such as neutrino-mediated
disassociation of deuterium.
Observations of solar neutrinos The first experiment to detect cosmic neutrinos was the 1968
Homestake experiment, led by Ray Davis, which used 100,000 gallons of dry-cleaning fluid (C2 Cl4 )
to detect neutrinos via the process 37 Cl + νe 7→37 Ar + e− . However, this only found about 1/3 as
many neutrinos as standard solar models predicted.
In 2002, the Sudbery Neutrino Observatory (SNO) measured the total neutrino flux and found
that once mu- and tau-neutrinos were accounted for, the total number of neutrinos was correct.
Thus, somehow electron neutrinos in the sun had become mu and tau neutrinos by the time they
reached the Earth.
The MSW effect It turns out that this puzzle can be resolved by a clever use of the adiabatic
theorem. Electron neutrinos scatter off of electrons and thus the Hamiltonian in (27) should be
22
modified to add a term proportional to the local density of electrons. Thus after some additional
rearranging, we obtain
−∆0 cos(2θ) ∆0 sin(2θ) CNe 0
H = E0 + + , (28)
∆0 sin(2θ) ∆0 cos(2θ) 0 0
where ∆0 , θ come from (27) (θ ≈ π/6 is the “mixing angle” that measures how far the flavor states
are from being eigenstates), C is a constant and Ne = Ne (~r) is the local electron density. If the
neutrino is traveling at speed ≈ c in direction x̂, then ~r ≈ ctx̂. Thus we can think of Ne as
time-dependent. We then can rewrite H as
CNe (t)
const · I + − ∆0 cos(2θ) σz + ∆0 sin(2θ)σx . (29)
2
This looks like the adiabatic Landau-Zener transition we studied in the last lecture, although here
the σz term is no longer being swept from −∞ to +∞. Instead, near the center of the sun, Ne (0)
is large and the eigenstates are roughly |νe i, |νµ i. For large t, the neutrinos are in vacuum, where
their eigenstates are |ν1 i, |ν2 i.
If the conditions of the adiabatic theorem are met, then neutrinos that start in state |νe i (in
the center of the sun) will emerge in state |ν2 i (at the surface of the sun). They will then remain
in this state as they propagate to the Earth. It turns out that this holds for neutrinos of energies
' 2M eV . In this case, the probability of observing the neutrino on Earth in the |νe i state (thinking
of neutrino detectors as making measurements in the flavor basis) is sin2 (θ), which gives more or
less the observed value of 0.31.
Here R = (R ~ 1, . . . , R
~ N ) denotes the positions of the N nuclei and Hel (R) includes all the other
terms, i.e. kinetic energy of the electrons as well as the potential energy terms which include
electron-electron, nuclei-nuclei and electron-nuclei interactions. Let r denote all of the coordinates
of the electrons. While (30) may be too hard to solve exactly, we can use a version of the adiabatic
theorem to derive an approximate solution.
We will consider a product ansatz:
(Typically this eigenstate wil be simply the ground state.) This is plausible because of the adiabatic
theorem. If the nuclei move slowly then as this happens the electrons can rapidly adjust to remain
in their ground states. Then once we have solved (32) we might imagine that we can substitute
23
back to solve for the nuclear eigenstates. We might guess that they are solutions to the following
eigenvalue equation
N 2
−
X ~ ~ 2 + Eel (R) γ(R) = Eγ(R).
∇ (33)
2Mj R~ j
j=1
~ ~ to (31) we obtain
However, this is not quite right. If we apply ∇ Rj
~ ~ Ψ(R, r) = (∇
∇ ~ ~ γ(R))ΦR (r) + γ(R)∇
~ ~ ΦR (r). (34)
Rj Rj Rj
Using the adiabatic approximation we neglect the overlap of ∇ ~ ~ Ψ(R, r) with all states to |ΦR i.
Rj
Equivalently we can multiply on the left by hΦR |. This results in
Z Z
d rΦR (r) ∇R~ j Ψ(R, r) = ∇R~ j γ(R) + γ(R) d3n rΦR (r)∗ ∇
3n ∗~ ~ ~ ~ ΦR (r)
Rj
~ ~ − iA
= (∇ ~ j )γ(R),
Rj
We conclude that the effective Hamiltonian actual experienced by the nuclei should be
N
X ~2 ~ ~ j )2 + Eel (R).
Heff = (∇ ~ − iA (36)
2Mj Rj
j=1
We will see these A ~ j terms again when we discuss electomagnetism later in the semester. In
systems of nuclei and atoms we need at least three nuclei before the A ~ j terms can have an effect,
for the same reason that we do not see a Berry phase unless we trace out a loop in a parameter
space of dimension ≥ 2.
The Born-Oppenheimer applies not just to nuclei and electrons but whenever we can divide a
system into fast and slow-moving degrees of freedom; e.g. we can treat a proton as a single particle
and ignore (or “integrate out”) the motion of the quarks within the proton. This is an important
principle that we often take for granted. Some more general versions of Born-Oppenheimer are
called “effective field theory” or the renormalization group.
4 Scattering
4.1 Preliminaries
One of the most important types of experiments in quantum mechanics is scattering. A beam of
particles is sent into a potential and scatters off it in various directions. The angular distribution
of scattered particles is then measured. In 8.04 we studied scattering in 1-d, and here we will
study scattering in 3-d. This is an enormous field, and we will barely scratch the surface of it. In
particular, we will focus on the following special case:
• Elastic scattering. The outgoing particle has the same energy as the incoming particle. This
means we can model the particles being scattered off semi-classically, as a static potential
V (~r). The other types of scattering are inelastic scattering, which can involve transformation
of the particles involved or creation of new particles, and absorption, in which there is no
outgoing particle.
24
• Non-relativistic scattering. This is by contrast with modern accelerators such as the LHC.
However, non-relativistic scattering is still relevant to many cutting-edge experiments, such
as modern search for cosmic dark matter (which is believed to be traveling at non-relativistic
speeds).
Even this special case can teach us a lot of interesting physics. For example, Rutherford scattering1
showed that atoms have nuclei, thereby refuting the earlier “plum pudding” model of atoms. This
led to a model of atoms in which electrons orbit nuclei like planets, and resolving the problems of
this model in turn was one of the early successes of quantum mechanics.
Scattering cross section: In scattering problems it is important to think about which physical
quantities can be observed. The incoming particles have a flux that is measured in terms of number
2N
of particles per unit area per unit time, i.e. ddAdt
in
. If we just count the total number of scattered
particles, then this is measured in terms of particles per time: dNdtscat . The ratio of these quantities
has units of area and is called the scattering cross section:
dNscat
dt
d2 Nin
= σ. (37)
dAdt
To get a sense of why these are the right units, consider scattering of classical particles off of a
classical hard sphere of radius a. If a particle hits the sphere it will scatter, and if it does not hit
the sphere it will not scatter. Assume that the beam of particles
p is much wider than the target,
i.e. each particle has trajectory ~r = (x0 , y0 , z0 + vt) with x20 + y02 given by a distribution
p with
standard deviation that is a. The particles that scatter will be the ones with x0 + y0 ≤ a 2 2
which corresponds to a region with area πa2 , which is precisely the cross-sectional area of the
2
sphere. Since we have dNdtscat = ddANdt
in
πa2 , it follows that σ = πa2 . This simple example is good to
keep in mind to have intuition about the meaning of scattering cross sections.
is defined to be
d2 Nscat
dσ
(θ, φ) ≡ ddΩ 2N
dt
(38)
dΩ in
dAdt
Here the spherical coordinates (θ, φ) denote the direction of the outgoing particles. It is conventional
to define the axes so that the incoming particles have momentum in the ẑ direction, so θ is the angle
between the scattered particle and the incoming beam (i.e. θ = 0 means no change in direction
and θ = π means backwards scattering) while φ is the azimuthal angle. Integrating over all angles
gives us the full cross-section, i.e. Z
dσ
σ = dΩ . (39)
dΩ
Quantum mechanical scattering: Assume that the incoming particle states are wavepackets
that are large relative to the target. This allows us to approximate the incoming particles as plane
wave, i.e.
iEt
ψin ∝ eikz− ~ , (40)
1
Rutherford scattering is named after Ernest Rutherford for his 1911 explanation of the 1909 experiment which
was carried out by Geiger and Marsden.
25
2 2
k
where E = ~2m . Here we need to assume that the potential V (~r) → 0 as r → ∞ so that plane
waves are solutions to the Schrödinger equation for large r. For the scattered wave, we should seek
solutions satisfying
~2 ~ 2
− ∇ ψscat = Eψscat as r → ∞ (41)
2m
1 d2
1 ˆ2
− 2
r + 2L ψscat = k 2 ψscat in spherical coordinates (42)
r dr r
A general solution can be written as a superposition of separable solutions. Separable solutions to
(42) can in turn be written as
rψ(r, θ, φ) = u(r)f (θ, φ), (43)
in terms of some functions u(r), f (θ, φ). In terms of these (42) becomes
1 ˆ2
u00 f + 2
uL f +k 2 uf = 0. (44)
r
| {z }
→0 as r→∞
Thus, for large r, we can cancel the f from each side and simply have u00 = −k 2 u, which has
solutions e±ikr . The eikr solution corresponds to outgoing waves and the e−ikr solution to incoming
waves. A scattered wave should be entirely outgoing, and so we obtain
r→∞ f (θ, φ) ikr− iEt
ψscat = e ~ (45)
r
or more precisely
f (θ, φ) ikr− iEt 1
ψscat = e ~ + O . (46)
r r2
Because the scattering is elastic, the k and E here are the same as for the incoming wave.
Time-independent formulation: As with 1-d scattering problems, the true scattering process
is of course time-dependent, but the quantities of interest (transmission/reflection in 1-d, differential
cross section in 3-d) can be extracted by solving the time-independent Schrödinger equation with
suitable boundary conditions. In the true process, the incoming wave should really be a wavepacket
with well-defined momentum ≈ (0, 0, k) and therefore delocalized position. The outgoing wave will
be a combination of an un-scattered part, which looks like the original wave packet continuing
forward in the ẑ direction, and a scattered part, which is a spherical outgoing wavepacket with a
f (θ, φ) angular dependence. However, we can treat the incoming wave instead as the static plane
wave eikz and the scattered wave instead as the static outgoing wave f (θ,φ) r e
ikr . (Both of these
are when r → ∞.) Thus we can formulate the entire scattering problem as a time-independent
boundary-value problem. The high-level strategy is then to solve the Schrödinger equation subject
to the boundary conditions
r→∞ f (θ, φ) ikr
ψ(~r) = eikz + e . (47)
r
This is analogous to what we did in 1-D scattering, where the boundary conditions were that ψ(x)
should approach eikx + Re−ikx for x → −∞ and should approach T eikx for x → ∞. As in the 1-D
case, we have to remember that this equation is an approximation for a time-dependent problem.
As a result when calculating observable quantities we have to remember not to include interference
terms between the incoming and reflected waves, since these never both exist at the same point in
time.
26
Relation to observables. In the 1-D case, the probabilities of transmission and reflection are
|T |2 and |R|2 respectively. In the 3-d case, the observable quantities are the differential cross
dσ
sections dΩ . To compute these, first evaluate the incoming flux
~ ikz = ~k ẑ = vẑ
~in = ~ Im ψin
S ∗ ~
∇ψin =
~
Im e−ikz ∇e (48)
m m m
In the last step we have used v = ~k/m. The units
√here are off because technically the wavefunction
ikz ikz
should be not e but something more like e / V , where V has units of volume. But neglecting
this factor in both the numerator and denominator of (38) will cause this to cancel out. Keeping
that in mind, we calculate the denominator to be
d2 Nin ~in | = v.
= |S (49)
dAdt
~ = ∂ 1 ∂ ˆ 1 ∂ ˆ
Similarly the outgoing flux is (using ∇ ∂r r̂ + r ∂θ θ + r sin θ ∂φ φ)
e−ikr ∗ eikr
~scat = ~ Im 1 r̂
S f ikr̂ f +O =v |f |2 + O(1/r3 ). (50)
m r r r3 r2
To relate this to the flux per solid angle we use d~a = r2 dΩr̂ to obtain
dσ
= |f (θ, φ)|2 . (52)
dΩ
The Optical Theorem. (52) is valid everywhere except at θ = 0. There we have to also
consider interference between the scattered and the unscattered wave. (Unlike the incoming wave,
the outgoing unscattered does coexist with the scattered wave.) The resulting flux is
f ∗ −ikr
~ ~ −ikz ikz ikr̂ ikr 2
Sout = Im e + e ik ẑe + f e + O(1/r ) (53a)
m r r
∗
r̂ ẑf ik(z−r) r̂f ik(r−z)
vẑ + v 2 |f |2 + v Re
= |{z} e + e (53b)
r {z } r r
~unscat
S
| | {z }
~scat
S ~interference
S
This last term can be thought of as the peffects of interference. We will evaluate it for large r and
ρ2
for θ ≈ 0. Here r̂ ≈ ẑ and we define ρ = x2 + y 2 so that (to leading order) r = z + 2z . Then
Z Z 2π Z ∞
~inter · d~a = v zˆ 2
∗ −ik ρ2z
2
ik ρ2z
S dφ ρdρ f e + fe (54a)
0 0 z
Z ∞ 2
v dρ ikρ2
= 4π Re f e 2z (54b)
z 2
Z0 ∞
2πv iky 4πv
= Re dye 2z f (0) = − Im f (0) (54c)
z 0 k
27
Since the outgoing flux should equal the incoming flux, we can define A to be the beam area and
find Z
4πv
Av = Av + v dΩ|f (θ, φ)|2 − Im f (0). (55)
k
Thus we obtain the following identity, known as the optical theorem:
Z
4π
dΩ|f (θ, φ)|2 = Im f (0) (56)
k
All this is well and good, but we have made no progress at all in computing f (θ, φ). We will
discuss two approximation methods: the partial wave method (which is exact, but yields nice
approximations when k is very small) and the Born approximation, which is a good approximation
when we are scattering off a weak potential. This is analogous to approximations we have seen
before (see Table 1). We will discuss the Born approximation in Section 4.2 and the partial-wave
~ 2 + k 2 )|ψi = U |ψi 2m
(∇ where U≡ V. (57)
~2
~ 2 + k 2 ) to obtain
This looks like a basic linear algebra question. Can we solve it by inverting (∇
? ~2
ψ = (∇ + k 2 )−1 U |ψi? (58)
To answer this, we first review some basic linear algebra. Suppose we want to solve the equation
A~x = ~b (59)
for some normal matrix A. We can write ~b = A−1 ~x only if A is invertible. Otherwise the solution
will not be uniquely defined. More generally, suppose that our vectors live on a space V . Then we
can divide up V as
If we restrict A to the subspace Im A then it is indeed invertible. The solutions to (59) are then
given by
~x = (A|Im A )−1~b + ~x0 where ~x0 ∈ ker A. (61)
Returning now to the quantum case, the operator (∇ ~ 2 + k 2 ) is certainly not invertible. States
~ 2 + k 2 )|ψ0 i = 0 exist, and are plane waves with momentum ~k. But if we restrict
satisfying (∇
28
~ 2 + k 2 ) to the subspace of states with momentum =
(∇ 6 ~k then it is invertible. Define the Green’s
~ 2 2 −1
operator G to be (∇ + k )|p=~k 6 . The calculation of G is rather subtle and the details can be
found in Griffiths. However, on general principles we can make a fair amount of progress. Since
~ 2 + k 2 ) is diagonal in the momentum basis, then G should be as well. Thus G should be written
(∇
as an integral over |p~ihp~| times some function of p~. By Fourier transforming this function we can
equivalently write G in terms of translation operators as
Z
r ·p
~ ~
G = d3~rG(~r)T~r where T~r ≡ e−i ~ . (62)
R 3 go through this more concretely. In the momentum basis we have the completeness relation
Let’s
d p~|p~ihp~| = I which implies
Z
~
∇ + k = d3 p~(−~2 p2 + k 2 )|p~ihp~|.
2 2
(63)
R0 R0
To invert this we might naively write G = d3 p~(−~2 p2 + k 2 )−1 |p~ihp~| where denotes the integral
over all p~ with p 6= ~k. To handle the diverging denominator, one method is to write
Z
G = lim d3 p~(−~2 p2 + k 2 + i)−1 |p~ihp~|. (64)
→0
Finally we can write this in the position basis according to (62) and obtain the position-space
Green’s function G(~r) by Fourier-transforming (−~2 p2 + k 2 + i)−1 . In Griffiths this integral is
carried out obtaining the answer:
eikr
G(~r) = − . (65)
4πr
This function G(~r) is called a Green’s function. We can thus write
−eikr
Z
G= d3~r T~r . (66)
4πr
for some free-particle solution |ψ0 i. Indeed for a scattering problem, we should have ψ0 (~r) = eikz .
(67) is exact, but not very useful because |ψi appears on both the LHS and RHS. However, it will
let us expand |ψi in powers of U .
The first Born approximation consists of replacing the |ψi on the RHS of (67) by |ψ0 i, thus
yielding
|ψi = |ψ0 i + GU |ψ0 i. (68)
The second Born approximation consists of using (68) to approximate |ψi in the RHS of (67), which
yields
|ψi = |ψ0 i + GU (|ψ0 i + GU |ψ0 i) = |ψ0 i + GU |ψ0 i + GU GU |ψ0 i. (69)
Of course we could also rewrite (67) as |ψi = (I − GU )−1 |ψ0 i = n≥0 (GU )n |ψ0 i (since (I − GU )
P
is formally invertible) and truncate this sum at some finite value of n.
29
These results so far have been rather abstract. Plugging in (65) and ψ0 (~r) = eikz we find that
the first Born approximation is
Z
ψ(~r) = ψ0 (~r) + d3~r0 ψ0 (~r0 )G(~r − ~r0 )U (~r0 ) (70a)
Z ~ r0 |
ik|r−~
3 0 ikz 0 e
ikz
= e − d ~r e U (~r0 ) (70b)
4π|~r − ~r0 |
If we assume that the potential is short range and we evaluate this quantity for r far outside the
range of U , then we will have r r0 for all the points where the integral has a nonzero contribution.
In this case |~r − ~r0 | ≈ r − r̂ · ~r0 . Let us further define
~k = kr̂ and ~k 0 = kẑ, (71)
corresponding to the outgoing and incoming wavevectors respectively. Then we have (still in the
first Born approximation)
0
eik|r−~
~ r|
Z
0
ψscat (~r) = − d3~r0 eikz U (~r0 ) (72a)
4π|~r − ~r0 |
Z ~0
ikr−ikr̂·r
0e
≈ − d3~r0 eikz U (~r0 ) (72b)
4πr
r0 ) eikr
Z
3 0 i(k ~ r0 U (~
~ 0 −k)·~
= − d ~r e (72c)
4π r
The quantity in parentheses is then f (θ, φ). If we define V˜ (~q) to be the Fourier transform of V (~r)
then we obtain
m ˜ ~0 ~
f1 (θ, φ) = − V (k − k), (73)
2π ~2
where f1 refers to the first Born approximation. One very simple example is when V (~r) = V0 δ(~r).
mV0
Then we simply have f1 = − 2π~ 2.
A further simplification occurs in the case when V (~r) is centrally symmetric. Then
Z Z ∞ Z 1 Z ∞
4π
Ṽ (~q) ≡ d3~rV (r)eiq~·~r = 2π dr dµr2 V (r)eiqrµ = dr rV (r) sin(qr). (74)
0 −1 q 0
2mβ 2
f1Yukawa (θ) = − .
~2 (µ2 + q 2 )
A good exercise is to rederive the Born approximation by using Fermi’s Golden Rule and
counting the number of outgoing states in the vicinity of a given ~k. See section 7.11 of Sakurai for
details. Another version is in Merzbacher, section 20.1.
30
Rigorous derivation of Green’s functions. The above derivation of G was somewhat informal.
However, once the form of G(~r) is derived informally or even guessed, it can be verified that
~ 2 + k 2 )G acts as the identity on all states with no component in the null space of (∇
(∇ ~ 2 + k 2 ).
This is the content of Griffiths problem 11.8. Implicit is that we are working in a Schwartz space
which rules out exponentially growing wavefunctions, and in turn implies that ∇ ~ 2 has only real
eigenvalues and is in fact Hermitian. A more rigorous derivation of the Born approximation can
also be obtained by using time-dependent perturbation theory as described by Sakurai section 7.11
and Merzbacher section 20.1. I particularly recommend the discussion in Merzbacher.
If we define ul (r) = rRl (r), then the eigenvalue equation becomes (for each l)
2mV l(l + 1)
− u00l + Veff ul = k 2 ul where Veff ≡ 2
+ . (77)
~ r2
If (75) holds, then for sufficiently large r, we can approximate Veff ≈ l(l + 1)/r2 . In this region
the solutions to (77) are given by the spherical Bessel functions. Redefining x ≡ kr and using the
assumption Veff ≈ l(l + 1)/r2 , (77) becomes
l(l + 1)
u00l − ul = −ul . (78)
x2
This has two linearly independent solutions: xjl (x) and xnl (x) where jl (x), nl (x) are the spherical
Bessel functions of the first and second kind respectively, and are defined as
1 d l sin(x) 1 d l cos(x)
l l
jl (x) = (−x) and nl (x) = −(−x) (79)
x dx x x dx x
These can be thought of as “sin-like” and “cos-like” respectively. For l = 0, 1, (79) becomes
sin(x) sin(x) cos(x)
j0 (x) = j1 (x) = − (80a)
x x2 x
cos(x) cos(x) sin(x)
n0 (x) = − n1 (x) = − − (80b)
x x2 x
2
R1
What are Legendre polynomials? One definition starts with the orthogonality condition −1 Pm (x)Pn (x)dx =
2 2
δ . (The δmn is the important term here; 2n+1
2n+1 mn
is a somewhat arbitrary convention.) Then if we apply the
Gram-Schmidt procedure to 1, x, x , . . . , we obtain the Legendre polynomials P0 = 1, P1 = 1, P2 = 12 (3x2 − 1), . . ..
2
Thus any degree-n polynomial can be written as a linear combination of P0 , . . . , Pn and vice-versa.
The reason for (76) is that if ψ(r, θ) is independent of φ then we can write it as a power series in r and z, or
equivalently r and zr = cos(θ). These power series can always be written in the form of (76).
31
One can check (by evaluating the derivatives in (79) repeatedly and keeping only the lowest power
of 1/x) that as x → ∞ we have the asymptotic behavior
1 1
jl (x) → sin(x − lπ/2) and nl (x) → − cos(x − lπ/2). (81)
x x
On the other hand, in the x → 0 limit we can keep track of only the lowest power of x to find that
as x → 0 we have
xl (2l − 1)!!
jl (x) → and nl (x) → − , (82)
(2l + 1)!! xl+1
where (2l + 1)!! ≡ (2l + 1)(2l − 1) · · · 3 · 1.
If jl and nl are sin-like and cos-like, it will be convenient to define functions that resemble
incoming and outgoing waves. These are the spherical Hankel functions of the first and second
kind:
(1) (2)
hl = jl + inl and hl = jl − inl (83)
(1) ikr (1)
For large r, hl (kr) → (−i)l+1 ekr , and so our scattered wave should be proportional to hl . More
(1)
precisely, Rl (r) (i.e. the angular-momentum-l component) should be proportional to hl (kr).
Putting this together we get
when V ≈0 ikz
X (1)
ψ(r, θ) = e + cl hl (kr)Pl (cos(θ)), (84)
l≥0
| {z }
ψscat
for some coefficients cl . It will be convenient to write the cl in terms of new coefficients al as
cl = kil+1 (2l + 1)al , so that we have
when V ≈0 (1)
X
ψscat (r, θ) = k il+1 (2l + 1)al hl (kr)Pl (cos(θ)) (85)
l≥0
when r→∞
X eikr
= (2l + 1)al Pl (cos(θ)) (86)
r
l≥0
| {z }
f (θ)
32
of observing something at a particular position, we will see interference effects between different
momenta, but if we integrate over all positions, these will go away.
The beauty of the partial-wave approach is that it reduces to a series of 1-d scattering problems,
one for each value of l. We have written down the form of the scattered wave. For the incoming
wave, we need to express
P eikz in the form of (76). From the arguments above we can see that eikz
should have form l (Al jl (kr) + Bl nl (kr))Pl (cos θ). The specific solution is given by Rayleigh’s
formula which we state without proof:
X
eikz = eikr cos θ = il (2l + 1)jl (kr)Pl (cos θ) (90)
l≥0
Now we have really expressed this as a series of 1-d scattering problems. Here comes the crucial
simplifying move. Because we are assuming that the collision is elastic, probability and angular-
momentum conservation means that ”what goes in must come out”; i.e.
|1 + 2ikal | = 1. (93)
The outgoing wave’s amplitude must have the same absolute value as the incoming wave’s ampli-
tude. We can rewrite (93) by introducing the phase shift δl defined by
1 + 2ikal = e2iδl . (94)
The factor of 2 is conventional. We can rewrite (94) to solve for al as
2eiδl − 1 eiδl sin(δl ) 1 1
al = = = . (95)
2ik k k cot(δl ) − i
Many equivalent expressions are also possible. In terms of the phase shifts, we can write
1X
f (θ) = (2l + 1)Pl (cos θ)eiδl sin(δl ) (96)
k
l
4π X
σ= 2 (2l + 1) sin2 (δl ). (97)
k
l
As an application we can verify that this satisfies the optical theorem. Observe that Pl (1) = 1 for
all l. Thus
4π 4π 1 X
Im f (0) = (2l + 1)Pl (1) sin2 (δl ) = σ. (98)
k k k
l
Another easy application is partial-wave unitarity, which bounds the total amount of scattered
wave with angular momentum l. Define σl = 4π 2
P
k2
(2l + 1) sin (δ l ), so that σ = l l Then using
σ .
sin2 (δl ) ≤ 1 we have
4π
σl ≤ 2 (2l + 1). (99)
k
33
This bound is called “partial-wave unitarity.”
How to compute phase shifts. Let us look again at the r → ∞ solution. We can use the fact
that overall phase and normalization don’t matter to obtain
(1) (2)
Rl (r) = hl (kr)e2iδl + hl (kr) (100a)
2iδl 2iδl
= (1 + e )jl (kr) + i(e − 1)nl (kr) (100b)
∝ cos(δl )jl (kr) − sin(δl )nl (kr) (100c)
∝ jl (kr) − tan(δl )nl (kr) (100d)
Suppose that we know the interior solution Rl (r) for r ≤ b and that V (r) ≈ 0 for r > b. Then we
R0 (r)
can compute the phase shift by matching Rll (r) for r = b ± .
Here is a simple example. Consider a hard sphere of radius b. Then Rl (r) = 0 for r ≤ b and we
have
−1 jl (kb)
jl (kb) − tan(δl )nl (kb) = 0 =⇒ δl = tan . (101)
nl (kb)
One particularly simple case is when l = 0. Then
sin(kb)
!
j0 (kb)
δ0 = tan−1 = tan−1 kb
= − tan−1 (tan(kb)) = −kb. (102)
n0 (kb) − cos(kb)
kb
It turns out that in general repulsive potentials yield negative phase shifts.
What about larger values of l? Suppose that kb 1. Then using the x → 0 approximations
for jl (x), nl (x), we obtain
(kb)2l+1 (kb)2l+1
k→0 −1
δl −−−→ tan ≈ ∼ (kb)2l+1 . (103)
(2l + 1)!!(2l − 1)!! 2l + 1!!(2l − 1)!!
For l = 0 this is 4πb2 which four times the classical value of πb2 , and for higher value of l this
drops exponentially (assuming kb 1). Even if kb > 1 this drops exponentially once l kb, which
confirms our intuition that the angular momentum should be on the order of ~kb.
Another way to think about this reason to favor low values of l is because the l(l + 1)/r2
term in Veff forms an angular momentum “barrier” that prevents low-energy incoming waves from
penetrating to small enough r to see the potential V (r).
The high-energy limit. The partial wave approximation is easiest to use in the low-k limit
because then we can restrict our attention to a few values of l, or even just l = 0. But for the hard
sphere we can also evaluate the kb 1 limit. In this case we expect to find angular momenta up
to lmax ≡ kb. Thus we approximate the total cross section by
lmax
4π X
σ≈ (2l + 1) sin2 (δl ). (105)
k2
l=0
34
The phases δl will vary over the entire range from 0 to 2π so we simply approximate sin2 (δl ) by its
average value of 1/2. Thus we obtain
kb
2π X
σ≈ (2l + 1) = 2πb2 . (106)
k2
l=0
This is now twice the classical result. Even though the particles are moving quickly they still
diffract like waves. One surprising consequence is that even though a hard sphere leaves a shadow,
there is a small bright spot at the center of the shadow. Indeed the optical theorem predicts that
k
Im f (0) = 4π σ ≈ kb2 /2. Thus |f (0)|2 ≥ (kb)2 b2 /4. For a further discussion of this bright spot and
the role it played in early 19th-century debates about whether light is a particle and/or a wave,
look up “Arago spot” on wikipedia.
Phase shifts. As we have seen, scattering can be understood in terms of phase shifts. Now we
describe a simple physical way of seeing this. If V = 0 then a plane wave has u0 (r) = sin(kr),
due to the u0 (0) = 0 boundary condition. When there is scattering, the phase shift δ0 will become
nonzero and we will have
u0 (r) = sin(kr + δ0 ).
If the potential is attractive then the phase will oscillate more rapidly in the scattering region and
so we will have δ0 > 0 while if it is repulsive then the phase will oscillate more slowly and we will
have δ0 < 0. See Fig. 2 for an illustration.
Figure 2: The phase shift δ0 is positive for attractive potentials and negative for repulsive potentials.
Scattering length. In the regime of low k it turns out that many potentials behave qualitatively
like what we have seen with the hard sphere, with a characteristic length scale called a “scattering
length.” To derive this suppose there is some b such that V (r) ≈ 0 for r ≥ b. In this region we
have u0 (r) ≈ sin(kb + δ0 ) (neglecting normalization). In the vicinity of b we have
35
If we extrapolate to smaller values of r our approximation hits 0 at r = a where
u0 (b) tan(kb + δ0 )
a=b− 0 =b− .
u0 (b) k
Using the tan addition formula and taking the limit k → 0 we find
36
MIT OpenCourseWare
http://ocw.mit.edu
For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.