EIGEN
EIGEN
EIGEN
EXAMPLE: The the standard vectors ~ei form an eigenbasis of −In . Their eigen-
values are −1. More generally, if D is diagonal, the standard vectors form an
eigenbasis with associated eigenvalues the corresponding entries on the diagonal.
EXAMPLE: If ~v is an eigenvector of A with eigenvalue λ, then ~v is an eigenvector
of A3 with eigenvalue λ3 .
EXAMPLE: 0 is an eigenvalue of A if and only if A is not invertible. Indeed, 0
is an eigenvalue ⇐⇒ there is a non-zero ~v so A~v = ~0 true ⇐⇒ ~v ∈ ker A so ker A
is non-trivial ⇐⇒ A not invertible.
1
2 EIGENVALUES AND EIGENVECTORS
2. Characteristic Equaiton
One of the hardest (computational) problems in linear algebra is to determine the
eigenvalues of a matrix. This is because, unlike everything else we have considered
so far, it is a non-linear problem. That being said, it is still a tractable problem
(especially for small matrices).
To understand the approach. Observe that if λ is an eigenvalue of A, then there
is a non-zero ~v so thatA~v = λ~v . That is,
A~v = (λIn )~v ⇐⇒ (A − λIn )~v = ~0 ⇐⇒ ~v ∈ ker(A − λIn ).
From which we conclude that A − λIn is not invertible and so det(A − λIn ) = 0.
In summary,
Theorem 2.1. λ is an eigenvalue of A if and only if
det(A − λIn ) = 0.
The equation det(A − λIn ) = 0 is called the characteristic equation of A.
EXAMPLE: Find the eigenvalues of
2 3
A= .
3 2
The characteristic equation is
2−λ 3
det(A − λI2 ) = det = λ2 − 4λ − 5 = (λ + 1)(λ − 5)
3 2−λ
Hence, the eigenvalues are λ = −1 and λ = 5. To find corresponding eigenvectors
we seek non-trivial solutions to
2 − (−1) 3 x1 2 − (5) 3 x1
= ~0 and = ~0
3 2 − (−1) x2 3 2 − (5) x2
By inspection the non-trivial solutions are
1 1
and .
−1 1
Hence,
2 3 1 1 1 1 −1 0
=
3 2 −1 1 −1 1 0 5
So we have diagonalized A.
EIGENVALUES AND EIGENVECTORS 3
3. Characteristic Polynomial
As we say for a 2 × 2 matrix, the characteristic equation reduces to finding the
roots of an associated quadratic polynomial. More generally, for a n × n matrix A,
the characteristic equation det(A − λIn ) = 0 reduces to finding roots of a degree n
polynomial o fthe form
fA (λ) = (−1)n λn + (−1)n−1 (trA)λn−1 + · + det A
this is called the characteristic polynomial of A. To see why this is true observe
that if P is the diagonal pattern of A − λIn , then
prod(P ) = (a11 − λ) · · · (ann − λ) = (−1)n λn + (−1)n−1 (trA)λn−2 + RP (λ)
where RP (λ) is a polynomial of degree at most n − 2 that depends on P (and of
course also on A and λ). If P is some other pattern of A − λIn , then at least two
entries are not on the diagonal. Hence,
prod(P ) = RP (λ)
for some RP (λ) that is a polynomial of degree at most n − 2 that depends on P .
Hence,
fA (λ) = det(A − λIn ) = (−1)n λn + (−1)n−1 (trA)λn−1 + R(λ)
where R(λ) has degree at most n − 2. Finally, fA (0) = det(A) and so the constant
term of fA is det A as claimed.
4 EIGENVALUES AND EIGENVECTORS
4. Eigenspaces
Consider an eigenvalue λ of A ∈ Rn×n . We define the eigenspace associated to
λ to be
Eλ = ker(A − λIn ) = {~v ∈ Rn : A~v = λ~v } ⊂ Rn .
Observe that dim Eλ ≥ 1. All non-zero elements of Eλ are eigenvectors of A with
eigenvalue λ.
1 0
EXAMPLE: A = has repeated eigenvalue 1. Clearly,
0 1
E1 = ker(A − I2 ) = ker(02×2 ) = R2
EIGENVALUES AND EIGENVECTORS 5
1 2
Similarly, the matrix B = has one repeated eigenvalue 1. However,
0 1
0 2 1
ker(B − I2 ) = ker = span( ).
0 0 0
Motivated by this example, define the geometric multiplicity of an eigenvalue λ
of A ∈ Rn×n tobe
gemu(λ) = null(A − λIn ) = n − rank(A − λIn ) ≥ 1.
5. Diagonalizable Matrices
We are now ready to give a computable condition that will allow us to determine
an answer to our central question in this part of the course: When is A ∈ Rn×n
diagonalizable?
Theorem 5.1. A matrix A ∈ Rn×n is diagonalizable if and only if the sum of the
geometric multiplicities of all of the eigenvalues of A is n.
Theorem 5.2. Fix a matrix A ∈ Rn×n and let ~v1 , . . . , ~vs be a set of vectors formed
by concatenating a basis of each non-trivial eigenspace of A. This set is linearly
independent (and so s ≤ n.)
Proof. If the vectors ~vi are not linearly independent, then they at least one is
redundant. Let ~vm be the first redundant vector on the list That is for some
1 ≤ m ≤ s we can write
m−1
X
~vm = ci~vi
i=1
and cannot do this for any smaller m. This means ~v1 , . . . , ~vm−1 are linearly inde-
pendent.
Let λm be the eigenvalue associated to ~vm . Observe, there must be some 1 ≤ k ≤
m − 1 so that λk 6= λm and ck 6= 0 as otherwise we would have a non-trivial linear
relation for of a set of linearly independent vectors in Eλm (which is impossible).
Clearly,
m−1
X
~0 = (A − λm In )~vm = ci (λi − λm )~vi
i=1
The main theorem follows easily form this. Indeed, the hypotheses gives n lin
indep vectors all which are eigenvectors of A. That is, an eigenbasis of A.
d kx
D(ekx ) = e = kekx
dx
so each fk (x) = ekx is an eigenfunction and every scalar k ∈ R is an eigenvalue.
EXAMPLE: Consider the map T : P2 → P2 given by T (p) = p(2x + 1). Is T
diagonalizable? As usual it is computationally more convenient to work in some
basis. To that end, let U = (1, x, x2 ) be the usual basis of P2 . As
1
[T (1)]U = [1]U = 0
0
1
[T (x)]U = [2x + 1]U = 2
0
8 EIGENVALUES AND EIGENVECTORS
1
[T (x2 )]U = [4x2 + 4x + 1]U = 4 ,
4
the associated matrix is
1 1 1
A = [T ]U = 0 2 4 .
0 0 4
This matrix is upper triangular, with distinct eigenvalues 1, 2 and 4 This means T
is also diagonalizable and has the same eigenvalues. We compute (for A)
0 1 1 1
E1 = ker(A − I3 ) = ker 0 1 4 = span(0)
0 0 3 0
−1 1 1 1
E2 = ker(A − 2I3 ) = 0 0 4 = span(1)
0 0 2 0
−3 1 1 1
E3 = ker(A − 4I3 ) = 0 −2 4 = span 2
0 0 0 1
1 1 1
Hence, A can be diagonalized by S = 0 1 2. Going back to T we check
0 0 1
T (1) = 1
T (1 + x) = 1 + (2x + 1) = 2(x + 1)
T (1 + 2x + x2 ) = 1 + 2(2x + 1) + (2x + 1)2 = 4(1 + 2x + x2 )
In particular, B = (1, 1 + x, 1 + 2x + x2 ) is an eigenbasis and
1 0 0
[T ]B = 0 2 0 .
0 0 4
Claim (3) can be shown as follows: By claim (1), fA (λ) = fB (λ) and so A and
B have the same eigenvalues with the same algebraic multiplicities. Furthermore,
if λ is an eigenvalue of both A and B, then A − λIn is similar to B − λIn . Hence,
by claim (2)
gemuA (λ) = null(A − λIn ) = null(B − λIn ) = gemuB (λ)
so the geometric multiplicities are the same as well. Notice Eλ (A) 6= Eλ (B) in
general. Finally, claim (4) Follows from claim (1) and the observation that the
characteristic polynomial encodes the trace and determinant.
2 3 3 2
EXAMPLE: The matrix is not similar to as their traces are
5 7 8 5
different.
Even if a matrix is not diagonalizable, it is still similar to one of a list of canonical
matrices. To do this is full generality is beyond the scope of this course. To illustrate
the idea we present the list for 2 × 2 matrices.
Theorem 8.2. Any A ∈ R2×2 is similar to one of the following:
λ1 0
(1) for λ1 , λ2 ∈ R. This occurs when A is diagonalizable.
0 λ2
λ 1
(2) . This occurs when A has repeated eigenvalue λ with gemu(λ) = 1.
0 λ
Such
A is sometimes called defective.
λ1 −λ2 p cos θ − sin θ
(3) = λ21 + λ22 where λ2 > 0 and π > θ > 0.
λ2 λ1 sin θ cos θ
This corresponds to an A with no real eigenvalues (in this case the complex
eigenvalues of A are λ1 ± iλ2 ).