Lect 3
Lect 3
Lect 3
Science
BITS Pilani MFDS Team
Pilani Campus
BITS Pilani
Pilani Campus
Lecture No. 3
Agenda
• Gerschgorin’s theorem
• Similarity transformation
• Diagonalization of matrices
• Quadratic forms
Solution. (continued 1)
(a) Eigenvalues. (continued 1)
Transferring the terms on the right to the left, we get
( −5 − λ ) x1 + 2 x2 = 0
(2*)
2 x1 + ( −2 − λ ) x2 = 0
This can be written in matrix notation
(3*) (A − λI)x = 0
Because (1) is Ax − λx = Ax − λIx = (A − λI)x = 0,
which gives (3*).
Solution. (continued 2)
(a) Eigenvalues. (continued 2)
We see that this is a homogeneous linear system. By
Cramer’s theorem in Sec. 7.7 it has a nontrivial solution (an
eigenvector of A we are looking for) if and only if its
coefficient determinant is zero, that is,
−5 − λ 2
D(λ ) = det( A − λI) =
2 −2 − λ
(4*)
= ( −5 − λ )( −2 − λ ) − 4 = λ 2 + 7 λ + 6 = 0.
Solution. (continued 3)
(a) Eigenvalues. (continued 3)
We call D(λ) the characteristic determinant or, if expanded,
the characteristic polynomial, and D(λ) = 0 the
characteristic equation of A. The solutions of this quadratic
equation are λ1 = −1 and λ2 = −6. These are the eigenvalues
of A.
(b1) Eigenvector of A corresponding to λ1. This vector is
obtained from (2*) with λ = λ1 = −1, that is,
−4 x1 + 2 x2 = 0
2 x1 − x2 = 0.
Solution. (continued 4)
(b1) Eigenvector of A corresponding to λ1. (continued)
A solution is x2 = 2x1, as we see from either of the two
equations, so that we need only one of them. This
determines an eigenvector corresponding to λ1 = −1 up to a
scalar multiple. If we choose x1 = 1, we obtain the
eigenvector
⎡ ⎤ ⎡ −5 2 ⎤⎡ 1 ⎤ ⎡ −1 ⎤
v = ⎢ 1
⎥ , Check: Av = ⎢ ⎥⎢ ⎥=⎢ ⎥ = (−1)v = λ1v.
⎣ 2 ⎦ ⎣ 2 −2 ⎦⎣ 2 ⎦ ⎣ −2 ⎦
Solution. (continued 5)
(b2) Eigenvector of A corresponding to λ2.
For λ = λ2 = −6, equation (2*) becomes
x1 + 2 x2 = 0
2 x1 + 4 x2 = 0.
A solution is x2 = −x1/2 with arbitrary x1. If we choose x1 = 2,
we get x2 = −1. Thus an eigenvector of A corresponding to
λ2 = −6 is
⎡ 2 ⎤ ⎡ −5 2 ⎤⎡ 2 ⎤ ⎡ −12 ⎤
w = ⎢ ⎥ , Check: Aw = ⎢ ⎥⎢ ⎥=⎢ ⎥ = (−6)w = λ2 w.
⎣ −1 ⎦ ⎣ 2 −2 ⎦⎣ −1 ⎦ ⎣ 6 ⎦
Theorem 1
Eigenvalues
The eigenvalues of a square matrix A are the roots of the
characteristic equation (4) of A.
Hence an n × n matrix has at least one eigenvalue and at most n
numerically different eigenvalues.
Theorem 2
Eigenvectors, Eigenspace
If w and x are eigenvectors of a matrix A corresponding to the
same eigenvalue λ, so are w + x (provided x ≠ −w) and kx for
any k ≠ 0.
Hence the eigenvectors corresponding to one and the same
eigenvalue λ of A, together with 0, form a vector space called the
eigenspace of A corresponding to that λ.
Solution. (continued 1)
To find eigenvectors, we apply the Gauss elimination (Sec.
7.3) to the system (A − λI)x = 0, first with λ = 5
and then with λ = −3 . For λ = 5 the characteristic matrix is
⎡ −7 2 −3 ⎤
A − λI = A − 5I = ⎢⎢ 2 −4 −6 ⎥⎥ .
⎢⎣ −1 −2 −5 ⎥⎦
It row-reduces to ⎡ −7 2 −3 ⎤
⎢ 0 −24 / 7 −48 / 7 ⎥ .
⎢ ⎥
⎢⎣ 0 0 0 ⎥⎦
Solution. (continued 2)
Hence it has rank 2. Choosing x3 = −1 we have x2 = 2 from
24 48
− x
2
− x 3 = 0 and then x1 = 1 from −7x1 + 2x2 − 3x3 = 0.
7 7
Hence an eigenvector of A corresponding to λ = 5 is
x1 = [1 2 −1]T.
For λ = −3 the characteristic matrix
⎡ 1 2 −3 ⎤
A − λI = A + 3I = ⎢ 2 4 −6 ⎥
⎢ ⎥
⎡ 1 2 −3 ⎤
⎢
⎣ − 1 −2 3 ⎥
⎦ ⎢0 0 0 ⎥ .
row-reduces to ⎢ ⎥
⎢⎣0 0 0 ⎥⎦
Solution. (continued 3)
Hence it has rank 1.
From x1 + 2x2 − 3x3 = 0 we have x1 = −2x2 + 3x3. Choosing
x2 = 1, x3 = 0 and x2 = 0, x3 = 1, we obtain two linearly
independent eigenvectors of A corresponding to λ = −3 [as
they must exist by (5), Sec. 7.5, with rank = 1 and n = 3],
⎡ −2 ⎤ ⎡3⎤
x 2 = ⎢⎢ 1 ⎥⎥ and x 3 = ⎢⎢0 ⎥⎥ .
⎢⎣ 0 ⎥⎦ ⎢⎣ 1 ⎥⎦
Diagonalization of a Matrix
Diagonalize
⎡ 7.3 0.2 −3.7 ⎤
A = ⎢⎢ −11.5 1.0 5.5 ⎥⎥ .
⎢⎣ 17.7 1.8 −9.3 ⎥⎦
Solution.
The characteristic determinant gives the characteristic
equation −λ3 −λ2 + 12λ = 0. The roots (eigenvalues of A)
are λ1 = 3, λ2 = −4, λ3 = 0. By the Gauss elimination applied
to (A − λI)x = 0 with λ = λ1, λ2, λ3 we find eigenvectors and
then X−1 by the Gauss–Jordan elimination
Solution. (continued 1)
The results are
⎡ −1⎤ ⎡ 1 ⎤ ⎡2⎤ ⎡ −1 1 2⎤
⎢ 3 ⎥ , ⎢ −1 ⎥ , ⎢1⎥ ,
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ X = ⎢⎢ 3 −1 1 ⎥⎥ ,
⎢⎣ −1⎥⎦ ⎢⎣ 3 ⎥⎦ ⎢⎣ 4 ⎥⎦ ⎢⎣ −1 3 4 ⎥⎦
⎡ −0.7 0.2 0.3 ⎤
X −1 = ⎢⎢ −1.3 −0.2 0.7 ⎥⎥ .
⎢⎣ 0.8 0.2 −0.2 ⎥⎦
Solution. (continued 2)
Calculating AX and multiplying by X−1 from the left, we
thus obtain
⎡ −0.7 0.2 0.3 ⎤ ⎡ −3 −4 0 ⎤ ⎡ 3 0 0 ⎤
D X −1
AX ⎢ −1.3 −0.2 0.7 ⎥⎢ 9 4 0 ⎥ = ⎢0 −4 0 ⎥ .
= = ⎢ ⎥⎢ ⎥ ⎢ ⎥
⎢⎣ 0.8 0.2 −0.2 ⎥⎦ ⎢⎣ −3 −12 0 ⎥⎦ ⎢⎣0 0 0 ⎥⎦
Q = x T Ax = ∑∑ a jk x j xk
j=1 k=1
= 3 x12 + 10 x1 x2 + 2 x .
2
2
Theorem 5
Principal Axes Theorem
Find out what type of conic section the following quadratic form represents
and transform it to principal axes:
Q = 17 x12 − 30x1x2 + 17 x2 2 = 128.
Solution. We have Q = xTAx, where
⎡ 17 −15 ⎤ ⎡ x1 ⎤
A=⎢ ⎥ , x = ⎢ ⎥.
⎣ −15 17 ⎦ ⎣ x2 ⎦
Solution. (continued 1)
This gives the characteristic equation (17 − λ)2 − 152 = 0. It
has the roots λ1 = 2, λ2 = 32. Hence (10) becomes
2 2
Q = 2 y1
+ 32 y 2
.
We see that Q = 128 represents the ellipse 2y12 + 32y22 = 128,
that is,
y12 y2 2
2
+ 2 = 1.
8 2
Solution. (continued 2)
If we want to know the direction of the principal axes in the
x1x2-coordinates, we have to determine normalized
eigenvectors from (A − λI)x = 0 with λ = λ1 = 2 and
λ = λ2 = 32 and then use (9). We get
⎡ ⎤ ⎡ ⎤
⎢ 1 / 2 ⎥ and ⎢ −1 / 2 ⎥
⎢⎣ 1 / 2 ⎥⎦ ⎢⎣ 1 / 2 ⎥⎦
Observe that the above vectors are orthonormal