Linear Algebra Chapter 5: Eigenvalues and Eigenvectors Section 5.2. - - PowerPoint PPT Presentation

linear algebra
SMART_READER_LITE
LIVE PREVIEW

Linear Algebra Chapter 5: Eigenvalues and Eigenvectors Section 5.2. - - PowerPoint PPT Presentation

Linear Algebra Chapter 5: Eigenvalues and Eigenvectors Section 5.2. DiagonalizationProofs of Theorems April 13, 2020 () Linear Algebra April 13, 2020 1 / 34 Table of contents Theorem 5.2. Matrix Summary of Eigenvalues of A 1 Corollary


slide-1
SLIDE 1

Linear Algebra

April 13, 2020 Chapter 5: Eigenvalues and Eigenvectors Section 5.2. Diagonalization—Proofs of Theorems

() Linear Algebra April 13, 2020 1 / 34

slide-2
SLIDE 2

Table of contents

1

Theorem 5.2. Matrix Summary of Eigenvalues of A

2

Corollary 1. A Criterion for Diagonalization

3

Corollary 2. Computation of Ak

4

Example 5.2.A

5

Theorem 5.3. Independence of Eigenvectors

6

Page 315 Number 6

7

Page 315 Number 18

8

Page 315 Number 10

9

Page 316 Number 22

10 Page 316 Number 24 11 Page 316 Number 26

() Linear Algebra April 13, 2020 2 / 34

slide-3
SLIDE 3

Theorem 5.2. Matrix Summary of Eigenvalues of A

Theorem 5.2

Theorem 5.2. Matrix Summary of Eigenvalues of A. Let A be an n × n matrix and let λ1, λ2, . . . , λn be (possibly complex) scalars and v1, v2, . . . , vn be nonzero vectors in n-space. Let C be the n × n matrix having vj as jth column vector and let D =        λ1 · · · λ2 · · · λ3 · · · . . . . . . . . . ... . . . · · · λn        . Then AC = CD if and only if λ1, λ2, . . . , λn are eigenvalues of A and vj is an eigenvector of A corresponding to λj for j = 1, 2, . . . , n.

() Linear Algebra April 13, 2020 3 / 34

slide-4
SLIDE 4

Theorem 5.2. Matrix Summary of Eigenvalues of A

Theorem 5.2 (continued)

  • Proof. We have

CD =     . . . . . . . . .

  • v1
  • v2

· · ·

  • vn

. . . . . . . . .            λ1 · · · λ2 · · · λ3 · · · . . . . . . . . . ... . . . · · · λn        =     . . . . . . . . . λ1 v1 λ2 v2 · · · λn vn . . . . . . . . .     . Also, AC = A     . . . . . . . . .

  • v1
  • v2

· · ·

  • vn

. . . . . . . . .     . Therefore, AC = CD if and only if A vj = λj vj.

() Linear Algebra April 13, 2020 4 / 34

slide-5
SLIDE 5

Corollary 1. A Criterion for Diagonalization

Corollary 1

Corollary 1. A Criterion for Diagonalization. An n × n matrix A is diagonalizable if and only if n-space has a basis consisting of eigenvectors of A.

  • Proof. Suppose A is diagonalizable. Then by Definition 5.3,

“Diagonalizable Matrix,” C −1AC = D for some invertible n × n matrix C.

() Linear Algebra April 13, 2020 5 / 34

slide-6
SLIDE 6

Corollary 1. A Criterion for Diagonalization

Corollary 1

Corollary 1. A Criterion for Diagonalization. An n × n matrix A is diagonalizable if and only if n-space has a basis consisting of eigenvectors of A.

  • Proof. Suppose A is diagonalizable. Then by Definition 5.3,

“Diagonalizable Matrix,” C −1AC = D for some invertible n × n matrix C. Then C(C −1AC) = CD or AC = CD and so by Theorem 5.3, “Matrix Summary of Eigenvalues of A,” the jth column of C is an eigenvector of A corresponding to eigenvalue λj, where the n (possibly complex) eigenvalues of A are λ1, λ2, . . . , λn.

() Linear Algebra April 13, 2020 5 / 34

slide-7
SLIDE 7

Corollary 1. A Criterion for Diagonalization

Corollary 1

Corollary 1. A Criterion for Diagonalization. An n × n matrix A is diagonalizable if and only if n-space has a basis consisting of eigenvectors of A.

  • Proof. Suppose A is diagonalizable. Then by Definition 5.3,

“Diagonalizable Matrix,” C −1AC = D for some invertible n × n matrix C. Then C(C −1AC) = CD or AC = CD and so by Theorem 5.3, “Matrix Summary of Eigenvalues of A,” the jth column of C is an eigenvector of A corresponding to eigenvalue λj, where the n (possibly complex) eigenvalues of A are λ1, λ2, . . . , λn. Since C is invertible then by Theorem 1.12, “Conditions for A−1 to Exist” (see part (v)) the span of the column vectors of C span n-space (that is, span either Rn as addressed in Theorem 1.12, or span Cn if we use complex scalars; we need a result like Theorem 1.12 valid in the complex setting, but such a result holds).

() Linear Algebra April 13, 2020 5 / 34

slide-8
SLIDE 8

Corollary 1. A Criterion for Diagonalization

Corollary 1

Corollary 1. A Criterion for Diagonalization. An n × n matrix A is diagonalizable if and only if n-space has a basis consisting of eigenvectors of A.

  • Proof. Suppose A is diagonalizable. Then by Definition 5.3,

“Diagonalizable Matrix,” C −1AC = D for some invertible n × n matrix C. Then C(C −1AC) = CD or AC = CD and so by Theorem 5.3, “Matrix Summary of Eigenvalues of A,” the jth column of C is an eigenvector of A corresponding to eigenvalue λj, where the n (possibly complex) eigenvalues of A are λ1, λ2, . . . , λn. Since C is invertible then by Theorem 1.12, “Conditions for A−1 to Exist” (see part (v)) the span of the column vectors of C span n-space (that is, span either Rn as addressed in Theorem 1.12, or span Cn if we use complex scalars; we need a result like Theorem 1.12 valid in the complex setting, but such a result holds).

() Linear Algebra April 13, 2020 5 / 34

slide-9
SLIDE 9

Corollary 1. A Criterion for Diagonalization

Corollary 1 (continued)

Proof (continued). Since n-space is dimension n and the column vectors

  • f C form a set of n vectors which span n-space then the vectors must be

linearly independent (by Theorem 2.3(3a), “Existence and Determination

  • f Bases”) and so are a basis for n-space by Definition 3.6, “Basis of a

Vector Space.” Conversely, suppose n-space has a basis consisting of eigenvectors of A, say v1, v2, . . . , vn where vj is an eigenvector corresponding to eigenvalue λj. Then by Definition 3.6, “Basis of a Vector Space,” the vectors are linearly independent. So if we form matrix C where the jth column of C is

  • vj then C is invertible by Theorem 1.12, “Conditions for A−1 to Exist.”

() Linear Algebra April 13, 2020 6 / 34

slide-10
SLIDE 10

Corollary 1. A Criterion for Diagonalization

Corollary 1 (continued)

Proof (continued). Since n-space is dimension n and the column vectors

  • f C form a set of n vectors which span n-space then the vectors must be

linearly independent (by Theorem 2.3(3a), “Existence and Determination

  • f Bases”) and so are a basis for n-space by Definition 3.6, “Basis of a

Vector Space.” Conversely, suppose n-space has a basis consisting of eigenvectors of A, say v1, v2, . . . , vn where vj is an eigenvector corresponding to eigenvalue λj. Then by Definition 3.6, “Basis of a Vector Space,” the vectors are linearly independent. So if we form matrix C where the jth column of C is

  • vj then C is invertible by Theorem 1.12, “Conditions for A−1 to Exist.” By

Theorem 5.2, “Matrix Summary of Eigenvalues of A,” with D as a diagonal matrix with djj = λj, then AC = CD. Since C is invertible, C −1AC = C −1CD or C −1AC = D. So A is diagonalizable.

() Linear Algebra April 13, 2020 6 / 34

slide-11
SLIDE 11

Corollary 1. A Criterion for Diagonalization

Corollary 1 (continued)

Proof (continued). Since n-space is dimension n and the column vectors

  • f C form a set of n vectors which span n-space then the vectors must be

linearly independent (by Theorem 2.3(3a), “Existence and Determination

  • f Bases”) and so are a basis for n-space by Definition 3.6, “Basis of a

Vector Space.” Conversely, suppose n-space has a basis consisting of eigenvectors of A, say v1, v2, . . . , vn where vj is an eigenvector corresponding to eigenvalue λj. Then by Definition 3.6, “Basis of a Vector Space,” the vectors are linearly independent. So if we form matrix C where the jth column of C is

  • vj then C is invertible by Theorem 1.12, “Conditions for A−1 to Exist.” By

Theorem 5.2, “Matrix Summary of Eigenvalues of A,” with D as a diagonal matrix with djj = λj, then AC = CD. Since C is invertible, C −1AC = C −1CD or C −1AC = D. So A is diagonalizable.

() Linear Algebra April 13, 2020 6 / 34

slide-12
SLIDE 12

Corollary 2. Computation of Ak

Corollary 2

Corollary 2. Computation of Ak. Let an n × n matrix A have n eigenvectors and eigenvalues, giving rise to the matrices C and D so that AC = CD, as described in Theorem 5.2. If the eigenvectors are independent, then C is an invertible matrix and C −1AC = D. Under these conditions, we have Ak = CDkC −1.

  • Proof. By Corollary 1, if the eigenvectors of A are independent, then A is

diagonalizable and so C is invertible. Now consider Ak = (CDC −1)(CDC −1) · · · (CDC −1)

  • k factors

= CD(C −1C)D(C −1C)D(C −1C) · · · (C −1C)DC −1 = CDIDID · · · IDC −1 = C DDD · · · D

  • k factors

C −1 = CDkC −1

() Linear Algebra April 13, 2020 7 / 34

slide-13
SLIDE 13

Corollary 2. Computation of Ak

Corollary 2

Corollary 2. Computation of Ak. Let an n × n matrix A have n eigenvectors and eigenvalues, giving rise to the matrices C and D so that AC = CD, as described in Theorem 5.2. If the eigenvectors are independent, then C is an invertible matrix and C −1AC = D. Under these conditions, we have Ak = CDkC −1.

  • Proof. By Corollary 1, if the eigenvectors of A are independent, then A is

diagonalizable and so C is invertible. Now consider Ak = (CDC −1)(CDC −1) · · · (CDC −1)

  • k factors

= CD(C −1C)D(C −1C)D(C −1C) · · · (C −1C)DC −1 = CDIDID · · · IDC −1 = C DDD · · · D

  • k factors

C −1 = CDkC −1

() Linear Algebra April 13, 2020 7 / 34

slide-14
SLIDE 14

Example 5.2.A

Example 5.2.A

Example 5.2.A. Diagonalize A =

  • 5

−3 −6 2

  • and calculate Ak.
  • Solution. We have

A − λI =

  • 5

−3 −6 2

  • − λ

1 1

  • =

5 − λ −3 −6 2 − λ

  • . So the

characteristic polynomial is p(λ) = det(A − λI) =

  • 5 − λ

−3 −6 2 − λ

  • = (5−λ)(2−λ)−(−3)(−6) = 10−7λ+λ2−18 = λ2−7λ−8 = (λ+1)(λ−8).

() Linear Algebra April 13, 2020 8 / 34

slide-15
SLIDE 15

Example 5.2.A

Example 5.2.A

Example 5.2.A. Diagonalize A =

  • 5

−3 −6 2

  • and calculate Ak.
  • Solution. We have

A − λI =

  • 5

−3 −6 2

  • − λ

1 1

  • =

5 − λ −3 −6 2 − λ

  • . So the

characteristic polynomial is p(λ) = det(A − λI) =

  • 5 − λ

−3 −6 2 − λ

  • = (5−λ)(2−λ)−(−3)(−6) = 10−7λ+λ2−18 = λ2−7λ−8 = (λ+1)(λ−8).

So the eigenvalues of A are λ1 = −1 and λ2 = 8. To find the eigenvectors corresponding to each eigenvalue we consider the formula A v = λ v or (A − λI) v = 0.

() Linear Algebra April 13, 2020 8 / 34

slide-16
SLIDE 16

Example 5.2.A

Example 5.2.A

Example 5.2.A. Diagonalize A =

  • 5

−3 −6 2

  • and calculate Ak.
  • Solution. We have

A − λI =

  • 5

−3 −6 2

  • − λ

1 1

  • =

5 − λ −3 −6 2 − λ

  • . So the

characteristic polynomial is p(λ) = det(A − λI) =

  • 5 − λ

−3 −6 2 − λ

  • = (5−λ)(2−λ)−(−3)(−6) = 10−7λ+λ2−18 = λ2−7λ−8 = (λ+1)(λ−8).

So the eigenvalues of A are λ1 = −1 and λ2 = 8. To find the eigenvectors corresponding to each eigenvalue we consider the formula A v = λ v or (A − λI) v = 0.

() Linear Algebra April 13, 2020 8 / 34

slide-17
SLIDE 17

Example 5.2.A

Example 5.2.A (continued 1)

Solution (continued). λ1 = −1. With v1 = [v1, v2] an eigenvector corresponding to the eigenvalue λ1 = −1 we need (A − (−1)I) v1 =

  • 0. So we consider the

augmented matrix 5 − (−1) −3 −6 2 − (−1)

  • =
  • 6

−3 −6 3 R2→R2+R1

  • 6

−3

  • .

() Linear Algebra April 13, 2020 9 / 34

slide-18
SLIDE 18

Example 5.2.A

Example 5.2.A (continued 1)

Solution (continued). λ1 = −1. With v1 = [v1, v2] an eigenvector corresponding to the eigenvalue λ1 = −1 we need (A − (−1)I) v1 =

  • 0. So we consider the

augmented matrix 5 − (−1) −3 −6 2 − (−1)

  • =
  • 6

−3 −6 3 R2→R2+R1

  • 6

−3

  • .

So we need 6v1 − 3v2 = = 0 or v1 = (1/2)v2 v2 = v2

  • r, with r = v2/2 as

a free variable, v1 = r v2 = 2r . So the collection of all eigenvectors of λ1 = −1 is v1 = r 1 2

  • where r ∈ R, r = 0.

() Linear Algebra April 13, 2020 9 / 34

slide-19
SLIDE 19

Example 5.2.A

Example 5.2.A (continued 1)

Solution (continued). λ1 = −1. With v1 = [v1, v2] an eigenvector corresponding to the eigenvalue λ1 = −1 we need (A − (−1)I) v1 =

  • 0. So we consider the

augmented matrix 5 − (−1) −3 −6 2 − (−1)

  • =
  • 6

−3 −6 3 R2→R2+R1

  • 6

−3

  • .

So we need 6v1 − 3v2 = = 0 or v1 = (1/2)v2 v2 = v2

  • r, with r = v2/2 as

a free variable, v1 = r v2 = 2r . So the collection of all eigenvectors of λ1 = −1 is v1 = r 1 2

  • where r ∈ R, r = 0.

() Linear Algebra April 13, 2020 9 / 34

slide-20
SLIDE 20

Example 5.2.A

Example 5.2.A (continued 2)

Solution (continued). λ2 = 8. As above, we consider (A − 8I) v2 = 0 and consider the augmented matrix 5 − (8) −3 −6 2 − (8)

  • =

−3 −3 −6 −6

  • R2→R2−2R1
  • −3

−3 R1→R1/(−3)

  • 1

1

  • .

So we need v1 + v2 = = 0 or v1 = −v2 v2 = v2

  • r, with s = v2 as a free

variable, v1 = −s v2 = s .

() Linear Algebra April 13, 2020 10 / 34

slide-21
SLIDE 21

Example 5.2.A

Example 5.2.A (continued 2)

Solution (continued). λ2 = 8. As above, we consider (A − 8I) v2 = 0 and consider the augmented matrix 5 − (8) −3 −6 2 − (8)

  • =

−3 −3 −6 −6

  • R2→R2−2R1
  • −3

−3 R1→R1/(−3)

  • 1

1

  • .

So we need v1 + v2 = = 0 or v1 = −v2 v2 = v2

  • r, with s = v2 as a free

variable, v1 = −s v2 = s . So the collection of all eigenvectors of λ2 = 8 is

  • v2 = s

−1 1

  • where s ∈ R, s = 0.

() Linear Algebra April 13, 2020 10 / 34

slide-22
SLIDE 22

Example 5.2.A

Example 5.2.A (continued 2)

Solution (continued). λ2 = 8. As above, we consider (A − 8I) v2 = 0 and consider the augmented matrix 5 − (8) −3 −6 2 − (8)

  • =

−3 −3 −6 −6

  • R2→R2−2R1
  • −3

−3 R1→R1/(−3)

  • 1

1

  • .

So we need v1 + v2 = = 0 or v1 = −v2 v2 = v2

  • r, with s = v2 as a free

variable, v1 = −s v2 = s . So the collection of all eigenvectors of λ2 = 8 is

  • v2 = s

−1 1

  • where s ∈ R, s = 0.

() Linear Algebra April 13, 2020 10 / 34

slide-23
SLIDE 23

Example 5.2.A

Example 5.2.A (continued 3)

Solution (continued). If we take r = s = 1 then we have the eigenvalues λ1 = −1 and λ2 = 8 with corresponding eigenvectors v1 = 1 2

  • and
  • v2 =

−1 1

  • , respectively. So by Theorem 5.2, “Matrix Summary of

Eigenvalues of A,” with C =     . . . . . .

  • v1
  • v2

. . . . . .     = 1 −1 2 1

  • and

D = λ1 λ2

  • =

−1 8

  • we have AC = CD. We find C −1 (by Note

1.5.A, “Computation of Inverses”): [C|I] = 1 −1 1 2 1 1 R2→R2−2R1

  • 1

−1 1 3 −2 1

  • ()

Linear Algebra April 13, 2020 11 / 34

slide-24
SLIDE 24

Example 5.2.A

Example 5.2.A (continued 3)

Solution (continued). If we take r = s = 1 then we have the eigenvalues λ1 = −1 and λ2 = 8 with corresponding eigenvectors v1 = 1 2

  • and
  • v2 =

−1 1

  • , respectively. So by Theorem 5.2, “Matrix Summary of

Eigenvalues of A,” with C =     . . . . . .

  • v1
  • v2

. . . . . .     = 1 −1 2 1

  • and

D = λ1 λ2

  • =

−1 8

  • we have AC = CD. We find C −1 (by Note

1.5.A, “Computation of Inverses”): [C|I] = 1 −1 1 2 1 1 R2→R2−2R1

  • 1

−1 1 3 −2 1

  • ()

Linear Algebra April 13, 2020 11 / 34

slide-25
SLIDE 25

Example 5.2.A

Example 5.2.A (continued 4)

Solution (continued). 1 −1 1 3 −2 1 R2→R2/3

  • 1

−1 1 1 −2/3 1/3

  • R1→R1+R2
  • 1

1/3 1/3 1 −2/3 1/3

  • ,

so C −1 =

  • 1/3

1/3 −2/3 1/3

  • = 1

3

  • 1

1 −2 1

  • . So A = CDC −1 where

C = 1 −1 2 1

  • , D =

−1 8

  • , and C −1 = 1

3

  • 1

1 −2 1

  • .

() Linear Algebra April 13, 2020 12 / 34

slide-26
SLIDE 26

Example 5.2.A

Example 5.2.A (continued 4)

Solution (continued). 1 −1 1 3 −2 1 R2→R2/3

  • 1

−1 1 1 −2/3 1/3

  • R1→R1+R2
  • 1

1/3 1/3 1 −2/3 1/3

  • ,

so C −1 =

  • 1/3

1/3 −2/3 1/3

  • = 1

3

  • 1

1 −2 1

  • . So A = CDC −1 where

C = 1 −1 2 1

  • , D =

−1 8

  • , and C −1 = 1

3

  • 1

1 −2 1

  • .

() Linear Algebra April 13, 2020 12 / 34

slide-27
SLIDE 27

Example 5.2.A

Example 5.2.A (continued 5)

Solution (continued). As in Corollary 2, Ak = CDkC −1 = 1 −1 2 1 −1 8 k 1 3

  • 1

1 −2 1

  • =

1 3 1 −1 2 1 (−1)k 8k 1 1 −2 1

  • =

1 3

  • (−1)k

−8k 2(−1)k 8k 1 1 −2 1

  • =

1 3

(−1)k + 2(8k) (−1)k − 8k 2(−1)k − 2(8k) 2(−1)k + 8k

  • .
  • ()

Linear Algebra April 13, 2020 13 / 34

slide-28
SLIDE 28

Theorem 5.3. Independence of Eigenvectors

Theorem 5.3

Theorem 5.3. Independence of Eigenvectors. Let A be an n × n matrix. If v1, v2, . . . , vn are eigenvectors of A corresponding to distinct eigenvalues λ1, λ2, . . . , λn, respectively, the set { v1, v2, . . . , vn} is linearly independent and A is diagonalizable.

  • Proof. We prove this by contradiction. Suppose that the conclusion is

false and the hypotheses are true. That is, suppose the eigenvectors

  • v1,

v2, . . . , vn are linearly dependent.

() Linear Algebra April 13, 2020 14 / 34

slide-29
SLIDE 29

Theorem 5.3. Independence of Eigenvectors

Theorem 5.3

Theorem 5.3. Independence of Eigenvectors. Let A be an n × n matrix. If v1, v2, . . . , vn are eigenvectors of A corresponding to distinct eigenvalues λ1, λ2, . . . , λn, respectively, the set { v1, v2, . . . , vn} is linearly independent and A is diagonalizable.

  • Proof. We prove this by contradiction. Suppose that the conclusion is

false and the hypotheses are true. That is, suppose the eigenvectors

  • v1,

v2, . . . , vn are linearly dependent. Then one of them is a linear combination of its predecessors (see page 203 number 37). Let vk be the first such vector, so that

  • vk = d1

v1 + d2 v2 + · · · + dk−1 vk−1 (2) and { v1, v2, . . . , vk−1} is independent.

() Linear Algebra April 13, 2020 14 / 34

slide-30
SLIDE 30

Theorem 5.3. Independence of Eigenvectors

Theorem 5.3

Theorem 5.3. Independence of Eigenvectors. Let A be an n × n matrix. If v1, v2, . . . , vn are eigenvectors of A corresponding to distinct eigenvalues λ1, λ2, . . . , λn, respectively, the set { v1, v2, . . . , vn} is linearly independent and A is diagonalizable.

  • Proof. We prove this by contradiction. Suppose that the conclusion is

false and the hypotheses are true. That is, suppose the eigenvectors

  • v1,

v2, . . . , vn are linearly dependent. Then one of them is a linear combination of its predecessors (see page 203 number 37). Let vk be the first such vector, so that

  • vk = d1

v1 + d2 v2 + · · · + dk−1 vk−1 (2) and { v1, v2, . . . , vk−1} is independent. Multiplying (2) by λk, we obtain λk vk = d1λk v1 + d2λk v2 + · · · + dk−1λk vk−1. (3)

() Linear Algebra April 13, 2020 14 / 34

slide-31
SLIDE 31

Theorem 5.3. Independence of Eigenvectors

Theorem 5.3

Theorem 5.3. Independence of Eigenvectors. Let A be an n × n matrix. If v1, v2, . . . , vn are eigenvectors of A corresponding to distinct eigenvalues λ1, λ2, . . . , λn, respectively, the set { v1, v2, . . . , vn} is linearly independent and A is diagonalizable.

  • Proof. We prove this by contradiction. Suppose that the conclusion is

false and the hypotheses are true. That is, suppose the eigenvectors

  • v1,

v2, . . . , vn are linearly dependent. Then one of them is a linear combination of its predecessors (see page 203 number 37). Let vk be the first such vector, so that

  • vk = d1

v1 + d2 v2 + · · · + dk−1 vk−1 (2) and { v1, v2, . . . , vk−1} is independent. Multiplying (2) by λk, we obtain λk vk = d1λk v1 + d2λk v2 + · · · + dk−1λk vk−1. (3)

() Linear Algebra April 13, 2020 14 / 34

slide-32
SLIDE 32

Theorem 5.3. Independence of Eigenvectors

Theorem 5.3 (continued)

Theorem 5.3. Independence of Eigenvectors. Let A be an n × n matrix. If v1, v2, . . . , vn are eigenvectors of A corresponding to distinct eigenvalues λ1, λ2, . . . , λn, respectively, the set { v1, v2, . . . , vn} is linearly independent and A is diagonalizable. Proof (continued). Also, multiplying (2) on the left by the matrix A yields λk vk = d1λ1 v1 + d2λ2 v2 + · · · + dk−1λk−1 vk−1 (4), since A vi = λi

  • vi. Subtracting (4) from (3), we see that
  • 0 = d1(λk − λ1)

v1 + d2(λk − λ2) v2 + · · · + dk−1(λk − λk−1) vk−1. But this equation is a dependence relation since not all di’s are 0 and the λ’s are hypothesized to be different. This contradicts the linear independence of the set { v1, v2, . . . , vk−1}. This contradiction shows that { v1, v2, . . . , vn} is independent. From Corollary 1 of Theorem 5.2 we see that A is diagonalizable.

() Linear Algebra April 13, 2020 15 / 34

slide-33
SLIDE 33

Theorem 5.3. Independence of Eigenvectors

Theorem 5.3 (continued)

Theorem 5.3. Independence of Eigenvectors. Let A be an n × n matrix. If v1, v2, . . . , vn are eigenvectors of A corresponding to distinct eigenvalues λ1, λ2, . . . , λn, respectively, the set { v1, v2, . . . , vn} is linearly independent and A is diagonalizable. Proof (continued). Also, multiplying (2) on the left by the matrix A yields λk vk = d1λ1 v1 + d2λ2 v2 + · · · + dk−1λk−1 vk−1 (4), since A vi = λi

  • vi. Subtracting (4) from (3), we see that
  • 0 = d1(λk − λ1)

v1 + d2(λk − λ2) v2 + · · · + dk−1(λk − λk−1) vk−1. But this equation is a dependence relation since not all di’s are 0 and the λ’s are hypothesized to be different. This contradicts the linear independence of the set { v1, v2, . . . , vk−1}. This contradiction shows that { v1, v2, . . . , vn} is independent. From Corollary 1 of Theorem 5.2 we see that A is diagonalizable.

() Linear Algebra April 13, 2020 15 / 34

slide-34
SLIDE 34

Page 315 Number 6

Page 315 Number 6

Page 315 Number 6. Find the eigenvalues λi and corresponding eigenvectors vi of A =   −3 5 −20 2 8 2 1 7  . Find an invertible matrix C and a diagonal matrix D such that D = C −1AC.

  • Solution. We show all computations and details, so this will take a
  • while. . .

() Linear Algebra April 13, 2020 16 / 34

slide-35
SLIDE 35

Page 315 Number 6

Page 315 Number 6

Page 315 Number 6. Find the eigenvalues λi and corresponding eigenvectors vi of A =   −3 5 −20 2 8 2 1 7  . Find an invertible matrix C and a diagonal matrix D such that D = C −1AC.

  • Solution. We show all computations and details, so this will take a
  • while. . .

We have A − λI =   −3 5 −20 2 8 2 1 7   − λ   1 1 1   =   −3 − λ 5 −20 2 0 − λ 8 2 1 7 − λ   .

() Linear Algebra April 13, 2020 16 / 34

slide-36
SLIDE 36

Page 315 Number 6

Page 315 Number 6

Page 315 Number 6. Find the eigenvalues λi and corresponding eigenvectors vi of A =   −3 5 −20 2 8 2 1 7  . Find an invertible matrix C and a diagonal matrix D such that D = C −1AC.

  • Solution. We show all computations and details, so this will take a
  • while. . .

We have A − λI =   −3 5 −20 2 8 2 1 7   − λ   1 1 1   =   −3 − λ 5 −20 2 0 − λ 8 2 1 7 − λ   .

() Linear Algebra April 13, 2020 16 / 34

slide-37
SLIDE 37

Page 315 Number 6

Page 315 Number 6 (continued 1)

Solution (continued). So the characteristic polynomial is p(λ) = det(A − λI) =

  • −3 − λ

5 −20 2 0 − λ 8 2 1 7 − λ

  • = (−3 − λ)
  • −λ

8 1 7 − λ

  • − (5)
  • 2

8 2 7 − λ

  • + (−20)
  • 2

−λ 2 1

  • = (−3 − λ) ((−λ)(7 − λ) − (8)(1)) − 5 ((2)(7 − λ) − (8)(2))

−20 ((2)(1) − (−λ)(2)) = (−3−λ)(λ2 −7λ−8)−5(−2λ−2)−20(2λ+2)

() Linear Algebra April 13, 2020 17 / 34

slide-38
SLIDE 38

Page 315 Number 6

Page 315 Number 6 (continued 1)

Solution (continued). So the characteristic polynomial is p(λ) = det(A − λI) =

  • −3 − λ

5 −20 2 0 − λ 8 2 1 7 − λ

  • = (−3 − λ)
  • −λ

8 1 7 − λ

  • − (5)
  • 2

8 2 7 − λ

  • + (−20)
  • 2

−λ 2 1

  • = (−3 − λ) ((−λ)(7 − λ) − (8)(1)) − 5 ((2)(7 − λ) − (8)(2))

−20 ((2)(1) − (−λ)(2)) = (−3−λ)(λ2 −7λ−8)−5(−2λ−2)−20(2λ+2) = (−3 − λ)(λ − 8)(λ + 1) + 10(λ + 1) − 40(λ + 1) = (λ+1) ((−3 − λ)(λ − 8) + 10 − 40) = (λ+1)(−3λ+24−λ2+8λ+10−40) = (λ+1)(−λ2+5λ−6) = −(λ+1)(λ2−5λ+6) = −(λ+1)(λ−2)(λ−3).

() Linear Algebra April 13, 2020 17 / 34

slide-39
SLIDE 39

Page 315 Number 6

Page 315 Number 6 (continued 1)

Solution (continued). So the characteristic polynomial is p(λ) = det(A − λI) =

  • −3 − λ

5 −20 2 0 − λ 8 2 1 7 − λ

  • = (−3 − λ)
  • −λ

8 1 7 − λ

  • − (5)
  • 2

8 2 7 − λ

  • + (−20)
  • 2

−λ 2 1

  • = (−3 − λ) ((−λ)(7 − λ) − (8)(1)) − 5 ((2)(7 − λ) − (8)(2))

−20 ((2)(1) − (−λ)(2)) = (−3−λ)(λ2 −7λ−8)−5(−2λ−2)−20(2λ+2) = (−3 − λ)(λ − 8)(λ + 1) + 10(λ + 1) − 40(λ + 1) = (λ+1) ((−3 − λ)(λ − 8) + 10 − 40) = (λ+1)(−3λ+24−λ2+8λ+10−40) = (λ+1)(−λ2+5λ−6) = −(λ+1)(λ2−5λ+6) = −(λ+1)(λ−2)(λ−3).

() Linear Algebra April 13, 2020 17 / 34

slide-40
SLIDE 40

Page 315 Number 6

Page 315 Number 6 (continued 2)

Solution (continued). Since p(λ) = −(λ + 1)(λ − 2)(λ − 3), then the eigenvalues are λ1 = −1, λ2 = 2, λ3 = 3. To find the eigenvectors corresponding to each eigenvalue we consider the formula A v = λ v or (A − λI) v = 0. λ1 = −1. With v1 = [v1, v2, v3] an eigenvector corresponding to eigenvalue λ1 = −1 we need (A − λI) v1 = 0.

() Linear Algebra April 13, 2020 18 / 34

slide-41
SLIDE 41

Page 315 Number 6

Page 315 Number 6 (continued 2)

Solution (continued). Since p(λ) = −(λ + 1)(λ − 2)(λ − 3), then the eigenvalues are λ1 = −1, λ2 = 2, λ3 = 3. To find the eigenvectors corresponding to each eigenvalue we consider the formula A v = λ v or (A − λI) v = 0. λ1 = −1. With v1 = [v1, v2, v3] an eigenvector corresponding to eigenvalue λ1 = −1 we need (A − λI) v1 =

  • 0. So we consider the augmented matrix

  −3 − (−1) 5 −20 2 0 − (−1) 8 2 1 7 − (−1)   =   −2 5 −20 2 1 8 2 1 8  

R3→R3−R2

 −2 5 −20 2 1 8  

R2→R2+R1

 −2 5 −20 6 −12  

() Linear Algebra April 13, 2020 18 / 34

slide-42
SLIDE 42

Page 315 Number 6

Page 315 Number 6 (continued 2)

Solution (continued). Since p(λ) = −(λ + 1)(λ − 2)(λ − 3), then the eigenvalues are λ1 = −1, λ2 = 2, λ3 = 3. To find the eigenvectors corresponding to each eigenvalue we consider the formula A v = λ v or (A − λI) v = 0. λ1 = −1. With v1 = [v1, v2, v3] an eigenvector corresponding to eigenvalue λ1 = −1 we need (A − λI) v1 =

  • 0. So we consider the augmented matrix

  −3 − (−1) 5 −20 2 0 − (−1) 8 2 1 7 − (−1)   =   −2 5 −20 2 1 8 2 1 8  

R3→R3−R2

 −2 5 −20 2 1 8  

R2→R2+R1

 −2 5 −20 6 −12  

() Linear Algebra April 13, 2020 18 / 34

slide-43
SLIDE 43

Page 315 Number 6

Page 315 Number 6 (continued 3)

Solution (continued).   −2 5 −20 6 −12  

R2→R2/6

 −2 5 −20 1 −2  

R1→R1−5R2

 −2 −10 1 −2  

R1→R1/(−2)

 1 5 1 −2   . So we need v1 + 5v3 = v2 − 2v3 = =

  • r

v1 = −5v3 v2 = 2v3 v3 = v3

  • r, with r = v3

as a free variable, v1 = −5r v2 = 2r v3 = r .

() Linear Algebra April 13, 2020 19 / 34

slide-44
SLIDE 44

Page 315 Number 6

Page 315 Number 6 (continued 3)

Solution (continued).   −2 5 −20 6 −12  

R2→R2/6

 −2 5 −20 1 −2  

R1→R1−5R2

 −2 −10 1 −2  

R1→R1/(−2)

 1 5 1 −2   . So we need v1 + 5v3 = v2 − 2v3 = =

  • r

v1 = −5v3 v2 = 2v3 v3 = v3

  • r, with r = v3

as a free variable, v1 = −5r v2 = 2r v3 = r .

() Linear Algebra April 13, 2020 19 / 34

slide-45
SLIDE 45

Page 315 Number 6

Page 315 Number 6 (continued 4)

Solution (continued). So the collection of all eigenvectors of λ1 = −1 is

  • v1 = r

  −5 2 1   where r ∈ R, r = 0. λ2 = 2. As above, we consider (A − 2I) v2 = 0 and consider the augmented matrix   −3 − (2) 5 −20 2 0 − (2) 8 2 1 7 − (2)   =   −5 5 −20 2 −2 8 2 1 5  

R1→R1/(−5)

  • R2 → R2/2

  1 −1 4 1 −1 4 2 1 5  

R2→R2−R1

  • R3 → R3 − 2R1

  1 −1 4 3 −3  

() Linear Algebra April 13, 2020 20 / 34

slide-46
SLIDE 46

Page 315 Number 6

Page 315 Number 6 (continued 4)

Solution (continued). So the collection of all eigenvectors of λ1 = −1 is

  • v1 = r

  −5 2 1   where r ∈ R, r = 0. λ2 = 2. As above, we consider (A − 2I) v2 = 0 and consider the augmented matrix   −3 − (2) 5 −20 2 0 − (2) 8 2 1 7 − (2)   =   −5 5 −20 2 −2 8 2 1 5  

R1→R1/(−5)

  • R2 → R2/2

  1 −1 4 1 −1 4 2 1 5  

R2→R2−R1

  • R3 → R3 − 2R1

  1 −1 4 3 −3  

() Linear Algebra April 13, 2020 20 / 34

slide-47
SLIDE 47

Page 315 Number 6

Page 315 Number 6 (continued 5)

Solution (continued).   1 −1 4 3 −3  

R3→R3/3

 1 −1 4 1 −1  

R2↔R3

 1 −1 4 1 −1  

R1→R1+R2

 1 3 1 −1   . So we need v1 + 3v3 = v2 − v3 = =

  • r

v1 = −3v3 v2 = v3 v3 = v3

  • r, with s = v3

as a free variable, v1 = −3s v2 = s v3 = s .

() Linear Algebra April 13, 2020 21 / 34

slide-48
SLIDE 48

Page 315 Number 6

Page 315 Number 6 (continued 5)

Solution (continued).   1 −1 4 3 −3  

R3→R3/3

 1 −1 4 1 −1  

R2↔R3

 1 −1 4 1 −1  

R1→R1+R2

 1 3 1 −1   . So we need v1 + 3v3 = v2 − v3 = =

  • r

v1 = −3v3 v2 = v3 v3 = v3

  • r, with s = v3

as a free variable, v1 = −3s v2 = s v3 = s .

() Linear Algebra April 13, 2020 21 / 34

slide-49
SLIDE 49

Page 315 Number 6

Page 315 Number 6 (continued 6)

Solution (continued). So the collection of all eigenvectors of λ2 = 2 is

  • v2 = s

  −3 1 1   where s ∈ R, s = 0. λ3 = 3. As above, we consider (A − 3I) v3 = 0 and consider the augmented matrix   −3 − (3) 5 −20 2 0 − (3) 8 2 1 7 − (3)   =   −6 5 −20 2 −3 8 2 1 4  

R1↔R3

 2 1 4 2 −3 8 −6 5 −20  

R2→R2−R1

  • R3 → R3 + 3R1

  2 1 4 −4 4 8 −8  

() Linear Algebra April 13, 2020 22 / 34

slide-50
SLIDE 50

Page 315 Number 6

Page 315 Number 6 (continued 6)

Solution (continued). So the collection of all eigenvectors of λ2 = 2 is

  • v2 = s

  −3 1 1   where s ∈ R, s = 0. λ3 = 3. As above, we consider (A − 3I) v3 = 0 and consider the augmented matrix   −3 − (3) 5 −20 2 0 − (3) 8 2 1 7 − (3)   =   −6 5 −20 2 −3 8 2 1 4  

R1↔R3

 2 1 4 2 −3 8 −6 5 −20  

R2→R2−R1

  • R3 → R3 + 3R1

  2 1 4 −4 4 8 −8  

() Linear Algebra April 13, 2020 22 / 34

slide-51
SLIDE 51

Page 315 Number 6

Page 315 Number 6 (continued 7)

Solution (continued).   2 1 4 −4 4 8 −8  

R2→R2/(−4)

  • R3 → R3/8

  2 1 4 1 −1 1 −1  

R1→R1−R2

  • R3 → R3 − R2

  2 5 1 −1  

R1↔R1/2

 1 5/2 1 −1   . So we need v1 + (5/2)v3 = v2 − v3 = =

  • r

v1 = −(5/2)v3 v2 = v3 v3 = v3

  • r, with

t = v3/2 as a free variable, v1 = −5t v2 = 2t v3 = 2t .

() Linear Algebra April 13, 2020 23 / 34

slide-52
SLIDE 52

Page 315 Number 6

Page 315 Number 6 (continued 7)

Solution (continued).   2 1 4 −4 4 8 −8  

R2→R2/(−4)

  • R3 → R3/8

  2 1 4 1 −1 1 −1  

R1→R1−R2

  • R3 → R3 − R2

  2 5 1 −1  

R1↔R1/2

 1 5/2 1 −1   . So we need v1 + (5/2)v3 = v2 − v3 = =

  • r

v1 = −(5/2)v3 v2 = v3 v3 = v3

  • r, with

t = v3/2 as a free variable, v1 = −5t v2 = 2t v3 = 2t .

() Linear Algebra April 13, 2020 23 / 34

slide-53
SLIDE 53

Page 315 Number 6

Page 315 Number 6 (continued 8)

Solution (continued). So the collection of all eigenvectors of λ3 = 3 is

  • v3 = t

  −5 2 2   where t ∈ R, t = 0. If we take r = s = t = 1 then we have the eigenvalues λ1 = −1, λ2 = 2, and λ3 = 3 with corresponding eigenvalues v1 =   −5 2 1  , v2 =   −3 1 1  , and v3 =   −5 2 2  , respectively.

() Linear Algebra April 13, 2020 24 / 34

slide-54
SLIDE 54

Page 315 Number 6

Page 315 Number 6 (continued 8)

Solution (continued). So the collection of all eigenvectors of λ3 = 3 is

  • v3 = t

  −5 2 2   where t ∈ R, t = 0. If we take r = s = t = 1 then we have the eigenvalues λ1 = −1, λ2 = 2, and λ3 = 3 with corresponding eigenvalues v1 =   −5 2 1  , v2 =   −3 1 1  , and v3 =   −5 2 2  , respectively. So by Theorem 5.2, “Matrix Summary of Eigenvalues of A,” with C =     . . . . . . . . .

  • v1
  • v2
  • v3

. . . . . . . . .     =   −5 −3 −5 2 1 2 1 1 2  , . . .

() Linear Algebra April 13, 2020 24 / 34

slide-55
SLIDE 55

Page 315 Number 6

Page 315 Number 6 (continued 8)

Solution (continued). So the collection of all eigenvectors of λ3 = 3 is

  • v3 = t

  −5 2 2   where t ∈ R, t = 0. If we take r = s = t = 1 then we have the eigenvalues λ1 = −1, λ2 = 2, and λ3 = 3 with corresponding eigenvalues v1 =   −5 2 1  , v2 =   −3 1 1  , and v3 =   −5 2 2  , respectively. So by Theorem 5.2, “Matrix Summary of Eigenvalues of A,” with C =     . . . . . . . . .

  • v1
  • v2
  • v3

. . . . . . . . .     =   −5 −3 −5 2 1 2 1 1 2  , . . .

() Linear Algebra April 13, 2020 24 / 34

slide-56
SLIDE 56

Page 315 Number 6

Page 315 Number 6 (continued 9)

Page 315 Number 6. Find the eigenvalues λi and corresponding eigenvectors vi of A =   −3 5 −20 2 8 2 1 7  . Find an invertible matrix C and a diagonal matrix D such that D = C −1AC. Solution (continued). . . . and D =   λ1 λ2 λ3   =   −1 2 3   we have AC = CD. By Theorem 5.3, “Independence of Eigenvalues,” we have that v1, v2, v3 are linearly independent vectors and A is diagonalizable (notice that C is invertible by Theorem 1.16, “The Square Case, m = n”). So D = C −1AC where C =   −5 −3 −5 2 1 2 1 1 2   and D =   −1 2 3  .

  • ()

Linear Algebra April 13, 2020 25 / 34

slide-57
SLIDE 57

Page 315 Number 6

Page 315 Number 6 (continued 9)

Page 315 Number 6. Find the eigenvalues λi and corresponding eigenvectors vi of A =   −3 5 −20 2 8 2 1 7  . Find an invertible matrix C and a diagonal matrix D such that D = C −1AC. Solution (continued). . . . and D =   λ1 λ2 λ3   =   −1 2 3   we have AC = CD. By Theorem 5.3, “Independence of Eigenvalues,” we have that v1, v2, v3 are linearly independent vectors and A is diagonalizable (notice that C is invertible by Theorem 1.16, “The Square Case, m = n”). So D = C −1AC where C =   −5 −3 −5 2 1 2 1 1 2   and D =   −1 2 3  .

  • ()

Linear Algebra April 13, 2020 25 / 34

slide-58
SLIDE 58

Page 315 Number 18

Page 315 Number 18

Page 315 Number 18. Prove that similar square matrices have the same eigenvalues with the same algebraic multiplicities.

  • Proof. (This is repetitious with Exercise 5.1.38.) Notice that

C −1AC − λI = C −1AC − λC −1C = C −1AC − C −1(λC) by Theorem 1.3.A(7), “Scalars Pull Through” = C −1(AC − λC) by Theorem 1.3.A(10), “Distribution Law of Matrix Multiplication” = C −1(A − λI)C by Theorem 1.3.A(10).

() Linear Algebra April 13, 2020 26 / 34

slide-59
SLIDE 59

Page 315 Number 18

Page 315 Number 18

Page 315 Number 18. Prove that similar square matrices have the same eigenvalues with the same algebraic multiplicities.

  • Proof. (This is repetitious with Exercise 5.1.38.) Notice that

C −1AC − λI = C −1AC − λC −1C = C −1AC − C −1(λC) by Theorem 1.3.A(7), “Scalars Pull Through” = C −1(AC − λC) by Theorem 1.3.A(10), “Distribution Law of Matrix Multiplication” = C −1(A − λI)C by Theorem 1.3.A(10).

() Linear Algebra April 13, 2020 26 / 34

slide-60
SLIDE 60

Page 315 Number 18

Page 315 Number 18 (continued)

Proof (continued). Recall that det(C −1) = 1/det(C) by Exercise 4.2.31. So the characteristic polynomial for C −1AC is det(C −1AC − λI) = det(C −1(A − λI)C) as just shown = det(C −1)det(A − λI)det(C) by Theorem 4.4, “The Multiplicative Property” = (1/det(C))det(A − λI)det(C) = det(A − λI).

() Linear Algebra April 13, 2020 27 / 34

slide-61
SLIDE 61

Page 315 Number 18

Page 315 Number 18 (continued)

Proof (continued). Recall that det(C −1) = 1/det(C) by Exercise 4.2.31. So the characteristic polynomial for C −1AC is det(C −1AC − λI) = det(C −1(A − λI)C) as just shown = det(C −1)det(A − λI)det(C) by Theorem 4.4, “The Multiplicative Property” = (1/det(C))det(A − λI)det(C) = det(A − λI). Now det(A − λI) is the characteristic polynomial of A, so A and C −1AC have the same characteristic polynomials. So these polynomials have the same roots with the same multiplicities (of course) and since the eigenvalues of a matrix are the roots of the characteristic polynomial, then A and C −1AC have the same eigenvalues with the same algebraic mulitplicities, as claimed.

() Linear Algebra April 13, 2020 27 / 34

slide-62
SLIDE 62

Page 315 Number 18

Page 315 Number 18 (continued)

Proof (continued). Recall that det(C −1) = 1/det(C) by Exercise 4.2.31. So the characteristic polynomial for C −1AC is det(C −1AC − λI) = det(C −1(A − λI)C) as just shown = det(C −1)det(A − λI)det(C) by Theorem 4.4, “The Multiplicative Property” = (1/det(C))det(A − λI)det(C) = det(A − λI). Now det(A − λI) is the characteristic polynomial of A, so A and C −1AC have the same characteristic polynomials. So these polynomials have the same roots with the same multiplicities (of course) and since the eigenvalues of a matrix are the roots of the characteristic polynomial, then A and C −1AC have the same eigenvalues with the same algebraic mulitplicities, as claimed.

() Linear Algebra April 13, 2020 27 / 34

slide-63
SLIDE 63

Page 315 Number 10

Page 315 Number 10

Page 315 Number 10. Determine whether A =   3 1 3 1 3   is diagonalizable.

  • Solution. First, we find the eigenvalues of A. Notice that

A − λI =   3 − λ 1 3 − λ 1 3 − λ   is upper triangular, so by Example 4.2.4, p(λ) = det(A − λI) = (3 − λ)3. So λ = 3 is the only eigenvalue of A and it is of algebraic multiplicity 3.

() Linear Algebra April 13, 2020 28 / 34

slide-64
SLIDE 64

Page 315 Number 10

Page 315 Number 10

Page 315 Number 10. Determine whether A =   3 1 3 1 3   is diagonalizable.

  • Solution. First, we find the eigenvalues of A. Notice that

A − λI =   3 − λ 1 3 − λ 1 3 − λ   is upper triangular, so by Example 4.2.4, p(λ) = det(A − λI) = (3 − λ)3. So λ = 3 is the only eigenvalue of A and it is of algebraic multiplicity 3. With v = [v1, v2, v3]T as an eigenvector corresponding to eigenvalue λ = 3, we need (A − λI) v = 0. So we consider the augmented matrix:   3 − (3) 1 3 − (3) 1 3 − (3)   =   1 1   .

() Linear Algebra April 13, 2020 28 / 34

slide-65
SLIDE 65

Page 315 Number 10

Page 315 Number 10

Page 315 Number 10. Determine whether A =   3 1 3 1 3   is diagonalizable.

  • Solution. First, we find the eigenvalues of A. Notice that

A − λI =   3 − λ 1 3 − λ 1 3 − λ   is upper triangular, so by Example 4.2.4, p(λ) = det(A − λI) = (3 − λ)3. So λ = 3 is the only eigenvalue of A and it is of algebraic multiplicity 3. With v = [v1, v2, v3]T as an eigenvector corresponding to eigenvalue λ = 3, we need (A − λI) v = 0. So we consider the augmented matrix:   3 − (3) 1 3 − (3) 1 3 − (3)   =   1 1   .

() Linear Algebra April 13, 2020 28 / 34

slide-66
SLIDE 66

Page 315 Number 10

Page 315 Number 10 (continued)

Page 315 Number 10. Determine whether A =   3 1 3 1 3   is diagonalizable. Solution (continued). So we need v2 = v3 = =

  • r

v1 = v1 v2 = v3 =

  • r,

with r = v1 as a free variable, v1 = r v2 = v3 = . So the collection of all eigenvectors of λ = 3 is v = r   1   where r ∈ R, r = 0. But then there can be only one vector in a set of linearly independent eigenvectors. That is, the dimension of Eλ is 1 and so λ = 3 is of geometric multiplicity 1. So, by Theorem 5.4, “A Criterion for Diagonalization,” A is not diagonalizable.

() Linear Algebra April 13, 2020 29 / 34

slide-67
SLIDE 67

Page 315 Number 10

Page 315 Number 10 (continued)

Page 315 Number 10. Determine whether A =   3 1 3 1 3   is diagonalizable. Solution (continued). So we need v2 = v3 = =

  • r

v1 = v1 v2 = v3 =

  • r,

with r = v1 as a free variable, v1 = r v2 = v3 = . So the collection of all eigenvectors of λ = 3 is v = r   1   where r ∈ R, r = 0. But then there can be only one vector in a set of linearly independent eigenvectors. That is, the dimension of Eλ is 1 and so λ = 3 is of geometric multiplicity 1. So, by Theorem 5.4, “A Criterion for Diagonalization,” A is not diagonalizable.

() Linear Algebra April 13, 2020 29 / 34

slide-68
SLIDE 68

Page 316 Number 22

Page 316 Number 22

Page 316 Number 22. Let A and C be n × n matrices, and let C be

  • invertible. Prove that, if

v is an eigenvector of A with corresponding eigenvalue λ, then C −1 v is an eigenvector of C −1AC with corresponding eigenvalue λ. Prove that all eigenvectors of C −1AC are of the form C −1 v, where v is an eigenvector of A.

  • Proof. If

v is an eigenvector of A corresponding to eigenvalue λ then A v = λ v.

() Linear Algebra April 13, 2020 30 / 34

slide-69
SLIDE 69

Page 316 Number 22

Page 316 Number 22

Page 316 Number 22. Let A and C be n × n matrices, and let C be

  • invertible. Prove that, if

v is an eigenvector of A with corresponding eigenvalue λ, then C −1 v is an eigenvector of C −1AC with corresponding eigenvalue λ. Prove that all eigenvectors of C −1AC are of the form C −1 v, where v is an eigenvector of A.

  • Proof. If

v is an eigenvector of A corresponding to eigenvalue λ then A v = λ

  • v. So

(C −1AC)(C −1 v) = C −1A(CC −1) v by Theorem 1.3.A(8), “Associativity of Matrix Multiplication” = C −1AI v = C −1A v = C −1(λ v) = λ(C −1 v) by Theorem 1.3.A(7), “Scalars Pull Through” and so λ is an eigenvalue of C −1AC with corresponding eigenvector C −1 v, as claimed.

() Linear Algebra April 13, 2020 30 / 34

slide-70
SLIDE 70

Page 316 Number 22

Page 316 Number 22

Page 316 Number 22. Let A and C be n × n matrices, and let C be

  • invertible. Prove that, if

v is an eigenvector of A with corresponding eigenvalue λ, then C −1 v is an eigenvector of C −1AC with corresponding eigenvalue λ. Prove that all eigenvectors of C −1AC are of the form C −1 v, where v is an eigenvector of A.

  • Proof. If

v is an eigenvector of A corresponding to eigenvalue λ then A v = λ

  • v. So

(C −1AC)(C −1 v) = C −1A(CC −1) v by Theorem 1.3.A(8), “Associativity of Matrix Multiplication” = C −1AI v = C −1A v = C −1(λ v) = λ(C −1 v) by Theorem 1.3.A(7), “Scalars Pull Through” and so λ is an eigenvalue of C −1AC with corresponding eigenvector C −1 v, as claimed.

() Linear Algebra April 13, 2020 30 / 34

slide-71
SLIDE 71

Page 316 Number 22

Page 316 Number 22 (continued)

Page 316 Number 22. Let A and C be n × n matrices, and let C be

  • invertible. Prove that, if

v is an eigenvector of A with corresponding eigenvalue λ, then C −1 v is an eigenvector of C −1AC with corresponding eigenvalue λ. Prove that all eigenvectors of C −1AC are of the form C −1 v, where v is an eigenvector of A. Proof (continued). Now suppose w is an eigenvector of C −1AC. Then C −1AC w = λ w for some λ ∈ R. Then C(C −1AC w) = Cλ w or (CC −1)AC w = λC w or A(C w) = λ(C w). So v = C w is an eigenvector

  • f A with corresponding eigenvalue λ. Then

w = C −1 v and so all eigenvectors of C −1AC are of the form C −1 v where v is an eigenvector of A, as claimed.

() Linear Algebra April 13, 2020 31 / 34

slide-72
SLIDE 72

Page 316 Number 22

Page 316 Number 22 (continued)

Page 316 Number 22. Let A and C be n × n matrices, and let C be

  • invertible. Prove that, if

v is an eigenvector of A with corresponding eigenvalue λ, then C −1 v is an eigenvector of C −1AC with corresponding eigenvalue λ. Prove that all eigenvectors of C −1AC are of the form C −1 v, where v is an eigenvector of A. Proof (continued). Now suppose w is an eigenvector of C −1AC. Then C −1AC w = λ w for some λ ∈ R. Then C(C −1AC w) = Cλ w or (CC −1)AC w = λC w or A(C w) = λ(C w). So v = C w is an eigenvector

  • f A with corresponding eigenvalue λ. Then

w = C −1 v and so all eigenvectors of C −1AC are of the form C −1 v where v is an eigenvector of A, as claimed.

() Linear Algebra April 13, 2020 31 / 34

slide-73
SLIDE 73

Page 316 Number 24

Page 316 Number 24

Page 316 Number 24. Prove that if λ1, λ2, . . . , λk are distinct real eigenvalues of an n × n real matrix A and if Bi is a basis for the eigenspace Eλi, then the union of the bases Bi is an independent set of vectors in Rn.

  • Proof. Let Bi = {

bi

1,

bi

2, . . . ,

bi

ni} for i = 1, 2, . . . , k, where ni is the

dimension of Eλi. Suppose a1

1

b1

1 + a1 2

b1

2 + · · · + a1 n1

b1

n1 + a2 1

b2

1 + a2 2

b2

2 + · · · + a2 n2

b2

n2 + · · ·

+ak

1

bk

1 + ak 2

bk

2 + · · · + ak nk

bk

nk =

0. (∗)

() Linear Algebra April 13, 2020 32 / 34

slide-74
SLIDE 74

Page 316 Number 24

Page 316 Number 24

Page 316 Number 24. Prove that if λ1, λ2, . . . , λk are distinct real eigenvalues of an n × n real matrix A and if Bi is a basis for the eigenspace Eλi, then the union of the bases Bi is an independent set of vectors in Rn.

  • Proof. Let Bi = {

bi

1,

bi

2, . . . ,

bi

ni} for i = 1, 2, . . . , k, where ni is the

dimension of Eλi. Suppose a1

1

b1

1 + a1 2

b1

2 + · · · + a1 n1

b1

n1 + a2 1

b2

1 + a2 2

b2

2 + · · · + a2 n2

b2

n2 + · · ·

+ak

1

bk

1 + ak 2

bk

2 + · · · + ak nk

bk

nk =

0. (∗) If we let wi = ai

1

bi

1 + ai 2

bi

2 + · · · + ai n1

bi

n1 then (∗) gives

  • w1 +

w2 + · · · + wk =

  • 0. But

w1, w2, . . . , wk are linearly independent by Theorem 5.3, “Independence of Eigenvectors.” This implies that each wi must in fact be the zero vector, wi = 0 (or else some nonzero wi is a linear combination of the other wi’s, contradicting the linear independence).

() Linear Algebra April 13, 2020 32 / 34

slide-75
SLIDE 75

Page 316 Number 24

Page 316 Number 24

Page 316 Number 24. Prove that if λ1, λ2, . . . , λk are distinct real eigenvalues of an n × n real matrix A and if Bi is a basis for the eigenspace Eλi, then the union of the bases Bi is an independent set of vectors in Rn.

  • Proof. Let Bi = {

bi

1,

bi

2, . . . ,

bi

ni} for i = 1, 2, . . . , k, where ni is the

dimension of Eλi. Suppose a1

1

b1

1 + a1 2

b1

2 + · · · + a1 n1

b1

n1 + a2 1

b2

1 + a2 2

b2

2 + · · · + a2 n2

b2

n2 + · · ·

+ak

1

bk

1 + ak 2

bk

2 + · · · + ak nk

bk

nk =

0. (∗) If we let wi = ai

1

bi

1 + ai 2

bi

2 + · · · + ai n1

bi

n1 then (∗) gives

  • w1 +

w2 + · · · + wk =

  • 0. But

w1, w2, . . . , wk are linearly independent by Theorem 5.3, “Independence of Eigenvectors.” This implies that each wi must in fact be the zero vector, wi = 0 (or else some nonzero wi is a linear combination of the other wi’s, contradicting the linear independence).

() Linear Algebra April 13, 2020 32 / 34

slide-76
SLIDE 76

Page 316 Number 24

Page 316 Number 24 (continued)

Page 316 Number 24. Prove that if λ1, λ2, . . . , λk are distinct real eigenvalues of an n × n real matrix A and if Bi is a basis for the eigenspace Eλi, then the union of the bases Bi is an independent set of vectors in Rn. Proof (continued). But then wi = ai

1

bi

1 + ai 2

bi

2 + · · · + ai n1

bi

n1 =

0 and since the bi

j’s are a basis for Eλi then the

bi

j are linear independent for a

given i and so each ai

j = 0 for given i. Since this holds for all

i = 1, 2, . . . , k then all ai

j = 0 and so we see from (∗) that

{ b1

1,

b1

2, . . . ,

bk

nk} = ∪k i=1Bi is a linearly independent set.

() Linear Algebra April 13, 2020 33 / 34

slide-77
SLIDE 77

Page 316 Number 26

Page 316 Number 26

Page 316 Number 26. Prove that the set {eλ1x, eλ2x, . . . , eλkx}, where the λi are distinct, is independent in the vector space D∞ of all functions mapping R into R and having derivatives of all orders (see Note 3.2.A).

  • Proof. We know that differentiation D is a linear transformation mapping

D∞ into D∞ (see Example 3.4.1). Now D(eλix) = d

dx [eλix] = λieλix, so

eλix is an eigenvector of D with corresponding eigenvalue λi.

() Linear Algebra April 13, 2020 34 / 34

slide-78
SLIDE 78

Page 316 Number 26

Page 316 Number 26

Page 316 Number 26. Prove that the set {eλ1x, eλ2x, . . . , eλkx}, where the λi are distinct, is independent in the vector space D∞ of all functions mapping R into R and having derivatives of all orders (see Note 3.2.A).

  • Proof. We know that differentiation D is a linear transformation mapping

D∞ into D∞ (see Example 3.4.1). Now D(eλix) = d

dx [eλix] = λieλix, so

eλix is an eigenvector of D with corresponding eigenvalue λi. Since λ1, λ2, . . . , λk are distinct by hypothesis, then by Theorem 5.3, “Independence of Eigenvectors,” the set of eigenvectors {eλ1x, eλ2x, . . . , eλkx} are linearly independent, as claimed.

() Linear Algebra April 13, 2020 34 / 34

slide-79
SLIDE 79

Page 316 Number 26

Page 316 Number 26

Page 316 Number 26. Prove that the set {eλ1x, eλ2x, . . . , eλkx}, where the λi are distinct, is independent in the vector space D∞ of all functions mapping R into R and having derivatives of all orders (see Note 3.2.A).

  • Proof. We know that differentiation D is a linear transformation mapping

D∞ into D∞ (see Example 3.4.1). Now D(eλix) = d

dx [eλix] = λieλix, so

eλix is an eigenvector of D with corresponding eigenvalue λi. Since λ1, λ2, . . . , λk are distinct by hypothesis, then by Theorem 5.3, “Independence of Eigenvectors,” the set of eigenvectors {eλ1x, eλ2x, . . . , eλkx} are linearly independent, as claimed.

() Linear Algebra April 13, 2020 34 / 34