Linear algebra and differential equations (Math 54): Lecture 3 - - PowerPoint PPT Presentation

linear algebra and differential equations math 54 lecture
SMART_READER_LITE
LIVE PREVIEW

Linear algebra and differential equations (Math 54): Lecture 3 - - PowerPoint PPT Presentation

Linear algebra and differential equations (Math 54): Lecture 3 Vivek Shende January 30, 2019 Hello and welcome to class! Hello and welcome to class! Last time Hello and welcome to class! Last time We learned more about row reduction, Hello


slide-1
SLIDE 1

Linear algebra and differential equations (Math 54): Lecture 3

Vivek Shende January 30, 2019

slide-2
SLIDE 2

Hello and welcome to class!

slide-3
SLIDE 3

Hello and welcome to class!

Last time

slide-4
SLIDE 4

Hello and welcome to class!

Last time

We learned more about row reduction,

slide-5
SLIDE 5

Hello and welcome to class!

Last time

We learned more about row reduction, and interpreted linear equations in terms of the linear span of vectors.

slide-6
SLIDE 6

Hello and welcome to class!

Last time

We learned more about row reduction, and interpreted linear equations in terms of the linear span of vectors.

Today

I’ll talk about the matrix-vector product,

slide-7
SLIDE 7

Hello and welcome to class!

Last time

We learned more about row reduction, and interpreted linear equations in terms of the linear span of vectors.

Today

I’ll talk about the matrix-vector product, and how linear equations can be formulated in these terms.

slide-8
SLIDE 8

Hello and welcome to class!

Last time

We learned more about row reduction, and interpreted linear equations in terms of the linear span of vectors.

Today

I’ll talk about the matrix-vector product, and how linear equations can be formulated in these terms. I’ll then explain more about the algebraic and geometric description of solution sets to linear equations.

slide-9
SLIDE 9

Hello and welcome to class!

Last time

We learned more about row reduction, and interpreted linear equations in terms of the linear span of vectors.

Today

I’ll talk about the matrix-vector product, and how linear equations can be formulated in these terms. I’ll then explain more about the algebraic and geometric description of solution sets to linear equations. Finally, I’ll start explaining the fundamental notions of linear dependence and independence.

slide-10
SLIDE 10

A parable

slide-11
SLIDE 11

Systems of linear equations

slide-12
SLIDE 12

Systems of linear equations

3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3

slide-13
SLIDE 13

Systems of linear equations

3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3 can be abbreviated via an augmented matrix   3 5 4 3 2 2 1 1 1 6 1 1 −1 1 −3  

slide-14
SLIDE 14

Systems of linear equations

3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3 can be abbreviated via an augmented matrix   3 5 4 3 2 2 1 1 1 6 1 1 −1 1 −3  

  • r written in vector form as

w   3 2 1   + x   5 1 1   + y   4 1 −1   + z   3 1 1   =   2 6 −3  

slide-15
SLIDE 15

The matrix-vector product

slide-16
SLIDE 16

The matrix-vector product

There is one more way to write systems of linear equations:

slide-17
SLIDE 17

The matrix-vector product

There is one more way to write systems of linear equations: 3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3

slide-18
SLIDE 18

The matrix-vector product

There is one more way to write systems of linear equations: 3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3 can be written as

slide-19
SLIDE 19

The matrix-vector product

There is one more way to write systems of linear equations: 3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3 can be written as   3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3  

slide-20
SLIDE 20

The matrix-vector product

In general, a column vector with c entries (in Rc)

slide-21
SLIDE 21

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left

slide-22
SLIDE 22

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns

slide-23
SLIDE 23

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns — the matrix has as many columns as the original vector has rows —

slide-24
SLIDE 24

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns — the matrix has as many columns as the original vector has rows — to obtain a vector with r rows (in Rr)

slide-25
SLIDE 25

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns — the matrix has as many columns as the original vector has rows — to obtain a vector with r rows (in Rr) — the new vector will have as many rows as the matrix.

slide-26
SLIDE 26

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns — the matrix has as many columns as the original vector has rows — to obtain a vector with r rows (in Rr) — the new vector will have as many rows as the matrix. [matrix with r rows and c columns]

slide-27
SLIDE 27

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns — the matrix has as many columns as the original vector has rows — to obtain a vector with r rows (in Rr) — the new vector will have as many rows as the matrix. [matrix with r rows and c columns] [vector with c rows]

slide-28
SLIDE 28

The matrix-vector product

In general, a column vector with c entries (in Rc) can be multiplied on the left by a matrix with r rows and c columns — the matrix has as many columns as the original vector has rows — to obtain a vector with r rows (in Rr) — the new vector will have as many rows as the matrix. [matrix with r rows and c columns] [vector with c rows] = [vector with r rows]

slide-29
SLIDE 29

The matrix-vector product

The formula

     a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      a11x1 + a12x2 + · · · + a1cxc a21x1 + a22x2 + · · · + a2cxc . . . ar1x1 + ar2x2 + · · · + arcxc     

slide-30
SLIDE 30

The matrix-vector product

The formula

     a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      a11x1 + a12x2 + · · · + a1cxc a21x1 + a22x2 + · · · + a2cxc . . . ar1x1 + ar2x2 + · · · + arcxc     

Example

  1 2 3 4 5 6  

  • 1

−1

  • =

  1 − 2 3 − 4 5 − 6  

slide-31
SLIDE 31

The matrix-vector product

The formula

     a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      a11x1 + a12x2 + · · · + a1cxc a21x1 + a22x2 + · · · + a2cxc . . . ar1x1 + ar2x2 + · · · + arcxc     

Example

  1 2 3 4 5 6  

  • 1

−1

  • =

  1 − 2 3 − 4 5 − 6   =   −1 −1 −1  

slide-32
SLIDE 32

Try it yourself!

  1 −1 −1 2 3 2     3 4 5  

slide-33
SLIDE 33

Try it yourself!

  1 −1 −1 2 3 2     3 4 5   These ones are the wrong size.   1 −1 −1 2 3 2   3 4

slide-34
SLIDE 34

Try it yourself!

  1 −1 −1 2 3 2     3 4 5   These ones are the wrong size.   1 −1 −1 2 3 2   3 4

  • =

  1 ∗ 3 + (−1) ∗ 4 (−1) ∗ 3 + 2 ∗ 4 3 ∗ 3 + 2 ∗ 4  

slide-35
SLIDE 35

Try it yourself!

  1 −1 −1 2 3 2     3 4 5   These ones are the wrong size.   1 −1 −1 2 3 2   3 4

  • =

  1 ∗ 3 + (−1) ∗ 4 (−1) ∗ 3 + 2 ∗ 4 3 ∗ 3 + 2 ∗ 4   =   −1 5 17  

slide-36
SLIDE 36

The matrix-vector product

So the equation      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      b1 b2 . . . br     

slide-37
SLIDE 37

The matrix-vector product

So the equation      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      b1 b2 . . . br      is equivalent to the equation      a11x1 + a12x2 + · · · + a1cxc a21x1 + a22x2 + · · · + a2cxc . . . ar1x1 + ar2x2 + · · · + arcxn      =      b1 b2 . . . br     

slide-38
SLIDE 38

An example with numbers

  3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3  

slide-39
SLIDE 39

An example with numbers

  3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3   Expanding the product on the left, we find   3w + 5x + 4y + 3z 2w + x + y + z w + x − y + z   =   2 6 −3  

slide-40
SLIDE 40

An example with numbers

  3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3   Expanding the product on the left, we find   3w + 5x + 4y + 3z 2w + x + y + z w + x − y + z   =   2 6 −3  

  • r in other words

3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3

slide-41
SLIDE 41

Ax = b

If we write A =      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc      , x =      x1 x2 . . . xc      , b =      b1 b2 . . . br     

slide-42
SLIDE 42

Ax = b

If we write A =      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc      , x =      x1 x2 . . . xc      , b =      b1 b2 . . . br      Then the equation      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      b1 b2 . . . br     

slide-43
SLIDE 43

Ax = b

If we write A =      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc      , x =      x1 x2 . . . xc      , b =      b1 b2 . . . br      Then the equation      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      b1 b2 . . . br      becomes simply Ax = b.

slide-44
SLIDE 44

Ax = b

If we write A =      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc      , x =      x1 x2 . . . xc      , b =      b1 b2 . . . br      Then the equation      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      =      b1 b2 . . . br      becomes simply Ax = b. Here A is called the matrix of coefficients.

slide-45
SLIDE 45

The matrix-vector product

The product Ax can also be written as      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      = x1      a11 a21 . . . ar1      + · · · + xc      a1c a2c . . . arc     

slide-46
SLIDE 46

The matrix-vector product

The product Ax can also be written as      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      = x1      a11 a21 . . . ar1      + · · · + xc      a1c a2c . . . arc      That is, Ax is a linear combination

slide-47
SLIDE 47

The matrix-vector product

The product Ax can also be written as      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      = x1      a11 a21 . . . ar1      + · · · + xc      a1c a2c . . . arc      That is, Ax is a linear combination of the columns of A,

slide-48
SLIDE 48

The matrix-vector product

The product Ax can also be written as      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      = x1      a11 a21 . . . ar1      + · · · + xc      a1c a2c . . . arc      That is, Ax is a linear combination of the columns of A, with coefficients given by the entries of the vector x. The equation Ax = b

slide-49
SLIDE 49

The matrix-vector product

The product Ax can also be written as      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      = x1      a11 a21 . . . ar1      + · · · + xc      a1c a2c . . . arc      That is, Ax is a linear combination of the columns of A, with coefficients given by the entries of the vector x. The equation Ax = b asserts that the vector b

slide-50
SLIDE 50

The matrix-vector product

The product Ax can also be written as      a11 a12 · · · a1c a21 a22 · · · a2c . . . . . . ... . . . ar1 ar2 · · · arc           x1 x2 . . . xc      = x1      a11 a21 . . . ar1      + · · · + xc      a1c a2c . . . arc      That is, Ax is a linear combination of the columns of A, with coefficients given by the entries of the vector x. The equation Ax = b asserts that the vector b is equal to this linear combination of the columns of A.

slide-51
SLIDE 51

An example with numbers

  3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3  

slide-52
SLIDE 52

An example with numbers

  3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3   The product on the left means a linear combination of the columns

  • f the matrix, weighted by the entries of the vector.

w   3 2 1   + x   5 1 1   + y   4 1 −1   + z   3 1 1   =   2 6 −3  

slide-53
SLIDE 53

An example with numbers

  3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   2 6 −3   The product on the left means a linear combination of the columns

  • f the matrix, weighted by the entries of the vector.

w   3 2 1   + x   5 1 1   + y   4 1 −1   + z   3 1 1   =   2 6 −3  

  • r in other words

3w + 5x + 4y + 3z = 2 2w + x + y + z = 6 w + x − y + z = −3

slide-54
SLIDE 54

A function from R4 to R3

    w x y z    

slide-55
SLIDE 55

A function from R4 to R3

    w x y z     →   3 5 4 3 2 1 1 1 1 1 −1 1       w x y z    

slide-56
SLIDE 56

A function from R4 to R3

    w x y z     →   3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   3w + 5x + 4y + 3z 2w + x + y + z w + x − y + z  

slide-57
SLIDE 57

A function from R4 to R3

    w x y z     →   3 5 4 3 2 1 1 1 1 1 −1 1       w x y z     =   3w + 5x + 4y + 3z 2w + x + y + z w + x − y + z   This map brought to you by the matrix   3 5 4 3 2 1 1 1 1 1 −1 1  

slide-58
SLIDE 58

Functions from matrices

If A is a matrix with

slide-59
SLIDE 59

Functions from matrices

If A is a matrix with r rows and c columns,

slide-60
SLIDE 60

Functions from matrices

If A is a matrix with r rows and c columns, and x ∈ Rc is any vector,

slide-61
SLIDE 61

Functions from matrices

If A is a matrix with r rows and c columns, and x ∈ Rc is any vector, then Ax ∈ Rr.

slide-62
SLIDE 62

Functions from matrices

If A is a matrix with r rows and c columns, and x ∈ Rc is any vector, then Ax ∈ Rr. So the matrix A defines a function (also called A):

slide-63
SLIDE 63

Functions from matrices

If A is a matrix with r rows and c columns, and x ∈ Rc is any vector, then Ax ∈ Rr. So the matrix A defines a function (also called A): A : Rc → Rr x → Ax

slide-64
SLIDE 64

Functions from matrices

If A is a matrix with r rows and c columns, and x ∈ Rc is any vector, then Ax ∈ Rr. So the matrix A defines a function (also called A): A : Rc → Rr x → Ax The equation Ax = b can be read as:

slide-65
SLIDE 65

Functions from matrices

If A is a matrix with r rows and c columns, and x ∈ Rc is any vector, then Ax ∈ Rr. So the matrix A defines a function (also called A): A : Rc → Rr x → Ax The equation Ax = b can be read as: which vectors x have the property that their image under the function A is the vector b?

slide-66
SLIDE 66

Functions from matrices

This is similar to the way that a number a determines a function “multiplication by a”

slide-67
SLIDE 67

Functions from matrices

This is similar to the way that a number a determines a function “multiplication by a” a : R → R x → ax

slide-68
SLIDE 68

Functions from matrices

This is similar to the way that a number a determines a function “multiplication by a” a : R → R x → ax Indeed, this is the case n = m = 1.

slide-69
SLIDE 69

Linearity

The matrix-vector product has the following properties: A(x + y) = Ax + Ay A(cx) = c(Ax)

slide-70
SLIDE 70

Linearity

The matrix-vector product has the following properties: A(x + y) = Ax + Ay A(cx) = c(Ax) Functions with these two properties are said to be linear.

slide-71
SLIDE 71

Linearity

The matrix-vector product has the following properties: A(x + y) = Ax + Ay A(cx) = c(Ax) Functions with these two properties are said to be linear. In fact, every linear function from Rc to Rr

slide-72
SLIDE 72

Linearity

The matrix-vector product has the following properties: A(x + y) = Ax + Ay A(cx) = c(Ax) Functions with these two properties are said to be linear. In fact, every linear function from Rc to Rr is multiplication by some matrix.

slide-73
SLIDE 73

Linearity

The matrix-vector product has the following properties: A(x + y) = Ax + Ay A(cx) = c(Ax) Functions with these two properties are said to be linear. In fact, every linear function from Rc to Rr is multiplication by some

  • matrix. We will see why this is next time.
slide-74
SLIDE 74

Thinking about linear equations

slide-75
SLIDE 75

Thinking about linear equations

We have seen that linear equations can be interpreted as

slide-76
SLIDE 76

Thinking about linear equations

We have seen that linear equations can be interpreted as

◮ Asking for the intersection locus of lines or planes

slide-77
SLIDE 77

Thinking about linear equations

We have seen that linear equations can be interpreted as

◮ Asking for the intersection locus of lines or planes ◮ Asking how to write one vector as a linear combination of

given others

slide-78
SLIDE 78

Thinking about linear equations

We have seen that linear equations can be interpreted as

◮ Asking for the intersection locus of lines or planes ◮ Asking how to write one vector as a linear combination of

given others

◮ Asking for vectors mapping to a given one under the linear

function associated to the coefficient matrix

slide-79
SLIDE 79

Thinking about linear equations

We have seen that linear equations can be interpreted as

◮ Asking for the intersection locus of lines or planes ◮ Asking how to write one vector as a linear combination of

given others

◮ Asking for vectors mapping to a given one under the linear

function associated to the coefficient matrix We know how to compute the solutions

slide-80
SLIDE 80

Thinking about linear equations

We have seen that linear equations can be interpreted as

◮ Asking for the intersection locus of lines or planes ◮ Asking how to write one vector as a linear combination of

given others

◮ Asking for vectors mapping to a given one under the linear

function associated to the coefficient matrix We know how to compute the solutions (row reduction).

slide-81
SLIDE 81

Thinking about linear equations

We have seen that linear equations can be interpreted as

◮ Asking for the intersection locus of lines or planes ◮ Asking how to write one vector as a linear combination of

given others

◮ Asking for vectors mapping to a given one under the linear

function associated to the coefficient matrix We know how to compute the solutions (row reduction). Now we’ll discuss qualitative properties of the solution set.

slide-82
SLIDE 82

Homogenous and inhomogenous equations

slide-83
SLIDE 83

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b,

slide-84
SLIDE 84

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2)

slide-85
SLIDE 85

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2

slide-86
SLIDE 86

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b

slide-87
SLIDE 87

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0

slide-88
SLIDE 88

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0 That is, the difference between two solutions to the inhomogenous equation Ax = b

slide-89
SLIDE 89

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0 That is, the difference between two solutions to the inhomogenous equation Ax = b is a solution to the homogenous equation Ax = 0.

slide-90
SLIDE 90

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0 That is, the difference between two solutions to the inhomogenous equation Ax = b is a solution to the homogenous equation Ax = 0. Differently said,

slide-91
SLIDE 91

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0 That is, the difference between two solutions to the inhomogenous equation Ax = b is a solution to the homogenous equation Ax = 0. Differently said, given one solution x = x0 to the inhomogenous equation Ax = b,

slide-92
SLIDE 92

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0 That is, the difference between two solutions to the inhomogenous equation Ax = b is a solution to the homogenous equation Ax = 0. Differently said, given one solution x = x0 to the inhomogenous equation Ax = b, all other solutions are given by the sum of x0

slide-93
SLIDE 93

Homogenous and inhomogenous equations

If Ax1 = b and Ax2 = b, then A(x1 − x2) = Ax1 − Ax2 = b − b = 0 That is, the difference between two solutions to the inhomogenous equation Ax = b is a solution to the homogenous equation Ax = 0. Differently said, given one solution x = x0 to the inhomogenous equation Ax = b, all other solutions are given by the sum of x0 and a solution to the homogenous equation Ax = 0.

slide-94
SLIDE 94

Try it yourself

Find all solutions to the homogenous equation x + y = 0 and to the inhomogenous equation x + y = 1. Graph your answers. How do they compare?

slide-95
SLIDE 95

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0,

slide-96
SLIDE 96

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2,

slide-97
SLIDE 97

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2, A(c1x1 + c2x2)

slide-98
SLIDE 98

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2, A(c1x1 + c2x2) = c1Ax1 + c2Ax2

slide-99
SLIDE 99

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2, A(c1x1 + c2x2) = c1Ax1 + c2Ax2 = 0

slide-100
SLIDE 100

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2, A(c1x1 + c2x2) = c1Ax1 + c2Ax2 = 0 In other words,

slide-101
SLIDE 101

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2, A(c1x1 + c2x2) = c1Ax1 + c2Ax2 = 0 In other words, any linear combination of solutions to a homogenous equation is again a solution.

slide-102
SLIDE 102

Homogenous equations and linearity

If Ax1 = 0 and Ax2 = 0, then for any scalars c1, c2, A(c1x1 + c2x2) = c1Ax1 + c2Ax2 = 0 In other words, any linear combination of solutions to a homogenous equation is again a solution. This is not true for inhomogenous equations! (Try x = 1.)

slide-103
SLIDE 103

Homogenous equations and linearity

The solution set to a homogenous equation can be described as a linear span.

slide-104
SLIDE 104

Homogenous equations and linearity

The solution set to a homogenous equation can be described as a linear span. To do this, as always, the first step is row reducing your system.

slide-105
SLIDE 105

Homogenous equations and linearity

The solution set to a homogenous equation can be described as a linear span. To do this, as always, the first step is row reducing your system. You are all expert row reducers now,

slide-106
SLIDE 106

Homogenous equations and linearity

The solution set to a homogenous equation can be described as a linear span. To do this, as always, the first step is row reducing your system. You are all expert row reducers now, so I’ll skip that step and just start with an already reduced one.

slide-107
SLIDE 107

Homogenous equations and linearity

    1 1 1 1 1           x1 x2 x3 x4 x5       =        

slide-108
SLIDE 108

Homogenous equations and linearity

    1 1 1 1 1           x1 x2 x3 x4 x5       =         This has the following augmented matrix:     1 1 1 1 1    

slide-109
SLIDE 109

Homogenous equations and linearity

We read off the solution from the augmented matrix     1 1 1 1 1    

slide-110
SLIDE 110

Homogenous equations and linearity

We read off the solution from the augmented matrix     1 1 1 1 1     Introduce free parameters for the variables whose column has no pivot;

slide-111
SLIDE 111

Homogenous equations and linearity

We read off the solution from the augmented matrix     1 1 1 1 1     Introduce free parameters for the variables whose column has no pivot; s for x1 and t for x5.

slide-112
SLIDE 112

Homogenous equations and linearity

We read off the solution from the augmented matrix     1 1 1 1 1     Introduce free parameters for the variables whose column has no pivot; s for x1 and t for x5. Now we read off the solutions:

slide-113
SLIDE 113

Homogenous equations and linearity

We read off the solution from the augmented matrix     1 1 1 1 1     Introduce free parameters for the variables whose column has no pivot; s for x1 and t for x5. Now we read off the solutions: these are (s, −t, −t, 0, t) for any s, t, or in other words,

slide-114
SLIDE 114

Homogenous equations and linearity

We read off the solution from the augmented matrix     1 1 1 1 1     Introduce free parameters for the variables whose column has no pivot; s for x1 and t for x5. Now we read off the solutions: these are (s, −t, −t, 0, t) for any s, t, or in other words,            s       1       + t       −1 −1 1       , any s, t            = Linear Span             1       ,       −1 −1 1            

slide-115
SLIDE 115

Inhomogenous equations

slide-116
SLIDE 116

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b,

slide-117
SLIDE 117

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty,

slide-118
SLIDE 118

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty, is a translate of the solution set of the homogenous equation Ax = 0.

slide-119
SLIDE 119

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty, is a translate of the solution set of the homogenous equation Ax = 0. Indeed, if Ax0 = b

slide-120
SLIDE 120

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty, is a translate of the solution set of the homogenous equation Ax = 0. Indeed, if Ax0 = b then for any x such that Ax = 0,

slide-121
SLIDE 121

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty, is a translate of the solution set of the homogenous equation Ax = 0. Indeed, if Ax0 = b then for any x such that Ax = 0, we have A(x0 + x) = Ax0 + Ax = b + 0 = b

slide-122
SLIDE 122

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty, is a translate of the solution set of the homogenous equation Ax = 0. Indeed, if Ax0 = b then for any x such that Ax = 0, we have A(x0 + x) = Ax0 + Ax = b + 0 = b Symbolically,

slide-123
SLIDE 123

Inhomogenous equations

The solution set of the inhomogenous equation Ax = b, if nonempty, is a translate of the solution set of the homogenous equation Ax = 0. Indeed, if Ax0 = b then for any x such that Ax = 0, we have A(x0 + x) = Ax0 + Ax = b + 0 = b Symbolically, x0 + {solutions of Ax = 0} = {solutions of Ax = b}

slide-124
SLIDE 124

Try it yourself!

Solve the following three systems of equations. 1 2 3 6 x y

  • =
  • 1

2 3 6 x y

  • =

3 4

  • 1

2 3 6 x y

  • =

2 6

slide-125
SLIDE 125

Try it yourself!

1 2 3 6 x y

  • =
  • has solution set the linear span of (−2, 1).
slide-126
SLIDE 126

Try it yourself!

1 2 3 6 x y

  • =
  • has solution set the linear span of (−2, 1).

1 2 3 6 x y

  • =

3 4

  • has the empty solution set
slide-127
SLIDE 127

Try it yourself!

1 2 3 6 x y

  • =
  • has solution set the linear span of (−2, 1).

1 2 3 6 x y

  • =

3 4

  • has the empty solution set

1 2 3 6 x y

  • =

2 4

  • has solution set (2, 0) + Linear span(−2, 1).
slide-128
SLIDE 128

Linear dependence and independence

slide-129
SLIDE 129

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent

slide-130
SLIDE 130

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.
slide-131
SLIDE 131

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.

Otherwise, it’s said to be linearly independent.

slide-132
SLIDE 132

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.

Otherwise, it’s said to be linearly independent. An equivalent characterization:

slide-133
SLIDE 133

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.

Otherwise, it’s said to be linearly independent. An equivalent characterization: the vectors are linearly dependent if there is some collection of scalars ai,

slide-134
SLIDE 134

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.

Otherwise, it’s said to be linearly independent. An equivalent characterization: the vectors are linearly dependent if there is some collection of scalars ai, not all zero, such that a1v1 + a2v2 + · · · akvk = 0

slide-135
SLIDE 135

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.

Otherwise, it’s said to be linearly independent. An equivalent characterization: the vectors are linearly dependent if there is some collection of scalars ai, not all zero, such that a1v1 + a2v2 + · · · akvk = 0

Example

The vectors   1   ,   1   ,   1  

slide-136
SLIDE 136

Linear dependence and independence

A collection of vectors v1, . . . , vk is said to be linearly dependent if

  • ne of the vi can be written as a linear combination of the others.

Otherwise, it’s said to be linearly independent. An equivalent characterization: the vectors are linearly dependent if there is some collection of scalars ai, not all zero, such that a1v1 + a2v2 + · · · akvk = 0

Example

The vectors   1   ,   1   ,   1   are linearly independent.

slide-137
SLIDE 137

The zero vector

slide-138
SLIDE 138

The zero vector

The set containing only the zero vector, {0} is

slide-139
SLIDE 139

The zero vector

The set containing only the zero vector, {0} is linearly dependent 1 × 0 = 0

slide-140
SLIDE 140

The zero vector

The set containing only the zero vector, {0} is linearly dependent 1 × 0 = 0 More generally, any collection of vectors which includes the zero vector is

slide-141
SLIDE 141

The zero vector

The set containing only the zero vector, {0} is linearly dependent 1 × 0 = 0 More generally, any collection of vectors which includes the zero vector is linearly dependent. 0 × v1 + 0 × v2 + · · · 0 × vn + 1 × 0 = 0

slide-142
SLIDE 142

Elementary operations do not change linear independence

slide-143
SLIDE 143

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent.

slide-144
SLIDE 144

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent. Then so too are any re-arrangement of these vectors

slide-145
SLIDE 145

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent. Then so too are any re-arrangement of these vectors and also any rescaling by nonzero vectors.

slide-146
SLIDE 146

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent. Then so too are any re-arrangement of these vectors and also any rescaling by nonzero vectors. Also, so are v1, v2 + cv1, v3, . . . , vn.

slide-147
SLIDE 147

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent. Then so too are any re-arrangement of these vectors and also any rescaling by nonzero vectors. Also, so are v1, v2 + cv1, v3, . . . , vn. Indeed, if a1v1 + a2v2 + · · · + anvn = 0

slide-148
SLIDE 148

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent. Then so too are any re-arrangement of these vectors and also any rescaling by nonzero vectors. Also, so are v1, v2 + cv1, v3, . . . , vn. Indeed, if a1v1 + a2v2 + · · · + anvn = 0 then so too (a1 − ca2)v1 + a2(v2 + cv1) + a3v3 + · · · + anvn = 0

slide-149
SLIDE 149

Elementary operations do not change linear independence

Suppose v1, v2, . . . , vn are linearly dependendent. Then so too are any re-arrangement of these vectors and also any rescaling by nonzero vectors. Also, so are v1, v2 + cv1, v3, . . . , vn. Indeed, if a1v1 + a2v2 + · · · + anvn = 0 then so too (a1 − ca2)v1 + a2(v2 + cv1) + a3v3 + · · · + anvn = 0 Moreover, a1 − ca2 and a2 are both zero if and only if a1 and a2 are both zero.

slide-150
SLIDE 150

Row reduction and linear independence

The rows of a reduced echelon matrix are linearly independent if and only if there is no zero row.

slide-151
SLIDE 151

Row reduction and linear independence

The rows of a reduced echelon matrix are linearly independent if and only if there is no zero row. This is because each pivot sits in a column with only zeros, so any non-zero linear combination of the rows will see one of the pivot entries.

slide-152
SLIDE 152

Row reduction and linear independence

The rows of a reduced echelon matrix are linearly independent if and only if there is no zero row. This is because each pivot sits in a column with only zeros, so any non-zero linear combination of the rows will see one of the pivot entries. So to check if a collection of vectors is linearly dependent or linearly independent

slide-153
SLIDE 153

Row reduction and linear independence

The rows of a reduced echelon matrix are linearly independent if and only if there is no zero row. This is because each pivot sits in a column with only zeros, so any non-zero linear combination of the rows will see one of the pivot entries. So to check if a collection of vectors is linearly dependent or linearly independent make them the rows of a matrix, row reduce, and then look for a zero row!.

slide-154
SLIDE 154

Row reduction and linear independence

The rows of a reduced echelon matrix are linearly independent if and only if there is no zero row. This is because each pivot sits in a column with only zeros, so any non-zero linear combination of the rows will see one of the pivot entries. So to check if a collection of vectors is linearly dependent or linearly independent make them the rows of a matrix, row reduce, and then look for a zero row!.

Is this true for an echelon matrix?

slide-155
SLIDE 155

Row reduction and linear independence

The rows of a reduced echelon matrix are linearly independent if and only if there is no zero row. This is because each pivot sits in a column with only zeros, so any non-zero linear combination of the rows will see one of the pivot entries. So to check if a collection of vectors is linearly dependent or linearly independent make them the rows of a matrix, row reduce, and then look for a zero row!.

Is this true for an echelon matrix?

Think about it!

slide-156
SLIDE 156

Try it yourself!

  • ,
slide-157
SLIDE 157

Try it yourself!

  • , linearly dependent

  1 2 4   ,   2 4 8  

slide-158
SLIDE 158

Try it yourself!

  • , linearly dependent

  1 2 4   ,   2 4 8   linearly dependent   1 2 3   ,   4 5 6   ,   5 7 9  

slide-159
SLIDE 159

Try it yourself!

  • , linearly dependent

  1 2 4   ,   2 4 8   linearly dependent   1 2 3   ,   4 5 6   ,   5 7 9   linearly dependent   1 2 3   ,   1 2   ,   1  

slide-160
SLIDE 160

Try it yourself!

  • , linearly dependent

  1 2 4   ,   2 4 8   linearly dependent   1 2 3   ,   4 5 6   ,   5 7 9   linearly dependent   1 2 3   ,   1 2   ,   1   linearly independent