overview
play

Overview Last time we introduced the notion of an orthonormal basis - PowerPoint PPT Presentation

Overview Last time we introduced the notion of an orthonormal basis for a subspace. We also saw that if a square matrix U has orthonormal columns, then U is invertible and U 1 = U T . Such a matrix is called an orthogonal matrix. At the


  1. Overview Last time we introduced the notion of an orthonormal basis for a subspace. We also saw that if a square matrix U has orthonormal columns, then U is invertible and U − 1 = U T . Such a matrix is called an orthogonal matrix. At the beginning of the course we developed a formula for computing the projection of one vector onto another in R 2 or R 3 . Today we’ll generalise this notion to higher dimensions. From Lay, §6.3 A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 1 / 24

  2. Review Recall from Stewart that if u � = 0 and y are vectors in R n , then proj u y = y · u u · uu is the orthogonal projection of y onto u . (Lay uses the notation “ ˆ y ” for this projection, where u is understood.) A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 2 / 24

  3. Review Recall from Stewart that if u � = 0 and y are vectors in R n , then proj u y = y · u u · uu is the orthogonal projection of y onto u . (Lay uses the notation “ ˆ y ” for this projection, where u is understood.) How would you describe the vector proj u y in words? A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 2 / 24

  4. Review Recall from Stewart that if u � = 0 and y are vectors in R n , then proj u y = y · u u · uu is the orthogonal projection of y onto u . (Lay uses the notation “ ˆ y ” for this projection, where u is understood.) How would you describe the vector proj u y in words? One possible answer: y can be written as the sum of a vector parallel to u and a vector orthogonal to u ; proj u y is the summand parallel to u . Or alternatively, y can be written as the sum of a vector in the line spanned by u and a vector orthogonal to u ; proj u y is the summand in Span { u } . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 2 / 24

  5. Review Recall from Stewart that if u � = 0 and y are vectors in R n , then proj u y = y · u u · uu is the orthogonal projection of y onto u . (Lay uses the notation “ ˆ y ” for this projection, where u is understood.) How would you describe the vector proj u y in words? One possible answer: y can be written as the sum of a vector parallel to u and a vector orthogonal to u ; proj u y is the summand parallel to u . Or alternatively, y can be written as the sum of a vector in the line spanned by u and a vector orthogonal to u ; proj u y is the summand in Span { u } . We’d like to generalise this, replacing Span { u } by an arbitrary subspace: Given y and a subspace W in R n , we’d like to write y as a sum of a vector in W and a vector in W ⊥ . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 2 / 24

  6. Example 1 Suppose that { u 1 , u 2 , u 3 } is an orthogonal basis for R 3 and let W = Span { u 1 , u 2 } . Write y as the sum of a vector ˆ y in W and a vector z in W ⊥ . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 3 / 24

  7. Example 1 Suppose that { u 1 , u 2 , u 3 } is an orthogonal basis for R 3 and let W = Span { u 1 , u 2 } . Write y as the sum of a vector ˆ y in W and a vector z in W ⊥ . y W ¶ z W u 2 y ˆ 0 u 1 A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 3 / 24

  8. Recall that for any orthogonal basis, we have y = y · u 1 u 1 + y · u 2 u 2 + y · u 3 u 3 . u 1 · u 1 u 2 · u 2 u 3 · u 3 A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 4 / 24

  9. Recall that for any orthogonal basis, we have y = y · u 1 u 1 + y · u 2 u 2 + y · u 3 u 3 . u 1 · u 1 u 2 · u 2 u 3 · u 3 It follows that y = y · u 1 u 1 + y · u 2 ˆ u 2 u 1 · u 1 u 2 · u 2 and z = y · u 3 u 3 . u 3 · u 3 Since u 3 is orthogonal to u 1 and u 2 , its scalar multiples are orthogonal to Span { u 1 , u 2 } . Therefore z ∈ W ⊥ All this can be generalised to any vector y and subspace W of R n , as we will see next. A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 4 / 24

  10. The Orthogonal Decomposition Theorem Theorem Let W be a subspace in R n . Then each y ∈ R n can be written uniquely in the form y = ˆ y + z (1) y ∈ W and z ∈ W ⊥ . where ˆ If { u 1 , . . . , u p } is any orthogonal basis of W , then y = y · u 1 u 1 + · · · + y · u p ˆ (2) u p u 1 · u 1 u p · u p The vector ˆ y is called the orthogonal projection of y onto W . Note that it follows from this theorem that to calculate the decomposition y = ˆ y + z , it is enough to know one orthogonal basis for W explicitly. Any orthogonal basis will do, and all orthogonal bases will give the same decomposition y = ˆ y + z . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 5 / 24

  11. Example 2 Given       1 1 0 1 0 − 1       u 1 =  , u 2 =  , u 3 =       0 1 1           − 1 1 − 1 let W be the subspace of R 4 spanned by { u 1 , u 2 , u 3 } .   2 − 3   Write y =  as the sum of a vector in W and a vector orthogonal to   4    1 W . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 6 / 24

  12. The orthogonal projection of y onto W is given by y · u 1 u 1 + y · u 2 u 2 + y · u 3 ˆ y = u 3 u 1 · u 1 u 2 · u 2 u 3 · u 3       1 1 0 − 2 1  + 7 0  + 6 − 1       =       0 1 1 3   3   3       − 1 1 − 1   5 1 − 8   =   13 3     3 A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 7 / 24

  13. The orthogonal projection of y onto W is given by y · u 1 u 1 + y · u 2 u 2 + y · u 3 y ˆ = u 3 u 1 · u 1 u 2 · u 2 u 3 · u 3       1 1 0 − 2 1  + 7 0  + 6 − 1       =       0 1 1 3   3   3       − 1 1 − 1   5 1 − 8   =   13 3     3 Also       2 5 1 − 3  − 1 − 8  = 1 − 1       z = y − ˆ y =       4 3 13 3 − 1           1 3 0 A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 7 / 24

  14. Thus the desired decomposition of y is = ˆ y + z y       2 5 1 − 3 1 − 8  + 1 − 1       =  .       4 13 − 1   3   3       1 3 0 y is in W ⊥ . The Orthogonal Decomposition Theorem ensures that z = y − ˆ However, verifying this is a good check against computational mistakes. A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 8 / 24

  15. Thus the desired decomposition of y is = ˆ y + z y       2 5 1 − 3 1 − 8  + 1 − 1       =  .       4 13 − 1   3   3       1 3 0 y is in W ⊥ . The Orthogonal Decomposition Theorem ensures that z = y − ˆ However, verifying this is a good check against computational mistakes. This problem was made easier by the fact that { u 1 , u 2 , u 3 } is an orthogonal basis for W . If you were given an arbitrary basis for W instead of an orthogonal basis, what would you do? A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 8 / 24

  16. Theorem (The Best Approximation Theorem) Let W be a subspace of R n , y any vector in R n , and ˆ y the orthogonal projection of y onto W . Then ˆ y is the closest vector in W to y , in the sense that � y − ˆ y � < � y − v � (3) for all v in W , v � = ˆ y . y ||y - v|| ||y - ŷ || 0 ŷ v || ŷ - v|| W A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 9 / 24

  17. Proof Let v be any vector in W , v � = ˆ y . Then ˆ y − v ∈ W . By the Orthogonal Decomposition Theorem, y − ˆ y is orthogonal to W . In particular y − ˆ y is orthogonal to ˆ y − v . Since y − v = ( y − ˆ y ) + (ˆ y − v ) the Pythagorean Theorem gives � y − v � 2 = � y − ˆ y � 2 + � ˆ y − v � 2 . Hence � y − v � 2 > � y − ˆ y � 2 . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 10 / 24

  18. We can now define the distance from a vector y to a subspace W of R n . Definition Let W be a subspace of R n and let y be a vector in R n . The distance from y to W is || y − ˆ y || where ˆ y is the orthogonal projection of y onto W . A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 11 / 24

  19. Example 3 Consider the vectors       3 1 − 4 − 1 − 2 1       y =  , u 1 =  , u 2 =  .       1 − 1 0          13 2 3 Find the closest vector to y in W = Span { u 1 , u 2 } . y · u 1 u 1 + y · u 2 ˆ y = u 2 u 1 · u 1 u 2 · u 2       1 − 4 − 1 30  + 26 − 2 1 − 5       =  =  .       − 1 0 − 3 10 26          2 3 9 A/Prof Scott Morrison (ANU) MATH1014 Notes Second Semester 2016 12 / 24

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend