on best approximations of matrix polynomials
play

On best approximations of matrix polynomials Petr Tich joint work - PowerPoint PPT Presentation

On best approximations of matrix polynomials Petr Tich joint work with Jrg Liesen Institute of Computer Science, Academy of Sciences of the Czech Republic September 12, 2008 Technische Universitt Hamburg-Harburg, Germany GAMM Workshop


  1. On best approximations of matrix polynomials Petr Tichý joint work with Jörg Liesen Institute of Computer Science, Academy of Sciences of the Czech Republic September 12, 2008 Technische Universität Hamburg-Harburg, Germany GAMM Workshop on Applied and Numerical Linear Algebra 1

  2. Ideal Arnoldi approximation problem p ∈P m � A m +1 − p ( A ) � , p ∈M m +1 � p ( A ) � = min min where M m +1 is the class of monic polynomials of degree m + 1 , P m is the class of polynomials of degree at most m . 2

  3. Ideal Arnoldi approximation problem p ∈P m � A m +1 − p ( A ) � , p ∈M m +1 � p ( A ) � = min min where M m +1 is the class of monic polynomials of degree m + 1 , P m is the class of polynomials of degree at most m . Introduced in [Greenbaum and Trefethen, 1994] , paper contains uniqueness result ( → story of the proof). The unique polynomial that solves the problem is called the ( m + 1) st ideal Arnoldi polynomial of A , or the ( m + 1) st Chebyshev polynomial of A . Some work on these polynomials in [Toh PhD thesis, 1996] , [Toh and Trefethen, 1998] , [Trefethen and Embree, 2005] . 2

  4. Matrix function best approximation problem We consider the matrix approximation problem p ∈P m � f ( A ) − p ( A ) � min � · � is the spectral norm (matrix 2-norm), A ∈ C n × n , f is analytic in neighborhood of A ’s spectrum. 3

  5. Matrix function best approximation problem We consider the matrix approximation problem p ∈P m � f ( A ) − p ( A ) � min � · � is the spectral norm (matrix 2-norm), A ∈ C n × n , f is analytic in neighborhood of A ’s spectrum. Well known: f ( A ) = p f ( A ) for a polynomial p f depending on values and possibly derivatives of f on A ’s spectrum. Without loss of generality we assume that f is a given polynomial. 3

  6. Matrix function best approximation problem We consider the matrix approximation problem p ∈P m � f ( A ) − p ( A ) � min � · � is the spectral norm (matrix 2-norm), A ∈ C n × n , f is analytic in neighborhood of A ’s spectrum. Well known: f ( A ) = p f ( A ) for a polynomial p f depending on values and possibly derivatives of f on A ’s spectrum. Without loss of generality we assume that f is a given polynomial. Does this problem have a unique solution p ∗ ∈ P m ? 3

  7. Outline General matrix approximation problems 1 Formulation of matrix polynomial approximation problems 2 Uniqueness results 3 Ideal Arnoldi versus ideal GMRES polynomials 4 4

  8. Outline General matrix approximation problems 1 Formulation of matrix polynomial approximation problems 2 Uniqueness results 3 Ideal Arnoldi versus ideal GMRES polynomials 4 5

  9. General matrix approximation problems Given m linearly independent matrices A 1 , . . . , A m ∈ C n × n , A ≡ span { A 1 , . . . , A m } , B ∈ C n × n \ A , � · � is a matrix norm. Consider the best approximation problem M ∈ A � B − M � . min 6

  10. General matrix approximation problems Given m linearly independent matrices A 1 , . . . , A m ∈ C n × n , A ≡ span { A 1 , . . . , A m } , B ∈ C n × n \ A , � · � is a matrix norm. Consider the best approximation problem M ∈ A � B − M � . min This problem has a unique solution if � · � is strictly convex. [see, e.g., Sreedharan, 1973] 6

  11. Strictly convex norms The norm � · � is strictly convex if for all X , Y , � X � = � Y � = 1 , � X + Y � = 2 ⇒ X = Y . 7

  12. Strictly convex norms The norm � · � is strictly convex if for all X , Y , � X � = � Y � = 1 , � X + Y � = 2 ⇒ X = Y . Which matrix norms are strictly convex? Let σ 1 ≥ σ 2 ≥ · · · ≥ σ n be singular values of X and 1 ≤ p ≤ ∞ . � n � 1 /p � σ p The c p -norm: � X � p ≡ . i i =1 p = 2 . . . Frobenius norm, p = ∞ . . . spectral norm, matrix 2 -norm, � X � ∞ = σ 1 , p = 1 . . . trace (nuclear) norm. 7

  13. Strictly convex norms The norm � · � is strictly convex if for all X , Y , � X � = � Y � = 1 , � X + Y � = 2 ⇒ X = Y . Which matrix norms are strictly convex? Let σ 1 ≥ σ 2 ≥ · · · ≥ σ n be singular values of X and 1 ≤ p ≤ ∞ . � n � 1 /p � σ p The c p -norm: � X � p ≡ . i i =1 p = 2 . . . Frobenius norm, p = ∞ . . . spectral norm, matrix 2 -norm, � X � ∞ = σ 1 , p = 1 . . . trace (nuclear) norm. Theorem. If 1 < p < ∞ then the c p -norm is strictly convex. [see, e.g., Zie ¸tak, 1988] 7

  14. Spectral norm (matrix 2 -norm) A useful matrix norm in many applications: spectral norm � X � ≡ σ 1 . 8

  15. Spectral norm (matrix 2 -norm) A useful matrix norm in many applications: spectral norm � X � ≡ σ 1 . This norm is not strictly convex: � � � � I I X = Y = ε, δ ∈ � 0 , 1 � . , , ε δ Then we have, for each ε, δ ∈ � 0 , 1 � , � X � = � Y � = 1 and � X + Y � = 2 but if ε � = δ then X � = Y . 8

  16. Spectral norm (matrix 2 -norm) A useful matrix norm in many applications: spectral norm � X � ≡ σ 1 . This norm is not strictly convex: � � � � I I X = Y = ε, δ ∈ � 0 , 1 � . , , ε δ Then we have, for each ε, δ ∈ � 0 , 1 � , � X � = � Y � = 1 and � X + Y � = 2 but if ε � = δ then X � = Y . Consequently: Best approximation problems in the spectral norm are not guaranteed to have a unique solution. 8

  17. Matrix approximation problems in spectral norm M ∈ A � B − M � = � B − A ∗ � min A ∗ ∈ A achieving the minimum is called a spectral approximation of B from the subspace A . Open question : When does this problem have a unique solution? 9

  18. Matrix approximation problems in spectral norm M ∈ A � B − M � = � B − A ∗ � min A ∗ ∈ A achieving the minimum is called a spectral approximation of B from the subspace A . Open question : When does this problem have a unique solution? Zi¸ etak’s sufficient condition ¸tak, 1993] . If the residual matrix B − A ∗ has an n -fold Theorem [Zie maximal singular value, then the spectral approximation A ∗ of B from the subspace A is unique. 9

  19. Matrix approximation problems in spectral norm M ∈ A � B − M � = � B − A ∗ � min A ∗ ∈ A achieving the minimum is called a spectral approximation of B from the subspace A . Open question : When does this problem have a unique solution? Zi¸ etak’s sufficient condition ¸tak, 1993] . If the residual matrix B − A ∗ has an n -fold Theorem [Zie maximal singular value, then the spectral approximation A ∗ of B from the subspace A is unique. Is this sufficient condition satisfied, e.g., for the ideal Arnoldi approximation problem? 9

  20. General characterization of spectral approximations General characterization by [Lau and Riha, 1981] and [Zie ¸tak, 1993, 1996] → based on the Singer’s theorem [Singer, 1970] . 10

  21. General characterization of spectral approximations General characterization by [Lau and Riha, 1981] and [Zie ¸tak, 1993, 1996] → based on the Singer’s theorem [Singer, 1970] . Define ||| · ||| (trace norm, nuclear norm, c 1 -norm) and �· , ·� by � Z , X � ≡ tr( Z ∗ X ) . ||| X ||| = σ 1 + · · · + σ n , 10

  22. General characterization of spectral approximations General characterization by [Lau and Riha, 1981] and [Zie ¸tak, 1993, 1996] → based on the Singer’s theorem [Singer, 1970] . Define ||| · ||| (trace norm, nuclear norm, c 1 -norm) and �· , ·� by � Z , X � ≡ tr( Z ∗ X ) . ||| X ||| = σ 1 + · · · + σ n , Characterization : [Zie ¸tak, 1996] A ∗ ∈ A is a spectral approximation of B from the subspace A iff there exists Z ∈ C n × n , s.t. ||| Z ||| = 1 , � Z , X � = 0 , ∀ X ∈ A , and Re � Z , B − A ∗ � = � B − A ∗ � . 10

  23. Chebyshev polynomials of Jordan blocks Theorem. Let J λ be the n × n Jordan block. Consider the ideal Arnoldi approximation problem p ∈M m � p ( J λ ) � = min min M ∈ A � B − M � , where B = J m λ , A = span { I , J λ , . . . , J m − 1 } . The minimum is λ attained by the polynomial p ∗ = ( z − λ ) m [Liesen and T., 2008] . 11

  24. Chebyshev polynomials of Jordan blocks Theorem. Let J λ be the n × n Jordan block. Consider the ideal Arnoldi approximation problem p ∈M m � p ( J λ ) � = min min M ∈ A � B − M � , where B = J m λ , A = span { I , J λ , . . . , J m − 1 } . The minimum is λ attained by the polynomial p ∗ = ( z − λ ) m [Liesen and T., 2008] . Proof. For p = ( z − λ ) m , the residual matrix B − M is given by B − M = p ( J λ ) = ( J λ − λ I ) m = J m 0 . 11

  25. Chebyshev polynomials of Jordan blocks Theorem. Let J λ be the n × n Jordan block. Consider the ideal Arnoldi approximation problem p ∈M m � p ( J λ ) � = min min M ∈ A � B − M � , where B = J m λ , A = span { I , J λ , . . . , J m − 1 } . The minimum is λ attained by the polynomial p ∗ = ( z − λ ) m [Liesen and T., 2008] . Proof. For p = ( z − λ ) m , the residual matrix B − M is given by B − M = p ( J λ ) = ( J λ − λ I ) m = J m 0 . Define Z ≡ e 1 e T m +1 . It holds that � Z , J k ||| Z ||| = 1 , λ � = 0 , k = 0 , . . . , m − 1 and � Z , B − M � = � Z , J m 0 � = 1 = � B − M � 11

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend