iterative methods for image processing lothar reichel
play

Iterative methods for Image Processing Lothar Reichel Como, May - PowerPoint PPT Presentation

Iterative methods for Image Processing Lothar Reichel Como, May 2018. Lecture 2: Tikhonov regularization and truncated SVD for large-scale problems. Outline of Lecture 2: Small to moderately-sized problems Tikhonov regularization in


  1. Iterative methods for Image Processing Lothar Reichel Como, May 2018.

  2. Lecture 2: Tikhonov regularization and truncated SVD for large-scale problems. Outline of Lecture 2: • Small to moderately-sized problems – Tikhonov regularization in standard form – Tikhonov regularization in general form – The generalized SVD • Large-scale problems – Tikhonov regularization based on Krylov subspace methods – Truncated SVD for large-scale problems

  3. Tikhonov regularization Solve the minimization problem x {� Ax − b � 2 2 + µ � Lx � 2 min 2 } , where µ > 0 is a regularization parameter (to be determined) and L ∈ R p × n is a regularization matrix chosen so that N ( A ) ∩ N ( L ) = { 0 } . Then the minimization problem has a unique solution for any µ > 0.

  4. Common choices of L : identity, discretizations of differential operator. In our applications A is a smoothing operator. Therefore, the Tikhonov minimization problem generally has a unique solution when L is a discrete differential operator. We would like L be such that important features of x exact are not damped. This is the case when they are in N ( L ).

  5. The normal equations associated with the Tikhonov minimization problem ( A T A + µL T L ) x = A T b have the unique solution x µ := ( A T A + µL T L ) − 1 A T b for any µ > 0. Generally, µ ց 0 x µ = A † b, lim µ →∞ x µ = 0 . lim Neither x 0 nor x ∞ are useful approximations of x exact . A proper choice of the value of µ is important. It involves computing x µ repeatedly for different µ -values. May be expensive.

  6. The discrepancy principle Assume that a fairly accurate estimate for δ := � b − b exact � 2 is known. The discrepancy principle prescribes that µ > 0 be chosen so that � Ax µ − b � 2 = ηδ for some constant η > 1 independent of δ . The computation of such a µ -value requires solution of the Tikhonov minimization problem for several values of µ .

  7. Methods for repeated Tikhonov minimization Assume that A ∈ R m × n is small and let L = I . Compute the SVD of A , A = U Σ V T , where U ∈ R m × m and V ∈ R n × n are orthogonal, and Σ = diag[ σ 1 , σ 2 , . . . , σ n ] ∈ R m × n with σ 1 ≥ σ 2 ≥ . . . ≥ σ n ≥ 0. The Tikhonov solution is given by x µ = V (Σ T Σ + µI ) − 1 Σ T U T b. The evaluation of � Ax µ − b � 2 requires only O ( m ) flops for every µ -value (without forming Ax µ ).

  8. The Generalized SVD (GSVD) Assume that A ∈ R m × n and L ∈ R p × n are small. (Here L � = I ). The GSVD of the matrix pair { A, L } are the factorizations A = U Σ X T , L = V MX T , where U ∈ R m × m and V ∈ R p × p are orthogonal, X ∈ R n × n is nonsingular, and Σ and M are diagonal.

  9. When m ≥ n ≥ p , diag[ σ 1 , σ 2 , . . . , σ p , 1 , 1 , . . . , 1] ∈ R m × n , Σ = [diag[ µ 1 , µ 2 , . . . , µ p ] , 0 , 0 , . . . , 0] ∈ R p × n , M = 0 ≤ σ 1 ≤ σ 2 ≤ . . . ≤ σ p ≤ 1 , 1 ≥ µ 1 ≥ µ 2 ≥ . . . ≥ µ p ≥ 0 , σ 2 j + µ 2 1 ≤ j ≤ p. j = 1 , The Tikhonov solution is given by x µ = X − T (Σ T Σ + µM T M ) − 1 Σ T U T b. The evaluation of � Ax µ − b � 2 requires only O ( m ) flops for every µ -value (without evaluating Ax µ ).

  10. When the matrices A and L are large, the computation of the SVD of A or GSVD of the matrix pair { A, L } is expensive. When A, L ∈ R n × n then, roughly, • the computation of the SVD of A requires about 10 n 3 flops, and • the computation of the GSVD of { A, L } requires about 25 n 3 flops. Therefore, the evaluation of these decompositions is impractical for large-scale problems.

  11. Methods for large-scale problems Zha described an iterative method for determining a few vectors of the GSVD of a pair of large matrices { A, L } . Kilmer, Hansen, and Espa˜ nol apply this method to Tikhonov regularization. Some properties: • It is an inner-outer iterative method. Generalized singular vectors are computed in the inner iteration. • Zha’s method may require fairly many iterations. We are interested in developing methods that require only few matrix-vector product evaluations with A .

  12. Application of standard Krylov subspace methods The Arnoldi process: Application of k steps to A ∈ R n × n with initial vector b gives the Arnoldi decomposition AV k = V k +1 H k +1 ,k , where the orthonormal columns of V k ∈ R n × k span the Krylov subspace K k ( A, b ) = span { b, Ab, A 2 b, . . . , A k − 1 b } with V k e 1 = b/ � b � 2 and H k +1 ,k ∈ R ( k +1) × k upper Hessenberg.

  13. We solve x ∈ K k ( A,b ) {� Ax − b � 2 2 + µ � Lx � 2 min 2 } by using the QR factorization LV k = Q k R k , where Q k ∈ R n × k has orthonormal columns and R k ∈ R k × k is upper triangular. Let x = V k y . Then y ∈ R k {� H k +1 ,k y − e 1 � b � 2 � 2 2 + µ � R k y � 2 2 } . min This reduced problem can be solved by using the GSVD of { H k +1 ,k , R k } .

  14. Some remarks: • The Arnoldi process can be replaced by a range restricted Arnoldi process that generates an orthonormal basis for the solution subspace K k ( A, A j b ) = span { A j b, A j +1 b, A j +2 b, . . . , A j + k − 1 b } . Typically, j = 1 or j = 2. • The Arnoldi process can be replaced by some other Krylov subspace method for reducing A , such as Golub–Kahan bidiagonalization. • The solution subspace is independent of L . For some problems this is a disadvantage.

  15. Reduction methods for matrix pairs { A, L } Reduction method by Li and Ye: Generalizes the Arnoldi process to matrix pairs: V 2 k H ( A ) AV k = 2 k,k , V 2 k +1 H ( L ) LV k = 2 k +1 ,k , where V 2 k +1 has orthonormal columns with V 2 k +1 e 1 = b/ � b � . The matrices H ( A ) 2 k,k and H ( L ) 2 k +1 ,k are upper “super Hessenberg”.

  16. Example: Matrices for k = 4.     * * * *   * * * *     * * * *       * * * *       * * * *       * * *       * * *       * * *   H ( A ) H ( L )   8 , 4 = , 9 , 4 = .   * * *     * *         * *     * *       * *       *       *   * *

  17. Solution subspace R ( V k ) generated by the Li–Ye method with initial vector b is of the form K k ( A, L, b ) = span { b, Ab, Lb, A 2 b, LAb, ALb, L 2 b, A 3 b, LA 2 b, ALAb, A 2 Lb, LALb, AL 2 b, L 3 b, . . . } The method alternatingly evaluates a matrix-vector product with A and a matrix-vector product with L . Generalized Arnoldi process for matrix pairs { A, L } :

  18. Given q 1 with � q 1 � = 1 ; 1. 2. N := 1; 3. for j = 1 , 2 , . . . , k do 4. if j > N then exit; 5. q := Aq j ; ˆ 6. for i = 1 , 2 , . . . N do h A ; i,j := q T q − q i h A ; i,j ; 7. i ˆ q ; ˆ q := ˆ 8. end for 9. h A ; N +1 ,j := � ˆ q � ; 10. if h A ; N +1 ,j > 0 then 11. N := N + 1; q N := ˆ q/h A ; N,j ; 12. end if

  19. 13. q := Lq j ; ˆ 14. for i = 1 , 2 , . . . N do h L ; i,j := q T q − q i h L ; i,j ; 15. i ˆ q ; ˆ q := ˆ 16. end for 17. h L ; N +1 ,j := � ˆ q � ; 18. If h L ; N +1 ,j > 0 then 19. N := N + 1; q N := ˆ q/h A ; N,j ; 20. end if 21. end for

  20. The scalar N in the algorithm tracks the number of vectors q i generated so far during the computations. Let α k and β k denote the values of N at the end of Lines 12 and 20, respectively, when j = k . AQ (: , 1: k ) = Q (: , 1: α k ) H A (1: α k , 1: k ) , LQ (: , 1: k ) = Q (: , 1: β k ) H L (1: β k , 1: k ) ;

  21. We solve x ∈ K k ( A,L,b ) {� Ax − b � 2 2 + µ � Lx � 2 min 2 } by using the generalized Arnoldi decompositions. Let x = V k y . Then we obtain the reduced problem y ∈ R k {� H ( A ) 2 + µ � H ( L ) 2 k,k y − e 1 � b � 2 � 2 2 k +1 ,k y � 2 2 } . min It can be solved by the GSVD.

  22. Example: We would like to determine the unavailable noise-free image represented by 412 × 412 pixels.

  23. The entries of the vector b ∈ R 412 2 store the pixel values, ordered column-wise, of the available blur- and noise-contaminated image.

  24. The blurring matrix A ∈ R 412 2 × 412 2 represents severe Gaussian blur. The image also has been contaminated by 30% Gaussian noise. We apply the Li–Ye method to solve x ∈ K k l ( A,L,b ) {� Ax − b � 2 2 + µ � Lx � 2 2 } min for two different regularization matrices L : • L = ∆, the standard discrete Laplace operator based on the five-point stencil. • L is a discretized and linearized Perona–Malik operator: 1 L ( x ) = div( g ( |∇ x | 2 ) ∇ x ) , ρ = 10 − 4 . g ( s ) = , 1 + s ρ

  25. Restored image using L = ∆. 6 generalized Arnoldi steps.

  26. Restored image with L determined by the Perona–Malik operator. Two step of GMRES give an approximate restoration with which L is defined.

  27. Edge map for restoration with Perona–Malik operator.

  28. Some remarks: • To work well with the discrepancy principle, e 1 should be replaced by P R ( H ( A ) 2 ℓ,ℓ ) e 1 , i.e., � H ( A ) � Ax µ − b � 2 = 2 k,k y µ − e 1 � b � 2 � 2 � H ( A ) ≥ 2 k,k y µ − P R ( H ( A ) 2 k,k ) e 1 � b � 2 � 2 . The discrepancy principle is applied to the right-hand side. • The method requires the generation of about twice as many orthonormal vectors as the dimension of the solution subspace.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend