extrapolation techniques and multiparameter treatment for
play

Extrapolation techniques and multiparameter treatment for Tikhonov - PowerPoint PPT Presentation

Extrapolation techniques and multiparameter treatment for Tikhonov regularization A review Michela Redivo Zaglia University of Padua - Italy Joint work with C. Brezinski (Lille-France) G. Rodriguez, S. Seatzu (Cagliari-Italy) Tikhonov


  1. Extrapolation techniques and multiparameter treatment for Tikhonov regularization A review Michela Redivo Zaglia University of Padua - Italy Joint work with C. Brezinski (Lille-France) G. Rodriguez, S. Seatzu (Cagliari-Italy) ➜ Tikhonov regularization ➜ Extrapolation techniques ➜ Multiparameter regularization 1

  2. A N APPLICATION TO REGULARIZATION When a p × p system Ax = b is ill-conditioned , its solution cannot be computed accurately. Tikhonov’s regularization consists in computing the vector x λ which minimizes the quadratic functional J ( λ, x ) = � Ax − b � 2 + λ � Hx � 2 over all vectors x , where λ ≥ 0 is a parameter, and H a given q × p ( q ≤ p ) matrix. � · � denotes the Euclidean norm. This vector x λ is the solution of the system ( C + λE ) x λ = A T b, where C = A T A and E = H T H . A N APPLICATION TO REGULARIZATION 2

  3. If λ is close to zero, then x λ is badly computed while, if λ is far away from zero, x λ is well computed but the norm of the error � x − x λ � is quite large. For decreasing values of λ , the norm of the error � x − x λ � first decreases, and then increases when λ approaches 0. Thus the norm of the error , which is the sum of the theoretical error and the error due to the computer’s arithmetic, passes through a minimum corresponding to the optimal choice of the regularization parameter λ . A N APPLICATION TO REGULARIZATION 3

  4. Several methods have been proposed to obtain an effective choice of λ (i.e. L -curve, Morozov discrepancy principle, GCV , . . . ). But each of these methods can fail. Idea: compute x λ for several values of λ , interpolate by a vector function of λ which mimics the exact behaviour, and then extrapolate at λ = 0 . The main problem is to use a conveniently chosen vector function. For that purpose, let us study the exact behaviour of x λ with respect to λ . A N APPLICATION TO REGULARIZATION 4

  5. We assume that H is a p × p non singular matrix, and we set y = Hx . Hence J ( λ, x ) = � AH − 1 y − b � 2 + λ � y � 2 Using the SVD of AH − 1 = U Σ V T , it can be proved that x λ = H − 1 y λ with p � σ i γ i y λ = i + λv i , and γ i = ( u i , b ) . σ 2 i =1 So, we will choose an interpolation function of the same form but with a sum running only from i = 1 to k , with k < p . Several extrapolation methods based on this idea exist. They depend whether the vectors v i are known or not . We assume, without loss of generality, that H = I (i.e. x λ = y λ ). A N APPLICATION TO REGULARIZATION 5

  6. E XTRAPOLATION TECHNIQUES R ESTRICTED CASE : If the vectors v 0 , . . . , v k are known , we interpolate by the rational function k � a i R k ( λ ) = b i + λ v i , i =1 where the a i ’s and the b i ’s are 2 k unknown scalars . For determining the parameters a i and b i , we impose the interpolation conditions ( x n = x λ n ) k � a i x n = v i b i + λ n i =1 k � a i x n +1 = v i . b i + λ n +1 i =1 E XTRAPOLATION TECHNIQUES 6

  7. Then, extrapolating at λ = 0 will give k � a j v j = y ( n ) x ≃ R k (0) = k . b j j =1 We assume that vectors w 1 , . . . , w k such that ( v i , w j ) = δ ij are known , and we obtain k � λ n +1 − λ n y ( n ) = ( x n , w j )( x n +1 , w j ) λ n +1 ( x n +1 , w j ) − λ n ( x n , w j ) v j . k j =1 Increasing the value of k , for n fixed we also have + a k +1 y ( n ) k +1 = y ( n ) v k +1 , k = 0 , 1 , . . . , p − 1 . k b k +1 E XTRAPOLATION TECHNIQUES 7

  8. If ( v i , v j ) = δ ij and w j = v j , the process is equivalent to the Truncated SVD (TSVD) , that is k � γ i y ( n ) = v i , γ i = ( u i , b ) . k σ i i =1 In this case, we can drop the superscript n since y ( n ) does not k depend on n . We set e k = x − y k and r k = b − Ay k . It holds � e k � 2 − γ 2 � e k +1 � 2 k +1 /σ 2 = k +1 � r k � 2 − γ 2 � r k +1 � 2 = k +1 . E XTRAPOLATION TECHNIQUES 8

  9. We have also the following identities, ∀ k , k � γ 2 � y k � 2 i = σ 2 i i =1 � y k � 2 + γ 2 k +1 � y k +1 � 2 = σ 2 k +1 p � γ 2 � e k � 2 i = σ 2 i i = k +1 ( y k , e k ) = ( e k − 1 − e k , e k ) = 0 � y k � 2 + � e k � 2 � x � 2 = p � � r k � 2 γ 2 = i . i = k +1 E XTRAPOLATION TECHNIQUES 9

  10. E XTRAPOLATION TECHNIQUES F ULL CASE : If the vectors v i are unknown , we base the extrapolation process of a rational function of the form k � 1 R k ( λ ) = b i + λ w i , k ≤ p, i =1 where the b i ’s are unknown numbers and the w i are unknown vectors . They are determined, as before, by imposing that x n = R k ( λ n ) for some values of n . Then we extrapolate at the point λ = 0 . E XTRAPOLATION TECHNIQUES 10

  11. It corresponds to computing k � 1 x ≃ y k = R k (0) = w i . b i i =1 R k can be written as R k ( λ ) = P k − 1 ( λ ) /Q k ( λ ) with α 0 + · · · + α k − 1 λ k − 1 , α i ∈ R p P k − 1 ( λ ) = k � ( b i + λ ) = β 0 + · · · + β k − 1 λ k − 1 + λ k , Q k ( λ ) = β i ∈ R . i =1 We gave 6 different algorithms for determining the two unknowns needed α 0 and β 0 ( R k (0) = α 0 /β 0 ). E XTRAPOLATION TECHNIQUES 11

  12. Let us describe one of them (the most satisfying one). We have to solve the interpolation problem Q k ( λ i ) x i = P k − 1 ( λ i ) , for i = 0 , . . . , k − 1 . Since Q k and P k − 1 are polynomials, we have, by Lagrange’s formula k � Q k ( λ ) = L i ( λ ) Q k ( λ i ) i =0 k − 1 � � P k − 1 ( λ ) = L i ( λ ) P k − 1 ( λ i ) i =0 k − 1 � � = L i ( λ ) Q k ( λ i ) x i i =0 E XTRAPOLATION TECHNIQUES 12

  13. with k k − 1 � � λ − λ j λ − λ j � L i ( λ ) = and L i ( λ ) = . λ i − λ j λ i − λ j j =0 j =0 j � = i j � = i Let λ k � = λ j , for j = 0 , . . . , k − 1 . We have k − 1 � � L i ( λ k ) Q k ( λ i ) x i = Q k ( λ k ) x k . i =0 Let s 1 , . . . , s p be linearly independent vectors . Setting c i = Q k ( λ i ) /Q k ( λ k ) and multiplying scalarly the preceding equation by s j , for j = 1 , . . . , p , leads to the following linear system k − 1 � � L i ( λ k )( x i , s j ) c i = ( x k , s j ) , j = 1 , . . . , p. i =0 Solving this system in the least squares sense gives c 0 , . . . , c k − 1 . E XTRAPOLATION TECHNIQUES 13

  14. Since the polynomial Q k ( λ ) = � k i =0 L i ( λ ) Q k ( λ i ) is monic and c k = 1 , we have a supplementary condition which gives the value Q k ( λ k ) . Thus Q k ( λ i ) = c i Q k ( λ k ) , and, finally, β 0 = Q k (0) is given by k � β 0 = L i (0) Q k ( λ i ) . i =0 From what precedes, we see that k − 1 � � α 0 = P k − 1 (0) = L i (0) Q k ( λ i ) x i i =0 and it follows that y k = R k (0) = P k − 1 (0) = β 0 . Q k (0) α 0 E XTRAPOLATION TECHNIQUES 14

  15. N UMERICAL EXAMPLES : A wide numerical experimentation has been performed. We used several kind of matrices A , heat, ilaplace, shaw, spikes Hansen Matlab Regularization Toolbox hilbert, lotkin, moler Matlab (gallery function) different solutions x t ( t means true solution ), given defined as in the Regularization Toolbox ones x i = 1 lin x i = i � � p �� 2 , quad x i = i − i = 1 , . . . , p � � 2 sin x i = sin 2 π ( i − 1) /p i = 1 , . . . , p various matrices H , I Identity matrix D 1 , D 2 , D 3 discrete approximations of the first, second and third derivative and we also try the case of a noised data vector. E XTRAPOLATION TECHNIQUES 15

  16. The tests show the effectiveness of the procedures , but that the best approximation , denoted by x opt , depends on the values of λ n chosen for interpolating and that the norm of the error � x t − x opt � can be strongly influenced by the choice of the regularizating matrix H . E XTRAPOLATION TECHNIQUES 16

  17. M ULTIPARAMETER REGULARIZATION A good choice of the matrix H often depends on the mathematical properties of the solution. Using several regularization terms avoids this difficult choice. We are looking for the vector x λ which minimizes the quadratic functional � � k � � Ax − b � 2 + λ i � H i x � 2 J ( λ, x ) = k , i =1 with λ = ( λ 1 , . . . , λ k ) T . M ULTIPARAMETER REGULARIZATION 17

  18. It is also the solution of the system � � k � x λ = A T b, C + λ i E i i =1 with C = A T A and E i = H T i H i . We have the following relation between x λ and x � � k � λ i C − 1 E i I + x λ = x. i =1 We note that x λ is also the vector minimizing k � � � Ax − b � 2 + kλ i � H i x � 2 � J ( λ, x ) = . i =1 M ULTIPARAMETER REGULARIZATION 18

  19. Hence, we consider the k vectors x λ i solving the minimization problems � � Ax − b � 2 + kλ i � H i x � 2 � min , i = 1 , . . . , k. x These vectors satisfy the linear systems ( C + kλ i E i ) x λ i = A T b, i = 1 , . . . , k. Thus we compute k one-parameter regularized solutions , and, after, we consider the approximation of x λ given by the linear combination k � x λ ( α ) = � α i x λ i , i =1 where α = ( α 1 , . . . , α k ) T and � k i =1 α i = 1 . M ULTIPARAMETER REGULARIZATION 19

  20. How to choose the vector α ? The following relation between x λ and � x λ ( α ) holds � � − 1 k � x λ = � x λ ( α ) + C + λ i E i ρ λ ( α ) , i =1 where   k k � �  kλ i E i −  x λ i . ρ λ ( α ) = α i λ j E j i =1 j =1 The vector α is chosen to minimize ρ λ ( α ) . It is the solution of an overdetermined system which is solved in the least-squares sense . M ULTIPARAMETER REGULARIZATION 20

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend