ECS 231 Gradient descent methods for solving large scale eigenvalue problems
1 / 17
ECS 231 Gradient descent methods for solving large scale eigenvalue - - PowerPoint PPT Presentation
ECS 231 Gradient descent methods for solving large scale eigenvalue problems 1 / 17 Generalized symmetric definite eigenvalue problem Generalized symmetric definite eigenvalue problem Au = Bu where A and B are n n symmetric, and B
1 / 17
◮ Generalized symmetric definite eigenvalue problem
◮ All eigenvalues and eigenvectors are real ◮ Denote the eigenvalues by λ1 ≤ λ2 ≤ · · · ≤ λn and their associated
◮ · B is defined through the B-inner product
2 / 17
◮ Rayleigh Quotient is defined by
◮ Minimization principle:
x ρ(x)
◮ In general for i > 1
x⊥Buj, 1≤j<i ρ(x)
3 / 17
◮ By the minimization principles, various optimization techniques can be
◮ Gradient (steepest) Decent (GD) and Conjugate Gradient (CG)
◮ Two useful quantities:
◮ The gradient of ρ(x):
◮ The Hessian of ρ(x):
4 / 17
◮ Minimizing the Rayleigh quotient along the direction of the gradient
t
5 / 17
◮ Given an approximation x to u1 and xB = 1 ◮ Compute the search direction:
◮ solve the problem
t ρ(x + tr) = ρ(x + toptr) ◮ update
6 / 17
0 Ax0, r0 = Ax0 − ρ0Bx0
i+1Axi+1, ri+1 = Axi+1 − ρi+1Bxi+1
7 / 17
8 / 17
◮ The case B = I, locally, the convergence rate is
[Faddeev and Faddeeva’63] and [Knyazev and Skorokhodov’91]: ◮ For the case B = I, [Yang’93].
z∈span{xi,ri} ρ(z)
z∈Km(A−ρiB,xi) ρ(z)
9 / 17
1 ), r(x(i) 2 ), . . . , r(x(i) k )]
Further reading: R.-C. Li, Rayleigh quotient based optimization methods for eigenvalue problems, in “Matrix Functions and Matrix Equations”, Z. Bai et al ed., World Scientific, 2015. 10 / 17
◮ The Conjugate Gradient (CG) method was originally proposed in
◮ In the 1960s, it was extended by Fletcher and Reeves as an iterative
◮ Because of the optimality properties of Rayleigh quotients, it is natural
11 / 17
◮ Define
◮ the gradient ∇φ(x) = Hx − b = r(x) (the residual vector)
◮ φ(x) is a quadratic functional in x. It is convex and has a unique local
◮ Given an initial guess x0, the CG method iteratively produces a
α φ(xi + αpi).
i Hpj = 0 for i = j,
12 / 17
i Ari
i Api
i+1ri+1
i ri
i+1Hpi = 0.
13 / 17
◮ In the absence of roundoff errors, it can be proved that
i rj = 0,
i Hpj = 0,
◮ The CG method converges in at most n steps, a direct method, is a
14 / 17
◮ In extending the CG method, the key is to recognize that the residual
◮ For the eigenproblem of A − λB, the objective function is the Rayleigh
15 / 17
0 Ax0, r0 = Ax0 − ρ0Bx0, p0 = r0;
i+1Axi+1, ri+1 = Axi+1 − ρi+1Bxi+1,
16 / 17
◮ Different choice of βi leads to the different version of the CG method.
i+1ri+1
i ri
i+1(ri+1 − ri)
i ri ◮ Choose βi, together with αi, to minimize the Rayleigh quotient on the
Ref.: A. Knyazev, Toward the optimal preconditioned eigensolver: locally optimal block preconditioned conjugate gradient method. SIAM J. Sci. Comput. 23(2):517-541, 2001
◮ Open problem: no quantitative estimate on the convergence rate of
17 / 17