SLIDE 19 RK vs Conjugate Gradient
We compare serial implementations of RK and CG. (The benefits of multicore implementation are similar for both.) Random A, δ = .1.
0.5 1 1.5 2 2.5 3 x 10
8
10
−8
10
−6
10
−4
10
−2
10
m=1000, n=500, λmin(ATA)=0.06937, λmax(ATA)=6.156
# of Operations ||Ax−b|| CG RK
0.5 1 1.5 2 2.5 3 x 10
8
10
−14
10
−12
10
−10
10
−8
10
−6
10
−4
10
−2
10
m=2000, n=500, λmin(ATA)=0.5616, λmax(ATA)=10.7
# of Operations ||Ax−b|| CG RK
CG does better in the more ill-conditioned case, probably due to nice distribution of dominant eigenvalues of ATA. (Note slower convergence in later stages.) RK is competitive in the well-conditioned case.
Wright (UW-Madison) Asynchronous Stochastic Optimization September 2014 19 / 44