robust control for analysis and design
- f large-scale optimization algorithms
robust control for analysis and design of large-scale optimization - - PowerPoint PPT Presentation
robust control for analysis and design of large-scale optimization algorithms Laurent Lessard University of WisconsinMadison Joint work with Ben Recht and Andy Packard LCCC Workshop on Large-Scale and Distributed Optimization Lund
2
3
x0 x1
20 40 60 80 100 10−2 10−4 Error
4
G∈G
f∈S cost(f, G)
5
2xTQx − pTx
6
100 101 102 103 104 0.2 0.4 0.6 0.8 1 Condition ratio L/m Convergence rate ρ Convergence rate on quadratic functions Gradient (quadratic) Nesterov (quadratic) Heavy ball (quadratic) 100 101 102 103 104 100 101 102 103 1 1/2 Condition ratio L/m Iterations to convergence Iterations to convergence for Gradient method
7
G∈G
f∈S cost(f, G)
8
algorithm (linear, known, decoupled) function (nonlinear, uncertain, coupled)
9
algorithm (linear, known, decoupled) function (nonlinear, uncertain, coupled)
10
−α 1
1 + β −β −α 1 1
1 + β −β −α 1 1 + β −β
11
∇ f y u
∇ f(x) x ∇ f(x) x ∇ f(x) x
∇ f(x) :
linear sector bounded + slope restricted sector bounded
x f(x) x f(x) x
f(x) :
quadratic strongly convex + Lipschitz gradients radially quasiconvex
∇ f y u
uk yk sector bounded
13
∇ f y u
m L uk yk sector bounded
14
∇ f y u
m L uk yk sector bounded + slope restricted
∇ f is sector-bounded + slope-restricted: constraint on (yk, uk) depends on history (y0, . . . , yk−1, u0, . . . , uk−1).
15
∇ f y u Ψ
ζ
z
k Mzk.
16
−α 1
−β −α 1 1
−β −α 1 1 + β −β
f(x) x ∇ f(x) x ∇ f(x) x
f is quadratic f is strongly convex f is quasiconvex
17
18
19
100 101 102 103 104 0.2 0.4 0.6 0.8 1 Condition ratio L/m Convergence rate ρ Convergence rate for Gradient method Gradient (all functions) Nesterov (quadratic) Heavy ball (quadratic) 100 101 102 103 104 100 101 102 103 1 1/2 Condition ratio L/m Iterations to convergence Iterations to convergence for Gradient method
20
100 101 102 103 104 0.2 0.4 0.6 0.8 1 Condition ratio L/m Convergence rate ρ Nesterov rate bounds IQC (quasiconvex) IQC (strongly convex) Nesterov (strongly convex) Nesterov (quadratic) Heavy ball (quadratic) 100 101 102 103 104 10−1 100 101 102 103 Condition ratio L/m Iterations to convergence Nesterov iterations
21
100 101 102 103 104 0.2 0.4 0.6 0.8 1 Condition ratio L/m Convergence rate ρ Heavy ball rate bounds IQC (quasiconvex) IQC (strongly convex) Nesterov (quadratic) Heavy ball (quadratic) 100 101 102 103 104 10−1 100 101 102 103 Condition ratio L/m Iterations to convergence Heavy ball iterations
22
25 2 x2
1 2x2 + 24x − 12
25 2 x2 − 24x + 36
−2 2 4 20 40 60 80 f(x)
23
24
G
ξ
∇ f ∆δ w u y
25
2 L+m (optimal stepsize with no noise)
100 101 102 103 0.2 0.4 0.6 0.8 1 Convergence rate ρ
Rates for different δ δ ∈ {.01, .02, .05, .1, .2, .5} Gradient method 100 101 102 103 10−1 100 101 102 103 Iterations to convergence Iterations for different δ δ ∈ {.01, .02, .05, .1, .2, .5} Gradient method
L (more conservative stepsize)
100 101 102 103 0.2 0.4 0.6 0.8 1 Convergence rate ρ
Rates for different δ δ ∈ {.01, .02, .05, .1, .2, .5} Gradient method 100 101 102 103 10−1 100 101 102 103 Iterations to convergence Iterations for different δ δ ∈ {.01, .02, .05, .1, .2, .5} Gradient method
26
100 101 102 103 104 0.2 0.4 0.6 0.8 1 Condition ratio L/m Convergence rate ρ
Rates for different δ δ ∈ {.05, .1, .2, .3, .4, .5} Nesterov (quadratic) 100 101 102 103 104 10−1 100 101 102 103 Condition ratio L/m Iterations to convergence Iterations for different δ δ ∈ {.05, .1, .2, .3, .4, .5} Nesterov (quadratic)
27
28
100 101 102 103 0.2 0.4 0.6 0.8 1 Condition ratio L/m Convergence rate ρ
Rates for different δ δ ∈ {.01, .1, .2, .3, .4, .5} Nesterov (quadratic) 100 101 102 103 10−1 100 101 102 103 Condition ratio L/m Iterations to convergence Iterations for different δ δ ∈ {.01, .1, .2, .3, .4, .5} Gradient Nesterov (quadratic)
29
G ∇ f ∇ g x proxλg(x) λ ∂g
+ −
30
31
32