ICML 2019 in Long Beach Model Function Based Conditional Gradient - - PowerPoint PPT Presentation

icml 2019 in long beach
SMART_READER_LITE
LIVE PREVIEW

ICML 2019 in Long Beach Model Function Based Conditional Gradient - - PowerPoint PPT Presentation

ICML 2019 in Long Beach Model Function Based Conditional Gradient Method with Armijo-like Line Search Peter Ochs Mathematical Optimization Group Saarland University 13.06.2019 joint work: Yura Malitsky 1 / 7 c 2019 Peter


slide-1
SLIDE 1

ICML 2019 in Long Beach

Model Function Based Conditional Gradient Method with Armijo-like Line Search

Peter Ochs Mathematical Optimization Group Saarland University — 13.06.2019 — joint work: Yura Malitsky

c 2019 — Peter Ochs Model Based Conditional Gradient

1 / 7

slide-2
SLIDE 2

Classic Conditional Gradient Method

Constrained Smooth Optimization Problem: min

x∈C f(x)

◮ C ⊂ RN compact and convex constraint set Conditional Gradient Method: Update step: y(k) ∈ argmin

y∈C

  • ∇f(x(k)), y
  • x(k+1) = γky(k) + (1 − γk)x(k)

Convergence mainly relies on: ◮ step size γk ∈ [0, 1] (we consider Armijo line search) ◮ Descent Lemma (implies curvature condition)

c 2019 — Peter Ochs Model Based Conditional Gradient

2 / 7

slide-3
SLIDE 3

Generalizing the Descent Lemma

Descent Lemma: |f(x) − f(¯ x) − ∇f(¯ x), x − ¯ x | ≤ L

2 x − ¯

x2 provides a measure for the linearization error quadratic growth ◮ f smooth non-convex ◮ L is the Lipschitz constant of ∇f

c 2019 — Peter Ochs Model Based Conditional Gradient

3 / 7

slide-4
SLIDE 4

Generalizing the Descent Lemma

Generalization of the Descent Lemma: |f(x) − f(¯ x) − ∇f(¯ x), x − ¯ x | ≤ ω(x − ¯ x) provides a measure for the linearization error growth given by ω ◮ f smooth non-convex ◮ ω: R+ → R+ is a growth function

c 2019 — Peter Ochs Model Based Conditional Gradient

3 / 7

slide-5
SLIDE 5

Generalizing the Descent Lemma

Generalization of the Descent Lemma: |f(x) − f¯

x(x)| ≤ ω(x − ¯

x) provides a measure for the approximation error growth given by ω ◮ f non-smooth non-convex ◮ ω: R+ → R+ is a growth function

c 2019 — Peter Ochs Model Based Conditional Gradient

3 / 7

slide-6
SLIDE 6

Model Assumption |f(x) − f¯

x(x)| ≤ ω(x − ¯

x)

f(x) f(x) + ω(x − ¯ x) f(x) − ω(x − ¯ x) ¯ x

c 2019 — Peter Ochs Model Based Conditional Gradient

4 / 7

slide-7
SLIDE 7

Model Function based Conditional Gradient Method

Model Function based Conditional Gradient Method:

y(k) ∈ argmin

y∈C

fx(k)(y) x(k+1) = γky(k) + (1 − γk)x(k) Examples for Model Assumption: |f(x) − f¯

x(x)| ≤ ω(x − ¯

x) ◮ additive composite problem: min

x∈C {f(x) =

g(x)

non-smooth

+ h(x)

smooth

} ◮ model function: f¯

x(x) = g(x) + h(¯

x) + ∇h(¯ x), x − ¯ x ◮ oracle: argmin

y∈C

g(y) +

  • ∇h(x(k)), y
  • c

2019 — Peter Ochs Model Based Conditional Gradient

5 / 7

slide-8
SLIDE 8

Examples

Examples for Model Assumption: |f(x) − f¯

x(x)| ≤ ω(x − ¯

x) ◮ hybrid Proximal–Conditional Gradient, example: min

x1∈C1 x2∈C2

{f(x1, x2) = g(x1)

non-smooth

+ h(x2)

smooth

} ◮ f¯

x(x1, x2) = h(¯

x2) + ∇h(¯ x2), x2 − ¯ x2 + g(x1) + 1

2λx1 − ¯

x12 ◮ oracle:      argmin

y1∈C1

g(y1) + 1

2λy1 − x(k) 1 2

argmin

y2∈C2

  • ∇h(x(k)

2 ), y2

  • c

2019 — Peter Ochs Model Based Conditional Gradient

6 / 7

slide-9
SLIDE 9

Examples

◮ composite problem ◮ second order Conditional Gradient Design model functions for your problem!

c 2019 — Peter Ochs Model Based Conditional Gradient

7 / 7