Blended Conditional Gradients: The unconditioning of conditional - - PowerPoint PPT Presentation

blended conditional gradients the unconditioning of
SMART_READER_LITE
LIVE PREVIEW

Blended Conditional Gradients: The unconditioning of conditional - - PowerPoint PPT Presentation

Blended Conditional Gradients: The unconditioning of conditional gradients Joint work with Gabor Braun, Sebastian Pokutta, Steve Wright Dan Tu 1/9 CONDITIONAL GRADIENTS: PROJECTION-FREE Given a polytope ! , solve the optimization problem: min


slide-1
SLIDE 1

1/9

Joint work with Gabor Braun, Sebastian Pokutta, Steve Wright

Dan Tu

Blended Conditional Gradients: The unconditioning of conditional gradients

slide-2
SLIDE 2

2/9

Problems: Ø LP Oracle can be computationally expensive. Ø The conditional gradient direction, as an approximation of the negative gradient, can be inefficient.

CONDITIONAL GRADIENTS: PROJECTION-FREE

Given a polytope !, solve the optimization problem: min % & s.t. & ∈ ! where the objective function % is smooth and strongly convex

Find a vertex through LP Oracle. Walk along the conditional gradient direction.

slide-3
SLIDE 3

3/9

BLENDED CONDITIONAL GRADIENT

!" !# !$ !% &' &'() &'()(" *+ &'()

Frank-Wolfe Phase

Once the progress over the simplex is too small, call the LP oracle to obtain a new vertex

Gradient Descent Phase

Perform gradient descent over the simplex (!", !#, !$) as long as it makes enough progress: ∇+ &' 0 !'

1234 − 6' 78 ≥ Φ.

slide-4
SLIDE 4

4/9

For a general simplex, decompose ! as a convex combination ! = ∑$%&

'

($)$, with ∑$%&

'

($ = 1 and ($ ≥ 0, - = 1, 2, … , 1 Treat ($ as variables à ! in a standard simplex with normal vector: 2 = (1, 1, … , 1)/ 1

GRADIENT DESCENT PHASE

SIMPLEX GRADIENT DESCENT ORACLE

slide-5
SLIDE 5

5/9

GRADIENT DESCENT PHASE

SIMPLEX GRADIENT DESCENT ORACLE

! "# Boundary −%& "# = ( + * ( * "#+,

If not acceptable

Perform line search on line segment between "# and !

Decompose −%& "#

−%& "# = ( + * ( ⊥ *

boundary point acceptable?

Set "#+, = ! if & "# ≥ &(!)

slide-6
SLIDE 6

6/9

BCG ALGORITHM

Gradient Descent Phase Frank-Wolfe Phase

slide-7
SLIDE 7

7/9

COMPUTATIONAL RESULTS

Fig 2: Sparse Signal Recovery

BCG outperforms several recent variants of Frank-Wolfe algorithm

Fig 1: Lasso Regression

slide-8
SLIDE 8

8/9

Theorem

If ! is a strongly convex and smooth function over the polytope " with geometric strong convexity # and simplicial curvature, then BCG algorithm ensures ! $% − ! x∗ ≤ * for some + that satisfies: + ≤ log /01

2

+ 8 log 01

/56 + 7856 9

log 856

2

= ;

56 9 log 01 2

CONVERGENCE

slide-9
SLIDE 9

9/9

THANKS!