Frank-Wolfe Algorithms for Saddle Point problems author: Gauthier - - PowerPoint PPT Presentation

frank wolfe algorithms for saddle point problems
SMART_READER_LITE
LIVE PREVIEW

Frank-Wolfe Algorithms for Saddle Point problems author: Gauthier - - PowerPoint PPT Presentation

Frank-Wolfe Algorithms for Saddle Point problems author: Gauthier Gidel, Supervisors: Simon Lacoste-Julien & Tony Jebara INRIA Paris, Sierra Team & Columbia University September 15 th 2016 September 15 th 2016 Gauthier Gidel, Simon


slide-1
SLIDE 1

Frank-Wolfe Algorithms for Saddle Point problems

author: Gauthier Gidel, Supervisors: Simon Lacoste-Julien & Tony Jebara INRIA Paris, Sierra Team & Columbia University September 15th 2016

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-2
SLIDE 2

Overview

◮ Machine Learning needs to tackle complicated optimization

problems ⇒ ML needs optimization.

◮ Frank-Wolfe algorithm (FW) gained in popularity in the

last couple of years.

◮ It is a convex optimization algorithm solving constrained

problems.

◮ We tried to extend FW to saddle point optimization which

is non trivial (we partially answered a 30 years old conjecture).

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-3
SLIDE 3

Motivations: Games

Zero-sum games with two players:

◮ Player 1 has actions {1, . . . , I} available. ◮ Player 2 has actions {1, . . . , J} available. ◮ If action i and action j, implies a reward Mij for Player 1 ◮ Two players play randomly, x ∈ ∆(|I|), y ∈ ∆(|J|),

E[Mij] = x⊤My Nash equilibrium: (x∗, y∗) ∈ X × Y, ∀(x, y) ∈ X × Y (x∗)⊤My ≤ (x∗)⊤My∗ ≤ x⊤My∗

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-4
SLIDE 4

Saddle point setting

Let L : X × Y → R, where X and Y are convex and compact.

  • Intuition from two players games:

◮ L is a score function. ◮ P1 chooses action in X and want to minimize the score. ◮ P2 chooses action in Y and want to maximize the score. ◮ The saddle point is the couple of best choice for each player.

  • L is said to be convex-concave if:
  • 1. ∀ y ∈ Y, x → L(x, y) is convex.
  • 2. ∀ x ∈ X, y → L(x, y) is concave.
  • A saddle point is a couple (x∗, y∗) such that,

∀(x, y) ∈ X × Y, L(x∗, y) ≤ L(x∗, y∗) ≤ L(x, y∗)

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-5
SLIDE 5

Motivations: mores applications

Robust learning:1 We want to learn min

θ∈Θ

1 n

n

  • i=1

ℓ (fθ(xi), yi) + λΩ(θ) (1) with an uncertainty regarding the data: min

θ∈Θ max w∈∆n n

  • i=1

ωiℓ (fθ(xi), yi) + λΩ(θ) (2)

1Junfeng Wen, Chun-Nam Yu, and Russell Greiner. “Robust Learning

under Uncertain Test Distributions: Relating Covariate Shift to Model Misspecification.” In: ICML. 2014, pp. 631–639.

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-6
SLIDE 6

Standard approaches in literature

The standard algorithm to solve Saddle point optimization is the projected gradient algorithm. x(t+1) = PX (x(t) − η∇xL(x(t), y(t))) y(t+1) = PY(y(t) + η∇yL(x(t), y(t))) When the gradient is uniformly bounded, 1 T

T

  • t=1
  • x(t), y(t)

− →

T→∞ (x∗, y∗)

(3)

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-7
SLIDE 7

The FW algorithm

Initialize x(0). For t = 0, . . . , T do

◮ Compute:

s(t) := argmin

s∈X

  • s, ∇f(x(t)

.

◮ Let γt =

2 2+t.

◮ Update:

x(t+1) = x(t) + γt(s(t) − x(t))

end

Figure: One step of the FW algorithm

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-8
SLIDE 8

SPFW

Then a Saddle point version of Frank Wolfe algorithm is

◮ Let z(0) = (x(0), y(0)) ∈ X × Y ◮ For t = 0 . . . T

◮ Compute G =

∇xL(x(t), y(t)) −∇yL(x(t), y(t))

  • ◮ Compute s(t) := argmin

s∈X×Y

s, G

◮ Let γt =

2 2+t

◮ Update z(t+1) := (1 − γt)z(t) + γts(t)

◮ return (x(T), y(T))

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-9
SLIDE 9

Advantages of SP-FW

Why would we use SP-FW ?

◮ Only a LMO (linear oracle). ◮ Gap certificate for free. ◮ Simplicity of implementation. ◮ Universal step size 2 2+k, adaptive step size gt 2CL , . . . ◮ Sparsity of the solution. ◮ Lots of improvement easily available. Block-coordinate,

Away Step... When the constraint set is a “complicated” polytope the projection can be super hard whereas the LMO might be tractable.

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-10
SLIDE 10

Problems with Hard projection

The structured SVM: min

ω λΩ(ω) + 1

n

n

  • i=1

˜ Hi(ω) where ˜ Hi(ω) = maxy∈Yi Li(y) − ω, φi(y) is the structured hinge loss. Then we can rewrite the problem as min

Ω(ω)≤R

1 n

n

  • i=1
  • max

yi∈Yi L⊤ i yi − ω⊤Miyi

  • but as the function is bilinear

min

Ω(ω)≤β

max

α∈∆(|Y|) bT α − ωT Mα

If Ω(·) is a group lasso norm with overlapping projection is

  • hard. Projecting on Y is intractable.

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-11
SLIDE 11

Problems with hard projection

University game:

  • 1. Game between two universities (A and B).
  • 2. Admitting d students and have to assign pairs of students

into dorms.

  • 3. The game has a payoff matrix M belonging to R(d(d−1)/2)2.
  • 4. Mij,kl is the expected tuition that B gets (or A gives up) if

A pairs student i with j and B pairs student k with l.

  • 5. Here the actions are both in the marginal polytope of all

perfect unipartite matchings. Hard to project on this polytope whereas the LMO can be solved efficiently with the blossom algorithm2.

  • 2J. Edmonds. “Paths, trees and flowers”.

In: Canadian Journal of Mathematics (1965).

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-12
SLIDE 12

Our contributions

Theoretical contributions:

◮ Introduced a SP extension of FW with away step and

proved its convergence over a polytope under some conditions (strong convexity of the function big enough). Partially answering a 30 years old conjecture3.

◮ With step size γt ∼ gt

ht = O

  • (1 − ρ)t/3

. (4)

3Janice H Hammond. “Solving asymmetric variational inequality

problems and systems of equations with generalized nonlinear programming algorithms”. PhD thesis. Massachusetts Institute of Technology, 1984.

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-13
SLIDE 13

Toy experiments

500 1000 1500 2000 10

−4

10

−2

10 Iteration Duality Gap

τ=0.037 γ = 2/(2+k) τ=0.037 γ adaptive τ=0.29 γ adaptive τ=0.38 γ adaptive τ=0.49 γ adaptive

Figure: SP-AFW on a toy example d = 30

100 200 300 400 500 600 10

−3

10

−2

10

−1

10 10

1

Iteration Duality Gap

τ=−0.22 γ heuristic τ=−2.4 γ heuristic τ=−46 γ heuristic τ=−4.6e+03 γ heuristic τ=−4.6e+03 γ = 2/(2+k)

Figure: SP-AFW on a toy example d = 30 with heuristic step-size

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-14
SLIDE 14

Experiments

10 10

1

10

2

10

−3

10

−2

10

−1

10 10

1

Iteration Duality gap

d=28 d=120 d=496 d=2016 d=8128 d=32640

Figure: SP-FW on the University game.

10

2

10

3

10

4

10

−5

10

−4

10

−3

10

−2

Effective Pass Primal Suboptimality

SP−FW, γ = 2/(2+k) SP−FW, γ = 1/(1+k) SP−BCFW, γ = 2n/(2n+k) Subgradient SSG, γ = 1/L(k+1)1/2 SSG γ2 = 0.1/L(k+1)1/2

Figure: Structural SVM with OCR dataset (highly regularized).

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-15
SLIDE 15

Conclusion

◮ There already exist a lot a saddle point problem in the

machine learning literature and they are most of the time solved by a trick.

◮ There exist a few number of algorithm to solve SP

problems directly ! (and they are not well known)

◮ SP-FW work on SPs and is the only algorithm existing

able to solve some of these problem.

Gauthier Gidel, Simon Lacoste-Julien Frank-Wolfe Algorithms for SP September 15th 2016

slide-16
SLIDE 16

Thank You !