constrained convex optimization virgil pavlu 1 convex set a set X - - PowerPoint PPT Presentation

constrained convex optimization
SMART_READER_LITE
LIVE PREVIEW

constrained convex optimization virgil pavlu 1 convex set a set X - - PowerPoint PPT Presentation

constrained convex optimization virgil pavlu 1 convex set a set X in a vector space is convex if for any w 1 , w 2 X and [0 , 1] we have w 1 + (1 ) w 2 X 2 convex function a function f is convex(concave) on X Dom ( f


slide-1
SLIDE 1

constrained convex optimization

virgil pavlu

1

slide-2
SLIDE 2

convex set

a set X in a vector space is convex if for any w1, w2 ∈ X and λ ∈ [0, 1] we have λw1 + (1 − λ)w2 ∈ X

2

slide-3
SLIDE 3

convex function

a function f is convex(concave) on X ⊆ Dom(f) if for any w1, w2 ∈ X and λ ∈ [0, 1] we have f(λw1 + (1 − λ)w2)≤ (≥)λf(w1) + (1 − λ)f(w2) if f is strict convex and twice differentiable on X then : ♦f = δwf(w) strict increasing ♦f ≥ 0 ♦f(x0) = 0 ⇔ x0 is a global minimum

3

slide-4
SLIDE 4

convex and differentiable

if f is convex and differentiable then for any w1,w2 we have 1.f(w2)(w2 − w1) ≥ f(w2) − f(w1) ≥ f(w1)(w2 − w1) 2.∃w = λw1 + (1 − λ)w2, λ ∈ [0, 1] such that f(w2) − f(w1) = f(w)(w2 − w1)

4

slide-5
SLIDE 5

unconstrained optimization

  • ne variable

interval cutting Newton’s Method several variables gradient descent conjugate gradient descent

5

slide-6
SLIDE 6

constrained optimization

given convex functions f, g1, g2, ...., gk, h1, h2, ...., hm on convex set X, the problem minimize f(w) subject to gi(w) ≤ 0 ,for all i hj(w) = 0 ,for all j has as its solution a convex set. If f is strict convex the solution is unique (if exists) we will assume all the good things one can imagine about functions f, g1, g2, ...., gk, h1, h2, ...., hm like convexity, differentiability etc.That will still not be enough though....

6

slide-7
SLIDE 7

equality constraints only

minimize f(w) subject to hj(w) = 0 ,for all j define the lagrangian function L(w, β) = f(w) +

  • j

βjhj(w) Lagrange theorem nec[essary] and suff[icient] conditions for a point

  • w to be an optimum (ie a solution for the problem above) are the

existence of β such that δwL(

w,

β) = 0 ; δβjL(

w,

β) = 0

7

slide-8
SLIDE 8

!10 !5 5 10 2 4 6 8 10 !1400 !1200 !1000 !800 !600 !400 !200 200

w alpha

8

slide-9
SLIDE 9

inequality constraints

minimize f(w) subject to gi(w) ≤ 0 ,for all i hj(w) = 0 ,for all j we can rewrite every inequality constraint hj(x) = 0 as two inequalities hj(w) ≤ 0 and hj(w) ≥ 0. so the problem becomes minimize f(w) subject to gi(w) ≤ 0 ,for all i

9

slide-10
SLIDE 10

Karush Kuhn Tucker theorem

minimize f(w) subject to gi(w) ≤ 0 ,for all i were gi are qualified constraints define the lagrangian function L(w, α) = f(w) +

  • i

αigi(w) KKT theorem nec and suff conditions for a point

w to be a solution

for the optimization problem are the existence of α such that δwL(

w,

α) = 0 ; αigi(

w) = 0

gi(

w) ≤ 0 ;

αi ≥ 0

10

slide-11
SLIDE 11

KKT - sufficiency

Assume (

w,

α) satisfies KKT conditions δwL(

w,

α) = δwf(

w) + k

i=1

αiδwgi(

w) = 0

δαiL(

w,

α) = gi(

w) ≤ 0

  • αigi(

w) = 0;

αi ≥ 0 Then f(w) − f(

w) ≥ (δwf( w))T(w − w) =

− k

i=1

αi(δwgi(

w))T(w − w) ≥ − k

i=1

αi(gi(w) − gi(

w)) =

− k

i=1

αigi(w) ≥ 0 so

w is solution

11

slide-12
SLIDE 12

saddle point

minimize f(w) subject to gi(w) ≤ 0 ,for all i and the lagrangian function L(w, α) = f(w) +

  • i

αigi(w) (

w,

α) with αi ≥ 0 is saddle point if ∀(w, α), αi ≥ 0 L(

w, α) ≤ L( w,

α) ≤ L(w, α)

12

slide-13
SLIDE 13

!10 !5 5 10 2 4 6 8 10 !1400 !1200 !1000 !800 !600 !400 !200 200

w alpha

13

slide-14
SLIDE 14

saddle point - sufficiency

minimize f(w) subject to gi(w) ≤ 0 ,for all i lagrangian function L(w, α) = f(w) +

i αigi(w)

(

w,

α) is saddle point ∀(w, α), αi ≥ 0 : L(

w, α) ≤ L( w,

α) ≤ L(w, α) then 1.

w is solution to optimization problem

2. αigi(

w) = 0 for all i

14

slide-15
SLIDE 15

saddle point - necessity

minimize f(w) subject to gi(w) ≤ 0 ,for all i were gi are qualified constraints lagrangian function L(w, α) = f(w) +

i αigi(w)

  • w is solution to optimization problem

then ∃ α ≥ 0 such that (

w,

α) is saddle point ∀(w, α), αi ≥ 0 : L(

w, α) ≤ L( w,

α) ≤ L(w, α)

15

slide-16
SLIDE 16

constraint qualifications

minimize f(w) subject to gi(w) ≤ 0 , for all i let Υ be the feasible region Υ = {w|gi(w) ≤ 0 ∀i} then the following additional conditions for functions gi are connected (i) ⇔ (ii) and (iii) ⇒ (i) : (i) there exists w ∈ Υ such that gi(w) ≤ 0 ∀i (ii) for all nonzero α ∈ [0, 1)k ∃w ∈ Υ such that αigi(w) = 0 ∀i (iii) the feasible region Υ contains at least two distinct elements, and ∃w ∈ Υ such that all gi are are strictly convex at w w.r.t. Υ

16

slide-17
SLIDE 17

KKT-gap

Assume

w is the solution for optimization problem.Then for any (w, α)

with α ≥ 0 and satisfying δwL(w, α) = 0 ; δαiL(w, α) ≥ 0 we have f(w) ≥ f(

w) ≥ f(w) +

k

  • i=1

αigi(w)

17

slide-18
SLIDE 18

duality

f(w) ≥ f(

w) ≥ f(w) +

k

  • i=1

αigi(w) dual maximization problem : maximize L(w, α) = f(w) + k

i=1 αigi(w)

subject to α ≥ 0 ; δwL(w, α) = 0 OR set θ(α) = infw L(w, α) maximize θ(α) subject to α ≥ 0 the primal and dual problems have the same objective value iff the gap can be vanished

18