Computational Optimization
Advance Topics NonSmooth Optimization Reference: Nonlinear Optimization, Ruszynski,2006
Computational Optimization Advance Topics NonSmooth Optimization - - PowerPoint PPT Presentation
Computational Optimization Advance Topics NonSmooth Optimization Reference: Nonlinear Optimization, Ruszynski,2006 Best Linear Separator: Supporting Plane Method Maximize distance Between two para supporting planes Distance = x
Advance Topics NonSmooth Optimization Reference: Nonlinear Optimization, Ruszynski,2006
Maximize distance Between two para supporting planes Distance = “Margin” =
|| || w δ β −
1 1 2 2 , ,
i i i wb z
=
1
1 1 2 2 , ,
w b i i i i z i i
=
at x.
t subgradien a is x)
g' f(x) f(y) such that g vector a function. convex a be : Let f R R R f
n n
+ ≥ ∈ →
Hinge loss 1
f(y) f(x) g'(y-x) ≥ +
) (x f g ∂ ∈
1
f(y) f(x) g'(y-x) ≥ +
1
k k k k k k k k k k k k
+ =
Contour plot
) (x f ∂
k
g −
But fixed stepsize schemes can still work
x t s x f ∈ . . ) ( min
1
k k k k k k k
+ =
1
k k k k
+ =
k k k
1 1 1
f(y) f(x ) g '(y-x ) ≥ +
2 2 2
f(y) f(x ) g '(y-x ) ≥ +
k k k
k
) (
k k
x f g ∂ ∈ arg min . . ( ) '( ) 1,..,
k i i i
x z s t z f x g y x i k ∈ ≥ + − =
1
( ) ( )
k k
if f x f x then stop optimal
+
=
Optimize convex program Lagrangian dual function Lagrangian dual problem
. . ) ( min X x b Ax t s x f ∈ =
) ( ' ) ( min ) ( Ax b x f
X x
− + =
∈
λ λ θ max ( )
λ
θ λ
≥
Subgradient found by solving for then
arg min ( ) '( )
k k x X
x f x b Ax λ
∈
∈ + − ( ) ( )
k k k
g b Ax θ λ = − ∈∂
arg min . . ( ) '( ) 1,..,
k i i i
z s t z g y i k C y C λ θ λ λ ∈ ≥ + − = − ≤ ≤
C constraints insure problem always has a solution
=
=
k i k kx
u x
1
*
2 ,
min 2 . . ( ) '( )
k z y X i i i k
z x w s t z f x g y x i J ρ
∈
+ − ≥ + − ∈
1.
Calculate f(xk) and gk if f(xk) < vk, add k to constraints in J
then wk=xk else wk = wk-1.
3.
Solve restricted master for (xk+1,vk+1)
4.
If fk(xk+1)=f(wk) , then stop xk+1 optimal
5.
Update J by removing cuts with negative multipliers from solving the subproblems.
No step size needed Nice check for optimality if a function achieves its lower bound it is optimal. Reduces to a series of nice convex quadratic subproblems Can remove constraints while still adding Finite convergence for piecewise linear convex function with polyhedral constraints. Can be extended to nonconvex nonsmooth
Still only uses first order information so can be slow.