CS/ECE/ISyE 524 Introduction to Optimization Spring 2017–18
- 25. NLP algorithms
❼ Overview ❼ Local methods ❼ Constrained optimization ❼ Global methods ❼ Black-box methods ❼ Course wrap-up
Laurent Lessard (www.laurentlessard.com)
25. NLP algorithms Overview Local methods Constrained optimization - - PowerPoint PPT Presentation
CS/ECE/ISyE 524 Introduction to Optimization Spring 201718 25. NLP algorithms Overview Local methods Constrained optimization Global methods Black-box methods Course wrap-up Laurent Lessard (www.laurentlessard.com)
CS/ECE/ISyE 524 Introduction to Optimization Spring 2017–18
Laurent Lessard (www.laurentlessard.com)
25-2
25-3
x
25-4
25-5
◮ unconstrained case ◮ constrained case
25-6
x
25-7
25-8
25-9
∂x ∂f ∂y
∂x2 ∂2f ∂x∂y ∂2f ∂x∂y ∂2f ∂y2
best linear approximation
25-10
◮ Pick a constant tk = t ◮ Pick a slowly decreasing stepsize, such as tk = 1/
◮ Exact line search: tk = arg mint f (xk − t∇
◮ A heuristic method (most common in practice).
25-11
λmin(Q) determines convergence.
5 5 2 1 1 2
Optimal step Shorter step Even shorter
100 101 102 103 number of iterations 10-14 10-12 10-10 10-8 10-6 10-4 10-2 100 distance to optimal point
Optimal step Shorter step Even shorter
5 5 2 1 1 2
Optimal step Shorter step Even shorter
100 101 102 103 number of iterations 10-14 10-12 10-10 10-8 10-6 10-4 10-2 100 distance to optimal point
Optimal step Shorter step Even shorter
25-12
25-13
25-14
i=1 fi(x). Use direction i∈S ∇
25-15
25-16
example by: L. El Ghaoui, UC Berkeley, EE127a 25-17
example by: L. El Ghaoui, UC Berkeley, EE127a 25-18
25-19
◮ DFP (Davidon–Fletcher–Powell) ◮ BFGS (Broyden–Fletcher–Goldfarb–Shanno)
25-20
2 4 6 8 x 3 2 1 1 2 3 y
Gradient Nesterov BFGS Newton 25-21
100 101 102 103 number of iterations 10-14 10-12 10-10 10-8 10-6 10-4 10-2 100 distance to optimal point
Gradient Nesterov BFGS Newton
25-22
25-23
25-24
x
x
m
25-25
4x2 + 1
25-26
x
25-27
25-28
LD_AUGLAG LD_AUGLAG_EQ LD_CCSAQ LD_LBFGS_NOCEDAL LD_LBFGS LD_MMA LD_SLSQP LD_TNEWTON LD_TNEWTON_RESTART LD_TNEWTON_PRECOND LD_TNEWTON_PRECOND_RESTART LD_VAR1 LD_VAR2 LN_AUGLAG LN_AUGLAG_EQ LN_BOBYQA LN_COBYLA LN_NEWUOA LN_NEWUOA_BOUND LN_NELDERMEAD LN_PRAXIS LN_SBPLX GD_MLSL GD_MLSL_LDS GD_STOGO GD_STOGO_RAND GN_CRS2_LM GN_DIRECT GN_DIRECT_L GN_DIRECT_L_RAND GN_DIRECT_NOSCAL GN_DIRECT_L_NOSCAL GN_DIRECT_L_RAND_NOSCAL GN_ESCH GN_ISRES GN_MLSL GN_MLSL_LDS GN_ORIG_DIRECT GN_ORIG_DIRECT_L
25-29
◮ Randomly pepper the space with initial points. ◮ Run your favorite local method starting from each point
◮ Compare the different local minima found.
25-30
◮ Systematically partition the space using a
◮ Search the smaller spaces using local gradient-based search.
25-31
25-32
25-33
◮ CS 525: linear programming methods ◮ CS 526: advanced linear programming
◮ CS 726: nonlinear optimization I ◮ CS 727: nonlinear optimization II ◮ CS 727: convex analysis
◮ CS 425: introduction to combinatorial optimization ◮ CS 577: introduction to algorithms ◮ CS 720: integer programming ◮ CS 728: integer optimization
25-34
25-35