newton type constrained optimization in a nutshell
play

Newton Type Constrained Optimization in a Nutshell Moritz Diehl - PowerPoint PPT Presentation

Newton Type Constrained Optimization in a Nutshell Moritz Diehl Optimization in Engineering Center (OPTEC) K.U. Leuven Overview Equality Constrained Optimization Optimality Conditions and Multipliers Newtons Method = SQP


  1. Newton Type Constrained Optimization in a Nutshell Moritz Diehl Optimization in Engineering Center (OPTEC) K.U. Leuven

  2. Overview • Equality Constrained Optimization • Optimality Conditions and Multipliers • Newton’s Method = SQP • Inequality Constraints • Constrained Gauss Newton Method • Relation to Optimal Control

  3. General Nonlinear Program (NLP) In direct methods, we have to solve the discretized optimal control problem, which is a Nonlinear Program (NLP) � G ( w ) = 0 , min w F ( w ) s.t. H ( w ) ≥ 0 . We first treat the case without inequalities. w F ( w ) s.t. min G ( w ) = 0 ,

  4. Lagrange Function and Optimality Conditions Introduce Lagrangian function L ( w, λ ) = F ( w ) − λ T G ( w ) Then for an optimal solution w ∗ exist multipliers λ ∗ such that ∇ w L ( w ∗ , λ ∗ ) = 0 , G ( w ∗ ) = 0 ,

  5. Newton’s Method on Optimality Conditions How to solve nonlinear equations ∇ w L ( w ∗ , λ ∗ ) = 0 , G ( w ∗ ) = 0 , ? Linearize! + ∇ 2 ∇ w L ( w k , λ k ) w L ( w k , λ k )∆ w −∇ w G ( w k )∆ λ = 0 , + ∇ w G ( w k ) T ∆ w G ( w k ) = 0 , This is equivalent, due to ∇L ( w k , λ k ) = ∇ F ( w k ) −∇ G ( w k ) λ k , with the shorthand λ + = λ k + ∆ λ , to ∇ w F ( w k ) + ∇ 2 w L ( w k , λ k )∆ w −∇ w G ( w k ) λ + = 0 , + ∇ w G ( w k ) T ∆ w G ( w k ) = 0 ,

  6. Newton Step = Quadratic Program Conditions ∇ w F ( w k ) + ∇ 2 w L ( w k , λ k )∆ w −∇ w G ( w k ) λ + = 0 , + ∇ w G ( w k ) T ∆ w G ( w k ) = 0 , are optimality conditions of a quadratic program (QP), namely: ∆ w ∇ F ( w k ) T ∆ w +1 2∆ w T A k ∆ w s.t. G ( w k ) + ∇ G ( w k ) T ∆ w min = 0 , with A k = ∇ 2 w L ( w k , λ k )

  7. Newton’s Method The full step Newton’s Method iterates by solving in each iteration the Quadratic Progam ∆ w ∇ F ( w k ) T ∆ w +1 2∆ w T A k ∆ w s.t. G ( w k ) + ∇ G ( w k ) T ∆ w min = 0 , w L ( w k , λ k ) . This obtains as solution the step ∆ w k and with A k = ∇ 2 QP = λ k + ∆ λ k . the new multiplier λ + Then we iterate: w k + ∆ w k w k +1 = λ k + ∆ λ k = λ + λ k +1 = QP This Newton’s method is also called “Sequential Quadratic Program- ming (SQP) for equality constrained optimization” (with “exact Hes- sian” and “full steps”)

  8. NLP with Inequalities Regard again NLP with both, equalities and inequalities: � G ( w ) = 0 , w F ( w ) s.t. min H ( w ) ≥ 0 . Introduce Lagrangian function L ( w, λ, µ ) = F ( w ) − λ T G ( w ) − µ T H ( w )

  9. Optimality Conditions with Inequalities THEOREM (Karush-Kuhn-Tucker (KKT) conditions) For an optimal solution w ∗ exist multipliers λ ∗ and µ ∗ such that ∇ w L ( w ∗ , λ ∗ , µ ∗ ) = 0 , G ( w ∗ ) = 0 , H ( w ∗ ) ≥ 0 , ≥ 0 , µ ∗ H ( w ∗ ) T µ ∗ = 0 , These contain nonsmooth conditions (the last three) which are called “complementarity conditions”. This system cannot be solved by New- ton’s Method. But still with SQP ...

  10. Sequential Quadratic Programming (SQP) By Linearizing all functions within the KKT Conditions, and setting λ + = λ k + ∆ λ and µ + = µ k + ∆ µ , we obtain the KKT conditions of a Quadratic Program (QP) (we omit these conditions). This QP is G ( w k ) + ∇ G ( w k ) T ∆ w � ∆ w ∇ F ( w k ) T ∆ w +1 = 0 , 2∆ w T A k ∆ w s.t. min H ( w k ) + ∇ H ( w k ) T ∆ w ≥ 0 , with A k = ∇ 2 w L ( w k , λ k , µ k ) and its solution delivers λ + µ + ∆ w k , QP , QP

  11. Constrained Gauss-Newton Method In special case of least squares objectives F ( w ) = 1 2 � R ( w ) � 2 2 can approximate Hessian ∇ 2 w L ( w k , λ k , µ k ) by much cheaper A k = ∇ R ( w ) ∇ R ( w ) T . Need no multipliers to compute A k ! QP= linear least squares: G ( w k ) + ∇ G ( w k ) T ∆ w � 1 = 0 , 2 � R ( w k )+ ∇ R ( w k ) T ∆ w � 2 min 2 s.t. H ( w k ) + ∇ H ( w k ) T ∆ w ≥ 0 , ∆ w Convergence: linear (better if � R ( w ∗ ) � small)

  12. Discrete Time Optimal Control Problem N − 1 � l i ( s i , q i ) + E ( s N ) minimize s, q i =0 subject to s 0 − x 0 = 0 , (initial value) s i +1 − f i ( s i , q i ) = 0 , i = 0 , . . . , N − 1 , (discrete system) (path constraints) h i ( s i , q i ) ≥ 0 , i = 0 , . . . , N, r ( s N ) ≥ 0 . (terminal constraints) Can arise also from direct multiple shooting parameterization of con- tinous optimal control problem. This NLP can be solved by SQP or Constrained Gauss-Newton method.

  13. Summary • Nonlinear Programs (NLP) have nonlinear but differentiable problem functions. • Sequential Quadratic Programming (SQP) is a Newton type method to solve NLPs that • solves in each iteration a Quadratic Program (QP) • obtains this QP by linearizing all nonlinear problem functions • an important SQP variant is the Constrained Gauss-Newton Method • SQP can be generalized to Sequential Convex Programming (SCP) • Discrete time optimal control problems are a special case of NLPs.

  14. Literature • J. Nocedal and S. Wright: Numerical Optimization, Springer, 2006 (2nd edition)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend