Optimization (ND Methods) What is the optimal solution? (ND) f- ( x - - PowerPoint PPT Presentation

optimization nd methods
SMART_READER_LITE
LIVE PREVIEW

Optimization (ND Methods) What is the optimal solution? (ND) f- ( x - - PowerPoint PPT Presentation

Optimization (ND Methods) What is the optimal solution? (ND) f- ( x ) ! * = min $ ! * HI ) = A (First-order) Necessary condition 1D: ! " = 0 Q - gives : If (E) stationary solution NI = * (Second-order) Sufficient condition 1D: !


slide-1
SLIDE 1

Optimization (ND Methods)

slide-2
SLIDE 2

What is the optimal solution? (ND)

(First-order) Necessary condition (Second-order) Sufficient condition

1D: !## " > 0

1D: !′ " = 0

! *∗ = min

$ ! * f-(x)

HI)

= A

NI

: If(E) =

Q - gives

stationary solution

×*

NI

i

t

)

is

positive

definite

→ x

* is

minimizer

slide-3
SLIDE 3

Taking derivatives…

f : IR

" → R

f-(Xn) = f-(Xi , Xz ,

  • , Xn)

¥

.

÷

.
  • :# ⇒

I:÷÷÷l

gradient

  • f

f

CnxD

¥

. If z

µ ,±,

Ein Eas

  • Tian Hi .
  • ¥

,

:*

. :*

.

:#

  • x.
  • o¥oxn a

¥÷⇒

K

YE'

  • ¥a

.. ¥¥

.

  • -
  • . g¥)

nth

slide-4
SLIDE 4

From linear algebra:

A symmetric ; ×; matrix = is positive definite if >&= > > @ for any > ≠ @ A symmetric ; ×; matrix = is positive semi-definite if >&= > ≥ @ for any > ≠ @ A symmetric ; ×; matrix = is negative definite if >&= > < @ for any > ≠ @ A symmetric ; ×; matrix = is negative semi-definite if >&= > ≤ @ for any > ≠ @ A symmetric ; ×; matrix = that is not negative semi-definite and not positive semi- definite is called indefinite

I.i

ta

  • O

y

. HyG EoII's. s

e::÷

.I :

.

slide-5
SLIDE 5

! *∗ = min

$ ! *

First order necessary condition: +! * = , Second order sufficient condition: - * is positive definite How can we find out if the Hessian is positive definite? la.eight)

THW→ ( X, y)

→ are eigenpairs of H

jyxyty

  • x " yn:

→ "5fa¥:*

.

* ki so

ti

/)

y

THy

> o tty ftp.defsxt is

minimizer

* Xi Lo fi ⇒ y '

Hy

< o

Hy → His neg def ⇒ x* is

maximizer

* Tgi,Yo }→ H is indefinite

→ x* is

psaddetde

  • int
slide-6
SLIDE 6

Types of optimization problems

Gradient-free methods Gradient (first-derivative) methods

Evaluate ! * , +! * , +%! *

Second-derivative methods

! *∗ = min

$ ! *

Evaluate ! ' Evaluate ! ' , !" #

5: nonlinear, continuous and smooth

+HH

slide-7
SLIDE 7

Consider the function " -E, -: = 2-E

9 + 4-: : + 2-: − 24-E

Find the stationary point and check the sufficient condition

Example (ND)me

E-if

"-244¥:(

"

  • "

: )

8×2+2

1)of -

  • e

→ (GI!:{

4]=fg]⇒

6×2--24

→ xf=4→x,=±z

8×2=-2

→ Xz=
  • O
  • 25

stationary points :

x# III.as]

x'

*
  • fo?⇐]

''et:

Hina

.fi#:.u.fttfo.-t-l::Ih::::i.n

.
slide-8
SLIDE 8

Optimization in ND:

Steepest Descent Method

Given a function ! * : ℛ7 → ℛ at a point *, the function will decrease its value in the direction of steepest descent: −+! *

" -E, -: = (-E − 1)O+(-: − 1)O

What is the steepest descent direction?

min FG)

x

EI

  • FIFA

£

Xz

  • -
  • • ×

⇒if

"

(Xo)

i

X1

slide-9
SLIDE 9

Steepest Descent Method

" -E, -: = (-E − 1)O+(-: − 1)O

Start with initial guess:

!M = 3 3

Check the update:

¥2

= Ii
  • Of

XoKD

es

I

It

Etan

.
  • IT
,]

missed

±

.

ft;]

  • i
slide-10
SLIDE 10

Steepest Descent Method

" -E, -: = (-E − 1)O+(-: − 1)O Update the variable with: !ILE = !I − PIQ" !I How far along the gradient should we go? What is the “best size” for PI? =

I ,

= Io
  • 0=5 OfGo)
.-
  • µ

12=0.57

How

can

we get

a ?

I

slide-11
SLIDE 11

i

:

"

÷÷÷÷÷:÷÷÷

.

.

05¥

slide-12
SLIDE 12

Steepest Descent Method

Algorithm: Initial guess: *3 Evaluate: ;4= −+! *4 Perform a line search to obtain <4 (for example, Golden Section Search) <4 = argmin

8

! *4 + < ;4 Update: *456 = *4 + <4 ;4

#

  • f
'"

=①

÷÷

.

D

O

ret ask

slide-13
SLIDE 13

Line Search

Xktt

= Xk
  • Ak Tf(

Xn)

we want to find

de St .

f-(Hett )

min f( Xn

  • ④DfC)

x

he

1st order condition ¥70

→ gives a)

DI

= ofArt,) . Tf(Xr)

= O

da

OXKH

pg(XnH)•Tf¥

f(Xkti) is

  • rthogonal

to

zigpoEEfrwnergefa-a.GR#

slide-14
SLIDE 14

Example

Consider minimizing the function ! "!, "" = 10("!)# − "" " + "! − 1 Given the initial guess "! = 2, ""= 2 what is the direction of the first step of gradient descent?

Miu far, , Xz)

X , gXz

=@ ±

.
  • I:]

a- =p

if

Iska

  • III ]

4un±÷÷t→fi;

slide-15
SLIDE 15

Newton’s Method

Using Taylor Expansion, we build the approximation:

f-(Ets)

  • f

t ELITE tf

It④

= ICI
  • -

t

nonlinear

quadratic

+ St order condition

: II - o

approx off

I⇒t⇒

→ His :p

.my#etric--/tfH.)s---TfGTf

→ solve

liusys

to find

  • Newton slip

s

slide-16
SLIDE 16

Newton’s Method

Algorithm: Initial guess: *3 Solve: -9 *4 ;4 = −+! *4 Update: *456 = *4 + ;4

Note that the Hessian is related to the curvature and therefore contains the information about how large the step should be.

,#ii=f¥

any

  • → solve

0¥ ⑤

  • T

G

slide-17
SLIDE 17

Try this out!

" -, R = 0.5-: + 2.5R:

When using the Newton’s Method to find the minimizer of this function, estimate the number of iterations it would take for convergence?

A) 1 B) 2-5 C) 5-10 D) More than 10 E) Depends on the initial guess

slide-18
SLIDE 18

Newton’s Method Summary

Algorithm: Initial guess: *3 Solve: -9 *4 ;4 = −+! *4 Update: *456 = *4 + ;4 About the method…

  • Typical quadratic convergence J
  • Need second derivatives L
  • Local convergence (start guess close to solution)
  • Works poorly when Hessian is nearly indefinite
  • Cost per iteration: >(@:)

=