Optimization with side constraints Philipp Warode October 2, 2019 - - PowerPoint PPT Presentation

optimization with side constraints
SMART_READER_LITE
LIVE PREVIEW

Optimization with side constraints Philipp Warode October 2, 2019 - - PowerPoint PPT Presentation

Optimization with side constraints Philipp Warode October 2, 2019 Mathematics Preparatory Course 2019 Philipp Warode Optimization with side constraints We want to maximize max x 2 y 2 with the side constraint x + y = 1


slide-1
SLIDE 1

Optimization with side constraints

Philipp Warode October 2, 2019

Mathematics Preparatory Course 2019 – Philipp Warode

slide-2
SLIDE 2

Optimization with side constraints

We want to maximize max −x2 − y2 with the side constraint x + y = 1

Mathematics Preparatory Course 2019 – Philipp Warode

slide-3
SLIDE 3

Simple solution

Reduce numbers of variables by inserting side constraints in the objective function Example: Side constraint: x + y = 1 ⇔ y = 1 − x Objective function: max −x2 − y2 = max −x2 − (1 − x)2 Not always possible, e.g. if the side constraint can’t be solved uniquely: x2 + y2 = 1

Mathematics Preparatory Course 2019 – Philipp Warode

slide-4
SLIDE 4

Lagrange Multipliers

Mathematics Preparatory Course 2019 – Philipp Warode

slide-5
SLIDE 5

Lagrange Multipliers

x y

1/4 2/4 3/4 4/4 5/4 6/4 7/4 8/4 9/4

f (x, y) = −x2 − y2 g(x, y) = x + y − 1 = 0

Mathematics Preparatory Course 2019 – Philipp Warode

slide-6
SLIDE 6

Lagrange Multipliers

x y

1/4 2/4 3/4 4/4 5/4 6/4 7/4 8/4 9/4

f (x, y) = −x2 − y2 g(x, y) = x + y − 1 = 0 ∇f ∇g In the maximum, ∇f (gradient of the objective) and ∇g (gradient of the constraint) are linear dependent

Mathematics Preparatory Course 2019 – Philipp Warode

slide-7
SLIDE 7

Lagrange Multipliers

x y

1/4 2/4 3/4 4/4 5/4 6/4 7/4 8/4 9/4

f (x, y) = −x2 − y2 g(x, y) = x + y − 1 = 0 ∇f ∇g In the maximum, ∇f (gradient of the objective) and ∇g (gradient of the constraint) are linear dependent Necessary condition for local optimum: ∇f = λ∇g

Mathematics Preparatory Course 2019 – Philipp Warode

slide-8
SLIDE 8

Lagrange Multipliers Theorem

Let x(0) = (x(0)

n , . . . , x(0) n ) be an optimal solution of

max f (x) s.t. gj(x) = 0 ∀j ∈ {1, . . . , m} and f , g partially differentiable with ∇gj(x(0)) = 0 for j = 1, . . . , m then there are λ1, . . . , λm such that ∇f (x(0)) +

m

  • j=1

λj∇gj(x(0)) = 0.

Mathematics Preparatory Course 2019 – Philipp Warode

slide-9
SLIDE 9

Lagrange Multipliers

1 Create Lagrange function

L(x, λ) := f (x) +

m

  • j=1

λjgj(x)

Mathematics Preparatory Course 2019 – Philipp Warode

slide-10
SLIDE 10

Lagrange Multipliers

1 Create Lagrange function

L(x, λ) := f (x) +

m

  • j=1

λjgj(x)

2 Compute the partial derivatives

∂ ∂xi L(x, λ) = ∂ ∂xi f (x) +

m

  • j=1

λj ∂ ∂xi gj(x) ∂ ∂λj L(x, λ) = gj(x)

Mathematics Preparatory Course 2019 – Philipp Warode

slide-11
SLIDE 11

Lagrange Multipliers

1 Create Lagrange function

L(x, λ) := f (x) +

m

  • j=1

λjgj(x)

2 Compute the partial derivatives

∂ ∂xi L(x, λ) = ∂ ∂xi f (x) +

m

  • j=1

λj ∂ ∂xi gj(x) ∂ ∂λj L(x, λ) = gj(x)

3 Solve the system of equations

∂ ∂xi L(x, λ) = 0 for all i and ∂ ∂λj L(x, λ) = 0 for all j

Mathematics Preparatory Course 2019 – Philipp Warode