Optimal Harvesting with Coupled Population and Price Dynamics Floyd - - PowerPoint PPT Presentation

optimal harvesting with coupled population and price
SMART_READER_LITE
LIVE PREVIEW

Optimal Harvesting with Coupled Population and Price Dynamics Floyd - - PowerPoint PPT Presentation

Optimal Harvesting with Coupled Population and Price Dynamics Floyd B. Hanson Laboratory for Advanced Computing University of Illinois at Chicago and Dennis Ryan Division of Science and Mathematics McKendree College SIAM 50th Anniversary


slide-1
SLIDE 1

Optimal Harvesting with Coupled Population and Price Dynamics

Floyd B. Hanson∗

Laboratory for Advanced Computing University of Illinois at Chicago and

Dennis Ryan

Division of Science and Mathematics McKendree College

SIAM 50th Anniversary Meeting 08-12 July 2002 in Philadelphia MS26 Control Applications in Mathematical Biology

∗Work supported in part by National Science Foundation Computational

Mathematics Program under grants DMS-93-01107, DMS-96-26692, DMS-99-73231, DMS-02-07081.

Hanson and Ryan — 1 — UIC and McKendree

slide-2
SLIDE 2

Overview

  • 1. Noninflationary, Deterministic Model.
  • 2. Inflationary, Stochastic Control Model.
  • 3. Numerical Approximations.
  • 4. Numerical Results.
  • 5. Conclusions.

Hanson and Ryan — 2 — UIC and McKendree

slide-3
SLIDE 3

Outline of Abstract

  • Optimal Control of Stochastic Resource in Continuous

Time.

  • Model Effects of Large Random Price Fluctuations.
  • Influence of Continuous growth and Jump Stochastic

Noise.

  • Computational Stochastic Dynamic Programming.
  • Pronounced Effect of Inflationary Prices on Optimal

Return.

Hanson and Ryan — 3 — UIC and McKendree

slide-4
SLIDE 4

Part 1. Noninflationary, Deterministic Model: Introduction.

1.1. Ordinary Differential Equation (ODE):

  • Nonlinear (Logistic) Dynamics:

dX(s) = [r1X(s)(1 − X(s)/K) − H(s)] ds, 0 < t < s < T.

  • Initial Conditions: X(0) = x0 ;

0 < t < T

  • State Variable (Resource Size): X(t) = [Xi(t)]1×1 ;
  • BiLinear Control-State Dynamics Assumption (for

Resource Harvesting): H(t) = qU(t)X(t) ;

  • q = Efficiency (Catchability) Coefficient;

Hanson and Ryan — 4 — UIC and McKendree

slide-5
SLIDE 5
  • Control Variable (Harvesting Effort):

U(t) = [Ui{X(t), t}]1×1 , Umin ≤ U(t) ≤ Umax < ∞ ;

  • Growth Parameters:

r1 = Resource Intrinsic Growth Rate; K = Environment Carrying (Saturation) Capacity.

Hanson and Ryan — 5 — UIC and McKendree

slide-6
SLIDE 6

1.2. Quadratic Performance Index: V (X, U, t) = T

t

e−δ(s−t) [pqU(s)X(s) − c(U(s))] ds , where

  • V (x, u, t) = Current Value of Future Resources (i.e.,

exp(δt) times Present Value);

  • T = Time Horizon (T ≥ t);
  • δ = Nominal Discount Rate (NOT adjusted for inflation);
  • p = Price of Resource per Unit Harvest Rate;
  • c(u) = c1u + c2u2 = Quadratic Costs (Assume Increasing,

Convex Quadratic Costs: c1 > 0 and c2 > 0);

  • Instantaneous Net Return: R(x, u) = pqux − c(u) .

Hanson and Ryan — 6 — UIC and McKendree

slide-7
SLIDE 7

1.3. Deterministic Dynamic Programming:

  • Optimization Goal = Maximize Total Return:

v∗(x, t) = V (x, u∗, t) = max

u

[V (x, u, t)] ;

  • PDE of Deterministic Dynamic Programming:

v∗

t (x, t) + r1x(1 − x/K)v∗ x (x, t) − δv∗(x, t) + S∗(x, t) = 0;

  • Control Switching Term:

S∗(x, t) = max

u

h“ p − v∗

x (x, t)

” qux − c1u − c2u2i ;

  • Regular (Unconstrained) Control:

uR(x, t) = (p − v∗

x(x, t))qx − c1

2 · c2 , c2 > 0;

Hanson and Ryan — 7 — UIC and McKendree

slide-8
SLIDE 8
  • Optimal (Constrained) Control:

u∗(x, t) =        Umax, Umax ≤ uR(x, t) uR(x, t), Umin ≤ uR(x, t) ≤ Umax Umin, uR(x, t) ≤ Umin        ;

  • Final Boundary Condition:

v∗(x, T) = 0;

  • Extinction Natural Boundary Condition:

v∗(0, t) = − (c1 + c2Umin)Umin δ “ 1 − e−δ(T −t)” , δ > 0.

Hanson and Ryan — 8 — UIC and McKendree

slide-9
SLIDE 9

Part 2. Inflationary, Stochastic Control Model

2.1. Stochastic Dynamics Equation (SDE (1)):

  • Nonlinear Dynamics with Gaussian (G) and Poisson

(Z) Noise: dX(s) = [r1X(s)(1 − X(s)/K) − H(s)] ds + σ1X(s) dW1(s) + X(s)

n

  • j=1

aj dZj(s, fj) , X(t) = x ,

  • Initial Conditions: X(0) = x0 ,

t0 < t < s < T ;

  • Gaussian (Wiener) Noise (Zero Mean and

Normalized): E[dW1(t)] = 0 , V ar[dW1(t)] = dt σ1 ≤ 0 ;

Hanson and Ryan — 9 — UIC and McKendree

slide-10
SLIDE 10
  • Poisson (Jump) Noise:

E[dZj(t, fj)] = fjdt , V ar[dZj(t, fj)] = fjdt , 1 ≤ j ≤ n, where fj = Jump Rate and aj = Jump Amplitude Coefficient (−1 < aj);

  • Independent (Uncorrelated) Processes Assumption:

Cov[dW1(t), dZj(t, fj)] = 0 , Cov[dZj(t, fj), dZj′(t, fj′)] = δj,j′fjdt ;

Hanson and Ryan — 10 — UIC and McKendree

slide-11
SLIDE 11

2.2. Inflationary Factor Model:

  • Nonlinear Supply–Demand Model Relation:

P(t) = p0 H (t) + p1

  • Y(t),

* P(t) · H(t) = Gross Return on Harvest; * p0 = Supply–Demand Price Coefficient; * p1 = Constant Price per Unit Harvest; * Y(t) = Fluctuating Inflationary Factor;

  • Linear Fluctuating Inflationary Factor SDE (2):

dY(s) = r2Y(s) ds + σ2Y(s) dW2(s) + Y(s)

m

X

j=1

bj dQj(s; gj) ,

* Y(t) = y; * r2 = Annual Rate of Inflation without Fluctuations; * gj = jth component of Inflationary Jump Rate; * bj = jth component of Jump Amplitude Coefficient;

Hanson and Ryan — 11 — UIC and McKendree

slide-12
SLIDE 12
  • Inflationary Gaussian (Wiener) Noise:

E[dW2(t)] = 0 , V ar[dW2(t)] = dt σ2 ≤ 0 ;

  • Inflationary Poisson (Jump) Noise:

E[dQj(t, gj)] = gjdt , V ar[dQj(t, gj)] = gjdt , 1 ≤ j ≤ m ;

  • Independent (Uncorrelated) Processes Assumption:

Cov[dW2(t), dQj(t, gj)] = 0 , Cov[dQj(t, gj), dQj′(t, gj′)] = δj,j′gjdt ;

Hanson and Ryan — 12 — UIC and McKendree

slide-13
SLIDE 13

1940 1950 1960 1970 1980

Year

1 2 3 4 5

P, Price (US Dollars per Kilogram)

Figure 1: Pacific halibut prices in USdollars per kilogram for each year from 1935 to 1985 (Raw Data: IPHC 1984 and 1985 Annual Reports).

Hanson and Ryan — 13 — UIC and McKendree

slide-14
SLIDE 14

1940 1950 1960 1970 1980

Year

5 10 15 20 25 30 35 40

H, Catch (Million Kilograms)

Figure 2: U.S.-Canadian catch in millions of kilograms for each year from 1935 to 1985 (Raw Data: IPHC 1984 and 1985 Annual Reports).

Hanson and Ryan — 14 — UIC and McKendree

slide-15
SLIDE 15

10 15 20 25 30 35

H, Catch (Million Kilograms)

1 2 3 4

P, Price (US Dollars per Kilogram)

Figure 3: Pacific halibut price in USdollars per kilogram versus catch in

millions of kilograms for years from 1935 to 1985. Linear regression for price times catch as a function of catch from 1980 to 1985 displayed as smooth hyperbolic price curve. (Raw Data: IPHC 1984 and 1985 Annual Reports). Hanson and Ryan — 15 — UIC and McKendree

slide-16
SLIDE 16

2.3. Mean Quadratic Performance Index:

V (x, y, u, t) = E »Z T

t

e−δ(s−t) [(p0 + p1qU(s)X(s))Y(s) − c(U(s))] ds | X(t) = x, Y(t) = y, U(t) = u – ,

  • V (x, y, u, t) = Expected Current Value of Future

Resources (i.e., exp(δt) times Present Value);

  • {x, y} = 2-Dim State of Inflationary Stochastic

Dynamics ;

  • T = Time Horizon (T ≥ t);
  • δ = Nominal Discount Rate (NOT adjusted for inflation);

Hanson and Ryan — 16 — UIC and McKendree

slide-17
SLIDE 17

2.4. Stochastic Dynamic Programming:

  • Optimization Goal = Maximize Total Return:

v∗(x, y, t) = V (x, y, u∗, t) = max

u

  • V (x, y, u, t)
  • ;
  • PDE of Stochastic Dynamic Programming:

= v∗

t (x, y, t) + r1x(1 − x/K)v∗ x(x, y, t) − δv∗(x, y, t)

+ σ2

1x2

2 v∗

xx +

X

j

fj ˆ v∗ ` (1 + aj)x, y, t ´ − v∗(x, y, t) ˜ + r2yv∗

y + σ2 2y2

2 v∗

yy +

X

j

gj ˆ v∗(x, (1 + bj)y, t) − v∗(x, y, t) ˜ + S∗(x, y, t),

by General Itˆ

  • Chain Rule;
  • Control Switching Term:

S∗(x, y, t) = max

u

ˆp0y + `p1y − v∗

x(x, y, t)´ qux − c1u − c2u2˜ ;

Hanson and Ryan — 17 — UIC and McKendree

slide-18
SLIDE 18

2.4.1. More Stochastic Dynamic Programming:

  • Regular (Unconstrained) Control:

uR(x, y, t) = (p1y − v∗

x (x, y, t))qx − c1

2c2 , c2 > 0;

  • Optimal (Constrained) Control:

u∗(x, y, t) = 8 > > < > > : Umax, Umax ≤ uR(x, y, t) uR(x, y, t), Umin ≤ uR(x, y, t) ≤ Umax Umin, uR(x, y, t) ≤ Umin 9 > > = > > ; ;

  • Final Boundary Condition:

v∗(x, y, T) = 0;

  • Extinction Natural Boundary Condition*:

v∗(0, 0, t) = − (c1 + c2Umin)Umin δ “ 1 − e−δ(T −t)” , δ > 0.

* see Kushner and Dupuis (1992) for proper handling of stochastic reflecting boundary conditions.

Hanson and Ryan — 18 — UIC and McKendree

slide-19
SLIDE 19

Part 3. Numerical Approximations

3.1 Basic Hybrid Numerical Procedures.

  • Extrapolated, Predictor-Corrector for Nonlinear

Iteration.

  • Crank-Nicolson Implicit for 2nd Order in Time and

State.

  • Modifications for Poisson Functional Terms.
  • Modifications for Optimization in Switching Term.

Hanson and Ryan — 19 — UIC and McKendree

slide-20
SLIDE 20

3.2. Numerical Discretizations:

  • State1: Xi ≡ (i − 1)∆x, i = 1, · · · , Nx, ∆x ≡ K/(Nx − 1);
  • State2: Yj ≡ (j − 1)∆y, j = 1, · · · , Ny, ∆y ≡ er2T /(Ny − 1);
  • Time: Tk ≡ T − (k − 1)∆t, k = 1, · · · , Nt, ∆t ≡ T/(Nt − 1);
  • Optimal Expected Value: v∗(xi, yj, tk) −

→ Vi,j,k;

  • v∗

x(Xi, Yj, Tk) −

→ DVXi,j,k ≡ 0.5(Vi+1,j,k − Vi−1,j,k)/∆x;

  • v∗

y (Xi, Yj, Tk) −

→ DVYi,j,k ≡ 0.5(Vi,j+1,k − Vi,j−1,k)/∆y;

  • v∗

xx(Xi, Yj, Tk) −

→ DDVXi,j,k ≡ (Vi+1,j,k − 2Vi,j,k + Vi−1,j,k)/(∆x)2;

  • v∗

yy(Xi, Yj, Tk) −

→ DDVYi,j,k ≡ (Vi,j+1,k − 2Vi,j,k + Vi,j−1,k)/(∆y)2;

  • v∗

t (Xi, Yj, Tk+0.5) −

→ DVTi,j,k ≡ −(Vi,j,k+1 − Vi,j,k)/∆t; with Error: O(∆x)2 + O(∆y)2 + O(∆t/2)2;

Hanson and Ryan — 20 — UIC and McKendree

slide-21
SLIDE 21

3.2.1. More Numerical Discretizations:

  • X-Poisson Term: v∗((1 + al)Xi, Yj, Tk) −

→ ZVi,j,k,l by 2nd order accurate interpolation between nearest nodes;

  • Y-Poisson Term: v∗(Xi, (1 + bl)Yj, Tk) −

→ QVi,j,k,l by 2nd order accurate interpolation between nearest nodes;

  • Regular Control:

uR(Xi, Yj, Tk) − → URi,j,k ≡ (p1Yj −DVXi,j,k ·q ·Xi −c1)/(2c2);

  • Optimal Control:

u∗(Xi, Yj, Tk) − → Ui,j,k ≡ same as exact composite expression;

Hanson and Ryan — 21 — UIC and McKendree

slide-22
SLIDE 22

3.3. Computational Stochastic Dynamic Programming: For k + 1 = 2 to Nt while i = 1 to Nx & j = 1 to Ny:

  • Accelerating Extrapolating Start:

VEi,j,k ≡ 0.5(3V (c,∗)

i,j,k − V (c,∗) i,j,k−1) ≃ Vi,j,k+0.5,

if k ≤ 2, which are used to get components DVXE, DVYE, DDVXE, DDVYE, ZVE, QVE, URE, UE & SE, and where V (c,∗)

i,j,k ) is the

final correction from step k;

  • Extrapolated-Predictor Step:

V (p)

i,j,k+1

= V (c,∗)

i,j,k + ∆t [r1Xi(1 − Xi/K)DVXEi,j,k

+ 1 2σ2

1X2 i DDVXEi,j,k − δVEi,j,k

+ Σlfl(ZVEi,j,k,l − VEi,j,k) + r2YjDVYEi,j,k + 1 2σ2

2Y 2 j DDVYEi,j,k

+ Σlgl(QVEi,j,k,l − VEi,j,k) + SEi,j,k ˜ ,

Hanson and Ryan — 22 — UIC and McKendree

slide-23
SLIDE 23
  • Predictor Evaluation (Crank-Nicolson Midpoint):

VM(p)

i,j,k ≡ 0.5(V (c,∗) i,j,k + V (p) i,j,k+1) ≃ Vi,j,k+0.5,

which are used to get predicted components of DVXM, DVYM, DDVXM, DDVYM, ZVM, QVM, URM, UM & SM;

Hanson and Ryan — 23 — UIC and McKendree

slide-24
SLIDE 24

3.3.1 More Computational Dynamic Programming:

  • (L + 1)st Corrector Step:

V (c,L+1)

i,j,k+1

= V (c,∗)

i,j,k + ∆t

  • r1xi(1 − xi/K)DVXM(c,L)

i,j,k

+ 1 2σ2

1x2 i DDVXM(c,L) i,j,k − δVM(c,L) i,j,k

+ Σlfl

  • ZVM(c,L)

i,j,k,l − VM(c,L) i,j,k

  • +

r2yjDVYM(c,L)

i,j,k + 1

2σ2

2y2 j DDVYM(c,L) i,j,k

+ Σlgl

  • QVM(c,L)

i,j,k,l − VM(c,L) i,j,k

  • + SM(c,L)

i,j,k

  • ,

for L + 1 = 1 to L∗, where VM(c,0)

i,j,k = VM(p) i,j,k; Hanson and Ryan — 24 — UIC and McKendree

slide-25
SLIDE 25
  • Corrector Evaluation:

VM(c,L)

i,j,k = 0.5(V (c,∗) i,j,k + V (c,L) i,j,k+1),

which are used to get corrected components of DVXM, DVYM, DDVXM, DDVYM, ZVM, QVM, URM, UM & SM;

Hanson and Ryan — 25 — UIC and McKendree

slide-26
SLIDE 26

3.3.2 More Computational Dynamic Programming:

  • Corrector Relative Stopping Criterion:

|V (c,L+1)

i,j,k+1

− V (c,L)

i,j,k+1| < ε|V (c,L) i,j,k+1|

for all {i, j} at fixed k + 1 and some relative tolerance ε > 0 with L + 1 = L∗

k and V (c,∗) i,j,k = V (c,L∗

k)

i,j,k

.

  • Mean Temporal-Spatial Mesh Corrector

Convergence Condition: ∆t < 1 2 1

  • (2A/(∆ξ)2)2 + (B/∆ξ)2

, where for example B/∆ξ = 0.5(Bx/∆x + By/∆y) represents some mean reciprocal of state meshes weighted by respective linear comparison coefficients Bx and By. This condition is a combined Parabolic-Hyperbolic (CFL) Mesh Ratio Condition.

Hanson and Ryan — 26 — UIC and McKendree

slide-27
SLIDE 27

Part 4. Numerical Results

Hanson and Ryan — 27 — UIC and McKendree

slide-28
SLIDE 28

0.2 0.4 0.6 0.8 1

y⋅exp(-r2⋅T), Inflationary Price Factor

50 100 150 200 250 300 350 400

V∗(K,y,t), Optimal Value (Million US Dollars)

Figure 4: Optimal current value, V ∗(K, y, t), in millions of USdollars versus

scaled price factor, y · exp(−r2 · T ), with time parameter t = 0.0, 2.0, 4.0, 6.0, 8.0, 10.0 for each curve ordered from top to bottom, respectively, and with population size fixed at carrying capacity x = K. Hanson and Ryan — 28 — UIC and McKendree

slide-29
SLIDE 29

0.2 0.4 0.6 0.8 1

y⋅exp(-r2⋅T), Inflationary Price Factor

0.002 0.004 0.006 0.008 0.01

q⋅E∗(K,y,t)/r1, Optimal Feedback Effort

Figure 5: Optimal feedback effort, q · E∗/r1(K, y, t), in dimensionless form

versus scaled price inflation factor, y · exp(−r2 · T ), with time parameter covering t = 0.0, 2.0, 4.0, 6.0, 8.0, 10.0 for each curve closely spaced from bottom to top, respectively, and with population size fixed at carrying ca- pacity x = K. Hanson and Ryan — 29 — UIC and McKendree

slide-30
SLIDE 30

0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2

r2, Inflationary Price Factor Rate

200 400 600 800 1000 1200

V∗(K,y,0), Optimal Value (Million US Dollars)

Figure 6: Sensitivity of optimal current value, V ∗(K, y, 0), to inflation price

factor rate r2, with curves parameterized by scaled inflation price factor, y · exp(−r2 · T ), ranging from 1.0 at top to 0.2 at bottom in steps of 0.2, with time fixed at initial value t = 0.0, and with population size fixed carrying capacity x = K. Hanson and Ryan — 30 — UIC and McKendree

slide-31
SLIDE 31

Part 5. Conclusions

  • Examined Effects of Random Price Fluctuations on

Optimal Policy and Optimal Return.

  • Successfully Applied Computational Stochastic

Dynamic Programming.

  • Random Price Jumps Strongly Affect Optimal Return.
  • Random Price Jumps have Less Impact on Optimal

Policy.

  • Random Price Jumps needed as Serious Consideration

as Hazardous Environments and other Environmental Effects.

Hanson and Ryan — 31 — UIC and McKendree