ordinary differential equations initial value problems
play

Ordinary Differential Equations Initial Value Problems The question - PowerPoint PPT Presentation

Ordinary Differential Equations Initial Value Problems The question of whether computers can think is just like the question of whether submarines can swim Edsger W. Dijkstra 1 Fall 2010 Topics to Be Discussed Topics to Be Discussed This


  1. Ordinary Differential Equations Initial Value Problems The question of whether computers can think is just like the question of whether submarines can swim Edsger W. Dijkstra 1 Fall 2010

  2. Topics to Be Discussed Topics to Be Discussed � This unit requires the knowledge of some very This unit requires the knowledge of some very basic ordinary equations. � The following topics will be presented: � The following topics will be presented: � Euler’s method � Heun’s method � Predictor-Corrector methods � Various Runge-Kutta methods with an emphasis on the second order methods emphasis on the second order methods 2

  3. What Is an Initial Value Problem? What Is an Initial Value Problem? � A first order differential equation has a form A first order differential equation has a form like this: = ' y y f x y f x y ( , ) ( ) � Here, y is a function of x , and y ’ is the derivative of y . Function f ( x , y ) gives the relation of x and y . f F ti f ( ) i th l ti f d � However, solution to the above ODE is not unique, and an initial condition is required: = = ' y x ( ) y y f x y ( , ) and 0 0 � The problem is: find find y given given y ( x 0 ) = y 0 step- step- by-ste y -step in an interval p in an interval [ x 0 , x n ] . [ 0 , n ] 3

  4. Euler’s Method: 1/6 Euler s Method: 1/6 � The easiest method is perhaps Euler’s. The easiest method is perhaps Euler s. � Since y ’ = f ( x , y ) and y 0 = y ( x 0 ), we know the tangent slope of y at ( x y ) i e y ’ = f ( x y ) tangent slope of y at ( x 0 , y 0 ), i.e ., y 0 = f ( x 0 , y 0 ). � If x 1 = x 0 + ∆ and ∆ is small enough, we may predict y ( x + ∆ ) using the tangent! ( + ∆ ) predict predict predict i th t t! y 0 ’= f ( x 0 , y 0 ) unknown y ( x ) the predicted y 1 h di d exact solution x 0 + ∆ x 0 4

  5. Euler’s Method: 2/6 Euler s Method: 2/6 predict” y ( x + ∆ ) ? predict y ( x ∆ ) ? � How How do w e “predict How How do do w e w e predict � Using the forward difference method for differentiation y ’( x ) can be approximated as differentiation, y ( x 0 ) can be approximated as + ∆ − ( ) ( ) y x y x ≈ ≈ 0 0 y x y x '( ( ) ) ∆ 0 � Replacing y ’( x 0 ) with f ( x 0 , y ( x 0 )) yields: 0 0 0 + ∆ − y x ( ) y x ( ) ≈ 0 0 f x ( , ( y x )) ∆ ∆ 0 0 � Therefore, the prediction is: + ∆ ≈ + ∆× y x ( ) y x ( ) f x ( , ( y x )) 0 0 0 0 5

  6. Euler’s Method: 3/6 Euler s Method: 3/6 � Recall the following formula: Recall the following formula: + ∆ ≈ + ∆× y x ( ) y x ( ) f x ( , ( y x )) 0 0 0 0 � W � We may start with an initial value ( x 0 , y 0 ). t t ith i iti l l ( ) Compute x 1 = x 0 + ∆ and y 1 = y 0 + ∆ × f ( x 0 , y 0 ). � Then, compute ( x 2 , y 2 ) from ( x 1 , y 1 ), etc. � The algorithm below traces out the locus of y . ∆ = (b-a)/n ! Initial values ! [a,b] x = a ! x 0 = a y = y(a) ! y 0 = y(a) = y(x 0 ) DO i = 1, n y = y + ∆ × f(x,y) ! n: # of subdivisions x = x + ∆ END 6

  7. Euler’s Method: 4/6 Euler s Method: 4/6 � Consider y ’ = x + y and y (0)=2 on [0,1]. Consider y x y and y (0) 2 on [0,1]. � Initial value is x 0 =0 and y 0 = y ( x 0 )=2. � If n = 4, ∆ =(1-0)/4=0.25 4 ∆ (1 0)/4 0 25 � If � Predicted (next) y = y + ∆ ( x + y ). Pred. y = y + ∆ ( x + y ) y ’= f ( x , y ) x y 2 5 2+0 25 × (0+2) 2.5=2+0.25 × (0+2) Init Init 0 0 2 0 2.0 2 0 2.0 3.1875=2.5+0.25 × 2.75 1 0.25 2.5 2.75 4 109375=3 1875+0 25 × 3 6875 4.109375 3.1875+0.25 × 3.6875 2 2 0 5 0.5 3 1875 3.1875 3 6875 3.6875 5.3245187=4.109375+0.25 × 4.859375 3 0.75 4.109375 4.859375 4 4 1 1 5 3245187 5.3245187 7

  8. Euler’s Method: 5/6 Euler s Method: 5/6 � If ∆ is sufficiently small, Euler’s method usually If ∆ is sufficiently small, Euler s method usually works fine. � However Euler’s method can produce large � However, Euler s method can produce large under- or over- predicted y . this is the expected solution hi i h d l i 8 x 0 x 1 x 2 x 3

  9. Euler’s Method: 6/6 Euler s Method: 6/6 � Consider y ’ = 1- y on [0,1] with y (0)=0. Consider y 1 y on [0,1] with y (0) 0. � The solution is y =1-e - x . � Th � The absolute error is increasing. b l t i i i exact y |error| % of exact x y 1 1/6 0.1666667 0.1535182 0.01314841 8.56 2 1/3 0.3055556 0.2834687 0.02208686 7.79 3 1/2 0.4212963 0.3934693 0.02782699 7.07 4 2/3 0.5177469 0.4865829 0.03116405 6.40 5 5/6 0.5981224 0.5654018 0.03272063 5.79 6 1 0.6651020 0.6321206 0.03298146 5.22 9

  10. Heun’s Method: 1/4 Heun s Method: 1/4 � Heun’s method tries to reduce the over- or Heun s method tries to reduce the over or under- shot of Euler’s method. � Since y ( x + ∆ ) may be inaccurate one may use � Since y ( x + ∆ ) may be inaccurate, one may use the average of y ’( x ) and y ’( x + ∆ ) as a corrector. y ’( x + ∆ ) ( y ’( x )+ y ’( x + ∆ ))/2 ( y ( x ) y ( x ∆ ))/2 y ’= f ( x y ) y = f ( x,y ) y ( x )+ ∆ × f ( x , y ) 10 x + ∆ x

  11. Heun s Method: 2/4 Heun’s Method: 2/4 � Heun’s method consists of two steps: Heun s method consists of two steps: Step 1: Use x and ∆ to compute y P ( x + ∆ ) � Step 1: with Euler’s method This is a predictor : with Euler s method. This is a predictor : + ∆ = + ∆ P y ( x ) y x ( ) f x y ( , ) tep 2: Use the ( y ’( x ) + y ’( x + ∆ ))/2 as the � Step S S 2 2 new derivative to correct over- or under- shot This is a corrector : + + + ∆ + ∆ + ∆ + ∆ P f x y f x y ( ( , ) ) f x f x ( ( , y y ( ( x x )) )) + ∆ = + ∆ y x ( ) y x ( ) 2 11

  12. Heun’s Method: 3/4 Heun s Method: 3/4 � Heun’s method is just a little more complex j p than Euler’s method. ∆ = (b-a)/n ! initialization ! [a,b]: ! [a,b]: interval te a x = a a ! x 0 = a y = y(a) ! y 0 = y(a) = y(x 0 ) DO i = 1, n y pred = y + ∆ × f(x y) ! Euler’s = y + ∆ × f(x,y) ! Euler s ! n : intervals ! n : intervals y p y = y + ∆ × (f(x,y) + f(x+ ∆ ,y pred ))/2 x = x + ∆ END DO END DO 12

  13. Heun s Method: 4/4 Heun’s Method: 4/4 � Consider y ’ = 1- y on [0,1] with y (0)=0, again. Consider y 1 y on [0,1] with y (0) 0, again. � The solution is y =1-e - x . � The absolute error is still increasing, but � Th b l t i till i i b t significantly smaller. exact y |error| % of exact x y 1 1/6 0.15277778 0.15351826 0.00074048 0.48 2 1/3 0.28221452 0.28346872 0.00125420 0.44 3 1/2 0.39187619 0.39346933 0.00159314 0.40 4 2/3 0.48478401 0.48658288 0.00179887 0.37 5 5/6 0.56349754 0.56540179 0.00190425 0.34 6 1 0.63018543 0.63212055 0.00193512 0.31 13

  14. Predictor Corrector Methods: 1/5 Predictor-Corrector Methods: 1/5 � Heun’s method is a special and the simplest Heun s method is a special and the simplest form of the predictor-corrector methods. � If we look at the corrector step and remove the � If we look at the corrector step and remove the superscript P from the predictor, we have + + ∆ + ∆ f x y ( , ) f x ( , ( y x )) + ∆ = + ∆ ( ) ( ) y x y x 2 � y ( x + ∆ ) is the unknown value that we want to compute, and it appears in both sides of the equation. 14

  15. Predictor-Corrector Methods: 2/5 Predictor Corrector Methods: 2/5 � To make this observation clearer, replacing p g y ( x + ∆ ) with a z yields: + + ∆ f x y ( , ) f x ( , ) z = y x + ∆ + ∆ z z y x ( ) ( ) 2 � Since z is an unknown, the above is an equation of variable z , and the desired z is a root! f i bl d th d i d i t! � If such a z can be found, say z *, the equation is perfectly balanced as follows: perfectly balanced as follows: + + ∆ * f x y ( , ) f x ( , z ) = + ∆ * z y x ( ) 2 2 � Therefore, this predictor-corrector method reduces to root finding. educes o oo d g. 15

  16. Predictor Corrector Methods: 3/5 Predictor-Corrector Methods: 3/5 � How to solve the following equation in z ? g q z + + ∆ f x y ( , ) f x ( , ) z = + ∆ ( ) z y x 2 2 � Since it is not easy to find the derivative with respect to z , Newton’s method is out. respect to z , Newton s method is out. � However, it is in a perfect form of fixed-point iteration ( i.e ., z = g ( z )). iteration ( i.e ., z g ( z )). � We may use y P as an initial value for z , followed by a number of iterations of fixed-point by a number of iterations of fixed point iteration to find a better corrector y C . � Note that fixed-point iteration method does not � Note that fixed point iteration method does not always converge. 16

  17. Predictor-Corrector Methods: 4/5 Predictor Corrector Methods: 4/5 � The following is the iterative computation for a g p better corrector with the fixed-point iteration. � MAX is the maximum number of iterations, and ε is a tolerance value. i t l l � There are more pow erful predictor- There are more pow erful predictor- corrector corrector methods corrector methods corrector methods methods. y pred = y + ∆ × f(x,y) y y × ( ,y) DO k = 1, MAX y correct = y + ∆ × (f(x,y) + f(x+ ∆ ,y pred ))/2 ! Fixed-point IF (|(y correct -y pred )/y correct | < ε ) EXIT | < ε ) EXIT IF (|(y -y )/y y pred = y correct END DO y y correct y = y correct 17

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend