stat 380 lecture 37 2019 04 05 introduction to brownian
play

STAT 380 Lecture 37 2019/04/05 Introduction to Brownian Motion - PDF document

STAT 380 Lecture 37 2019/04/05 Introduction to Brownian Motion Physical motivation: Pollen grains under microscope. Observed to jump about. Explanation: tiny grains hit by individual water molecules. Random fluctuation in net impulse due


  1. STAT 380 Lecture 37 – 2019/04/05 Introduction to Brownian Motion Physical motivation: Pollen grains under microscope. Observed to jump about. Explanation: tiny grains hit by individual water molecules. Random fluctuation in net impulse due to collisions. Particles too small for Law of Large Num- bers to hold them still. 1

  2. Brownian Motion For fair random walk Y n = number of heads minus number of tails, Y n = U 1 + · · · + U n where the U i are independent and P ( U i = 1) = P ( U i = − 1) = 1 2 Notice: E( U i ) = 0 Var( U i ) = 1 Recall central limit theorem: U 1 + · · · + U n ⇒ N (0 , 1) √ n Now: rescale time axis so that n steps take 1 time unit and vertical axis so step size is 1 / √ n . 2

  3. n=16 n=64 n=256 n=1024 3

  4. We now turn these pictures into a stochas- tic process: n ≤ t < k +1 For k we define n X n ( t ) = U 1 + · · · + U k √ n Notice: E( X n ( t )) = 0 and Var( X n ( t )) = k n As n → ∞ with t fixed we see k/n → t . Moreover: U 1 + · · · + U k � n √ = kX n ( t ) k converges to N (0 , 1) by the central limit theorem. Thus X n ( t ) ⇒ N (0 , t ) 4

  5. X n ( t + s ) − X n ( t ) independent of X n ( t ): 2 rvs involve sums of different U i . As n → ∞ the processes X n converge to a process X with the properties: 1. X ( t ) has a N (0 , t ) distribution. 2. X has independent increments: if 0 = t 0 < t 1 < t 2 < · · · < t k then X ( t 1 ) − X ( t 0 ) , . . . , X ( t k ) − X ( t k − 1 ) are independent. 3. Increments are stationary : for all s X ( t + s ) − X ( s ) ∼ N (0 , t ) 4. X (0) = 0. 5

  6. Definition : Any process satisfying 1-4 above is a Brownian motion. Properties of Brownian motion Suppose t > s . Then E( X ( t ) | X ( s )) = E { X ( t ) − X ( s ) + X ( s ) | X ( s ) } = E { X ( t ) − X ( s ) | X ( s ) } + E { X ( s ) | X ( s ) } = 0 + X ( s ) = X ( s ) Notice the use of independent increments and of E( Y | Y ) = Y . • Again if t > s : Var { X ( t ) | X ( s ) } = Var { X ( t ) − X ( s ) + X ( s ) | X ( s ) } = Var { X ( t ) − X ( s ) | X ( s ) } = Var { X ( t ) − X ( s ) } = t − s 6

  7. If t < s then X ( s ) = X ( t ) + { X ( s ) − X ( t ) } is sum of 2 independent normals. So: X ∼ N (0 , σ 2 ), and Y ∼ N (0 , τ 2 ) indepen- dent. Z = X + Y . Conditional distribution of X given Z : f X | Z ( x | z ) = f X,Z ( x, z ) f Z ( z ) = f X,Y ( x, z − x ) f Z ( z ) = f X ( x ) f Y ( z − x ) f Z ( z ) Now Z is N (0 , γ 2 ) where γ 2 = σ 2 + τ 2 so 2 π e − x 2 / (2 σ 2 ) 2 π e − ( z − x ) 2 / (2 τ 2 ) 1 1 √ √ σ τ f X | Z ( x | z ) = 2 π e − z 2 / (2 γ 2 ) 1 √ γ γ 2 π exp {− ( x − a ) 2 / (2 b 2 ) } √ = τσ for suitable choices of a and b . To find them compare coefficients of x 2 , x and 1. 7

  8. Coefficient of x 2 : b 2 = 1 1 σ 2 + 1 τ 2 so b = τσ/γ . Coefficient of x : b 2 = z a τ 2 so that σ 2 a = b 2 z/τ 2 = σ 2 + τ 2 z Finally you should check that a 2 b 2 = z 2 τ 2 − z 2 γ 2 to make sure the coefficients of 1 work out as well. Conclusion: given Z = z the conditional distribution of X is N ( a, b 2 ) with a and b as above. 8

  9. Application to Brownian motion: • For t < s let X be X ( t ) and Y be X ( s ) − X ( t ) so Z = X + Y = X ( s ). Then σ 2 = t , τ 2 = s − t and γ 2 = s . Thus b 2 = ( s − t ) t s and a = t sX ( s ) SO: E( X ( t ) | X ( s )) = t sX ( s ) and Var( X ( t ) | X ( s )) = ( s − t ) t s 9

  10. The Reflection Principle Tossing a fair coin: 5 more HTHHHTHTHHTHHHTTHTH heads than tails 5 more THTTTHTHTTHTTTHHTHT tails than heads Both sequences have the same probability. 10

  11. So: starting at stopping time: Sequence with k more heads than tails in next m tosses matched to sequence with k more tails than heads. Both sequences have same prob. Suppose Y n is a fair ( p = 1 / 2) random walk. Define M n = max { Y k , 0 ≤ k ≤ n } Compute P ( M n ≥ x )? Trick: Compute P ( M n ≥ x, Y n = y ) 11

  12. First: if y ≥ x then { M n ≥ x, Y n = y } = { Y n = y } Second: if M n ≥ x then T ≡ min { k : Y k = x } ≤ n Fix y < x . Consider a sequence of H’s and T’s which leads to say T = k and Y n = y . Switch the results of tosses k + 1 to n to get a sequence of H’s and T’s which has T = k and Y n = x + ( x − y ) = 2 x − y > x . This proves P ( T = k, Y n = y ) = P ( T = k, Y n = 2 x − y ) 12

  13. This is true for each k so P ( M n ≥ x, Y n = y ) = P ( M n ≥ x, Y n = 2 x − y ) = P ( Y n = 2 x − y ) Finally, sum over all y to get � P ( M n ≥ x ) = P ( Y n = y ) y ≥ x � + P ( Y n = 2 x − y ) y<x Make the substitution k = 2 x − y in the second sum to get � P ( M n ≥ x ) = P ( Y n = y ) y ≥ x � + P ( Y n = k ) k>x � = 2 P ( Y n = k ) + P ( Y n = x ) k>x 13

  14. Brownian motion version: M t = max { X ( s ); 0 ≤ s ≤ t } T x = min { s : X ( s ) = x } (called hitting time for level x ). Then { T x ≤ t } = { M t ≥ x } Any path with T x = s < t and X ( t ) = y < x is matched to an equally likely path with T x = s < t and X ( t ) = 2 x − y > x . So for y > x P ( M t ≥ x, X ( t ) > y ) = P ( X ( t ) > y ) while for y < x P ( M t ≥ x, X ( t ) < y ) = P ( X ( t ) > 2 x − y ) 14

  15. 15

  16. Let y → x to get P ( M t ≥ x, X ( t ) > x ) = P ( M t ≥ x, X ( t ) < x ) = P ( X ( t ) > x ) Adding these together gives P ( M t > x ) = 2 P ( X ( t ) > x ) √ = 2 P ( N (0 , 1) > x/ t ) Hence M t has the distribution of | N (0 , t ) | . 16

  17. On the other hand in view of { T x ≤ t } = { M t ≥ x } the density of T x is √ d dt 2 P ( N (0 , 1) > x/ t ) Use the chain rule to compute this. First d dyP ( N (0 , 1) > y ) = − φ ( y ) where φ is the standard normal density φ ( y ) = e − y 2 / 2 √ 2 π because P ( N (0 , 1) > y ) is 1 minus the stan- dard normal cdf. 17

  18. So √ d dt 2 P ( N (0 , 1) > x/ t ) √ √ t ) d = − 2 φ ( x/ dt ( x/ t ) x 2 πt 3 / 2 exp {− x 2 / (2 t ) } = √ This density is called the Inverse Gaussian density. T x is called a first passage time NOTE: the preceding is a density when viewed as a function of the variable t . Martingales A stochastic process M ( t ) indexed by ei- ther a discrete or continuous time param- eter t is a martingale if: E { M ( t ) | M ( u ); 0 ≤ u ≤ s } = M ( s ) whenever s < t . 18

  19. Examples • A fair random walk is a martingale. • If N ( t ) is a Poisson Process with rate λ then N ( t ) − λt is a martingale. • Standard Brownian motion (defined above) is a martingale. Note: Brownian motion with drift is a pro- cess of the form X ( t ) = σB ( t ) + µt where B is standard Brownian motion, in- troduced earlier. X is a martingale if µ = 0. We call µ the drift 19

  20. • If X ( t ) is a Brownian motion with drift then Y ( t ) = e X ( t ) is a geometric Brownian motion. For suitable µ and σ we can make Y ( t ) a martingale. • If a gambler makes a sequence of fair bets and M n is the amount of money s/he has after n bets then M n is a mar- tingale – even if the bets made depend on the outcomes of previous bets, that is, even if the gambler plays a strategy. 20

  21. Some evidence for some of the above: Random walk: U 1 , U 2 , . . . iid with P ( U i = 1) = P ( U i = − 1) = 1 / 2 and Y k = U 1 + · · · + U k with Y 0 = 0. Then E( Y n | Y 0 , . . . , Y k ) = E( Y n − Y k + Y k | Y 0 , . . . , Y k ) = E( Y n − Y k | Y 0 , . . . , Y k ) + Y k n � = E( U j | U 1 , . . . , U k ) + Y k k +1 n � = E( U j ) + Y k k +1 = Y k 21

  22. Things to notice: Y k treated as constant given Y 1 , . . . , Y k . Knowing Y 1 , . . . , Y k is equivalent to knowing U 1 , . . . , U k . For j > k we have U j independent of U 1 , . . . , U k so conditional expectation is unconditional expectation. Since Standard Brownian Motion is limit of such random walks we get martingale property for standard Brownian motion. 22

  23. Poisson Process : X ( t ) = N ( t ) − λt . Fix t > s . E( X ( t ) | X ( u ); 0 ≤ u ≤ s ) = E( X ( t ) − X ( s ) + X ( s ) |H s ) = E( X ( t ) − X ( s ) |H s ) + X ( s ) = E( N ( t ) − N ( s ) − λ ( t − s ) |H s ) + X ( s ) = E( N ( t ) − N ( s )) − λ ( t − s ) + X ( s ) = λ ( t − s ) − λ ( t − s ) + X ( s ) = X ( s ) Things to notice: I used independent increments. H s is shorthand for the conditioning event. Similar to random walk calculation. 23

  24. Black Scholes We model the price of a stock as X ( t ) = x 0 e Y ( t ) where Y ( t ) = σB ( t ) + µt is a Brownian motion with drift ( B is stan- dard Brownian motion). If annual interest rates are e α − 1 we call α the instantaneous interest rate; if we invest $1 at time 0 then at time t we would have e αt . In this sense an amount of money x ( t ) to be paid at time t is worth only e − αt x ( t ) at time 0 (because that much money at time 0 will grow to x ( t ) by time t ). 24

  25. Present Value : If the stock price at time t is X ( t ) per share then the present value of 1 share to be delivered at time t is Z ( t ) = e − αt X ( t ) With X as above we see Z ( t ) = x 0 e σB ( t )+( µ − α ) t Now we compute E { Z ( t ) | Z ( u ); 0 ≤ u ≤ s } = E { Z ( t ) | B ( u ); 0 ≤ u ≤ s } for s < t . Write Z ( t ) = x 0 e σB ( s )+( µ − α ) t × e σ ( B ( t ) − B ( s )) Since B has independent increments we find E { Z ( t ) | B ( u ); 0 ≤ u ≤ s } = x 0 e σB ( s )+( µ − α ) t × E � e σ { B ( t ) − B ( s ) } � 25

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend