18 175 lecture 14 weak convergence and characteristic
play

18.175: Lecture 14 Weak convergence and characteristic functions - PowerPoint PPT Presentation

18.175: Lecture 14 Weak convergence and characteristic functions Scott Sheffield MIT 1 18.175 Lecture 14 Outline Weak convergence Characteristic functions 2 18.175 Lecture 14 Outline Weak convergence Characteristic functions 3 18.175 Lecture 14


  1. 18.175: Lecture 14 Weak convergence and characteristic functions Scott Sheffield MIT 1 18.175 Lecture 14

  2. Outline Weak convergence Characteristic functions 2 18.175 Lecture 14

  3. Outline Weak convergence Characteristic functions 3 18.175 Lecture 14

  4. Convergence results � Theorem: If F n → F ∞ , then we can find corresponding random variables Y n on a common measure space so that Y n → Y ∞ almost surely. � Proof idea: Take Ω = (0 , 1) and Y n = sup { y : F n ( y ) < x } . � Theorem: X n = ⇒ X ∞ if and only if for every bounded continuous g we have Eg ( X n ) → Eg ( X ∞ ). � Proof idea: Define X n on common sample space so converge a.s., use bounded convergence theorem. � Theorem: Suppose g is measurable and its set of discontinuity points has µ X measure zero. Then X n = ⇒ X ∞ implies g ( X n ) = ⇒ g ( X ). � Proof idea: Define X n on common sample space so converge a.s., use bounded convergence theorem. 4 18.175 Lecture 14

  5. Compactness Theorem: Every sequence F n of distribution has subsequence � � converging to right continuous nondecreasing F so that lim F n ( k ) ( y ) = F ( y ) at all continuity points of F . Limit may not be a distribution function. � � Need a “tightness” assumption to make that the case. Say µ n � � are tight if for every E we can find an M so that µ n [ − M , M ] < E for all n . Define tightness analogously for corresponding real random variables or distributions functions. Theorem: Every subsequential limit of the F n above is the � � distribution function of a probability measure if and only if the F n are tight. 5 18.175 Lecture 14

  6. Total variation norm If we have two probability measures µ and ν we define the � � total variation distance between them is || µ − ν || := sup B | µ ( B ) − ν ( B ) | . Intuitively, it two measures are close in the total variation � � sense, then (most of the time) a sample from one measure looks like a sample from the other. Corresponds to L 1 distance between density functions when � � these exist. Convergence in total variation norm is much stronger than � � weak convergence. Discrete uniform random variable U n on (1 / n , 2 / n , 3 / n , . . . , n / n ) converges weakly to uniform random variable U on [0 , 1]. But total variation distance between U n and U is 1 for all n . 6 18.175 Lecture 14

  7. Outline Weak convergence Characteristic functions 7 18.175 Lecture 14

  8. Outline Weak convergence Characteristic functions 8 18.175 Lecture 14

  9. Characteristic functions Let X be a random variable. � � The characteristic function of X is defined by � � itX ]. φ ( t ) = φ X ( t ) := E [ e Recall that by definition e it = cos( t ) + i sin( t ). � � Characteristic function φ X similar to moment generating � � function M X . φ X + Y = φ X φ Y , just as M X + Y = M X M Y , if X and Y are � � independent. And φ aX ( t ) = φ X ( at ) just as M aX ( t ) = M X ( at ). � � ( m ) m ] = i m φ And if X has an m th moment then E [ X (0). � � X Characteristic functions are well defined at all t for all random � � variables X . 9 18.175 Lecture 14

  10. Characteristic function properties φ (0) = 1 � � φ ( − t ) = φ ( t ) � � | φ ( t ) | = | Ee itX | ≤ E | e itX | = 1. � � | φ ( t + h ) − φ ( t ) | ≤ E | e ihX − 1 | , so φ ( t ) uniformly continuous � � on ( −∞ , ∞ ) Ee it ( aX + b ) itb φ ( at ) = e � � 10 18.175 Lecture 14

  11. Characteristic function examples Coin: If P ( X = 1) = P ( X = − 1) = 1 / 2 then � � it + e − it ) / 2 = cos t . φ X ( t ) = ( e That’s periodic. Do we always have periodicity if X is a � � random integer? Poisson: If X is Poisson with parameter λ then � � ∞ − λ λ k e itk = exp( λ ( e it − 1)). φ X ( t ) = k =0 e k ! Why does doubling λ amount to squaring φ X ? � � − t 2 / 2 Normal: If X is standard normal, then φ X ( t ) = e . � � Is φ X always real when the law of X is symmetric about zero? � � Exponential: If X is standard exponential (density e − x on � � (0 , ∞ )) then φ X ( t ) = 1 / (1 − it ). Bilateral exponential: if f X ( t ) = e −| x | / 2 on R then � � φ X ( t ) = 1 / (1 + t 2 ). Use linearity of f X → φ X . 11 18.175 Lecture 14

  12. MIT OpenCourseWare http://ocw.mit.edu 18.175 Theory of Probability Spring 2014 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend