optimal estimation of certain random quantities
play

Optimal estimation of certain random quantities Mark Podolskij Risk - PowerPoint PPT Presentation

Optimal estimation of certain random quantities Mark Podolskij Risk and Statistics, Ulm joint work with J. Ivanovs Aarhus University, Denmark 1 Topic of the talk Let ( X t ) t [0 , 1] be a stochastic process (Brownian motion, Lvy


  1. Optimal estimation of certain random quantities Mark Podolskij Risk and Statistics, Ulm joint work with J. Ivanovs Aarhus University, Denmark 1

  2. Topic of the talk Let ( X t ) t ∈ [0 , 1] be a stochastic process (Brownian motion, Lévy process, SDE etc.). Given the observations X 0 , X ∆ n , X 2∆ n , . . . , X ⌊ 1 / ∆ n ⌋ ∆ n with ∆ n → 0 and the random parameter of interest Q , what is the optimal estimator of Q ? 2

  3. Low vs. high frequency data Low frequency data High frequency data Observed data Observed data X 0 ( ω ) , X ∆ n ( ω ) , ..., X ⌊ 1 / ∆ n ⌋ ∆ n ( ω ) X 1 , X 2 , ..., X n i . i . d . ∼ F Asymptotic knowledge Asymptotic knowledge ( X t ( ω )) t ∈ [0 , 1] distribution function F Identifiable objects Identifiable objects functionals of ( X t ( ω )) t ∈ [0 , 1] functionals of F 3

  4. Background � In the classical test theory the model parameters are deterministic objects. There exist numerous approaches to access the optimality of estimators: Cramer-Rao bounds, maximum likelihood theory, minimax approach, Le Cam theory, etc. � However, in the high frequency setting the objects of interests are often random . Examples include quadratic variation, realised jumps, supremum/infimum of a process, local times, occupation time measures etc. � In this framework very little is known about how to construct optimal estimates. 4

  5. Example: Estimation of the quadratic variation Let X be a continuous semimartingale of the form � t � t t ≥ 0 X t = X 0 + a s ds + σ s dW s 0 0 where a and σ are stochastic processes, and W is a Brownian motion. An important result in the theory of high frequency data is the following theorem. Theorem (Jacod(94)) It holds that   � 1 � � 1 � ⌊ 1 / ∆ n ⌋ � � � 2 −  d st ∆ − 1 / 2  σ 2 σ 4 X i ∆ n − X ( i − 1)∆ n s ds → MN 0 , 2 s ds n 0 0 i =1 Recently, Clement, Delattre & Gloter (13) have proved that the above estimator is asymptotically efficient applying an infinite dimensional LAMN property. 5

  6. Introduction � The results of Clement, Delattre & Gloter (13) only cover estimation problems for volatility functionals. In this talk we will rather focus on the following random objects: X := sup X s s ∈ [0 , 1] � 1 1 l ( x ) := lim 1 ( − ǫ,ǫ ) ( X s − x ) ds 2 ǫ ǫ ↓ 0 0 � 1 L ( x ) := 1 ( x , ∞ ) ( X s ) ds 0 which is the supremum, local time and occupation time measure of the process X , respectively. � We are interested in optimal estimation of these objects given high frequency data ( X i ∆ n ) 0 ≤ i ≤⌊ 1 / ∆ n ⌋ . 6

  7. A remark on optimality We will see that many naive estimators are rate optimal, but not efficient! In fact, efficient estimators are easy to introduce. Let Q = Φ(( X s ) s ∈ [0 , 1] ) be a random variable of interest. An optimal estimator of Q is given as (i) in L 2 -sense: E [ Q | ( X i ∆ n ) 0 ≤ i ≤⌊ 1 / ∆ n ⌋ ] (ii) in L 1 -sense: median[ Q | ( X i ∆ n ) 0 ≤ i ≤⌊ 1 / ∆ n ⌋ ] We will investigate the asymptotic theory for these type of estimates in the setting of supremum, local time and occupation time measure of the process X , where X is a Brownian motion, stable Lévy process or a continuous diffusion process. 7

  8. Naive estimator of the supremum � It is rather simple to propose the following estimate for the supremum P M n := i =1 ,..., ⌊ 1 / ∆ n ⌋ X i ∆ n max → X where the consistency holds for all Lévy processes X . � The asymptotic theory for the maximum has been studied in several papers including Asmussen, Glynn & Pitman (95) (Brownian motion) and Ivanovs (18) (general Lévy processes). � Since M n < X , the estimator M n is downward biased and there were several attempts to correct the bias. 8

  9. A result on zooming-in at supremum The following result from the theory of Lévy processes will be extremely useful for our asymptotic theory. Theorem (Ivanovs (18)) Let X be an α -stable Lévy process with α ∈ (0 , 2] . Denote by τ the time of the supremum of X on the interval [0 , 1] . Then we obtain the functional stable convergence � � � � d st � ( Z n ∆ − 1 /α ( X τ + t ∆ n − X τ ) → t ) t ∈ R := X t n t ∈ R t ∈ R where � X is the so called Lévy process conditioned to stay negative , which is independent of F . When X is a Brownian motion, we deduce the identity � X t = −� B t � where B is a 3 -dimensional Brownian motion. 9

  10. Application to estimation of the supremum The previous result has the following consequence. Theorem (Ivanovs (18)) Let X be an α -stable Lévy process with α ∈ (0 , 2] . Then it holds that � � d j ∈ Z ( � ∆ − 1 /α M n − X → max X j + U ) n where U ∼ U (0 , 1) is independent of � X and F . Sketch of proof: Note that � � ∆ − 1 /α = Z n X ( ⌈ τ/ ∆ n ⌉ + i )∆ n − X τ n i + { τ/ ∆ n } → U ∼ U (0 , 1). Since Z n d st d st → � Recall that { τ/ ∆ n } X , we conclude that � � d j ∈ Z ( � ∆ − 1 /α M n − X → max X j + U ) n ✷ 10

  11. Computation of the optimal estimator: The Brownian case The basis of our approach is the computation of the conditional probability � � H n ( x ) := P X ≤ x | ( X i ∆ n ) 0 ≤ i ≤⌊ 1 / ∆ n ⌋ x > 0 . Due to Markov and self-similarity property of X , we easily see that � � � n ∆ − 1 / 2 n ) , ∆ − 1 / 2 ∆ n H n ( x ) = ( x − X i − 1 F i X n n i =1 � � where F ( x , y ) = P X ≤ x | X 1 = y = 1 − exp( − 2 x ( x − y )). After rescaling we deduce the stable convergence � � � � � ∆ 1 / 2 x + ∆ − 1 / 2 ( M n − X ( i − 1)∆ n ) , ∆ − 1 / 2 ∆ n H n x + M n = F i X n n n i ∈ Z � � � d st X j + U − � � X i + U , � X i +1+ U − � → G ( x ) := x + max . F X i + U j ∈ Z i ∈ Z 11

  12. Conditional mean and conditional median � � � For the conditional mean T (2) := E X | ( X i ∆ n ) i we obtain the n formula � ∞ � � �� T (2) − X = ( M n − X ) + ∆ 1 / 2 ∆ 1 / 2 1 − H n x + M n dx n n n 0 Hence, the probabilistic structure of X only affects the second order term. � � � Similarly, for the conditional median T (1) X | ( X i ∆ n ) i := median we n deduce the identity � � − 1 T (1) − X = ( M n − X ) + ∆ 1 / 2 ∆ 1 / 2 H n · + M n (1 / 2) n n n and again the probabilistic structure of X only affects the second order term. 12

  13. Asymptotic theory for the optimal estimators: Brownian case Theorem (Ivanovs & P. (19)) Define the estimates � � � � T (1) T (2) X | ( X i ∆ n ) i X | ( X i ∆ n ) i = median , = E . n n (i) It holds that � � d j ∈ Z ( � ∆ − 1 / 2 T (1) X j + U ) + G − 1 (1 / 2) . − X → max n n (ii) Furthermore, � ∞ � � d j ∈ Z ( � ∆ − 1 / 2 T (2) − X → max (1 − G ( y )) dy . X j + U ) + n n 0 In particular, we have that MSE ( M n ) ≈ 6 . 25 ! MSE ( T (2) n ) 13

  14. Simulation of asymptotic distributions 1.5 1.0 type density conditional expectation conditional median old 0.5 0.0 0 1 2 error 14

  15. Asymptotic theory: The α -stable case Theorem (Ivanovs & P. (19)) Let X be a α -stable Lévy motion with α ∈ (0 , 2) . (i) Define T (1) = median [ X | ( X i ∆ n ) i ] . Then we obtain n � � d j ∈ Z ( � ∆ − 1 /α T (1) X j + U ) + G − 1 (1 / 2) . − X → max n n and the estimator is L 1 -optimal for α ∈ (1 , 2) . (ii) Define T (2) = E [ X | ( X i ∆ n ) i ] for α ∈ (1 , 2) . Then it holds that n � ∞ � � d j ∈ Z ( � ∆ − 1 /α T (2) − X → max X j + U ) + (1 − G ( y )) dy . n n 0 15

  16. Naive estimators for the local time � In this chapter we assume that X is a Brownian motion. Recall the definition of local time: � 1 1 l ( x ) = lim 1 ( − ǫ,ǫ ) ( X s − x ) ds 2 ǫ ǫ ↓ 0 0 where x ∈ R . � A straightforward estimator of l ( x ) is given as ⌊ 1 / ∆ n ⌋ � P l n ( x ) := a n ∆ n g ( a n ( X i ∆ n − x )) → l ( x ) i =1 � where g is a kernel satisfying R g ( x ) dx = 1, and a n → ∞ with a n ∆ n → 0. � We will focus on a more general class of statistics: ⌊ 1 / ∆ n ⌋ � � � V ( h , x ) n := a n ∆ n a n ( X i ∆ n − x ) , ∆ − 1 / 2 ∆ n h i X n i =1 16

  17. Asymptotic theory for V ( h , x ) n Theorem (Borodin (86), Jacod (98)) Assume that a n = ∆ − 1 / 2 and h satisfies the condition n � R | y | p h 1 ( y ) dy < ∞ for | h ( y , z ) | ≤ h 1 ( y ) exp( λ | z | ) for some λ > 0 and some p > 3 . Then it holds that P V ( h , x ) n → c h l ( x ) � �� � where c h = R h ( y , z ) ϕ ( z ) dz dy and ϕ denotes the density of the R standard normal distribution. Furthermore, we obtain the stable convergence ( V ( h , x ) n − c h l ( x )) d st ∆ − 1 / 4 → MN (0 , v h l ( x )) n for a certain constant v h > 0 . An interesting example is the number of crossings at level 0 which corresponds to x = 0 and h ( y , z ) = 1 ( −∞ , 0) ( y ( y + z )). 17

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend