quantitative clts via martingale embeddings
play

Quantitative CLTs via Martingale Embeddings Dan Mikulincer Weizmann - PowerPoint PPT Presentation

Quantitative CLTs via Martingale Embeddings Dan Mikulincer Weizmann Institute of Science Joint work with Ronen Eldan and Alex Zhai 1 The central limit theorem i =1 be i.i.d. copies of a random vector X in R d with Let { X i } E [ X ] = 0


  1. Quantitative CLTs via Martingale Embeddings Dan Mikulincer Weizmann Institute of Science Joint work with Ronen Eldan and Alex Zhai 1

  2. The central limit theorem i =1 be i.i.d. copies of a random vector X in R d with Let { X i } ∞ E [ X ] = 0 and Cov ( X ) = Σ . n 1 If S n := � X i and G ∼ N (0 , Σ) then √ n i =1 S n − → n →∞ G , in an appropriate sense. • We usually normalize X to be isotropic, that is, Σ = I d . • We are interested in bounding the convergence rate. 2

  3. The central limit theorem i =1 be i.i.d. copies of a random vector X in R d with Let { X i } ∞ E [ X ] = 0 and Cov ( X ) = Σ . n 1 If S n := � X i and G ∼ N (0 , Σ) then √ n i =1 S n − → n →∞ G , in an appropriate sense. • We usually normalize X to be isotropic, that is, Σ = I d . • We are interested in bounding the convergence rate. 2

  4. The central limit theorem i =1 be i.i.d. copies of a random vector X in R d with Let { X i } ∞ E [ X ] = 0 and Cov ( X ) = Σ . n 1 If S n := � X i and G ∼ N (0 , Σ) then √ n i =1 S n − → n →∞ G , in an appropriate sense. • We usually normalize X to be isotropic, that is, Σ = I d . • We are interested in bounding the convergence rate. 2

  5. Quantitative central limit theorem Berry-Esseen is an early examples of a quantitative bound. Theorem (Berry-Esseen) In the 1 -dimensional case, for any t ∈ R , � | X | 3 � | P ( S n ≤ t ) − P ( G ≤ t ) | ≤ E √ n . This estimate is sharp. 3

  6. Quantitative central limit theorem Berry-Esseen is an early examples of a quantitative bound. Theorem (Berry-Esseen) In the 1 -dimensional case, for any t ∈ R , � | X | 3 � | P ( S n ≤ t ) − P ( G ≤ t ) | ≤ E √ n . This estimate is sharp. 3

  7. Quantitative central limit theorem In higher dimensions the current best known result is due to Bentkus. Theorem (Bentkus, 2003) In the d-dimensional case, for any convex set K ⊂ R d , 1 � || X || 3 � 4 E | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d √ n . 1 4 term is the maximal Gaussian surface area of a convex • The d set in R d . If K ε is the ε enlargement of K then P ( G ∈ K ε \ K ) ≤ 4 ε d 1 4 . 1 4 remains an open question. • Whether one can omit d 4

  8. Quantitative central limit theorem In higher dimensions the current best known result is due to Bentkus. Theorem (Bentkus, 2003) In the d-dimensional case, for any convex set K ⊂ R d , 1 � || X || 3 � 4 E | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d √ n . 1 4 term is the maximal Gaussian surface area of a convex • The d set in R d . If K ε is the ε enlargement of K then P ( G ∈ K ε \ K ) ≤ 4 ε d 1 4 . 1 4 remains an open question. • Whether one can omit d 4

  9. Quantitative central limit theorem In higher dimensions the current best known result is due to Bentkus. Theorem (Bentkus, 2003) In the d-dimensional case, for any convex set K ⊂ R d , 1 � || X || 3 � 4 E | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d √ n . 1 4 term is the maximal Gaussian surface area of a convex • The d set in R d . If K ε is the ε enlargement of K then P ( G ∈ K ε \ K ) ≤ 4 ε d 1 4 . 1 4 remains an open question. • Whether one can omit d 4

  10. Other metrics We consider stronger notions of distance Definition (Relative entropy between X and G ) Ent ( X || G ) := E [ln( f ( X ))] , where f is the density of X with respect to G . Definition (Wasserstein distance between X and G ) || X − G || 2 � � 1 / 2 � � W 2 ( X , G ) := inf E π π where π ranges over all possible couplings of X and G . 5

  11. Other metrics We consider stronger notions of distance Definition (Relative entropy between X and G ) Ent ( X || G ) := E [ln( f ( X ))] , where f is the density of X with respect to G . Definition (Wasserstein distance between X and G ) || X − G || 2 � � 1 / 2 � � W 2 ( X , G ) := inf E π π where π ranges over all possible couplings of X and G . 5

  12. Relative entropy For relative entropy, if A ⊂ R d is any measurable set, then by Pinsker’s inequality, � | P ( S n ∈ A ) − P ( G ∈ A ) | ≤ Ent ( S n || G ) . • In 84’ Barron showed that if Ent ( X || G ) < ∞ then n →∞ Ent ( S n || G ) = 0 . lim • In 2011, Bobkov, Chistyakov and G¨ otze showed that if, in addition, X has a finite fourth moment then Ent ( S n || G ) ≤ C n . • The above constant may depend on X as well as the dimension. 6

  13. Relative entropy For relative entropy, if A ⊂ R d is any measurable set, then by Pinsker’s inequality, � | P ( S n ∈ A ) − P ( G ∈ A ) | ≤ Ent ( S n || G ) . • In 84’ Barron showed that if Ent ( X || G ) < ∞ then n →∞ Ent ( S n || G ) = 0 . lim • In 2011, Bobkov, Chistyakov and G¨ otze showed that if, in addition, X has a finite fourth moment then Ent ( S n || G ) ≤ C n . • The above constant may depend on X as well as the dimension. 6

  14. Relative entropy For relative entropy, if A ⊂ R d is any measurable set, then by Pinsker’s inequality, � | P ( S n ∈ A ) − P ( G ∈ A ) | ≤ Ent ( S n || G ) . • In 84’ Barron showed that if Ent ( X || G ) < ∞ then n →∞ Ent ( S n || G ) = 0 . lim • In 2011, Bobkov, Chistyakov and G¨ otze showed that if, in addition, X has a finite fourth moment then Ent ( S n || G ) ≤ C n . • The above constant may depend on X as well as the dimension. 6

  15. Relative entropy For relative entropy, if A ⊂ R d is any measurable set, then by Pinsker’s inequality, � | P ( S n ∈ A ) − P ( G ∈ A ) | ≤ Ent ( S n || G ) . • In 84’ Barron showed that if Ent ( X || G ) < ∞ then n →∞ Ent ( S n || G ) = 0 . lim • In 2011, Bobkov, Chistyakov and G¨ otze showed that if, in addition, X has a finite fourth moment then Ent ( S n || G ) ≤ C n . • The above constant may depend on X as well as the dimension. 6

  16. Wasserstein distance The approximation error on a convex set K ⊂ R d , can be related to the Wasserstein distance using the following inequality by Zhai 1 2 6 W 2 ( S n , G ) 3 . | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d Proof. = W 2 ( S n , G ) 2 . || S n − G || 2 � � Take the optimal coupling, so E P ( S n ∈ K ) ≤ P ( || S n − G || ≤ ε, S n ∈ K ) + P ( || S n − G || > ε ) ≤ P ( G ∈ K ε ) + ε − 2 W 2 ( S n , G ) 2 1 4 + ε − 2 W 2 ( S n , G ) 2 . ≤ P ( G ∈ K ) + ε d Now, optimize over ε . 7

  17. Wasserstein distance The approximation error on a convex set K ⊂ R d , can be related to the Wasserstein distance using the following inequality by Zhai 1 2 6 W 2 ( S n , G ) 3 . | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d Proof. = W 2 ( S n , G ) 2 . || S n − G || 2 � � Take the optimal coupling, so E P ( S n ∈ K ) ≤ P ( || S n − G || ≤ ε, S n ∈ K ) + P ( || S n − G || > ε ) ≤ P ( G ∈ K ε ) + ε − 2 W 2 ( S n , G ) 2 1 4 + ε − 2 W 2 ( S n , G ) 2 . ≤ P ( G ∈ K ) + ε d Now, optimize over ε . 7

  18. Wasserstein distance Theorem (Zhai) If || X || ≤ β almost surely then √ d β log( n ) √ n W 2 ( S n , G ) ≤ . • Plugging this into the previous inequality shows 1 2 2 β | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 3 . 1 n 3 for β 3 in Bentkus’ bound gives || X || 3 � • Substituting E � 1 4 β 3 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d . 1 n 2 • the bounds are not comparable. 8

  19. Wasserstein distance Theorem (Zhai) If || X || ≤ β almost surely then √ d β log( n ) √ n W 2 ( S n , G ) ≤ . • Plugging this into the previous inequality shows 1 2 2 β | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 3 . 1 n 3 for β 3 in Bentkus’ bound gives || X || 3 � • Substituting E � 1 4 β 3 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d . 1 n 2 • the bounds are not comparable. 8

  20. Wasserstein distance Theorem (Zhai) If || X || ≤ β almost surely then √ d β log( n ) √ n W 2 ( S n , G ) ≤ . • Plugging this into the previous inequality shows 1 2 2 β | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 3 . 1 n 3 for β 3 in Bentkus’ bound gives || X || 3 � • Substituting E � 1 4 β 3 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d . 1 n 2 • the bounds are not comparable. 8

  21. Wasserstein distance Theorem (Zhai) If || X || ≤ β almost surely then √ d β log( n ) √ n W 2 ( S n , G ) ≤ . • Plugging this into the previous inequality shows 1 2 2 β | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 3 . 1 n 3 for β 3 in Bentkus’ bound gives || X || 3 � • Substituting E � 1 4 β 3 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d . 1 n 2 • the bounds are not comparable. 8

  22. Wasserstein distance √ √ Consider X , distributed uniformly on ± de i . In this case, β = d and Zhai’s bound gives 5 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 6 . 1 n 3 5 So, we can expect the CLT to hold whenever d 2 << n . On the other hand, Bentkus’ bound gives 7 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 4 . 1 n 2 7 2 << n for convergence. In this case, we would require d 9

  23. Wasserstein distance √ √ Consider X , distributed uniformly on ± de i . In this case, β = d and Zhai’s bound gives 5 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 6 . 1 n 3 5 So, we can expect the CLT to hold whenever d 2 << n . On the other hand, Bentkus’ bound gives 7 | P ( S n ∈ K ) − P ( G ∈ K ) | ≤ d 4 . 1 n 2 7 2 << n for convergence. In this case, we would require d 9

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend