Efficiency of Bayesian procedures in some high dimensional problems - - PowerPoint PPT Presentation

efficiency of bayesian procedures in some high
SMART_READER_LITE
LIVE PREVIEW

Efficiency of Bayesian procedures in some high dimensional problems - - PowerPoint PPT Presentation

Outline Example 1 key issues Dirichlet-Laplace prior Example 2 Efficiency of Bayesian procedures in some high dimensional problems Natesh S. Pillai Dept. of Statistics, Harvard University pillai@fas.harvard.edu May 16, 2013 DIMACS


slide-1
SLIDE 1

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Efficiency of Bayesian procedures in some high dimensional problems

Natesh S. Pillai

  • Dept. of Statistics, Harvard University

pillai@fas.harvard.edu May 16, 2013 DIMACS Workshop

slide-2
SLIDE 2

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Joint Work: Collaborators

Anirban Bhattacharya, Debdeep Pati and David Dunson (Duke University and Florida State) Christian Robert, Jean-Michel Marin, Judith Rousseau (Paris 9) Jun Yin (University of Wisconsin)

slide-3
SLIDE 3

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Outline

Goal: Understand Bayesian methods in high dimensions. Example 1: Covariance matrix estimation Example 2: Bayesian model choice via ABC Implications, Frequentist-Bayes connection in high dimensions.

slide-4
SLIDE 4

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Conversation with Peter E. Huybers

Motivation: Time variability in covariance patterns: stationarity? Instrumental measurements, only for the past n = 150 years. Measurements on p = 2000 latitude-longitude points. Estimate O(p2) parameters. Need judicious modeling.

slide-5
SLIDE 5

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Covariance Matrix Estimation: Why Shrinkage?

We observe y1, . . . yn

i.i.d

∼ Npn(0, Σ0n) and set y(n) = (y1, . . . , yn) For pn = p, fixed, the sample covariance estimator Σsample = 1 n

n

  • i=1

yiyT

i

is consistent for population eigenvalues. ˆ λi are consistent for population eigenvalues: √ n(ˆ λi − λi) ⇒ N(0, V(λi))

slide-6
SLIDE 6

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Covariance Matrix in high dimensions

Simplest Case: Σ0n = I Take p = pn = c n, c ∈ (0, 1).

  • λ1,

λpn largest and smallest (non-zero) eigenvalues of Σsample = 1 n

n

  • i=1

yiyT

i

Then as n → ∞ (and thus pn also grows), (Marcenko-Pastur, 1967) almost surely! lim

n→∞

  • λ1 = (1 +

√ c)2 lim

n→∞

  • λpn = (1 −

√ c)2 MLE is not consistent!

slide-7
SLIDE 7

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Covariance Matrix in high dimensions

limn→∞ λ1 = (1 + √c)2 = λ+. Confidence Interval: n 2/3( λ1 − λ+) ⇒ TW1 where TW1 is the Tracy-Widom law (Johnstone 2000). Universality phenomenon: Results go beyond the case of Gaussian (Tao and Vu, 2009; P . and Yin, 2011)

slide-8
SLIDE 8

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Correlation Matrix

Johnstone (2001): Correlation Matrices for PCA. Theorem (P . and Yin, 2012, AoS) Largest eigenvalue of sample correlation matrices still

  • inconsistent. All of the problems from covariance matrices

persist.

slide-9
SLIDE 9

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Understanding Asymptotics

20 century n → ∞. Now: both p, n → ∞. Why should we bother? Because the above asymptotics is remarkably accurate for ‘small’ n, ‘small’ p!

slide-10
SLIDE 10

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Sample covariance matrix plot, n = 100, p = 25

n=100, p= 25

Max Eigenvalue of Sample Covariance Matrix Frequency 1.8 2.0 2.2 2.4 2.6 50 100 150 200 250 300

slide-11
SLIDE 11

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Sample covariance matrix plot, n = 500, p = 125

n=500, p= 125

Max Eigenvalue of Sample Covariance Matrix Frequency 2.05 2.10 2.15 2.20 2.25 2.30 2.35 2.40 100 200 300 400

slide-12
SLIDE 12

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Factor Models: Motivation

Interest in estimating dependence in high-dim obs. + prediction and classification from high-dim correlated markers such as gene expression, SNPs. Center prior on a “sparse” structure, while allowing uncertainty and flexibility. Latent factor methods (West, 2003; Lucas et al., 2006; Carvalho et al., 2008). Huge applications (economics, finance, signal processing..)

slide-13
SLIDE 13

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Gaussian factor models

Explain dependence through shared dependence on fewer latent factors yi ∼ N(0, Σp×p) , 1 ≤ i ≤ n . Focus on the case p = pn ≫ n. Factor models assume the “decomposition" Σ = ΛΛT + σ2Ip Λ is a p × k matrix, k ≪ n.

slide-14
SLIDE 14

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Gaussian factor models

Explain dependence through shared dependence on fewer latent factors yi = µ + Ληi + ǫi, ǫi ∼ Np(0, Σ), i = 1, . . . , n µ ∈ Rp, a vector of means, with µ = 0. ηi ∈ Rk, latent factors, Λ a p × k matrix of factor loadings with k ≪ p. ǫi are i.i.d with N(0, σ2).

slide-15
SLIDE 15

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Factor models for covariance estimation

Unstructured Σ has O(p2) free elements Factor models Σ = ΛΛT + σ2Ip . Still O(p) elements to estimate!

slide-16
SLIDE 16

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

High-dimensional covariance estimation

‘Frequentist’ solution– MLE doesn’t work. Start with sample covariance matrix: Σsample = 1 n

n

  • i=1

yiyT

i

. Great interest in regularized estimation (Bickel & Levina, 2008a, b; Wu and Pourahmadi, 2010, Cai and Liu, 2011 ...) Estimator which achieves the ‘minimax’ rate: ˆ Σij = Σsample

ij

1|Σsample

ij

|>tn .

Unstable; Confidence intervals..

slide-17
SLIDE 17

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Sparse factor modeling

A natural bayesian alternative: sparse factor modeling (West, 2003); also (Lucas et al., 2006; Carvalho et al., 2008) and many others Allow zeros in loadings through point mass mixture priors: Λij given point mass priors or shrinkage priors. Prior assigns Λij = 0 with non-zero probability. Why care about this prior? Bayesian analogue of thresholding. Assume k to be known (but easy to relax this).

slide-18
SLIDE 18

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Important questions

Can Bayes methods produce estimators which are comparable to frequentist estimators? Can one do computation in reasonable time? How to address Statistical efficiency-Computational efficiency trade off?

slide-19
SLIDE 19

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Our objective

Bayesian counterpart lacks a theoretical framework in terms of posterior convergence rates. A prior Π(Λ ⊗ σ2) induces a prior distribution Π(Ω) How does the posterior behave assuming data sampled from fixed truth? Huge literature on frequentist properties of the posterior distribution

slide-20
SLIDE 20

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Questions need to be addressed

Does the posterior measure concentrate around the truth increasingly with sample size? What role does the prior play? How does the dimensionality affect the rate of contraction?

slide-21
SLIDE 21

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Preliminaries

We consider the operator norm ( · 2) A2 = sup

x∈Sr−1 Ax2 = s(1)

Largest Eigenvalue of A, for symmetric A.

slide-22
SLIDE 22

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Setup

We observe y1, . . . yn

i.i.d

∼ Npn(0, Σ0n) and set y(n) = (y1, . . . , yn), Σ0n = Λ0Λt

0 + σ2Ipn×pn

Want to find a minimum sequence ǫn → 0 such that lim

n→∞ P

  • Σ − Σ0n2 > ǫn | y(n)

= 0 Can we find such ǫn even if pn ≫ n? What is the role of the prior?

slide-23
SLIDE 23

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Assumptions on truth

“Realistic Assumption:" (A1) Sparsity: Each column of Λ0n has at most sn non-zero entries, with sn = O(log pn).

slide-24
SLIDE 24

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Prior choice & a key result

Prior (PL) Let Λij ∼ (1 − π)δ0 + πg(·), π ∼ Beta(1, pn + 1). g(·) has Laplace like or heavier tails Theorem (Pati, Bhattacharya, P . and Dunson, 2012) For the high-dimensional factor model rn =

  • log7(pn)/n,

lim

n→∞ P(Σ − Σ02 > rn | y(n)) = 0 .

slide-25
SLIDE 25

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Prior choice & a key result

Prior (PL) Let Λij ∼ (1 − π)δ0 + πg(·), π ∼ Beta(1, pn + 1). g(·) has Laplace like or heavier tails Theorem (Pati, Bhattacharya, P . and Dunson, 2012) For the high-dimensional factor model rn =

  • log7(pn)/n,

lim

n→∞ P(Σ − Σ02 > rn | y(n)) = 0 .

slide-26
SLIDE 26

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Implication of the result

Rate ǫn =

  • log2(pn)/n.

We will get consistency if lim

n→∞

log7 pn n = 0 . Ultra-High dimensions, pn = en1/7.

slide-27
SLIDE 27

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Important Implication for Asymptotics

This rate we get is similar to the minimax rate for similar problems Cai and Zhou (2011), but not the same! rn = minimax rate ×

  • log pn

The above phenomenon is similar to what happens in mixture modeling! Ghosal (2001): Bayesian nonparametric modeling doesn’t match frequentist rates. If true: Serious implications.

slide-28
SLIDE 28

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

A couple of Implications

Minimax theory will tell only half the story. Heuristics based on bayes. BIC? Frequentist-Bayes agreement/disagreement?

slide-29
SLIDE 29

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Interesting Challenges in Mathematical Statistics

Need to have 2 things to show Bayesian methods work well. Show prior is not too “dogmatic”. Likelihood is able to “separate points". Neymann-Pearson Lemma Separation of points: Traditional Likelihood Ratio doesn’t work!

slide-30
SLIDE 30

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Example : Intuition and Tools from Random Matrix Theory

Intuition from random matrix theory (RMT) - “tall” matrices properly normalized look like identity matrices. If entries of Λ0 were drawn i.i.d. N(0, 1), Vershynin (2011) tells us 1 pΛT

0Λ0 − Ik2 ≤ C

√ k √p with high probability.

slide-31
SLIDE 31

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Computationally easier priors

We need to construct prior distribution for a pn × 1 vector Λ. Conjugate priors – easier to update Many popular ones. Many ‘loss functions’ are prior distributions; thus point estimates are posterior modes.

slide-32
SLIDE 32

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Regularization: Statistical flavor of the decade

Estimates of the form ˆ Λ = arg min

Λ n

  • i=1

(Yi − Λi)2 + θ

n

  • i=1

|Λi|k . Gazillion papers; not a SINGLE one constructs confidence intervals or uncertainty estimation. Two special cases: k = 2: (Ridge regression, James-Stein type) ˆ Λ = arg min

Λ n

  • i=1

(Yi − Λi)2 + θ

n

  • i=1

|Λi|2 . k = 1: (LASSO) ˆ Λ = arg min

Λ n

  • i=1

(Yi − Λi)2 + θ

n

  • i=1

|Λi| .

slide-33
SLIDE 33

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Prior choice & another key result

Prior Let the columns Λi = LASSO or RIDGE prior. Theorem (Pati, Bhattacharya, P . and Dunson, 2012) For a large class of models, the above, the convergence rate is strictly slower than the point mass priors.

slide-34
SLIDE 34

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Prior choice & another key result

Prior Let the columns Λi = LASSO or RIDGE prior. Theorem (Pati, Bhattacharya, P . and Dunson, 2012) For a large class of models, the above, the convergence rate is strictly slower than the point mass priors.

slide-35
SLIDE 35

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Intuition?

Independence! Stein phenomenon.

slide-36
SLIDE 36

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Dirichlet Laplace prior & properties

We propose a simple dependent modification leading to

  • ptimal concentration & efficient computation

Λj ∼ DE(φjτ), φ = (φ1, . . . , φp)T ∈ Sp−1, τ > 0 DE = Double exponential Constraining φ to the simplex crucial - allows for dependence We let φ ∼ Diri(α, . . . , α) - α < 1 favors small # dominant values with remaining ≈ 0 Computation easy! Take advantage of Conjugacy

slide-37
SLIDE 37

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Dirichlet-Laplace prior - motivation

Theorem (Pati, Bhattacharya, P . and Dunson, 2013) The Dirichlet-Laplace priors produce convergence rates identical to that of the point mass priors.

slide-38
SLIDE 38

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm

ABC: Approximate Bayes Computation. Rubin(1984) Generate θ∗ ∼ π Generate pseudo-data Ypseudo from fθ∗. Accept θ∗ as posterior, if Ypseudo = Yobs . Repeat.

slide-39
SLIDE 39

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm

Exactly matching the observed data - Impossible, even in 1 dimension! Key Idea: Approximately match. Choose a distance d, and tolerance ǫ. Accept θ∗ if d(Ypseudo, Yobs) < ǫ . For a given d, accuracy of the procedure can be improved by choosing ǫ smaller and smaller and smaller...

slide-40
SLIDE 40

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm: Twist

In real examples, it is still expensive/impossible to compute d(Ypseudo, Yobs). Twist: Use some function η of the data: called the “summary statistic" and accept if d

  • η(Ypseudo), η(Yobs)
  • < ǫ .

Why no sufficient statistics? Recall the Pitman-Koopman-Darmois theorem, for exponential families. Dimension of the sufficient statistic necessarily increases with the sample size!

slide-41
SLIDE 41

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm

The above version, re-discovered in population genetics (Tavare et.al, 1997). Literally 100’s of papers! How to choose d and ǫ? Fearnhead and Prangle, 2012, JRSS-B discussion.

slide-42
SLIDE 42

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm for Model Selection

Compare 2 models: compute the Bayes factors. Bayes Factor ∝ Ratio of Marginal Likelihoods. Jeffreys’ interpretation, as strength of evidence. Easy to perform, using the ABC algorithm!

slide-43
SLIDE 43

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm for Model Selection

Choose Model 1 or 2 according to the prior. Given the model, generate (θ∗, Ypseudo) from the prior distribution of the corresponding model. Accept θ∗, and the Model, if d(Ypseudo, Yobs) < ǫ . Estimate for Bayes Factor = # of timesModel 1 is accepted

# of timesModel 2 is accepted

slide-44
SLIDE 44

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm for Model Selection using η

The above algorithm = Recipe for Disaster! High Profile papers! Miller, N. et al, (2005) Science. Multiple transatlantic introductions of the Western corn rootworm.

slide-45
SLIDE 45

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Lots of popular software

Donoho (2002). DIY-ABC ABCToolbox PopABC ABC-SysBio

slide-46
SLIDE 46

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Result

Theorem (Robert, Jean-Marie, Jean-Michel, P ., 2011, PNAS) Bayes Model selection based on a summary statistic η can be INCONSISTENT.

slide-47
SLIDE 47

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm for Model Selection using η

“Popular beliefs" in the field. Accuracy can be increased with choosing ǫ very small: thus increase in computing power leads to more accurate results. If gives reasonable answers for parameter estimation, no reason why it should go wrong for model selection!

slide-48
SLIDE 48

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

ABC algorithm for Model Selection

What goes wrong for model selection? Marginal likelihood based on η(Y) :=

  • Θ f(η(Y)|θ)π(θ)dθ.

BF(η(Y)) := Bayes Factor based on the single observation η(Y). Sufficiency vs. Ancilliarity!

slide-49
SLIDE 49

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Example

A statistic can be sufficient for two models, but cannot be “sufficient" across the models. Ancilliarity......? Suppose, we observe Y = (y1, y2, · · · , yn) integer valued data. Two competing models: Poisson(λ) vs. Geometric(p). Statistic η(Y) = n

i=1 yi.

slide-50
SLIDE 50

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Example

Almost surely, as the sample size goes to infinity, the Bayes Factor based on η converges to θ−1

0 (θ0 + 1)2e−θ0 ,

where θ0 = E(yi) > 0.

slide-51
SLIDE 51

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Ilustration

yi vs. BF plot.

slide-52
SLIDE 52

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Another Example

Consider two models: Model 1: N(θ1, 1), Model 2: Laplace(θ2,

1 √ 2)

¯ Y Median(Y) Sample variance mad(Y) = Median(|Y - Median(Y)|)

slide-53
SLIDE 53

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Conclusions

Shrinkage priors = serious business in high dimensions. Innocent looking priors may look “dogmatic". Frequentist-Bayes agreement may not hold, implications? Ad-hoc methods often don’t work, but opportunity for statistical theory. Lots of open problems, virtually nothing is known!

slide-54
SLIDE 54

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

References

Universality of Correlation matrices ( P ., Yin, J., 2012), Annals of Statistics. Lack for confidence in ABC model selection, (Robert, Jean-Marie, Jean-Michel, P ., 2011),PNAS. Bayesian Shrinkage, (Pati, Bhattacharya, P ., Dunson, 2012) (2012) Bayesian high dimensional covariance estimation using factor models (Pati, Bhattacharya, P ., Dunson, 2012) Universality of Covariance matrices (P ., Yin, J., 2013), Annals of Applied Probability

slide-55
SLIDE 55

Outline Example 1 key issues Dirichlet-Laplace prior Example 2

Remarks

Thank you!