smoothing sample extremes the mixed model approach
play

Smoothing sample extremes: the mixed model approach Francesco Pauli - PowerPoint PPT Presentation

Smoothing sample extremes: the mixed model approach Francesco Pauli Fabrizio Laurini Dept of Statistical Sciences, Dept of Economics, University of Padova University of Parma fpauli@stat.unipd.it flaurini@stat.unipd.it EVA 2005 F. Pauli


  1. Smoothing sample extremes: the mixed model approach Francesco Pauli Fabrizio Laurini Dept of Statistical Sciences, Dept of Economics, University of Padova University of Parma fpauli@stat.unipd.it flaurini@stat.unipd.it EVA 2005 F. Pauli & F. Laurini - p. 1/34

  2. Outline of this talk ■ Smoothing semipararametric tools & splines Basics Outline EVA 2005 F. Pauli & F. Laurini - p. 2/34

  3. Outline of this talk ■ Smoothing semipararametric tools & splines Basics Outline ■ Generalized linear mixed models (GLMM) for spline estimation EVA 2005 F. Pauli & F. Laurini - p. 2/34

  4. Outline of this talk ■ Smoothing semipararametric tools & splines Basics Outline ■ Generalized linear mixed models (GLMM) for spline estimation ■ GLMM for smoothing sample extremes EVA 2005 F. Pauli & F. Laurini - p. 2/34

  5. Outline of this talk ■ Smoothing semipararametric tools & splines Basics Outline ■ Generalized linear mixed models (GLMM) for spline estimation ■ GLMM for smoothing sample extremes ■ Some bits of Bayesian inference and MCMC EVA 2005 F. Pauli & F. Laurini - p. 2/34

  6. Outline of this talk ■ Smoothing semipararametric tools & splines Basics Outline ■ Generalized linear mixed models (GLMM) for spline estimation ■ GLMM for smoothing sample extremes ■ Some bits of Bayesian inference and MCMC ■ Simulations and applications to pollutants EVA 2005 F. Pauli & F. Laurini - p. 2/34

  7. Parametric modeling Popular models for response y i , i = 1 , . . . , n have form Smoothing modeling y i = f ( x i ) + error i Regression stick ■ f can be any parametric function spline knots placement ■ many way to characterize error; examples: penalized splines ◆ Gaussian iid (simple regression) optimization generalization ◆ Correlated errors LMM ◆ Exponential family (generalized linear models) spline & LMM extensions breath EVA 2005 F. Pauli & F. Laurini - p. 3/34

  8. Regression Take simple linear regression Smoothing modeling f ( x ) = β 0 + β 1 x Regression stick Smoothing features: spline ■ f is smooth; often unsuited to model real data knots placement penalized splines ■ f is a linear combination of basis functions optimization generalization � � LMM 1 x spline & LMM extensions breath EVA 2005 F. Pauli & F. Laurini - p. 4/34

  9. Regression Take simple linear regression Smoothing modeling f ( x ) = β 0 + β 1 x Regression stick Smoothing features: spline ■ f is smooth; often unsuited to model real data knots placement penalized splines ■ f is a linear combination of basis functions optimization generalization � � LMM 1 x spline & LMM extensions An extension is polynomial regression breath f ( x ) = β 0 + β 1 x + β 2 x 2 + . . . + β p x p Smoothing features: ■ f is smooth and suited to model non-linearity ■ f is a linear combination of basis functions � x p � x 2 · · · 1 x EVA 2005 F. Pauli & F. Laurini - p. 4/34

  10. Broken linear stick modeling Linear model for a “structural change” at time κ τ (broken Smoothing modeling stick) Regression stick f ( x ) = β 0 + β 1 x + b τ ( x − κ τ ) + () + the positive part of ( x − κ τ ) spline knots placement penalized splines Smoothing features: optimization ■ f is rough and suited to explain structural changes generalization LMM ■ f is a linear combination of basis functions spline & LMM extensions � � 1 x ( x − κ τ ) + breath ( x − κ τ ) + is a linear spline basis function EVA 2005 F. Pauli & F. Laurini - p. 5/34

  11. Linear spline modeling Extension to linear spline regression by adding knots Smoothing modeling κ 1 , . . . , κ K Regression K stick � b k ( x − κ k ) + spline f ( x ) = β 0 + β 1 x + knots placement k =1 penalized splines Smoothing features: optimization generalization ■ f is piecewise linear, and more flexible than broken stick LMM spline & LMM ■ f is a linear combination of basis functions extensions � � breath 1 x ( x − κ 1 ) + . . . ( x − κ K ) + EVA 2005 F. Pauli & F. Laurini - p. 6/34

  12. Linear spline modeling Extension to linear spline regression by adding knots Smoothing modeling κ 1 , . . . , κ K Regression K stick � b k ( x − κ k ) + spline f ( x ) = β 0 + β 1 x + knots placement k =1 penalized splines Smoothing features: optimization generalization ■ f is piecewise linear, and more flexible than broken stick LMM spline & LMM ■ f is a linear combination of basis functions extensions � � breath 1 x ( x − κ 1 ) + . . . ( x − κ K ) + “Definition” ■ The set of functions { ( x − κ j ) + } , j = 1 , . . . , K is a linear spline basis ■ A linear combination of such basis functions is a piecewise linear function ■ Commonly called spline with knots at κ 1 , . . . , κ K EVA 2005 F. Pauli & F. Laurini - p. 6/34

  13. How do we choose knots? Knots selection and their placement have drawbacks Smoothing modeling ■ Somehow ad hoc solution Regression ■ Overfitting stick spline ■ Might be time consuming knots placement penalized splines optimization generalization LMM spline & LMM extensions breath EVA 2005 F. Pauli & F. Laurini - p. 7/34

  14. Penalized spline regression Let β = ( β 0 , β 1 , b 1 , . . . , b K ). Smoothing modeling Wiggly fit is avoided by constraints on b k such as Regression K stick � b 2 spline k < C finite C knots placement k =1 penalized splines Minimization is written as optimization generalization LMM β T Dβ ≤ C minimize � y − Xβ � 2 subject to spline & LMM extensions where D is a squared positive matrix with K + 2 rows breath � � 0 2 × 2 0 D = 0 I K × K Smoothing features: ■ f is smooth and pleasing ■ The amount of smoothness is controlled by C , and does not depend on number/placement of knots EVA 2005 F. Pauli & F. Laurini - p. 8/34

  15. Solution of constrained optimization With Lagrange multiplier argument, for some λ ≥ 0 , choose β Smoothing modeling to minimize Regression stick � y − Xβ � 2 + λ 2 β T Dβ β λ = ( X T X + λ 2 D ) − 1 X T y. with solution ˆ spline knots placement ■ λ 2 β T Dβ is the roughness penalty term penalized splines optimization ■ λ is the smoothing parameter generalization LMM ■ Connections with ridge regression spline & LMM extensions breath λ small λ big EVA 2005 F. Pauli & F. Laurini - p. 9/34

  16. Generalization of spline models Generalization may involve Smoothing modeling Use of different basis Regression stick ■ Use of B -spline (numerical stability) spline ■ Use of natural cubic splines (arise as solution of knots placement penalized splines optimization problem) optimization generalization Use of different penalties LMM ■ Penalize “some difference” in spline coefficients spline & LMM extensions ■ Penalizing degree of spline function. Example: q -th breath derivative of f EVA 2005 F. Pauli & F. Laurini - p. 10/34

  17. Penalized splines as Linear Mixed Model Take the Linear Mixed Model (LMM) Smoothing modeling y = Xβ + Zu + ε Regression stick and assume spline knots placement penalized splines � � � � � � � � G = σ 2 u u 0 G 0 u I optimization E = , COV = and R = σ 2 generalization ε ε 0 0 R ε I LMM spline & LMM ■ Xβ is the fixed component extensions breath ■ u is the random component or random effect EVA 2005 F. Pauli & F. Laurini - p. 11/34

  18. Spline embedded in LMM For LMM Smoothing modeling Regression y = Xβ + Zu + ε stick spline K knots placement � u k ( x i − κ k ) + + ε i y i = β 0 + β 1 x i + penalized splines k =1 optimization generalization splines are the Best Linear Unbiased Predictor where u is a LMM spline & LMM vector of random coefficients. extensions Details: breath     ( x 1 − κ 1 ) + · · · ( x 1 − κ K ) + 1 x 1 . . . . ...  . .   . .  X = and Z = . . . .         ( x n − κ 1 ) + · · · ( x n − κ K ) + 1 x n ■ In general u k i.i.d. N (0 , σ 2 u ) ■ σ u = ∞ leads to wiggly (over)fit ■ Finite σ u shrinks u k (smooth fit) EVA 2005 F. Pauli & F. Laurini - p. 12/34

  19. Sketch of extensions to LMM (increasing complexity) ■ Semiparametric models: one covariate is nonparametric Smoothing modeling i ) + X [ − 1] y i = f ( X 1 β + ε i , f smooth Regression i stick spline ■ Semiparamtric mixed models: add random effects u knots placement ■ Additive models: covariates as penalized linear splines, e.g. penalized splines optimization generalization y i = c + f 1 ( x i ) + f 2 ( t i ) + ε i ; f 1 and f 2 are smooth LMM spline & LMM extensions ■ Generalized parametric regression with random effects breath (GLMM), e.g. exp { ( Xβ + Zu ) i } � � y i | u ∼ Ber u ∼ MV N ( 0 , G θ ) 1 + exp { ( Xβ + Zu ) i } ■ Combine all above ingredients, eventually with a Bayesian approach. EVA 2005 F. Pauli & F. Laurini - p. 13/34

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend