estimating distributional
play

Estimating Distributional Parameters in Hierarchical Models - PowerPoint PPT Presentation

Estimating Distributional Parameters in Hierarchical Models Introduction: Variability in Hierarchical Models Linear Models = 0 + 1 + (0, 2 ) Modelling central


  1. Estimating Distributional Parameters in Hierarchical Models

  2. Introduction: Variability in Hierarchical Models

  3. Linear Models 𝑧 π‘—π‘˜ = 𝛾 0 + 𝛾 1 π‘Œ π‘—π‘˜ + 𝑓 π‘—π‘˜ 𝑓 π‘—π‘˜ ∼ 𝑂(0, Οƒ 2 ) β€’ Modelling central tendency β€’ Response ( 𝑧 π‘—π‘˜ ) is a sum of intercept ( 𝛾 0 ), slopes ( 𝛾 1 , 𝛾 2 , … ), and error ( 𝑓 π‘—π‘˜ ) β€’ Error is assumed to be normally distributed around zero

  4. Linear Models lm(y ~ pred) β€’ Modelling central tendency β€’ Response (y) is a sum of intercept (implicit), slopes (pred), and error (implicit) β€’ Error is assumed to be normally distributed around zero

  5. Linear Mixed Effects Models 𝑧 π‘—π‘˜ = 𝛾 0 + ΞΌ 0𝑗 + (𝛾 1 + ΞΌ 1𝑗 )π‘Œ π‘—π‘˜ + 𝑓 π‘—π‘˜ ΞΌ 0𝑗 ∼ 𝑂(0, Οƒ 2 ) ΞΌ 1𝑗 ∼ 𝑂(0, Οƒ 2 ) 𝑓 π‘—π‘˜ ∼ 𝑂(0, Οƒ 2 ) β€’ Modelling central tendency β€’ Response ( 𝑧 π‘—π‘˜ ) is a sum of intercept ( 𝛾 0 ), slopes ( 𝛾 1 , 𝛾 2 , … ), random unit intercepts ( ΞΌ 0𝑗 ), random unit slopes ( ΞΌ 1𝑗 ), and error ( 𝑓 π‘—π‘˜ ) β€’ Error, random intercepts, and random slopes are assumed to be normally distributed around zero

  6. Linear Mixed Effects Models lmer(y ~ pred + (pred | rand_unit)) β€’ Modelling central tendency β€’ Response (y) is a sum of intercept (implicit), slopes (pred), random unit intercepts (pred || rand_unit), random unit slopes (pred | rand_unit), and error (implicit) β€’ Error, random intercepts, and random slopes are assumed to be normally distributed around zero

  7. Example Non-Gaussian Data: RT β€’ 2AFC: does the word match the picture? β€’ Congruency (2) x Predictability (12% – 100%) β€’ 35 Subjects, 200 trials bandage + sardine

  8. Gamma Family GLMM m_glmer <- glmer( rt ~ cong * pred + (cong * pred | subj) + (cong | image) + (1 | word), family = Gamma(identity), control = glmerControl( optimizer = β€œ bobyqa ”, optCtrl = list(maxfun = 2e5) ) )

  9. GLMM Results

  10. GLMM Results – Random Effects summary(m_glmer)

  11. GLMM Results – Random Effects ranef(m_glmer)

  12. GLMM Results – Random Effects m_glmer %>% ranef() %>% as.data.frame()

  13. GLMM Results – Random Effects ranef(m_glmer) %>% as_tibble() %>% filter(grpvar == β€œsubj") %>% mutate(grp = fct_reorder2(grp, term, condval)) %>% ggplot(aes( x = grp, y = condval, ymin = condval - condsd, ymax = condval + condsd )) + geom_pointrange(size=0.25) + facet_wrap(vars(term), scales="free", nrow=2)

  14. GLMM Results – Random Effects – Subject

  15. GLMM Results – Random Effects – Image

  16. GLMM Results – Random Effects – Word

  17. Estimating Distributional Parameters in Hierarchical Models

  18. What if Meaningful Effects on Variance? β€’ All glm variants model single parameters (i.e. central tendency) β€’ What if your effect looks like this?

  19. What if Meaningful Effects on Variance? β€’ Mu is higher F(1, 1998) = 3237, p <.001 β€’ Sigma is higher Levene’s F(1, 1998) = 550, p <.001

  20. Assumption-free Distribution Comparison β€’ Within a single model? β€’ Assumption free distribution comparison (e.g. Kolmogorov – Smirnov) could be one approach! β€’ Overlapping index (Pastore & Calcagni, 2019) from 0 (no overlap) to 1 (identical distribution)

  21. Assumption-free Distribution Comparison x <- rnorm(1000, 10, 1), y <- rnorm(1000, 10.5, 1.5)

  22. Assumption-free Distribution Comparison

  23. Overlap Index Mu * Sigma Parameter Space

  24. Overlap Index Mu * Sigma Parameter Space

  25. Overlap Index Mu * Sigma Parameter Space

  26. Weirder Distribution Example

  27. Weirder Distribution Example

  28. Summary so far β€’ Assumption- free approaches are flexible but don’t allow us to test/make any specific predictions β€’ Equivalent of shrugging and saying β€œyeah idk probs something going on there” (though useful for very weird distributions) β€’ Explicitly modelling multiple parameters of an assumed distribution can give us more meaningful info

  29. Distributional Parameters in brms brm( bf( dv ~ Intercept + iv + (iv | rand_unit), sigma ~ Intercept + iv + (iv | rand_unit) ), control = list( adapt_delta = 0.999, max_treedepth = 12 ), sample_all_pars = TRUE )

  30. Shifted Log-Normal Distribution

  31. Shifted Log-Normal Distribution

  32. Bayesian Shifted Log-Normal Mixed Effects Model with Distributional Parameters brms::bf( rt ~ Intercept + cong * pred + (cong * pred | subj) + (cong | image) + (1 | word), sigma ~ rt ~ Intercept + cong * pred + (cong * pred | subj) + (cong | image) + (1 | word), ndt ~ rt ~ Intercept + cong * pred + (cong * pred | subj) + (cong | image) + (1 | word) )

  33. Bayesian Results – Random Effects ranef(m_bme) ID (e.g. subj_01, subj_02…) * value ( est, err, Q2.5, Q97.5) * fixed parameter

  34. Caveats β€’ Computationally intensive if using non- informative priors for complex hierarchical formulae β€’ Have to avoid temptation to try over-infer about mechanisms unless using more cognitively informed models (e.g. drift diffusion)

  35. Summary Hierarchical models with maximal structures for distributional parameters are a robust and appropriate way of looking at or accounting for subject/item/etc variability in fixed effects when you’re interested in more than central tendency. But , if you can assume no systematic differences in distributional parameters, GLMMs will suffice (and save you a lot of time and effort)!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend