a consistent approach to inconsistencies
play

A consistent approach to inconsistencies Fabian Khlinger (Kavli - PowerPoint PPT Presentation

A consistent approach to inconsistencies Fabian Khlinger (Kavli IPMU) in collaboration with Benjamin Joachimi (UCL) SCLSS workshop Oxford, 19 th April 2018 I. Motivation Typical questions arising in a (LSS) data analysis: 1. Is model 0


  1. A consistent approach to inconsistencies Fabian Köhlinger (Kavli IPMU) in collaboration with Benjamin Joachimi (UCL) SCLSS workshop Oxford, 19 th April 2018

  2. I. Motivation

  3. Typical questions arising in a (LSS) data analysis: 1. Is model 0 (e.g. wCDM) more likely than my fiducial model 1 (e.g. 𝚳 CDM)? 2. Is data set 1 (e.g. Planck) consistent with data set 0 (e.g. cosmic shear)? 3. Is split 1 of my data set (e.g. z-bin X) consistent with another split 0 of the same data set (e.g. all other z-bins)?

  4. Typical questions arising in a (LSS) data analysis: 1. Is model 0 (e.g. wCDM) more likely than my fiducial model 1 (e.g. 𝚳 CDM)? 2. Is data set 1 (e.g. Planck) consistent with data set 0 (e.g. cosmic shear)? 3. Is split 1 of my data set (e.g. z-bin X) consistent with another split 0 of the same data set (e.g. all other z-bins)?

  5. II. Bayesian approach to (in)consistency

  6. 1. Bayesian evidence: data hypothesis, model parameters evidence likelihood prior The evidence is the average of the likelihood over the prior, so it automatically implements Occam’s razor. 6

  7. 2. Bayes factor: Calculate the ratio of probabilities that each model is correct (given the data): typically set to 1 Bayes’ theorem Bayes factor a priori H 0 : ‘hypothesis for model 1’ H 1 : ‘hypothesis for model 0’ d: data 7

  8. 2. Bayes factor: H 0 is more likely to be true than H 1 > 1 Data set comparison: (Nested) model comparison: H 0 : ‘there is one common set of parameters describing e.g. H 0 : ‘wCDM' Planck and cosmic shear’ vs. vs. H 1 : ‘ 𝚳 CDM' H 1 : ‘each data set requires its own set of parameters’ e.g. Marshall, Rajguru & Slosar (2006) 8

  9. 2. Bayes factor: H 0 is more likely to be true than H 1 > 1 Data set comparison: H 0 : ‘one common set of parameters is sufficient for This does NOT hold for describing the fiducial (= split 1 + … + split N) data set’ correlated data sets! vs. H 1 : ‘each split i of the data set requires its own set of parameters’ 9

  10. 2. Bayes factor: H 0 is more likely to be true than H 1 > 1 Data set comparison: H 0 : ‘one common set of parameters is sufficient for This does NOT hold for describing the fiducial (= split 1 + … + split N) data set’ correlated data sets! vs. H 1 : ‘each split i of the data set requires its own set of parameters’ 10

  11. 3. Posterior predictive distribution (PPD): PPD likelihood of new data posterior sample The PPD is the average of the likelihood of the new data over the posterior of the parameters of a given model. : original data d Can the model(s) describe the data? ˆ : PPD split samples 
 d s Are ‘split’ models consistent? ˆ : PPD joint sample d j quantify this by: - comparing the difference between joint and split PPDs to zero - comparing the (Gaussian) data distribution to the corresponding PPDs 11

  12. 3. Posterior predictive distribution (PPD): Quantify tension between Gaussian data distribution and PPDs by calculating overlap with m 𝜏 -region. FK+ in prep. 12

  13. III. Test case: cosmic shear correlation functions from KiDS-450

  14. a) Systematics in z-bin 3?

  15. 1. Data and PPDs: + z-bin 3 (incl. cross-correlations) vs. all other correlations black: data from KiDS-450 (Hildebrandt+ 2017) red: mode of joint PPD blue: modes of split PPDs FK+ in prep. 15

  16. 2. Comparison of key parameters: z-bin 3 (incl. cross-correlations) vs. all other correlations amplitude of intrinsic alignment model FK+ in prep. 16

  17. 3. Comparison in data space: + z-bin 3 (incl. cross-correlations) vs. all other correlations red: mode of joint PPD blue: modes of split PPDs FK+ in prep. 17

  18. 4. Comparing difference of PPDs: + z-bin 3 (incl. cross-correlations) vs. all other correlations FK+ in prep. red: mode of joint PPD blue: modes of split PPDs FK+ in prep. 18

  19. b) Scale-dependent systematics?

  20. 1. Data and PPDs: + Large scales vs. small scales black: data from KiDS-450 (Hildebrandt+ 2017) red: mode of joint PPD blue: modes of split PPDs FK+ in prep. 20

  21. 2. Comparison of key parameters: Large scales vs. small scales amplitude of intrinsic alignment model FK+ in prep. 21

  22. 3. Comparison in data space: + Large scales vs. small scales red: mode of joint PPD blue: modes of split PPDs FK+ in prep. 22

  23. 4. Comparing difference of PPDs: + Large scales vs. small scales FK+ in prep. red: mode of joint PPD blue: modes of split PPDs FK+ in prep. 23

  24. IV. Summary

  25. Summary: 1. Bayesian evidence and the Bayes factor are powerful concepts for model comparison • can be expanded to consistency checks of (correlated) datasets 2. Quantification of consistency with Bayes factor is not optimal: • all information compressed into one number • no hints to from where systematics arise • mind the priors… 3. Complementary tool: PPDs • systematics apparent in data space • can be compressed into various numbers ( 𝛕 -levels) 25

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend