inference
play

Inference Connecting models to data The problem with infection data - PowerPoint PPT Presentation

Inference Connecting models to data The problem with infection data Often only observe a proportion of reality Hospitalised case data gives you those who had severe infection Symptom onsets are observed but infection times are not Or


  1. Inference Connecting models to data

  2. The problem with infection data Often only observe a proportion of reality • Hospitalised case data gives you those who had severe infection • Symptom onsets are observed but infection times are not Or only observe a measure of infection • antibody response at one time point • result of imperfect diagnostic test We use this data to infer the ‘truth’.

  3. 𝜄 Model In a perfect world, we would directly observe the Truth ‘truth’. Observation process Data square=observed, circle=unobserved

  4. Diagnostic testing results 𝜄 • Susceptible-Infected model Model • Predicted number of susceptible Truth ( 𝑇 ) and infected ( 𝐽 ) animals • Binomial( 𝐽 , sensitivity).Binomial( 𝑇 , specificity) Observation process • Diagnostic test results : test Data positives and test negatives square=observed, circle=unobserved

  5. Serological data 𝜄 • Antibody process model Model • Predicted log antibody titre Truth • Normally distributed error around Observation predicted log antibody titre process • Laboratory based assay (measure Data of log antibody titre) square=observed, circle=unobserved

  6. Imperfect reporting of incidence data 𝜄 • 𝜄 = 𝑆 * , 𝐸 -./ , 𝐸 012 , 𝐸 033 , 𝛽, 𝜍 • Deterministic/Stochastic model SEITL model Model • Predicted incidence 𝐽𝑜𝑑 Truth • We assumed data were recorded according to a Poisson process : Poisson( 𝜍𝐽𝑜𝑑) with reporting Observation process rate 𝜍 and predicted incidence 𝐽𝑜𝑑 Data • Reported incidence over time square=observed, circle=unobserved

  7. 𝜄 Model Connecting your models to data relies on distinguishing how you predict the ‘ truth ’ ( model ) and how you connect this ‘truth’ to your data ( observation Truth process ). Observation process Data

  8. Examples • Kucharski AJ, Lessler J, Cummings DAT, Riley S (2018) Timescales of influenza A/H3N2 antibody dynamics. PLOS Biology 16(8): e2004974.https://doi.org/10.1371/journal.pbio.2004974 • Brooks-Pollock E, Roberts G.O, Keeling, M.J (2014) A dynamic model of bovine tuberculosis spread and control in Great Britain. Nature, 511, pp. 228-231

  9. Approximate Bayesian Computation

  10. Outline 1. What is Approximate Bayesian Computation? 2. When do we use ABC instead of other methods would we use it? 3. How do we use it? a) Choices in the ABC-rejection algorithm b) Short introduction to more advanced ABC

  11. 1. What is Approximate Bayesian Computation?

  12. Bayesian inference is based on the idea of updating belief with new evidence • Belief : Prior distribution. Parameters are random variables instead of fixed quantities (they have there own distribution) • Evidence : Likelihood function tells you the probability of the data given the parameters

  13. Bayesian inference 𝜄 : Mathematical model parameter, 𝐸 : Data 𝑄 𝜄 𝐸 = 𝑄 𝐸 𝜄 𝑄 𝜄 𝑄 𝐸

  14. Bayesian inference 𝜄 : Mathematical model parameter, 𝐸 : Data 𝑄 𝜄 𝐸 ∝𝑄 𝐸 𝜄 𝑄 𝜄

  15. Bayesian inference 𝜄 : Mathematical model parameter, 𝐸 : Data 𝑄 𝜄 𝐸 ∝𝑄 𝐸 𝜄 𝑄 𝜄 Probability of data given 𝜄 (likelihood) EVIDENCE

  16. Bayesian inference Prior 𝜄 : Mathematical model parameter, 𝐸 : Data probability BELIEF 𝑄 𝜄 𝐸 ∝𝑄 𝐸 𝜄 𝑄 𝜄 Probability of data given 𝜄 (likelihood) EVIDENCE

  17. Bayesian inference Prior 𝜄 : Mathematical model parameter, 𝐸 : Data probability BELIEF Posterior 𝑄 𝜄 𝐸 ∝𝑄 𝐸 𝜄 𝑄 𝜄 probability Probability of data given 𝜄 (likelihood) EVIDENCE

  18. Bayesian inference Prior 𝜄 : Mathematical model parameter, 𝐸 : Data probability BELIEF Posterior 𝑄 𝜄 𝐸 ∝𝑄 𝐸 𝜄 𝑄 𝜄 probability Probability of data What if we can’t use a given 𝜄 (likelihood) likelihood function? EVIDENCE

  19. ABC rejection algorithm Figure from https://doi.org/10.1371/journal.pcbi.1002803

  20. ABC rejection algorithm 1. Sample 𝜄 ∗ from the prior distribution 𝑄 𝜄 Figure from https://doi.org/10.1371/journal.pcbi.1002803

  21. ABC rejection algorithm 1. Sample 𝜄 ∗ from the prior distribution 𝑄 𝜄 2. Simulate a dataset 𝐸 ∗ from your model using 𝜄 ∗ Figure from https://doi.org/10.1371/journal.pcbi.1002803

  22. ABC rejection algorithm 1. Sample 𝜄 ∗ from the prior distribution 𝑄 𝜄 2. Simulate a dataset 𝐸 ∗ from your model using 𝜄 ∗ 3. If 𝑒 𝐸, 𝐸 ∗ ≤ 𝜗 accept 𝜄 ∗ , otherwise reject Figure from https://doi.org/10.1371/journal.pcbi.1002803

  23. ABC rejection algorithm 1. Sample 𝜄 ∗ from the prior distribution 𝑄 𝜄 2. Simulate a dataset 𝐸 ∗ from your model using 𝜄 ∗ 3. If 𝑒 𝐸, 𝐸 ∗ ≤ 𝜗 accept 𝜄 ∗ , otherwise reject 4. Repeat until you have accepted 𝑂 accepted samples Figure from https://doi.org/10.1371/journal.pcbi.1002803

  24. ABC rejection algorithm 1. Sample 𝜄 ∗ from 𝑄 𝜄 2. Simulate a dataset 𝐸 ∗ from your model using 𝜄 ∗ 3. Calculate the summary statistic for the observed data 𝝂 = 𝑻(𝑬) and simulated data 𝝂 = 𝑻 𝑬 ∗ 4. If 𝒆 𝑻(𝑬), 𝑻 𝑬 ∗ ≤ 𝝑 accept 𝜾 ∗ , otherwise reject 5. Repeat until you have accepted 𝑂 accepted samples

  25. ABC rejection algorithm 1. Sample 𝜄 ∗ from 𝑄 𝜄 2. Simulate a dataset 𝐸 ∗ from your model using 𝜄 ∗ Summary statistic for 3. Calculate the summary statistic for model trajectory the observed data 𝝂 = 𝑻(𝑬) and simulated data 𝝂 = 𝑻 𝑬 ∗ Distance measure between 4. If 𝒆 𝑻(𝑬), 𝑻 𝑬 ∗ ≤ 𝝑 accept 𝜾 ∗ , summary statistic and data otherwise reject 5. Repeat until you have accepted 𝑂 accepted samples

  26. 1. What is Approximate Bayesian Computation? A method to approximate the posterior distribution 𝑄 𝜄 𝐸 without a likelihood function 𝑄 𝜄 𝐸 ≈ 𝑄(𝜄|𝑒 𝑇 𝐸 , 𝑇 𝐸 ∗ ≤ 𝜗)

  27. 2. When do we use ABC instead of other methods?

  28. 2. When do we use ABC instead of other methods? • Data quality is poor, which means we have to aggregate it • The likelihood function might be costly to evaluate (it takes a long time) • Large data sets • Complicated likelihood function

  29. 2. When do we use ABC instead of other methods? • Data quality is poor, which means we have to aggregate it • The likelihood function might be costly to evaluate (it takes a long time) • Large data sets • Complicated likelihood function • Intuitive method of model fitting • Parameter -> model trajectory -> accept or reject

  30. 3. How do we use ABC? a. Choices in the ABC- rejection algorithm

  31. Choice of summary statistic(s) 𝑻(𝑬) • This is how we choose whether to accept or reject parameter values • Sufficient summary statistic will give the same result as the likelihood • "no other statistic that can be calculated from the same sample provides any additional information as to the value of the parameter" s Fisher, R.A. (1922). "On the mathematical foundations of theoretical statistics". Philosophical Transactions of the Royal Society A. 222 : 309–368. doi:10.1098/rsta.1922.0009. JFM 48.1280.02. JSTOR 91208.

  32. Choice of summary statistic(s) 𝑻(𝑬) • This is how we choose whether to accept or reject parameter values • Sufficient summary statistic will give the same result as the likelihood • "no other statistic that can be calculated from the same sample provides any additional information as to the value of the parameter" • If we haven’t written down a likelihood then we can’t know if our summary statistics are sufficient… s Fisher, R.A. (1922). "On the mathematical foundations of theoretical statistics". Philosophical Transactions of the Royal Society A. 222 : 309–368. doi:10.1098/rsta.1922.0009. JFM 48.1280.02. JSTOR 91208.

  33. Choice of summary statistic(s) 𝑻(𝑬) • This is how we choose whether to accept or reject parameter values • Sufficient summary statistic will give the same result as the likelihood • "no other statistic that can be calculated from the same sample provides any additional information as to the value of the parameter" • In practice • Look at published model fitting studies using ABC methods for ideas for sufficient statistics • Check with simulated data ! Fisher, R.A. (1922). "On the mathematical foundations of theoretical statistics". Philosophical Transactions of the Royal Society A. 222 : 309–368. doi:10.1098/rsta.1922.0009. JFM 48.1280.02. JSTOR 91208.

  34. Number of particles (N) • The more the better, but computation time must be taken into account

  35. Tolerance value 𝝑 • Determines whether you accept or reject parameter(s) based on how closely the model prediction matches you data • Too small and the algorithm will take a long time to run • Too big and the final distribution of particles will be too wide

  36. Tolerance value 𝝑 • Determines whether you accept or reject parameter(s) based on how closely the model prediction matches you data • Too small and the algorithm will take a long time to run • Too big and the final distribution of particles will be too wide • The magnitude of the tolerance value 𝝑 will depend on your distance measure

Recommend


More recommend