the value of bayesian statistics
play

The value of Bayesian statistics for assessing comparability - PowerPoint PPT Presentation

The value of Bayesian statistics for assessing comparability Timothy Mutsvari (Arlenda) on behalf of EFSPI Working Group Agenda Bayesian Methods: General Principles Direct Probability Statements Posterior Predictive Distribution


  1. The value of Bayesian statistics for assessing comparability Timothy Mutsvari (Arlenda) on behalf of EFSPI Working Group

  2. Agenda • Bayesian Methods: General Principles • Direct Probability Statements • Posterior Predictive Distribution • Biosimilarity Model formulation • Sample Size Justification • Multiplicity • Multiple CQAs • Assurance (not Power)

  3. Bayesian Methods: General Principles

  4. Two different ways to make a decision based on Pr 𝐩𝐜𝐭𝐟𝐬𝐰𝐟𝐞 𝐞𝐛𝐮𝐛 𝐨𝐩𝐮 𝐜𝐣𝐩𝐭𝐣𝐧𝐣𝐦𝐛𝐬 ) A  Better known as the p-value concept  Used in the null hypothesis test (or decision)  This is the likelihood of the data assuming an hypothetical explanation (e.g. the “null hypothesis”)  Classical statistics perspective (Frequentist) Pr 𝐜𝐣𝐩𝐭𝐣𝐧𝐣𝐦𝐛𝐬 𝐩𝐜𝐭𝐟𝐬𝐰𝐟𝐞 𝐞𝐛𝐮𝐛 ) B  Bayesian perspective  It is the probability of similarity given the data 3 / 18

  5. Bayesian Principle • After having observed the data of the study, the prior distribution of the treatment effect is updated to obtain the posterior distribution PRIOR distribution STUDY data POSTERIOR distribution 0.5 1.2 1.0 0.4 0.8 0.3  0.6 0.2 + 0.4 0.1 0.2 0.0 0.0 0 2 4 6 8 10 0 2 4 6 8 10 P(treatment effect > 5.5)= P(success) • Instead of having a point estimate (+/- standard deviation), we have a complete distribution for any parameter of interest

  6. Posterior Predictive Distribution • Given the model and the posterior distribution of its parameters, what are the plausible values for a future observation 𝑧 ? • This can be answered by computing the plausibility of the possible values of 𝑧 conditionally on the available information: 𝑞 𝑧 𝑒𝑏𝑢𝑏 = 𝑞 𝑧 𝜄 𝑞 𝜄 𝑒𝑏𝑢𝑏 𝑒𝜄 • The factors in the integrant are - 𝑞 𝑧 𝜄 : it is given by the model for given values of the parameters - 𝑞(𝜄|𝑒𝑏𝑢𝑏) : it is the posterior distribution of the model parameter

  7. Posterior Predictive Distribution - Illustration 3rd , repeat this 2nd , draw an observation 1 st , draw a mean and operation a large number from the resulting a variance from: of time to obtain the distribution predictive distribution Y~ Normal(µ i , σ ² i ) Posterior of mean µ i X X X X Posterior of Variance σ ² i given mean drawn

  8. Difference: Simulations vs Predictions Bayesian Predictions Monte Carlo Simulations the uncertainty of parameter the “new observations” are drawn estimates (location and dispersion) is from distribution “centered” on taken into account before drawing estimated location and dispersion “new observations” from relevant parameters (treated as “true distribution values”).

  9. Difference: Simulations vs Predictions Bayesian Predictions Monte Carlo Simulations the uncertainty of parameter the “new observations” are drawn estimates (location and dispersion) is from distribution “centered” on taken into account before drawing estimated location and dispersion “new observations” from relevant parameters (treated as “true distribution values”).

  10. Why Bayesian for Biosimilarity? • What is the question: • what is the probability of being biosimilar given available data? • what is the probability of having future lots within the limits given available data?   Pr 𝐩𝐜𝐭𝐟𝐬𝐰𝐟𝐞 𝐞𝐛𝐮𝐛 𝐨𝐩𝐮 𝐜𝐣𝐩𝐭𝐣𝐧𝐣𝐦𝐛𝐬) Pr 𝐂𝐣𝐩𝐭𝐣𝐧𝐣𝐦𝐛𝐬 𝐩𝐜𝐭𝐟𝐬𝐰𝐟𝐞 𝐞𝐛𝐮𝐛) vs Pr 𝐆𝐯𝐮𝐯𝐬𝐟 𝐦𝐩𝐮𝐭 𝐣𝐨 𝐦𝐣𝐧𝐣𝐮𝐭 𝐩𝐜𝐭𝐟𝐬𝐰𝐟𝐞 𝐞𝐛𝐮𝐛) • The question becomes naturally Bayesian • Many decisions can be deduced from the posterior and predictive distributions • In addition • leverage historical data (e.g. on assay variability) • Bayesian approach can easily handle multivariate problems

  11. Biosimilar Model formulation

  12. Biosimilarity Model - Univariate Case 2  Model for Biosimilar 𝐷𝑅𝐵 𝑈𝑓𝑡𝑢 ~ 𝑂 𝜈 𝑈𝑓𝑡𝑢 , 𝜏 𝑈𝑓𝑡𝑢 2 𝐷𝑅𝐵 𝑆𝑓𝑔 ~ 𝑂 𝜈 𝑆𝑓𝑔 , 𝜏 𝑠𝑓𝑔  Model for Ref 2  Test will not extremely different from Ref 2 𝜏 𝑈𝑓𝑡𝑢 = 𝛽 0 ∗ 𝜏 𝑠𝑓𝑔 𝛽 0 ~ Uniform (𝑏, 𝑐) , for well chosen 𝑏 & 𝑐 , e.g. 1/10 to 10 • From this model: • directly derive the PI/TI from predictive distributions • easily extendable to multivariate model • power computations are straight forward from predictive distributions

  13. Model performance (compare to true pars)

  14. Biosimilarity Model - Univariate Case • Variability can be decomposed to: 2 2 𝜏 𝑈𝑓𝑡𝑢 + 𝜏 𝑏𝑡𝑡𝑏𝑧 2 2 𝜏 𝑆𝑓𝑔 + 𝜏 𝑏𝑡𝑡𝑏𝑧 • Synthesize assay historical data into informative prior for variability (all other pars being non-informative)

  15. Bayesian PI/TI – Illustration (1) Predictive Distributions (non-informative Prior on all parameters) Likelihood Ref Predictive Distribution  Tolerance Intervals (e.g. Wolfinger)

  16. Bayesian PI/TI – Illustration (2) Predictive Distributions (non-informative Prior on all parameters) Likelihood Ref Test Predictive Distribution  Prediction Interval Predictive Distribution  Tolerance Intervals (e.g. Wolfinger)

  17. Bayesian PI/TI – Illustration (3) Prior Predictive Distributions (informative on validated Likelihood assay variance) Ref Test Predictive Distribution  Prediction Interval Predictive Distribution  Tolerance Intervals (e.g. Wolfinger)

  18. Sample Size Calculation

  19. Sample size for Biosimilarity Evaluation • Sample Test data from the predictive • How many new batches given past results to be within specs?

  20. Multiplicity Extension of the univariate case

  21. Bayesian - Multivariate CQA Model • Let • 𝒀 be 𝑜 × 𝑙 matrix of observations for test. • 𝒁 be 𝑛 × 𝑙 matrix of observations for ref. 𝒀 𝒁 ~ 𝑵𝑾𝑶 𝝂 𝑼 𝜯 𝑼 𝜯 𝑺𝑼 𝝂 𝑺 , 𝜯 𝑺 𝜯 𝑺𝑼 • Any test FDA Tier1, FDA Tier2 or PI/TI can be easily computed • Pr [𝐔𝐟𝐭𝐮 − 𝐒𝐟𝐠]|𝐄𝐛𝐮𝐛 ~ 𝑵𝑾𝑶 𝝂 𝑼 − 𝝂 𝑺 , [𝜯 𝑼 +𝜯 𝑺 − 𝟑 ∗ 𝜯 𝑺𝑼 ]

  22. Multivariate CQA Model • Use Ref. predictive to compute the limits of k CQAs • Compare the Test data from k CQAs to the limits • To get the joint test: • Calculate the joint acceptance probability

  23. Assurance (not Power)

  24. Assurance (Bayesian Power) • Unconditional probability of significance given prior - O’Hagan et al. (2005) • Expectation of the power averaged over the prior distribution • ‘ True Probability of Success ’ of a trial • In Frequentist power is based on a particular value of the effect • A very ‘ strong ’ prior

  25. Power vs assurance independent samples t-test ( H 0 : 𝜈 1 = 𝜈 2 vs H 1 : 𝜈 1 ≠ 𝜈 2 ) bayesian approach (assurance) • In order to reflect the uncertainty, a large number of effect sizes, i.e. (𝜈 1 −𝜈 2 )/𝜏 pooled , are generated power assurance using the prior distributions. • A power curve is obtained for each effect size • the expected (weighted by prior beliefs) power curve is calculated

  26. Conclusions • Using Bayesian approach: • I can directly derive probabilities of interest • Uncertainties are well propagated • Bayesian predictive distribution answers the very objective • probability of biosimilar given data • future lots to remain within specs • leverage historical data  save costs • Informative priors can be justified and recommended • Correlated CQAs • Easily compute joint acceptance probability

  27. When SIM IMILAR is not the SAME!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend