tutorials
play

Tutorials 3 tutorials: Day 1: introduction to Bayesian analysis and - PowerPoint PPT Presentation

Tutorials 3 tutorials: Day 1: introduction to Bayesian analysis and BAT, basic examples Day 2: limit setting, including systematics and priors Day 3: hypothesis testing, counting experiment Basics of Bayesian approach Procedure of learning


  1. Tutorials 3 tutorials: Day 1: introduction to Bayesian analysis and BAT, basic examples Day 2: limit setting, including systematics and priors Day 3: hypothesis testing, counting experiment

  2. Basics of Bayesian approach Procedure of learning from experiment: (for a fixed model) ∣  D  ∝ p   p  D ∣  p 0   p 0    - probability of the parameters  — Prior  containing all the knowledge before the experiment  defines allowed ranges for parameters  contains preconceptions about possible values of parameters p     D ∣   D - probability of the data given the parameters — Likelihood  how likely is the data a representation of the model  contains information from theory as well as modelling of experimental conditiona ∣    p   - probability for parameters given the data — Posterior D  D  result of the analysis containing all the knowledge about the parameters after the experiment p   D ∣  p 0   After adding a normalization  ∣  p  D  = ∫ p   D ∣  p 0   d   Bayes formula IMPRS Block course, 19.10.2009 Daniel Kollár #2

  3. Posterior probability Posterior probability density contains all information Typically, several quantities are extracted: mode, mean, variance, ... ● marginalized posterior distributions ● posterior integrated over all but one (two, ...) parameter(s) – D  ∏ D  = ∫ p  p  i ∣  ∣  d  j ≠ i p  1 ∣  D  j  2 generally non-trivial and in many cases – practically impossible to do using standard techniques key tool: – p  2 ∣  D  Markov Chain Monte Carlo  1 IMPRS Block course, 19.10.2009 Daniel Kollár #3

  4. MCMC — Metropolis algorithm f (y) accepted with probability f ( y )/ f ( x i ) f ( x i ) always In BAT implemented Metropolis algorithm ● accepted Map positive function f ( x ) by random walk f (y) ● towards higher probabilities y x i y Algorithm: ● Start at some randomly chosen x i – Randomly generate y – If f ( y ) ≥ f ( x i ), set x i + 1 = y – p = f  y  If f ( y ) < f ( x i ), set to x i + 1 = y with probability – f  x i  If y not accepted, stay where you are, i.e. set x i + 1 = x i – Start over – Sampling is enhanced in regions with higher values of f(x) ● IMPRS Block course, 19.10.2009 Daniel Kollár #4

  5. MCMC: an example linear scale log scale mapping an arbitrary function: ● f  x  = x 4 sin 2 x number of iterations distribution sampled by MCMC in ● 1000 this case quickly converges towards the underlying distribution mapping of complicated ● shapes with multiple minima and maxima 10000 Note: MCMC has to become stationary ● to sample from underlying 1000000 distribution in general the convergence is a ● non-trivial problem IMPRS Block course, 19.10.2009 Daniel Kollár #5

  6. Scanning parameter space with MCMC Use MCMC to scan p   ● D ∣  p 0     ∣  p  parameter space of  D  = ∫ p   D ∣  p 0   d    = p   f  D ∣  p 0   ● MCMC converges ● p  1 ∣  D  towards underlying distribution  2 Determining of the – overall probability p  2 ∣  D  distribution of the p  ∣  parameters D  Marginalize wrt. individual ● parameters while walking → obtain D  ∏ D  = ∫ p  p  i ∣  ∣  d  j ≠ i j Find maximum (mode) ●  1 Uncertainty propagation ● IMPRS Block course, 19.10.2009 Daniel Kollár #6

  7. Idea behind BAT ● Provide flexible environment to phrase arbitrary problems ● Provide set of numerical tools ● C++ based framework (flexible, modular) ● Interfaces to ROOT, Cuba, Minuit, user defined, ... ● create model Define MODEL USER DEFINED  ● define parameters ● read-in data  ● define likelihood p   D ∣  ● define priors p 0   ● normalize MODEL ● find mode / fit INDEPENDENT Read DATA ● test the fit (common tools) ● from text file, ROOT ● marginalize wrt. one tree, user defined or two parameters ● compare models ● nice output IMPRS Block course, 19.10.2009 Daniel Kollár #7

  8. BAT distribution download from ● http://www.mppmu.mpg.de/bat comes in the form of shared library ● (plus a .rootmap file for interactive ROOT session) depends on ROOT I/O and drawing functionality ● can in principle be removed if there's need – can be compiled with Cuba support ● BAT contains 15 classes at the moment which ● provide: main infrastructure – algorithms – output and logging – extension classes to solve specific (frequent) fitting problems – IMPRS Block course, 19.10.2009 Daniel Kollár #8

  9. Tutorials – DAY 1 IMPRS Block course, 19.10.2009 Daniel Kollár #9

  10. BAT examples examples are part of BAT distribution ● demonstrate basic functionality and program/macro structure ● ROOT macros: 01 simple GraphFitter function fitting assuming Gaussian uncertainties – 02 HistogramFitter function fitting assuming Poisson uncertainties – 03 advanced GraphFitter as 01 but for more functions at the same time – model comparison – 04 EfficiencyFitter function fitting assuming Binomial uncertainties – Compiled program: 05 advanced fitter function fitting assuming arbitrary asymmetric uncertainties – IMPRS Block course, 19.10.2009 Daniel Kollár #10

  11. example01 - GraphFitter Simple ROOT macro: Data to fit are in a TGraphErrors TGraphErrors * graph = ...; object (uncertainties are necessary) Function to fit is defined via TF1 TF1 * f1 = new TF1(“f1”,”[0]+[1]*x”, 0., 100.); object f1->SetParLimits(0, -20., 30.); Parameter ranges have to be f1->SetParLimits(1, .5, 1.5); specified ! // BAT fitting bellow Define BCGraphFitter , assume BCGraphFitter * gf = new BCGraphFitter(graph, f1); gaussian uncertainties gf->Fit(); Fit the TGraphErrors with TF1 function using BAT Draw the data, the best fit and the gf->DrawFit( "" , true); error band gf->PrintAllMarginalized(“file.ps”); Print all Marginalized distributions IMPRS Block course, 19.10.2009 Daniel Kollár #11

  12. example01 - GraphFitter Output Fit and the error band representing the central 68%probability interval Running 100000 iterations for 5 chains ● approximately 30 seconds ● ! very simple example ! Marginalized posterior probability densities IMPRS Block course, 19.10.2009 Daniel Kollár #12

  13. example02 - HistogramFitter Simple ROOT macro: takes histogram as an input ● uses Poissonian uncertainties ● define data histogram TH1D * histo = ...; define fit function and parameter ranges TF1 * f1 = ...; BCHistogramFitter * hf = new BCHistogramFitter(histo, f1); hf -> Fit(); hf -> DrawFit( "" , true); ... Use whenever you want to fit a histogram. IMPRS Block course, 19.10.2009 Daniel Kollár #13

  14. example03 - GraphFitter The example from the BAT paper Fit data set using: ● I. 2nd order polynomial (no peak) II. gaussian peak + constant III. gaussian peak + straight line IV. gaussian peak + 2nd order pol. Assume flat a priori probabilities ● in certain ranges of parameters, p 0  i.e.  = const. Search for peak in range from 2. to 18. with maximum sigma of 4. ● Data were generated as gaussian peak + 2nd order polynomial ● IMPRS Block course, 19.10.2009 Daniel Kollár #14

  15. Posterior probability density 2nd order polynomial Marginalized probability wrt. one or two parameters D  = ∫ p   p  i ∣  D ∣  p 0   d   j ≠ i Integrated over all other parameters ● All calculated during single MCMC ● run Stored as TH1D and TH2D ● In general, mode of the marginalized ● distribution not equal to global mode Extracted values left to the user ● Default output: ● Mean – Central 68% interval – Limits – Mode – Median – All information about the posterior PDF is in the Markov chain (stored as ROOT Tree) IMPRS Block course, 19.10.2009 Daniel Kollár #15

  16. Posterior probability density peak + straight line Central interval not always optimal ● Multiple maxima in parameter space ● MCMC follows probability ● distributions with complicated shapes Optionally calculate smallest interval ● IMPRS Block course, 19.10.2009 Daniel Kollár #16

  17. Fits + Uncertainty bands Uncertainty band ● calculated during Markov Chain run calculate f ( x ) at many – different x using sampled  fill 2D histogram in ( x , y ) – after run look at – distribution of y at given x and calculate central 68% interval Fit can lie outside of the ● central 68% interval again, central 68% not – optimal IMPRS Block course, 19.10.2009 Daniel Kollár #17

  18. Uncertainty multi-band Calculating f(x) at many x for every set of parameters sampled in the ● Markov chain ⇔ uncertainty propagation Can lead to multiple uncertainty bands ● Probability density for true y at x=5.0 p(y) at x=5.0 Data Best fit Smallest 68% Smallest 68% f(x=5.0) for the best fit IMPRS Block course, 19.10.2009 Daniel Kollár #18

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend