likelihood based statistical decisions
play

Likelihood-Based Statistical Decisions Marco Cattaneo Seminar for - PowerPoint PPT Presentation

Likelihood-Based Statistical Decisions Marco Cattaneo Seminar for Statistics ETH Z urich, Switzerland July 23, 2005 Likelihood Function set of statistical models { P : } observation A likelihood function lik : P


  1. Likelihood-Based Statistical Decisions Marco Cattaneo Seminar for Statistics ETH Z¨ urich, Switzerland July 23, 2005

  2. Likelihood Function set of statistical models { P θ : θ ∈ Θ } observation A ❀ likelihood function lik : θ �→ P θ ( A )

  3. Likelihood Function set of statistical models { P θ : θ ∈ Θ } observation A ❀ likelihood function lik : θ �→ P θ ( A ) The likelihood function lik measures the relative plausibility of the models P θ , on the basis of the observation A alone. The likelihood function lik is not calibrated : only ratios lik ( θ 1 ) /lik ( θ 2 ) are well determined.

  4. Likelihood Function set of statistical models { P θ : θ ∈ Θ } observation A ❀ likelihood function lik : θ �→ P θ ( A ) The likelihood function lik measures the relative plausibility of the models P θ , on the basis of the observation A alone. The likelihood function lik is not calibrated : only ratios lik ( θ 1 ) /lik ( θ 2 ) are well determined. Example. X ∼ Binomial ( n, θ ) n = 5 , θ ∈ Θ = [0 , 1] lik ( θ ) ∝ θ 3 (1 − θ ) 2 x = 3 ⇒ 0 0.2 0.4 0.6 0.8 1 θ

  5. Statistical Decision Problem set of statistical models { P θ : θ ∈ Θ } set of possible decisions D loss function L : Θ × D → [0 , ∞ ) L ( θ, d ) is the loss we would incur, according to the model P θ , by making the decision d .

  6. Statistical Decision Problem set of statistical models { P θ : θ ∈ Θ } set of possible decisions D loss function L : Θ × D → [0 , ∞ ) L ( θ, d ) is the loss we would incur, according to the model P θ , by making the decision d . observation A ❀ likelihood function lik on Θ MPL criterion: minimize sup θ lik ( θ ) L ( θ, d )

  7. Statistical Decision Problem set of statistical models { P θ : θ ∈ Θ } set of possible decisions D loss function L : Θ × D → [0 , ∞ ) L ( θ, d ) is the loss we would incur, according to the model P θ , by making the decision d . observation A ❀ likelihood function lik on Θ MPL criterion: minimize sup θ lik ( θ ) L ( θ, d ) minimax criterion: minimize sup θ L ( θ, d ) MPL = minimax if lik is constant (i.e., complete ignorance about Θ ) MPL: Minimax Plausibility-weighted Loss

  8. Example lik ( θ ) ∝ θ 3 (1 − θ ) 2 L ( θ, d ) = | d − θ 2 | d ML = 0 . 36 , d MP L ≈ 0 . 385 , d BU ≈ 0 . 335 0 0.2 0.4 0.6 0.8 1 θ

  9. Example lik ( θ ) ∝ θ 3 (1 − θ ) 2 L ( θ, d ) = | d − θ 2 | d ML = 0 . 36 , d MP L ≈ 0 . 385 , d BU ≈ 0 . 335 0 0.2 0.4 0.6 0.8 1 θ 2 (1 − √ τ ) 2 3 τ = θ 2 , lik ( τ ) ∝ τ L ( τ, d ) = | d − τ | d ML = 0 . 36 , d MP L ≈ 0 . 385 , d BU ≈ 0 . 404 0 0.2 0.4 0.6 0.8 1 τ

  10. Example lik ( θ ) ∝ θ 3 (1 − θ ) 2 L ( θ, d ) = | d − θ 2 | d ML = 0 . 36 , d MP L ≈ 0 . 385 , d BU ≈ 0 . 335 0 0.2 0.4 0.6 0.8 1 θ 2 (1 − √ τ ) 2 3 τ = θ 2 , lik ( τ ) ∝ τ L ( τ, d ) = | d − τ | d ML = 0 . 36 , d MP L ≈ 0 . 385 , d BU ≈ 0 . 404 0 0.2 0.4 0.6 0.8 1 τ � 2 | d − τ | if d ≤ τ L ( τ, d ) = | d − τ | if d ≥ τ d ML = 0 . 36 , d MP L ≈ 0 . 468 , d BU ≈ 0 . 502 ( d BU ≈ 0 . 435 using θ )

  11. Relative Plausibility The likelihood function can be easily updated by multiplying it with the (conditional) likelihood functions based on the new observations. Prior information can be encoded in a “prior likelihood function” assumed to be based on past (independent) observations.

  12. Relative Plausibility The likelihood function can be easily updated by multiplying it with the (conditional) likelihood functions based on the new observations. Prior information can be encoded in a “prior likelihood function” assumed to be based on past (independent) observations. The relative plausibility is the extension of the likelihood function to the subsets H of Θ by means of the supremum: rp ( H ) ∝ sup θ ∈H lik ( θ ) .

  13. Relative Plausibility The likelihood function can be easily updated by multiplying it with the (conditional) likelihood functions based on the new observations. Prior information can be encoded in a “prior likelihood function” assumed to be based on past (independent) observations. The relative plausibility is the extension of the likelihood function to the subsets H of Θ by means of the supremum: rp ( H ) ∝ sup θ ∈H lik ( θ ) . The relative plausibility is thus a quantitative description of the uncertain knowledge about the models P θ , that can start with complete ignorance or with prior information, that can be easily updated when new data are observed, and that can be used for inference and decision making.

  14. Imprecise Probabilities The relative plausibility is a non-calibrated possibility measure on Θ .

  15. Imprecise Probabilities The relative plausibility is a non-calibrated possibility measure on Θ . MPL criterion: minimize sup θ rp { θ } L ( θ, d ) � �� � Shilkret integral of L ( · , d ) with respect to rp

  16. Imprecise Probabilities The relative plausibility is a non-calibrated possibility measure on Θ . MPL criterion: minimize sup θ rp { θ } L ( θ, d ) � �� � Shilkret integral of L ( · , d ) with respect to rp If Γ is a set of probability measures on Θ , the consideration of the (second- order) relative plausibility on Γ leads to a non-calibrated possibilistic hierarchical model , which allows non-vacuous conclusions even if Γ is the set of all probability measures on Θ .

  17. Properties The relative plausibility and the MPL criterion:

  18. Properties The relative plausibility and the MPL criterion: • are simple and intuitive.

  19. Properties The relative plausibility and the MPL criterion: • are simple and intuitive. • are parametrization invariant.

  20. Properties The relative plausibility and the MPL criterion: • are simple and intuitive. • are parametrization invariant. • lead to decision functions that are equivariant (if the problem is invariant) and asymptotic optimal (if some regularity conditions are satisfied).

  21. Properties The relative plausibility and the MPL criterion: • are simple and intuitive. • are parametrization invariant. • lead to decision functions that are equivariant (if the problem is invariant) and asymptotic optimal (if some regularity conditions are satisfied). • satisfy the strong likelihood principle.

  22. Properties The relative plausibility and the MPL criterion: • are simple and intuitive. • are parametrization invariant. • lead to decision functions that are equivariant (if the problem is invariant) and asymptotic optimal (if some regularity conditions are satisfied). • satisfy the strong likelihood principle. • can use pseudo likelihood functions.

  23. Properties The relative plausibility and the MPL criterion: • are simple and intuitive. • are parametrization invariant. • lead to decision functions that are equivariant (if the problem is invariant) and asymptotic optimal (if some regularity conditions are satisfied). • satisfy the strong likelihood principle. • can use pseudo likelihood functions. • can represent complete (or partial) ignorance.

  24. Properties The relative plausibility and the MPL criterion: • are simple and intuitive. • are parametrization invariant. • lead to decision functions that are equivariant (if the problem is invariant) and asymptotic optimal (if some regularity conditions are satisfied). • satisfy the strong likelihood principle. • can use pseudo likelihood functions. • can represent complete (or partial) ignorance. • can handle prior information in a natural way.

  25. Example Estimation of the variance components in the 3 × 3 random effect one-way layout, under normality assumptions and weighted squared error loss. ve � va � ( SSa + SSe ) ( SSa + SSe ) 0.16 0.15 0.1 0.12 0.05 0.08 0 0 0.2 0.4 0.6 0.8 1 0.04 SSa/(SSa+SSe) -0.05 0 0 0.2 0.4 0.6 0.8 1 MPL SSa/(SSa+SSe) MPL ANOVA ANOVA = ANOVA+ = MINQU ML ML ReML = ANOVA+ ReML nonneg. MINQ min. bias

  26. Example ve − ve )2] va − va )2] 3 E [( � E [( � ve 2 ( va +1 3 ve )2 1 1.6 0.95 1.2 0.9 0.85 0.8 0.8 0.4 0.75 0.7 0 0.2 0.4 0.6 0.8 1 Va/(Va+Ve) 0 0.2 0.4 0.6 0.8 1 MPL Va/(Va+Ve) ANOVA MPL ANOVA = ANOVA+ = MINQU ML ReML = ANOVA+ ML nonneg. MINQ min. bias ReML

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend