Bayesian Fitting Probabilistic Morphable Models Summer School, June - - PowerPoint PPT Presentation

bayesian fitting
SMART_READER_LITE
LIVE PREVIEW

Bayesian Fitting Probabilistic Morphable Models Summer School, June - - PowerPoint PPT Presentation

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL Bayesian Fitting Probabilistic Morphable Models Summer School, June 2017 Sandro Schnborn University of Basel > DEPARTMENT OF


slide-1
SLIDE 1

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Bayesian Fitting

Probabilistic Morphable Models Summer School, June 2017 Sandro Schönborn University of Basel

slide-2
SLIDE 2

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Uncertainty: Probability Distributions

  • Probabilistic Models
  • Uncertain Observation (noise, outlier, occlusion, …)
  • Fitting: Model explanation of observed data – probabilistic?

2

Tells us about the outcome’s certainty!

Observations Fit & Certainty Ground truth

Bishop PRML, 2006

slide-3
SLIDE 3

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Probability: An Example

  • Dentist example: Does the patient have a cavity?

3

Bu But t th the pati tient t eith ther has a cavity ty or does not

  • There is no 80% cavity!
  • Having a cavity should not depend on whether the

patient has a toothache or gum problems

All these statements do not contradict each other, they summarize th the denti tist’s know

  • wledge about the patient

Certainty 𝑄 cavity = 0.1 𝑄 cavity toothache) = 0.8 𝑄 cavity toothache, gum problems) = 0.4

AIMA: Russell & Norvig, Artificial Intelligence. A Modern Approach, 3rd edition,

slide-4
SLIDE 4

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Uncertainty: Bayesian Probability

  • How are probabilities to be interpreted?

They are sometimes contradictory: Why does the distribution change when we have more data? Shouldn’t there be a real distribution of 𝑄 𝜄 ?

  • Bayesian probabilities rely on a subjective perspective:

Probability is used to express our current knowledge. It can change when we learn or see more: With more data, we are more certain about our result.

  • Not subjective in the sense that it is arbitrary!

There are quantitative rules to follow mathematically

  • Probability expresses an observers certainty, often called be

belie lief

4

Subjectivity: There is no single, real underlying distribution. A probability distribution expresses our knowledge – It is different in different situations and for different observers since they have different knowledge.

slide-5
SLIDE 5

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Towards Bayesian Inference

5

Posterior Models: Gaussian Process Regression Probabilistic Fit: Probabilistic interpretation of data Observed Points Posterior Model Update of prior to posterior model: Bayesian Inference

slide-6
SLIDE 6

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Belief Updates

6

Mo Model Face distribution Ob Obser ervation Concrete points Possibly uncertain Po Posterior Face distribution consistent with observation Prior belief More knowledge Posterior belief Consistency: Laws of probability calculus!

slide-7
SLIDE 7

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Joint Distribution

Marginal

Distribution of certain points only

Conditional

Distribution of points conditioned on known values of others

7

Probabilistic model: joint distribution of points

𝑄 𝑦>|𝑦@ = 𝑄 𝑦>, 𝑦@ 𝑄 𝑦@ 𝑄 𝑦> = A 𝑄(𝑦>, 𝑦@)

  • DE

𝑄 𝑦>, 𝑦@

Both can be easily calculated for Gaussian models

slide-8
SLIDE 8

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Certain Observation

  • Observations are known values
  • Distribution of 𝑦> after
  • bserving 𝑦@, … , 𝑦G:

𝑄 𝑦>|𝑦@ … 𝑦G = 𝑄 𝑦>, 𝑦@, … , 𝑦G 𝑄 𝑦@, … , 𝑦G

  • Conditional probability

8

slide-9
SLIDE 9

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Towards Bayesian Inference

  • Update belief about 𝑦> by observing 𝑦@, … , 𝑦G

𝑄 𝑦> → 𝑄 𝑦> 𝑦@ … 𝑦G

  • Factorize joint distribution

𝑄 𝑦>, 𝑦@, … , 𝑦G = 𝑄 𝑦@, … , 𝑦G|𝑦> 𝑄 𝑦>

  • Rewrite conditional distribution

𝑄 𝑦>|𝑦@ … 𝑦G = 𝑄 𝑦>, 𝑦@, … , 𝑦G 𝑄 𝑦@, … , 𝑦G = 𝑄 𝑦@, … , 𝑦G|𝑦> 𝑄 𝑦> 𝑄 𝑦@, … , 𝑦G

  • General: Query (𝑅) and Evidence (𝐹)

𝑄 𝑅|𝐹 = 𝑄 𝑅, 𝐹 𝑄 𝐹 = 𝑄 𝐹|𝑅 𝑄 𝑅 𝑄 𝐹

9

slide-10
SLIDE 10

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Uncertain Observation

  • Observations with uncertainty

Model needs to describe how

  • bservations are distributed

with joint distribution 𝑄 𝑅, 𝐹

  • Still conditional probability

But joint distribution is more complex

  • Joint distribution factorized

𝑄 𝑅, 𝐹 = 𝑄 𝐹|𝑅 𝑄 𝑅

  • Likelihood 𝑄 𝐹|𝑅
  • Prior 𝑄 𝑅

10

slide-11
SLIDE 11

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Likelihood

𝑄 𝑅, 𝐹 = 𝑄 𝐹|𝑅 𝑄 𝑅

  • Likelihood x prior: factorization is more flexible than full joint
  • Prior: distribution of core model without observation
  • Likelihood: describes how observations are distributed
  • Common example: Gaussian distributed points

11

Pr Prior Li Likelihood Joi Joint 𝑄 𝑅 𝑄 𝐹|𝑅 𝑹 𝑭

slide-12
SLIDE 12

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Bayesian Inference

  • Conditional/Bayes rule: method to update beliefs

𝑄 𝑅|𝐹 = 𝑄 𝐹|𝑅 𝑄 𝑅 𝑄 𝐹

  • Each observation updates our belief (changes knowledge!)

𝑄 𝑅 → 𝑄 𝑅 𝐹 → 𝑄 𝑅 𝐹, 𝐺 → 𝑄 𝑅 𝐹, 𝐺, 𝐻 → ⋯

  • Bayesian Inference: How beliefs evolve with observation
  • Recursive: Posterior becomes prior of next inference step

12

Pr Prior Li Likelihood Po Posterior Ma Marginal Likelihood

slide-13
SLIDE 13

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Marginalization

  • Models contain irrelevant/hidden variables

e.g. points on chin when nose is queried

  • Marginalize over hidden variables (𝐼)

13

𝑄 𝑅 𝐹 = A 𝑄 𝑅, 𝐼 𝐹

  • Q

= A 𝑄 𝐹, 𝐼|𝑅 𝑄 𝑅 𝑄 𝐹, 𝐼

  • Q
slide-14
SLIDE 14

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

General Bayesian Inference

  • Observation of additional variables
  • Common case, e.g. face rendering, landmark locations
  • Coupled to core model via likelihood factorization
  • General Bayesian inference case:
  • Distribution of data 𝐸 (formerly Evidence)
  • Parameters 𝜄 (formerly Query)

𝑄 𝜄|𝐸 = 𝑄 𝐸|𝜄 𝑄 𝜄 𝑄 𝐸 𝑄 𝜄|𝐸 ∝ 𝑄 𝐸|𝜄 𝑄 𝜄

14

slide-15
SLIDE 15

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Example: Bayesian Curve Fitting

  • Curve Fitting: Data interpretation with a model
  • Posterior distribution expresses certainty
  • in parameter space
  • in the predictive distribution

15

slide-16
SLIDE 16

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Posterior of Regression Parameters

16

No data N=1 N=2 N=19

𝑄 𝑥 𝐸>

Bishop PRML, 2006

𝑄 𝑥 𝑄 𝑥 𝐸>, 𝐸@ 𝑄 𝑥 𝐸>, 𝐸@, …

𝑄 𝐸>|𝑥 𝑄 𝐸@|𝑥

slide-17
SLIDE 17

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

More Bayesian Inference Examples

17

Bishop PRML, 2006

Classification

e.g. Bayes classifier

Non-Linear Curve Fitting

e.g. Gaussian Process Regression

slide-18
SLIDE 18

> DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE PROBABILISTIC MORPHABLE MODELS | JUNE 2017 | BASEL

Summary: Bayesian Inference

  • Belief: formal expression of an observer’s knowledge
  • Subjective state of knowledge about the world
  • Beliefs are expressed as probability distributions
  • Formally not arbitrary: Consistency requires laws of probability
  • Observations change knowledge and thus beliefs
  • Bayesian inference formally updates prior beliefs to posteriors
  • Conditional Probability
  • Integration of observation via likelihood x prior factorization

𝑄 𝜄|𝐸 = 𝑄 𝐸|𝜄 𝑄 𝜄 𝑄 𝐸

18