Introduction to the theory of imprecise probability Erik Quaeghebeur - - PowerPoint PPT Presentation

introduction to the theory of imprecise probability
SMART_READER_LITE
LIVE PREVIEW

Introduction to the theory of imprecise probability Erik Quaeghebeur - - PowerPoint PPT Presentation

Introduction to the theory of imprecise probability Erik Quaeghebeur TU Delft, the Netherlands UTOPIAE Training School 2018, Durham, England Why would you want your probability to be imprecise? 2 versus 3 Uncertainty about outcome of. . .


slide-1
SLIDE 1

Introduction to the theory of imprecise probability

Erik Quaeghebeur

TU Delft, the Netherlands

UTOPIAE Training School 2018, Durham, England

slide-2
SLIDE 2

Why would you want your probability to be imprecise?

2

slide-3
SLIDE 3

versus

3

slide-4
SLIDE 4

Uncertainty about outcome of. . .

versus

Agents (Gamblers)

Wiske Yoko Tsuno

4

slide-5
SLIDE 5

Assessment (gambles accepted)

1 if ⊗ 5 if ⊗4 if + 1 if

Agents (Gamblers)

Wiske Yoko Tsuno

5

slide-6
SLIDE 6

Natural extension (λ, µ ⊙ 0)

λ ⊗ 5λ +µ + µ ⊗4λ + λ +µ + µ

Rational agents

Wiske Yoko Tsuno

6

slide-7
SLIDE 7

Natural extension (λ, µ ⊙ 0)

λ ⊗ 5λ +µ + µ

COHERENCE

⊗4λ + λ +µ + µ

Rational agents

Wiske Yoko Tsuno

7

slide-8
SLIDE 8

Assessment (gambles accepted)

⊗ 5 , ⊗4 +

Agents (Gamblers)

Heroine pool

8

slide-9
SLIDE 9

Natural extension (λ, µ ⊙ 0)

(λW ⊗ 4λY + µ ) + (⊗5λW + λY + µ )

Irrational agents

Heroine pool

9

slide-10
SLIDE 10

Natural extension (λ, µ ⊙ 0)

(λW ⊗ 4λY + µ ) + (⊗5λW + λY + µ )

SURE LOSS!

Irrational agents

Heroine pool

10

slide-11
SLIDE 11

Assessment (gambles accepted)

Agents (Gamblers)

Heroine pool

11

slide-12
SLIDE 12

Natural extension (λ, µ ⊙ 0)

µ + µ

Rational agents

Heroine pool

12

slide-13
SLIDE 13

Natural extension (λ, µ ⊙ 0)

µ + µ

VACUOUS

Rational agents

Heroine pool

13

slide-14
SLIDE 14

Basic concepts

◮ Agent reasoning about experiment with uncertain outcome ◮ Possibility space 𝒴 of outcomes ◮ Gambles are real-valued functions of the outcomes;

ℒ = 𝒴 ⊃ R (ℒ is assumed to be a linear space)

◮ Assessment is a description of a set of acceptable gambles ◮ Natural extension of an assessment is

the set of all acceptable gambles implied by the agent’s rationality criteria (and other assumptions)

14

slide-15
SLIDE 15

Coherence, the classical rationality criteria

Constructive Positive scaling If 𝑔 is acceptable and λ > 0, then λ𝑔 is acceptable. Addition If 𝑔 and 𝑕 are acceptable, then 𝑔 + 𝑕 is acceptable. Background Accepting gain If 𝑔 is nonnegative for all outcomes, then 𝑔 is acceptable. Avoiding sure loss If 𝑕 is negative for all outcomes, then 𝑕 is not acceptable.

15

slide-16
SLIDE 16

Where are the imprecise probabilities I came here for!!!

16

slide-17
SLIDE 17

Previsions/Expectations are prices for gambles

◮ ‘Prevision’ and ‘Expectation’ are synomyms ◮ Prices are real values interpreted as constant gambles ◮ Lower prevision 𝑄(𝑔) is the

supremum acceptable buying price of 𝑔: 𝑄(𝑔) = sup¶ν ∈ R : 𝑔 ⊗ ν is acceptable♢ Upper prevision 𝑄(𝑔) is the infimum acceptable selling price of 𝑔

◮ Conjugacy of coherent lower and upper previsions:

𝑄(𝑔) = ⊗𝑄(⊗𝑔)

◮ If 𝑄(𝑔) = 𝑄(𝑔), then 𝑄(𝑔) = 𝑄(𝑔) is the prevision of 𝑔

17

slide-18
SLIDE 18

Probabilities are previsions of indicator gambles

◮ Event 𝐵 is a subset of 𝒴 ◮ Indicator gamble

1A(𝑦) =

∮︂

1, 𝑦 ∈ 𝐵, 0, 𝑦 ̸∈ 𝐵

◮ Lower probability 𝑄(𝐵) = 𝑄(1A)

Upper probability 𝑄(𝐵) = 𝑄(1A)

◮ Conjugacy of coherent lower and upper probabilities (𝐵c = 𝒴 \ 𝐵):

𝑄(𝐵) = 1 ⊗ 𝑄(𝐵c)

◮ If 𝑄(𝐵) = 𝑄(𝐵), then 𝑄(𝐵) = 𝑄(𝐵) is the probability of 𝐵

18

slide-19
SLIDE 19

Natural extension (λ, µ ⊙ 0)

(λ + µ ) +(⊗5λ + µ ) (⊗4λ + µ ) (λ + µ ) (λW ⊗ 4λY + µ ) + (⊗5λW + λY + µ ) µ + µ

Agents

Wiske Yoko Tsuno Irrational pool Rational pool

19

slide-20
SLIDE 20

Wiske’s lower probability that Belgium will win

𝑄( ) = sup

∮︂

ν :

1 ⊗ ν ⊗ν

=

λ + µ ⊗5λ + µ

, µ ⊙ 0, λ ⊙ 0

⨀︁

20

slide-21
SLIDE 21

Wiske’s lower probability that Belgium will win

𝑄( ) = sup

∮︂

ν :

1 ⊗ ν ⊗ν

=

λ + µ ⊗5λ + µ

, µ ⊙ 0, λ ⊙ 0

⨀︁

= sup

{︁5λ + µ

: 1 ⊗ 5λ + µ = λ + µ , µ ⊙ 0, λ ⊙ 0

21

slide-22
SLIDE 22

Wiske’s lower probability that Belgium will win

𝑄( ) = sup

∮︂

ν :

1 ⊗ ν ⊗ν

=

λ + µ ⊗5λ + µ

, µ ⊙ 0, λ ⊙ 0

⨀︁

= sup

{︁5λ + µ

: 1 ⊗ 5λ + µ = λ + µ , µ ⊙ 0, λ ⊙ 0

= sup

5λ + µ : λ = 1 6(1 + µ ⊗ µ ), µ ⊙ 0, λ ⊙ 0

22

slide-23
SLIDE 23

Wiske’s lower probability that Belgium will win

𝑄( ) = sup

∮︂

ν :

1 ⊗ ν ⊗ν

=

λ + µ ⊗5λ + µ

, µ ⊙ 0, λ ⊙ 0

⨀︁

= sup

{︁5λ + µ

: 1 ⊗ 5λ + µ = λ + µ , µ ⊙ 0, λ ⊙ 0

= sup

5λ + µ : λ = 1 6(1 + µ ⊗ µ ), µ ⊙ 0, λ ⊙ 0

= sup

⎭5

6 ⊗ 1 6µ ⊗ 5 6µ : µ ⊙ 0

= 5 6

23

slide-24
SLIDE 24

P( ), P( ) P( ), P( )

5 6, 1

0, 1

6

0, 1

5 4 5, 1

+∞, ⊗∞ + ∞, ⊗∞ 0, 1 0, 1

Agents

Wiske Yoko Tsuno Irrational pool Rational pool

24

slide-25
SLIDE 25

Assessments of lower previsions

◮ Assume a lower prevision 𝑄 with values assessed for a set of

gambles 𝒧

◮ How can we apply the theory we have seen? ◮ Translate the lower prevision 𝑄(𝑔) for a gamble 𝑔 ∈ 𝒧 into a set

¶𝑔 ⊗ 𝑄(𝑔) + 𝜁 : 𝜁 > 0♢

  • f acceptable gambles

◮ The gambles 𝑔 ⊗ 𝑄(𝑔) are called marginal gambles

25

slide-26
SLIDE 26

Expressions for assessments of lower previsions

Avoiding sure loss sup

x∈𝒴 n

∑︂

k=1

(𝑔k(𝑦) ⊗ 𝑄(𝑔k)) ⊙ 0 for all 𝑜 ⊙ 0 and 𝑔k ∈ 𝒧

26

slide-27
SLIDE 27

Expressions for assessments of lower previsions

Avoiding sure loss sup

x∈𝒴 n

∑︂

k=1

(𝑔k(𝑦) ⊗ 𝑄(𝑔k)) ⊙ 0 for all 𝑜 ⊙ 0 and 𝑔k ∈ 𝒧 Coherence sup

x∈𝒴

⎠ n ∑︂

k=1

(𝑔k(𝑦) ⊗ 𝑄(𝑔k)) ⊗ 𝑛(𝑔0 ⊗ 𝑄(𝑔0))

⊙ 0 for all 𝑜, 𝑛 ⊙ 0 and 𝑔k ∈ 𝒧

27

slide-28
SLIDE 28

Expressions for assessments of lower previsions

Avoiding sure loss sup

x∈𝒴 n

∑︂

k=1

(𝑔k(𝑦) ⊗ 𝑄(𝑔k)) ⊙ 0 for all 𝑜 ⊙ 0 and 𝑔k ∈ 𝒧 Coherence sup

x∈𝒴

⎠ n ∑︂

k=1

(𝑔k(𝑦) ⊗ 𝑄(𝑔k)) ⊗ 𝑛(𝑔0 ⊗ 𝑄(𝑔0))

⊙ 0 for all 𝑜, 𝑛 ⊙ 0 and 𝑔k ∈ 𝒧 Natural extension 𝐹(𝑔) = sup

∮︂

inf

x∈𝒴

𝑔(𝑦)⊗

n

∑︂

k=1

λk

(︁𝑔k(𝑦)⊗𝑄(𝑔k) )︁⎨

: 𝑜 ⊙ 0, 𝑔k ∈ 𝒧, λk > 0

⨀︁

28

slide-29
SLIDE 29

Does it really have to be so involved?

29

slide-30
SLIDE 30

Lower previsions on linear spaces

If the lower prevision 𝑄 is defined for all gambles in a linear space ℒ, the coherence criteria simplify: Accepting sure gains 𝑄(𝑔) ⊙ inf 𝑔 for all 𝑔 ∈ ℒ Super-linearity 𝑄(𝑔 + 𝑕) ⊙ 𝑄(𝑔) + 𝑄(𝑕) for all 𝑔, 𝑕 ∈ ℒ Positive homogeneity 𝑄(λ𝑔) = λ𝑄(𝑔) for all 𝑔 ∈ ℒ and λ > 0

30

slide-31
SLIDE 31

Upper previsions on linear spaces

If the upper prevision 𝑄 is defined for all gambles in a linear space ℒ, the coherence criteria simplify: Accepting sure gains 𝑄(𝑔) ⊘ sup 𝑔 for all 𝑔 ∈ ℒ Sub-linearity 𝑄(𝑔 + 𝑕) ⊘ 𝑄(𝑔) + 𝑄(𝑕) for all 𝑔, 𝑕 ∈ ℒ Positive homogeneity 𝑄(λ𝑔) = λ𝑄(𝑔) for all 𝑔 ∈ ℒ and λ > 0

31

slide-32
SLIDE 32

Coherent lower & upper previsions

For a coherent lower prevision 𝑄 and its conjugate upper prevision 𝑄 many useful properties can be derived; we present a few: Upper dominates lower 𝑄(𝑔) ⊙ 𝑄(𝑔) for all 𝑔 ∈ ℒ Constants 𝑄(µ) = µ for all µ ∈ R Constant additivity 𝑄(𝑔 + µ) = 𝑄(𝑔) + µ for all 𝑔 ∈ ℒ and µ ∈ R Gamble dominance if 𝑔 ⊙ 𝑕 + µ then 𝑄(𝑔) ⊙ 𝑄(𝑕) + µ for all 𝑔, 𝑕 ∈ ℒ and µ ∈ R Mixed sub/super-additivity 𝑄(𝑔 + 𝑕) ⊘ 𝑄𝑔 + 𝑄(𝑕) ⊘ 𝑄(𝑔 + 𝑕) for all 𝑔, 𝑕 ∈ ℒ

32

slide-33
SLIDE 33

I heard that imprecise probabilities are just sets of probabilities?

33

slide-34
SLIDE 34

The probability simplex

𝑞 = (𝑞 , 𝑞 )

34

slide-35
SLIDE 35

The probability simplex

(1

2, 1 2)

(1, 0) (0, 1)

35

slide-36
SLIDE 36

The probability simplex

1 𝑄p( ) = 0 ≤ 𝑞 + 1 ≤ 𝑞

36

slide-37
SLIDE 37

The probability simplex

1 𝑄p( ) = 0 ≤ 𝑞 + 1 ≤ 𝑞 𝑄 Y( ) = 4

5

(0, 4

5)

ℳY

37

slide-38
SLIDE 38

The probability simplex

1 𝑄p( ) = 0 ≤ 𝑞 + 1 ≤ 𝑞 𝑄 Y( ) = 4

5

(0, 4

5)

ℳY

CREDAL SET

38

slide-39
SLIDE 39

From lower previsions to credal sets

◮ The prevision of a gamble

is a linear function over the probability simplex

◮ The lower prevision of a gamble

can be seen as bounding the prevision of that gamble so constraining the possible probability mass functions

◮ A lower prevision corresponds to a set of constraints,

defining a credal set (closed convex set) ℳ = ¶𝑞 : 𝑄p(𝑔) ⊙ 𝑄(𝑔) for all 𝑔 ∈ 𝒧♢

◮ All this generalizes to infinite 𝒴

39

slide-40
SLIDE 40

A larger probability simplex

P(enalties)

40

slide-41
SLIDE 41

A larger probability simplex

P(enalties) 𝑄( ) = 𝑄(P) 𝑄( ) = 𝑄(P)

41

slide-42
SLIDE 42

A larger probability simplex

P(enalties) 𝑄( ⊗ P) = 0 𝑄( ⊗ P) = 0 ℳPool

42

slide-43
SLIDE 43

A larger probability simplex

P(enalties) (1, 0, 0) (0, 1, 0) (1

3, 1 3, 1 3)

ℳPool

43

slide-44
SLIDE 44

A larger probability simplex

P(enalties) (1, 0, 0) (0, 1, 0) (1

3, 1 3, 1 3)

ℳPool 𝑄(P) = 1

3

44

slide-45
SLIDE 45

From credal sets to lower previsions

◮ A credal set is a closed convex set of probability mass functions

(or more generally, previsions)

◮ A credal set is determined completely by its set of extreme points ℳ* ◮ A nonempty credal set is equivalent to a coherent lower prevision

Lower envelope theorem 𝑄(𝑔) = min¶𝑄p(𝑔) : 𝑞 ∈ ℳ♢ = min¶𝑄p(𝑔) : 𝑞 ∈ ℳ*♢

45

slide-46
SLIDE 46

Where are the conditional models? We need them to learn!

46

slide-47
SLIDE 47

Basics of conditioning and updating

◮ Conditioning acceptable gambles is done by restriction to the

subspace of gambles that are zero outside the conditioning event

◮ Conditioning lower previsions is a form of natural extension

𝐹(𝑔 ♣ 𝐵) =

∮︂

infx∈A 𝑔(𝑦) if 𝑄(𝐵) = 0, max¶µ ∈ R : 𝑄(1A(𝑔 ⊗ µ))♢ if 𝑄(𝐵) > 0

◮ Conditional credal set =

credal set of conditional probability mass functions ℳ♣𝐵 =

∮︂

whole simplex if ∃𝑞 ∈ ℳ : 𝑄p(𝐵) = 0,

{︁𝑞(≤ ♣ 𝐵) : 𝑞 ∈ ℳ ⟨

if ∀𝑞 ∈ ℳ : 𝑄p(𝐵) > 0

◮ Natural extension often gives vacuous conditionals;

regular extension is a less imprecise updating rule: it removes those 𝑞 such that 𝑄p(𝐵) = 0 from ℳ

47

slide-48
SLIDE 48

Conditioning using natural extension

P(enalties) (1, 0, 0) (0, 1, 0) (1

3, 1 3, 1 3)

ℳPool ℳPool♣¶ , 𝑄♢

48

slide-49
SLIDE 49

Conditioning using regular extension

P(enalties) (0, 1, 0) (1

3, 1 3, 1 3)

ℳ′

Pool

ℳ′

Pool♣¶

, 𝑄♢

49

slide-50
SLIDE 50

What about imprecise probabilities

  • n continuous spaces?

50

slide-51
SLIDE 51

Too few remarks about imprecise probabilities on continuous spaces

◮ Mostly defined using credal sets

  • f parametric distributions

where the parameters vary in a set

◮ Also common:

defined using probability mass assignments to subsets of the space

◮ Examples:

◮ Imprecise Dirichlet model ◮ P-boxes ◮ lower density functions

◮ Calculating lower and upper previsions (natural extension)

can easily become difficult optimization problems

51

slide-52
SLIDE 52

Nothing in this lecture had anything to do with aerospace engineering!

52

slide-53
SLIDE 53

53

slide-54
SLIDE 54

𝑌 ⊗⃗ 1Y cos ψ ⊗⃗ 1Z sin ψ Ω𝑠 + 𝑉r⊤ 𝑉r⊥ φ α 𝜄

54

slide-55
SLIDE 55

𝑎 𝑍 𝑆 ψ ⊗⃗ 1Y cos ψ ⊗⃗ 1Z sin ψ 𝑠 𝑎 ⃗ 1X cos β ⊗⃗ 1Y sin β ℎ0 𝑉∞ 𝑉 ℎ 𝑌 𝑍 ⃗ 𝑉 β

55

slide-56
SLIDE 56

Relationship between turbulence and angle of attack

𝑉r⊥(1 + δ) φ α 𝜄 αδ = α ⊗ α0 = atan ((1 + δ) tan φ0) ⊗ φ0 δ = tan(φ0 + αδ) ⊗ tan φ0 tan φ0

56

slide-57
SLIDE 57

Is this all you have?

57