A game-theoretic ergodic theorem for imprecise Markov chains Gert - - PowerPoint PPT Presentation

a game theoretic ergodic theorem for imprecise markov
SMART_READER_LITE
LIVE PREVIEW

A game-theoretic ergodic theorem for imprecise Markov chains Gert - - PowerPoint PPT Presentation

A game-theoretic ergodic theorem for imprecise Markov chains Gert de Cooman Ghent University, SYSTeMS gert.decooman@UGent.be http://users.UGent.be/gdcooma gertekoo.wordpress.com GTP 2014 CIMAT, Guanajuato 13 November 2014 My boon


slide-1
SLIDE 1

A game-theoretic ergodic theorem for imprecise Markov chains

Gert de Cooman

Ghent University, SYSTeMS gert.decooman@UGent.be http://users.UGent.be/˜gdcooma gertekoo.wordpress.com

GTP 2014 CIMAT, Guanajuato 13 November 2014

slide-2
SLIDE 2

My boon companions

FILIP HERMANS ENRIQUE MIRANDA JASPER DE BOCK

slide-3
SLIDE 3

Jean Ville and martingales

slide-4
SLIDE 4

The original definition of a martingale

Étude critique de la notion de collectif, 1939, p. 83

slide-5
SLIDE 5

In a (perhaps) more modern notation

Ville’s definition of a martingale

A martingale s is a sequence of real functions so, s1(X1), s2(X1,X2), . . . such that

1 so = 1; 2 sn(X1,...,Xn) ≥ 0 for all n ∈ N; 3 E(sn+1(x1,...,xn,Xn+1)|x1,...,xn) = sn(x1,...,xn) for all n ∈ N0 and all

x1,...,xn. It represents the outcome of a fair betting scheme, without borrowing (or bankruptcy).

slide-6
SLIDE 6

Ville’s theorem

The collection of all (locally defined!) martingales determines the probability P on the sample space Ω: P(A) = sup{λ ∈ R: s martingale and limsup

n→+∞

λsn(X1,...,Xn) ≤ IA} = inf{λ ∈ R: s martingale and liminf

n→+∞ λsn(X1,...,Xn) ≥ IA}

slide-7
SLIDE 7

Ville’s theorem

The collection of all (locally defined!) martingales determines the probability P on the sample space Ω: P(A) = sup{λ ∈ R: s martingale and limsup

n→+∞

λsn(X1,...,Xn) ≤ IA} = inf{λ ∈ R: s martingale and liminf

n→+∞ λsn(X1,...,Xn) ≥ IA}

Turning things around

Ville’s theorem suggests that we could take a convex set of martingales as a primitive notion, and probabilities and expectations as derived notions. That we need an convex set of them, elucidates that martingales are examples of partial probability assessments.

slide-8
SLIDE 8

Imprecise probabilities: dealing with partial probability assessments

slide-9
SLIDE 9

Partial probability assessments

lower and/or upper bounds for – the probabilities of a number of events, – the expectations of a number of random variables

slide-10
SLIDE 10

Partial probability assessments

lower and/or upper bounds for – the probabilities of a number of events, – the expectations of a number of random variables

Imprecise probability models

A partial assessment generally does not determine a probability measure uniquely, only a convex closed set of them.

slide-11
SLIDE 11

Partial probability assessments

lower and/or upper bounds for – the probabilities of a number of events, – the expectations of a number of random variables

Imprecise probability models

A partial assessment generally does not determine a probability measure uniquely, only a convex closed set of them.

IP Theory

systematic way of dealing with, representing, and making conservative inferences based on partial probability assessments

slide-12
SLIDE 12

Lower and upper expectations

slide-13
SLIDE 13

Lower and upper expectations

A Subject is uncertain about the value that a variable X assumes in X .

Gambles:

A gamble f : X → R is an uncertain reward whose value is f(X). G (X ) denotes the set of all gambles on X .

slide-14
SLIDE 14

Lower and upper expectations

A Subject is uncertain about the value that a variable X assumes in X .

Gambles:

A gamble f : X → R is an uncertain reward whose value is f(X). G (X ) denotes the set of all gambles on X .

Lower and upper expectations:

A lower expectation is a real functional that satisfies:

  • E1. E(f) ≥ inf f

[bounds]

  • E2. E(f +g) ≥ E( f)+E(g)

[superadditivity]

  • E3. E(λ f) = λE(f) for all real λ ≥ 0

[non-negative homogeneity] E( f) := −E(−f) defines the conjugate upper expectation.

slide-15
SLIDE 15

Sub- and supermartingales

slide-16
SLIDE 16

An event tree and its situations

Situations are nodes in the event tree, and the sample space Ω is the set of all terminal situations: t ω initial terminal non-terminal

slide-17
SLIDE 17

Events

An event A is a subset of the sample space Ω: s Γ(s) := {ω ∈ Ω: s ⊑ ω}

slide-18
SLIDE 18

Local, or immediate prediction, models

In each non-terminal situation s, Subject has a belief model Q(·|s). s c2 c1 t Q(·|s) on G (D(s)) Q(·|t) on G (D(t)) D(s) = {c1,c2} is the set of daughters of s.

slide-19
SLIDE 19

Sub- and supermartingales

We can use the local models Q(·|s) to define sub- and supermartingales:

A submartingale M

is a real process such that in all non-terminal situations s: Q(M (s·)|s) ≥ M (s).

A supermartingale M

is a real process such that in all non-terminal situations s: Q(M (s·)|s) ≤ M (s).

slide-20
SLIDE 20

Lower and upper expectations

The most conservative lower and upper expectations on G (Ω) that coincide with the local models and satisfy a number of additional continuity criteria (cut conglomerability and cut continuity):

Conditional lower expectations:

E( f|s) := sup{M (s): limsupM ≤ f on Γ(s)}

Conditional upper expectations:

E(f|s) := inf{M (s): liminfM ≥ f on Γ(s)}

slide-21
SLIDE 21

Test supermartingales and strictly null events

A test supermartingale T

is a non-negative supermartingale with T () = 1. (Very close to Ville’s definition of a martingale.)

An event A is strictly null

if there is some test supermartingale T that converges to +∞ on A: limT (ω) = lim

n→∞T (ωn) = +∞ for all ω ∈ A.

If A is strictly null then P(A) = E(IA) = inf{M (): liminfM ≥ IA} = 0.

slide-22
SLIDE 22

A few basic limit results

Supermartingale convergence theorem [Shafer and Vovk, 2001]

A supermartingale M that is bounded below converges strictly almost surely to a real number: liminfM (ω) = limsupM (ω) ∈ R strictly almost surely.

slide-23
SLIDE 23

A few basic limit results

Strong law of large numbers for submartingale differences [De Cooman and De Bock, 2013]

Consider any submartingale M such that its difference process ∆M (s) = M (s·)−M (s) ∈ G (D(s)) for all non-terminal s is uniformly bounded. Then liminfM ≥ 0 strictly almost surely, where M (ωn) = 1 nM (ωn) for all ω ∈ Ω and n ∈ N

slide-24
SLIDE 24

A few basic limit results

Lévy’s zero–one law [Shafer, Vovk and Takemura, 2012]

For any bounded real gamble f on Ω: limsup

n→+∞

E( f|ωn) ≤ f(ω) ≤ liminf

n→+∞ E( f|ωn) strictly almost surely.

slide-25
SLIDE 25

Imprecise Markov chains

slide-26
SLIDE 26

A simple discrete-time finite-state stochastic process

a (a,a) (a,a,a) (a,a,b) (a,b) (a,b,a) (a,b,b) b (b,a) (b,a,a) (b,a,b) (b,b) (b,b,a) (b,b,b) Q(·|) Q(·|a) Q(·|b) Q(·|a,a) Q(·|b,b) Q(·|b,a) Q(·|a,b)

slide-27
SLIDE 27

An imprecise IID model

a (a,a) (a,a,a) (a,a,b) (a,b) (a,b,a) (a,b,b) b (b,a) (b,a,a) (b,a,b) (b,b) (b,b,a) (b,b,b) Q(·|) Q(·|) Q(·|) Q(·|) Q(·|) Q(·|) Q(·|)

slide-28
SLIDE 28

An imprecise Markov chain

a (a,a) (a,a,a) (a,a,b) (a,b) (a,b,a) (a,b,b) b (b,a) (b,a,a) (b,a,b) (b,b) (b,b,a) (b,b,b) Q(·|) Q(·|a) Q(·|b) Q(·|a) Q(·|b) Q(·|a) Q(·|b)

slide-29
SLIDE 29

Stationarity and ergodicity

The lower expectation En for the state Xn at time n: En( f) = E(f(Xn)) The imprecise Markov chain is Perron–Frobenius-like if for all marginal models E1 and all f: En( f) → E∞( f). and if E1 = E∞ then En = E∞, and the imprecise Markov chain is stationary. In any Perron–Frobenius-like imprecise Markov chain: lim

n→+∞

1 n

n

k=1

En(f) = E∞(f) and E∞( f) ≤ liminf

n→+∞

1 n

n

k=1

f(Xk) ≤ limsup

n→+∞

1 n

n

k=1

f(Xk) ≤ E∞( f) str. almost surely.

slide-30
SLIDE 30

A more general ergodic theorem: the basics

Introduce a shift operator: θω = θ(x1,x2,x3,...) := (x2,x3,x4,...) for all ω ∈ Ω, and for any gamble f on Ω a shifted gamble θ f := f ◦θ: (θ f)(ω) := f(θω) for all ω ∈ Ω. For any bounded gamble f on Ω, the bounded gambles: g = liminf

n→+∞

1 n

n−1

k=0

θ k f and g = limsup

n→+∞

1 n

n−1

k=0

θ k f are shift-invariant: θg = g.

slide-31
SLIDE 31

A more general ergodic theorem: use Lévy’s zero–one law

In any Perron–Frobenius-like imprecise Markov chain, for any shift-invariant gamble g = θg on Ω: lim

n→+∞E(g|ωn) = E∞(g) and

lim

n→+∞E(g|ωn) = E∞(g)

and therefore E∞(g) ≤ g ≤ E∞(g) strictly almost surely.

slide-32
SLIDE 32

New books

slide-33
SLIDE 33
slide-34
SLIDE 34