Minimax strategy for prediction with expert advice under stochastic - - PowerPoint PPT Presentation

minimax strategy for prediction with expert advice under
SMART_READER_LITE
LIVE PREVIEW

Minimax strategy for prediction with expert advice under stochastic - - PowerPoint PPT Presentation

Minimax strategy for prediction with expert advice under stochastic assumptions Wojciech Kot lowski Pozna n University of Technology, Poland Learning Faster from Easy Data II NIPS 2015 Workshop 1 / 3 Prediction with expert advice In


slide-1
SLIDE 1

Minimax strategy for prediction with expert advice under stochastic assumptions

Wojciech Kot lowski

Pozna´ n University of Technology, Poland

Learning Faster from Easy Data II NIPS 2015 Workshop

1 / 3

slide-2
SLIDE 2

Prediction with expert advice

In trials t = 1, 2, . . . , T: – Algorithm predicts with wt ∈ ∆K. – Loss vector ℓt ∈ [0, 1]K is revealed. – Algorithm incurs loss wt · ℓt. Regret of a strategy ω = (w1, . . . , wT ): R =

  • t

wt · ℓt − min

k

  • t≤T

ℓt,k

LT,k

Goal: find ω minimizing the worst-case regret over all sequences.

2 / 3

slide-3
SLIDE 3

Prediction with expert advice

In trials t = 1, 2, . . . , T: – Algorithm predicts with wt ∈ ∆K. – Loss vector ℓt ∈ [0, 1]K is revealed. – Algorithm incurs loss wt · ℓt. Regret of a strategy ω = (w1, . . . , wT ): R =

  • t

wt · ℓt − min

k

  • t≤T

ℓt,k

LT,k

Goal: find ω minimizing the worst-case regret over all sequences. Too pessimistic: minimax ω has the same regret on all sequences!

2 / 3

slide-4
SLIDE 4

Prediction with expert advice

In trials t = 1, 2, . . . , T: – Algorithm predicts with wt ∈ ∆K. – Loss vector ℓt ∈ [0, 1]K is revealed. – Algorithm incurs loss wt · ℓt. Regret of a strategy ω = (w1, . . . , wT ): R =

  • t

wt · ℓt − min

k

  • t≤T

ℓt,k

LT,k

Goal: find ω minimizing the worst-case regret over all sequences. Too pessimistic: minimax ω has the same regret on all sequences! Drop the minimax principle? Drop the worst-case assumptions?

2 / 3

slide-5
SLIDE 5

Prediction with expert advice

In trials t = 1, 2, . . . , T: – Algorithm predicts with wt ∈ ∆K. – Loss vector ℓt ∈ [0, 1]K is revealed. – Algorithm incurs loss wt · ℓt. Regret of a strategy ω = (w1, . . . , wT ): R =

  • t

wt · ℓt − min

k

  • t≤T

ℓt,k

LT,k

Goal: find ω minimizing the worst-case regret over all sequences. Too pessimistic: minimax ω has the same regret on all sequences! Drop the minimax principle? Drop the worst-case assumptions?

2 / 3

slide-6
SLIDE 6

Stochastic setting: optimal strategy for “easy” data

Assumption: Each expert k = 1, . . . , K generates losses i.i.d. from a fixed distribution Pk. Goal: Find the minimax strategy ω w.r.t. all choices of distributions P = (P1, . . . , Pk).

3 / 3

slide-7
SLIDE 7

Stochastic setting: optimal strategy for “easy” data

Assumption: Each expert k = 1, . . . , K generates losses i.i.d. from a fixed distribution Pk. Goal: Find the minimax strategy ω w.r.t. all choices of distributions P = (P1, . . . , Pk). Minimax in terms of what?

3 / 3

slide-8
SLIDE 8

Stochastic setting: optimal strategy for “easy” data

Assumption: Each expert k = 1, . . . , K generates losses i.i.d. from a fixed distribution Pk. Goal: Find the minimax strategy ω w.r.t. all choices of distributions P = (P1, . . . , Pk). Minimax in terms of what? Expected regret: Reg(ω, P) = E

t wt · ℓt − mink LT,k

  • Redundancy:

Red(ω, P) = E

t wt · ℓt

  • − mink E [LT,k]

Excess risk: Risk(ω, P) = E [wT · ℓT ] − mink E [ℓT,k]

3 / 3

slide-9
SLIDE 9

Stochastic setting: optimal strategy for “easy” data

Assumption: Each expert k = 1, . . . , K generates losses i.i.d. from a fixed distribution Pk. Goal: Find the minimax strategy ω w.r.t. all choices of distributions P = (P1, . . . , Pk). Minimax in terms of what? Expected regret: Reg(ω, P) = E

t wt · ℓt − mink LT,k

  • Redundancy:

Red(ω, P) = E

t wt · ℓt

  • − mink E [LT,k]

Excess risk: Risk(ω, P) = E [wT · ℓT ] − mink E [ℓT,k] We give a strategy ω∗, which is minimax with respect to all three measures simultaneously: sup

P

R(ω∗, P) = inf

ω sup P

R(ω, P), where R is either Reg, Red, or Risk.

3 / 3