Follow the leader if you can, Hedge if you must
Tim van Erven
NIPS, 2013 Joint work with: Steven de Rooij Peter Grünwald Wouter Koolen
Follow the leader if you can, Hedge if you must Tim van Erven - - PowerPoint PPT Presentation
Follow the leader if you can, Hedge if you must Tim van Erven NIPS, 2013 Joint work with: Steven de Rooij Peter Gr nwald Wouter Koolen Outline Follow-the-Leader: works well for `easy' data : few leader changes, i.i.d. but not
Tim van Erven
NIPS, 2013 Joint work with: Steven de Rooij Peter Grünwald Wouter Koolen
– works well for `easy' data: few leader changes, i.i.d. – but not robust to worst-case data
– robust, but does not exploit easy data
– robust against worst case + can exploit i.i.d. data – but do not exploit few leader changes in general
expert on average
– online convex optimization – predicting electricity consumption – predicting air pollution levels – spam detection – ...
Goal: minimize regret where
Loss of the best expert
predicted best in the past:
where
– If the leader does not change, our loss is the
same as the loss of the leader, so the regret stays the same
– If the leader does change, our regret
increases at most by 1 (range of losses)
changes only finitely many times w.h.p.
losses
from uniform distribution as a regularizer
(more conservative learning)
– approximate our loss: – by the mix loss: – and bound the approximation error:
Regret
Balances the two terms
i.i.d. losses
– = conservative learning – In practice, better when learning rate does
not go to 0 with ! [DGGS, 2013]
– Lost advantages of FTL!
– robust against worst-case losses; but – if the data are `easy', we should learn faster!
– [CBMS, 2007] and AdaHedge: – Related bound by [HK, 2008]
Regret variance of
– [CBMS, 2007] and AdaHedge: – Related bound by [HK, 2008]
Regret variance of
expert: Regret variance of
Recover FTL benefits for i.i.d. data
Regret
balancing
Regret
balancing
Regret
NB Bernstein's bound is pretty sharp, so in practice CBMS ≈ AdaHedge up to constants. balancing
with
Lemma [KV, 2005]: If , then
data with a small number of leader changes
– [CBMS, 2007] and AdaHedge: – Related bound by [HK, 2008]
– “Follow the leader if you can, Hedge if you must”
– Regret best of AdaHedge and FTL
Regret
– Flip: Tune
like FTL
– Flop: Tune
like AdaHedge
Regret FTL Regret AdaHedge Regret Bound
– Flip: Tune
like FTL
– Flop: Tune
like AdaHedge
regimes
errors balanced: Regret
FTL Bound AdaHedge Bound
AdaHedge does not
– works well for `easy' data: i.i.d., few leader changes – but not robust to worst-case data
– robust against worst case + can exploit i.i.d. data – but do not exploit few leader changes in general
– “Follow the leader if you can, Hedge if you must”
– Regret best of AdaHedge and FTL
and FTL?
chosen in hindsight?
expert advice. Machine Learning, 66(2/3):321–352, 2007.
aggregating specialized experts. Machine Learning, 90(2):231-260, 2013.
You Must. Accepted by the Journal of Machine Learning Research, 2013.
translating and rescaling the losses
– Extension so this is not necessary.
Important when range of losses is unknown!
– Invariant under rescaling and translation of
losses, so get this for free.
Regret variance of
(like FTL) then we could be wrong all the time: Regret:
loss 1. Then