1
CS/CNS/EE 253
Online Algorithms: Learning & Optimization with No Regret. - - PowerPoint PPT Presentation
Online Algorithms: Learning & Optimization with No Regret. CS/CNS/EE 253 Daniel Golovin 1 CS/CNS/EE 253 The Setup Optimization: Model the problem (objective, constraints) Pick best decision from a feasible set. Learning:
1
CS/CNS/EE 253
2
CS/CNS/EE 253
3
CS/CNS/EE 253
4
CS/CNS/EE 253
e2xt ct(e)
Dealing with Limited Feedback: later in the course.
5
CS/CNS/EE 253
6
CS/CNS/EE 253
7
CS/CNS/EE 253
i wit.
8
CS/CNS/EE 253
[Littlestone & Warmuth '89]
i wit.
4Wt
9
CS/CNS/EE 253
Experts\Time 1
10
CS/CNS/EE 253
T
t=1
T
t=1
t=1 ct(x)
“Maybe all one can do is hope to end up with the right regrets.” – Arthur Miller
11
CS/CNS/EE 253
j
* Pedantic note: Hedge is often called “Randomized Weighted Majority”, and abbreviated “WMR”, though WMR was published in the context of binary classification, unlike Hedge.
[Freund & Schapire '97]
12
CS/CNS/EE 253
j
13
CS/CNS/EE 253
t=1
t=1 ct(x(ei; t)).
14
CS/CNS/EE 253
[def of pt(i)] [Bernoulli's ineq]
If x > ¡1; r 2 (0; 1) then (1 + x)r · 1 + rx
i
i
i
i wit. Then W0 = n and WT +1 ¸ (1 ¡ ²)OPT.
15
CS/CNS/EE 253
T
t=1
T
t=1
t=1
16
CS/CNS/EE 253
t=1 ct(x(ei; t)).
2
17
CS/CNS/EE 253
18
CS/CNS/EE 253
– All distributions over experts – All sequences of experts that have K transitions [Auer et al '02] – Various classes of functions of input features [Blum & Mansour '05]
– Arbitrary convex set of experts, metric space of
[Zinkevich '03, Kleinberg et al '08]
– All policies of a K-state initially unknown Markov
– Arbitrary sets of strategies in with linear costs that
19
CS/CNS/EE 253
– get low regret on mondays, rainy days, etc.
– if rule “if(P) then predict Q” is right 90% of the time it
applies, be right 89% of the time P applies.
– If you played x1, ..., xT then have no regret against
g(x1), ..., g(xT) for every g:X→X
20
CS/CNS/EE 253
right 89% of the time P applies. Get this for every rule simultaneously.
inputs, some good, some lousy, and combine them into a great classifier.
– if (“physics” in D) then classify D as “science”.
[Freund et al '97, Blum '97, Blum & Mansour '05]
21
CS/CNS/EE 253
enrolled in?
– if(ML-101, CS-201 in C) then CS – if(ML-101, Stats-201 in C) then Stats
and Stats-201?
22
CS/CNS/EE 253
e;f = 1 for all e 2 E; f 2 F.
e = P f f(t)wt e;f.
e wt e.
e = wt e=W t.
[Algorithm from Blum & Mansour '05]
23
CS/CNS/EE 253
[Algorithm from Blum & Mansour '05]
Ensures total sum of weights can never increase.
e;f
e;f · nm for all t
e;f =
t¸0
P
t¸0[f(t)(ct(e)¡¯E[ct(et)])]
24
CS/CNS/EE 253
t=1 f(t) ¢ ct(e)
t=1 f(t) ¢ ct(et)