Active Learning and
Optimized Information Gathering
Lecture 6 – Gaussian Process Optimization
CS 101.2 Andreas Krause
Announcements Homework 1: out tomorrow Due Thu Jan 29 Project - - PowerPoint PPT Presentation
Active Learning and Optimized Information Gathering Lecture 6 Gaussian Process Optimization CS 101.2 Andreas Krause Announcements Homework 1: out tomorrow Due Thu Jan 29 Project Proposal due Tue Jan 27 Office hours Come to office
CS 101.2 Andreas Krause
2
Due Thu Jan 29
Proposal due Tue Jan 27
Come to office hours before your presentation! Andreas: Friday 1:30-3pm, 260 Jorgensen Ryan: Wednesday 4:00-6:00pm, 109 Moore
3
1.
2.
3.
4
εn greedy, UCB1 have regret O(log(T) K)
Have to make assumptions!
5
6
7
“strong” assumptions Regret O(T2/3 n)
“weak” assumptions Regret O(C(n) Tn/(n+1)) Curse of dimensionality!
8
9
10
11
12
13
14
15
0.5 1 1.5 2
1 2 0.1 0.2 0.3 0.4
0.5 1 1.5 2
1 2 0.05 0.1 0.15 0.2
16
17
18
19
0.5 1 1.5 2
1 2 0.1 0.2 0.3 0.4
20
21
(infinite) set of random variables, indexed by some set V i.e., for each x∈ V there’s a RV Yx Let A ⊆ V, |A|= {x1,…,xk} < ∞ Then YA ~ N(µA,ΣAA) where K: V× V → R is called kernel (covariance) function µ: V → R is called mean function
22
23
24
25
0.2 0.4 0.6 0.8 1
1 2
0.2 0.4 0.6 0.8 1
0.5 1 1.5 2 2.5 3
26
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0.5 1 1.5
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0.5 1 1.5 2 2.5
27
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6
28
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0.5 1 1.5 2 2.5
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0.5 1 1.5 2 2.5
29
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
1 2 3 4
30
31
32
“Nonparametric regression” Can fit any data set!! ☺
33
34
Can do gradient descent, conjugate gradient, etc.
35
36
(infinite) set of random variables, indexed by some set V i.e., for each x∈ V there’s a RV Yx Let A ⊆ V, |A|= {x1,…,xk} < ∞ Then YA ~ N(µA,ΣAA) where K: V× V → R is called kernel (covariance) function µ: V → R is called mean function
37
Text (strings) Graphs Sets …
38
39
Nonparametric generalization of logistic regression Like SVMs (but give confidence on predicted labels!)
Model count data over space, …
40
41
42
43
44
Nonparametric generalization of linear regression Flexible ways to encode prior assumptions about mean payoffs
Combination of regression and optimization Use confidence bands for selecting samples