Shrinkage
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Maximilian Kasy
Department of Economics, Harvard University
1 / 37
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel - - PowerPoint PPT Presentation
Shrinkage Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Shrinkage Agenda 6 equivalent representations of the posterior mean
Shrinkage
Department of Economics, Harvard University
1 / 37
Shrinkage
2 / 37
Shrinkage
3 / 37
Shrinkage Normal posterior means – equivalent representations
i
4 / 37
Shrinkage Normal posterior means – equivalent representations
5 / 37
Shrinkage Normal posterior means – equivalent representations
t(·)
6 / 37
Shrinkage Normal posterior means – equivalent representations
t
7 / 37
Shrinkage Normal posterior means – equivalent representations
8 / 37
Shrinkage Normal posterior means – equivalent representations
t n
i=1
9 / 37
Shrinkage Normal posterior means – equivalent representations
10 / 37
Shrinkage Normal posterior means – equivalent representations
11 / 37
Shrinkage Normal posterior means – equivalent representations
◮ First order condition for quadratic loss function, ◮ pull derivative inside, ◮ and switch order of integration.
◮ X and θ are jointly Normal, ◮ conditional expectations for multivariate Normals are linear.
◮ Posterior is symmetric unimodal ⇒ posterior mean is posterior
◮ Posterior mode = maximizer of posterior log-likelihood =
◮ since denominator fX does not depend on θ.
12 / 37
Shrinkage Normal posterior means – equivalent representations
◮ Any penalty of the form
◮ corresponds to the log of a Normal prior
◮ Change of coordinates turns
◮ Diagonality implies
13 / 37
Shrinkage Gaussian process regression
14 / 37
Shrinkage Gaussian process regression
15 / 37
Shrinkage Gaussian process regression
16 / 37
Shrinkage Gaussian process regression
i
i
i
i Xi +σ 2Ω−1
i Yi.
17 / 37
Shrinkage Gaussian process regression
18 / 37
Shrinkage Gaussian process regression
19 / 37
Shrinkage Gaussian process regression
20 / 37
Shrinkage Gaussian process regression
21 / 37
Shrinkage Gaussian process regression
◮ c(x) is the n vector with entries C(x,Xi), ◮ and C is the n × n matrix with entries Ci,j = C(Xi,Xj).
◮ is a linear combination of the functions C(·,Xi) ◮ with weights
22 / 37
Shrinkage Gaussian process regression
2l x − x′2
2|det(Cη +σ 2I)|− 1 2Y ′(Cη +σ 2I)−1Y
23 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
f
i
f
i
24 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
t n
i=1
C,
C = t′C−1t.
◮ X|θ ∼ N(θ,Ik), ◮ θ ∼ N(0,C).
25 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
26 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
27 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
28 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
i,j
29 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
i,j
i,j
i
j
i
i
j
i,j
30 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
◮ For each regression penalty, ◮ when function evaluation is continuous w.r.t. the penalty norm ◮ there exists a corresponding prior.
◮ The solution to the penalized regression problem ◮ is the posterior mean for this prior.
31 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
f
i
C.
◮ Write
◮ where ρ is orthogonal to C(Xi,·) for all i. ◮ Show that ρ = 0. ◮ Solve the resulting least squares problem in a1,...,an.
32 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
i
C
i
C
i
j
i
C
i
j
i,j
C
C
33 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
2
2
34 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
35 / 37
Shrinkage Splines and Reproducing Kernel Hilbert Spaces
x (y)g′′(y)dy
x
0 (x − y)g′′(y)dy
y=0
36 / 37
Shrinkage References
37 / 37