Machine Learning - MT 2017
- 4. Maximum Likelihood
Machine Learning - MT 2017 4. Maximum Likelihood Varun Kanade - - PowerPoint PPT Presentation
Machine Learning - MT 2017 4. Maximum Likelihood Varun Kanade University of Oxford October 16, 2017 Outline Probabilistic Perspective of Machine Learning Probabilistic Formulation of the Linear Model Maximum Likelihood Estimate
◮ Probabilistic Formulation of the Linear Model ◮ Maximum Likelihood Estimate ◮ Relation to the Least Squares Estimate
1
2σ2
−∞
−∞
−∞
2
σ
−∞
2 dt
3
1) and X2 ∼ N(µ2, σ2 2) are independent
1
2
1σ2 2)1/2 · exp
1
2
1
2
4
5
cov(x) = E
= var(X1) cov(X1, X2) · · · cov(X1, XD) cov(X2, X1) var(X2) · · · cov(X2, XD) . . . . . . ... . . . cov(XD, X1) cov(XD, X2) · · · var(XD) .
7
8
i=1.
9
i=1.
N
N
N
10
N
N
11
12
ML = 1
13
i=1, we can obtain the MLE wML and σML.
ML) 14
◮ Linear model: y = w · x + ǫ ◮ Explicitly model ǫ ∼ N(0, σ2)
◮ Every w, σ defines a probability distribution over observed data ◮ Pick w and σ that maximise the likelihood of observing the data
◮ As in the previous lecture, we have closed form expressions ◮ Algorithm simply implements elementary matrix operations
15
16
i=1, let us express the likelihood of observing the data
N
N
N
17
◮ Beyond Linearity: Basis Expansion, Kernels ◮ Regularization: Ridge Regression, LASSO ◮ Overfitting, Model Complexity, Cross Validation
18