EE613 Machine Learning for Engineers
NONLINEAR REGRESSION II
Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute
- Dec. 20, 2017
1
NONLINEAR REGRESSION II Sylvain Calinon Robot Learning & - - PowerPoint PPT Presentation
EE613 Machine Learning for Engineers NONLINEAR REGRESSION II Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Dec. 20, 2017 1 First, lets recap some useful properties and approaches presented in previous
1
2
GMR can cover a large spectrum
Both and can be multidimensional encoded in Gaussian mixture model (GMM) retrieved by Gaussian mixture regression (GMR)
Nadaraya-Watson kernel regression Least squares linear regression
3
Nonlinear regression I
4
Linear regression Nonlinear regression I
5
Linear regression
6
HMMs
7
[C. K. I. Williams and C. E. Rasmussen. Gaussian processes for regression. In Advances in Neural Information Processing Systems (NIPS), pages 514–520, 1996] [S. Roberts, M. Osborne, M. Ebden, S. Reece, N. Gibson, and S. Aigrain. Gaussian processes for time-series modelling. Philosophical Trans. of the Royal Society A, 371(1984):1–25, 2012]
8
marginal distributions P(y1) and P(y2) that are unidimensional.
conditional distribution.
variable based on the observation of the other.
9
number of variables.
imagined as a single point sampled from a multivariate Gaussian distribution.
10
11
Note that with GPs, we do not build joint distributions on {x1,x2,…xN}!
x can be multivariate
12
equivalent to a distribution over a function space y = f(x).
a function, it turns out that we only need to be able to define a distribution over the function values at a finite, but arbitrary, set of points.
y = {y1,..., yN} should be imagined as a single point sampled from an N-variate Gaussian.
13
benefiting from properties inherited from multivariate normal distributions.
the distributions of various derived quantities can be
estimating the average using sample values at a small set of times. → Usually more powerful than just selecting a model type, such as selecting the degree of a polynomial to fit a dataset, as we have seen in the lecture about linear regression.
14
Polynomial fitting with least squares and nullspace optimization
an example of parametric modeling technique, where we provided the degree of the polynomial.
generate multiple solutions of a fitting problem, thus obtaining a family
a valid explanation for the observed data. Bayesian modeling will be the central tool to working with such distribution over curves.
15
covariance over an arbitrarily large set of variables can be defined through the covariance kernel function k(xi, xj), providing the covariance elements between any two sample locations xi and xj. If xN is closer to x3 than x1, we also expect yN to be closer to y3 than y1
16
17
process that is smooth, that is continuous, that has typical amplitude,
(e.g., within a typical dynamic range), etc.
→ We will work mathematically with the infinite space of all functions
that have these characteristics.
these hyperparameters govern characteristics that are more generic such as the scale of a distribution rather than acting explicitly on the structure or functional form of the signals.
18
for points xi and xj that are similar, we expect the corresponding output
basic aspects that can be defined through the covariance function k are the process stationarity, isotropy, smoothness or periodicity.
that past observations can be informative about current data as a function of how long ago they were observed.
Euclidean distance |xi - xj|.
directions between xi and xj.
19
20
21
22
23
Monte-Carlo Grid-based 2nd order model
Several approaches exist to estimate the hyperparameters of the covariance function: Maximum Likelihood Estimation (MLE), cross-validation (CV), Bayesian approaches involving sampling algorithms such as MCMC, etc.
For example, given an expression for the log marginal likelihood and its derivative, we can estimate the kernel parameters using standard gradient-based optimizer. Note that since the
24
1 100 1 100 1 100 1 100
25
[C. K. I. Williams and C. E. Rasmussen. Gaussian processes for regression. In Advances in Neural Information Processing Systems (NIPS), pages 514–520, 1996] [S. Roberts, M. Osborne, M. Ebden, S. Reece, N. Gibson, and S. Aigrain. Gaussian processes for time-series modelling. Philosophical Trans. of the Royal Society A, 371(1984):1–25, 2012]
26
27
Kriging can also be understood as a form of Bayesian inference: It starts with a prior distribution over functions, that takes the form of a Gaussian process. Namely, N samples from a function will be normally distributed, where the covariance between any two samples is the covariance function (or kernel) of the Gaussian process evaluated at the spatial location of two points.
28
method of interpolation for which the interpolated values are modeled by a Gaussian process governed by prior covariances.
unbiased prediction of the intermediate values.
Master's thesis of Danie G. Krige, a South African statistician and mining engineer.
(RKHS), with the reproducing kernel given by the covariance function. Interpretation: the spline is motivated by a minimum norm interpolation based on a Hilbert space structure, while standard kriging is motivated by an expected squared prediction error based on a stochastic model.
29
Interpretation:
A set of values y is first observed, each value associated with a spatial/temporal location x. Now, a new value y* can be predicted at any new spatial/temporal location x*, by combining the Gaussian prior with a Gaussian likelihood function for each of the observed values. The resulting posterior distribution is also Gaussian, with a mean and covariance that can be simply computed from the observed values, their variance, and the kernel matrix derived from the prior.
30
31
32
33
34
35
36
37
38
39
40
The Wiener process is a simple continuous-time stochastic process often put in connection to the Brownian motion.
41
42
43
probabilistic models.
k(xi, xj) =P(xi) P(xj), which can be interpreted as an inner product in the
probabilities to belong to the model.
discriminative setting, thus combining the performance of both generative and discriminative models.
the capability of handling missing data or partial sequences of various lengths (e.g., with HMM).
44
covariance function can be defined as a linear combination of
different insights about the dataset.
kernel parameters (also known as multiple kernel learning). The idea is to define the kernel as a weighted sum of basis kernels, and then to optimize the weights instead of the kernel parameters.
Dictionary of basis kernel functions
45
that the computational complexity of prediction is O(N3)
→ not feasible for large data sets!
Sparse Gaussian processes circumvent this issue by building a representative set for the given process y = f(x).
density function over positive definite matrices. The generalised Wishart process (GWP) is a collection of positive semi-definite random matrices indexed by any arbitrary dependent variable. It can for example be used to model time varying covariance matrices Σ(t). A draw from a Wishart process is then a collection of matrices indexed by time, similarly to a draw from a GP being a collection of function values indexed by time. Similarly to GP, GWP can capture diverse covariance structures and it can easily handle missing data.
46
GPs to find a lower dimensional non-linear embedding of high dimensional data.
probabilistically, the latent variables are marginalized and the parameters are obtained by maximizing the likelihood.
mapping (in the form of a GP). However, in GPLVM, the mapping is from the embedded (latent) space to the data space, whereas in kernel PCA, the mapping is in the opposite direction.
but has been extended to construct various forms of shared manifold models between two observation spaces.
GPR
C.K.I. Williams and C.E. Rasmussen. Gaussian processes for regression. In Advances in Neural Information Processing Systems (NIPS), pages 514–520, 1996 C.E. Rasmussen and C.K.I. Williams. Gaussian processes for machine learning. MIT Press, Cambridge, MA, USA, 2006
processes for time-series modelling. Philosophical Trans. of the Royal Society A, 371(1984):1–25, 2012 GPLVM
process latent variable models. Journal of machine learning research, 6:1783-1816, 2005 GWP A.G. Wilson and Z. Ghahramani. Generalised Wishart processes. Uncertainty in Artificial Intelligence, 2011
47