SLIDE 1
Estimating Gaussian Mixture Models from Data with Missing Features - - PowerPoint PPT Presentation
Estimating Gaussian Mixture Models from Data with Missing Features - - PowerPoint PPT Presentation
Estimating Gaussian Mixture Models from Data with Missing Features by Daniel McMichael CSSIP Missing Data In classification we frequently see to classify objects using vectors of measured features. Sometimes these features are missing: [ 0
SLIDE 2
SLIDE 3
Gaussian Mixture Models (GMMs)
A probability density model (a weighted sum of Gaussians):
p(y jf i ;- i
- i
- i exp
- i
- 1
- i
- i
(1) : Qualities of GMMs:
Can model any density (given enough components) Can be applied to classification Widely used “Easy” to analyse SLIDE 4
Heteroscedastic GMMs (HGMMs)
Conventionally, GMMs are homoscedastic: all data are modelled with the same Gaussian distribution:
p(y j j i ;- i
- i
- i
- i
(2) Introduce a heteroscedastic variant, where the response to each datum is different:
p(y j jy j ; M j ;- i
- i
- i
- i
- i
- i
- y
(3)
SLIDE 5
Uses for HGMMs
estimation of GMM parameters from data with missing features; estimation and prediction of indirectly observed mixture processes; modelling heteroscedastic data.Need only a simplified HGMM:
p(y j j) = n X i=1- i
- i
- i
(4) The gain matrices of each of the
N data, fM j g N j =1, contain only 1s and 0s,and are formed by deleting the rows from an identity matrix corresponding to the missing features of each datum. This is the marginal distribution of the remaining features.
SLIDE 6
The EM Algorithm
[Refer: Dempster, Laird and Rubin, 1977] Aim: to maximise a likelihood or posterior, over the parameters
, whilstintegrating out the nuisance parameters
Z.To maximise the likelihood
p(Y j; Z ), start with the guess,- =
- :
- )
- )
- = arg max
- Q(j
- )
SLIDE 7
The E-step for HGMMs
Assume conditionally independent data
Y = fy j g N j =1, and group theheteroscedastic parameters
fM j g N j =1 together into a set M.The E-step is the calculation of
P (ijy j ;- ;
and HGMM components:
P (ijy j ;- ;
- ;
- ;
- ;
- ;
i.e.
P (ijy j ;- ;
- ;
- i
- ;
- l
SLIDE 8
The M-step for HGMMs
Maximise
Q(j- ) with respect to
- i
1
N P (ijy j ;- ;
- i
- P
- ))H
- 1
- P
- )H
- iterate to find
- i :
- i
- i
- i
- i
- i
2
N X j =1 P (ijy j ;- )H
- M
- i
- M
- i
- I
- i
If
- < 2 then
- i
SLIDE 9
Results
Figure 1: Left: each feature missing 10% of the time. Figure 2: Right: each feature missing 60% of the time.
SLIDE 10