em algorithm
play

EM Algorithm 09-09-2019 For Mixture Gaussian Models Instructor - - PowerPoint PPT Presentation

E9 205 Machine Learning for Signal Processing EM Algorithm 09-09-2019 For Mixture Gaussian Models Instructor - Sriram Ganapathy (sriramg@iisc.ac.in) Teaching Assistant - Prachi Singh (prachisingh@iisc.ac.in). Basics of Information Theory


  1. E9 205 Machine Learning for Signal Processing EM Algorithm 09-09-2019 For Mixture Gaussian Models Instructor - Sriram Ganapathy (sriramg@iisc.ac.in) Teaching Assistant - Prachi Singh (prachisingh@iisc.ac.in).

  2. Basics of Information Theory • Entropy of distribution • KL divergence • Jensen’s inequality • Expectation Maximization Algorithm for MLE

  3. Gaussian Mixture Models A Gaussian Mixture Model (GMM) is defined as The weighting coefficients have the property

  4. Gaussian Mixture Models • Properties of GMM • Can model multi-modal data. • Identify data clusters. • Can model arbitrarily complex data distributions The set of parameters for the model are The number of parameters is

  5. MLE for GMM • The log-likelihood function over the entire data in this case will have a logarithm of a summation • Solving for the optimal parameters using MLE for GMM is not straight forward. • Resort to the Expectation Maximization (EM) algorithm

  6. Basics of Information Theory • Entropy of distribution • KL divergence • Jensen’s inequality • Expectation Maximization Algorithm for MLE

  7. Expectation Maximization Algorithm • Iterative procedure. • Assume the existence of hidden variable associated with each data sample • Let the current estimate (at iteration n) be Define the Q function as

  8. Expectation Maximization Algorithm • It can be proven that if we choose then • In many cases, finding the maximum for the Q function may be easier than likelihood function w.r.t. the parameters. • Solution is dependent on finding a good choice of the hidden variables which eases the computation • Iteratively improve the log-likelihood function.

  9. EM Algorithm Summary • Initialize with a set of model parameters (n=1) • Compute the conditional expectation (E-step) • Maximize the conditional expectation w.r.t. parameter. (M-step) (n = n+1) • Check for convergence • Go back to E-step if model has not converged.

  10. EM Algorithm for GMM • The hidden variables will be the index of the mixture component which generated • Re-estimation formulae

  11. EM Algorithm for GMM E-step M-step

  12. EM Algorithm for GMM ANEMIA PATIENTS AND CONTROLS 4.4 4.3 Red Blood Cell Hemoglobin Concentration 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  13. EM Algorithm for GMM EM ITERATION 1 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  14. EM Algorithm for GMM EM ITERATION 3 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  15. EM Algorithm for GMM EM ITERATION 5 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  16. EM Algorithm for GMM EM ITERATION 10 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  17. EM Algorithm for GMM EM ITERATION 15 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  18. EM Algorithm for GMM EM ITERATION 25 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  19. EM Algorithm for GMM ANEMIA DATA WITH LABELS 4.4 Red Blood Cell Hemoglobin Concentration 4.3 4.2 4.1 4 3.9 3.8 3.7 3.3 3.4 3.5 3.6 3.7 3.8 3.9 4 Red Blood Cell Volume Cadez, Igor V., et al. "Hierarchical models for screening of iron deficiency anemia." ICML . 1999.

  20. K-means Algorithm for Initialization

  21. Other Considerations • Initialization - random or k-means • Number of Gaussians • Type of Covariance matrix • Spherical covariance

  22. Other Considerations • Initialization - random or k-means • Number of Gaussians • Type of Covariance matrix • Diagonal covariance

  23. Other Considerations • Initialization - random or k-means • Number of Gaussians • Type of Covariance matrix • Full covariance

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend