mixture models
play

Mixture Models Bhiksha Raj 27 Oct 2016 11755/18797 1 Learning - PowerPoint PPT Presentation

Machine Learning for Signal Processing Expectation Maximization Mixture Models Bhiksha Raj 27 Oct 2016 11755/18797 1 Learning Distributions for Data Problem: Given a collection of examples from some data, estimate its distribution


  1. Machine Learning for Signal Processing Expectation Maximization Mixture Models Bhiksha Raj 27 Oct 2016 11755/18797 1

  2. Learning Distributions for Data • Problem: Given a collection of examples from some data, estimate its distribution • Solution: Assign a model to the distribution – Learn parameters of model from data • Models can be arbitrarily complex – Mixture densities, Hierarchical models. 11755/18797 2

  3. A Thought Experiment 6 3 1 5 4 1 2 4 … • A person shoots a loaded dice repeatedly • You observe the series of outcomes • You can form a good idea of how the dice is loaded – Figure out what the probabilities of the various numbers are for dice • P(number) = count(number)/count(rolls) • This is a maximum likelihood estimate – Estimate that makes the observed sequence of numbers most probable 11755/18797 3

  4. The Multinomial Distribution • A probability distribution over a discrete collection of items is a Multinomial  ( : belongs to a discrete set ) ( ) P X X P X • E.g. the roll of dice – X : X in (1,2,3,4,5,6) • Or the toss of a coin – X : X in (head, tails) 11755/18797 4

  5. Maximum Likelihood Estimation n 2 n 4 n 1 n 5 n 6 n 3 p 6 p 3 p 4 p 1 p 2 p 2 p 4 p 5 p 1 p 5 p 6 p 3 • Basic principle: Assign a form to the distribution – E.g. a multinomial – Or a Gaussian • Find the distribution that best fits the histogram of the data 11755/18797 5

  6. Defining “Best Fit” • The data are generated by draws from the distribution – I.e. the generating process draws from the distribution • Assumption: The world is a boring place – The data you have observed are very typical of the process • Consequent assumption: The distribution has a high probability of generating the observed data – Not necessarily true • Select the distribution that has the highest probability of generating the data – Should assign lower probability to less frequent observations and vice versa 11755/18797 6

  7. Maximum Likelihood Estimation: Multinomial • Probability of generating (n 1 , n 2 , n 3 , n 4 , n 5 , n 6 )   n ( , , , , , ) P n n n n n n Const p i 1 2 3 4 5 6 i i • Find p 1 ,p 2 ,p 3 ,p 4 ,p 5 ,p 6 so that the above is maximized • Alternately maximize        log ( , , , , , ) log( ) log P n n n n n n Const n p 1 2 3 4 5 6 i i i – Log() is a monotonic function – argmax x f(x) = argmax x log(f(x)) • Solving for the probabilities gives us EVENTUALLY n – Requires constrained optimization to  i p ITS JUST  i ensure probabilities sum to 1 n COUNTING! j j 11755/18797 7

  8. Segue: Gaussians   1  X  m Q    m Q  m 1 T ( ) ( ; , ) exp 0 . 5 ( ) ( ) P X N X X  Q d ( 2 ) | | • Parameters of a Gaussian: – Mean m , Covariance Q 11755/18797 8

  9. Maximum Likelihood: Gaussian  Given a collection of observations ( X 1 , X 2 ,…), estimate mean m and covariance Q   1      m Q  m 1 T ( , ,...) exp 0 . 5 ( ) ( ) P X X X X 1 2 i i  Q d ( 2 ) | | i           Q   m Q  m 1 T log ( , ,...) 0 . 5 log | | ( ) ( ) P X X C X X 1 2 i i i • Maximizing w.r.t m and Q gives us ITS STILL 1 1      JUST m  Q   m  m T X X X i i i COUNTING! N N i i 11755/18797 9

  10. Laplacian  m   1 | | x  m     ( ) ( ; , ) exp P x L x b   2 b b • Parameters: Median m , scale b ( b > 0) – m is also the mean, but is better viewed as the median 11755/18797 10

  11. Maximum Likelihood: Laplacian  Given a collection of observations ( x 1 , x 2 ,…), estimate mean m and scale b  m | |   x     i log ( , ,...) log( ) P x x C N b 1 2 b i • Maximizing w.r.t m and b gives us Still just counting 1  m    m ({ }) | | median x b x i i N i 11755/18797 11

  12. Dirichlet (from wikipedia) log of the density as we change α from α=(0.3, 0.3, 0.3) to (2.0, 2.0, 2.0), keeping all the individual αi's equal to each other.  K =3. Clockwise from top left:  a ( ) α =(6, 2, 2), (3, 7, 5), (6, 2, 6), (2, 3, 4) i  a   a  1 ( ) ( ; ) i P X D X x i   i    a  i • Parameters are a s i   i – Determine mode and curvature • Defined only of probability vectors – X = [x 1 x 2 .. x K ] , S i x i = 1, x i >= 0 for all i 11755/18797 12

  13. Maximum Likelihood: Dirichlet  Given a collection of observations ( X 1 , X 2 ,…), estimate a                 a    a   a   log ( , ,...) ( 1 ) log( ) log log P X X X N N   1 2 , i j i i i     j i i i • No closed form solution for a s. – Needs gradient ascent • Several distributions have this property: the ML estimate of their parameters have no closed form solution 11755/18797 13

  14. Continuing the Thought Experiment 6 3 1 5 4 1 2 4 … 4 4 1 6 3 2 1 2 … • Two persons shoot loaded dice repeatedly – The dice are differently loaded for the two of them • We observe the series of outcomes for both persons • How to determine the probability distributions of the two dice? 11755/18797 14

  15. Estimating Probabilities 6 4 5 1 2 3 4 5 2 2 1 4 3 4 6 2 1 6 … • Observation: The sequence of numbers from the two dice – As indicated by the colors, we know who rolled what number 11755/18797 15

  16. Estimating Probabilities 6 4 5 1 2 3 4 5 2 2 1 4 3 4 6 2 1 6 … • Observation: The sequence of numbers from the two dice – As indicated by the colors, we know who rolled what number 4 1 3 5 2 4 4 2 6.. 6 5 2 4 2 1 3 6 1.. • Segregation: Separate the blue Collection of “blue” Collection of “red” observations from the red numbers numbers 11755/18797 16

  17. Estimating Probabilities • Observation: The sequence of 6 4 5 1 2 3 4 5 2 2 1 4 3 4 6 2 1 6 … numbers from the two dice – As indicated by the colors, we know who rolled what number • Segregation: Separate the blue 4 1 3 5 2 4 4 2 6.. 6 5 2 4 2 1 3 6 1.. observations from the red • From each set compute probabilities for each of the 6 0.3 0.3 0.25 0.25 possible outcomes 0.2 0.2 0.15 0.15 0.1 0.1 no. of times number was rolled 0.05 0.05  ( ) 0 0 P number 1 2 3 4 5 6 1 2 3 4 5 6 total number of observed rolls 11755/18797 17

  18. A Thought Experiment 6 4 1 5 3 2 2 2 … 6 3 1 5 4 1 2 4 … 4 4 1 6 3 2 1 2 … • Now imagine that you cannot observe the dice yourself • Instead there is a “ caller ” who randomly calls out the outcomes – 40% of the time he calls out the number from the left shooter, and 60% of the time, the one from the right (and you know this) • At any time, you do not know which of the two he is calling out • How do you determine the probability distributions for the two dice? 18 11755/18797

  19. A Thought Experiment 6 4 1 5 3 2 2 2 … 6 3 1 5 4 1 2 4 … 4 4 1 6 3 2 1 2 … • How do you now determine the probability distributions for the two sets of dice … • .. If you do not even know what fraction of time the blue numbers are called, and what fraction are red? 11755/18797 19

  20. A Mixture Multinomial • The caller will call out a number X in any given callout IF – He selects “RED”, and the Red die rolls the number X – OR – He selects “BLUE” and the Blue die rolls the number X • P(X) = P(Red)P(X|Red) + P(Blue)P(X|Blue) – E.g. P(6) = P(Red)P(6|Red) + P(Blue)P(6|Blue) • A distribution that combines (or mixes ) multiple multinomials is a mixture multinomial   ( ) ( ) ( | ) P X P Z P X Z Z Mixture weights Component multinomials 11755/18797 20

  21. Mixture Distributions Mixture Gaussian    m Q  ( ) ( ) ( ; , ) P X P Z N X ( ) ( ) ( | ) P X P Z P X Z z z Z Z Mixture weights Component distributions Mixture of Gaussians and Laplacians     m Q  m ( ) ( ) ( ; , ) ( ) ( ; , ) P X P Z N X P Z L X b , z z i z z i Z Z i • Mixture distributions mix several component distributions – Component distributions may be of varied type • Mixing weights must sum to 1.0 • Component distributions integrate to 1.0 • Mixture distribution integrates to 1.0 11755/18797 21

  22. Maximum Likelihood Estimation   ( ) ( ) ( | ) P X P Z P X Z • For our problem: Z – Z = color of dice n   X        n ( , , , , , ) ( ) ( ) ( | ) P n n n n n n Const P X Const P Z P X Z X 1 2 3 4 5 6   X X Z • Maximum likelihood solution: Maximize         log( ( , , , , , )) log( ) log ( ) ( | ) P n n n n n n Const n P Z P X Z 1 2 3 4 5 6 X   X Z • No closed form solution (summation inside log)! – In general ML estimates for mixtures do not have a closed form – USE EM! 11755/18797 22

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend