probabilistic classifiers generative
play

Probabilistic Classifiers -- Generative Naive Bayes Announcements - PowerPoint PPT Presentation

Probabilistic Classifiers -- Generative Naive Bayes Announcements Math for Visual Computing Course Agenda: First classifier using our knowledge of probabilities Quick overview of Linear Algebra Least Square


  1. Probabilistic Classifiers -- Generative Naive Bayes

  2. Announcements ● Math for Visual Computing Course ● Agenda: ○ First classifier using our knowledge of probabilities ○ Quick overview of Linear Algebra ○ Least Square Regression

  3. Definition Classifier outputs a probability distribution over the class labels: X → P(y|x)

  4. Types of Classifiers ● Generative: learns to generate the examples ○ Learn P(x|y) and P(y) from the training data and then apply Bayes rule to find P(y|x)

  5. Types of Classifiers ● Conditional Classifiers: Model conditional distribution P(y|x) directly. ○ Example: logistic regression classifier ○ Neural Networks.

  6. Naive Bayes Classifier: A Generative Classifer Each attribute � is conditionally independent given the class label Formula:

  7. Notion of conditional independence

  8. Training a Naïve Bayesian Classifier ● Given training data: ● Estimating P(y): A multinomial distribution. ● Estimating � ○ If j-th attribute is categorical: P(xj|y) is estimated as the relative freq of samples having value di as j-th attribute in class y ○ If j-th attribute is continuous: P(xj|y) is estimated through a continuous density function: eg. Gaussian density function ● Computationally easy in both cases

  9. Play-tennis example: estimating P(x i |C) outlook P(sunny|p) = 2/9 P(sunny|n) = 3/5 P(overcast|p) = 4/9 P(overcast|n) = 0 P(rain|p) = 3/9 P(rain|n) = 2/5 temperature P(hot|p) = 2/9 P(hot|n) = 2/5 P(mild|p) = 4/9 P(mild|n) = 2/5 P(cool|p) = 3/9 P(cool|n) = 1/5 humidity P(high|p) = 3/9 P(high|n) = 4/5 P(normal|p) = 6/9 P(normal|n) = 2/5 P(p) = 9/14 windy P(n) = 5/14 P(true|p) = 3/9 P(true|n) = 3/5 P(false|p) = 6/9 P(false|n) = 2/5

  10. Naive Bayesian Classifier (II) ● Given a training set, we can compute the probabilities

  11. Play-tennis example: classifying X ● An unseen sample X = <rain, hot, high, false> ● P(X|p)·P(p) = P(rain|p)·P(hot|p)·P(high|p)·P(false|p)·P(p) = 3/9·2/9·3/9·6/9·9/14 = 0.010582 ● ● P(X|n)·P(n) =P(rain|n)·P(hot|n)·P(high|n)·P(false|n)·P(n) = 2/5·2/5·4/5·2/5·5/14 = 0.018286 ● ● Sample X is classified in class n (don’t play)

  12. Example: MNIST digit classification Jupyter Notebook: http://d2l.ai/chapter_crashcourse/naive-bayes.html

  13. Naive Bayes for Spam Filtering.

  14. Summary of Naïve Bayes Classifier ● Creates a distribution over class label given x P(y|x) by applying Bayes rule. ○ Requires estimating P(x|y) for each class y and P(y) ● Estimates P(x|y) by assuming that each attributes of x are conditionally independent given the class label ○ Very easy computationally. ● Many applications in spite of simplistic assumption. ● Can we do better?

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend