autonomous intelligent robotics
play

Autonomous Intelligent Robotics Instructor: Shiqi Zhang - PowerPoint PPT Presentation

Spring 2017 CIS 493, EEC 492, EEC 592: Autonomous Intelligent Robotics Instructor: Shiqi Zhang http://eecs.csuohio.edu/~szhang/teaching/17spring/ Machine learning: an introduction Slides adapted from Ray Mooney, Pedro Domingos, James Hays, and


  1. Spring 2017 CIS 493, EEC 492, EEC 592: Autonomous Intelligent Robotics Instructor: Shiqi Zhang http://eecs.csuohio.edu/~szhang/teaching/17spring/

  2. Machine learning: an introduction Slides adapted from Ray Mooney, Pedro Domingos, James Hays, and Yi-Fan Chang

  3. A Few Quotes ● “A breakthrough in machine learning would be worth ten Microsofts” (Bill Gates, Chairman, Microsoft) ● “Machine learning is the next Internet” (Tony Tether, Director, DARPA) ● Machine learning is the hot new thing” (John Hennessy, President, Stanford) ● “Web rankings today are mostly a matter of machine learning” (Prabhakar Raghavan, Dir. Research, Yahoo) ● “Machine learning is going to result in a real revolution” (Greg Papadopoulos, CTO, Sun) ● “Machine learning is today’s discontinuity” (Jerry Yang, CEO, Yahoo)

  4. So What Is Machine Learning? ● Automating automation ● Getting computers to program themselves ● Writing software is the bottleneck ● Let the data do the work instead!

  5. Traditional Programming Data Output Computer Program Machine Learning Data Program Computer Output

  6. Magic? No, more like gardening • Seeds = Algorithms • Nutrients = Data • Gardener = You • Plants = Programs

  7. Why Study Machine Learning? Engineering Better Computing Systems ● Develop systems that are too difficult/expensive to construct manually because they require specific detailed skills or knowledge tuned to a specific task ( knowledge engineering bottleneck ). ● Develop systems that can automatically adapt and customize themselves to individual users. Personalized news or mail filter – – Personalized tutoring ● Discover new knowledge from large databases ( data mining ). Market basket analysis (e.g. diapers and beer) – Medical text mining (e.g. migraines to calcium channel blockers to – magnesium) 7

  8. Why Study Machine Learning? The Time is Ripe ● Many basic effective and efficient algorithms available. ● Large amounts of on-line data available. ● Large amounts of computational resources available. 8

  9. Related Disciplines ● Artificial Intelligence ● Data Mining ● Probability and Statistics ● Information theory ● Numerical optimization ● Computational complexity theory ● Control theory (adaptive) ● Psychology (developmental, cognitive) ● Neurobiology ● Linguistics ● Philosophy 9

  10. Defining the Learning Task Improve on task, T, with respect to performance metric, P, based on experience, E. T: Playing checkers P: Percentage of games won against an arbitrary opponent E: Playing practice games against itself T: Recognizing hand-written words P: Percentage of words correctly classified E: Database of human-labeled images of handwritten words T: Driving on four-lane highways using vision sensors P: Average distance traveled before a human-judged error E: A sequence of images and steering commands recorded while observing a human driver. T: Categorize email messages as spam or legitimate. P: Percentage of email messages correctly classified. 10 E: Database of emails, some with human-given labels

  11. ML in a Nutshell ● Tens of thousands of machine learning algorithms ● Hundreds new every year ● Every machine learning algorithm has three components: – Representation – Evaluation – Optimization

  12. Representation ● Decision trees ● Sets of rules / Logic programs ● Instances ● Graphical models (Bayes/Markov nets) ● Neural networks ● Support vector machines ● Model ensembles ● Etc.

  13. Evaluation ● Accuracy ● Precision and recall ● Squared error ● Likelihood ● Posterior probability ● Cost / Utility ● Margin ● Entropy ● K-L divergence ● Etc.

  14. Optimization ● Combinatorial optimization – E.g.: Greedy search ● Convex optimization – E.g.: Gradient descent ● Constrained optimization – E.g.: Linear programming

  15. Types of Learning ● Supervised (inductive) learning – Training data includes desired outputs ● Unsupervised learning – Training data does not include desired outputs ● Semi-supervised learning – Training data includes a few desired outputs ● Reinforcement learning – Rewards from sequence of actions

  16. Inductive Learning ● Given examples of a function (X, F(X)) ● Predict function F(X) for new examples X – Discrete F(X) : Classification – Continuous F(X) : Regression – F(X) = Probability( X ): Probability estimation

  17. Supervised vs. Unsupervised Unsupervised Supervised learning learning 17 Semi-supervised learning

  18. Supervised learning

  19. Supervised learning: Steps Training Training Labels Training Images Image Learned Training Features model Testing Image Learned Prediction Features model Test Image Slide credit: D. Hoiem and L. Lazebnik

  20. Unsupervised learning

  21. Fully supervised “Weakly” supervised Unsupervised Definition depends on task Slide credit: L. Lazebnik

  22. Generalization Training set (labels known) Test set (labels unknown) How well does a learned model generalize from the data it was trained on to ● a new test set? Slide credit: L. Lazebnik

  23. Generalization ● Components of generalization error Bias: how much the average model over all training sets differ from – the true model? ● Error due to inaccurate assumptions/simplifications made by the model Variance: how much models estimated from different training sets – differ from each other ● Underfitting: model is too “simple” to represent all the relevant class characteristics High bias and low variance – High training error and high test error – ● Overfitting: model is too “complex” and fits irrelevant characteristics (noise) in the data Low bias and high variance – Overfitting Thriller! Low training error and high test error – https://youtu.be/DQWI1kvmwRg Slide credit: L. Lazebnik

  24. Generative vs. Discriminative Classifiers Generative Models Discriminatjve Models ● Represent both the data • Learn to directly predict the and the labels labels from the data ● Often, makes use of • Ofuen, assume a simple conditional independence boundary (e.g., linear) and priors • Examples ● Examples – Logistjc regression Naïve Bayes classifier – – SVM Bayesian network – – Boosted decision trees ● Models of data may apply • Ofuen easier to predict a to future prediction label from the data than to problems model the data Slide credit: D. Hoiem

  25. N e a r e s t N e i g h b o r C l a s s i fj e r •A s s i g n l a b e l o f n e a r e s t t r a i n i n g d a t a p o i n t t o e a c h t e s t d a t a p o i n t from Duda et al. Voronoi partitioning of feature space for two-category 2D and 3D data Source: D. Lowe

  26. K-nearest neighbor x x o x x x x + o x o x o + o o o x2 x1

  27. 1-nearest neighbor x x o x x x x + o x o x o + o o o x2 x1

  28. 3-nearest neighbor x x o x x x x + o x o x o + o o o x2 x1

  29. 5-nearest neighbor x x o x x x x + o x o x o + o o o x2 x1

  30. Using K-NN • Simple, a good one to try fjrst • With infjnite examples, 1-NN provably has error that is at most twice Bayes optjmal error

  31. Sample Applications ● Web search ● Computational biology ● Finance ● E-commerce ● Space exploration ● Robotics (reasoning, planning, control, etc) ● Information extraction ● Social networks ● Debugging ● [Your favorite area]

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend