Autonomous Intelligent Robotics Instructor: Shiqi Zhang - - PowerPoint PPT Presentation

autonomous intelligent robotics
SMART_READER_LITE
LIVE PREVIEW

Autonomous Intelligent Robotics Instructor: Shiqi Zhang - - PowerPoint PPT Presentation

Spring 2017 CIS 493, EEC 492, EEC 592: Autonomous Intelligent Robotics Instructor: Shiqi Zhang http://eecs.csuohio.edu/~szhang/teaching/17spring/ Machine learning: an introduction Slides adapted from Ray Mooney, Pedro Domingos, James Hays, and


slide-1
SLIDE 1

Spring 2017 CIS 493, EEC 492, EEC 592:

Autonomous Intelligent Robotics

Instructor: Shiqi Zhang

http://eecs.csuohio.edu/~szhang/teaching/17spring/

slide-2
SLIDE 2

Machine learning: an introduction

Slides adapted from Ray Mooney, Pedro Domingos, James Hays, and Yi-Fan Chang

slide-3
SLIDE 3

A Few Quotes

  • “A breakthrough in machine learning would be worth

ten Microsofts” (Bill Gates, Chairman, Microsoft)

  • “Machine learning is the next Internet”

(Tony Tether, Director, DARPA)

  • Machine learning is the hot new thing”

(John Hennessy, President, Stanford)

  • “Web rankings today are mostly a matter of machine learning”

(Prabhakar Raghavan, Dir. Research, Yahoo)

  • “Machine learning is going to result in a real revolution” (Greg

Papadopoulos, CTO, Sun)

  • “Machine learning is today’s discontinuity”

(Jerry Yang, CEO, Yahoo)

slide-4
SLIDE 4

So What Is Machine Learning?

  • Automating automation
  • Getting computers to program themselves
  • Writing software is the bottleneck
  • Let the data do the work instead!
slide-5
SLIDE 5

Traditional Programming Machine Learning Computer Data Program Output Computer Data Output Program

slide-6
SLIDE 6

Magic?

No, more like gardening

  • Seeds = Algorithms
  • Nutrients = Data
  • Gardener = You
  • Plants = Programs
slide-7
SLIDE 7

7

Why Study Machine Learning? Engineering Better Computing Systems

  • Develop systems that are too difficult/expensive to

construct manually because they require specific detailed skills or knowledge tuned to a specific task (knowledge engineering bottleneck).

  • Develop systems that can automatically adapt and

customize themselves to individual users.

Personalized news or mail filter

Personalized tutoring

  • Discover new knowledge from large databases (data

mining).

Market basket analysis (e.g. diapers and beer)

Medical text mining (e.g. migraines to calcium channel blockers to magnesium)

slide-8
SLIDE 8

8

Why Study Machine Learning? The Time is Ripe

  • Many basic effective and efficient

algorithms available.

  • Large amounts of on-line data available.
  • Large amounts of computational resources

available.

slide-9
SLIDE 9

9

Related Disciplines

  • Artificial Intelligence
  • Data Mining
  • Probability and Statistics
  • Information theory
  • Numerical optimization
  • Computational complexity theory
  • Control theory (adaptive)
  • Psychology (developmental, cognitive)
  • Neurobiology
  • Linguistics
  • Philosophy
slide-10
SLIDE 10

10

Defining the Learning Task

Improve on task, T, with respect to

performance metric, P, based on experience, E.

T: Playing checkers P: Percentage of games won against an arbitrary opponent E: Playing practice games against itself T: Recognizing hand-written words P: Percentage of words correctly classified E: Database of human-labeled images of handwritten words T: Driving on four-lane highways using vision sensors P: Average distance traveled before a human-judged error E: A sequence of images and steering commands recorded while

  • bserving a human driver.

T: Categorize email messages as spam or legitimate. P: Percentage of email messages correctly classified. E: Database of emails, some with human-given labels

slide-11
SLIDE 11

ML in a Nutshell

  • Tens of thousands of machine learning

algorithms

  • Hundreds new every year
  • Every machine learning algorithm has three

components:

– Representation – Evaluation – Optimization

slide-12
SLIDE 12

Representation

  • Decision trees
  • Sets of rules / Logic programs
  • Instances
  • Graphical models (Bayes/Markov nets)
  • Neural networks
  • Support vector machines
  • Model ensembles
  • Etc.
slide-13
SLIDE 13

Evaluation

  • Accuracy
  • Precision and recall
  • Squared error
  • Likelihood
  • Posterior probability
  • Cost / Utility
  • Margin
  • Entropy
  • K-L divergence
  • Etc.
slide-14
SLIDE 14

Optimization

  • Combinatorial optimization

– E.g.: Greedy search

  • Convex optimization

– E.g.: Gradient descent

  • Constrained optimization

– E.g.: Linear programming

slide-15
SLIDE 15

Types of Learning

  • Supervised (inductive) learning

– Training data includes desired outputs

  • Unsupervised learning

– Training data does not include desired outputs

  • Semi-supervised learning

– Training data includes a few desired outputs

  • Reinforcement learning

– Rewards from sequence of actions

slide-16
SLIDE 16

Inductive Learning

  • Given examples of a function (X, F(X))
  • Predict function F(X) for new examples X

– Discrete F(X): Classification – Continuous F(X): Regression – F(X) = Probability(X): Probability estimation

slide-17
SLIDE 17

Supervised vs. Unsupervised

17 Supervised learning Unsupervised learning Semi-supervised learning

slide-18
SLIDE 18

Supervised learning

slide-19
SLIDE 19

Prediction

Supervised learning: Steps

Training Labels Training Images Training

Training

Image Features Image Features

Testing

Test Image Learned model Learned model

Slide credit: D. Hoiem and L. Lazebnik

slide-20
SLIDE 20

Unsupervised learning

slide-21
SLIDE 21

Unsupervised “Weakly” supervised Fully supervised Definition depends on task

Slide credit: L. Lazebnik

slide-22
SLIDE 22

Generalization

  • How well does a learned model generalize from the data it was trained on to

a new test set? Training set (labels known) Test set (labels unknown)

Slide credit: L. Lazebnik

slide-23
SLIDE 23

Generalization

  • Components of generalization error

Bias: how much the average model over all training sets differ from the true model?

  • Error due to inaccurate assumptions/simplifications made by the

model

Variance: how much models estimated from different training sets differ from each other

  • Underfitting: model is too “simple” to represent all the

relevant class characteristics

High bias and low variance

High training error and high test error

  • Overfitting: model is too “complex” and fits irrelevant

characteristics (noise) in the data

Low bias and high variance

Low training error and high test error

Slide credit: L. Lazebnik

Overfitting Thriller! https://youtu.be/DQWI1kvmwRg

slide-24
SLIDE 24

Generative vs. Discriminative Classifiers

Generative Models

  • Represent both the data

and the labels

  • Often, makes use of

conditional independence and priors

  • Examples

Naïve Bayes classifier

Bayesian network

  • Models of data may apply

to future prediction problems

Discriminatjve Models

  • Learn to directly predict the

labels from the data

  • Ofuen, assume a simple

boundary (e.g., linear)

  • Examples

– Logistjc regression – SVM – Boosted decision trees

  • Ofuen easier to predict a

label from the data than to model the data

Slide credit: D. Hoiem

slide-25
SLIDE 25

N e a r e s t N e i g h b

  • r

C l a s s i fj e r

  • A

s s i g n l a b e l

  • f

n e a r e s t t r a i n i n g d a t a p

  • i

n t t

  • e

a c h t e s t d a t a p

  • i

n t

Voronoi partitioning of feature space for two-category 2D and 3D data

from Duda et al.

Source: D. Lowe

slide-26
SLIDE 26

K-nearest neighbor

x x x x x x x x

  • x2

x1 + +

slide-27
SLIDE 27

1-nearest neighbor

x x x x x x x x

  • x2

x1 + +

slide-28
SLIDE 28

3-nearest neighbor

x x x x x x x x

  • x2

x1 + +

slide-29
SLIDE 29

5-nearest neighbor

x x x x x x x x

  • x2

x1 + +

slide-30
SLIDE 30

Using K-NN

  • Simple, a good one to try fjrst
  • With infjnite examples, 1-NN provably has error

that is at most twice Bayes optjmal error

slide-31
SLIDE 31

Sample Applications

  • Web search
  • Computational biology
  • Finance
  • E-commerce
  • Space exploration
  • Robotics (reasoning, planning, control, etc)
  • Information extraction
  • Social networks
  • Debugging
  • [Your favorite area]
slide-32
SLIDE 32
slide-33
SLIDE 33