Introduction to Machine Learning
CMSC 422 MARINE CARPUAT
marine@cs.umd.edu
Machine Learning CMSC 422 M ARINE C ARPUAT marine@cs.umd.edu What - - PowerPoint PPT Presentation
Introduction to Machine Learning CMSC 422 M ARINE C ARPUAT marine@cs.umd.edu What is this course about? Machine learning studies algorithms for learning to do stuff By finding (and exploiting) patterns in data What can we do with
CMSC 422 MARINE CARPUAT
marine@cs.umd.edu
Analyze genomics data Recognize objects in images Analyze text & speech Teach robots how to cook from youtube videos
Question Answering system beats Jeopardy champion Ken Jennings at Quiz bowl!
– Replace ``human writing code'' with ``human supplying data''
– How to abstract from ``training'' examples to ``test'' examples?
– Finance, robotics, vision, machine translation, medicine, etc.
– Look at a problem – Identify if ML is an appropriate solution – If so, identify what types of algorithms might be applicable – Apply those algorithms
– A survey of ML algorithms – A tutorial on ML toolkits such as Weka, TensorFlow, …
Foundations of Supervised Learning
Advanced Supervised Learning
Unsupervised learning
We are here to help you learn by
– Introducing concepts from multiple perspectives
– Providing opportunities to practice, and feedback to help you stay on track
Teaching Assistants: Ryan Dorson Joe Yue-Hei Ng
– Do a lot of math (calculus, linear algebra, probability) – Do a fair amount of programming
– Do the required readings!
Grading
almost weekly
(30%), 3 of them, in teams
class
cumulative, in class.
2:59pm
http://www.cs.umd.edu/ class/spring2016/cmsc4 22//syllabus/
– Please use piazza instead of email
What does it mean to “learn by example”?
– Memorizing the training examples is not enough! – Need to generalize to make good predictions on test examples
– Many classifier hypotheses are plausible – Need assumptions about the nature of the relation between examples and classes
Problem setting
Input
} of unknown target function 𝑔 Output
𝑚(𝑧, 𝑧) where 𝑧 is the truth and 𝑧 the system’s prediction e.g. 𝑚 𝑧, 𝑔(𝑦) = 0 𝑗𝑔 𝑧 = 𝑔(𝑦) 1 𝑝𝑢ℎ𝑓𝑠𝑥𝑗𝑡𝑓 Captures our notion of what is important to learn
– Data generating distribution: Probability distribution 𝐸 over (𝑦, 𝑧) pairs – We don’t know what 𝐸 is! – We get a random sample from it: our training data
– as measured by loss 𝑚 – on future examples that are also draw from 𝐸
– 𝜁 , the expected loss of 𝑔 over 𝐸 with respect to 𝑚 should be small
𝜁 ≜ 𝔽 𝑦,𝑧 ~𝐸 𝑚(𝑧, 𝑔(𝑦)) =
(𝑦,𝑧)
𝐸 𝑦, 𝑧 𝑚(𝑧, 𝑔(𝑦))
don’t know what 𝐸 is
– training examples { 𝑦 1 , 𝑧 1 , … 𝑦 𝑂 , 𝑧 𝑂 }
𝜁 ≜
𝑜=1 𝑂 1
𝑂 𝑚(𝑧 𝑜 , 𝑔(𝑦 𝑜 ))
– a loss function 𝑚 – a sample from some unknown data distribution 𝐸
(𝑦,𝑧)
What does it mean to “learn by example”?
– Function approximation – Learning as minimizing expected loss
– due Wednesday 2:59pm