CSCI 4520: Introduction to Machine Learning
Mehdi Allahyari Georgia Southern University
(some slides from Pedro Domingos, University of Washington)
Spring 2020
1
CSCI 4520: Introduction to Machine Learning Spring 2020 Mehdi - - PowerPoint PPT Presentation
CSCI 4520: Introduction to Machine Learning Spring 2020 Mehdi Allahyari Georgia Southern University (some slides from Pedro Domingos, University of Washington) 1 Mehdi Allahyari n Call me Dr. Yari n PhD from UGA n Machine Learning, Semantic
Mehdi Allahyari Georgia Southern University
(some slides from Pedro Domingos, University of Washington)
1
n Call me Dr. Yari n PhD from UGA n Machine Learning, Semantic Web, Topic Modeling,
2
3
CSCI 4520 - Introduction to Machine Learning
Spring 2020 – CRN: 17186
Instructor: Dr. Mehdi Allahyari <mallahyari@georgiasouthern.edu> Office: CEIT 2321 Course Webpage: https://sci2lab.github.io/mehdi/teaching/cs4520 Lecture Times Tuesday Thursday CRN 17186 12:30 PM – 01:45 PM 12:30 PM – 01:45 PM CEIT 2207 CEIT 2207 Office Hours Monday Wednesday 11:00 AM – 12:00 PM 11:00 AM – 12:00 PM
Tuesday Thursday 11:00 AM – 12:00 PM 11:00 AM – 12:00 PM CEIT 2321 CEIT 2321
4
5
n Course page: https://sci2lab.github.io/mehdi/teaching/cs4520/
n is where you get course content (syllabus, lectures, etc.)
n Folio:
n Folio is the website where you will submit all assignments (if
6
n This introductory course is designed:
n to give upper level undergraduate a broad overview of many
n covers the theory and practical algorithms for machine learning
n Topics:
n Statistical and probabilistic methods, generative and
n In addition, the course covers fundamental concepts such as
7
n Basics knowledge of probability, linear algebra and
n For example, standard probability distributions and also how to
n Significance programming experience
n Python is the language we’ll use
8
n Attendance and participation: 5% n Homeworks: 20%
n Four homeworks: submitted hardcopy and on Folio n Some exercises, significant programming
n Midterm 03/05/2020 20% n Final exam 05/05/2020 12:30 – 2:30 PM. 25% n Final project: 30%
n Teams of 2 n Define a task, create/acquire data for the task, train ML algorithm(s),
n Check the course website for details
9
n 0-24 hours late
n 24-48 hours late
n >48 hours late
n Double check your submission n Save your confirmation email
10
n Homeworks to be done individually
n Your own answer and code
n can collaborate in figuring out answers and helping each
n No sharing of pseudocode or code or simulation results n Academic Honesty:
n abide by the University’s academic honesty policy and the
n Lack of knowledge of the academic honesty policy is not a
11
n Familiarize yourself with the syllabus n This PowerPoint highlights key information from the
n You are responsible for the entirety of the syllabus
n Your HW is to go home & read the syllabus thoroughly
12
n “A breakthrough in machine learning would be worth
n “Machine learning is the next Internet”
n Machine learning is the hot new thing”
n “Machine learning is Google’s top priority”
n “Machine learning is Microsoft Research’s largest investment area” (Peter Lee, Head,
n “‘Data scientist’ is the hottest job title in Silicon Valley”
13
n How can we solve a specific problem? n As computer scientists we write a program that encodes a set of
n In many cases is very difficult to specify those rules, e.g., given a
14
n Learning systems are not directly programmed to solve a
n Examples of how they should behave n From trial-and-error experience trying to solve the problem
n Different than standard CS n Want to implement unknown function, only have access e.g.,
n Learning simply means incorporating information from the
15
n “The study of computer programs thatimprove automatically
n T. Mitchell Machine Learning n Getting computers to program themselves n Writing software is the bottleneck, Let the data do the work
16
17
} Seeds = Algorithms } Nutrients = Data } Gardener = Y
} Plants = Programs
18
n Web search n Computational biology n Finance n E-commerce n Space exploration n Robotics n Information extraction n Social networks n Debugging n [Your favorite area]
19
20
n It is very hard to write programs that solve problems like
n What distinguishes a 2 from a 7? n How does our brain do it?
n Instead of writing a program by hand, we collect examples that
n A machine learning algorithm then takes these examples and
n The program produced by the learning algorithm may look very
n If we do it right, the program works for new cases as well as the
21
n Statistics n Analytics / Data Science n Data Mining n Artificial Intelligence
22
n Understanding domain, prior knowledge, and goals n Data integration, selection, cleaning n pre-processing, etc. n Learning models n Interpreting results n Consolidating and deploying discovered knowledge n Loop
23
n Hundreds of machine learning algorithms n Tens new every year n Every machine learning algorithm has three components:
n Representation n Evaluation n Optimization
24
25
n Given some data, how can we tell if a function is
26
n E.g.: Greedy search
n E.g.: Gradient descent
n E.g.: Linear programming
27
28
29
30
31
32
33
34
Figure: Siri: https://www.youtube.com/watch?v=8ciagGASro0
35
Figure: Photomath: https://photomath.net/
36
37
38
39
40
41
42
43
44
https://blog.insightdatascience.com/generating-custom-photo-realistic-faces-using-ai-d170b1b59255
https://www.youtube.com/watch?v=CrnLINlbfFA Fig: https://i.ytimg.com/vi/CrnLINlbfFA/hqdefault.jpg
45
[Gatys, Ecker, Bethge. A Neural Algorithm of Artistic Style. Arxiv’15.]
46
47
48
Figure: Video: https://www.youtube.com/watch?v=YQIMGV5vtd4
49
1.
2.
3.
4.
5.
6.
7.
50
Figure: Video: https://www.youtube.com/watch?v=V1eYniJ0Rnk
51
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
52
n Training data includes desired outputs
n Training data does not include desired outputs
n Training data includes a few desired outputs
n Rewards from sequence of actions
53
n Given examples of a function (X, F(X)) n Predict function F(X) for new examples X
n Discrete F(X): Classification (1-of-N output (speech
n Continuous F(X): Regression (real-valued output (predicting
n F(X) = Probability(X): Probability estimation
54
n Clustering: Discover groups of similar inputs
55
n Decision tree induction n Instance-based learning n Linear Regression n Bayesian learning n Logistic Regression n Neural networks
n Clustering n Dimensionality reduction
56