CS480/680 Machine Learning Lecture 1: January 7th, 2020
Course Introduction Zahra Sheikhbahaee
University of Waterloo CS480/680 Winter 2020 Zahra Sheikhbahaee 1
CS480/680 Machine Learning Lecture 1: January 7 th , 2020 Course - - PowerPoint PPT Presentation
CS480/680 Machine Learning Lecture 1: January 7 th , 2020 Course Introduction Zahra Sheikhbahaee CS480/680 Winter 2020 Zahra Sheikhbahaee University of Waterloo 1 Outline Introduction to Machine Learning Course website and details:
University of Waterloo CS480/680 Winter 2020 Zahra Sheikhbahaee 1
CS480/680 Winter 2020 Zahra Sheikhbahaee 2
University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 3
University of Waterloo
Postdoctoral Researcher
§ Gaurav Gupta g27gupta@uwaterloo.ca § Zeou Hu z97hu@uwaterloo.ca § Arash Mollajafari Sohi amollaja@uwaterloo.ca § Zahra Rezapour Siahgourabi zrezapou@uwaterloo.ca § Colin Vandenhof cm5vande@uwaterloo.ca
University of Waterloo CS480/680 Winter 2020 Zahra Sheikhbahaee 4
University of Waterloo CS480/680 Winter 2020 Zahra Sheikhbahaee 5
University of Waterloo CS480/680 Winter 2020 Zahra Sheikhbahaee 6
CS480/680 Winter 2020 Zahra Sheikhbahaee 7
University of Waterloo
computers the ability to learn without being explicitly programmed.
with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E.
performance criterion using example data or past experience. We need learning in cases where we cannot directly write a computer program to solve a given problem, but need example data or experience.
In statistics, going from particular observations to general descriptions is called inference and learning is called estimation and classification is called discriminant analysis.
CS480/680 Winter 2020 Zahra Sheikhbahaee 8 University of Waterloo
Supervised learning
Ø Classification Ø Regression
Reinforcement learning Unsupervised learning
Ø Clustering Ø reducing dimensionality
CS480/680 Winter 2020 Zahra Sheikhbahaee 9 University of Waterloo
𝑔: ℝ$ ⟶ {1, … , 𝑙}
CS480/680 Winter 2020 Zahra Sheikhbahaee 10 University of Waterloo
regression problems.
CS480/680 Winter 2020 Zahra Sheikhbahaee 11 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 12 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 13 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 14 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 15 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 16 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 17 University of Waterloo
capacity: overfit by memorizing properties of the training set).
in 2N ways as positive and negative and 2N different learning problems can be defined by N data points. If for any of these problems, we can find a hypothesis h ∈ H that separates the positive examples from the negative, then we say H shatters N points. The maximum number of points that can be arranged so that classifier H can shatter them and it is called the Vapnik- Chervonekis (VC) dimension of H, is denoted as VC(H), and measures the capacity of the hypothesis class H.
CS480/680 Winter 2020 Zahra Sheikhbahaee 18 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 19 University of Waterloo
– Clustering: e.g. K-mean clustering
Compressed representation, features, generative model:
CS480/680 Winter 2020 Zahra Sheikhbahaee 20 University of Waterloo
https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/38115.pdf CS480/680 Winter 2020 Zahra Sheikhbahaee 21 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 22
When the output of the system is a sequence of actions. In such a case, a single action is not important; what is important is the policy that is the sequence of correct actions to reach the goal. The reward is a numerical signal which indicates how good actions are.
University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 23
– Game 2 move 37: AlphaGo plays unexpected move (odds 1/10,000)
University of Waterloo
The theories that incorporate constraints on the information processing capacities of an agent are called theories of bounded rationality (Herbert Simon).
without taking into account its limited computational resources.
resources such as time, access to information, capacity for information, and processing power and can only be rational to a certain extent. Agents modeled with unbounded rationality act to maximize utility, while agents modeled with bounded rationality can only aim for some satisfactory amount of utility (a regularized expected utility known as the free energy, where the regularizer is given by the information divergence from a prior to a posterior policy).
University of Waterloo CS480/680 Winter 2020 Zahra Sheikhbahaee 24
CS480/680 Winter 2020 Zahra Sheikhbahaee 25 University of Waterloo
CS480/680 Winter 2020 Zahra Sheikhbahaee 26 University of Waterloo