Intro to ML March 10, 2020 Data Science CSCI 1951A Brown - - PowerPoint PPT Presentation

intro to ml
SMART_READER_LITE
LIVE PREVIEW

Intro to ML March 10, 2020 Data Science CSCI 1951A Brown - - PowerPoint PPT Presentation

Intro to ML March 10, 2020 Data Science CSCI 1951A Brown University Instructor: Ellie Pavlick HTAs: Josh Levin, Diane Mutako, Sol Zitter 1 Announcements This class is going viral! (Funny? No? Too soon?) Not officially, but starting to


slide-1
SLIDE 1

Intro to ML

March 10, 2020 Data Science CSCI 1951A Brown University Instructor: Ellie Pavlick HTAs: Josh Levin, Diane Mutako, Sol Zitter

1

slide-2
SLIDE 2

Announcements

  • This class is going viral! (Funny? No? Too soon?)
  • Not officially, but starting to prep just in case
  • Trial run on Thursday
  • Quizzes and Clickers will remain both valid until

further notice

  • Questions?

2

slide-3
SLIDE 3

Today

  • ML “preliminaries”—terminology, basic building

blocks, conceptual background

  • The two faces of linear regression
  • Training with Stochastic Gradient Descent

3

slide-4
SLIDE 4

Today

  • ML “preliminaries”—terminology, basic building

blocks, conceptual background

  • The two faces of linear regression
  • Training with Stochastic Gradient Descent

4

slide-5
SLIDE 5

Quick Clicker Q!

How much ML experience have you had?

5

(a) None at all. I have obviously heard of ML but I’ve never really dealt with it. (b) Small amount of informal experience. I’ve read articles/blog posts and gotten the gist of how it works. (c) Like (b), but I’ve followed along an coded some models myself (d) Comfortable. I’ve taken an ML class. (e) Very comfortable. I’ve taken an ML class/classes and I’ve built models myself for research projects or internships.

5

slide-6
SLIDE 6

Quick Clicker Q!

Characterize your knowledge of ML:

6

(a) Mostly “conventional” ML (b) Mostly deep learning (c) Equally comfortable with both (d) Not comfortable with either

6

slide-7
SLIDE 7

Oversimplified ML

Goal/Task Data Model

7

slide-8
SLIDE 8

Oversimplified ML

Goal/Task Data Model

Prediction of some kind, e.g.:

  • price of a stock (number)
  • sentiment of a piece of text (discrete label)
  • objects in an image (tagging)
  • strategy for a video game (sequence)
  • parse tree of a sentence (tree structure)

8

slide-9
SLIDE 9

Oversimplified ML

Goal/Task Data Model

Can be anything. Usually data size and/or representation is the limiting factor.

9

slide-10
SLIDE 10

Oversimplified ML

Goal/Task Data Model

Decisions about how the problem is structured AND how to estimate parameters

  • linear/logistic regression
  • SVMs
  • Naive Bayes, Bayesian Networks
  • Neural Networks
  • ….

10

slide-11
SLIDE 11

Goal/Task Data Model

Defining an ML problem

11

slide-12
SLIDE 12

Defining an ML problem

https://youtu.be/bq2_wSsDwkQ?t=682

12

slide-13
SLIDE 13

Defining an ML problem

13

slide-14
SLIDE 14

Data Model

Defining an ML problem

Goal/Task

14

slide-15
SLIDE 15

Data Model

Defining an ML problem

Task = Increase Consumption

15

slide-16
SLIDE 16

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

16

slide-17
SLIDE 17

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

??? ???

17

slide-18
SLIDE 18

Defining an ML problem

  • What is “machine learnable”?
  • Like…basically everything, right? WRONG!! (kind of)
  • Input features need to be concrete and
  • representable. Definition of “success” needs to be

quantifiable (and, right now, usually differentiable).

18

slide-19
SLIDE 19

Defining an ML problem

  • What is “machine learnable”?
  • Like…basically everything, right? WRONG!! (kind of)
  • Input features need to be concrete and
  • representable. Definition of “success” needs to be

quantifiable (and, right now, usually differentiable).

19

slide-20
SLIDE 20

Defining an ML problem

  • What is “machine learnable”?
  • Like…basically everything, right? WRONG!! (kind of)
  • Input features need to be concrete and
  • representable. Definition of “success” needs to be

quantifiable (and, right now, usually differentiable).

20

slide-21
SLIDE 21

Defining an ML problem

  • What is “machine learnable”?
  • Like…basically everything, right? WRONG!! (kind of)
  • Input features need to be concrete and
  • representable. Definition of “success” needs to be

quantifiable (and, right now, usually differentiable).

21

slide-22
SLIDE 22

Defining an ML problem

  • What is “machine learnable”?
  • Like…basically everything, right? WRONG!! (kind of)
  • Input features need to be concrete and
  • representable. Definition of “success” needs to be

quantifiable (and, right now, usually differentiable).

22

slide-23
SLIDE 23

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Objective/Loss Function = ???

23

slide-24
SLIDE 24

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Objective/Loss Function = ??? Features = ???

24

slide-25
SLIDE 25

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Objective/Loss Function = ??? Features = ???

25

slide-26
SLIDE 26

Prediction Target

  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Objective function….ideas?

26

slide-27
SLIDE 27
  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Prediction target….ideas?
  • Objective/Loss Function…ideas?

Prediction Target

27

slide-28
SLIDE 28

Clicker Question!

28 28

slide-29
SLIDE 29
  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Prediction target….ideas?
  • Objective/Loss Function…ideas?
  • Time spent on site (avg. per user/total)
  • Number of users
  • Number of articles read (need to define “read”)
  • Number of articles clicked on
  • Time per article
  • Articles shared…

Prediction Target

29

slide-30
SLIDE 30
  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Prediction target….ideas?
  • Objective/Loss Function…ideas?
  • Time spent on site (avg. per user/total)
  • Number of users
  • Number of articles read (need to define “read”)
  • Number of articles clicked on
  • Time per article
  • Articles shared…

Prediction Target

30

slide-31
SLIDE 31

Clicker Question!

31 31

slide-32
SLIDE 32
  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Prediction target….ideas?
  • Objective/Loss Function…ideas?

Prediction Target

  • Time spent on site (avg. per user/total)
  • Number of users
  • Number of articles read (need to define “read”)
  • Number of articles clicked on
  • Time per article
  • Articles shared…

32

slide-33
SLIDE 33
  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Prediction target….ideas?
  • Objective/Loss Function…ideas?

Prediction Target

  • Time spent on site (avg. per user/total)
  • Number of users
  • Number of articles read (need to define “read”)
  • Number of articles clicked on
  • Time per article
  • Articles shared…
  • Difference between predicted and true value
  • Squared difference between predicted and

true value

  • Predicted probability of true value
  • Whether you were right or wrong (binary)

33

slide-34
SLIDE 34
  • Goal = Increase consumption of “content” NOS for

your clickbait farm pulitzer-prize worthy publication

  • Prediction target….ideas?
  • Objective/Loss Function…ideas?

Prediction Target

  • Time spent on site (avg. per user/total)
  • Number of users
  • Number of articles read (need to define “read”)
  • Number of articles clicked on
  • Time per article
  • Articles shared…
  • Difference between predicted and true value
  • Squared difference between predicted

and true value

  • Predicted probability of true value
  • Whether you were right or wrong (binary)

34

slide-35
SLIDE 35

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Objective/Loss Function = ??? Features = ???

35

slide-36
SLIDE 36

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks Features = ???

36

slide-37
SLIDE 37

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Features = ??? Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

37

slide-38
SLIDE 38

Features

  • Data = Reading habits collected via unauthorized

ever-present cookies and remote control of webcam user-consented GDPR-compliant data usage agreements

  • Features….ideas?

38

slide-39
SLIDE 39

Features

  • Data = Reading habits collected via unauthorized

ever-present cookies and remote control of webcam user-consented GDPR-compliant data usage agreements

  • Features….ideas?

39

slide-40
SLIDE 40

Clicker Question!

40 40

slide-41
SLIDE 41

Features

  • Data = Reading habits collected via unauthorized

ever-present cookies and remote control of webcam user-consented GDPR-compliant data usage agreements

  • Features….ideas?

41

slide-42
SLIDE 42

Features

  • Data = Reading habits collected via unauthorized

ever-present cookies and remote control of webcam user-consented GDPR-compliant data usage agreements

  • Features….ideas?
  • Article topic
  • Recency (minutes since release)
  • Words in title/snippet
  • Presence of photo
  • Reading level
  • Fonts/layouts
  • User location
  • Topics of articles the user has read previously
  • Number of likes

42

slide-43
SLIDE 43

Features

  • Data = Reading habits collected via unauthorized

ever-present cookies and remote control of webcam user-consented GDPR-compliant data usage agreements

  • Features….ideas?
  • Article topic
  • Recency (minutes since release)
  • Words in title/snippet
  • Presence of photo
  • Reading level
  • Fonts/layouts
  • User location
  • Topics of articles the user has read previously
  • Number of likes

43

slide-44
SLIDE 44

Features

  • Recency: Float
  • Words in title: String
  • Presence of photo: Boolean
  • Reading level: Integer

44

slide-45
SLIDE 45

Features

Clicks Recency Reading Level Photo Title 10 1.3 11 1 “New Tax Guidelines” 1000 1.7 3 1 “This 600lb baby…” 1000000 2.4 2 1 “18 reasons you should never look at this cat unless you…” 1 5.9 19 “The Brothers Karamazov: a neo- post-globalist perspective”

45

slide-46
SLIDE 46

Features

Clicks Recency Reading Level Photo Title 10 1.3 11 1 “New Tax Guidelines” 1000 1.7 3 1 “This 600lb baby…” 1000000 2.4 2 1 “18 reasons you should never look at this cat unless you…” 1 5.9 19 “The Brothers Karamazov: a neo- post-globalist perspective”

y

46

slide-47
SLIDE 47

Features

Clicks Recency Reading Level Photo Title 10 1.3 11 1 “New Tax Guidelines” 1000 1.7 3 1 “This 600lb baby…” 1000000 2.4 2 1 “18 reasons you should never look at this cat unless you…” 1 5.9 19 “The Brothers Karamazov: a neo- post-globalist perspective”

x

47

slide-48
SLIDE 48

Features

Clicks Recency Reading Level Photo Title 10 1.3 11 1 “New Tax Guidelines” 1000 1.7 3 1 “This 600lb baby…” 1000000 2.4 2 1 “18 reasons you should never look at this cat unless you…” 1 5.9 19 “The Brothers Karamazov: a neo- post-globalist perspective”

numeric features — defined for (nearly) every row

48

slide-49
SLIDE 49

Features

Clicks Recency Reading Level Photo Title 10 1.3 11 1 “New Tax Guidelines” 1000 1.7 3 1 “This 600lb baby…” 1000000 2.4 2 1 “18 reasons you should never look at this cat unless you…” 1 5.9 19 “The Brothers Karamazov: a neo- post-globalist perspective”

boolean features — 0 or 1 (“dummy” variables)

49

slide-50
SLIDE 50

Features

Clicks Recency Reading Level Photo Title 10 1.3 11 1 “New Tax Guidelines” 1000 1.7 3 1 “This 600lb baby…” 1000000 2.4 2 1 “18 reasons you should never look at this cat unless you…” 1 5.9 19 “The Brothers Karamazov: a neo- post-globalist perspective”

strings = boolean features — 0 or 1 (“dummy” variables)

50

slide-51
SLIDE 51

Features

Clicks Recency Reading Level Photo Title: “new” Title: “tax” Title: “this” Title: “…” … 10 1.3 11 1 1 … 1000 1.7 3 1 1 1 … 1000000 2.4 2 1 1 1 … 1 5.9 19 …

strings = boolean features — 0 or 1 (“dummy” variables)

51

slide-52
SLIDE 52

Features

Clicks Recency Reading Level Photo Title: “new” Title: “tax” Title: “this” Title: “…” … 10 1.3 11 1 1 … 1000 1.7 3 1 1 1 … 1000000 2.4 2 1 1 1 … 1 5.9 19 …

“sparse features” — 0 for most rows

52

slide-53
SLIDE 53

Clicker Question!

53 53

slide-54
SLIDE 54

Clicker Question!

For the problem set up, how many features will there be? I.e. how many columns in our X matrix, (not including Y)?

(a)112,000 (b)5 (c) 27 (d)110,000

54

Y: happiness X1: day of week (“monday”, “tuesday”, … “sunday”) X2: bank account balance (real value) X3: breakfast (yes,no) X4: whether you have found your inner peace (yes,no) X5: words from last week’s worth of tweets (assuming tweets are at most 15 words long and there are 100K words in the English vocabulary)

54

slide-55
SLIDE 55

Clicker Question!

For the problem set up, how many features will there be? I.e. how many columns in our X matrix, (not including Y)?

(a)100,012 (b)5

Y: happiness X1: day of week (“monday”, “tuesday”, … “sunday”) X2: bank account balance (real value) X3: breakfast (yes,no) X4: whether you have found your inner peace (yes,no) X5: words from last week’s worth of tweets (assuming tweets are at most 15 words long and there are 100K words in the English vocabulary)

(c) 27 (d)100,010

55

7 1 1 1 100,000

55

slide-56
SLIDE 56

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Features = ??? Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

56

slide-57
SLIDE 57

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Features = {Recency:float, ReadingLevel:Int, Photo:Bool, Title_New:Bool, Title_Tax:Bool, …} Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

57

slide-58
SLIDE 58

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Features = {Recency:float, ReadingLevel:Int, Photo:Bool, Title_New:Bool, Title_Tax:Bool, …} Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

58

slide-59
SLIDE 59

ML = Function Approximation

Model

59

slide-60
SLIDE 60

ML = Function Approximation

Model

60

slide-61
SLIDE 61

ML = Function Approximation

Model

61

slide-62
SLIDE 62

ML = Function Approximation

Model

62

slide-63
SLIDE 63

ML = Function Approximation

Model

63

slide-64
SLIDE 64

ML = Function Approximation

You define inputs and outputs.

Model

64

slide-65
SLIDE 65

ML = Function Approximation

You define inputs and outputs. (The really hard part)

Model

65

slide-66
SLIDE 66

ML = Function Approximation

The machine will (ideally) learn the function (with a lot of help from you)

Model

66

slide-67
SLIDE 67

ML = Function Approximation

The machine will (ideally) learn the function (with a lot of help from you) (The part that gets the most attention.)

Model

67

slide-68
SLIDE 68

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1

68

slide-69
SLIDE 69

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1

69

slide-70
SLIDE 70

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1

70

slide-71
SLIDE 71

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1

71

slide-72
SLIDE 72

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1

72

slide-73
SLIDE 73

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1 # 2

73

slide-74
SLIDE 74

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1 # 2

74

slide-75
SLIDE 75

Model

clicks reading level

75

slide-76
SLIDE 76

Model

Regression: continuous (infinite) output f(reading level) = # of clicks clicks reading level

76

slide-77
SLIDE 77

Model

Classification: discrete (finite) output f(reading level) = {clicked, not clicked} clicks reading level

77

slide-78
SLIDE 78

Model

clicks = m(reading_level) + b m = -2.4

clicks reading level

Linear Regression —> The specific “model” we are using here.

78

slide-79
SLIDE 79

Model

clicks = m(reading_level) + b m = -2.4

clicks reading level

clicks —> output/labels/target

79

slide-80
SLIDE 80

Model

clicks = m(reading_level) + b m = -2.4

clicks reading level

reading level —> The “feature” which is observed/derived from the data

80

slide-81
SLIDE 81

Model

clicks = m(reading_level) + b m = -2.4

clicks reading level

m and b —> The “parameters” which need to be set (by looking at data)

81

slide-82
SLIDE 82

Model

clicks = m(reading_level) + b m = cov(rl, c)/var(rl)

clicks reading level

“setting parameters”, “learning”, “training”, “estimation”

82

slide-83
SLIDE 83

Model

clicks = m(reading_level) + b m = -2.4

clicks reading level

parameter values, “weights”, “coefficients”

83

slide-84
SLIDE 84

Data = Reading Habits Model

Defining an ML problem

Task = Increase Consumption

Features = {Recency:float, ReadingLevel:Int, Photo:Bool, Title_New:Bool, Title_Tax:Bool, …} Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks Linear Regression

84

slide-85
SLIDE 85

Features = {Recency:float, ReadingLevel:Int, Photo:Bool, Title_New:Bool, Title_Tax:Bool, …}

Linear Regression

Defining an ML problem

Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

85

slide-86
SLIDE 86

Features = {Recency:float, ReadingLevel:Int, Photo:Bool, Title_New:Bool, Title_Tax:Bool, …}

Linear Regression

Defining an ML problem

Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

86

slide-87
SLIDE 87

Features = {Recency:float, ReadingLevel:Int, Photo:Bool, Title_New:Bool, Title_Tax:Bool, …}

Linear Regression

Defining an ML problem

Objective/Loss Function = squared difference between predicted total number of clicks and actual total number of clicks

Soooo….how do I know if my model is good?

87

slide-88
SLIDE 88

Train/Test Splits

88

slide-89
SLIDE 89

Train/Test Splits

MSE = 10

89

slide-90
SLIDE 90

Train/Test Splits

MSE = 10

90

slide-91
SLIDE 91

Train/Test Splits

Train

91

slide-92
SLIDE 92

Train/Test Splits

Train MSE = 6

92

slide-93
SLIDE 93

Train/Test Splits

Test

93

slide-94
SLIDE 94

Clicker Question!

94

slide-95
SLIDE 95

Test

Clicker Question! (a) Go up (b) Go down (c) Stay the same (modulo random variation)

What should we expect MSE to do?

95

slide-96
SLIDE 96

Test

Clicker Question! (a) Go up (b) Go down (c) Stay the same (modulo random variation)

What should we expect MSE to do? If your model isn’ t “right” yet (i.e. in practice, most

  • f the time)

96

slide-97
SLIDE 97

Test

Clicker Question! (a) Go up (b) Go down (c) Stay the same (modulo random variation)

What should we expect MSE to do? If your model is “right”

  • r is not yet powerful

enough (i.e. can’ t memorize training data).

97

slide-98
SLIDE 98

Train/Test Splits

Test

98

slide-99
SLIDE 99

Train/Test Splits

Test MSE = 12

99

slide-100
SLIDE 100

Train/Test Splits

Train MSE = 4

Problem gets worse as models get more powerful/flexible

100

slide-101
SLIDE 101

Train/Test Splits

MSE = 14

Problem gets worse as models get more powerful/flexible

101

slide-102
SLIDE 102

102

slide-103
SLIDE 103

Today

  • ML “preliminaries”—terminology, basic building

blocks, conceptual background

  • The two faces of linear regression
  • Training with Stochastic Gradient Descent

103

slide-104
SLIDE 104

Regression Analysis in Stats Regression in ML

104

slide-105
SLIDE 105

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y

105

slide-106
SLIDE 106

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs

106

slide-107
SLIDE 107

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity

107

slide-108
SLIDE 108

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine

108

slide-109
SLIDE 109

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size

109

slide-110
SLIDE 110

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set

110

slide-111
SLIDE 111

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values

111

slide-112
SLIDE 112

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

112

slide-113
SLIDE 113

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

But! These are the same model. These difference are “in general”/“by convention”, not anything fundamental.

113

slide-114
SLIDE 114

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

Different scientific communities with different goals.

114

slide-115
SLIDE 115

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

Different scientific communities with different goals. (and different software packages :)) <- R, stats_models, STATA sklearn, matlab, pytorch ->

115

slide-116
SLIDE 116

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

116

slide-117
SLIDE 117

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

In the limit, I think these goals are the same. Even if we care about prediction (and we want to do it using as few models as possible), shouldn’t we get the best performance by modeling the “true” underlying process? Isn’t it the case that correct explanatory/causal models necessarily make right predictions, but not vice-versa?

117

slide-118
SLIDE 118

Regression Analysis in Stats Regression in ML

Make claims about whether there is a meaningful relationship between X and Y Given X, predict Y; deploy a model to make predictions for new inputs (Often) interested in causation; focus on controls and removing colinearity Focused on prediction accuracy; exploiting correlation is totally fine A “result” is typically in the form of a significant relationship and/or practically relevant effect size A “result” is typically in the form

  • f an improvement in prediction

performance on a (held out) test set Avoid overfitting by preferring simple models; avoid

  • verclaiming by accounting

for “degrees of freedom” when computing p values Avoid overfitting through regularization; avoid

  • verclaiming by maintaining

train/test splits and reporting test performance

Counter argument: You can get perfect* predictive performance with the wrong model. We were extremely good at predicting whether objects would fall or float long before we knew about gravity. Explanatory/causal models are hard! We might never get

  • there. Maybe empirically accurate predictions

should lead, and theory/explanation will follow?

118

slide-119
SLIDE 119

Today

  • ML “preliminaries”—terminology, basic building

blocks, conceptual background

  • The two faces of linear regression
  • Training with Stochastic Gradient Descent

119

slide-120
SLIDE 120

Model

  • Make assumptions about the problem domain.
  • How is the data generated?
  • How is the decision-making procedure structured?
  • What types of dependencies exist?
  • Trending buzzword: “inductive biases”
  • How to train the model?

# 1 # 2

120

slide-121
SLIDE 121

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize

121

slide-122
SLIDE 122

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize

122

slide-123
SLIDE 123

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize

123

slide-124
SLIDE 124

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize Cov(X, Y ) V ar(X)

<latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit><latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit><latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit><latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit>

m =

= ¯ Y − m ¯ X

<latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit><latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit><latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit><latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit>

b =

124

slide-125
SLIDE 125

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize

125

slide-126
SLIDE 126

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize

∂Q ∂m =

n

X

i=1

−2Xi(Yi − b − mXi) = 0

<latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit>

126

slide-127
SLIDE 127

Training with Gradient Descent

=

n

X

i=1

(Yi − ˆ Y )2

<latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit><latexit sha1_base64="r/S0kziZA4N57u3eiAn1RfFZ4w=">ACDHicbVDLSsNAFJ3UV62vqks3g0WoC0tSBN0Uim5ctmBfNGmYTCft0MkzEyEvIBbvwVNy4UcesHuPNvnLZaOuBgcM53LnHi9iVCrT/DZya+sbm1v57cLO7t7+QfHwqC3DWGDSwiELRdDkjDKSUtRxUg3EgQFHiMdb3I78zsPREga8ns1jYgToBGnPsVIacktlpqwBm0ZB25Ca1Y6SHha7rkUXkB7jFTS8HVZ0yK+YcJVYGSmBDA23+GUPQxwHhCvMkJR9y4yUkyChKGYkLdixJBHCEzQifU05Coh0kvkxKTzTyhD6odCPKzhXf08kKJByGng6GSA1lsveTPzP68fKv3YSyqNYEY4Xi/yYQRXCWTNwSAXBik01QVhQ/VeIx0grHR/BV2CtXzyKmlXK5ZsZqXpfpNVkcenIBTUAYWuAJ1cAcaoAUweATP4BW8GU/Gi/FufCyiOSObOQZ/YHz+ADvfmdM=</latexit>

minimize

∂Q ∂m =

n

X

i=1

−2Xi(Yi − b − mXi) = 0

<latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit>

127

slide-128
SLIDE 128

Training with Gradient Descent

https://independentseminarblog.com/2018/01/12/moving-below-the-surface-3-gradient-descent-william/ 128

slide-129
SLIDE 129

Training with Gradient Descent

https://independentseminarblog.com/2018/01/12/moving-below-the-surface-3-gradient-descent-william/

∂Q ∂b =

n

X

i=1

−2(Yi − mXi − b) = 0

<latexit sha1_base64="9cIJCpfvS+XsZ9OkPHz9MbmKXfE=">ACLnicbVBdS8MwFE3n15xfVR9CQ5hPjaIeiLIrg4wZuTtZa0ix1YUlaklQYpb/IF/+KPgq4qs/w3QO1OmBkM593LvPWHCqNKO82yVZmbn5hfKi5Wl5ZXVNXt9o6PiVGLSxjGLZTdEijAqSFtTzUg3kQTxkJHLcHha+Je3RCoaiws9SojP0Y2gEcVIGymwz7xIpx5CZKaIgZb+TcPc3gEPZXyIKNHbn6diRzuNWpXAYV7kHfHX7hrapzArjp1Zwz4l7gTUgUTNAP70evHOVEaMyQUj3XSbSfFYMxI3nFSxVJEB6iG9IzVCBOlJ+Nz83hjlH6MIqleULDsfqzI0NcqRE36+9wpAdq2ivE/7xeqNDP6MiSTUR+GtQlDKoY1hkB/tUEqzZyBCEJTW7QjxAJj9tEq6YENzpk/+STqPuOnW3tV89PpnEUQZbYBvUgAsOwDE4B03QBhjcgQfwAl6te+vJerPev0pL1qRnE/yC9fEJufumUQ=</latexit><latexit sha1_base64="9cIJCpfvS+XsZ9OkPHz9MbmKXfE=">ACLnicbVBdS8MwFE3n15xfVR9CQ5hPjaIeiLIrg4wZuTtZa0ix1YUlaklQYpb/IF/+KPgq4qs/w3QO1OmBkM593LvPWHCqNKO82yVZmbn5hfKi5Wl5ZXVNXt9o6PiVGLSxjGLZTdEijAqSFtTzUg3kQTxkJHLcHha+Je3RCoaiws9SojP0Y2gEcVIGymwz7xIpx5CZKaIgZb+TcPc3gEPZXyIKNHbn6diRzuNWpXAYV7kHfHX7hrapzArjp1Zwz4l7gTUgUTNAP70evHOVEaMyQUj3XSbSfFYMxI3nFSxVJEB6iG9IzVCBOlJ+Nz83hjlH6MIqleULDsfqzI0NcqRE36+9wpAdq2ivE/7xeqNDP6MiSTUR+GtQlDKoY1hkB/tUEqzZyBCEJTW7QjxAJj9tEq6YENzpk/+STqPuOnW3tV89PpnEUQZbYBvUgAsOwDE4B03QBhjcgQfwAl6te+vJerPev0pL1qRnE/yC9fEJufumUQ=</latexit><latexit sha1_base64="9cIJCpfvS+XsZ9OkPHz9MbmKXfE=">ACLnicbVBdS8MwFE3n15xfVR9CQ5hPjaIeiLIrg4wZuTtZa0ix1YUlaklQYpb/IF/+KPgq4qs/w3QO1OmBkM593LvPWHCqNKO82yVZmbn5hfKi5Wl5ZXVNXt9o6PiVGLSxjGLZTdEijAqSFtTzUg3kQTxkJHLcHha+Je3RCoaiws9SojP0Y2gEcVIGymwz7xIpx5CZKaIgZb+TcPc3gEPZXyIKNHbn6diRzuNWpXAYV7kHfHX7hrapzArjp1Zwz4l7gTUgUTNAP70evHOVEaMyQUj3XSbSfFYMxI3nFSxVJEB6iG9IzVCBOlJ+Nz83hjlH6MIqleULDsfqzI0NcqRE36+9wpAdq2ivE/7xeqNDP6MiSTUR+GtQlDKoY1hkB/tUEqzZyBCEJTW7QjxAJj9tEq6YENzpk/+STqPuOnW3tV89PpnEUQZbYBvUgAsOwDE4B03QBhjcgQfwAl6te+vJerPev0pL1qRnE/yC9fEJufumUQ=</latexit><latexit sha1_base64="9cIJCpfvS+XsZ9OkPHz9MbmKXfE=">ACLnicbVBdS8MwFE3n15xfVR9CQ5hPjaIeiLIrg4wZuTtZa0ix1YUlaklQYpb/IF/+KPgq4qs/w3QO1OmBkM593LvPWHCqNKO82yVZmbn5hfKi5Wl5ZXVNXt9o6PiVGLSxjGLZTdEijAqSFtTzUg3kQTxkJHLcHha+Je3RCoaiws9SojP0Y2gEcVIGymwz7xIpx5CZKaIgZb+TcPc3gEPZXyIKNHbn6diRzuNWpXAYV7kHfHX7hrapzArjp1Zwz4l7gTUgUTNAP70evHOVEaMyQUj3XSbSfFYMxI3nFSxVJEB6iG9IzVCBOlJ+Nz83hjlH6MIqleULDsfqzI0NcqRE36+9wpAdq2ivE/7xeqNDP6MiSTUR+GtQlDKoY1hkB/tUEqzZyBCEJTW7QjxAJj9tEq6YENzpk/+STqPuOnW3tV89PpnEUQZbYBvUgAsOwDE4B03QBhjcgQfwAl6te+vJerPev0pL1qRnE/yC9fEJufumUQ=</latexit>

Q =

n

X

i=1

(Yi − (mXi + b))2

<latexit sha1_base64="q3o/vVOTz7GUx+RiBjGcRGq1f0o=">ACD3icbVDLSgMxFM3UV62vUZdugkVpEctMEXRTKLpx2YJ9SB9DJk3b0CQzJBmhDP0DN/6KGxeKuHXrzr8xbWehrQcu93DOvST3+CGjSjvOt5VaWV1b30hvZra2d3b37P2DugoiUkNByQTR8pwqgNU01I81QEsR9Rhr+6GbqNx6IVDQd3ockg5HA0H7FCNtJM8+rcISbKuIezEtuZNuLCa5e4/Cc5jTdPoJ/Pd4uenXUKzgxwmbgJyYIEFc/+avcCHEiNGZIqZbrhLoTI6kpZmSaUeKhAiP0IC0DBWIE9WJZ/dM4IlRerAfSFNCw5n6eyNGXKkx980kR3qoFr2p+J/XinT/qhNTEUaCDx/qB8xqAM4DQf2qCRYs7EhCEtq/grxEmEtYkwY0JwF09eJvViwXUKbvUiW75O4kiDI3AMcsAFl6AMbkEF1AGj+AZvI368l6sd6tj/loykp2DsEfWJ8/aLmZsg=</latexit><latexit sha1_base64="q3o/vVOTz7GUx+RiBjGcRGq1f0o=">ACD3icbVDLSgMxFM3UV62vUZdugkVpEctMEXRTKLpx2YJ9SB9DJk3b0CQzJBmhDP0DN/6KGxeKuHXrzr8xbWehrQcu93DOvST3+CGjSjvOt5VaWV1b30hvZra2d3b37P2DugoiUkNByQTR8pwqgNU01I81QEsR9Rhr+6GbqNx6IVDQd3ockg5HA0H7FCNtJM8+rcISbKuIezEtuZNuLCa5e4/Cc5jTdPoJ/Pd4uenXUKzgxwmbgJyYIEFc/+avcCHEiNGZIqZbrhLoTI6kpZmSaUeKhAiP0IC0DBWIE9WJZ/dM4IlRerAfSFNCw5n6eyNGXKkx980kR3qoFr2p+J/XinT/qhNTEUaCDx/qB8xqAM4DQf2qCRYs7EhCEtq/grxEmEtYkwY0JwF09eJvViwXUKbvUiW75O4kiDI3AMcsAFl6AMbkEF1AGj+AZvI368l6sd6tj/loykp2DsEfWJ8/aLmZsg=</latexit><latexit sha1_base64="q3o/vVOTz7GUx+RiBjGcRGq1f0o=">ACD3icbVDLSgMxFM3UV62vUZdugkVpEctMEXRTKLpx2YJ9SB9DJk3b0CQzJBmhDP0DN/6KGxeKuHXrzr8xbWehrQcu93DOvST3+CGjSjvOt5VaWV1b30hvZra2d3b37P2DugoiUkNByQTR8pwqgNU01I81QEsR9Rhr+6GbqNx6IVDQd3ockg5HA0H7FCNtJM8+rcISbKuIezEtuZNuLCa5e4/Cc5jTdPoJ/Pd4uenXUKzgxwmbgJyYIEFc/+avcCHEiNGZIqZbrhLoTI6kpZmSaUeKhAiP0IC0DBWIE9WJZ/dM4IlRerAfSFNCw5n6eyNGXKkx980kR3qoFr2p+J/XinT/qhNTEUaCDx/qB8xqAM4DQf2qCRYs7EhCEtq/grxEmEtYkwY0JwF09eJvViwXUKbvUiW75O4kiDI3AMcsAFl6AMbkEF1AGj+AZvI368l6sd6tj/loykp2DsEfWJ8/aLmZsg=</latexit><latexit sha1_base64="q3o/vVOTz7GUx+RiBjGcRGq1f0o=">ACD3icbVDLSgMxFM3UV62vUZdugkVpEctMEXRTKLpx2YJ9SB9DJk3b0CQzJBmhDP0DN/6KGxeKuHXrzr8xbWehrQcu93DOvST3+CGjSjvOt5VaWV1b30hvZra2d3b37P2DugoiUkNByQTR8pwqgNU01I81QEsR9Rhr+6GbqNx6IVDQd3ockg5HA0H7FCNtJM8+rcISbKuIezEtuZNuLCa5e4/Cc5jTdPoJ/Pd4uenXUKzgxwmbgJyYIEFc/+avcCHEiNGZIqZbrhLoTI6kpZmSaUeKhAiP0IC0DBWIE9WJZ/dM4IlRerAfSFNCw5n6eyNGXKkx980kR3qoFr2p+J/XinT/qhNTEUaCDx/qB8xqAM4DQf2qCRYs7EhCEtq/grxEmEtYkwY0JwF09eJvViwXUKbvUiW75O4kiDI3AMcsAFl6AMbkEF1AGj+AZvI368l6sd6tj/loykp2DsEfWJ8/aLmZsg=</latexit>

∂Q ∂m =

n

X

i=1

−2Xi(Yi − b − mXi) = 0

<latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit><latexit sha1_base64="XMJx4vTyfUZ86d4j9mLjvojhdw=">ACMXicbVDLSgMxFM34rPVdekmWIS6aJkpgm4KRTdtmAf0qlDJs20oUlmSDJCGeaX3Pgn4qYLRdz6E6YPUFsPXDg515u7vEjRpW27Ym1tr6xubWd2cnu7u0fHOaOjlsqjCUmTRyUHZ8pAijgjQ1Yx0IkQ9xlp+6Pbqd9+JFLRUNzpcUR6HA0EDShG2kheruYGEuHEjZDUFDHYSH84T2EFuirmXkIrTvqQiBQWyx2PFu49CovQN8XN8K02V4ub5fsGeAqcRYkDxaoe7kXtx/imBOhMUNKdR070r1kuhszkmbdWJEI4REakK6hAnGiesns4hSeG6UPg1CaEhrO1N8TCeJKjblvOjnSQ7XsTcX/vG6sg+teQkUayLwfFEQM6hDOI0P9qkWLOxIQhLav4K8RCZCLUJOWtCcJZPXiWtcsmxS07jMl+9WcSRAafgDBSA65AFdRAHTQBk/gFbyBd+vZmlgf1ue8dc1azJyAP7C+vgE9U6ea</latexit>

Helpful equations for following along in the jupyter notebook = Cov(X, Y ) V ar(X)

<latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit><latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit><latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit><latexit sha1_base64="WPNm1YuVK0X591E2DeKZG6XKYA=">ACB3icbVBNS8NAEN3Urxq/oh4FWSxC1ISEfQiFHvxWMG2kTaUzXbTLt1swu6mUEJuXvwrXjwo4tW/4M1/47bNQVsfDzem2Fmnh8zKpVtfxuFldW19Y3iprm1vbO7Z+0ftGSUCEyaOGKRcH0kCaOcNBVjLixICj0GWn7o/rUb4+JkDTi92oSEy9EA04DipHSUs869uE17AYC4dSsR+OyewYfKlqtpAou5WsZ5Xsqj0DXCZOTkogR6NnfX7EU5CwhVmSMqOY8fKS5FQFDOSmd1EkhjhERqQjqYchUR6eyPDJ5qpQ+DSOjiCs7U3xMpCqWchL7uDJEaykVvKv7ndRIVXHkp5XGiCMfzRUHCoIrgNBTYp4JgxSaICyovhXiIdKhKB2dqUNwFl9eJq3zqmNXnbuLUu0mj6MIjsAJKAMHXIauAUN0AQYPIJn8ArejCfjxXg3PuatBSOfOQR/YHz+ADg1lvo=</latexit>

m =

<latexit sha1_base64="mxdaQB8GmfgkBu9RKA5ZbS6fR8=">ACkHicdVHdSsMwGE3r/ybeulNcAjbxUY7Bb1wOPVGvJrgdLJuJc1SDSZpSVJhlD6P7+Odb2O6daCbfhA4Od/5TpKTIGZUacf5suyl5ZXVtfWN0ubW9s5ueW/UWJxKSLIxbJXoAUYVSQrqakV4sCeIBI0/B203ef3onUtFIPOhxTAYcvQgaUoy0ofzyB4ct6IUS4dRTCfdT2nKzYSoyWH32ac+nsA69AMn0OTObWjavMuSwOdP0ck32j2E1NzOetToUhbowXjCtTlxr9Zlu2Mz8csVpOJOCi8AtQAU1fHLn94owgknQmOGlOq7TqwHKZKaYkaykpcoEiP8hl5I30CBOFGDdBJoBo8NM4JhJM0SGk7YnxMp4kqNeWCUHOlXNd/Lyb96/USH54OUijRODpQWHCoI5g/jtwRCXBmo0NQFhSc1eIX5EJU5s/LJkQ3PknL4LHZsN1Gu79aV9XcSxDg7BEagCF5yBNrgFHdAF2Nq2TqwLq2Xv2+f2pX01ldpWMXMAfpV9w3+Zca5</latexit><latexit sha1_base64="mxdaQB8GmfgkBu9RKA5ZbS6fR8=">ACkHicdVHdSsMwGE3r/ybeulNcAjbxUY7Bb1wOPVGvJrgdLJuJc1SDSZpSVJhlD6P7+Odb2O6daCbfhA4Od/5TpKTIGZUacf5suyl5ZXVtfWN0ubW9s5ueW/UWJxKSLIxbJXoAUYVSQrqakV4sCeIBI0/B203ef3onUtFIPOhxTAYcvQgaUoy0ofzyB4ct6IUS4dRTCfdT2nKzYSoyWH32ac+nsA69AMn0OTObWjavMuSwOdP0ck32j2E1NzOetToUhbowXjCtTlxr9Zlu2Mz8csVpOJOCi8AtQAU1fHLn94owgknQmOGlOq7TqwHKZKaYkaykpcoEiP8hl5I30CBOFGDdBJoBo8NM4JhJM0SGk7YnxMp4kqNeWCUHOlXNd/Lyb96/USH54OUijRODpQWHCoI5g/jtwRCXBmo0NQFhSc1eIX5EJU5s/LJkQ3PknL4LHZsN1Gu79aV9XcSxDg7BEagCF5yBNrgFHdAF2Nq2TqwLq2Xv2+f2pX01ldpWMXMAfpV9w3+Zca5</latexit><latexit sha1_base64="mxdaQB8GmfgkBu9RKA5ZbS6fR8=">ACkHicdVHdSsMwGE3r/ybeulNcAjbxUY7Bb1wOPVGvJrgdLJuJc1SDSZpSVJhlD6P7+Odb2O6daCbfhA4Od/5TpKTIGZUacf5suyl5ZXVtfWN0ubW9s5ueW/UWJxKSLIxbJXoAUYVSQrqakV4sCeIBI0/B203ef3onUtFIPOhxTAYcvQgaUoy0ofzyB4ct6IUS4dRTCfdT2nKzYSoyWH32ac+nsA69AMn0OTObWjavMuSwOdP0ck32j2E1NzOetToUhbowXjCtTlxr9Zlu2Mz8csVpOJOCi8AtQAU1fHLn94owgknQmOGlOq7TqwHKZKaYkaykpcoEiP8hl5I30CBOFGDdBJoBo8NM4JhJM0SGk7YnxMp4kqNeWCUHOlXNd/Lyb96/USH54OUijRODpQWHCoI5g/jtwRCXBmo0NQFhSc1eIX5EJU5s/LJkQ3PknL4LHZsN1Gu79aV9XcSxDg7BEagCF5yBNrgFHdAF2Nq2TqwLq2Xv2+f2pX01ldpWMXMAfpV9w3+Zca5</latexit><latexit sha1_base64="mxdaQB8GmfgkBu9RKA5ZbS6fR8=">ACkHicdVHdSsMwGE3r/ybeulNcAjbxUY7Bb1wOPVGvJrgdLJuJc1SDSZpSVJhlD6P7+Odb2O6daCbfhA4Od/5TpKTIGZUacf5suyl5ZXVtfWN0ubW9s5ueW/UWJxKSLIxbJXoAUYVSQrqakV4sCeIBI0/B203ef3onUtFIPOhxTAYcvQgaUoy0ofzyB4ct6IUS4dRTCfdT2nKzYSoyWH32ac+nsA69AMn0OTObWjavMuSwOdP0ck32j2E1NzOetToUhbowXjCtTlxr9Zlu2Mz8csVpOJOCi8AtQAU1fHLn94owgknQmOGlOq7TqwHKZKaYkaykpcoEiP8hl5I30CBOFGDdBJoBo8NM4JhJM0SGk7YnxMp4kqNeWCUHOlXNd/Lyb96/USH54OUijRODpQWHCoI5g/jtwRCXBmo0NQFhSc1eIX5EJU5s/LJkQ3PknL4LHZsN1Gu79aV9XcSxDg7BEagCF5yBNrgFHdAF2Nq2TqwLq2Xv2+f2pX01ldpWMXMAfpV9w3+Zca5</latexit>

b = ¯ Y − m ¯ X

<latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit><latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit><latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit><latexit sha1_base64="hwHBCHxlXmNAb1i+1nKjhGPDZ7s=">ACAHicbZDLSgMxFIbP1Fut1EXLtwEi+DGMiOCboSiG5cV7EXaoWTSTBuaZIYkI5RhNr6KGxeKuPUx3Pk2peFtv4Q+PjPOZycP0w408bzvp3C0vLK6lpxvbSxubW94+7uNXScKkLrJOaxaoVYU84krRtmOG0limIRctoMhzfjevORKs1ieW9GCQ0E7ksWMYKNtbruQYiuUCfEKnvI0SkSU27lXbfsVbyJ0CL4MyjDTLWu+9XpxSQVBrCsdZt30tMkGFlGOE0L3VSTRNMhrhP2xYlFlQH2eSAHB1bp4eiWNknDZq4vycyLQeidB2CmwGer42Nv+rtVMTXQYZk0lqCTRVHKkYnROA3UY4oSw0cWMFHM/hWRAVaYGJtZyYbgz5+8CI2ziu9V/LvzcvV6FkcRDuEITsCHC6jCLdSgDgRyeIZXeHOenBfn3fmYthac2cw+/JHz+QOthJUt</latexit>

129

slide-130
SLIDE 130

👌

130