CS 559: Machine Learning Fundamentals and Applications 8 th Set of - - PowerPoint PPT Presentation

cs 559 machine learning fundamentals and applications 8
SMART_READER_LITE
LIVE PREVIEW

CS 559: Machine Learning Fundamentals and Applications 8 th Set of - - PowerPoint PPT Presentation

1 CS 559: Machine Learning Fundamentals and Applications 8 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Project Proposal Dataset How many


slide-1
SLIDE 1

CS 559: Machine Learning Fundamentals and Applications 8th Set of Notes

Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215

1

slide-2
SLIDE 2

Project Proposal

  • Dataset

– How many instances

  • Classes (or what is being predicted)
  • Inputs

– Include feature extraction, if needed – If your inputs are images or financial data, this must be addressed

  • Methods

– At least one simple classifier (MLE with Gaussian model, Naïve Bayes, kNN) – At least one advanced classifier (SVM, Boosting, Random Forest, CNN)

2

slide-3
SLIDE 3

Project Proposal

  • Typical experiments

– Measure benefits due to advanced classifier compared to simple classifier – Compare different classifier settings

  • k in kNN
  • Different SVM kernels
  • AdaBoost vs. cascade
  • Different CNN architectures

– Measure effects of amount of training data available – Evaluate accuracy as a function of the degree of dimensionality reduction using PCA

3

slide-4
SLIDE 4

Project Proposal

  • Email me a pdf with all these
  • I must say “approved” in my response,
  • therwise address my comments and

resubmit

4

slide-5
SLIDE 5

Overview

  • Linear Regression

– Barber Ch. 17 – HTF Ch. 3

5

slide-6
SLIDE 6

Simple Linear Regression

  • How does a single variable of interest

relate to another (single) variable?

– Y = outcome variable (response, dependent...) – X = explanatory variable (predictor, feature, independent...)

  • Data: n pairs of continuous observations

(X1,Y1) … (Xn,Yn)

6

slide-7
SLIDE 7

Example

  • How does systolic blood pressure (SBP) relate to age?
  • Graph suggests that Y relates to X in an approximately

linear way

7

slide-8
SLIDE 8

Regression: Step by Step

  • 1. Assume a linear model: Y = β0 + β1 X
  • 2. Find the line which “best” fits the data, i.e.

estimate parameters β0 and β1

  • 3. Does variation in X help describe

variation in Y ?

  • 4. Check assumptions of model
  • 5. Draw inferences and make predictions

8

slide-9
SLIDE 9

Straight-line Plots

9

slide-10
SLIDE 10

Assumptions of Linear Regression

  • Five basic assumptions
  • 1. Existence: for each fixed value of X, Y is

a random variable with finite mean and variance

  • 2. Independence: the set of Yi are

independent random variables given Xi

10

slide-11
SLIDE 11

Assumptions of Linear Regression

  • 3. Linearity: the mean value of Y is a linear

function of X

|

  • 11
slide-12
SLIDE 12

Assumptions of Linear Regression

  • 4. Homoscedasticity: the variance of Y is the

same for any X

  • 5. Normality: For each fixed value of X, Y

has a normal distribution (by assumption 4, σ2 does not depend on X)

12

slide-13
SLIDE 13

Formulation

  • Yi are linear function of Xi plus some random

error

13

slide-14
SLIDE 14

Linear Regression

14

slide-15
SLIDE 15

Estimating β0 and β1

  • Find “best” line
  • Criterion for “best”: estimate β0 and β1 to

minimize:

  • This is the residual sum of squares, sum of

squares due to error, or sum of squares about regression line

  • Least Squares estimator

15

slide-16
SLIDE 16

Rationale for LS Estimates

  • 2 measures the “deviance” of Yi from the

estimated model

  • The “best” model is the one from which the data

deviate the least

16

slide-17
SLIDE 17

Least Squares Estimators

  • Taking derivatives with respect to β, we obtain
  • The residual variance is

17

slide-18
SLIDE 18

Example: SBP/age data

18

slide-19
SLIDE 19

Using the Model

  • Using the parameter estimates, our best guess

for any Y given X is

  • Hence at
  • Every regression line goes through ( ,

)

  • Also

19

slide-20
SLIDE 20

Correlation and Regression Coefficient

20

slide-21
SLIDE 21

Example

21

slide-22
SLIDE 22

Example

22

slide-23
SLIDE 23

Example

23

slide-24
SLIDE 24

Example

24