Empirical Confidence Models for Supervised Machine Learning - - PowerPoint PPT Presentation

empirical confidence models for supervised machine
SMART_READER_LITE
LIVE PREVIEW

Empirical Confidence Models for Supervised Machine Learning - - PowerPoint PPT Presentation

Empirical Confidence Models for Supervised Machine Learning Margarita Castro 1 , Meinolf Sellmann 2 , Zhaoyuan Yang 2 , Nurali Virani 2 1 University of Toronto, Mechanical and Industrial Engineering 2 General Electric, Global Research Center May,


slide-1
SLIDE 1

Empirical Confidence Models for Supervised Machine Learning

Margarita Castro1, Meinolf Sellmann2, Zhaoyuan Yang2, Nurali Virani2

1 University of Toronto, Mechanical and Industrial Engineering 2 General Electric, Global Research Center

May, 2020

slide-2
SLIDE 2

Motivation

ML in high-stake context Main Issues:

u We can’t expect the models to be

perfect.

u Summarize statistics (e.g.,

accuracy) can be misleading to assess a specific prediction.

2 Self-Driving Cars Healthcare diagnosis Cyber security

slide-3
SLIDE 3

Empirical Confidence for Regression

We propose:

A model that can declare its

  • wn incompetence.

“We develop techniques that learn when models generated by certain learning techniques on a particular data set can be expected to perform well, and when not.” 3

𝑌 Run time instance

Regression Model + Competence Assessor

𝑍′ Prediction 𝐷 Competence Level Trusted, Cautioned

  • r Not Trusted
slide-4
SLIDE 4

Outline of the Talk

Part 1: Competence Assessor

u Overall framework. u Meta-features. u Meta Training Data.

Part 2: Numerical Evaluation

u Experimental Setting. u Results. u Conclusions.

4

slide-5
SLIDE 5

Empirical Competence Assessor

PART 1 5

slide-6
SLIDE 6

Competence Assessor Pipeline

6 Regressor Competence Assessor Meta Feature Builder

Input for Competence Assessor (𝑌, 𝑍) Training Set 𝑦 Run-time input y′ Prediction 𝐷 Competence Level (𝑌, 𝑍)

Technique

(e.g., Random Forest)

Regressor

Training Set

Primary Model

1 2

slide-7
SLIDE 7

Meta Feature Builder

Relate run-time input with training data and regressor technique Relating Input and Training set:

u Different distances measures depending

  • n the regressor technique

𝑒: 𝐺×𝐺 → ℝ!

u Neighborhood 𝑂(𝑦) based on the distance

measure 𝑒 ⋅,⋅ .

u We consider 𝑙 nearest neighbors

with 𝑙 = 5. 7 Meta Feature Builder

(𝑌, 𝑍)

Training Set

𝑦 Run-time input 𝑔 𝑦 = 𝑧′ Prediction 𝑦 y′

6 meta features

slide-8
SLIDE 8

Our Six Meta Features

1. Average Distance to the Neighborhood 𝑁" 𝑦 ≔ 3

#!,%! ∈' #

𝑒 𝑦, 𝑦( 𝑙

u Measure how far the run-time input

from the training data set. 2. Average Prediction Distance 3. Deviation from regressor’s prediction 𝑁) 𝑦 ≔ 3

#!,%! ∈' #

𝑔 𝑦 − 𝑔(𝑦() 𝑙 𝑁* 𝑦 ≔ 𝑔 𝑦 − 3

#!,%! ∈' #

𝑧 𝑡(𝑦) 𝑒(𝑦, 𝑦′) u Relationship between predictions at the vicinity of current input. 8

slide-9
SLIDE 9

Our Six Meta Features

4. Average training error on 𝑂(𝑦) 5. Variance training error on 𝑂(𝑦) 𝑁+ 𝑦 ≔ 3

#!,%! ∈' #

𝑔 𝑦( − 𝑧′ 𝑡 𝑦 𝑒 𝑦(, 𝑦

𝑁! 𝑦 ≔ $

"!,$! ∈& "

𝑔 𝑦' − 𝑧' − 𝑁( 𝑦

)

𝑙 − 1

u Accuracy of regressor in the immediate vicinity. 6. Target value variability on 𝑂(𝑦) 𝑁, 𝑦 ≔ 3

#!,%! ∈' #

𝑧( − 9 𝑧 ) 𝑙 − 1 u Variance of true value in 𝑂(𝑦). 9

slide-10
SLIDE 10

Training Data For Competence Assesor

10 Regressor Technique

Training Set

Splitter

Base Validation

Meta Feature Builder

𝑍′ 𝑍

Training Data for Competence Assessor

𝐷

slide-11
SLIDE 11

Splitter Procedure

Standard Cross-Validation

u

Random splitting into ℎ ∈ {3,5,10} buckets.

u

One validation bucket and the rest as base.

Projection Splits

u

Assess i.i.d. assumption of the technique.

u

Create interpolation and extrapolation scenarios.

u

Project over 1st and 2nd PC dimension and sort the training data before splitting.

11

Training Set Base Validation Training Set Base Validation Projected and sorted data

slide-12
SLIDE 12

Training Meta Model

Classification Label (C)

u Based on the true error of the learned

model.

u Sort the absolute residual values in

ascending order and set the labels as:

u 80% smaller

à Trusted

u 80-95%

à Cautioned

u Last 5%

à Not trusted

Note: the labeling can be modified for specific applications

Training Techniques

u Off-the-shelf SVM and Random Forest

Classifier.

u Our goal is to test the framework in several

datasets.

Note: More sophisticated techniques can be used for specific applications. 12

slide-13
SLIDE 13

Numerical Evaluation

PART 2 13

slide-14
SLIDE 14

Experimental Setting

Objective:

Evaluate our Empirical Competence Model (ECM) over different scenarios.

u Six UCI benchmark data-sets. u Regressors: Linear, Random Forest, and

  • SVR. (Off-the-shelf)

u Task: standard, interpolation, and

extrapolation.

Cross-Validation Tasks

u Standard cross-validation. u Interpolation and Extrapolation: u Cluster data and take complete

clusters as test set.

u PC projections (1st and 3rd).

14

slide-15
SLIDE 15

Proof-of-Concept Experiment

Setting:

u 1-dimension data following a linear

regression with random noise.

u Interpolation task. u Regressors:

u Linear regression. u Random forest.

15

Linear Regression Model Random Forest Model ECM Predictions

slide-16
SLIDE 16

Evaluating ECM over Airfoil Dataset

16

Trusted Cautious Not Trusted

Bigger MSE for C and NT classes.

slide-17
SLIDE 17

Evaluate Effectiveness of Pipeline

Baseline: Competence assessor trained over original data (only standard splitting and no meta features) 17

Trusted Warned

ECM has lower MSE for Trusted class and higher MSE for Warned class.

slide-18
SLIDE 18

Conclusions & Future Works

u We present an Empirical Confidence Model (ECM) that assess the

reliability of the regression model predictions.

u We show the effectives of ECM for i.i.d. and non-i.i.d. train/test splits. u Future works: u Study other reliability measures as meta-features. u Integrate our methodology in an active learning setting.

18

slide-19
SLIDE 19

Empirical Confidence Models for Supervised Machine Learning

Thank You!