Complexity vs. Performance: Empirical Analysis of Machine Learning - - PowerPoint PPT Presentation

complexity vs performance empirical analysis of machine
SMART_READER_LITE
LIVE PREVIEW

Complexity vs. Performance: Empirical Analysis of Machine Learning - - PowerPoint PPT Presentation

Complexity vs. Performance: Empirical Analysis of Machine Learning as a Service Yuanshun Yao , Zhujun Xiao, Bolun Wang*, Bimal Viswanath, Haitao Zheng and Ben Y. Zhao The University of Chicago *University of California, Santa Barbara


slide-1
SLIDE 1

Complexity vs. Performance: Empirical Analysis of Machine Learning as a Service

Yuanshun Yao, Zhujun Xiao, Bolun Wang*, Bimal Viswanath, Haitao Zheng and Ben Y. Zhao The University of Chicago *University of California, Santa Barbara ysyao@cs.uchicago.edu

slide-2
SLIDE 2

ML in Network Research

congestion control protocols

  • Sivaraman et al.,

SIGCOMM’14

  • Winstein & Balakrishnan,

SIGCOMM’13

network link prediction

  • Liu et al., IMC’16
  • Zhao et al., IMC’12

user behavior analysis

  • Wang et al., IMC’14
  • Zannettouet al., IMC’17

slide-3
SLIDE 3

Running ML is Hard

dataset model

Solution: Machine Learning as a Service (ML-as-a-Service)

slide-4
SLIDE 4

ML-as-a-Service

ML-as-a-Service

training data user input (model, parameter etc.)

slide-5
SLIDE 5

Is my model good enough?

Why Study ML-as-a-Service?

Q: How well do they perform? Q: How much does the amount of user control impact ML performance?

slide-6
SLIDE 6

ML-as-a-Service Platforms

Google Prediction Amazon ML Microsoft ML PIO ABM BigML less amount of user input more

slide-7
SLIDE 7

Control in ML

training data trained model

?

slide-8
SLIDE 8

Control in ML

training data trained model

Data Cleaning

  • Invalid/dup/missing

data

?

slide-9
SLIDE 9

Control in ML

training data trained model

Data Cleaning

  • Invalid/dup/missing

data

Feature Selection

  • Mutual Info,Pearson,

Chi…

?

slide-10
SLIDE 10

Control in ML

training data

Classifier Choice

  • Logistic Regression,

Decision Tree, kNN…

trained model

Data Cleaning

  • Invalid/dup/missing

data

Feature Selection

  • Mutual Info,Pearson,

Chi_square…

?

slide-11
SLIDE 11

Control in ML

training data

Classifier Choice

  • Logistic Regression,

Decision Tree, kNN…

trained model

Data Cleaning

  • Invalid/dup/missing

data

Feature Selection

  • Mutual Info,Pearson,

Chi_square…

Parameter Tuning

  • Logistic Regression: L1,

L2, max_iter…

slide-12
SLIDE 12

Control in ML-as-a-Service

Google ABM

✖ ✖ ✖ ✖ ✖ ✖ ✖ ✖

Amazon

✖ ✖ ✖

PIO BigML

✖ ✖

✔ ✔

✖ ✖

Microsoft

✔ ✔

low user control/complexity high

Data Cleaning Feature Selection Classifier Choice Parameter Tuning

Complexity vs. Performance?

slide-13
SLIDE 13

Performance Measurement

slide-14
SLIDE 14

Characterizing Performance

  • Theoretical modeling is hard
  • Output of ML model depends on dataset
  • No access to implementation details
  • Empirical data-driven analysis
  • Simulate a real-world scenario from end to end
  • Need a large number of diverse datasets
  • Focus on binary classification
slide-15
SLIDE 15

Dataset

  • 119 datasets
  • From diverse application domains
  • Sample size: 15 - 245K, number of features: 1 - 4K
  • 79% of them are from UCI ML Repository

Life Science 37% Computer Applications 15% Artificial Test 14% Social Science 9% Physical Science 8% Financial & Business 6% Other 11%

slide-16
SLIDE 16

Methodology

  • Tune all available control dimensions

training data trained model

Feature Selection Classifier Choice Parameter Tuning

✖ ✔ ✔

API

  • Logistic Regression
  • KNN
  • SVM

API

  • L1_reg
  • L2_reg
  • Max_iter

API

slide-17
SLIDE 17

Methodology

  • Tune all available control dimensions

training data trained model

Feature Selection Classifier Choice Parameter Tuning

✖ ✔ ✔

API

testing data

API

slide-18
SLIDE 18

Trade-offs between Complexity and Performance

slide-19
SLIDE 19

Complexity vs. Performance

complexity low high

  • Q: How does the complexity correlate with performance?
  • High complexity -> high performance

0.5 0.6 0.7 0.8 0.9 1 ABM Google Amazon BigML PIO Microsoft Scikit

Average F-Score

Optimized

slide-20
SLIDE 20

Complexity vs. Risk

  • Q: How does the risk correlate with complexity?
  • High complexity -> high risk

complexity low high

0.1 0.2 0.3 0.4 0.5 ABM Google Amazon BigML PIO Microsoft Scikit Performance Variance (F-Score)

slide-21
SLIDE 21

Understanding Server-side Optimization

slide-22
SLIDE 22

Reverse-engineering Optimization

  • 1

1 2

  • 1.5 -1 -0.5

0.5 1 1.5 Feature #2 Feature #1 Class 0 Class 1

  • 6
  • 3

3 6

  • 3
  • 2
  • 1

1 2 3 Feature #2 Feature #1 Class 0 Class 1

Circular Linear

  • Q: Does server-side adapt to different datasets?
  • Reverser-engineering using datasets
  • Create synthetic datasets
  • Use prediction results to infer classifier information
slide-23
SLIDE 23

Understanding Optimization

Google decision boundaries

  • 1

1 2

  • 1.5 -1 -0.5

0.5 1 1.5 Feature #2 Feature #1 Class 0 Class 1

  • 6
  • 3

3 6

  • 3
  • 2
  • 1

1 2 3 Feature #2 Feature #1 Class 0 Class 1

  • Google switches between classifiers based on the dataset
  • Use supervised learning to infer classifier family used
slide-24
SLIDE 24

Takeaways

  • ML-as-a-Service is an attractive tool to reduce workload
  • But user control still has a large impact on performance
  • Fully automated systems are less risky
slide-25
SLIDE 25

Thank you! Questions?