Human-Centered Machine Learning Saleema a Amershi hi Machine T - - PowerPoint PPT Presentation

human centered machine learning
SMART_READER_LITE
LIVE PREVIEW

Human-Centered Machine Learning Saleema a Amershi hi Machine T - - PowerPoint PPT Presentation

Human-Centered Machine Learning Saleema a Amershi hi Machine T eaching Group, Microsoft Research UW CSE 510 Lecture, March 1, 2016 What is Machine T eaching? Can improve learning with better learning strategies: Note taking


slide-1
SLIDE 1

Human-Centered Machine Learning

Saleema a Amershi hi Machine T eaching Group, Microsoft Research

UW CSE 510 Lecture, March 1, 2016

slide-2
SLIDE 2

What is Machine T eaching?

slide-3
SLIDE 3

Can improve learning with better learning strategies:

  • Note taking
  • Self-explanation
  • Practice
  • Mnemonic devices
  • ….
slide-4
SLIDE 4
slide-5
SLIDE 5
slide-6
SLIDE 6

Machine Learning Machine T eaching

Images from http://thetomatos.com

slide-7
SLIDE 7

What is Machine Learning?

slide-8
SLIDE 8

What is Machine Learning?

“Process by which a system improves performance from experience.” – Herbert Simon “Study of algorithms that improve their performance P at some task T with experience E” – T

  • m Mitchell

“Field of study that gives computers the ability to learn without being explicitly programmed” – Arthur Samuel

slide-9
SLIDE 9

Programming

6 5 3 1 8 7 2 1 2 3 5 6 7 8

slide-10
SLIDE 10

Programming

6 5 3 1 8 7 2 1 2 3 5 6 7 8

slide-11
SLIDE 11

Programming

slide-12
SLIDE 12

Programming

slide-13
SLIDE 13

f(x)≈y

slide-14
SLIDE 14

f( )≈2

slide-15
SLIDE 15

How do Machines Learn?

slide-16
SLIDE 16

How do Machines Learn?

Clean Data Model

Apply Machine Learning Algorithms

slide-17
SLIDE 17

How do Machines Learn?

Apply Machine Learning Algorithms

Images from: https://www.toptal.com/machine-learning/machine-learning-theory-an-introductory-primer

slide-18
SLIDE 18

How do Machines Learn?

Clean Data Model

Apply Machine Learning Algorithms Where do you get this data? How should it be represented? Which algorithm should you use? How do you know if its working?

slide-19
SLIDE 19

Investigating Statistical Machine Learning as a T

  • ol for Software

Developers

Patel, K., Fogarty, J., Landay, J., and Harrison, B. CHI 2008.

slide-20
SLIDE 20

Methodology

Semi-structured interviews with 11 researchers. 5 hour think-aloud study with 10 participants. Digit recognition task.

slide-21
SLIDE 21

Applied Machine Learning is a Process

Collect Data Create Features Select Model Evaluate

Slide content from Kayur Patel

slide-22
SLIDE 22
slide-23
SLIDE 23

Applied Machine Learning is a Process

Collect Data Create Features Select Model Evaluate

slide-24
SLIDE 24

Applied Machine Learning is a Process

Collect Data Create Features Select Model Evaluate

slide-25
SLIDE 25

What About Music Recommendation?

Collect Data Create Features Select Model Evaluate Genre: Rock T empo: Fast Drums: Yes Time of day: Afternoon Recently heard: No ….

slide-26
SLIDE 26

Problems with current tools

Don’t support machine learning as an iterative and exploratory process.

Image from: http://www.crowdflower.com/blog/the-data-science-ecosystem

slide-27
SLIDE 27

Problems with current tools

Don’t support machine learning as an iterative and exploratory process. Don’t support relating data to behaviors of the algorithm.

LogitBoostWith8T

  • 18EvenWindow-Iter=10.model

LogitBoostWith8T

  • 18EvenWindow-Iter=20.model

SVMWith8T

  • 18EvenWindow-Iter=10.model

….

slide-28
SLIDE 28

Problems with current tools

Don’t support machine learning as an iterative and exploratory process. Don’t support relating data to behaviors of the algorithm. Don’t support evaluation in context of use.

slide-29
SLIDE 29

Model Performance is Important

Clean Data Model

Apply Machine Learning Algorithms

slide-30
SLIDE 30

What are other considerations?

Collect Data Create Features Select Model Evaluate

slide-31
SLIDE 31

Computational Efficiency?

Collect Data Create Features Select Model Evaluate

slide-32
SLIDE 32

Data processing efficiency?

Collect Data Create Features Select Model Evaluate

“Data scientists, according to interviews and expert estimates, spend 50 percen ent to 80 percent nt

  • f
  • f their

ir time mired in this more mundane labor

  • f collecting and preparing unruly digital data.”
  • New York Times, 2014
slide-33
SLIDE 33

Understandability?

Collect Data Create Features Select Model Evaluate

“TAP9 initially used a decision tree algorithm because it allowed TAP9 to easily see what features were being used…Later in the study…they transitioned to using more complex models in search of increased performance.”

slide-34
SLIDE 34

Considerations for Machine Learning

Model performance Computational efficiency Iteration efficiency Ease of experimentation Understandability …. New opportunities for HCI research! Need to make tradeoffs!

slide-35
SLIDE 35

Interactive Machine Learning

Fails, J.A. and Olsen, D.R. IUI 2003.

slide-36
SLIDE 36

Crayons: IML for Pixel Classifiers

slide-37
SLIDE 37

What tradeoffs did Crayons make?

Collect Data Create Features Select Model Evaluate

slide-38
SLIDE 38

What tradeoffs did Crayons make?

Collect Data Create Features Select Model Evaluate

slide-39
SLIDE 39

“Classical” ML Interactive ML

slide-40
SLIDE 40

What tradeoffs did Crayons make?

Rapid iteration

  • Fast training
  • Integrated environment

Simplicity Model performance Flexibility

  • Automatic featuring
  • No model selection
slide-41
SLIDE 41

When are these tradeoffs appropriate?

Rapid iteration Simplicity Novices Large set of available features Data can be efficiently viewed and labeled Model performance Flexibility Experts Custom features needed Data types that can’t be viewed at a glance Labels obtained from external sources

slide-42
SLIDE 42

Flock: Hybrid Crowd-Machine Learning Classifiers

Cheng, J. and Bernstein, S. CSCW 2015.

slide-43
SLIDE 43

Collect Data Create Features Select Model Evaluate

“At the end of the day, some machine learning projects succeed and some fail. What makes the differences? Easil ily the e most important ant factor

  • r is

the features ures used ed.” [Domingos, CACM 2012]

slide-44
SLIDE 44

How do people come up with features?

Look for features used in related domains. Use intuition or domain knowledge. Apply automated techniques. Featur ture e ideati ation

  • n – think of and experiment with custom features.
slide-45
SLIDE 45

A Brainstorming Exercise

slide-46
SLIDE 46

How do people come up with features?

Look for features used in related domains. Use intuition or domain knowledge. Apply automated techniques. Featur ture e ideati ation

  • n – think of and experiment with custom features.

“The novelty of generated ideas increases as participants ideate, reaching a peak after their 18th instance.” [Krynicki, F. R., 2014]

slide-47
SLIDE 47

Workflow

User specifies a concept and uploads some unlabeled data. Crowd views data and suggests features.

slide-48
SLIDE 48

What makes a cat a cat?

slide-49
SLIDE 49

What makes a cat a cat?

slide-50
SLIDE 50

Workflow

User specifies a concept and uploads some unlabeled data. Crowd compares and contrasts positive and negative examples and suggests “why” they are different. Reasons become features. Reasons are clustered. User vets, edits, and adds to features. Crowd implements feature by labeling data. Features used to build classifiers.

slide-51
SLIDE 51

Collect Data Create Features Select Model Evaluate Collect Data Create Features Select Model Evaluate Crayons Flock

slide-52
SLIDE 52

Collect Data Create Features Select Model Evaluate

Positives Negatives Standard Ranked List Split T echnique (Best/Worst Matches) [Fogarty et al., CHI 2007]

slide-53
SLIDE 53

Collect Data Create Features Select Model Evaluate

Traditional Labeling Grouping and tagging surfaces decision making. Moving, merging and splitting groups helps with revising decisions. [Kulesza et al., CHI 2014] Structured Labeling

slide-54
SLIDE 54

Collect Data Create Features Select Model Evaluate

[Amershi et al., CHI 2015] Summary Stats ModelTracker

slide-55
SLIDE 55

Collect Data Create Features Select Model Evaluate

[Patel et al., IJCAI 2011]

slide-56
SLIDE 56

Collect Data Create Features Select Model Evaluate

Rule-based explanation Keyword-based explanation Similarity-based explanation

[Stumpf et al, IUI 2007]

slide-57
SLIDE 57

Collect Data Create Features Select Model Evaluate

slide-58
SLIDE 58

Experts Everyday People Practitioners

slide-59
SLIDE 59

Experts Everyday People Practitioners

How are these scenarios different?

User experience impacts what you can expose. Interaction focus impacts attention and feedback. Accuracy requirements impacts expected time and effort. …..

slide-60
SLIDE 60

Principles for human-centered ML?

Tradit ditiona nal User Interfa faces ces

Visibility and feedback Consistency and standards Predictability Actionability Error prevention and recovery ….

Intelligent/ gent/ML ML-Base ased d Interfa faces ces

Safety Trust Manage expectations Degrade gracefully under uncertainty ….

slide-61
SLIDE 61

Traditional Machine Learning

Clean Data Model

Apply Machine Learning Algorithms

slide-62
SLIDE 62

Human-Centered Machine Learning

Collect Data Create Features Select Model Evaluate

slide-63
SLIDE 63

Human-Centered Machine Learning

Machine Learning Machine T eaching

+

samershi@microsoft.com