Ground Truth, Machine Learning, and the Mechanical Turk Bob - - PowerPoint PPT Presentation

ground truth machine learning and the mechanical turk
SMART_READER_LITE
LIVE PREVIEW

Ground Truth, Machine Learning, and the Mechanical Turk Bob - - PowerPoint PPT Presentation

Ground Truth, Machine Learning, and the Mechanical Turk Bob Carpenter (w. Emily Jamison, Breck Baldwin) Alias-i, Inc. Supervised Machine Learning 1. Define coding standard mapping inputs to outputs, e.g.: English word stem


slide-1
SLIDE 1

Ground Truth, Machine Learning, and the Mechanical Turk

Bob Carpenter (w. Emily Jamison, Breck Baldwin)

Alias-i, Inc.

slide-2
SLIDE 2

Supervised Machine Learning

  • 1. Define coding standard mapping inputs to outputs,

e.g.:

  • English word → stem
  • newswire text → person name spans
  • biomedical text → genes mentioned
  • 2. Collect inputs and code “gold standard” training data
  • 3. Develop and train statistical model using data
  • 4. Apply to unseen inputs
slide-3
SLIDE 3

Coding Bottleneck

  • Bottleneck is collecting training corpus
  • Commericial data’s expensive (e.g. LDA, ELRA)
  • Academic corpora typically restrictively licensed
  • Limited to existing corpora
  • For new problems, use: self, grad students, temps,

interns, . . .

  • Mechanical Turk to the rescue
slide-4
SLIDE 4

Case Studies

slide-5
SLIDE 5

Case 1: Named Entities

slide-6
SLIDE 6

Named Entities Worked

  • Conveying the coding standard

– official MUC-6 standard dozens of pages – examples are key – (maybe a qualifying exam)

  • User Interface Problem

– highlighting with mouse too fiddly (c.f. Fitt’s Law) – one entity type at a time (vs. pulldown menus) – checkboxes (vs. highlighting spans)

slide-7
SLIDE 7

Discussion: Named Entities

  • 190K tokens, 64K capitalized, 4K names
  • Less than a week at 2 cents/400 tokens
  • Turkers overall better than LDC data

– Correctly Rejected: Webster’s, Seagram, Du Pont, Buick-Cadillac, Moon, erstwhile Phineas Foggs – Incorrectly Accepted: Tass – Missed Punctuation: J E. ‘‘Buster’’ Brown

  • Many Turkers no better than chance

(c.f. social psych by Yochai Benkler, Harvard)

slide-8
SLIDE 8

Case 2: Morphological Stemming

slide-9
SLIDE 9

Morphological Stemming Worked

  • Three iterations on coding standard

– simplified task to one stem

  • Four iterations on final standard instructions

– added previously confusing examples

  • Added qualifying test
slide-10
SLIDE 10

Case 3: Gene Linkage

slide-11
SLIDE 11

Gene Linkage Failed

  • Could get Turkers to pass qualifier
  • Could not get Turkers to take task even at $1/hit
  • Doing coding ourselves (5-10 minutes/HIT)
  • How to get Turkers do these complex tasks?

– Low concentration tasks done quickly – Compatible with studies of why Turkers Turk

slide-12
SLIDE 12

Inferring Gold Standards

slide-13
SLIDE 13

Voted Gold Standard

  • Turkers vote
  • Label with majority category
  • Censor if no majority
slide-14
SLIDE 14

Some Labeled Data

  • Seed the data with cases with known labels
  • Use known cases to estimate coder accuracy
  • Vote with adjustment for accuracy
  • Requires relatively large amount of items for

– estimating accuracies well – liveness for new items

  • Gold may not be as pure as requesters think
  • Some preference tasks have no “right” answer

– e.g. Bing vs. Google, Facestat, Colors, ...

slide-15
SLIDE 15

Estimate Everything

  • Gold standard labels
  • Coder accuracies

– sensitivity (false negative rate; misses) – specificity (false positive rate; false alarms) – imbalance indicates bias; high values accuracy

  • Coding standard difficulty

– average accuracies – variation among coders

  • Item difficulty (important, but not enough data)
slide-16
SLIDE 16

Benefits of Estimation

  • Full Bayesian posterior inference

– probabilistic “gold standard” – compatible with Bayesian machine learning

  • Works better than voting with threshold

– largest benefit with few Turkers/item – evaluated with known “gold standard”

  • Compatible with adding gold standard cases
slide-17
SLIDE 17

Why we Need Task Difficulty

  • What’s your estimate for:

– a baseball player who goes 12 for 20? – a market that goes down 9 out of 10 days? – a coin that lands heads 3 out of 10 times? – . . .

  • Smooth estimates for coders with few items
  • Hierarchical model of accuracy prior
slide-18
SLIDE 18

Soft Gold Standard

  • Is 24 karat gold standard even possible?
  • Some items are really marginal
  • Traditional approach

– censoring disagreements – adjudicating disagreements (revise standard) – adjudication may not converge

  • Posterior uncertainty can be modeled
slide-19
SLIDE 19

Active Learning

  • Choose most useful items to code next
  • Balance two criteria

– high uncertainty – high typicality (how to measure?)

  • Can get away with fewer coders/item
  • May introduce sampling bias
slide-20
SLIDE 20

Code-a-Little, Learn-a-Little

  • Semi-automated coding
  • System suggests labels
  • Coders correct labels
  • Much faster coding
  • But may introduce bias
slide-21
SLIDE 21

Statistical Inference Model

slide-22
SLIDE 22

5 Dentists Diagnosing Cavities

Dentists Count Dentists Count Dentists Count 00000 1880 10000 22 00001 789 10001 26 00010 43 10010 6 00011 75 10011 14 00100 23 10100 1 00101 63 10101 20 00110 8 10110 2 00111 22 10111 17 01000 188 11000 2 01001 191 11001 20 01010 17 11010 6 01011 67 11011 27 01100 15 11100 3 01101 85 11101 72 01110 8 11110 1 01111 56 11111 100

slide-23
SLIDE 23

Posteriors for Dentist Accuracies

Annotator Specificities a.0 0.6 0.7 0.8 0.9 1.0 50 100 150 200 Annotator Key 1 2 3 4 5 Annotator Sensitivities a.1 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 50 100 150 200

  • Posterior density vs. point estimates (e.g. mean)
slide-24
SLIDE 24

Posteriors for Dentistry Data Items

00000 0.0 0.5 1.0 00001 0.0 0.5 1.0 00010 0.0 0.5 1.0 00011 0.0 0.5 1.0 00100 0.0 0.5 1.0 00101 0.0 0.5 1.0 00110 0.0 0.5 1.0 00111 0.0 0.5 1.0 01000 0.0 0.5 1.0 01001 0.0 0.5 1.0 01010 0.0 0.5 1.0 01011 0.0 0.5 1.0 01100 0.0 0.5 1.0 01101 0.0 0.5 1.0 01110 0.0 0.5 1.0 01111 0.0 0.5 1.0 10000 0.0 0.5 1.0 10001 0.0 0.5 1.0 10010 0.0 0.5 1.0 10011 0.0 0.5 1.0 10100 0.0 0.5 1.0 10101 0.0 0.5 1.0 10110 0.0 0.5 1.0 10111 0.0 0.5 1.0 11000 0.0 0.5 1.0 11001 0.0 0.5 1.0 11010 0.0 0.5 1.0 11011 0.0 0.5 1.0 11100 0.0 0.5 1.0 11101 0.0 0.5 1.0 11110 0.0 0.5 1.0 11111 0.0 0.5 1.0

Accounts for bias, so very different from simple vote!

slide-25
SLIDE 25

Beta-Binomial “Random Effects”

✒✑ ✓✏ α0 ✒✑ ✓✏ β0 ✒✑ ✓✏ α1 ✒✑ ✓✏ β1 ✒✑ ✓✏ θ0,j ✒✑ ✓✏ θ1,j ✒✑ ✓✏ π ✒✑ ✓✏ ci ✒✑ ✓✏ xk ❅ ❅ ❘

❅ ❅ ❘

✲ ✲ ❅ ❅ ❅ ❅ ❅ ❘

I K J

slide-26
SLIDE 26

Sampling Notation

Coders don’t all code the same items

ci ∼ Bernoulli(π) a0,j ∼ Beta(α0, β0) a1,j ∼ Beta(α1, β1) xk ∼ Bernoulli(cika1,jk + (1 − cik)(1 − a0,jk)) π ∼ Beta(1, 1) α0/(α0 + β0) ∼ Beta(1, 1) α0 + β0 ∼ Polynomial(−5/2) α1/(α1 + β1) ∼ Beta(1, 1) α1 + β1 ∼ Polynomial(−5/2)

slide-27
SLIDE 27

Posterior for Coding Difficulty/Variance

These are posteriors from a simulation with cross-hairs giving “true” answer:

  • 0.60

0.65 0.70 0.75 0.80 0.85 0.90 20 40 60 80 100

Posterior: Sensitivity Mean & Scale

alpha.1 / (alpha.1 + beta.1) alpha.1 + beta.1

  • 0.60

0.65 0.70 0.75 0.80 0.85 0.90 20 40 60 80 100

Posterior: Specificity Mean & Scale

alpha.0 / (alpha.0 + beta.0) alpha.0 + beta.0

slide-28
SLIDE 28

Extending Model

  • Task difficulty

– Logistic Item-Response models – Need more than 5-10 coders/item for tight posterior

  • Easy to extend coding types

– Multinomial responses – Ordinal responses (e.g. Likert 1-5 scale) – Scalar responses

slide-29
SLIDE 29

The End

  • References

– http://lingpipe-blog.com

  • Contact

– carp@alias-i.com

  • R/BUGS Code Subversion Repository

– http://alias-i.com/lingpipe/sandbox