Strategic Classification with Crowdsourcing Yang Liu ( joint work - - PowerPoint PPT Presentation

strategic classification with crowdsourcing
SMART_READER_LITE
LIVE PREVIEW

Strategic Classification with Crowdsourcing Yang Liu ( joint work - - PowerPoint PPT Presentation

Strategic Classification with Crowdsourcing Yang Liu ( joint work with Yiling Chen) yangl@seas.harvard.edu Harvard University Nov. 2016 Introduction Preliminary Our results Conclusion (Non-strategic) Classification Non-strategic


slide-1
SLIDE 1

Strategic Classification with Crowdsourcing

Yang Liu ( joint work with Yiling Chen) yangl@seas.harvard.edu Harvard University

  • Nov. 2016
slide-2
SLIDE 2

Introduction Preliminary Our results Conclusion

(Non-strategic) Classification

Non-strategic classification yi = f ∗(xi), f ∗ : Rd → {−1, +1}

  • Observing a set of training data, to learn f

˜ f = argminf ∈F

n

  • i=1

l(f (xi), yi).

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

2 / 11

slide-3
SLIDE 3

Introduction Preliminary Our results Conclusion

Strategic classification

When data comes from strategic data sources...

  • Outsource xi to get a label ˜

yi.

  • Crowdsourcing, survey, human reports etc.

Such training data carries noise

  • Intrinsic: due to limited worker expertise.
  • Strategic: lack of incentives.

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

3 / 11

slide-4
SLIDE 4

Introduction Preliminary Our results Conclusion

Strategic classification

When data comes from strategic data sources...

  • Outsource xi to get a label ˜

yi.

  • Crowdsourcing, survey, human reports etc.

Such training data carries noise

  • Intrinsic: due to limited worker expertise.
  • Strategic: lack of incentives.

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

3 / 11

slide-5
SLIDE 5

Introduction Preliminary Our results Conclusion

Goal to achieve

The leaner wants to learn a good, unbiased classifier

  • Workers’ observations come from a flipping error model p+, p−.
  • Workers are effort sensitive.
  • Elicit high quality data from workers. (better performance)

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

4 / 11

slide-6
SLIDE 6

Introduction Preliminary Our results Conclusion

Goal to achieve

The leaner wants to learn a good, unbiased classifier

  • Workers’ observations come from a flipping error model p+, p−.
  • Workers are effort sensitive.
  • Elicit high quality data from workers. (better performance)

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

4 / 11

slide-7
SLIDE 7

Introduction Preliminary Our results Conclusion

Goal to achieve

The leaner wants to learn a good, unbiased classifier

  • Workers’ observations come from a flipping error model p+, p−.
  • Workers are effort sensitive.
  • Elicit high quality data from workers. (better performance)

Information elicitation without verification

  • Peer prediction: SCORE(˜

yi, ˜ yj)

  • DG13, RF15, SAFP16, KS16...
  • Exerting effort to have a high quality data is

usually a good equilibria.

SCORE(˜ yi, ˜ yj)

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

4 / 11

slide-8
SLIDE 8

Introduction Preliminary Our results Conclusion

Our method

Joint learning and information elicitation:

  • SCORE(˜

yi, ˜ yj) ⇒ SCORE(˜ yi, Machine)

  • ”Machine Prediction”
  • How to obtain a good machine answer?

SCORE(˜ yi, Machine)

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

5 / 11

slide-9
SLIDE 9

Introduction Preliminary Our results Conclusion

Classification with flipping errors [Natarajan et al. 13]

  • Suppose workers are truthfully reporting, how to de-bias?

˜ l(t, y) := (1 − p−y)l(t, y) − pyl(t, −y) 1 − p+ − p− , p+ + p− < 1.

  • Why does it work? [un-biased in expectation]

y[˜

l(t, ˜ y)] = l(t, y), ∀t.

  • Find ˜

f ∗

˜ l

via minimizing the empirical risk w.r.t. ˜ l(t, y): ˜ f ∗

˜ l

= argminf ˆ R˜

l(f ) := 1

N

N

  • j=1

˜ l(f (xj), ˆ yj).

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

6 / 11

slide-10
SLIDE 10

Introduction Preliminary Our results Conclusion

Classification with flipping errors [Natarajan et al. 13]

  • Suppose workers are truthfully reporting, how to de-bias?

˜ l(t, y) := (1 − p−y)l(t, y) − pyl(t, −y) 1 − p+ − p− , p+ + p− < 1.

  • Why does it work? [un-biased in expectation]

y[˜

l(t, ˜ y)] = l(t, y), ∀t.

  • Find ˜

f ∗

˜ l

via minimizing the empirical risk w.r.t. ˜ l(t, y): ˜ f ∗

˜ l

= argminf ˆ R˜

l(f ) := 1

N

N

  • j=1

˜ l(f (xj), ˆ yj).

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

6 / 11

slide-11
SLIDE 11

Introduction Preliminary Our results Conclusion

Classification with flipping errors [Natarajan et al. 13]

  • Suppose workers are truthfully reporting, how to de-bias?

˜ l(t, y) := (1 − p−y)l(t, y) − pyl(t, −y) 1 − p+ − p− , p+ + p− < 1.

  • Why does it work? [un-biased in expectation]

y[˜

l(t, ˜ y)] = l(t, y), ∀t.

  • Find ˜

f ∗

˜ l

via minimizing the empirical risk w.r.t. ˜ l(t, y): ˜ f ∗

˜ l

= argminf ˆ R˜

l(f ) := 1

N

N

  • j=1

˜ l(f (xj), ˆ yj).

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

6 / 11

slide-12
SLIDE 12

Introduction Preliminary Our results Conclusion

Our mechanism

For each worker i:

  • Estimate flipping errors ˜

pi,+, ˜ pi,− based on {xj, ˜ yj}j=i.

  • Train ˜

f ∗

˜ l,−i using [Natarajan et al. 13] with data from j = i.

ˆ yi

˜ f ∗

˜ l,−i(xi) Agree?

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

7 / 11

slide-13
SLIDE 13

Introduction Preliminary Our results Conclusion

How to estimate error rate

How do we estimate without ground-truth?

P+[p2

i,+ + (1 − pi,+)2] + P−[p2 i,− + (1 − pi,−)2] = Pr(mathcing)

P+pi,+ + P−(1 − pi,−) = Fraction of -1 labels observed

  • Lemma: There is a unique pair of ˜

pi,+, ˜ pi,− s.t. ˜ pi,+ + ˜ pi,− < 1 ⇒Bayesian informative: ⇔ Pr(yi = s|˜ yi = s) > Prior(s), s ∈ {+, −}

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

8 / 11

slide-14
SLIDE 14

Introduction Preliminary Our results Conclusion

How to estimate error rate

How do we estimate without ground-truth?

P+[p2

i,+ + (1 − pi,+)2] + P−[p2 i,− + (1 − pi,−)2] = Pr(mathcing)

P+pi,+ + P−(1 − pi,−) = Fraction of -1 labels observed

  • Lemma: There is a unique pair of ˜

pi,+, ˜ pi,− s.t. ˜ pi,+ + ˜ pi,− < 1 ⇒Bayesian informative: ⇔ Pr(yi = s|˜ yi = s) > Prior(s), s ∈ {+, −}

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

8 / 11

slide-15
SLIDE 15

Introduction Preliminary Our results Conclusion

How to estimate error rate

How do we estimate without ground-truth?

P+[p2

i,+ + (1 − pi,+)2] + P−[p2 i,− + (1 − pi,−)2] = Pr(mathcing)

P+pi,+ + P−(1 − pi,−) = Fraction of -1 labels observed

  • Lemma: There is a unique pair of ˜

pi,+, ˜ pi,− s.t. ˜ pi,+ + ˜ pi,− < 1 ⇒Bayesian informative: ⇔ Pr(yi = s|˜ yi = s) > Prior(s), s ∈ {+, −}

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

8 / 11

slide-16
SLIDE 16

Introduction Preliminary Our results Conclusion

Results

Effort exertion is a BNE. Benefits of doing so?

  • Less redundant assignment: not all tasks are re-assigned ⇒

budget efficient.

  • Better incentive: Reporting symmetric uninformative signal &

permutation signal is not an equilibrium.

  • More learning flavor: no requirement of knowing workers’ data

distribution.

  • Better privacy preserving etc...

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

9 / 11

slide-17
SLIDE 17

Introduction Preliminary Our results Conclusion

Results

Effort exertion is a BNE. Benefits of doing so?

  • Less redundant assignment: not all tasks are re-assigned ⇒

budget efficient.

  • Better incentive: Reporting symmetric uninformative signal &

permutation signal is not an equilibrium.

  • More learning flavor: no requirement of knowing workers’ data

distribution.

  • Better privacy preserving etc...

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

9 / 11

slide-18
SLIDE 18

Introduction Preliminary Our results Conclusion

Results

Effort exertion is a BNE. Benefits of doing so?

  • Less redundant assignment: not all tasks are re-assigned ⇒

budget efficient.

  • Better incentive: Reporting symmetric uninformative signal &

permutation signal is not an equilibrium.

  • More learning flavor: no requirement of knowing workers’ data

distribution.

  • Better privacy preserving etc...

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

9 / 11

slide-19
SLIDE 19

Introduction Preliminary Our results Conclusion

Results

Effort exertion is a BNE. Benefits of doing so?

  • Less redundant assignment: not all tasks are re-assigned ⇒

budget efficient.

  • Better incentive: Reporting symmetric uninformative signal &

permutation signal is not an equilibrium.

  • More learning flavor: no requirement of knowing workers’ data

distribution.

  • Better privacy preserving etc...

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

9 / 11

slide-20
SLIDE 20

Introduction Preliminary Our results Conclusion

Results

Effort exertion is a BNE. Benefits of doing so?

  • Less redundant assignment: not all tasks are re-assigned ⇒

budget efficient.

  • Better incentive: Reporting symmetric uninformative signal &

permutation signal is not an equilibrium.

  • More learning flavor: no requirement of knowing workers’ data

distribution.

  • Better privacy preserving etc...

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

9 / 11

slide-21
SLIDE 21

Introduction Preliminary Our results Conclusion

A case study: collusion is not an equlibria

Suppose j = i collude by reporting −1

P+(p2

i,+ + (1 − pi,+)2] + P−(p2 i,− + (1 − pi,−)2] = 1.

P+pi,+ + P−1(1 − pi,−) = 1.

⇒ ˜ pi,+ = 1 ⇒ the solution interprets the missing of +1 as high error rate. ˜ l(t, y = −1) := (1 − ˜ pi,+)l(t, −1) − ˜ pi,−l(t, +1) 1 − ˜ pi,+ − ˜ pi,− = l(t, +1) ⇒ the surrogate loss punishes this particular class ⇒ better to report +1 to match.

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

10 / 11

slide-22
SLIDE 22

Introduction Preliminary Our results Conclusion

A case study: collusion is not an equlibria

Suppose j = i collude by reporting −1

P+(p2

i,+ + (1 − pi,+)2] + P−(p2 i,− + (1 − pi,−)2] = 1.

P+pi,+ + P−1(1 − pi,−) = 1.

⇒ ˜ pi,+ = 1 ⇒ the solution interprets the missing of +1 as high error rate. ˜ l(t, y = −1) := (1 − ˜ pi,+)l(t, −1) − ˜ pi,−l(t, +1) 1 − ˜ pi,+ − ˜ pi,− = l(t, +1) ⇒ the surrogate loss punishes this particular class ⇒ better to report +1 to match.

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

10 / 11

slide-23
SLIDE 23

Introduction Preliminary Our results Conclusion

A case study: collusion is not an equlibria

Suppose j = i collude by reporting −1

P+(p2

i,+ + (1 − pi,+)2] + P−(p2 i,− + (1 − pi,−)2] = 1.

P+pi,+ + P−1(1 − pi,−) = 1.

⇒ ˜ pi,+ = 1 ⇒ the solution interprets the missing of +1 as high error rate. ˜ l(t, y = −1) := (1 − ˜ pi,+)l(t, −1) − ˜ pi,−l(t, +1) 1 − ˜ pi,+ − ˜ pi,− = l(t, +1) ⇒ the surrogate loss punishes this particular class ⇒ better to report +1 to match.

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

10 / 11

slide-24
SLIDE 24

Introduction Preliminary Our results Conclusion

Summary

What we achieve

  • A classification problem with strategic data sources.
  • A classification aided approach to elicit information.
  • Enjoy several favorable properties.

Hope to see more on how machine learning can help information elicitation

Thank you!

Liu (Harvard) Mathematical Fundation for Crowdsourcing HCOMP16

  • Nov. 2016

11 / 11