Informed Truthfulness for Multi-Task Peer Prediction Victor Shnayder - - PowerPoint PPT Presentation

informed truthfulness for multi task peer prediction
SMART_READER_LITE
LIVE PREVIEW

Informed Truthfulness for Multi-Task Peer Prediction Victor Shnayder - - PowerPoint PPT Presentation

Informed Truthfulness for Multi-Task Peer Prediction Victor Shnayder , Arpit Agarwal, Rafael Frongillo, David C. Parkes Nov 3, 2016; HCOMP16 Foundations Workshop 1 Lets talk about crowdsourcing 2 Task: How many cows in this image? 3


slide-1
SLIDE 1 1

Informed Truthfulness for Multi-Task Peer Prediction

Victor Shnayder, Arpit Agarwal, Rafael Frongillo, David C. Parkes

Nov 3, 2016; HCOMP’16 Foundations Workshop

slide-2
SLIDE 2 2

Let’s talk about crowdsourcing

slide-3
SLIDE 3

Task: How many cows in this image?

3
slide-4
SLIDE 4

Count: 23

4
slide-5
SLIDE 5

… or just guess: 42

5
slide-6
SLIDE 6

Get feedback: “23 is good. $0.02”

6
slide-7
SLIDE 7

Get feedback: “42 is way off! $0.00”

7
slide-8
SLIDE 8

Get feedback: “42 is way off! $0.00” Where did feedback come from?

8
slide-9
SLIDE 9

Get feedback: “42 is way off! $0.00” Where did feedback come from?

Gold standard: professional cow counter said 23

9
slide-10
SLIDE 10

Get feedback: “42 is way off! $0.00” Where did feedback come from?

Machine vision: automated cow counter said 20-25

10
slide-11
SLIDE 11

Get feedback: “42 is way off! $0.00” Where did feedback come from?

Peers: 3 others said 20, 22, 24

11
slide-12
SLIDE 12

Peer prediction: giving feedback based on peer reports

12
slide-13
SLIDE 13

Evaluate

“7 cows” “good!” “1 cow” “1 cow” “good!” “bad!” Observe Observe Observe

13
slide-14
SLIDE 14

Applications beyond s

Gather location-specific info Image and video labeling Search result evaluation Academic peer review Participatory sensing Evaluations for peer assessment in massive courses

14
slide-15
SLIDE 15

Goals:

Ensure truthful equilibrium exists and is attractive Impossibility results

15
slide-16
SLIDE 16

Task model

Agent 1 Task i

16

Signal: 1 … n

slide-17
SLIDE 17

Task model

Agent 1 Task i

17

Prior probability P(i)

slide-18
SLIDE 18

Task model

Agent 1 Task Agent 2 i j

18

Joint probability: P(i,j)

slide-19
SLIDE 19

Task model

Agent 1 Task Agent 2 i j

19

People can misreport!

slide-20
SLIDE 20

Task model

Agent 1 Task Agent 2 i j

20

Goal: design scores to encourage effort, truthful reports

slide-21
SLIDE 21

Output agreement (von Ahn, Dabbish ‘04)

Agent 1 Task Agent 2

1 1

Agree! Score 1

21
slide-22
SLIDE 22

Agent 1 Task Agent 2

1 2

Don’t agree: score 0

22

Output agreement

slide-23
SLIDE 23

Output agreement

Agent 1 Task Agent 2 Honest reporting is a correlated equil if my signal predicts yours

23
slide-24
SLIDE 24

Output agreement

Agent 1 Task Agent 2

24

To manipulate: all agents always report same thing

slide-25
SLIDE 25

Ensuring truthful reporting is best

[Kamble et. al. ’15, Radanovic et. al. ’16]

Agent 1 Task Agent 2

1 1

Agree! Score 1/(scaling factor) Scaling factor learned from reports on many similar tasks. Truthfulness is an equilibrium, guarantees highest payoff.

25
slide-26
SLIDE 26

Multi-task approach

[Dasgupta-Ghosh’13] Agent 1

Task 1 Task 2 1 1

26
slide-27
SLIDE 27

Task 3

Agent 1

Task 1 Task 2

Agent 2

1 1 2 2

27

Multi-task approach

[Dasgupta-Ghosh’13]

slide-28
SLIDE 28

Task 3

Agent 1

Task 1 Task 2

Agent 2

1 2 1 2

Likely to match

28

Key idea

slide-29
SLIDE 29

Task 3

Agent 1

Task 1 Task 2

Agent 2

1 2 1 2

L e s s l i k e l y t

  • m

a t c h

29

Key idea

slide-30
SLIDE 30

Task 3

Agent 1

Task 1 Task 2

Agent 2

1 2 1 2

Reward matching on shared tasks. Punish matching on non-shared tasks.

30

Key idea

slide-31
SLIDE 31

Task 3

Agent 1

Task 1 Task 2

Agent 2

1 2 1 2

31

P(agree on shared) – P(agree on non-shared)

Why doesn’t this Just Work with >2 signals?

slide-32
SLIDE 32

Our multi-signal mechanism: Correlated Agreement

  • 1. Split tasks into shared and non-shared.
  • 2. Score = (agree on shared) – (agree on non-shared)

“Agree” when reports aren’t equal, but positively correlated.

Constant or random reporting still has expected score 0

32
slide-33
SLIDE 33

The Correlated Agreement mechanism

  • 1. Informed truthful—being truthful is optimal, better

than constant or random reports

  • 2. Works with minimal information with few tasks
  • 3. Works with no information with many tasks
33
slide-34
SLIDE 34

Connection to information theory

Truthful score ~ DTV(P(⋅,⋅) - P(⋅)P(⋅)) More agreement ⇒ higher scores. Other rules correspond to different distance functions. [Kong-Schoenebeck ‘16]

34
slide-35
SLIDE 35

Open questions

Is peer prediction practical as primary incentive? When? Combine peer prediction with other incentive models in a single system? Heterogeneous agents Non-binary effort models Non-random task assignment (e.g. maps) Unintended correlated “signals”

35
slide-36
SLIDE 36

Thank you!

shnayder@post.harvard.edu

36
slide-37
SLIDE 37

Extra slides

37
slide-38
SLIDE 38

Setup

Agents 1, 2, tasks k Signals i,j (require effort) Shared tasks, agent 1 tasks, agent 2 tasks Signal prior P(i), joint P(i,j) Strategies: F, G probability of reporting r given signal i. Informed strategy: depend on the signal somehow Truthful strategy: F*

38
slide-39
SLIDE 39

Solution concepts

E(F, G): expected payment for a shared task (Strict) Proper: E(F*, G*) ≥ E(F, G*), for all F != F* Strong-truthful: E(F*, G*) ≥ E(F, G), for all F, G (if expected payment tied, F and G must be permutations) Informed-truthful: E(F*, G*) ≥ E(F, G), for all F, G (if expected payment tied, F and G must be informed)

39
slide-40
SLIDE 40

A useful matrix

40
slide-41
SLIDE 41

Example Delta

Joint:

41

Delta: Sign(Delta): Prior:

slide-42
SLIDE 42

Scoring matrix

42
slide-43
SLIDE 43

Expected payment

43

(Lemma: can restrict to deterministic strategies)

slide-44
SLIDE 44

Key theorem The Correlated Agreement mechanism is informed truthful for all* models.

44
slide-45
SLIDE 45

Proof sketch

45
slide-46
SLIDE 46

Other results

Detail-free version of mechanism: learn scoring matrix from reports Correlated Agreement is maximal among large class of mechanisms that use this scoring structure Much simpler analysis of Dasgupta-Ghosh’13 mechanism

46
slide-47
SLIDE 47

Binary mechanism (Dasgupta-Ghosh ’13):

  • 1. Split tasks into shared and non-shared.
  • 2. Score: (agree on shared) – (agree on non-shared)

Expected score: P(agree on shared) – P(agree on non-shared) Have to agree based on properties of shared task. Constant or random reporting has expected score = 0

47