Eliciting Informative Feedback: The Peer-Prediction Method Colin - - PowerPoint PPT Presentation

β–Ά
eliciting informative feedback the peer prediction method
SMART_READER_LITE
LIVE PREVIEW

Eliciting Informative Feedback: The Peer-Prediction Method Colin - - PowerPoint PPT Presentation

Eliciting Informative Feedback: The Peer-Prediction Method Colin Zheng and Kenneth Wang 1 The Setting Camera Quality: dist over {, } 2 ~ 1 ~ : dist over , jointly Rater Rater w/ 1 2 Report


slide-1
SLIDE 1

1

Eliciting Informative Feedback: The Peer-Prediction Method

Colin Zheng and Kenneth Wang

slide-2
SLIDE 2

2 2

Rater 1

Camera Quality: dist 𝐸 over {𝑀, 𝐼}

Rater 2 𝑑2~𝑇 𝑑1~𝑇

𝑇: dist over π‘š, β„Ž jointly w/ 𝐸

Center:

  • Ebay
  • NetFlix
  • Amazon

Report 𝑠1 Report 𝑠2

The Setting

slide-3
SLIDE 3

3 3

Rater 1

Camera Quality: dist 𝐸 over {𝑀, 𝐼}

Rater 2 𝑑2~𝑇 𝑑1~𝑇

𝑇: dist over π‘š, β„Ž jointly w/ 𝐸

Center:

  • Ebay
  • NetFlix
  • Amazon

Report 𝑠1 Report 𝑠2

The Setting

Task:

How to make rational raters report honestly (𝑠𝑗 = 𝑑𝑗)

Naive Attempt: reward 𝜐 = $1 if 𝑠1 = 𝑠2, 𝜐 = $0 otherwise Problem: rater 1 will report β€œmore likely signal” of rater 2

slide-4
SLIDE 4

4 4

Rater 1

Camera Quality: dist 𝐸 over {𝑀, 𝐼}

Rater 2 𝑑2~𝑇 𝑑1~𝑇

𝑇: dist over π‘š, β„Ž jointly w/ 𝐸

Center:

  • Ebay
  • NetFlix
  • Amazon

Report 𝑠1 Report 𝑠2

The Setting

Task (formulation):

Reporting 𝑠𝑗 = 𝑑𝑗 is a Nash Equilibrium βˆ€ rater 𝑗, signal 𝑛, 𝑦 β‰  𝑛 𝔽𝑑2 𝜐1 𝑑1, 𝑑2 𝑑1 = 𝑛 β‰₯ 𝔽𝑑2 𝜐1 𝑦, 𝑑2 𝑑1 = 𝑛 𝔽𝑑1 𝜐2 𝑑2, 𝑑1 𝑑2 = 𝑛 β‰₯ 𝔽𝑑1 𝜐2 𝑦, 𝑑1 𝑑2 = 𝑛

slide-5
SLIDE 5

5 5

Rater 1

Camera Quality: dist 𝐸 over {𝑀, 𝐼}

Rater 2 𝑑2~𝑇 𝑑1~𝑇

𝑇: dist over π‘š, β„Ž jointly w/ 𝐸

Center:

  • Ebay
  • NetFlix
  • Amazon

Report 𝑠1 Report 𝑠2

Solution using Proper Scoring Rule

Task:

choose payoff 𝜐1 such that 𝔽𝑑2 𝜐1 𝑦, 𝑑2 𝑑1 = 𝑛 is maximized at 𝑦 = 𝑛 Equivalently, 𝔽𝑨~π‘Ž 𝜐1 𝑦, 𝑨 where π‘Ž = 𝑑2 𝑑1=𝑛

Recall: 𝑔 is Proper Scoring Rule if

𝔽𝑨~π‘Ž[𝑔(𝑄, 𝑨)] is maximized at 𝑄 = π‘Ž

So let 𝜐1 𝑦, 𝑨 = 𝑔(𝑑2 𝑑1=𝑦,𝑨)

slide-6
SLIDE 6

6 6

Issue: costs to the rater

In reality,

  • Evaluating and reporting honestly incur a

relative cost 𝑑 > 0

(e.g. testing is time consuming, opportunity cost, ...)

  • Can be offset by scaling up the payoffs
  • Problems:

– Paying too much for truthful information? – What if 𝑑 is unknown? [PRGJ08]

slide-7
SLIDE 7

7 7

Issue: costs to the center

To reduce center’s cost:

  • Budget balancing:

– pair up raters, make each pair a zero-sum game

  • Use linear optimization to find cost-optimal

payoff function [JF06]

slide-8
SLIDE 8

8 8

Issue: risk aversion

  • When the center knows the raters’ utility

function(s)

  • When the center does not know the utility

function(s)

  • Using multiple reference raters reduces risk
slide-9
SLIDE 9

9 9

Issue: collusion

  • Using an honest reference rater
  • Randomized reference rater
  • Outside experts...
slide-10
SLIDE 10

10 10

Issue: Unknown/diff. priors

  • In reality, raters have
  • diff. beliefs about

(𝐸, 𝑇), unknown to the center

  • Addressed in [WP12]

for binary signals 𝑇

Rater 1

Camera Quality: dist 𝐸 over {𝑀, 𝐼}

Rater 2 𝑑2~𝑇 𝑑1~𝑇 𝑇: dist over

π‘š, β„Ž jointly w/ 𝐸

Center:

  • Ebay
  • NetFlix
  • Amazon

Report 𝑠1 Report 𝑠2