1
Announcements
ØHW 3 due next Tuesday ØNo HW 4
Announcements HW 3 due next Tuesday No HW 4 1 CS6501: T opics in - - PowerPoint PPT Presentation
Announcements HW 3 due next Tuesday No HW 4 1 CS6501: T opics in Learning and Game Theory (Fall 2019) Crowdsourcing Information and Peer Prediction Instructor: Haifeng Xu Outline Eliciting Information without Verification
1
ØHW 3 due next Tuesday ØNo HW 4
CS6501: T
(Fall 2019) Crowdsourcing Information and Peer Prediction
Instructor: Haifeng Xu
3
Ø Eliciting Information without Verification Ø Equilibrium Concept and Peer Prediction Mechanism Ø Bayesian Truth Serum
4
ØRecruit AMT workers to label images
5
ØRecruit AMT workers to label images
ØPeer grading (of, e.g., essays) on MOOC
6
ØRecruit AMT workers to label images
ØPeer grading (of, e.g., essays) on MOOC
ØElicit ratings for various entities (e.g., on Yelp or Google)
7
ØRecruit AMT workers to label images
ØPeer grading (of, e.g., essays) on MOOC
ØElicit ratings for various entities (e.g., on Yelp or Google)
ØAnd many other applications…
8
ØWe (the designer) elicit information from population ØCannot or too costly to know ground truth
ØAgents/experts may misreport
Challenge: cannot verify the report/prediction Solution: let multiple agents compete for the same task, and score them against each other (thus the name “peer prediction”) Where else did we see a similar idea?
9
ØElicit Alice’s and Bob’s truthful rating 𝐵, 𝐶 about UVA dinning
) 𝐵, 𝐶 = [𝐼, 𝑀] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝐼] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝑀] = 0.02
Let’s try to understand this distribution …
Ø It is symmetric among Alice and Bob Ø 𝑄 𝐵 = 𝐼 = 0.5 + 0.24 = 0.74
Ø 𝑄 𝐵 = 𝐼|𝐶 = 𝐼 =
>(?@A,B@A) >(B@A)
=
C.D C.EF = GD HE
Ø 𝑄 𝐵 = 𝐼|𝐶 = 𝑀 = >(?@A,B@I)
>(B@I)
= C.GF
C.GJ = KG KH
10
ØElicit Alice’s and Bob’s truthful rating 𝐵, 𝐶 about UVA dinning
) 𝐵, 𝐶 = [𝐼, 𝑀] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝐼] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝑀] = 0.02
GD HE ; 𝑄 𝐵 = 𝐼|𝐶 = 𝑀 = KG KH
Q: What are some natural peer comparison and rewarding mechanisms?
ØOne simple idea is to reward agreement
𝐵 , L 𝐶 (may misreport)
𝐵 = L 𝐶 , otherwise reward 0
11
ØElicit Alice’s and Bob’s truthful rating 𝐵, 𝐶 about UVA dinning
) 𝐵, 𝐶 = [𝐼, 𝑀] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝐼] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝑀] = 0.02
GD HE ; 𝑄 𝐵 = 𝐼|𝐶 = 𝑀 = KG KH
Q: What are some natural peer comparison and rewarding mechanisms?
ØOne simple idea is to reward agreement
𝐵 , L 𝐶 (may misreport)
𝐵 = L 𝐶 , otherwise reward 0
ØDoes this work?
12
ØElicit Alice’s and Bob’s truthful rating 𝐵, 𝐶 about UVA dinning
) 𝐵, 𝐶 = [𝐼, 𝑀] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝐼] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝑀] = 0.02
GD HE ; 𝑄 𝐵 = 𝐼|𝐶 = 𝑀 = KG KH
Q: What are some natural peer comparison and rewarding mechanisms?
ØOne simple idea is to reward agreement
𝐵 , L 𝐶 (may misreport)
𝐵 = L 𝐶 , otherwise reward 0
ØDoes this work?
Truthful report is not an equilibrium!
13
ØElicit Alice’s and Bob’s truthful rating 𝐵, 𝐶 about UVA dinning
) 𝐵, 𝐶 = [𝐼, 𝑀] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝐼] = 0.24; 𝑄 𝐵, 𝐶 = [𝑀, 𝑀] = 0.02
GD HE ; 𝑄 𝐵 = 𝐼|𝐶 = 𝑀 = KG KH
Q: What are some natural peer comparison and rewarding mechanisms?
ØBoth players always report 𝐼 (i.e., L
𝐵 = L 𝐶 = 𝐼) is a Nash Equ.
ØWhy?
possible
14
Ø Eliciting Information without Verification Ø Equilibrium Concept and Peer Prediction Mechanism Ø Bayesian Truth Serum
15
ØTwo experts Alice and Bob, each holding a signal 𝐵 ∈ {𝐵K, ⋯ , 𝐵O}
and 𝐶 ∈ {𝐶K, ⋯ , 𝐶P} respectively
ØWe would like to elicit Alice’s and Bob’s true signals
A seemingly richer but equivalent model
ØWe want to estimate distribution of random var 𝐹 ØJoint prior distribution 𝑞 of (𝐵, 𝐶, 𝐹) is publicly known
ØGoal: elicit 𝐵, 𝐶 to refine our estimation of 𝐹
16
Eliciting signals vs distributions
ØIn prediction markets, we asked experts to report distributions ØHere, could have done the same thing
A seemingly richer but equivalent model
ØWe want to estimate distribution of random var 𝐹 ØJoint prior distribution 𝑞 of (𝐵, 𝐶, 𝐹) is publicly known
ØGoal: elicit 𝐵, 𝐶 to refine our estimation of 𝐹
17
Eliciting signals vs distributions
ØIn prediction markets, we asked experts to report distributions ØHere, could have done the same thing
≠ 𝑞 𝐹 𝐵′ for any 𝐵 ≠ 𝐵′
A seemingly richer but equivalent model
ØWe want to estimate distribution of random var 𝐹 ØJoint prior distribution 𝑞 of (𝐵, 𝐶, 𝐹) is publicly known
ØGoal: elicit 𝐵, 𝐶 to refine our estimation of 𝐹
18
Eliciting signals vs distributions
ØIn prediction markets, we asked experts to report distributions ØHere, could have done the same thing
≠ 𝑞 𝐹 𝐵′ for any 𝐵 ≠ 𝐵′
ØDrawback: have to assume an accurate and known prior
A seemingly richer but equivalent model
ØWe want to estimate distribution of random var 𝐹 ØJoint prior distribution 𝑞 of (𝐵, 𝐶, 𝐹) is publicly known
ØGoal: elicit 𝐵, 𝐶 to refine our estimation of 𝐹
19
Info Elicitation Mechanisms and Equilibrium
ØRecall, we elicit info by asking Alice’s and Bob’s signal L
𝐵 , L 𝐶
ØAs before, will design rewards 𝑠
?(L
𝐵 , L 𝐶 ) and 𝑠
B(L
𝐵 , L 𝐶 )
ØAlice’s action is a report strategy 𝜏
? 𝐵 ∈ {𝐵K, ⋯ , 𝐵O} [Bob similar]
? and 𝑠 B so
that there is a good pure equilibrium
? 𝐵 = 𝐵, 𝜏B(𝐶) = 𝐶
ØThen, what outcome is expected to occur? ØGenerally, it is a Bayesian Nash equilibrium (BNE)
à equilibrium outcome
20
Info Elicitation Mechanisms and Equilibrium
ØRecall, we elicit info by asking Alice’s and Bob’s signal L
𝐵 , L 𝐶
ØAs before, will design rewards 𝑠
?(L
𝐵 , L 𝐶 ) and 𝑠
B(L
𝐵 , L 𝐶 )
ØAlice’s action is a report strategy 𝜏
? 𝐵 ∈ {𝐵K, ⋯ , 𝐵O} [Bob similar]
? 𝐵 , 𝜏B(𝐶) is a Bayesian Nash equilibrium if the
following holds 𝔽B|? 𝑠
? 𝜏 ? 𝐵 , 𝜏B 𝐶
≥ 𝔽B|? 𝑠
? 𝜏Z? 𝐵 , 𝜏B 𝐶
, ∀𝐵 𝔽?|B 𝑠
B 𝜏 ? 𝐵 , 𝜏B 𝐶
≥ 𝔽?|B 𝑠
B 𝜏 ? 𝐵 , 𝜏′B 𝐶
, ∀𝐶.
We say it is a strict BNE if both “≥” are “>”
21
ØDesign objective: choose 𝑠
?, 𝑠 B so that truth-telling is an Equ.
Any ideas? Ø Use proper scoring rules, but don’t know signal distributions… Ø Alice’s signal can be used to estimate a distribution of Bob’s signal, and vice versa
22
Note: step 2 relies on the prior distribution 𝑞 Information Elicitation without Verification
“Parameter”: any strict proper scoring rule 𝑇(𝑗; 𝑞)
1.
Elicit Alice’s signal L 𝐵 and Bob’s signal L 𝐶
2.
Calculate 𝑞 ̅
? = dist of 𝐶 conditioned on
̅ 𝐵, and similarly 𝑞 h
B
3.
Award Alice 𝑠
? L
𝐵 , L 𝐶 = 𝑇( L 𝐶 ; 𝑞 ̅
?) and Bob 𝑠 B L
𝐵 , L 𝐶 = 𝑇( L 𝐵 ; 𝑞 h
B)
23
Information Elicitation without Verification
“Parameter”: any strict proper scoring rule 𝑇(𝑗; 𝑞)
1.
Elicit Alice’s signal L 𝐵 and Bob’s signal L 𝐶
2.
Calculate 𝑞 ̅
? = dist of 𝐶 conditioned on
̅ 𝐵, and similarly 𝑞 h
B
3.
Award Alice 𝑠
? L
𝐵 , L 𝐶 = 𝑇( L 𝐶 ; 𝑞 ̅
?) and Bob 𝑠 B L
𝐵 , L 𝐶 = 𝑇( L 𝐵 ; 𝑞 h
B)
Proof: show 𝜏
? 𝐵 = 𝐵 is a best response to 𝜏B(𝐶) = 𝐶, and vice versa
Ø If Bob reports 𝐶 truthfully, Alice receives 𝑇(𝐶; 𝑞 ̅
?) by reporting ̅
𝐵 Ø With true signal 𝐵, what is Alice’s best response report ̅ 𝐵?
? to be exactly her true belief of dist. of 𝐶
𝐵 = 𝐵.
24
ØMechanism is only described for two experts, but no difficult to
generalize to 𝑜 experts
ØSerious issues are the following
Issue 1: there are many other equilibria in the game
ØDinning rating example with slightly different numbers
0.1; 𝑄 𝐵, 𝐶 = [𝑀, 𝐼] = 0.1; 𝑄 𝐵, 𝐶 = [𝑀, 𝑀] = 0.4
ØBoth always report 𝐼 is also an equilibrium
?) for whatever
true 𝐵
𝐵 = 𝐼 makes 𝑞 ̅
? 𝐼 = 𝑄 𝐶 = 𝐼
̅ 𝐵 = 𝐼 = 4/5
𝐵 = 𝑀 makes 𝑞 ̅
? 𝐼 = 𝑄 𝐶 = 𝐼
̅ 𝐵 = 𝑀 = 1/5
25
ØMechanism is only described for two experts, but no difficult to
generalize to 𝑜 experts
ØSerious issues are the following
Issue 1: there are many other equilibria in the game
ØMore generally, reporting quantities that are easy to coordinate
likely forms an equilibrium
ØThis is a fundamental issue of peer prediction
Open question: how to design mechanisms where truth- telling is unique (or the most plausible) equilibrium
26
ØMechanism is only described for two experts, but no difficult to
generalize to 𝑜 experts
ØSerious issues are the following
Issue 2: Designer has to know the joint distribution of (𝐵, 𝐶) Ø Not very realistic, as designer usually has little knowledge Ø But, there are remedies for this
27
Ø Eliciting Information without Verification Ø Equilibrium Concept and Peer Prediction Mechanism Ø Bayesian Truth Serum
28
Designed for a Special yet Realistic Setting
ØWe, the designer, want to predict distribution of 𝐹 Ø𝑜 experts, each 𝑗 has a signal 𝑇j ∼ 𝑞(𝑇|𝐹) i.i.d.
ØObjective: elicit true signals 𝑇K, ⋯ , 𝑇O
Key design ideas
29
Designed for a Special yet Realistic Setting
ØWe, the designer, want to predict distribution of 𝐹 Ø𝑜 experts, each 𝑗 has a signal 𝑇j ∼ 𝑞(𝑇|𝐹) i.i.d.
ØObjective: elicit true signals 𝑇K, ⋯ , 𝑇O
Key design ideas
Ø Cannot compute posterior distribution conditioned on any expert’s signal anymore, but still need it to score him Ø So, will elicit both his signal and his posterior belief of others’ signals
30
Bayesian Truth Serum [Prelec, Science’04]
The Protocol
1. For each 𝑗, elicit her signal L 𝑇j and her prediction h
𝑞j ∈ Δ|m| of the
distribution of any other expert’s signal (agents are i.i.d. a-priori) 2. Calculate (geometric) mean prediction ̅
𝑞 where log ̅
𝑞m =
K O ∑j log h
𝑞m
j for any signal 𝑇
3. Compute ̅
𝜇 to the empirical distribution of reported signals L
𝑇j’s. 4. Reward agent 𝑗 the following (𝐻 is any proper scoring rule) log
h 𝜇h
ms
̅ 𝑞 ̅
ms
+ 𝔽m∼L
t 𝐻(𝑇; ̅
𝑞j)
31
Bayesian Truth Serum [Prelec, Science’04]
The Protocol
1. For each 𝑗, elicit her signal L 𝑇j and her prediction h
𝑞j ∈ Δ|m| of the
distribution of any other expert’s signal (agents are i.i.d. a-priori) 2. Calculate (geometric) mean prediction ̅
𝑞 where log ̅
𝑞m =
K O ∑j log h
𝑞m
j for any signal 𝑇
3. Compute ̅
𝜇 to the empirical distribution of reported signals L
𝑇j’s. 4. Reward agent 𝑗 the following (𝐻 is any proper scoring rule) log
h 𝜇h
ms
̅ 𝑞 ̅
ms
+ 𝔽m∼L
t 𝐻(𝑇; ̅
𝑞j) Score of 𝑗’s signal report 𝑇j (good if ̅ 𝜇 ̅
ms ≥
̅ 𝑞 ̅
ms)
Ø That is, 𝑗’s reported type is surprisingly more common than predicted probability
32
Bayesian Truth Serum [Prelec, Science’04]
The Protocol
1. For each 𝑗, elicit her signal L 𝑇j and her prediction h
𝑞j ∈ Δ|m| of the
distribution of any other expert’s signal (agents are i.i.d. a-priori) 2. Calculate (geometric) mean prediction ̅
𝑞 where log ̅
𝑞m =
K O ∑j log h
𝑞m
j for any signal 𝑇
3. Compute ̅
𝜇 to the empirical distribution of reported signals L
𝑇j’s. 4. Reward agent 𝑗 the following (𝐻 is any proper scoring rule) log
h 𝜇h
ms
̅ 𝑞 ̅
ms
+ 𝔽m∼L
t 𝐻(𝑇; ̅
𝑞j) Score of 𝑗’s prediction ̅ 𝑞j, against the true signal distribution ̅ 𝜇 Ø By properness, want ̅ 𝑞j to be close to ̅ 𝜇
33
Bayesian Truth Serum [Prelec, Science’04]
equilibrium in the previous protocol. Ø That is, expert 𝑗 should report his true signal 𝑇j and his true posterior belief of other expert’s signals Ø 𝑜 → ∞ is needed because in that case ̅ 𝜇 → the exact signal distribution (under truthful signal report)
Ø Proof is a bit intricate (see the Science paper) Ø Very insightful, particularly, the design of rewarding “surprisingly common” signals, which is not clear before at all Ø The issue of existence of multiple equilibria is still there
Haifeng Xu
University of Virginia hx4ad@virginia.edu