1
Announcements
ØHW 3 and proposal due today
Announcements HW 3 and proposal due today 1 CS6501: T opics in - - PowerPoint PPT Presentation
Announcements HW 3 and proposal due today 1 CS6501: T opics in Learning and Game Theory (Fall 2019) Selling Information Instructor: Haifeng Xu Outline Bayesian Persuasion and Information Selling Sell to a Single Decision Maker
1
ØHW 3 and proposal due today
CS6501: T
(Fall 2019) Selling Information
Instructor: Haifeng Xu
3
Ø Bayesian Persuasion and Information Selling Ø Sell to a Single Decision Maker Ø Sell to Multiple Decision Makers
4
ØOne of the two primarily ways to influence agents’ behaviors
ØAccounts for a significant share in economic activities
Persuasion is the act of exploiting an informational advantage in
5
ØTwo players: a sender (she) and a receiver (he)
ØReceiver takes an action 𝑗 ∈ 𝑜 = {1,2, ⋯ , 𝑜}
ØBoth players know prior 𝑞, but sender additionally observes 𝜄
6
ØTwo players: a sender (she) and a receiver (he)
ØReceiver takes an action 𝑗 ∈ 𝑜 = {1,2, ⋯ , 𝑜}
ØBoth players know prior 𝑞, but sender additionally observes 𝜄 ØSender reveals partial information via a signaling scheme to
influence receiver’s decision and maximize her utility Definition: A signaling scheme is a mapping 𝜌: Θ → Δ; where Σ is the set of all possible signals. 𝜌 is fully described by 𝜌 𝜏, 𝜄
>∈?,@∈; where 𝜌 𝜏, 𝜄 = prob. of
sending 𝜏 when observing 𝜄 (so ∑@∈; 𝜌 𝜏, 𝜄 = 1 for any 𝜄)
7
ØSender = advisor, receiver = recruiter ØΘ = {𝑓𝑦𝑑𝑓𝑚𝑚𝑓𝑜𝑢, 𝑏𝑤𝑓𝑠𝑏𝑓}, 𝜈 𝑓𝑦𝑑𝑓𝑚𝑚𝑓𝑜𝑢 = 1/3 ØReceiver decides Hire or NotHire
ØOptimal strategy is a signaling scheme
8
ØOptimal signaling scheme is computed by an LP
Revelation Principle. There always exists an optimal signaling scheme that uses at most 𝑜(= # receiver actions) signals, where signal 𝜏L induce optimal receiver action 𝑗
9
ØProsecutor persuades judge
10
ØProsecutor persuades judge ØLobbyists persuade politicians
11
ØProsecutor persuades judge ØLobbyists persuade politicians ØElection candidates persuade voters
12
ØProsecutor persuades judge ØLobbyists persuade politicians ØElection candidates persuade voters ØSellers persuade buyers
13
ØProsecutor persuades judge ØLobbyists persuade politicians ØElection candidates persuade voters ØSellers persuade buyers
14
ØProsecutor persuades judge ØLobbyists persuade politicians ØElection candidates persuade voters ØSellers persuade buyers ØExecutives persuade stockholders
15
ØProsecutor persuades judge ØLobbyists persuade politicians ØElection candidates persuade voters ØSellers persuade buyers ØExecutives persuade stockholders Ø. . .
Many persuasion models built upon Bayesian persuasion
ØPersuading many receivers, voters, attackers, drivers on road
network, buyers in auctions, etc..
ØPrivate vs public persuasion ØSelling information is also a variant
16
ØSender = seller, Receiver = buyer who is a decision maker ØBuyer takes an action 𝑗 ∈ 𝑜 = {1, ⋯ , 𝑜} ØBuyer has a utility function 𝑣(𝑗, 𝜄; 𝜕) where
17
ØSender = seller, Receiver = buyer who is a decision maker ØBuyer takes an action 𝑗 ∈ 𝑜 = {1, ⋯ , 𝑜} ØBuyer has a utility function 𝑣(𝑗, 𝜄; 𝜕) where
Remarks:
Ø 𝑣, 𝑞, 𝑔 are public knowledge ØAssume 𝜄, 𝜕 are independent ØIn mechanism design, seller also does not know buyer’s value
18
ØSender = seller, Receiver = buyer who is a decision maker ØBuyer takes an action 𝑗 ∈ 𝑜 = {1, ⋯ , 𝑜} ØBuyer has a utility function 𝑣(𝑗, 𝜄; 𝜕) where
Remarks:
Ø 𝑣, 𝑞, 𝑔 are public knowledge ØAssume 𝜄, 𝜕 are independent ØIn mechanism design, seller also does not know buyer’s value
Q: How to price the item if seller knowns buyer’s value of it?
19
ØSender = seller, Receiver = buyer who is a decision maker ØBuyer takes an action 𝑗 ∈ 𝑜 = {1, ⋯ , 𝑜} ØBuyer has a utility function 𝑣(𝑗, 𝜄; 𝜕) where
ØSeller observes the state 𝜄; Buyer knows his private type 𝜕 ØSeller would like to sell her information about 𝜄 to maximize
revenue Key differences from Bayesian persuasion
ØSeller does not have a utility fnc – instead maximize revenue ØBuyer here has private info 𝜕, which is unknown to seller
20
Ø Bayesian Persuasion and Information Selling Ø Sell to a Single Decision Maker Ø Sell to Multiple Decision Makers
21
Warm-up: What if Buyer Has no Private Info
Ø 𝑣(𝑗, 𝜄; 𝜕) where sate 𝜄 ∼ 𝑒𝑗𝑡𝑢. 𝑞 and buyer type 𝜕 ∼ 𝑒𝑗𝑡𝑢. 𝑔 ØWhen seller also observes 𝜕 . . .
Q: How to sell information optimally?
22
Warm-up: What if Buyer Has no Private Info
Ø 𝑣(𝑗, 𝜄; 𝜕) where sate 𝜄 ∼ 𝑒𝑗𝑡𝑢. 𝑞 and buyer type 𝜕 ∼ 𝑒𝑗𝑡𝑢. 𝑔 ØWhen seller also observes 𝜕 . . .
Q: How to sell information optimally?
ØSeller knows exactly how much the buyer values “any amount” of
her information à should charge him just that amount
23
Warm-up: What if Buyer Has no Private Info
Ø 𝑣(𝑗, 𝜄; 𝜕) where sate 𝜄 ∼ 𝑒𝑗𝑡𝑢. 𝑞 and buyer type 𝜕 ∼ 𝑒𝑗𝑡𝑢. 𝑔 ØWhen seller also observes 𝜕 . . .
Q: How to sell information optimally?
ØSeller knows exactly how much the buyer values “any amount” of
her information à should charge him just that amount
ØHow to charge the most?
Payment = ∑>∈? 𝑞 𝜄 ⋅ [max
L
𝑣(𝑗, 𝜄; 𝜕)] − max
L
∑>∈? 𝑞 𝜄 ⋅ 𝑣(𝑗, 𝜄; 𝜕)
24
Warm-up: What if Buyer Has no Private Info
Ø 𝑣(𝑗, 𝜄; 𝜕) where sate 𝜄 ∼ 𝑒𝑗𝑡𝑢. 𝑞 and buyer type 𝜕 ∼ 𝑒𝑗𝑡𝑢. 𝑔 ØWhen seller also observes 𝜕 . . .
Q: How to sell information optimally?
ØSeller knows exactly how much the buyer values “any amount” of
her information à should charge him just that amount
ØHow to charge the most?
Payment = ∑>∈? 𝑞 𝜄 ⋅ [max
L
𝑣(𝑗, 𝜄; 𝜕)] − max
L
∑>∈? 𝑞 𝜄 ⋅ 𝑣(𝑗, 𝜄; 𝜕) Buyer expected utility if learns 𝜄
25
Warm-up: What if Buyer Has no Private Info
Ø 𝑣(𝑗, 𝜄; 𝜕) where sate 𝜄 ∼ 𝑒𝑗𝑡𝑢. 𝑞 and buyer type 𝜕 ∼ 𝑒𝑗𝑡𝑢. 𝑔 ØWhen seller also observes 𝜕 . . .
Q: How to sell information optimally?
ØSeller knows exactly how much the buyer values “any amount” of
her information à should charge him just that amount
ØHow to charge the most?
Payment = ∑>∈? 𝑞 𝜄 ⋅ [max
L
𝑣(𝑗, 𝜄; 𝜕)] − max
L
∑>∈? 𝑞 𝜄 ⋅ 𝑣(𝑗, 𝜄; 𝜕) Buyer expected utility without knowing 𝜄
26
Warm-up: What if Buyer Has no Private Info
Ø 𝑣(𝑗, 𝜄; 𝜕) where sate 𝜄 ∼ 𝑒𝑗𝑡𝑢. 𝑞 and buyer type 𝜕 ∼ 𝑒𝑗𝑡𝑢. 𝑔 ØWhen seller also observes 𝜕 . . .
Q: How to sell information optimally?
ØSeller knows exactly how much the buyer values “any amount” of
her information à should charge him just that amount
ØHow to charge the most?
Payment = ∑>∈? 𝑞 𝜄 ⋅ [max
L
𝑣(𝑗, 𝜄; 𝜕)] − max
L
∑>∈? 𝑞 𝜄 ⋅ 𝑣(𝑗, 𝜄; 𝜕)
More interesting and realistic is when buyer has private info
27
The class of mechanisms is too broad
ØThe mechanism will: (1) elicit private info from buyer; (2) reveal
info based on realized 𝜄; (3) charge buyer
ØMay interact with buyer for many rounds ØBuyer may misreport his private info of 𝜕
28
The class of mechanisms is too broad . . . but, at the end of the day, the buyer of type 𝜕 is charged some amount 𝑢] in expectation and learns a posterior belief about 𝜄
29
The class of mechanisms is too broad . . . but, at the end of the day, the buyer of type 𝜕 is charged some amount 𝑢] in expectation and learns a posterior belief about 𝜄
Theorem (Revelation Principle). Any information selling mechanism can be “simulated” by a direct and truthful revelation mechanism:
Ø Proof: similar to proof of revelation principle for mechanism design Ø Optimal mechanism reduces to an incentive compatible menu 𝑢], 𝜌] ]
30
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
31
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕
32
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕
depends only
33
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕
depends on 𝜌], but will not change due to our way of merging
34
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕 So merging signals in 𝜌] retains this constraint
depends on 𝜌], but will not change due to our way of merging
35
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕a ≥ 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 Incentive compatibility constraint for any 𝜕a (≠ 𝜕)
36
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕a ≥ 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 Incentive compatibility constraint for any 𝜕a (≠ 𝜕)
This will change! Why?
37
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕a ≥ 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 Incentive compatibility constraint for any 𝜕a (≠ 𝜕)
This will change! Why?
38
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 ≥ 𝑉] 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕′ Incentive compatibility constraint for 𝜕 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕a ≥ 𝑉]b 𝑠𝑓𝑞𝑝𝑠𝑢 𝜕 Incentive compatibility constraint for any 𝜕a (≠ 𝜕)
Key idea: this term will only decrease since 𝜕′ gets less info due to merging of signals
39
Signaling scheme 𝜌] is still complicated
ØFor any fixed buyer type 𝜕, how many signals needed for 𝜌]?
response is not valid any more – why?
Theorem (Simplifying Signaling Schemes). There always exists an optimal incentive compatible menu 𝑢], 𝜌] ], such that 𝜌] uses at most 𝑜 signals with 𝜏L recommending action 𝑗
Such an information-selling mechanism is like consulting – buyer reports type 𝜕, seller charges him 𝑢]
40
ØWill be incentive compatible – reporting true 𝜕 is optimal ØThe recommended action is guaranteed to be the optimal
action for buyer 𝜕 given his information
Ø 𝑢], 𝜌] ] is public knowledge, and computed by LP
The Consulting Mechanism 1. Elicit buyer type 𝜕 2. Charge buyer 𝑢] 3. Observe realized state 𝜄 and recommend action 𝑗 to the buyer with probability 𝜌](𝜏L, 𝜄)
41
ØWill be incentive compatible – reporting true 𝜕 is optimal ØThe recommended action is guaranteed to be the optimal
action for buyer 𝜕 given his information
Ø 𝑢], 𝜌] ] is public knowledge, and computed by LP
The Consulting Mechanism 1. Elicit buyer type 𝜕 2. Charge buyer 𝑢] 3. Observe realized state 𝜄 and recommend action 𝑗 to the buyer with probability 𝜌](𝜏L, 𝜄)
computed by the following program.
42
Optimal 𝑠
], 𝜌] ] can be computed by a convex program
43
Optimal 𝑠
], 𝜌] ] can be computed by a convex program
Expected revenue
44
Optimal 𝑠
], 𝜌] ] can be computed by a convex program
Reporting true 𝜕 is optimal
45
Optimal 𝑠
], 𝜌] ] can be computed by a convex program
Similar to constraints in persuasion
46
Optimal 𝑠
], 𝜌] ] can be computed by a convex program
Ø A convex fnc of variables Ø Can be converted to an LP
47
Ø Bayesian Persuasion and Information Selling Ø Sell to a Single Decision Maker Ø Sell to Multiple Decision Makers
48
ØFor single decision maker, more information always helps
ØA fundamental challenge for selling to multiple buyers is that
information does not necessarily help them
49
Example: More Information Hurts Buyers
ØInsurance industry: insurance company and customer
ØTwo types of customers: Healthy and Unhealthy
ØSeller is an information holder, who knows whether any customer
is healthy or not
Sell Not Sell Buy (-10, 10) (-0, 0) Not Buy (0 , 0) (0 , 0)
Insurance company customer
Healthy customer Sell Not Sell Buy (-10, -50) (-110, 0) Not Buy (-111 , 0) (-111 , 0)
Insurance company
Unhealthy customer
50
Example: More Information Hurts Buyers
Ø Customer and insurance company will look at expectation
Sell Not Sell Buy (-10, 10) (-0, 0) Not Buy (0 , 0) (0 , 0)
Insurance company customer
Healthy customer, prob = 0.9 Sell Not Sell Buy (-10, -50) (-110, 0) Not Buy (-111 , 0) (-111 , 0)
Insurance company
Unhealthy customer
Q: What happens without seller’s information ?
Sell Not Sell Buy (-10, 4) (-11 , 0) Not Buy (-11.1, 0) (-11.1, 0)
51
Example: More Information Hurts Buyers
E.g., customer wants to buy info from seller to decide whether he should buyer insurance or not Q: What if seller tells (only) customer her health status ?
Sell Not Sell Buy (-10, 10) (-0, 0) Not Buy (0 , 0) (0 , 0)
Insurance company customer
Healthy customer, prob = 0.9 Sell Not Sell Buy (-10, -50) (-110, 0) Not Buy (-111 , 0) (-111 , 0)
Insurance company
Unhealthy customer
52
Example: More Information Hurts Buyers
ØIf Healthy, customer will not buy ØIf Unhealthy, customer will buy ØCustomer’s reaction reveals his healthy status
Q: What if seller tells (only) customer her health status ?
Sell Not Sell Buy (-10, 10) (-0, 0) Not Buy (0 , 0) (0 , 0)
Insurance company customer
Healthy customer, prob = 0.9 Sell Not Sell Buy (-10, -50) (-110, 0) Not Buy (-111 , 0) (-111 , 0)
Insurance company
Unhealthy customer
53
Example: More Information Hurts Buyers
ØIf Healthy, customer will not buy ØIf Unhealthy, customer will buy ØCustomer’s reaction reveals his healthy status
Q: What if seller tells (only) customer her health status ?
Sell Not Sell Buy (-10, 10) (-0, 0) Not Buy (0 , 0) (0 , 0)
Insurance company customer
Healthy customer, prob = 0.9 Sell Not Sell Buy (-10, -50) (-110, 0) Not Buy (-111 , 0) (-111 , 0)
Insurance company
Unhealthy customer
à utility (0,0) for both à Will not sell, utility (-110,0)
54
Example: More Information Hurts Buyers
ØIf Healthy, customer will not buy ØIf Unhealthy, customer will buy ØCustomer’s reaction reveals his healthy status ØIn expectation (-11, 0)
Q: What if seller tells (only) customer her health status ? à utility (0,0) for both à Will not sell, utility (-110,0) Recall previously (-10,4)
Sell Not Sell Buy (-10, 10) (-0, 0) Not Buy (0 , 0) (0 , 0)
Insurance company customer
Healthy customer, prob = 0.9 Sell Not Sell Buy (-10, -50) (-110, 0) Not Buy (-111 , 0) (-111 , 0)
Insurance company
Unhealthy customer
Haifeng Xu
University of Virginia hx4ad@virginia.edu