CS6501: T
- pics in Learning and Game Theory
(Fall 2019) How Can Classifiers Induce Right Efforts?
Instructor: Haifeng Xu
CS6501: T opics in Learning and Game Theory (Fall 2019) How Can - - PowerPoint PPT Presentation
CS6501: T opics in Learning and Game Theory (Fall 2019) How Can Classifiers Induce Right Efforts? Instructor: Haifeng Xu Outline Motivations and Model Examples and Results 2 Decisions and Incentives Often today, ML is used to assist
CS6501: T
(Fall 2019) How Can Classifiers Induce Right Efforts?
Instructor: Haifeng Xu
2
Ø Motivations and Model Ø Examples and Results
3
Often today, ML is used to assist decisions about human beings
4
ØEducation
Often today, ML is used to assist decisions about human beings
5
ØEducation ØWhen a measure becomes a target, gaming behaviors happen
(Goodhart’s Law) Often today, ML is used to assist decisions about human beings
6
ØEducation ØWhen a measure becomes a target, gaming behaviors happen
(Goodhart’s Law)
ØMany other applications: recommender systems, hiring, finance…
reviews or checkins
Often today, ML is used to assist decisions about human beings
7
ØEducation ØWhen a measure becomes a target, gaming behaviors happen
(Goodhart’s Law)
ØMany other applications: recommender systems, hiring, finance…
reviews or checkins
ØParticularly an issue when transparency is required
Often today, ML is used to assist decisions about human beings
Chief scientist of Obama 2012 Campaign
8
Strategic Behaviors Goal/score (determined by some measure)
9
Strategic Behaviors Goal/score (determined by some measure)
Desirable behavior
10
Strategic Behaviors
Undesirable behavior
Goal/score (determined by some measure)
11
ØSome strategic behaviors are desirable, and some are not
I think it’s best to. . . distinguish between seven different types of test preparation: Working more effectively; Teaching more; Working harder; Reallocation; Alignment; Coaching; Cheating. The first three are what proponents of high-stakes testing want to see
12
ØSome strategic behaviors are desirable, and some are not
The Main Question How to design decision rules to induce desirable strategic behaviors?
ØUsually not possible to keep the rule confidential ØShould not simply use a rule that cannot be affected at all ØSo, this requires careful design
13
Ø𝑛 available actions (e.g., study hard, cheating) Ø𝑜 different features (e.g., HW grade, midterm grade) ØEach unit effort on action 𝑘 results in 𝛽%&(≥ 0) increase in feature 𝑗
1 𝑘 𝑛
. . . . . .
𝐺
.
𝐺& 𝐺
/
. . . . . .
𝛽.. 𝛽%. 𝛽%& 𝛽0%
14
ØAgent’s action: allocation (𝑦., ⋯ , 𝑦0) of 1 unit of effort to actions
1 𝑘 𝑛
. . . . . .
𝐺
.
𝐺& 𝐺
/
. . . . . .
𝛽.. 𝛽%. 𝛽%& 𝛽0%
15
ØAgent’s action: allocation (𝑦., ⋯ , 𝑦0) of 1 unit of effort to actions
𝐺& = 𝑔
&(∑% 𝑦%𝛽%&)
(an increasing concave fnc)
𝑦. 𝑦% 𝑦0
. . . . . .
𝐺
.
𝐺& 𝐺
/
. . . . . .
𝛽.. 𝛽%. 𝛽%& 𝛽0%
∑% 𝑦% ≤ 1
16
ØAgent’s action: allocation (𝑦., ⋯ , 𝑦0) of 1 unit of effort to actions
𝐺& = 𝑔
&(∑% 𝑦%𝛽%&)
(an increasing concave fnc)
ØPrincipal’s action: design the evaluation rule 𝐼(𝐺
., ⋯ , 𝐺 /)
𝑦. 𝑦% 𝑦0
. . . . . .
𝐺
.
𝐺& 𝐺
/
. . . . . .
𝛽.. 𝛽%. 𝛽%& 𝛽0% 𝐼
Evaluation rule 𝐼(𝐺
., ⋯ , 𝐺 /)
∑% 𝑦% ≤ 1
17
ØAgent’s action: allocation (𝑦., ⋯ , 𝑦0) of 1 unit of effort to actions
𝐺& = 𝑔
&(∑% 𝑦%𝛽%&)
(an increasing concave fnc)
ØPrincipal’s action: design the evaluation rule 𝐼(𝐺
., ⋯ , 𝐺 /)
ØPrincipal has a desirable effort profile 𝑦∗ (e.g., 𝑦∗ = “work hard”) ØAgent goal: choose 𝑦 to maximize 𝐼
Q: Can the principal design 𝐼 to induce her desirable 𝑦∗?
18
Relation to problems we studied before
ØThis is a Stackelberg game
ØThis is a mechanism design problem
ØMore generally, this a principal-agent mechanism design problem
Q: Can the principal design 𝐼 to induce her desirable 𝑦∗?
19
Ø Motivations and Model Ø Examples and Results
20
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 1 2 2 1 𝐼
𝑦∗ = (0, 1, 0) Q: Can the principal induce the desirable 𝑦∗ = (0,1,0)?
cheating studying copying
𝐼 = 0.6 𝐺= + 0.4𝐺
D
21
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 1 2 2 1 𝐼
𝑦∗ = (0, 1, 0) Q: Can the principal induce the desirable 𝑦∗ = (0,1,0)?
ØAns: Yes
spend it on studying
cheating studying copying
𝐼 = 0.6 𝐺= + 0.4𝐺
D
22
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 2 1 1 1.5 𝐼
𝐼 = 0.6 𝐺= + 0.4𝐺
D
Q: What about this setting?
cheating studying copying
𝑦∗ = (0, 1, 0)
23
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 2 1 1 1.5 𝐼
𝐼 = 0.6 𝐺= + 0.4𝐺
D
Q: What about this setting?
ØAns: No
cheating studying copying
𝑦∗ = (0, 1, 0)
24
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 2 1 1 1.5 𝐼
𝐼 = 0.4 𝐺= + 0.6𝐺
D
Q: What about changing 𝐼 to our class’s rule?
cheating studying copying
𝑦∗ = (0, 1, 0)
25
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 2 1 1 1.5 𝐼
𝐼 = 0.4 𝐺= + 0.6𝐺
D
Q: What about changing 𝐼 to our class’s rule?
ØAns: Yes
cheating studying copying
𝑦∗ = (0, 1, 0)
26
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 3 1 1 3 𝐼
𝐼 = 0.4 𝐺= + 0.6𝐺
D
Q: What about these effort transition values?
cheating studying copying
𝑦∗ = (0, 1, 0)
27
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 3 1 1 3 𝐼
𝐼 = 0.4 𝐺= + 0.6𝐺
D
Q: What about these effort transition values?
ØAns: No, regardless of what 𝐼 you choose
GH ; , 0, 𝑦< + IH ; ) is better for agent
cheating studying copying
𝑦∗ = (0, 1, 0)
28
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 𝐼
Q: In general, when would it be impossible to induce 𝑦∗?
cheating studying copying
𝑦∗ = (0, 1, 0)
𝛽.= 𝛽;= 𝛽;> 𝛽<>
𝐼 = 𝛾=𝐺= + 𝛾>𝐺
D
29
𝑦. 𝑦; 𝑦< 𝐺= 𝐺> 𝐼
Q: In general, when would it be impossible to induce 𝑦∗?
ØWith 𝐶 = 1 effort on studying, we get 𝐺=, 𝐺> = (𝛽;=, 𝛽;>) ØIf ∃ (𝑦., 𝑦;, 𝑦<) such that: (1) 𝑦. + 𝑦; + 𝑦< < 1; but (2) 𝑦.𝛽.= + 𝑦;𝛽;= ≥
𝛽;= and 𝑦;𝛽;> + 𝑦<𝛽<> ≥ 𝛽;>, then cannot induce effort on studying
cheating studying copying
𝑦∗ = (0, 1, 0)
𝛽.= 𝛽;= 𝛽;> 𝛽<>
𝐼 = 𝛾=𝐺= + 𝛾>𝐺
D
30
ØLet’s focus on the special case 𝑦∗ = 𝑓
% for some 𝑘
ØPrevious argument shows a necessary condition
There is no 𝑦., ⋯ , 𝑦0 ≥ 0 such that: 1. ∑% 𝑦% < 1 2. 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅) Note: 𝑦 here is a row vector
Which Effort Profile Can Be Incentivized, and How?
31
ØLet’s focus on the special case 𝑦∗ = 𝑓
% for some 𝑘
ØPrevious argument shows a necessary condition
There is no 𝑦., ⋯ , 𝑦0 ≥ 0 such that: 1. ∑% 𝑦% < 1 2. 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅) Note: 𝑦 here is a row vector Define 𝜆% ≔ min
I
∑% 𝑦% subject to (1) 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅); (2) 𝑦 ≥ 0. A necessary condition is 𝜆% ≥ 1.
Which Effort Profile Can Be Incentivized, and How?
32
ØLet’s focus on the special case 𝑦∗ = 𝑓
% for some 𝑘
ØPrevious argument shows a necessary condition
Note: 𝜆% ≤ 1 always because 𝑦 = 𝑓
% is feasible
Define 𝜆% ≔ min
I
∑% 𝑦% subject to (1) 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅); (2) 𝑦 ≥ 0. A necessary condition is 𝜆% ≥ 1.
Which Effort Profile Can Be Incentivized, and How?
33
ØLet’s focus on the special case 𝑦∗ = 𝑓
% for some 𝑘
ØPrevious argument shows a necessary condition
Note: 𝜆% ≤ 1 always because 𝑦 = 𝑓
% is feasible
Define 𝜆% ≔ min
I
∑% 𝑦% subject to (1) 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅); (2) 𝑦 ≥ 0. A necessary condition is 𝜆% = 1.
Which Effort Profile Can Be Incentivized, and How?
34
Which Effort Profile Can Be Incentivized, and How?
ØLet’s focus on the special case 𝑦∗ = 𝑓
% for some 𝑘
ØPrevious argument shows a necessary condition
Define 𝜆% ≔ min
I
∑% 𝑦% subject to (1) 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅); (2) 𝑦 ≥ 0. A necessary condition is 𝜆% = 1. Theorem: (1) There is a way to incentivize 𝑓
% if and only if
𝜆% = 1. (2) Whenever 𝑓
% can be incentivized, there is a linear
𝐼 of form 𝐼 = ∑& 𝛾& 𝐺& that incentivizes 𝑓
%.
35
Which Effort Profile Can Be Incentivized, and How?
ØLet’s focus on the special case 𝑦∗ = 𝑓
% for some 𝑘
ØPrevious argument shows a necessary condition
Define 𝜆% ≔ min
I
∑% 𝑦% subject to (1) 𝑦 ⋅ 𝛽 ≥ 𝛽(𝑘,⋅); (2) 𝑦 ≥ 0. A necessary condition is 𝜆% = 1. Theorem: (1) There is a way to incentivize 𝑓
% if and only if
𝜆% = 1. (2) Whenever 𝑓
% can be incentivized, there is a linear
𝐼 of form 𝐼 = ∑& 𝛾& 𝐺& that incentivizes 𝑓
%.
Proof
ØWe know if 𝜆% < 1, we cannot incentivize 𝑓
%, so 𝜆% = 1 is necessary
ØTo prove sufficiency, we construct a linear 𝐼 that indeed induce 𝑓
% when
𝜆% = 1
36
%
ØConsider 𝐼 = ∑& 𝛾& 𝐺&, agent’s optimization problem
max
I∈XY 𝐼 = ∑& 𝛾& ⋅ 𝑔 & ∑Z 𝑦Z𝛽Z&
Value of feature 𝑗
37
%
ØConsider 𝐼 = ∑& 𝛾& 𝐺&, agent’s optimization problem
max
I∈XY 𝐼 = ∑& 𝛾& ⋅ 𝑔 & ∑Z 𝑦Z𝛽Z&
ØWhen would the optimal solution be 𝑦∗ = 𝑓
%?
[I] |I_I∗ ≥ [\ [I]` |I_I∗ for all 𝑘′ (verify it after class)
∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
Eq.(1)
38
%
ØConsider 𝐼 = ∑& 𝛾& 𝐺&, agent’s optimization problem
max
I∈XY 𝐼 = ∑& 𝛾& ⋅ 𝑔 & ∑Z 𝑦Z𝛽Z&
ØWhen would the optimal solution be 𝑦∗ = 𝑓
%?
[I] |I_I∗ ≥ [\ [I]` |I_I∗ for all 𝑘′ (verify it after class)
∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
Eq.(1)
Q: Given 𝜐% = 1, do there exist 𝛾 ≠ 0 so that Eq. (1) holds? Ø Eq (1) is also a set of linear constraints on 𝛾 Ø Ans: yes, through an elegant duality argument
39
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
𝐵 𝑘,⋅ ⋅ 𝛾= ≥ 𝐵 𝑘b,⋅ ⋅ 𝛾=, ∀𝑘′ 𝛾 ≥ 0 ∃𝛾 ≠ 0 such that
40
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
𝐵 𝑘,⋅ ⋅ 𝛾= ≥ 𝐵 𝑘b,⋅ ⋅ 𝛾=, ∀𝑘′ 𝛾 ≥ 0 ∃𝛾 ≠ 0 such that s.t. 𝟐 ≥ 𝐵 ⋅ 𝛾=, ∀𝑙 𝛾 ≥ 0 max
i
𝐵 𝑘,⋅ ⋅ 𝛾=
41
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
s.t. 𝟐 ≥ 𝐵 ⋅ 𝛾=, ∀𝑙 𝛾 ≥ 0 max
i
𝐵 𝑘,⋅ ⋅ 𝛾=
s.t. 𝑧 ⋅ 𝐵 ≥ 𝐵(𝑘, : ) 𝑧 ≥ 0 min
m
𝟐 ⋅ 𝑧= Dual LP
42
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
s.t. 𝟐 ≥ 𝐵 ⋅ 𝛾=, ∀𝑙 𝛾 ≥ 0 max
i
𝐵 𝑘,⋅ ⋅ 𝛾=
s.t. 𝑧 ⋅ 𝐵 ≥ 𝐵(𝑘, : ) 𝑧 ≥ 0 min
m
𝟐 ⋅ 𝑧= Dual LP Ø The constraint is ∑Z 𝑧Z 𝛽Z& ⋅ 𝑔
& b ≥ 𝛽%& ⋅ 𝑔 & b, ∀𝑗
i.e., ∑Z 𝑧Z 𝛽Z& ≥ 𝛽%&, ∀𝑗
43
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
s.t. 𝟐 ≥ 𝐵 ⋅ 𝛾=, ∀𝑙 𝛾 ≥ 0 max
i
𝐵 𝑘,⋅ ⋅ 𝛾=
s.t. 𝑧 ⋅ 𝐵 ≥ 𝐵(𝑘, : ) 𝑧 ≥ 0 min
m
𝟐 ⋅ 𝑧= Dual LP Ø The constraint is ∑Z 𝑧Z 𝛽Z& ⋅ 𝑔
& b ≥ 𝛽%& ⋅ 𝑔 & b, ∀𝑗
i.e., ∑Z 𝑧Z 𝛽Z& ≥ 𝛽%&, ∀𝑗 Ø Dual opt is exactly the def of 𝜆%(= 1)
44
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
s.t. 𝟐 ≥ 𝐵 ⋅ 𝛾=, ∀𝑙 𝛾 ≥ 0 max
i
𝐵 𝑘,⋅ ⋅ 𝛾=
s.t. 𝑧 ⋅ 𝐵 ≥ 𝐵(𝑘, : ) 𝑧 ≥ 0 min
m
𝟐 ⋅ 𝑧= Dual LP Ø The constraint is ∑Z 𝑧Z 𝛽Z& ⋅ 𝑔
& b ≥ 𝛽%& ⋅ 𝑔 & b, ∀𝑗
i.e., ∑Z 𝑧Z 𝛽Z& ≥ 𝛽%&, ∀𝑗 Ø Dual opt is exactly the def of 𝜆%(= 1) Primal opt = 1 𝛾 can be easily constructed
45
ØGoal:∑& 𝛾& ⋅ 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& ≥ ∑& 𝛾& ⋅ 𝛽%`& ⋅ 𝑔 & b ∑Z 𝑦Z ∗𝛽Z& , ∀𝑘′
ØLet 𝐵%,& = 𝛽%& ⋅ 𝑔
& b ∑Z 𝑦Z ∗𝛽Z& which is a constant (𝑦∗ is given)
ØNeed to check the linear system
s.t. 𝟐 ≥ 𝐵 ⋅ 𝛾=, ∀𝑙 𝛾 ≥ 0 max
i
𝐵 𝑘,⋅ ⋅ 𝛾=
s.t. 𝑧 ⋅ 𝐵 ≥ 𝐵(𝑘, : ) 𝑧 ≥ 0 min
m
𝟐 ⋅ 𝑧= Dual LP Ø The constraint is ∑Z 𝑧Z 𝛽Z& ⋅ 𝑔
& b ≥ 𝛽%& ⋅ 𝑔 & b, ∀𝑗
i.e., ∑Z 𝑧Z 𝛽Z& ≥ 𝛽%&, ∀𝑗 Ø Dual opt is exactly the def of 𝜆%(= 1) Primal opt = 1 𝛾 can be easily constructed
46
ØSimilar conclusion holds with similar proof ØIt turns out that the condition depends on 𝑇∗, the support of 𝑦∗
Theorem: (1) There is a way to incentivize 𝑦∗ if and only if 𝜆o∗ = 1 for some suitably defined 𝜆o∗. (2) Whenever 𝑦∗ can be incentivized, there is a linear 𝐼 that incentivizes 𝑦∗.
47
ØPreviously, principal has a single 𝑦∗ to induce
ØA natural optimization version of the problem
48
ØPreviously, principal has a single 𝑦∗ to induce
ØA natural optimization version of the problem
ØProblem: maximize (𝑦) subject to 𝑦 is incentivizable
Theorem: The above problem is NP-hard, even when is concave.
Open question: Ø What kind of can be optimized? Linear? Ø What kind effort transition graph makes the problem more tractable?
Haifeng Xu
University of Virginia hx4ad@virginia.edu