CS6501: T
- pics in Learning and Game Theory
Inherent Trade-Offs in Algorithmic Fairness Instructor: Haifeng Xu - - PowerPoint PPT Presentation
CS6501: T opics in Learning and Game Theory (Fall 2019) Inherent Trade-Offs in Algorithmic Fairness Instructor: Haifeng Xu COMPAS: A Risk Prediction T ool to Criminal Justice Correctional Offender Management Profiling for Alternative
2
ØCorrectional Offender Management Profiling for Alternative
ØStill many issues
3
ØCorrectional Offender Management Profiling for Alternative
ØStill many issues
4
ØIn a ProPublica investigation of the algorithm…
5
ØAdvertising and commercial contents
6
ØAdvertising and commercial contents
ØMedical testing and diagnosis
ØHiring or admission
Ø…
7
ØAlgorithms may encode pre-existing bias
8
ØAlgorithms may encode pre-existing bias
9
ØAlgorithms may encode pre-existing bias
ØAlgorithms may create bias when serving its own objective
ØInput data are biased
ØBiased algorithm may get biased feedback and further strengthen
10
ØAlgorithms may encode pre-existing bias
ØAlgorithms may create bias when serving its own objective
ØInput data are biased
ØBiased algorithm may get biased feedback and further strengthen
11
ØIn many applications, we classify whether people possess some
ØNext: an abstract model to capture this process
12
ØThere is a collection of people, each of whom is either a positive
13
ØThere is a collection of people, each of whom is either a positive
ØEach person has an associated feature vector 𝜏
14
ØThere is a collection of people, each of whom is either a positive
ØEach person has an associated feature vector 𝜏
ØEach person belongs to one of two groups
15
ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”
16
ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”
ØThe score assignment process: put 𝜏 into bins (possibly randomly)
17
ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”
ØThe score assignment process: put 𝜏 into bins (possibly randomly)
18
ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”
ØThe score assignment process: put 𝜏 into bins (possibly randomly)
19
ØA very subjective perception ØYet, for algorithm design, need a concrete and objective definition Ø> 20 different definitions of fairness so far
ØThis raises many questions
20
ØA very subjective perception ØYet, for algorithm design, need a concrete and objective definition Ø> 20 different definitions of fairness so far
ØThis raises many questions
21
22
23
24
25
26
27
28
Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏
29
Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0
30
Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0
31
Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0
ØCalibration:
32
Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0
ØCalibration: yes, all the ratio is 1 or 0 for each group ØBalance of positive class: yes, both groups have average score 1 ØBalance of negative class: yes, both groups have average score 0
33
Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0
ØBut, this is not really a realistic setting… Ø𝑞# = 0 𝑝𝑠 1 means we know for sure each individual’s label
34
ØAverage 𝑞# (over 𝜏’s) is the same among two groups
35
ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#
36
ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#
37
ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#
ØCalibration: yes, since 𝑤 = average 𝑞# is exactly the probability of
ØBalance of positive class: trivial, as scores are the same ØBalance of negative class: trivial as well
38
ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#
ØBut, this score assignment is not useful and has low accuracy ØThere may exist a more accurate score assignment in this case
39
ØThe two (degenerated) examples are the only cases where you
40
ØAssume there is a score assignment satisfying all three defs ØWill derive contradictions, unless the instance is the previous
41
Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢
ØTotal score of all group-t people in bin 𝑐 (i.e., 𝑤* ⋅ 𝑂0,*) equal expected
42
Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢
ØTotal score of all group-t people in bin 𝑐 (i.e., 𝑤* ⋅ 𝑂0,*) equal expected
ØSumming over all bins à total score of all group-t people equals
43
Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢
Ø𝑦 = average score of a person in negative class Ø𝑧 = average score of a person in positive class
44
Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢
Ø𝑦 = average score of a person in negative class Ø𝑧 = average score of a person in positive class ØTotal score in group 𝑢 is 𝑧 𝑂0 − 𝑜0 + 𝑦𝑜0 ØRe-arranging 𝑦 = (1 − 𝑧)
TU VUWTU
45
Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢
Ø𝑦 = average score of a person in negative class Ø𝑧 = average score of a person in positive class ØTotal score in group 𝑢 is 𝑧 𝑂0 − 𝑜0 + 𝑦𝑜0 ØRe-arranging 𝑦 = (1 − 𝑧)
TU VUWTU
ØTo make sure 𝑦, 𝑧 are the same for both groups, the two lines must
46
Ø“Equality of Opportunity in Supervised Learning [NeurIPS’16]”
Ø“On Fairness and Calibration [NeurIPS’17]”
47
ØShow similar negative results, but for classification
48