Inherent Trade-Offs in Algorithmic Fairness Instructor: Haifeng Xu - - PowerPoint PPT Presentation

inherent trade offs in algorithmic fairness
SMART_READER_LITE
LIVE PREVIEW

Inherent Trade-Offs in Algorithmic Fairness Instructor: Haifeng Xu - - PowerPoint PPT Presentation

CS6501: T opics in Learning and Game Theory (Fall 2019) Inherent Trade-Offs in Algorithmic Fairness Instructor: Haifeng Xu COMPAS: A Risk Prediction T ool to Criminal Justice Correctional Offender Management Profiling for Alternative


slide-1
SLIDE 1

CS6501: T

  • pics in Learning and Game Theory

(Fall 2019) Inherent Trade-Offs in Algorithmic Fairness

Instructor: Haifeng Xu

slide-2
SLIDE 2

2

COMPAS: A Risk Prediction T

  • ol to Criminal Justice

ØCorrectional Offender Management Profiling for Alternative

Sanctions (COMPAS)

  • Used by states of New York, Wisconsin, Cali, Florida, etc.
  • A software that assesses likelihood of a defendant of reoffending

ØStill many issues

  • Not interpretable
  • Low accuracy
  • Bias/unfairness
slide-3
SLIDE 3

3

COMPAS: A Risk Prediction T

  • ol to Criminal Justice

ØCorrectional Offender Management Profiling for Alternative

Sanctions (COMPAS)

  • Used by states of New York, Wisconsin, Cali, Florida, etc.
  • A software that assesses likelihood of a defendant of reoffending

ØStill many issues

  • Not interpretable
  • Low accuracy
  • Bias/unfairness (this lecture)
slide-4
SLIDE 4

4

COMPAS: A Risk Prediction T

  • ol to Criminal Justice

ØIn a ProPublica investigation of the algorithm…

“…blacks are almost twice as likely as whites to be labeled a higher risk but not actually re-offend” -- unequal false positive rate “… whites are much more likely than blacks to be labeled lower-risk but go on to commit other crimes” -- unequal false negative rate Algorithms seem unfair!!

slide-5
SLIDE 5

5

Other Examples

ØAdvertising and commercial contents

Searching names that are likely assigned to black babies generates more ads suggestive of an arrest

slide-6
SLIDE 6

6

Other Examples

ØAdvertising and commercial contents

  • If a male and female user are equally interested in a product, will they

be equally likely to be shown an ad for it?

  • Will women in aggregate be shown ads for lower-paying jobs?

ØMedical testing and diagnosis

  • Will treatment be applied uniformly across different groups of

patients?

ØHiring or admission

  • Will students or job candidates from different groups be admitted with

equal probability?

Ø…

slide-7
SLIDE 7

7

Why Algorithms May Be “Unfair”?

ØAlgorithms may encode pre-existing bias

  • E.g., British Nationality act program, designed to automate evaluation
  • f new UK citizens
  • It accurately reflects tenets of the law “a man is the father of only his

legitimate children, whereas a woman is the mother of all her children, legitimate or not”

slide-8
SLIDE 8

8

Why Algorithms May Be “Unfair”?

ØAlgorithms may encode pre-existing bias

  • Easier to handle
slide-9
SLIDE 9

9

Why Algorithms May Be “Unfair”?

ØAlgorithms may encode pre-existing bias

  • Easier to handle

ØAlgorithms may create bias when serving its own objective

  • E.g., search engines try to show your favorite contents but not the

most fair contents

ØInput data are biased

  • E.g., ML may classify based on sensitive features in biased data
  • Can we simply remove these sensitive features during training?

ØBiased algorithm may get biased feedback and further strengthen

the issue

slide-10
SLIDE 10

10

Why Algorithms May Be “Unfair”?

ØAlgorithms may encode pre-existing bias

  • Easier to handle

ØAlgorithms may create bias when serving its own objective

  • E.g., search engines try to show your favorite contents but not the

most fair contents

ØInput data are biased

  • E.g., ML may classify based on sensitive features in biased data
  • Can we simply remove these sensitive features during training?

ØBiased algorithm may get biased feedback and further strengthen

the issue This lecture: there is another reason – some basic definitions of fairness are intrinsically not compatible

slide-11
SLIDE 11

11

The Problem of Predicting Risk Scores

ØIn many applications, we classify whether people possess some

property by predicting a score based on their features

  • Criminal justice
  • Loan lending
  • University admission

ØNext: an abstract model to capture this process

slide-12
SLIDE 12

12

The Problem of Predicting Risk Scores

ØThere is a collection of people, each of whom is either a positive

  • r negative instance
  • Positive/negative describe the true label of each individual

positive negative

slide-13
SLIDE 13

13

The Problem of Predicting Risk Scores

ØThere is a collection of people, each of whom is either a positive

  • r negative instance
  • Positive/negative describe the true label of each individual

ØEach person has an associated feature vector 𝜏

  • 𝑞# = fraction of people with 𝜏 who are positive

positive negative 𝜏 𝜏 𝑞# = 1/3 𝜏

slide-14
SLIDE 14

14

The Problem of Predicting Risk Scores

ØThere is a collection of people, each of whom is either a positive

  • r negative instance
  • Positive/negative describe the true label of each individual

ØEach person has an associated feature vector 𝜏

  • 𝑞# = fraction of people with 𝜏 who are positive

ØEach person belongs to one of two groups

positive negative 𝜏 𝜏 Group 1 Group 2 𝜏 publicly known 𝑞# = 1/3

slide-15
SLIDE 15

15

The Problem of Predicting Risk Scores

ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”

  • Naturally, the score should only depend on 𝜏, not individual’s group
slide-16
SLIDE 16

16

The Problem of Predicting Risk Scores

ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”

  • Naturally, the score should only depend on 𝜏, not individual’s group

ØThe score assignment process: put 𝜏 into bins (possibly randomly)

  • Only depend on 𝜏 (label is unknown in advance)

. . . . . . bin 𝑐 score 𝑤* bin 1 score 𝑤+ 𝜏

slide-17
SLIDE 17

17

The Problem of Predicting Risk Scores

ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”

  • Naturally, the score should only depend on 𝜏, not individual’s group

ØThe score assignment process: put 𝜏 into bins (possibly randomly)

  • Only depend on 𝜏 (label is unknown in advance)
  • Example 1: assign all 𝜏 to the same bin; give that bin score 𝑞#
  • Example 2: assign all people to one bin; give score 1

. . . . . . bin 𝑐 score 𝑤* bin 1 score 𝑤+ 𝜏

slide-18
SLIDE 18

18

The Problem of Predicting Risk Scores

ØTask: assign risk score to each individual ØObjective: accuracy (of course) and “fair”

  • Naturally, the score should only depend on 𝜏, not individual’s group

ØThe score assignment process: put 𝜏 into bins (possibly randomly)

  • Only depend on 𝜏 (label is unknown in advance)
  • Example 1: assign all 𝜏 to the same bin; give that bin score 𝑞#
  • Example 2: assign all people to one bin; give score 1

. . . . . . bin 𝑐 score 𝑤* bin 1 score 𝑤+ 𝜏 Note: may have very bad accuracy but good fairness, as they are different

slide-19
SLIDE 19

19

Well…What Does “Fair” Really Mean?

ØA very subjective perception ØYet, for algorithm design, need a concrete and objective definition Ø> 20 different definitions of fairness so far

  • See a survey paper “Fairness Definitions Explained”

ØThis raises many questions

  • Are they all reasonable? Can we satisfy all of them?
  • Which one/subset of them we should use when designing algorithms?
  • Do I have to sacrifice accuracy to achieve fairness?
slide-20
SLIDE 20

20

Well…What Does “Fair” Really Mean?

ØA very subjective perception ØYet, for algorithm design, need a concrete and objective definition Ø> 20 different definitions of fairness so far

  • See a survey paper “Fairness Definitions Explained”

ØThis raises many questions

  • Are they all reasonable? Can we satisfy all of them?
  • Which one/subset of them we should use when designing algorithms?
  • Do I have to sacrifice accuracy to achieve fairness?

Some basic definitions of fairness are already not compatible, regardless how much accuracy you are willing to sacrifice

slide-21
SLIDE 21

21

Fairness Def 1: Calibration

Definition [Calibration within groups]. For each bin 𝑐, let

  • 𝑂0,* = # of people assigned to 𝑐 from group 𝑢
  • 𝑜0,* = # of positive people assigned to 𝑐 from group 𝑢

We should have 𝑜0,* = 𝑤* ⋅ 𝑂0,* for each 𝑢, 𝑐

slide-22
SLIDE 22

22

Fairness Def 1: Calibration

Definition [Calibration within groups]. For each bin 𝑐, let

  • 𝑂0,* = # of people assigned to 𝑐 from group 𝑢
  • 𝑜0,* = # of positive people assigned to 𝑐 from group 𝑢

We should have 𝑜0,* = 𝑤* ⋅ 𝑂0,* for each 𝑢, 𝑐

Group 1 𝑤* = 0.75

slide-23
SLIDE 23

23

Fairness Def 1: Calibration

Definition [Calibration within groups]. For each bin 𝑐, let

  • 𝑂0,* = # of people assigned to 𝑐 from group 𝑢
  • 𝑜0,* = # of positive people assigned to 𝑐 from group 𝑢

We should have 𝑜0,* = 𝑤* ⋅ 𝑂0,* for each 𝑢, 𝑐

Group 1 𝑤* = 0.75

slide-24
SLIDE 24

24

Fairness Def 1: Calibration

Definition [Calibration within groups]. For each bin 𝑐, let

  • 𝑂0,* = # of people assigned to 𝑐 from group 𝑢
  • 𝑜0,* = # of positive people assigned to 𝑐 from group 𝑢

We should have 𝑜0,* = 𝑤* ⋅ 𝑂0,* for each 𝑢, 𝑐

In practice, we do not know who are positive so cannot check the condition, but the definition still applies Group 1 𝑤* = 0.75

slide-25
SLIDE 25

25

Fairness Def 2: Balance of Negative Class

𝐹 𝑤 𝜏 | σ negative and in group 1 = 𝐹 𝑤 𝜏 | σ negative and in group 2

Definition [Balance of Negative Class]. Average scores assigned to people of group 1 who are negative should be the same as average scores assigned to people of group 2 who are negative.

positive negative Group 1 Group 2 𝑤(𝜏) 𝑤(𝜏′)

slide-26
SLIDE 26

26

Fairness Def 3: Balance of Positive Class

𝐹 𝑤 𝜏 | σ positive and in group 1 = 𝐹 𝑤 𝜏 | σ positive and in group 2

Definition [Balance of Negative Class]. Average scores assigned to people of group 1 who are positive should be the same as average scores assigned to people of group 2 who are positive.

positive negative Group 1 Group 2 𝑤(𝜏) 𝑤(𝜏′)

slide-27
SLIDE 27

27

Is It Possible to Achieve All Three?

slide-28
SLIDE 28

28

Is It Possible to Achieve All Three?

Yes: Example 1

Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏

𝜏 𝜏 Group 1 Group 2 𝜏 𝜏′ 𝜏′ 𝜏′

slide-29
SLIDE 29

29

Is It Possible to Achieve All Three?

Yes: Example 1

Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0

and all 𝜏 with 𝑞# = 1 to bin 1

𝜏 𝜏 Group 1 Group 2 𝜏 𝜏′ 𝜏′ 𝜏′ bin 1 𝑤+ = 1 bin 0 𝑤M = 0

slide-30
SLIDE 30

30

Is It Possible to Achieve All Three?

Yes: Example 1

Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0

and all 𝜏 with 𝑞# = 1 to bin 1 Claim: This score assignment satisfies all 3 fairness defs.

slide-31
SLIDE 31

31

Is It Possible to Achieve All Three?

Yes: Example 1

Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0

and all 𝜏 with 𝑞# = 1 to bin 1 Claim: This score assignment satisfies all 3 fairness defs.

ØCalibration:

Group 1 𝑤+ = 1

yes, all the ratio is 1 or 0 for each group

slide-32
SLIDE 32

32

Is It Possible to Achieve All Three?

Yes: Example 1

Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0

and all 𝜏 with 𝑞# = 1 to bin 1 Claim: This score assignment satisfies all 3 fairness defs.

ØCalibration: yes, all the ratio is 1 or 0 for each group ØBalance of positive class: yes, both groups have average score 1 ØBalance of negative class: yes, both groups have average score 0

slide-33
SLIDE 33

33

Is It Possible to Achieve All Three?

Yes: Example 1

Ø𝑞# = 1 𝑝𝑠 0 for all 𝜏 ØTwo bins with 𝑤M = 0 and 𝑤+ = 1; assign all 𝜏 with 𝑞# = 0 to bin 0

and all 𝜏 with 𝑞# = 1 to bin 1 Claim: This score assignment satisfies all 3 fairness defs. Caveats

ØBut, this is not really a realistic setting… Ø𝑞# = 0 𝑝𝑠 1 means we know for sure each individual’s label

slide-34
SLIDE 34

34

Is It Possible to Achieve All Three?

Yes: Example 2

ØAverage 𝑞# (over 𝜏’s) is the same among two groups

𝜏′ 𝜏 Group 1 Group 2 𝜏 𝜏′ 𝜏 𝜏′ 𝐹 𝑞#|𝜏 ∈ Group 1 𝐹 𝑞#|𝜏 ∈ Group 2 =

slide-35
SLIDE 35

35

Is It Possible to Achieve All Three?

Yes: Example 2

ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#

𝜏′ 𝜏 Group 1 Group 2 𝜏 𝜏′ 𝜏 𝜏′ 𝐹 𝑞#|𝜏 ∈ Group 1 𝐹 𝑞#|𝜏 ∈ Group 2 = 𝑤 = 𝐹 𝑞#|𝜏 ∈ Group 1

slide-36
SLIDE 36

36

Is It Possible to Achieve All Three?

Yes: Example 2

ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#

Claim: This score assignment satisfies all 3 fairness defs.

slide-37
SLIDE 37

37

Is It Possible to Achieve All Three?

Yes: Example 2

ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#

Claim: This score assignment satisfies all 3 fairness defs.

ØCalibration: yes, since 𝑤 = average 𝑞# is exactly the probability of

positive instances in both groups

ØBalance of positive class: trivial, as scores are the same ØBalance of negative class: trivial as well

slide-38
SLIDE 38

38

Is It Possible to Achieve All Three?

Yes: Example 2

ØAverage 𝑞# (over 𝜏’s) is the same among two groups ØOne bin, with 𝑤 equal the above average 𝑞#

Claim: This score assignment satisfies all 3 fairness defs. Caveats

ØBut, this score assignment is not useful and has low accuracy ØThere may exist a more accurate score assignment in this case

that still satisfy three definitions

  • Bad news: it is NP-hard to find
slide-39
SLIDE 39

39

Inherent Trade-offs of Algorithmic Fairness

ØThe two (degenerated) examples are the only cases where you

can possibly satisfy all three fairness definitions Theorem: For the problem of risk score assignment, if there is a risk assignment that satisfies all the three fairness definitions before, the problem must be one of the previous two example cases.

slide-40
SLIDE 40

40

Proof Sketch

ØAssume there is a score assignment satisfying all three defs ØWill derive contradictions, unless the instance is the previous

degenerated settings

slide-41
SLIDE 41

41

Proof Sketch

Definition [Calibration]. For each bin 𝑐, let

  • 𝑂0,* = # of people assigned to 𝑐 from group 𝑢
  • 𝑜0,* = # of positive people assigned to 𝑐 from group 𝑢

We should have 𝑜0,* = 𝑤* ⋅ 𝑂0,* for each 𝑢, 𝑐

Notations

Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢

Calibration condition implies

ØTotal score of all group-t people in bin 𝑐 (i.e., 𝑤* ⋅ 𝑂0,*) equal expected

number of positive group-t people in bin 𝑐 (i.e., 𝑜0,*)

slide-42
SLIDE 42

42

Proof Sketch

Notations

Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢

Calibration condition implies

ØTotal score of all group-t people in bin 𝑐 (i.e., 𝑤* ⋅ 𝑂0,*) equal expected

number of positive group-t people in bin 𝑐 (i.e., 𝑜0,*)

ØSumming over all bins à total score of all group-t people equals

expected number of positive group-t people

slide-43
SLIDE 43

43

Proof Sketch

Notations

Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢

Another way to calculate total scores

Ø𝑦 = average score of a person in negative class Ø𝑧 = average score of a person in positive class

slide-44
SLIDE 44

44

Proof Sketch

Notations

Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢

Another way to calculate total scores

Ø𝑦 = average score of a person in negative class Ø𝑧 = average score of a person in positive class ØTotal score in group 𝑢 is 𝑧 𝑂0 − 𝑜0 + 𝑦𝑜0 ØRe-arranging 𝑦 = (1 − 𝑧)

TU VUWTU

= 𝑜0 by calibration Group 1 Group 2

slide-45
SLIDE 45

45

Proof Sketch

Notations

Ø𝑂0 = total number of people in group 𝑢 Ø𝑜0 = total number of positive people in group 𝑢

Another way to calculate total scores

Ø𝑦 = average score of a person in negative class Ø𝑧 = average score of a person in positive class ØTotal score in group 𝑢 is 𝑧 𝑂0 − 𝑜0 + 𝑦𝑜0 ØRe-arranging 𝑦 = (1 − 𝑧)

TU VUWTU

ØTo make sure 𝑦, 𝑧 are the same for both groups, the two lines must

intersect

  • Unless slopes are the same, only intersect at (0,1)

= 𝑜0 by calibration Group 1 Group 2

slide-46
SLIDE 46

46

Can Achieve Two Definitions

Ø“Equality of Opportunity in Supervised Learning [NeurIPS’16]”

  • Can achieve balance of positive and negative class, but no

requirement for calibration

  • Objective: find most accurate prediction subject to fairness constraints

Ø“On Fairness and Calibration [NeurIPS’17]”

  • Can achieve calibration and any linear combination of balance of

positive and negative class

slide-47
SLIDE 47

47

Similar Negative Results

ØShow similar negative results, but for classification

“Fair prediction with disparate impact: A study of bias in recidivism prediction instruments” “Algorithmic decision making and the cost of fairness”

slide-48
SLIDE 48

48

Happy Thanksgiving