Bounding the fairness and accuracy of classifiers from population - - PowerPoint PPT Presentation

bounding the fairness and accuracy of classifiers from
SMART_READER_LITE
LIVE PREVIEW

Bounding the fairness and accuracy of classifiers from population - - PowerPoint PPT Presentation

Bounding the fairness and accuracy of classifiers from population statistics ICML 2020 Sivan Sabato and Elad Yom-Tov Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 1 / 15 The 1-slide summary We show how to study a


slide-1
SLIDE 1

Bounding the fairness and accuracy of classifiers from population statistics ICML 2020

Sivan Sabato and Elad Yom-Tov

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 1 / 15

slide-2
SLIDE 2

The 1-slide summary

We show how to study a classifier without even a black box access to the classifier and without validation data.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 2 / 15

slide-3
SLIDE 3

The 1-slide summary

We show how to study a classifier without even a black box access to the classifier and without validation data. Our methodology makes provable inferences about classifier quality.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 2 / 15

slide-4
SLIDE 4

The 1-slide summary

We show how to study a classifier without even a black box access to the classifier and without validation data. Our methodology makes provable inferences about classifier quality. The quality combines the accuracy and the fairness of the classifier.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 2 / 15

slide-5
SLIDE 5

The 1-slide summary

We show how to study a classifier without even a black box access to the classifier and without validation data. Our methodology makes provable inferences about classifier quality. The quality combines the accuracy and the fairness of the classifier. We make inferences using a small number of aggregate statistics.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 2 / 15

slide-6
SLIDE 6

The 1-slide summary

We show how to study a classifier without even a black box access to the classifier and without validation data. Our methodology makes provable inferences about classifier quality. The quality combines the accuracy and the fairness of the classifier. We make inferences using a small number of aggregate statistics. We demonstrate in experiments a wide range of possible applications.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 2 / 15

slide-7
SLIDE 7

The 1-slide summary

We show how to study a classifier without even a black box access to the classifier and without validation data. Our methodology makes provable inferences about classifier quality. The quality combines the accuracy and the fairness of the classifier. We make inferences using a small number of aggregate statistics. We demonstrate in experiments a wide range of possible applications.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 2 / 15

slide-8
SLIDE 8

Introduction

Classifiers affect many aspects of our lives. But some of these classifiers cannot be directly validated:

◮ Unavailability of representative individual-level validation data ◮ Company of government secret: not even black-box access

What can we infer about a classifier using only aggregate statistics?

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 3 / 15

slide-9
SLIDE 9

What can we tell about an unpublished classifier?

A motivating example: A health insurance company classifies whether a client is as “at risk” for some medical condition. We do not know how this classification is done; We have no individual classification data.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 4 / 15

slide-10
SLIDE 10

What can we tell about an unpublished classifier?

A motivating example: A health insurance company classifies whether a client is as “at risk” for some medical condition. We do not know how this classification is done; We have no individual classification data. But we would still like to study the properties of the classifier:

◮ Accuracy ◮ Fairness Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 4 / 15

slide-11
SLIDE 11

What can we tell about an unpublished classifier?

A motivating example: A health insurance company classifies whether a client is as “at risk” for some medical condition. We do not know how this classification is done; We have no individual classification data. But we would still like to study the properties of the classifier:

◮ Accuracy ◮ Fairness

Can this be done with minimal information about the classifier?

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 4 / 15

slide-12
SLIDE 12

Fairness

Fairness is defined with respect to some attribute of the individual.

◮ E.g., race, age, gender, state of residence

We will be interested in attributes with several different values. A sub-population includes the individual who share the attribute value (e.g., same race/age bracket/state, etc.).

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 5 / 15

slide-13
SLIDE 13

Fairness

Fairness is defined with respect to some attribute of the individual.

◮ E.g., race, age, gender, state of residence

We will be interested in attributes with several different values. A sub-population includes the individual who share the attribute value (e.g., same race/age bracket/state, etc.). A fair classifier treats all sub-populations the same. Equalized Odds [Hardt et. al, 2016]: The false positive rate (FPR) and the false negative rate (FNR) are fixed across all sub-populations.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 5 / 15

slide-14
SLIDE 14

Using population statistics

Back to the example: Use available information

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 6 / 15

slide-15
SLIDE 15

Using population statistics

Back to the example: Use available information Size of each sub-population Prevalence rate of the condition in each sub-population Fraction of positive predictions in each sub-population.

State Population Fraction Have condition Classified as positive California 12.2% 0.3% 0.4% Texas 8.6% 1.2% 5% ... ... ... ...

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 6 / 15

slide-16
SLIDE 16

Using population statistics

Back to the example: Use available information Size of each sub-population Prevalence rate of the condition in each sub-population Fraction of positive predictions in each sub-population.

State Population Fraction Have condition Classified as positive California 12.2% 0.3% 0.4% Texas 8.6% 1.2% 5% ... ... ... ...

What is the accuracy of this classifier? What is the fairness?

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 6 / 15

slide-17
SLIDE 17

Using population statistics

Back to the example: Use available information Size of each sub-population Prevalence rate of the condition in each sub-population Fraction of positive predictions in each sub-population.

State Population Fraction Have condition Classified as positive California 12.2% 0.3% 0.4% Texas 8.6% 1.2% 5% ... ... ... ...

What is the accuracy of this classifier? What is the fairness? Without individual data, there are many possibilities:

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 6 / 15

slide-18
SLIDE 18

Using population statistics

Back to the example: Use available information Size of each sub-population Prevalence rate of the condition in each sub-population Fraction of positive predictions in each sub-population.

State Population Fraction Have condition Classified as positive California 12.2% 0.3% 0.4% Texas 8.6% 1.2% 5% ... ... ... ...

What is the accuracy of this classifier? What is the fairness? Without individual data, there are many possibilities:

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 6 / 15

slide-19
SLIDE 19

Using population statistics

Back to the example: Use available information Size of each sub-population Prevalence rate of the condition in each sub-population Fraction of positive predictions in each sub-population.

State Population Fraction Have condition Classified as positive California 12.2% 0.3% 0.4% Texas 8.6% 1.2% 5% ... ... ... ...

What is the accuracy of this classifier? What is the fairness? Without individual data, there are many possibilities:

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 6 / 15

slide-20
SLIDE 20

The relationship between accuracy and fairness

If fairness or error are constrained, this also constrains the other.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 7 / 15

slide-21
SLIDE 21

The relationship between accuracy and fairness

If fairness or error are constrained, this also constrains the other. Example:

Population Fraction Have condition Classified as positive State A 1/2 1/3 1/2 State B 1/2 2/3 2/3 Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 7 / 15

slide-22
SLIDE 22

The relationship between accuracy and fairness

If fairness or error are constrained, this also constrains the other. Example:

Population Fraction Have condition Classified as positive State A 1/2 1/3 1/2 State B 1/2 2/3 2/3 ◮ True positives:

.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 7 / 15

slide-23
SLIDE 23

The relationship between accuracy and fairness

If fairness or error are constrained, this also constrains the other. Example:

Population Fraction Have condition Classified as positive State A 1/2 1/3 1/2 State B 1/2 2/3 2/3 ◮ True positives:

.

◮ Which are the predicted positives? Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 7 / 15

slide-24
SLIDE 24

The relationship between accuracy and fairness

If fairness or error are constrained, this also constrains the other. Example:

Population Fraction Have condition Classified as positive State A 1/2 1/3 1/2 State B 1/2 2/3 2/3 ◮ True positives:

.

◮ Which are the predicted positives? ◮ Smallest error:

. Error of 12.5%, unfair.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 7 / 15

slide-25
SLIDE 25

The relationship between accuracy and fairness

If fairness or error are constrained, this also constrains the other. Example:

Population Fraction Have condition Classified as positive State A 1/2 1/3 1/2 State B 1/2 2/3 2/3 ◮ True positives:

.

◮ Which are the predicted positives? ◮ Smallest error:

. Error of 12.5%, unfair.

◮ Fair solution:

. 25% error.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 7 / 15

slide-26
SLIDE 26

Balancing accuracy and fairness

The two measures:

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-27
SLIDE 27

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-28
SLIDE 28

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-29
SLIDE 29

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Combine both desiderata: For β ∈ [0, 1], discrepancyβ := β · unfairness + (1 − β) · error,

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-30
SLIDE 30

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Combine both desiderata: For β ∈ [0, 1], discrepancyβ := β · unfairness + (1 − β) · error, β defines a trade-off between (un)fairness and error.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-31
SLIDE 31

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Combine both desiderata: For β ∈ [0, 1], discrepancyβ := β · unfairness + (1 − β) · error, β defines a trade-off between (un)fairness and error. By lower-bounding discrepancyβ, we can answer:

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-32
SLIDE 32

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Combine both desiderata: For β ∈ [0, 1], discrepancyβ := β · unfairness + (1 − β) · error, β defines a trade-off between (un)fairness and error. By lower-bounding discrepancyβ, we can answer:

◮ What is the minimal unfairness that the classifier must have,

given an upper bound on its error?

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-33
SLIDE 33

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Combine both desiderata: For β ∈ [0, 1], discrepancyβ := β · unfairness + (1 − β) · error, β defines a trade-off between (un)fairness and error. By lower-bounding discrepancyβ, we can answer:

◮ What is the minimal unfairness that the classifier must have,

given an upper bound on its error?

◮ What is the minimal error that the classifier must have,

given an upper bound on its unfairness?

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-34
SLIDE 34

Balancing accuracy and fairness

The two measures:

◮ error: Fraction of the population classified with the wrong label. ◮ unfairness: Fraction of the population treated differently than a

common baseline. We expand on this next.

Combine both desiderata: For β ∈ [0, 1], discrepancyβ := β · unfairness + (1 − β) · error, β defines a trade-off between (un)fairness and error. By lower-bounding discrepancyβ, we can answer:

◮ What is the minimal unfairness that the classifier must have,

given an upper bound on its error?

◮ What is the minimal error that the classifier must have,

given an upper bound on its unfairness?

◮ What is the minimal combined cost of this classifier? Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 8 / 15

slide-35
SLIDE 35

Quanitifying unfairness

Decompose the conditional distribution of predictions given labels:

◮ A baseline distribution which is common to all sub-populations;

FPR = α1 and FNR = α0,

◮ A nuisance distribution for each sub-population s;

FPR = α1

s and FNR = α0 s,

◮ The distribution for sub-population s is a mixture:

ηs · Nuisances + (1 − ηs) · Baseline.

η

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 9 / 15

slide-36
SLIDE 36

Quanitifying unfairness

Decompose the conditional distribution of predictions given labels:

◮ A baseline distribution which is common to all sub-populations;

FPR = α1 and FNR = α0,

◮ A nuisance distribution for each sub-population s;

FPR = α1

s and FNR = α0 s,

◮ The distribution for sub-population s is a mixture:

ηs · Nuisances + (1 − ηs) · Baseline.

◮ Define unfairness as the fraction of the population that is treated

differently from the baseline treatment =

s ηs.

η

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 9 / 15

slide-37
SLIDE 37

Quanitifying unfairness

Decompose the conditional distribution of predictions given labels:

◮ A baseline distribution which is common to all sub-populations;

FPR = α1 and FNR = α0,

◮ A nuisance distribution for each sub-population s;

FPR = α1

s and FNR = α0 s,

◮ The distribution for sub-population s is a mixture:

ηs · Nuisances + (1 − ηs) · Baseline.

◮ Define unfairness as the fraction of the population that is treated

differently from the baseline treatment =

s ηs.

◮ The decomposition to baseline and nuisance is unobserved.

η

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 9 / 15

slide-38
SLIDE 38

Quanitifying unfairness

Decompose the conditional distribution of predictions given labels:

◮ A baseline distribution which is common to all sub-populations;

FPR = α1 and FNR = α0,

◮ A nuisance distribution for each sub-population s;

FPR = α1

s and FNR = α0 s,

◮ The distribution for sub-population s is a mixture:

ηs · Nuisances + (1 − ηs) · Baseline.

◮ Define unfairness as the fraction of the population that is treated

differently from the baseline treatment =

s ηs.

◮ The decomposition to baseline and nuisance is unobserved. ◮ Set ηs to the minimum consistent with the input statistics.

η

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 9 / 15

slide-39
SLIDE 39

Quanitifying unfairness

Decompose the conditional distribution of predictions given labels:

◮ A baseline distribution which is common to all sub-populations;

FPR = α1 and FNR = α0,

◮ A nuisance distribution for each sub-population s;

FPR = α1

s and FNR = α0 s,

◮ The distribution for sub-population s is a mixture:

ηs · Nuisances + (1 − ηs) · Baseline.

◮ Define unfairness as the fraction of the population that is treated

differently from the baseline treatment =

s ηs.

◮ The decomposition to baseline and nuisance is unobserved. ◮ Set ηs to the minimum consistent with the input statistics.

η(αy, αy

s ) =

     1 − αy

s /αy

αy

s < αy

1 − (1 − αy

s )/(1 − αy)

αy

s > αy

αy

s = αy.

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.2 0.4 0.6 0.8 1 η(a,b)

b

b = 0.01 b = 0.5 b = 0.5 b = 0.75 b = 0.99

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 9 / 15

slide-40
SLIDE 40

Lower-bounding discrepancyβ

Given known FPRs and FNRs {αy

s } in each sub-population,

discrepancyβ({αy

s }) =

β · min

(α0,α1)∈[0,1]2

  • g∈G

ws

  • y∈Y

πy

s η(αy, αy s ) + (1 − β) ·

  • g∈G

ws

  • y∈Y

πy

s αy s .

ws := P(attribute value is s) πs := P(positive label | s) ˆ ps := P(positive prediction | s)

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 10 / 15

slide-41
SLIDE 41

Lower-bounding discrepancyβ

Given known FPRs and FNRs {αy

s } in each sub-population,

discrepancyβ({αy

s }) =

β · min

(α0,α1)∈[0,1]2

  • g∈G

ws

  • y∈Y

πy

s η(αy, αy s ) + (1 − β) ·

  • g∈G

ws

  • y∈Y

πy

s αy s .

ws := P(attribute value is s) πs := P(positive label | s) ˆ ps := P(positive prediction | s) We derive a lower bound on min{αy

s } discrepancyβ({αy

s }) subject to

the constraints imposed by {ws, πs, ˆ ps}.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 10 / 15

slide-42
SLIDE 42

Lower-bounding discrepancyβ

Given known FPRs and FNRs {αy

s } in each sub-population,

discrepancyβ({αy

s }) =

β · min

(α0,α1)∈[0,1]2

  • g∈G

ws

  • y∈Y

πy

s η(αy, αy s ) + (1 − β) ·

  • g∈G

ws

  • y∈Y

πy

s αy s .

ws := P(attribute value is s) πs := P(positive label | s) ˆ ps := P(positive prediction | s) We derive a lower bound on min{αy

s } discrepancyβ({αy

s }) subject to

the constraints imposed by {ws, πs, ˆ ps}. Theorem The minimum of discrepancyβ({αy

s }) subject to the constraints imposed

by {ws, πs, ˆ ps} is obtained by an assignment in a small number of

  • ne-dimensional solution sets.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 10 / 15

slide-43
SLIDE 43

Experiments: Tightness of lower bound

(In all experiments, sub-populations are defined by state of residence.)

β

β

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 11 / 15

slide-44
SLIDE 44

Experiments: Tightness of lower bound

(In all experiments, sub-populations are defined by state of residence.) We obtain a lower bound; how tight is it in practice?

β

β

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 11 / 15

slide-45
SLIDE 45

Experiments: Tightness of lower bound

(In all experiments, sub-populations are defined by state of residence.) We obtain a lower bound; how tight is it in practice? Generated hundreds of classifiers from the US Census data set. The classifiers are known and we can calculate their true properties.

β

β

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 11 / 15

slide-46
SLIDE 46

Experiments: Tightness of lower bound

(In all experiments, sub-populations are defined by state of residence.) We obtain a lower bound; how tight is it in practice? Generated hundreds of classifiers from the US Census data set. The classifiers are known and we can calculate their true properties. Left plot: Compared the lower bound on discrepancy1 ≡ unfairness with the true unfairness. Right plot: For randomly selected classifiers, the ratio between the true value and the lower bound for β ∈ [0, 1].

0% 20% 40% 60% 80% 100% >99%>90%>80%>70%>60%>50%>40%>30%>20%>10%

Percent with ratio above threshold

0.2 0.4 0.6 0.8 1

0.2 0.4 0.6 0.8 1 lower bound on discβ/true value β

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 11 / 15

slide-47
SLIDE 47

Experiments: Making inferences in the wild (1)

In the following experiments, discrepancyβ is unknown. We calculate (unfairness,error) Pareto-curves as a function of β.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 12 / 15

slide-48
SLIDE 48

Experiments: Making inferences in the wild (1)

In the following experiments, discrepancyβ is unknown. We calculate (unfairness,error) Pareto-curves as a function of β. Experiment 1: Identify if anonymous individuals have a certain cancer from their search queries in Bing. Classify as positive if user searched for said cancer. True positive rates per state from CDC data. Results lower-bound the quality of these classifiers.

0.2 0.4 0.6 0.8 1

0.2 0.4 0.6 0.8 1 Error/true positives Unfairness/true positives

Breast Cervical Colon Liver Lung Skin Stomach testicular Thyroid Bladder Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 12 / 15

slide-49
SLIDE 49

Experiments: Making inferences in the wild (2)

Experiment (2): Studied 10 pre-election polls from the 2016 US presidential elections. Treat each poll as a classifier from individual to vote. How biased are these polls in their treatment of different states?

0.05 0.1 0.15 0.2

0.02 0.04 0.06 0.08 0.1 0.12 0.14 Error/true positives Unfairness/true positives

Sep-01 (a) Sep-01 (b) Sep-13 Oct-13 Oct-26 Oct-27 Nov-06 (a) Nov-06 (b) Nov-07 (a) Nov-07 (b)

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 13 / 15

slide-50
SLIDE 50

Experiments: Making inferences in the wild (3)

Experiment (3): Compare cancer mortality rates in different states “True positive” rates: cancer mortality rates in each state “Predicted” rates: expected mortality in the state based on cancer prevalence and overall US mortality. “Classifier” maps an individual to an outcome (living/deceased) Error and unfairness can speculatively point to patterns in health care access or in cancer strains.

0.5 1 1.5 2

0.2 0.4 0.6 0.8 1 Error/true positives Unfairness/true positives

Brain Breast Colon Kidney Liver Lung Oral Pancreatic Stomach Thyroid

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 14 / 15

slide-51
SLIDE 51

Summary

We showed how a small set of aggregate statistics can be used to make strong inferences about the quality of the classifier. The methodology can be applied to a range of applications:

◮ Estimating the quality of a classifier during development stages ◮ Studying classifiers of public importance ◮ Analysis of statistical phenomena by defining an appropriate classifier

Extending this toolbox is an important research direction with many open problems.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 15 / 15

slide-52
SLIDE 52

Summary

We showed how a small set of aggregate statistics can be used to make strong inferences about the quality of the classifier. The methodology can be applied to a range of applications:

◮ Estimating the quality of a classifier during development stages ◮ Studying classifiers of public importance ◮ Analysis of statistical phenomena by defining an appropriate classifier

Extending this toolbox is an important research direction with many open problems.

Sabato & Yom-Tov (Microsoft & BGU) Bounding fairness and accuracy 15 / 15