Why is My Classifier Discriminatory?
Irene Y. Chen, Fredrik D. Johansson, David Sontag Massachusetts Institute of Technology (MIT)
NeurIPS 2018, Poster #120 Thurs 12/6 10:45am β 12:45pm @ 210 & 230
Why is My Classifier Discriminatory? Irene Y. Chen, Fredrik D. - - PowerPoint PPT Presentation
Why is My Classifier Discriminatory? Irene Y. Chen, Fredrik D. Johansson, David Sontag Massachusetts Institute of Technology (MIT) NeurIPS 2018, Poster #120 Thurs 12/6 10:45am 12:45pm @ 210 & 230 It is su y to make a surprisi singly y
Irene Y. Chen, Fredrik D. Johansson, David Sontag Massachusetts Institute of Technology (MIT)
NeurIPS 2018, Poster #120 Thurs 12/6 10:45am β 12:45pm @ 210 & 230
Source: Shutterstock
0.16 0.18 0.20 0.22
White Other Hispanic Black Asian
resource allocation.
resource allocation.
noise.
resource allocation.
noise.
augmentation and training data collection to fix unfairness.
Model
Ligett, 2017
2016; Corbett-Davies et al, 2017
Model
Ligett, 2017
2016; Corbett-Davies et al, 2017
Data
Model
Ligett, 2017
2016; Corbett-Davies et al, 2017
Data
2013; Feldman et al, 2015
Model
Ligett, 2017
2016; Corbett-Davies et al, 2017
Data
2013; Feldman et al, 2015
True data function
Learned model
Learned model
Learned model True data function
Learned model True data function
Learned model
Learned model
Learned model Orange dot model error
Learned model Orange dot model error Blue dot model error
True data function π = π. πππ
π = π β π
True data function π = π. πππ
π = π β π
Learned model
Learned model Orange dot model error
Learned model Orange dot model error Blue dot model error
Learned model Orange dot model error Blue dot model error
We define fairness in the context of loss like false positive rate, false negative rate, etc. For example, zero-one loss for data D and prediction π +: πΏ- π +, π, πΈ βΆ= π2 π + β π π΅ = π) We can then formalize unfairness as group differences. Ξ 9 π + βΆ= | πΏ; β πΏ<| We rely on accurate Y labels and focus on algorithmic error
We define fairness in the context of loss like false positive rate, false negative rate, etc. For example, zero-one loss for data D and prediction π +: πΏ- π +, π, πΈ βΆ= π2 π + β π π΅ = π) We can then formalize unfairness as group differences. Ξ 9 π + βΆ= | πΏ; β πΏ<| We rely on accurate Y labels and focus on algorithmic error.
Theorem 1: For error over group a given predictor π
+: πΏΜ - π + = πΆ 9- π + + π 9
+) + π C- Note that π- indicates the expectation of π- over X and data D. Accordingly, the expected discrimination level Ξ 9: = |πΏ; C β πΏΜ <| can be decomposed into differences in bias, differences in variance, and differences in noise. Ξ 9 = (πΆ 9; β πΆ 9< + (π 9;βπ 9<) + (π C;βπ C<)|
Theorem 1: For error over group a given predictor π
+: πΏΜ - π + = πΆ 9- π + + π 9
+) + π C- Note that π- indicates the expectation of π- over X and data D. Accordingly, the expected discrimination level Ξ 9: = |πΏ; C β πΏΜ <| can be decomposed into differences in bias, differences in variance, and differences in noise. Ξ 9 = (πΆ 9; β πΆ 9< + (π 9;βπ 9<) + (π C;βπ C<)|
0.16 0.18 0.20 0.22
Zero-one loss
White Other Hispanic Black Asian
Asian Black Hispanic Other White
1. We found statistically significant racial differences in zero-one loss.
5000 10000 15000
Training data size
0.27 0.25 0.23 0.21 0.19 Zero-one loss
Asian Black Hispanic Other White
1. We found statistically significant racial differences in zero-one loss. 2. By subsampling data, we fit inverse power laws to estimate the benefit of more data and reducing variance.
Cancer patients Cardiac patients
0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Error enrichment
1106 1877 619 2564 19711 736 2100 1211 4181 17649
Asian Black Hispanic Other White
1. We found statistically significant racial differences in zero-one loss. 2. By subsampling data, we fit inverse power laws to estimate the benefit of more data and reducing variance. 3. Using topic modeling, we identified subpopulations to gather more features to reduce noise.
applications, both the data and model should be considered.
will check their algorithms for bias, variance, and noise-- which will guide further efforts to reduce unfairness.
applications, both the data and model should be considered.
will check their algorithms for bias, variance, and noise-- which will guide further efforts to reduce unfairness.
applications, both the data and model should be considered.
will check their algorithms for bias, variance, and noise-- which will guide further efforts to reduce unfairness.