Fairness and bias in Machine Learning
Thierry Silbermann, Tech Lead Data Science at Nubank
QCon 2019
A quick review on tools to detect biases in machine learning model
- thierry.silbermann@nubank.com.br
Fairness and bias in Machine Learning A quick review on tools to - - PowerPoint PPT Presentation
QCon 2019 Fairness and bias in Machine Learning A quick review on tools to detect biases in machine learning model Thierry Silbermann, Tech Lead Data Science at Nubank thierry.silbermann@nubank.com.br Data collection Todays
Thierry Silbermann, Tech Lead Data Science at Nubank
QCon 2019
A quick review on tools to detect biases in machine learning model
personal information.
challenges.
major threat to traditional notions of individual privacy.
making can have unintended and harmful consequences, such as unfair or discriminatory treatment of users.
personal information.
challenges.
major threat to traditional notions of individual privacy.
making can have unintended and harmful consequences, such as unfair or discriminatory treatment of users.
learning models are used to support decision making in high-stakes applications such as:
http://fairware.cs.umass.edu/papers/Verma.pdf
same time [Kleinberg et al., 2017]
which bias metrics and bias mitigation strategies are best is yet to be achieved [Friedler et al., 2018]
different bias handling algorithms address different parts
contribution, how, when and why to use it is challenging even for experts in algorithmic fairness.
Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
True Negative False Positive False Negative True Positive
Did not recidivate Recidivate Label low-risk Label high-risk
Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
True Negative False Positive False Negative True Positive
Did not recidivate Recidivate Label low-risk Label high-risk Decision maker: Of those I’ve labeled high-risk, how many will recidivate ? Predictive value
Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
True Negative False Positive False Negative True Positive
Did not recidivate Recidivate Label low-risk Label high-risk Decision maker: Of those I’ve labeled high-risk, how many will recidivate ? Predictive value Defendant: What’s the probability I’ll be incorrectly classifying high-risk ? False positive rate
Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk
True Negative False Positive False Negative True Positive
Did not recidivate Recidivate Label low-risk Label high-risk Decision maker: Of those I’ve labeled high-risk, how many will recidivate ? Predictive value Defendant: What’s the probability I’ll be incorrectly classifying high-risk ? False positive rate Society [think hiring rather than criminal justice]: Is the selected set demographically balanced ? Demography
https://en.wikipedia.org/wiki/Confusion_matrix
to the recipient.
terms of benefit received
advantage
treatments or outcomes
unwanted bias that places privileged groups at a systematic advantage and unprivileged groups at a systematic disadvantage.
training data or models.
unwanted bias in training data or models.
application !
https://demographics.virginia.edu/DotMap/index.html
https://demographics.virginia.edu/DotMap/index.html
Chicago Area, IL, USA
, FP , TN, FN, TPR, FPR, TNR, FNR
== benchmarking
unprotected groups have equal probability of being assigned to the positive predicted class.
applicants to have good predicted credit score:
equivalent opportunity to obtain a good credit score, regardless of their gender.
X=0 X=1 Predicted condition FALSE A B TRUE C D
The 80% test was originally framed by a panel of 32 professionals assembled by the State of California Fair Employment Practice Commission (FEPC) in 1971
The 80% rule can then be quantified as:
X=0 X=1 Predicted condition FALSE A B TRUE C D
https://dsapp.uchicago.edu/projects/aequitas/
Disparate impact remover Relabelling Learning Fair representation
Disparate impact remover Prejudice remover regulariser Optimised Preprocessing Relabelling Reject Option Classification Learning Fair representation Adversarial Debiasing
Disparate impact remover Prejudice remover regulariser Additive counterfactually fair estimator Optimised Preprocessing Equalised Odds Post-processing Relabelling Reweighing Reject Option Classification Calibrated Equalised Odds Post-processing Learning Fair representation Adversarial Debiasing Meta-Algorithm for Fair Classification
predictions:
AIF360, https://arxiv.org/abs/1810.01943
(group, label) combination differently to ensure fairness before classification.
transformation that edits the features and labels in the data with group fairness, individual distortion, and data fidelity constraints and
representation that encodes the data well but obfuscates information about protected attributes.
increase group fairness while preserving rank-ordering within groups.
classifier to maximize prediction accuracy and simultaneously reduce an adversaries ability to determine the protected attribute from the predictions. This approach leads to a fair classifier as the predictions cannot carry any group discrimination information that the adversary can exploit.
discrimination-aware regularization term to the learning
linear program to find probabilities with which to change
2017) optimizes over calibrated classifier score outputs to find probabilities with which to change output labels with an equalized odds objective.
favorable outcomes to unprivileged groups and unfavorable
the decision boundary with the highest uncertainty.
Datasets Adult Census Income, German Credit, COMPAS Metrics Disparate impact Statistical parity difference Average odds difference Equal opportunity difference Classifiers Logistic Regression (LR), Random Forest Classifier (RF), Neural Network (NN) Pre-processing Algorithms Re-weighing (Kamiran & Calders, 2012) Optimized pre-processing (Calmon et al., 2017) Learning fair representations (Zemel et al., 2013) Disparate impact remover (Feldman et al., 2015) In-processing Algorithms Adversarial debasing (Zhang et al., 2018) Prejudice remover (Kamishima et al., 2012) Post-processing Algorithms Equalized odds post-processing (Hardt et al., 2016) Calibrated eq. odds post-processing (Pleiss et al., 2017) Reject option classification (Kamiran et al., 2012)
AIF360, https://arxiv.org/abs/1810.01943
SPD Fair Value is 0
AIF360, https://arxiv.org/abs/1810.01943
DI Fair Value is 1
AIF360, https://arxiv.org/abs/1810.01943
AIF360, https://arxiv.org/abs/1810.01943
Adult census dataset Protected attribute: race
AIF360, https://arxiv.org/abs/1810.01943
Transparency (ACM FAT*) https://fatconference.org/
Intelligence (XAI) http://home.earthlink.net/~dwaha/ research/meetings/ijcai17-xai/
interpretable.ml/
papers/Verma.pdf
and Mitigating Unwanted Algorithmic Bias https://arxiv.org/ pdf/1810.01943.pdf
pdf/1811.05577.pdf
Driven Applications: https://arxiv.org/pdf/1510.02377.pdf
https://www.youtube.com/watch?v=jIXIuYdnyyk
www.youtube.com/watch?v=XCFDckvyC0M