Fairness and bias in Machine Learning A quick review on tools to - - PowerPoint PPT Presentation

fairness and bias in machine learning
SMART_READER_LITE
LIVE PREVIEW

Fairness and bias in Machine Learning A quick review on tools to - - PowerPoint PPT Presentation

QCon 2019 Fairness and bias in Machine Learning A quick review on tools to detect biases in machine learning model Thierry Silbermann, Tech Lead Data Science at Nubank thierry.silbermann@nubank.com.br Data collection Todays


slide-1
SLIDE 1

Fairness and bias in Machine Learning

Thierry Silbermann, Tech Lead Data Science at Nubank

QCon 2019

A quick review on tools to detect biases in machine learning model

  • thierry.silbermann@nubank.com.br
slide-2
SLIDE 2

Data collection

  • Today’s applications collect and mine vast quantities of

personal information.

  • The collection and use of such data raise two important

challenges.

  • First, massive data collection is perceived by many as a

major threat to traditional notions of individual privacy.

  • Second, the use of personal data for algorithmic decision-

making can have unintended and harmful consequences, such as unfair or discriminatory treatment of users.

slide-3
SLIDE 3

Data collection

  • Today’s applications collect and mine vast quantities of

personal information.

  • The collection and use of such data raise two important

challenges.

  • First, massive data collection is perceived by many as a

major threat to traditional notions of individual privacy.

  • Second, the use of personal data for algorithmic decision-

making can have unintended and harmful consequences, such as unfair or discriminatory treatment of users.

slide-4
SLIDE 4
slide-5
SLIDE 5
slide-6
SLIDE 6
slide-7
SLIDE 7
  • Fairness is increasingly important concern as machine

learning models are used to support decision making in high-stakes applications such as:

  • Mortgage lending
  • Hiring
  • Prison sentencing
  • (Approve customers, increase credit line)

Fairness

slide-8
SLIDE 8

Definitions of fairness

http://fairware.cs.umass.edu/papers/Verma.pdf

slide-9
SLIDE 9

Definitions of fairness

  • It is impossible to satisfy all definitions of fairness at the

same time [Kleinberg et al., 2017]

  • Although fairness research is a very active field, clarity on

which bias metrics and bias mitigation strategies are best is yet to be achieved [Friedler et al., 2018]

  • In addition to the multitude of fairness definitions,

different bias handling algorithms address different parts

  • f the model life-cycle, and understanding each research

contribution, how, when and why to use it is challenging even for experts in algorithmic fairness.

Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk

slide-10
SLIDE 10

Example: Prison sentencing

Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk

True Negative False Positive False Negative True Positive

Did not recidivate Recidivate Label low-risk Label high-risk

slide-11
SLIDE 11

Example: Prison sentencing

Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk

True Negative False Positive False Negative True Positive

Did not recidivate Recidivate Label low-risk Label high-risk Decision maker: Of those I’ve labeled high-risk, how many will recidivate ? Predictive value

slide-12
SLIDE 12

Example: Prison sentencing

Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk

True Negative False Positive False Negative True Positive

Did not recidivate Recidivate Label low-risk Label high-risk Decision maker: Of those I’ve labeled high-risk, how many will recidivate ? Predictive value Defendant: What’s the probability I’ll be incorrectly classifying high-risk ? False positive rate

slide-13
SLIDE 13

Example: Prison sentencing

Tutorial: 21 fairness definitions and their politics: https://www.youtube.com/watch?v=jIXIuYdnyyk

True Negative False Positive False Negative True Positive

Did not recidivate Recidivate Label low-risk Label high-risk Decision maker: Of those I’ve labeled high-risk, how many will recidivate ? Predictive value Defendant: What’s the probability I’ll be incorrectly classifying high-risk ? False positive rate Society [think hiring rather than criminal justice]: Is the selected set demographically balanced ? Demography

slide-14
SLIDE 14

https://en.wikipedia.org/wiki/Confusion_matrix

18 scores/metrics

slide-15
SLIDE 15

Terminology

  • Favorable label: a label whose value corresponds to an outcome that provides an advantage

to the recipient.

  • receiving a loan, being hired for a job, and not being arrested
  • Protected attribute: attribute that partitions a population into groups that have parity in

terms of benefit received

  • race, gender, religion
  • Protected attributes are not universal, but are application specific
  • Privileged value of a protected attribute: group that has historically been at a systematic

advantage

  • Group fairness: the goal of groups defined by protected attributes receiving similar

treatments or outcomes

  • Individual fairness: the goal of similar individuals receiving similar treatments or outcomes
slide-16
SLIDE 16

Terminology

  • Bias: systematic error
  • In the context of fairness, we are concerned with

unwanted bias that places privileged groups at a systematic advantage and unprivileged groups at a systematic disadvantage.

  • Fairness metric: a quantification of unwanted bias in

training data or models.

  • Bias mitigation algorithm: a procedure for reducing

unwanted bias in training data or models.

slide-17
SLIDE 17

But wait !

  • I’m not using any feature that is discriminatory for my

application !

  • I’ve never used gender or even race !
slide-18
SLIDE 18

But wait !

https://demographics.virginia.edu/DotMap/index.html

slide-19
SLIDE 19

But wait !

https://demographics.virginia.edu/DotMap/index.html

Chicago Area, IL, USA

slide-20
SLIDE 20

Fairness metric

  • Confusion matrix
  • TP

, FP , TN, FN, TPR, FPR, TNR, FNR

  • Prevalence, accuracy, PPV, FDR, FOR, NPV
  • LR+, LR-, DOR, F1
slide-21
SLIDE 21

Fairness metric

  • Difference of Means
  • Disparate Impact
  • Statistical Parity
  • Odd ratios
  • Consistency
  • Generalized Entropy Index
slide-22
SLIDE 22

Statistical parity difference

  • Group fairness == statistical parity difference == equal acceptance rate

== benchmarking

  • A classifier satisfies this definition if subjects in both protected and

unprotected groups have equal probability of being assigned to the positive predicted class.

  • Example, this would imply equal probability for male and female

applicants to have good predicted credit score:

  • P(d = 1 | G = male) = P (d = 1 | G = female)
  • The main idea behind this definition is that applicants should have an

equivalent opportunity to obtain a good credit score, regardless of their gender.

slide-23
SLIDE 23

Disparate impact

X=0 X=1 Predicted condition FALSE A B TRUE C D

The 80% test was originally framed by a panel of 32 professionals assembled by the State of California Fair Employment Practice Commission (FEPC) in 1971

slide-24
SLIDE 24

Disparate impact

The 80% rule can then be quantified as:

X=0 X=1 Predicted condition FALSE A B TRUE C D

slide-25
SLIDE 25

Aequitas approach

https://dsapp.uchicago.edu/projects/aequitas/

slide-26
SLIDE 26

How about some solutions?

slide-27
SLIDE 27

Disparate impact remover Relabelling Learning Fair representation

slide-28
SLIDE 28

Disparate impact remover Prejudice remover regulariser Optimised Preprocessing Relabelling Reject Option Classification Learning Fair representation Adversarial Debiasing

slide-29
SLIDE 29

Disparate impact remover Prejudice remover regulariser Additive counterfactually fair estimator Optimised Preprocessing Equalised Odds Post-processing Relabelling Reweighing Reject Option Classification Calibrated Equalised Odds Post-processing Learning Fair representation Adversarial Debiasing Meta-Algorithm for Fair Classification

slide-30
SLIDE 30

Tools

slide-31
SLIDE 31
  • There are three main paths to the goal of making fair

predictions:

  • fair pre-processing,
  • fair in-processing, and
  • fair post-processing

How about fixing predictions?

slide-32
SLIDE 32

AIF360, https://arxiv.org/abs/1810.01943

slide-33
SLIDE 33

Pre-Processing

  • Reweighing generates weights for the training examples in each

(group, label) combination differently to ensure fairness before classification.

  • Optimized preprocessing (Calmon et al., 2017) learns a probabilistic

transformation that edits the features and labels in the data with group fairness, individual distortion, and data fidelity constraints and

  • bjectives.
  • Learning fair representations (Zemel et al., 2013) finds a latent

representation that encodes the data well but obfuscates information about protected attributes.

  • Disparate impact remover (Feldman et al., 2015) edits feature values to

increase group fairness while preserving rank-ordering within groups.

slide-34
SLIDE 34

In-Processing

  • Adversarial debiasing (Zhang et al., 2018) learns a

classifier to maximize prediction accuracy and simultaneously reduce an adversaries ability to determine the protected attribute from the predictions. This approach leads to a fair classifier as the predictions cannot carry any group discrimination information that the adversary can exploit.

  • Prejudice remover (Kamishima et al., 2012) adds a

discrimination-aware regularization term to the learning

  • bjective
slide-35
SLIDE 35

Post-Processing

  • Equalized odds postprocessing (Hardt et al., 2016) solves a

linear program to find probabilities with which to change

  • utput labels to optimize equalized odds.
  • Calibrated equalized odds post-processing (Pleiss et al.,

2017) optimizes over calibrated classifier score outputs to find probabilities with which to change output labels with an equalized odds objective.

  • Reject option classification (Kamiran et al., 2012) gives

favorable outcomes to unprivileged groups and unfavorable

  • utcomes to privileged groups in a confidence band around

the decision boundary with the highest uncertainty.

slide-36
SLIDE 36

Experiments

Datasets Adult Census Income, German Credit, COMPAS Metrics Disparate impact Statistical parity difference Average odds difference Equal opportunity difference Classifiers Logistic Regression (LR), Random Forest Classifier (RF), Neural Network (NN) Pre-processing Algorithms Re-weighing (Kamiran & Calders, 2012) Optimized pre-processing (Calmon et al., 2017) Learning fair representations (Zemel et al., 2013) Disparate impact remover (Feldman et al., 2015) In-processing Algorithms Adversarial debasing (Zhang et al., 2018) Prejudice remover (Kamishima et al., 2012) Post-processing Algorithms Equalized odds post-processing (Hardt et al., 2016) Calibrated eq. odds post-processing (Pleiss et al., 2017) Reject option classification (Kamiran et al., 2012)

AIF360, https://arxiv.org/abs/1810.01943

slide-37
SLIDE 37

Results - Statistical Parity Difference (SPD)

SPD Fair Value is 0

AIF360, https://arxiv.org/abs/1810.01943

slide-38
SLIDE 38

Results - Disparate Impact (DI)

DI Fair Value is 1

AIF360, https://arxiv.org/abs/1810.01943

slide-39
SLIDE 39

Results

AIF360, https://arxiv.org/abs/1810.01943

Adult census dataset Protected attribute: race

slide-40
SLIDE 40

Results

AIF360, https://arxiv.org/abs/1810.01943

slide-41
SLIDE 41

Thank you

slide-42
SLIDE 42

References

  • Conference
  • ACM Conference on Fairness, Accountability, and

Transparency (ACM FAT*) https://fatconference.org/

  • IJCAI 2017 Workshop on Explainable Artificial

Intelligence (XAI) http://home.earthlink.net/~dwaha/ research/meetings/ijcai17-xai/

  • Interpretable ML Symposium - NIPS 2017 http://

interpretable.ml/

slide-43
SLIDE 43

References

  • Books
  • https://fairmlbook.org/
  • Course materials
  • Berkeley CS 294: Fairness in machine learning
  • Cornell INFO 4270: Ethics and policy in data science
  • Princeton COS 597E: Fairness in machine learning
slide-44
SLIDE 44

References

  • Papers
  • Fairness Definitions Explained: http://fairware.cs.umass.edu/

papers/Verma.pdf

  • AIF360: An Extensible Toolkit for Detecting, Understanding,

and Mitigating Unwanted Algorithmic Bias https://arxiv.org/ pdf/1810.01943.pdf

  • Aequitas: A Bias and Fairness Audit Toolkit: https://arxiv.org/

pdf/1811.05577.pdf

  • FairTest: Discovering Unwarranted Associations in Data-

Driven Applications: https://arxiv.org/pdf/1510.02377.pdf

slide-45
SLIDE 45

References

  • Videos
  • Tutorial: 21 fairness definitions and their politics

https://www.youtube.com/watch?v=jIXIuYdnyyk

  • AI Fairness 360 Tutorial at ACM FAT* 2019 https://

www.youtube.com/watch?v=XCFDckvyC0M