Leveraging GANs for fairness evaluations Emily Denton Research - - PowerPoint PPT Presentation

leveraging gans for fairness evaluations
SMART_READER_LITE
LIVE PREVIEW

Leveraging GANs for fairness evaluations Emily Denton Research - - PowerPoint PPT Presentation

Leveraging GANs for fairness evaluations Emily Denton Research Scientist, Google Brain Emily Denton Margaret Mitchell Timnit Gebru Ben Hutchinson Background ML Fairness seeks to address algorithmic unfairness , with a focus on machine


slide-1
SLIDE 1

Leveraging GANs for fairness evaluations

Ben Hutchinson Margaret Mitchell Timnit Gebru Emily Denton Emily Denton Research Scientist, Google Brain

slide-2
SLIDE 2

P • 2

Background

ML Fairness seeks to address algorithmic unfairness, with a focus on machine learning systems

Very broad research area! I will be focusing on one specifjc component: detecting undesirable bias in computer vision systems

slide-3
SLIDE 3

P • 3

Bias in Computer Vision

The Coded Gaze: Unmasking Algorithmic Bias Joy Buolamwini

Unrepresentative training data can lead to disparities in accuracy for difgerent demographics

slide-4
SLIDE 4

P • 4

[Wilson et al. Predictive inequity in object detection. arXiv:1902.11097, 2019]

Bias in Computer Vision

slide-5
SLIDE 5

P • 5

[Zhao et al. Men Also Like Shopping: Reducing Gender Bias Amplifjcation using Corpus-level Constraints. EMNLP, 2017.]

Bias in Computer Vision

Social biases embedded in data distribution can be reproduced and/or amplifjed

[Hendricks et al. Women also snowboard: Overcoming bias in captioning models. ECCV, 2018]

slide-6
SLIDE 6

P • 6

[Misra et al. Seeing through the Human Reporuing Bias: Visual Classifjers from Noisy Human-Centric Labels. CVPR 2016]

Bias in Computer Vision

Human reporuing bias can afgect annotations

slide-7
SLIDE 7

P • 7

[Misra et al. Seeing through the Human Reporuing Bias: Visual Classifjers from Noisy Human-Centric Labels. CVPR 2016]

Bias in Computer Vision

Human reporuing bias can afgect annotations

“Green bananas”

slide-8
SLIDE 8

P • 8

Bias in Computer Vision

“doctor” “female doctor”

  • r

“nurse”

[Misra et al. Seeing through the Human Reporuing Bias: Visual Classifjers from Noisy Human-Centric Labels. CVPR 2016]

Social biases can afgect annotations and propagate through ML system

slide-9
SLIDE 9

P • 9

Bias in Computer Vision

Social biases can afgect annotations and propagate through ML system

[Rhue. Racial Infmuence on Automated Perceptions of Emotions. 2019]

slide-10
SLIDE 10

P • 10

How can GANs help?

High quality photo realistic images

[Karras et al. Progressive growing of gans for improved quality, stability, and variation. ICLR, 2018]

slide-11
SLIDE 11

P • 11

How can GANs help?

High quality photo realistic images

[Karras et al. Progressive growing of gans for improved quality, stability, and variation. ICLR, 2018]

Controllable image synthesis

slide-12
SLIDE 12

P • 12

How can GANs help?

Generative techniques provide tools for testing a classifjer’s sensitivity to difgerent image features Can answer questions of the form: How does the classifjer’s output change as some characteristic of the image is systematically varied? Is the classifjer sensitive to a characteristic that should be irrelevant for the task?

slide-13
SLIDE 13

P • 13

GANs can help uncover undesirable bias

CNN

P(Smile | x)

x

slide-14
SLIDE 14

P • 14

GANs can help uncover undesirable bias

CNN

P(Smile | x)

x x’

Manipulate facial hair

slide-15
SLIDE 15

P • 15

GANs can help uncover undesirable bias

CNN

P(Smile | x)

x

CNN

P(Smile | x’)

x’

Did it change?

Manipulate facial hair

slide-16
SLIDE 16

P • 16

P(Smile | x)

Can observe the efgect on a classifjers of systematically manipulating factors of variation in an image

P(Smile | x)

slide-17
SLIDE 17

P • 17

P(Smile | x) P(Smile | x)

All else being equal, the presence of facial hair should be irrelevant to the classifjer

Can observe the efgect on a classifjers of systematically manipulating factors of variation in an image

X

slide-18
SLIDE 18

P • 18

Experimental setup

Smiling classifjer trained on CelebA (128x128 resolution images)

slide-19
SLIDE 19

P • 19

Experimental setup

Smiling classifjer trained on CelebA (128x128 resolution images) Standard progressive GAN trained to generate 128x128 CelebA images

slide-20
SLIDE 20

P • 20

Experimental setup

Smiling classifjer trained on CelebA (128x128 resolution images) Standard progressive GAN trained to generate 128x128 CelebA images Encoder trained to infer latent codes that generated an images

slide-21
SLIDE 21

P • 21

Aturibute vectors

Directions in latent space that manipulate a paruicular factor of variation in the image

da

Latent codes corresponding to images without aturibute a Latent codes corresponding to images with aturibute a Aturibute vector

slide-22
SLIDE 22

P • 22

Aturibute vectors

Eyeglasses = 1 Eyeglasses = 0 Mustache = 1 Mustache = 0 Blond_Hair = 1 Blond_Hair = 0

We infer aturibute vectors using binary CelebA annotations

slide-23
SLIDE 23

P • 23

CelebA aturibute vectors

E( ) E( ) E( ) E( ) E( ) E( ) E( ) E( )

dMustache

Mustache = 1 Mustache = 0

slide-24
SLIDE 24

P • 24

A note on CelebA aturibute vectors

Many of the atuributes are subjective or ill-defjned Interpretation of category boundaries is contingent on the annotators The resulting manipulations refmect how the paruicular atuributes were operationalized and measured within the CelebA dataset

slide-25
SLIDE 25

P • 25

Manipulating images with CelebA aturibute vectors

slide-26
SLIDE 26

P • 26

Manipulating images with CelebA aturibute vectors

slide-27
SLIDE 27

P • 27

Manipulating images with CelebA aturibute vectors

slide-28
SLIDE 28

P • 28

Manipulating images with CelebA aturibute vectors

slide-29
SLIDE 29

P • 29

Quantifying classifjer sensitivity

Model f outputs the probability of a smile being present in the image: Sensitivity of the continuous valued output of f to changes defjned by the aturibute vector d:

Difgerence in classifjers’ output that results from moving in direction d in latent space

slide-30
SLIDE 30

P • 30

Quantifying classifjer sensitivity

Given a threshold, 0 ≤ c ≤ 1, binary classifjcations are

  • btained:

Sensitivity of the discrete classifjcation decision to peruurbations along an vector d as:

Frequency with which classifjcation fmips from smiling to not smiling Frequency with which classifjcation fmips from not smiling to smiling

slide-31
SLIDE 31

P • 31

Quantifying classifjer sensitivity

Given a threshold, 0 ≤ c ≤ 1, binary classifjcations are

  • btained:

Sensitivity of the discrete classifjcation decision to peruurbations along an vector d as:

slide-32
SLIDE 32

P • 32

Quantifying classifjer sensitivity

Given a threshold, 0 ≤ c ≤ 1, binary classifjcations are

  • btained:

Sensitivity of the discrete classifjcation decision to peruurbations along an vector d as:

slide-33
SLIDE 33

P • 33

Quantifying classifjer sensitivity

Given a threshold, 0 ≤ c ≤ 1, binary classifjcations are

  • btained:

Sensitivity of the discrete classifjcation decision to peruurbations along an vector d as:

slide-34
SLIDE 34

P • 34

What have the aturibute vectors encoded?

~12% of images initially classifjed as not smiling get classifjed as smiling afuer Heavy_Makeup augmentation

slide-35
SLIDE 35

P • 35

What have the aturibute vectors encoded?

~12% of images initially classifjed as not smiling get classifjed as smiling afuer Heavy_Makeup augmentation

slide-36
SLIDE 36

P • 36

What have the aturibute vectors encoded?

~7% of images initially classifjed as smiling get classifjed as not smiling afuer Young augmentation

slide-37
SLIDE 37

P • 37

BUT, need to be careful the aturibute vector hasn’t actually encoded something that should be relevant to smiling classifjcation!

Mouth expression has defjnitely changed ~40% of images initially classifjed as not smiling get classifjed as smiling afuer High_Cheekbones augmentation

slide-38
SLIDE 38

P • 38

BUT, need to be careful the aturibute vector hasn’t actually encoded something that should be relevant to smiling classifjcation!

So far we’re verifjed makeup, facial hair and age related aturibute directions leave basic mouth shape/smile unchanged In process of running more of these studies on complete set of atuributes

slide-39
SLIDE 39

P • 39

Social context is imporuant

Generative techniques can be used to detect unintended and undesirable bias in facial analysis Equalizing error statistics across difgerent groups (defjned along cultural, demographic, phenotypical lines) is imporuant but not suffjcient for building fair, equitable, just or inclusive technology This analysis should be paru of a larger, socially contextualized, project to critically assess broader ethical concerns relating to facial analysis technology

slide-40
SLIDE 40

P • 40

Future work

  • GAN can be trained on difgerent dataset than classifjer
  • Increased disentanglement of latent space
  • Extend beyond faces
  • Other ways of leveraging synthetic data for evaluation (or training?) purposes

○ i.e. mine GANs for data, not people

slide-41
SLIDE 41

P • 41

Related work

Countergactual fairness

Kilberuus et al. Avoiding discrimination through causal reasoning. NIPS, 2017. Kusner et al. Countergactual fairness. NIPS, 2017.

Countergactual fairness for text

Garg et al. Countergactual Fairness in Text Classifjcation through Robustness. AIES, 2019

Individual fairness

Dwork et al. Fairness Through Awareness. ITCS, 2012.

Model interpretability

Kim et al. Interpretability beyond feature aturibution: Quantitative testing with concept activation vectors (tcav). ICML, 2018. Chang et al. Explaining image classifjers by countergactual generation. ICLR, 2019. Fong and Vedaldi. Interpretable explanations of black boxes by meaningful peruurbation. ICCV, 2017. Dabkowski and Gal. Real time image saliency for black box classifjers. NIPS, 2017 Simonyan et al. Deep inside convolutional networks: Visualising image classifjcation models and saliency maps. 2013

slide-42
SLIDE 42

P • 42

Thanks!

Denton et al. Detecting Bias with Generative Countergactual Face Aturibute Augmentation. CVPR Workshop on Fairness, Accountability, Transparency and Ethics in Computer Vision, 2019.