Building Fair and Robust Representations for Vision and Language - - PowerPoint PPT Presentation

building fair and robust representations for vision and
SMART_READER_LITE
LIVE PREVIEW

Building Fair and Robust Representations for Vision and Language - - PowerPoint PPT Presentation

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong Building Fair and Robust Representations for Vision and Language Vicente Ordez-Romn Assistant Professor Department of Computer Science Outline Issues


slide-1
SLIDE 1

Building Fair and Robust Representations for Vision and Language

Vicente Ordóñez-Román

Assistant Professor Department of Computer Science

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

slide-2
SLIDE 2

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Outline

  • Issues identified in biased representations
  • Metrics and findings
  • Solutions that have been proposed

2

slide-3
SLIDE 3

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Annotated Data + Machine Learning / Deep Learning

f(x)

Words, Text, Linguistic Structure

slide-4
SLIDE 4

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Case Study 1: Most Basic form of Grounding: Image to Words

f(x)

kitchen no-kitchen

Protected variable: Gender

slide-5
SLIDE 5

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Case Study 1: Most Basic form of Grounding: Image to Words

f(x)

kitchen no-kitchen

Protected variable: Gender For any pair of gender types: P(kitchen = 1 / gender = m) = P(kitchen = 1 / gender = f) P(kitchen = 0 / gender = m) = P(kitchen = 0 / gender = f)

slide-6
SLIDE 6

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

ICML 2013

slide-7
SLIDE 7

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

X: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

slide-8
SLIDE 8

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

y = f(x)

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

X: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

slide-9
SLIDE 9

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

Instead

X: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

Z: Representations

slide-10
SLIDE 10

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

X: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

y = f(z)

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

Z: Representations

Instead

̂ x = ∑

i

zivi

x y

slide-11
SLIDE 11

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

X+: Images

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

Instead

X-: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

y = f(z)

Z: Representations

̂ x = ∑

i

zivi

y

slide-12
SLIDE 12

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

X+: Images

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

Instead

X-: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

y = f(z)

Z: Representations

̂ x = ∑

i

zivi

y P(zi|x+) = P(zi|x−)

slide-13
SLIDE 13

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1: Feature Invariant Learning

Learning Fair Representations Zemel, Wu, Swersky, Pitassi, and Dwork. ICML 2013

L = ∑

k

CrossEntropy(y(k), ̂ y(k)) + α∑

k

x(k) − ̂ x(k) + β 1 |X + | ∑

X+

z(k)

i

− 1 |X − | ∑

X−

z(k)

i

Classifications should be good Reconstructions should be good Intermediate Representations should be indistinguishable across values of the protected variable

slide-14
SLIDE 14

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

y = f(x)

X: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

slide-15
SLIDE 15

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

y = f(x)

X: Images Y: Labels

kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen no-kitchen kitchen

z

slide-16
SLIDE 16

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

y = f(x)

X: Images

z

kitchen / no-kitchen

  • bjective

gender prediction adversarial objective

slide-17
SLIDE 17

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

y = f(x)

X: Images

z

Person identification

  • bjective

illumination type

slide-18
SLIDE 18

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

slide-19
SLIDE 19

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Case Study: Visual Semantic Role Labeling (vSRL)

Commonly Uncommon: Semantic Sparsity in Situation Recognition Mark Yatskar, Vicente Ordonez, Luke Zettlemoyer, Ali Farhadi CVPR 2017

slide-20
SLIDE 20

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Lots of Images of People Carrying Backpacks Not Many Images of People Carrying Tables But Lots of Images of Tables in Other Images

Compositionality: How to learn what looks like carrying?

slide-21
SLIDE 21

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Deep Neural Network + Compositional Conditional Random Field (CRF)

Commonly Uncommon: Semantic Sparsity in Situation Recognition Mark Yatskar, Vicente Ordonez, Luke Zettlemoyer, Ali Farhadi CVPR 2017

slide-22
SLIDE 22

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Situation Recognition: CVPR 2017 Compositional Shared Learning of Underlying Concepts

Commonly Uncommon: Semantic Sparsity in Situation Recognition Mark Yatskar, Vicente Ordonez, Luke Zettlemoyer, Ali Farhadi CVPR 2017

http://imsitu.org/demo/

slide-23
SLIDE 23

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

However we kept running into this…

Commonly Uncommon: Semantic Sparsity in Situation Recognition Mark Yatskar, Vicente Ordonez, Luke Zettlemoyer, Ali Farhadi CVPR 2017

http://imsitu.org/demo/

slide-24
SLIDE 24

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

However we kept running into this…

Commonly Uncommon: Semantic Sparsity in Situation Recognition Mark Yatskar, Vicente Ordonez, Luke Zettlemoyer, Ali Farhadi CVPR 2017

http://imsitu.org/demo/

slide-25
SLIDE 25

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Model? Dataset?

Key Finding: Models Amplify Biases in the Dataset

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

slide-26
SLIDE 26

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Model? Dataset?

Images of People Cooking

Key Finding: Models Amplify Biases in the Dataset

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

slide-27
SLIDE 27

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Model? Dataset?

Men Cooking: 33% Women Cooking: 66%

Key Finding: Models Amplify Biases in the Dataset

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

slide-28
SLIDE 28

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Model? Dataset?

Men Cooking: 33% Women Cooking: 66% Test Images

Key Finding: Models Amplify Biases in the Dataset

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

slide-29
SLIDE 29

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Model? Dataset?

Men Cooking: 33% Women Cooking: 66% Men Cooking: 16% Women Cooking: 84%

Key Finding: Models Amplify Biases in the Dataset

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

slide-30
SLIDE 30

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Model Dataset

Men Cooking: 33% Women Cooking: 66% Men Cooking: 16% Women Cooking: 84%

Model*

*Our solution: RBA: Optimize for accuracy but also to match data distribution.

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

slide-31
SLIDE 31

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Men Also Like Shopping: Reducing Gender Bias Amplification Using Corpus Level Constraints Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, Kai-Wei Chang. EMNLP 2017

Model Dataset

Men Cooking: 33% Women Cooking: 66% Men Cooking: 16% Women Cooking: 84%

Model*

Men Cooking: 20% Women Cooking: 80% *Our solution: RBA: Optimize for accuracy but also to match data distribution.

slide-32
SLIDE 32

Reducing Bias Amplification (RBA)

Integer Linear Program <= margin

Training Ratio - Predicted Ratio

∀ points

f(y1 … yn)

Lagrangian Relaxation

Sontag et al., 2011; Rush and Collins, 2012; Chang and Collins, 2011; Peng et al., 2015, Chang et al., 2013; Dalvi, 2015

s(yi , image) max yi

X

i

constraints inference

slide-33
SLIDE 33

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Our most recent work on this topic:

Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias in Deep Image

  • Representations. Tianlu Wang, Jieyu Zhao, Mark

Yatskar, Kai-Wei Chang, Vicente Ordonez. ICCV 2019

  • Biases are present even in more generic and widespread Image Classifiers
  • Biases are present even when gender is not one of the target variables
  • Biases are present even when a best effort is placed on

balancing the dataset for gender

Key Findings

slide-34
SLIDE 34

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Elazar and Goldberg (2018) introduced a notion of leakage from feature representations

Adversarial Removal of Demographic Attributes from Text Data Yanai Elazar, Yoav Goldberg. EMNLP 2018

y = f(x)

X: Text

Tweet Sentiment Objective Can I predict gender or age from these features?

slide-35
SLIDE 35

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Task: Multi-label Prediction

Knife Carrot Table Kitchen Utensils Annotations Man/Woman Classifier

slide-36
SLIDE 36

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Definition: Dataset Leakage

Knife Carrot Table Kitchen Utensils Annotations Man/Woman Classifier

Gender Leakage from the Dataset/Task

slide-37
SLIDE 37

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Definition: Dataset Leakage vs Model Leakage

Knife Carrot Table Kitchen Utensils Annotations (acc=100%) Man/Woman Classifier Knife Carrot Table Kitchen Pineapple Predictions (acc = 58%) Man/Woman Classifier

Gender Leakage from the Model Predictions

Model

Gender Leakage from the Dataset/Task

slide-38
SLIDE 38

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Definition: Dataset Leakage vs Model Leakage

Knife Carrot Table Kitchen Utensils Annotations (acc=100%) Man/Woman Classifier Knife Carrot Table Kitchen Pineapple Predictions (acc = 58%) Man/Woman Classifier

Model Leakage

Model

Dataset Leakage

slide-39
SLIDE 39

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Definition: Dataset Leakage @ 58% vs Model Leakage @ 58%

Knife Carrot Table Kitchen Baseball Annotations (acc=58%) Man/Woman Classifier

Dataset Leakage

Knife Carrot Table Kitchen Pineapple Predictions (acc = 58%) Man/Woman Classifier

Model Leakage

Model

Random Perturbations

slide-40
SLIDE 40

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Definition: Bias Augmentation Definition: Model Leakage @ K - Dataset Leakage @ K

slide-41
SLIDE 41

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Key Finding: Models Leak even when Dataset doesn’t

Women 27% Men 73%

50 56.25 62.5 68.75 75 Dataset Leakage @ 100% Model Leakage Dataset Leakage mAP 100% mAP 58% mAP 58%

Bias Amplification: 10%

Training set size: 28k Task: Classify 80 objects

Gender Leakage Object Classification

slide-42
SLIDE 42

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Key Finding: Models Leak even when Dataset doesn’t

Women 38% Men 63%

50 56.25 62.5 68.75 75 Dataset Leakage @ 100% Model Leakage Dataset Leakage mAP 100% mAP 56% mAP 56%

Bias Amplification: 9%

Training set size: 16k Task: Classify 80 objects

Gender Leakage Object Classification

slide-43
SLIDE 43

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Key Finding: Models Leak even when Dataset doesn’t

Women 50% Men 50%

50 56.25 62.5 68.75 75 Dataset Leakage @ 100% Model Leakage Dataset Leakage mAP 100% mAP 48% mAP 48%

Bias Amplification: 10%

Training set size: 6k Task: Classify 80 objects

Gender Leakage Object Classification

slide-44
SLIDE 44

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Issues Revelaed

  • Models are again shown to not only replicate but

also amplify effects of protected variables.

  • Balancing a dataset is hard - and not effective to

mitigate bias as it is hard to balance against latent variables

44

slide-45
SLIDE 45

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

y = f(x)

X: Images

z

Person identification

  • bjective

illumination type

slide-46
SLIDE 46

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Controllable Invariance through Adversarial Feature Learning Qizhe Xie, Zihang Dai, Yulun Du, Eduard Hovy, Graham Neubig. NeurIPS 2017

y = f(x)

X: Images

z

kitchen / no-kitchen

  • bjective

gender prediction adversarial objective

slide-47
SLIDE 47

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach 1I: Adversarial Feature Learning

Adversarial Removal of Demographic Attributes from Text Data Yanai Elazar, Yoav Goldberg. EMNLP 2018

y = f(x)

X: Text

z

Tweet Sentiment Objective adversarial demographic prediction: age, gender

slide-48
SLIDE 48

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach: Deep Adversarial Feature Learning

y = f(x)

X: Images

z

kitchen / no-kitchen

  • bjective

gender prediction adversarial objective

Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias in Deep Image Representations. Tianlu Wang, Jieyu Zhao, Mark Yatskar, Kai-Wei Chang, Vicente Ordonez. ICCV 2019

slide-49
SLIDE 49

EMNLP 2019 Tutorial on Bias and Fairness in Natural Language Processing, Hong Kong

Approach: Deep Adversarial Feature Learning

y = f(x)

X: Images

z

kitchen / no-kitchen

  • bjective

gender prediction adversarial objective

Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias in Deep Image Representations. Tianlu Wang, Jieyu Zhao, Mark Yatskar, Kai-Wei Chang, Vicente Ordonez. ICCV 2019

slide-50
SLIDE 50

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Adversarial Removal of Sensitive Features

i.e. Predict Objects while trying to

  • bscure gender
slide-51
SLIDE 51

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

i.e. Predict Objects while trying to

  • bscure gender

Adversarial Removal of Sensitive Features

slide-52
SLIDE 52

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

i.e. Predict Objects while trying to

  • bscure gender

Adversarial Removal of Sensitive Features

slide-53
SLIDE 53

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

i.e. Predict Objects while trying to

  • bscure gender

Adversarial Removal of Sensitive Features

slide-54
SLIDE 54

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

i.e. Predict Objects while trying to

  • bscure gender

Adversarial Removal of Sensitive Features

slide-55
SLIDE 55

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Results

slide-56
SLIDE 56

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Case Study: Image Captioning

Deep Convolutional Neural Network

  • Recurrent Neural Text Decoder
slide-57
SLIDE 57

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Case Study: Image Captioning

Women also Snowboard: Overcoming Bias in Captioning Models Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach. ECCV 2018

A woman cooking a meal A man wearing a black hat is snowboarding

slide-58
SLIDE 58

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Approach I: Add a Confusion Loss

Women also Snowboard: Overcoming Bias in Captioning Models Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach. ECCV 2018

Idea: Augment the data by removing people artificially, and keep a set of gendered reference words where a different loss will be applied Words for every pair of genders should be equally probable

slide-59
SLIDE 59

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Approach II: Add a Confidence Loss

Women also Snowboard: Overcoming Bias in Captioning Models Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach. ECCV 2018

Idea: Discourage the following from happening at the same time: P(word = man) = 0.95 and P(word = woman) = 0.92 Take into account mutual exclusion among groups of words

slide-60
SLIDE 60

ICCV 2019 Linguistics Meets Image and Video Retrieval Workshop, Seoul, South Korea

Students and Collaborators

Tianlu Wang Jieyu Zhao Mark Yatskar Kai-Wei Chang Paola Cascante Ziyan Yang Fuwen Tan Song Feng Baishakhi Ray Hui Wu Xiaoxiao Guo