Man is to Computer Programmer as Woman is to Homemaker? Debiasing - - PowerPoint PPT Presentation

man is to computer programmer as woman is to homemaker
SMART_READER_LITE
LIVE PREVIEW

Man is to Computer Programmer as Woman is to Homemaker? Debiasing - - PowerPoint PPT Presentation

Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings A paper presentation by Giacomo Alliata Computational Social Media 22/05/2020 Presentation of the paper Authors: Tolga Bolukbasi, Kai-Wei Chang, James Zou,


slide-1
SLIDE 1

Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings

A paper presentation by Giacomo Alliata

Computational Social Media 22/05/2020

slide-2
SLIDE 2

Presentation of the paper

  • Authors: Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh

Saligrama and Adam Kalai

  • Presented at NIPS 2016
  • Outline of the paper:

○ Introduction and related works ○ Geometry of Gender and Bias in Word Embeddings ○ Debiasing algorithms ○ Determining gender neutral words ○ Debiasing results ○ Discussion

slide-3
SLIDE 3

What are word embeddings?

A word embedding, trained on word co-occurrence in text corpora, represents each word (or common phrase) as a d-dimensional word vector. Main properties:

  • words that have a similar semantic meaning tend to have vectors that

are close together (measured with cosine similarity)

  • vectors differences between word embeddings represent relationships

between words Analogy: “man is to king as woman is to x” should return x = queen

In this paper, the authors primarily use word2vec 300-dimensional embedding trained on a corpus of Google News texts consisting of 3 million English words.

slide-4
SLIDE 4

Sexism in word embeddings

  • “man is to computer programmer as woman is to x” returns x = housekeeper
  • “father is to a doctor as mother is to a x” returns x = nurse
slide-5
SLIDE 5

Quantifying bias

To quantify bias, the authors compute the similarity between the target word and a pair of gender-specific words.

Ex: nurse is both close to woman and man (since all are humans) but the distance (nurse, woman) is smaller than the distance (nurse, man), thus suggesting bias.

Important distinction:

  • gender-specific words: brother, king (associated with a gender by

definition)

  • gender-neutral words: nurse, babysit, shoes, flight attendant

Gender pairs: set of words that differ mainly in gender (“she-he”, “mother-father”)

slide-6
SLIDE 6

Geometry of Gender and Bias in Word Embeddings

Occupational stereotypes: crowdworkers are asked to evaluate whether an occupation is considered female-stereotypic, male-stereotypic or neutral. Results: strong correlation (Spearman coefficient: 0.51) between the projection of the occupation words onto the she-he axis and the stereotypicality estimates

(similar results both on w2cNEWS and GloVe trained on web-crawl corpus)

slide-7
SLIDE 7

Geometry of Gender and Bias in Word Embeddings

Analogies: take a seed pair (a, b) determining a direction and score all pairs

  • f words (x, y) using the following matrix (with threshold δ)

Evaluation with US-based crowd-workers, with 2 questions:

  • whether the pairing makes sense as an analogy
  • whether it reflects a gender stereotype

Results: (rated by 5+ people out of 10)

  • 72/150 analogies are gender-appropriate
  • 29/150 analogies exhibit gender stereotypes
slide-8
SLIDE 8

Identifying the Gender Subspace

Identify 10 gender-pairs directions to more robustly estimate bias

  • different biases associated with different gender pairs
  • polysemy (ex: “to grandfather a regulation”)

PCA on these 10 vector differences identifies the gender subspace in a single direction

10 selected gender-pairs Percentage of variance explained by the PCA (on gender pairs and on 10 random vectors)

slide-9
SLIDE 9

Debiasing Algorithms

First step: Identify the Gender Subspace that captures the bias Second step: Neutralize and Equalize or Soften

  • Neutralize ensures that gender neutral words are 0 in the gender

subspace

  • Equalize equalizes sets of words outside the subspace (thus any neutral

word is equidistant to all words in each equality set)

○ Ex: if {grandmother, grandfather} and {guy, gal} were two equality sets, then after equalization babysit would be equidistant to grandmother and grandfather and also equidistant to gal and guy

  • Soften reduces the difference between these sets while maintaining as

much similarity as possible (with a parameter to control the trade-off)

○ Ex: grandfather would preserve its other meaning in to grandfather a situation

slide-10
SLIDE 10

Debiasing Algorithms

slide-11
SLIDE 11

Determining Gender Neutral Words

In practice, it is easier to identify the gender-specific words and take the complementary set of words as the gender-neutral words. Classification task:

  • produce a training set (using

dictionary definitions) of gender-specific words

  • train the classification model

(gender-specific or gender-neutral) ○ linear SVM in the paper ○ F-score: 0.627 ± 0.102

Results of the classificer

slide-12
SLIDE 12

Debiasing Results

Evaluate using analogy task as before (pair (she, he) as input and pair (x, y) as

  • utput, rated as gender stereotypical or not by crowd-workers). Results are

also confirmed by testing on standard benchmarks the quality of the analogies

slide-13
SLIDE 13

Discussion

Main results:

  • a single direction largely captures gender bias and projecting

gender-neutral words on this direction helps to quantify gender bias

  • the hard-debiasing algorithm significantly reduces the gender bias

while preserving appropriate analogies

  • the soft-debiasing algorithm could be useful in certain

applications Overall, bias in word embeddings simply reflects bias in society but, if we are not able to directly reduce it, we should at least ensure machine learning does not increase it.

slide-14
SLIDE 14

Thanks for your attention!

Giacomo Alliata Computational Social Media 22/05/2020