Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings
A paper presentation by Giacomo Alliata
Computational Social Media 22/05/2020
Man is to Computer Programmer as Woman is to Homemaker? Debiasing - - PowerPoint PPT Presentation
Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings A paper presentation by Giacomo Alliata Computational Social Media 22/05/2020 Presentation of the paper Authors: Tolga Bolukbasi, Kai-Wei Chang, James Zou,
A paper presentation by Giacomo Alliata
Computational Social Media 22/05/2020
Saligrama and Adam Kalai
○ Introduction and related works ○ Geometry of Gender and Bias in Word Embeddings ○ Debiasing algorithms ○ Determining gender neutral words ○ Debiasing results ○ Discussion
A word embedding, trained on word co-occurrence in text corpora, represents each word (or common phrase) as a d-dimensional word vector. Main properties:
are close together (measured with cosine similarity)
between words Analogy: “man is to king as woman is to x” should return x = queen
In this paper, the authors primarily use word2vec 300-dimensional embedding trained on a corpus of Google News texts consisting of 3 million English words.
To quantify bias, the authors compute the similarity between the target word and a pair of gender-specific words.
Ex: nurse is both close to woman and man (since all are humans) but the distance (nurse, woman) is smaller than the distance (nurse, man), thus suggesting bias.
Important distinction:
definition)
Gender pairs: set of words that differ mainly in gender (“she-he”, “mother-father”)
Occupational stereotypes: crowdworkers are asked to evaluate whether an occupation is considered female-stereotypic, male-stereotypic or neutral. Results: strong correlation (Spearman coefficient: 0.51) between the projection of the occupation words onto the she-he axis and the stereotypicality estimates
(similar results both on w2cNEWS and GloVe trained on web-crawl corpus)
Analogies: take a seed pair (a, b) determining a direction and score all pairs
Evaluation with US-based crowd-workers, with 2 questions:
Results: (rated by 5+ people out of 10)
Identify 10 gender-pairs directions to more robustly estimate bias
PCA on these 10 vector differences identifies the gender subspace in a single direction
10 selected gender-pairs Percentage of variance explained by the PCA (on gender pairs and on 10 random vectors)
First step: Identify the Gender Subspace that captures the bias Second step: Neutralize and Equalize or Soften
subspace
word is equidistant to all words in each equality set)
○ Ex: if {grandmother, grandfather} and {guy, gal} were two equality sets, then after equalization babysit would be equidistant to grandmother and grandfather and also equidistant to gal and guy
much similarity as possible (with a parameter to control the trade-off)
○ Ex: grandfather would preserve its other meaning in to grandfather a situation
In practice, it is easier to identify the gender-specific words and take the complementary set of words as the gender-neutral words. Classification task:
dictionary definitions) of gender-specific words
(gender-specific or gender-neutral) ○ linear SVM in the paper ○ F-score: 0.627 ± 0.102
Results of the classificer
Evaluate using analogy task as before (pair (she, he) as input and pair (x, y) as
also confirmed by testing on standard benchmarks the quality of the analogies
Main results:
gender-neutral words on this direction helps to quantify gender bias
while preserving appropriate analogies
applications Overall, bias in word embeddings simply reflects bias in society but, if we are not able to directly reduce it, we should at least ensure machine learning does not increase it.
Giacomo Alliata Computational Social Media 22/05/2020