Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings
@inproceedings{Bolukbasi2016ManIT, title={Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings}, author={Tolga Bolukbasi and Kai-Wei Chang and James Y. Zou and Venkatesh Saligrama and Adam Tauman Kalai}, booktitle={NIPS}, year={2016} }
The blind application of machine learning runs the risk of amplifying biases present in data. [] Key Method Geometrically, gender bias is first shown to be captured by a direction in the word embedding. Second, gender neutral words are shown to be linearly separable from gender definition words in the word embedding.
Figures and Tables from this paper
1,911 Citations
Replicating ’Man is to Computer Programmer as Woman is to Homemakers? Debiasing Word Embeddings’
- Computer Science
- 2022
This work replicates, testing, and visualizes a methodology to remove gender stereotypes from word embeddings to affirm the results reported by the original researchers and test their model’s robustness.
Fair Is Better than Sensational: Man Is to Doctor as Woman Is to Doctor
- Computer ScienceCL
- 2020
A series of clarifications are provided that should put well-known, and potentially new analogies into the right perspective, which might have yielded a distorted picture of bias in word embeddings.
Adversarial Approaches to Debiasing Word Embeddings
- Computer Science
- 2021
This work proposes that bias can be mitigated through the use of Generative Adversarial Networks and experiments with two different problem formulations, showing preliminary results on the WEAT scoring system.
Understanding the Origins of Bias in Word Embeddings
- Computer ScienceICML
- 2019
Given a word embedding trained on a corpus, this work develops a technique for understanding the origins of bias in word embeddings that identifies how perturbing the corpus will affect the bias of the resulting embedding.
What are the Biases in My Word Embedding?
- Computer ScienceAIES
- 2019
An algorithm for enumerating biases in word embeddings that outputs a number of Word Embedding Association Tests (WEATs) that capture various biases present in the data, which makes it easier to identify biases against intersectional groups, which depend on combinations of sensitive features.
Bias in word embeddings
- Computer ScienceFAT*
- 2020
A new technique for bias detection for gendered languages is developed and used to compare bias in embeddings trained on Wikipedia and on political social media data, and it is proved that existing biases are transferred to further machine learning models.
Nurse is Closer to Woman than Surgeon? Mitigating Gender-Biased Proximities in Word Embeddings
- Computer ScienceTransactions of the Association for Computational Linguistics
- 2020
RAN-Debias is proposed, a novel gender debiasing methodology that not only eliminates the bias present in a word vector but also alters the spatial distribution of its neighboring vectors, achieving a bias-free setting while maintaining minimal semantic offset.
Impact of Gender Debiased Word Embeddings in Language Modeling
- Computer ScienceArXiv
- 2021
This paper studies how an state-of-the-art recurrent neural language model behaves when trained on data, which under-represents females, using pre-trained standard and debiased word embeddings, to address gender, race and social biases.
A Source-Criticism Debiasing Method for GloVe Embeddings
- Computer ScienceArXiv
- 2021
A simple yet effective method for debiasing GloVe word embeddings which works by incorporating explicit information about training set bias rather than removing biased data outright, and reduces the effect size on Word Embedding Association Test (WEAT) sets without sacrificing training data or TOP-1 performance.
Word embeddings are biased. But whose bias are they reflecting?
- PsychologyAI & SOCIETY
- 2022
From Curriculum Vitae parsing to web search and recommendation systems, Word2Vec and other word embedding techniques have an increasing presence in everyday interactions in human society. Biases,…
References
SHOWING 1-10 OF 47 REFERENCES
Linguistic Regularities in Continuous Space Word Representations
- Computer ScienceNAACL
- 2013
The vector-space word representations that are implicitly learned by the input-layer weights are found to be surprisingly good at capturing syntactic and semantic regularities in language, and that each relationship is characterized by a relation-specific vector offset.
Linguistic Regularities in Sparse and Explicit Word Representations
- Computer ScienceCoNLL
- 2014
It is demonstrated that analogy recovery is not restricted to neural word embeddings, and that a similar amount of relational similarities can be recovered from traditional distributional word representations.
Deep Recursive Neural Networks for Compositionality in Language
- Computer ScienceNIPS
- 2014
The results show that deep RNNs outperform associated shallow counterparts that employ the same number of parameters and outperforms previous baselines on the sentiment analysis task, including a multiplicative RNN variant as well as the recently introduced paragraph vectors.
Improving Document Ranking with Dual Word Embeddings
- Computer ScienceWWW
- 2016
This paper investigates the popular neural word embedding method Word2vec as a source of evidence in document ranking and proposes the proposed Dual Embedding Space Model (DESM), which provides evidence that a document is about a query term.
Automated Experiments on Ad Privacy Settings
- Computer ScienceProc. Priv. Enhancing Technol.
- 2015
AdFisher, an automated tool that explores how user behaviors, Google's ads, and Ad Settings interact, finds that the Ad Settings was opaque about some features of a user’s profile, that it does provide some choice on advertisements, and that these choices can lead to seemingly discriminatory ads.
Certifying and Removing Disparate Impact
- Computer ScienceKDD
- 2015
This work links disparate impact to a measure of classification accuracy that while known, has received relatively little attention and proposes a test for disparate impact based on how well the protected class can be predicted from the other attributes.
Semi-supervised Question Retrieval with Gated Convolutions
- Computer ScienceNAACL
- 2016
This paper designs a recurrent and convolutional model (gated convolution) to effectively map questions to their semantic representations and demonstrates that the model yields substantial gains over a standard IR baseline and various neural network architectures (including CNNs, LSTMs and GRUs).
It's a Man's Wikipedia? Assessing Gender Inequality in an Online Encyclopedia
- SociologyICWSM
- 2015
This paper presents and applies a computational method for assessing gender bias on Wikipedia along multiple dimensions and finds that while women on Wikipedia are covered and featured well in many Wikipedia language editions, the way women are portrayed starkly differs from the way men are portrayed.
Efficient Estimation of Word Representations in Vector Space
- Computer ScienceICLR
- 2013
Two novel model architectures for computing continuous vector representations of words from very large data sets are proposed and it is shown that these vectors provide state-of-the-art performance on the authors' test set for measuring syntactic and semantic word similarities.
Learning with Annotation Noise
- Computer ScienceACL
- 2009
It is shown that there is a hazard that some of the training instances are actually hard cases with unreliable annotations, which is relatively unproblematic for an algorithm operating under the 0--1 loss model, whereas for the commonly used voted perceptron algorithm, hard training cases could result in incorrect prediction on the uncontroversial cases at test time.