• Corpus ID: 59158788

Attenuating Bias in Word Vectors

@inproceedings{Dev2019AttenuatingBI,
  title={Attenuating Bias in Word Vectors},
  author={Sunipa Dev and J. M. Phillips},
  booktitle={AISTATS},
  year={2019}
}
Word vector representations are well developed tools for various NLP and Machine Learning tasks and are known to retain significant semantic and syntactic structure of languages. [...] Key Method We verify how names are masked carriers of gender bias and then use that as a tool to attenuate bias in embeddings. Further, we extend this property of names to show how names can be used to detect other types of bias in the embeddings such as bias based on race, ethnicity, and age.Expand
Bias in word embeddings
TLDR
A new technique for bias detection for gendered languages is developed and used to compare bias in embeddings trained on Wikipedia and on political social media data, and it is proved that existing biases are transferred to further machine learning models.
OSCaR: Orthogonal Subspace Correction and Rectification of Biases in Word Embeddings
TLDR
OSCaR (Orthogonal Subspace Correction and Rectification), a bias-mitigating method that focuses on disentangling biased associations between concepts instead of removing concepts wholesale, is proposed.
On Measuring and Mitigating Biased Inferences of Word Embeddings
TLDR
A mechanism for measuring stereotypes using the task of natural language inference is designed and a reduction in invalid inferences via bias mitigation strategies on static word embeddings (GloVe), and it is shown that for gender bias, these techniques extend to contextualizedembeddings when applied selectively only to the static components of contextualized embeddeds.
A General Framework for Implicit and Explicit Debiasing of Distributional Word Vector Spaces
TLDR
Experimental findings across three embedding methods suggest that the proposed debiasing models are robust and widely applicable: they often completely remove the bias both implicitly and explicitly without degradation of semantic information encoded in any of the input distributional spaces.
Evaluating Gender Bias in Hindi-English Machine Translation
TLDR
This work attempts to evaluate and quantify the gender bias within a Hindi-English machine translation system by implementing a modified version of the existing TGBI metric based on the grammatical considerations for Hindi.
Socially Aware Bias Measurements for Hindi Language Representations
TLDR
This work investigates the biases present in Hindi language representations such as caste and religion associated biases and demonstrates how biases are unique to specific language representations based on the history and culture of the region they are widely spoken in.
Word Embeddings via Causal Inference: Gender Bias Reducing and Semantic Information Preserving
  • Lei Ding, Dengdeng Yu, +7 authors Bei Jiang
  • Computer Science
    ArXiv
  • 2021
TLDR
This work proposes a novel methodology that leverages a causal inference framework to effectively remove gender bias and achieves state-of-the-art results in genderdebiasing tasks.
A Survey on Bias in Deep NLP
TLDR
Bias is introduced in a formal way and how it has been treated in several networks, in terms of detection and correction, and a strategy to deal with bias in deep NLP is proposed.
Monolingual and Multilingual Reduction of Gender Bias in Contextualized Representations
TLDR
DensRay, an analytical method for obtaining interpretable dense subspaces, performs on-par with prior approaches, but arguments that it is more robust and indications that it preserves language model performance better are provided.
Joint Multiclass Debiasing of Word Embeddings
TLDR
A joint multiclass debiasing approach that is capable ofdebiasing multiple bias dimensions simultaneously and can both reduce or even completely eliminate bias, while maintaining meaningful relationships between vectors in word embeddings is presented.
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 17 REFERENCES
Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings
TLDR
This work empirically demonstrates that its algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks.
Quantifying and Reducing Stereotypes in Word Embeddings
TLDR
A novel gender analogy task is created and combined with crowdsourcing to systematically quantify the gender bias in a given embedding, and an efficient algorithm is developed that reduces gender stereotype using just a handful of training examples while preserving the useful geometric properties of the embedding.
GloVe: Global Vectors for Word Representation
TLDR
A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Linguistic Regularities in Sparse and Explicit Word Representations
TLDR
It is demonstrated that analogy recovery is not restricted to neural word embeddings, and that a similar amount of relational similarities can be recovered from traditional distributional word representations.
Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints
TLDR
This work proposes to inject corpus-level constraints for calibrating existing structured prediction models and design an algorithm based on Lagrangian relaxation for collective inference to reduce the magnitude of bias amplification in multilabel object classification and visual semantic role labeling.
Women also Snowboard: Overcoming Bias in Captioning Models
TLDR
A new Equalizer model is introduced that ensures equal gender probability when gender Evidence is occluded in a scene and confident predictions when gender evidence is present and has lower error than prior work when describing images with people and mentioning their gender and more closely matches the ground truth ratio of sentences including women to sentences including men.
Efficient Estimation of Word Representations in Vector Space
TLDR
Two novel model architectures for computing continuous vector representations of words from very large data sets are proposed and it is shown that these vectors provide state-of-the-art performance on the authors' test set for measuring syntactic and semantic word similarities.
Distributed Representations of Words and Phrases and their Compositionality
TLDR
This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Semantics derived automatically from language corpora contain human-like biases
TLDR
It is shown that machines can learn word associations from written texts and that these associations mirror those learned by humans, as measured by the Implicit Association Test (IAT), and that applying machine learning to ordinary human language results in human-like semantic biases.
Neural Word Embedding as Implicit Matrix Factorization
TLDR
It is shown that using a sparse Shifted Positive PMI word-context matrix to represent words improves results on two word similarity tasks and one of two analogy tasks, and conjecture that this stems from the weighted nature of SGNS's factorization.
...
1
2
...