Semantics derived automatically from language corpora contain human-like biases

@article{Caliskan2017SemanticsDA,
  title={Semantics derived automatically from language corpora contain human-like biases},
  author={Aylin Caliskan and Joanna J. Bryson and Arvind Narayanan},
  journal={Science},
  year={2017},
  volume={356},
  pages={183 - 186}
}
Machines learn what people know implicitly AlphaGo has demonstrated that a machine can learn how to do things that people spend many years of concentrated study learning, and it can rapidly learn how to do them better than any human can. Caliskan et al. now show that machines can learn word associations from written texts and that these associations mirror those learned by humans, as measured by the Implicit Association Test (IAT) (see the Perspective by Greenwald). Why does this matter… 
Corpus-based Comparison of Distributional Models of Language and Knowledge Graphs
TLDR
A corpus-based comparative model is introduced that allows us to compare representations of different sources generated under the distributional semantic theory, and it is shown that it can also deal with entity names and not just words, overcoming some problems that follow from the ambiguity of natural language.
Beyond Big Data: What Can We Learn from AI Models?: Invited Keynote
TLDR
It is shown that humans have unique linguistic fingerprints, which can be de-anonymize humans that have written certain text, source code, or even executable binaries of compiled code, which is a serious privacy threat for individuals that would like to remain anonymous.
Semantic projection: recovering human knowledge of multiple, distinct object features from word embeddings
TLDR
A powerful, domain-general solution: "semantic projection" of word-vectors onto lines that represent various object features, like size, intelligence, and danger, which recovers human judgments across a range of object categories and properties.
What are the Biases in My Word Embedding?
TLDR
An algorithm for enumerating biases in word embeddings that outputs a number of Word Embedding Association Tests (WEATs) that capture various biases present in the data, which makes it easier to identify biases against intersectional groups, which depend on combinations of sensitive features.
Predicting High-Level Human Judgment Across Diverse Behavioral Domains
Recent advances in machine learning, combined with the increased availability of large natural language datasets, have made it possible to uncover semantic representations that characterize what
Discovering and Interpreting Conceptual Biases in Online Communities
TLDR
A general, data-driven approach to automatically discover and help interpret conceptual biases encoded in word embeddings is proposed and applied to study the conceptual biases present in the language used in online communities.
Semantics Derived Automatically from Language Corpora Contain Human-like Moral Choices
TLDR
It is shown that applying machine learning to human texts can extract deontological ethical reasoning about "right" and "wrong" conduct, and indicates that text corpora contain recoverable and accurate imprints of the authors' social, ethical and even moral choices.
Distributional social semantics: Inferring word meanings from communication patterns
TLDR
Through the analysis of the communication patterns of over 330,000 individuals on the online forum Reddit, the findings of the current article demonstrate that social information about word usage allows for unique aspects of a word's meaning to be acquired, providing a new pathway for distributional model development.
Word Embeddings Inherently Recover the Conceptual Organization of the Human Mind
TLDR
It is found that machine learning can recover the conceptual organization of the human mind when applied to the natural language use of millions of people, using text from billions of webpages.
Evolution of emotion semantics
TLDR
Empirical semantics are suggested to be evolving over time, with prototypical emotion words remaining semantically stable, while other emotion words evolve more freely.
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 76 REFERENCES
Extracting Semantics from the Enron Corpus
Indirect measures must be used when analysing attitudes, as individuals are unlikely to voluntarily express beliefs that are opposed by social norms. The IAT indirectly assesses attitudes through the
The Distributional Hypothesis
TLDR
There is a correlation between distributional similarity and meaning similarity, which allows us to utilize the former in order to estimate the latter, and one can pose two very basic questions concerning the distributional hypothesis: what kind of distributional properties the authors should look for, and what — if any — the differences are between different kinds of Distributional properties.
Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings
TLDR
This work empirically demonstrates that its algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks.
Extracting semantic representations from word co-occurrence statistics: A computational study
TLDR
This article presents a systematic exploration of the principal computational possibilities for formulating and validating representations of word meanings from word co-occurrence statistics and finds that, once the best procedures are identified, a very simple approach is surprisingly successful and robust over a range of psychologically relevant evaluation measures.
A Neural Probabilistic Language Model
TLDR
This work proposes to fight the curse of dimensionality by learning a distributed representation for words which allows each training sentence to inform the model about an exponential number of semantically neighboring sentences.
Distributed Representations of Words and Phrases and their Compositionality
TLDR
This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
GloVe: Global Vectors for Word Representation
TLDR
A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
From Frequency to Meaning: Vector Space Models of Semantics
TLDR
The goal in this survey is to show the breadth of applications of VSMs for semantics, to provide a new perspective on VSMs, and to provide pointers into the literature for those who are less familiar with the field.
The Direct Route: Mediated Priming in Semantic Space
McKoon and Ratcliff (1992) presented a theory of mediated priming where the priming effect is due to a direct but weak relatedness between prime and target. They also introduced a quantitative
Accessing Different Types of Lexical Semantic Information: Evidence From Priming
The types of semantic information that are automatically retrieved from the mental lexicon on hearing a word were investigated in 3 semantic priming experiments. The authors probed for activation of
...
1
2
3
4
5
...