Semantics derived automatically from language corpora contain human-like biases

  title={Semantics derived automatically from language corpora contain human-like biases},
  author={Aylin Caliskan and Joanna J. Bryson and Arvind Narayanan},
  pages={183 - 186}
Machines learn what people know implicitly AlphaGo has demonstrated that a machine can learn how to do things that people spend many years of concentrated study learning, and it can rapidly learn how to do them better than any human can. Caliskan et al. now show that machines can learn word associations from written texts and that these associations mirror those learned by humans, as measured by the Implicit Association Test (IAT) (see the Perspective by Greenwald). Why does this matter… 

Context Matters: Recovering Human Semantic Structure from Machine Learning Analysis of Large‐Scale Text Corpora

This work trained state‐of‐the‐art machine learning algorithms using contextually‐constrained text corpora (domain‐specific subsets of Wikipedia articles, 50+ million words each) and showed that this procedure greatly improved predictions of empirical similarity judgments and feature ratings of contextually relevant concepts.

Large pre-trained language models contain human-like biases of what is right and wrong to do

The capabilities of the “moral direction” for guiding (even other) LMs towards producing normative text and showcase it on RealToxicityPrompts testbed, preventing the neural toxic degeneration in GPT-2.

Corpus-based Comparison of Distributional Models of Language and Knowledge Graphs

A corpus-based comparative model is introduced that allows us to compare representations of different sources generated under the distributional semantic theory, and it is shown that it can also deal with entity names and not just words, overcoming some problems that follow from the ambiguity of natural language.

Beyond Big Data: What Can We Learn from AI Models?: Invited Keynote

It is shown that humans have unique linguistic fingerprints, which can be de-anonymize humans that have written certain text, source code, or even executable binaries of compiled code, which is a serious privacy threat for individuals that would like to remain anonymous.

Semantic projection: recovering human knowledge of multiple, distinct object features from word embeddings

A powerful, domain-general solution: "semantic projection" of word-vectors onto lines that represent various object features, like size, intelligence, and danger, which recovers human judgments across a range of object categories and properties.

Word Embeddings Reveal How Fundamental Sentiments Structure Natural Language

Central to affect control theory are culturally shared meanings of concepts. That these sentiments overlap among members of a culture presumably reflects their roots in the language use that members

What are the Biases in My Word Embedding?

An algorithm for enumerating biases in word embeddings that outputs a number of Word Embedding Association Tests (WEATs) that capture various biases present in the data, which makes it easier to identify biases against intersectional groups, which depend on combinations of sensitive features.

The Generative Nature of Commonsense Knowledge: Insights from Machine Learning

The main finding of this paper is that the knowledge base that directly facilitates both human agreement and the model’s measure of fit is by its very nature generative, and only truly exists in representation as it is applied.

Semantic projection recovers rich human knowledge of multiple object features from word embeddings.

A domain-general method is used to extract context-dependent relationships from word embeddings: 'semantic projection' of word-vectors onto lines that represent features such as size or danger ('safe' to 'dangerous'), analogous to 'mental scales'.

Predicting High-Level Human Judgment Across Diverse Behavioral Domains

It is shown that word embeddings can be used to predict complex theoretically- and practically- relevant human perceptions and evaluations in domains as diverse as social cognition, health behavior, risk perception, organizational behavior, and marketing.



Extracting Semantics from the Enron Corpus

Indirect measures must be used when analysing attitudes, as individuals are unlikely to voluntarily express beliefs that are opposed by social norms. The IAT indirectly assesses attitudes through the

The Distributional Hypothesis

There is a correlation between distributional similarity and meaning similarity, which allows us to utilize the former in order to estimate the latter, and one can pose two very basic questions concerning the distributional hypothesis: what kind of distributional properties the authors should look for, and what — if any — the differences are between different kinds of Distributional properties.

Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings

This work empirically demonstrates that its algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks.

Extracting semantic representations from word co-occurrence statistics: A computational study

This article presents a systematic exploration of the principal computational possibilities for formulating and validating representations of word meanings from word co-occurrence statistics and finds that, once the best procedures are identified, a very simple approach is surprisingly successful and robust over a range of psychologically relevant evaluation measures.

A Neural Probabilistic Language Model

This work proposes to fight the curse of dimensionality by learning a distributed representation for words which allows each training sentence to inform the model about an exponential number of semantically neighboring sentences.

From Frequency to Meaning: Vector Space Models of Semantics

The goal in this survey is to show the breadth of applications of VSMs for semantics, to provide a new perspective on VSMs, and to provide pointers into the literature for those who are less familiar with the field.

Word and Object

This edition offers a new preface by Quine's student and colleague Dagfinn Follesdal that describes the never-realized plans for a second edition of Word and Object, in which Quine would offer a more unified treatment of the public nature of meaning, modalities, and propositional attitudes.

The Direct Route: Mediated Priming in Semantic Space

McKoon and Ratcliff (1992) presented a theory of mediated priming where the priming effect is due to a direct but weak relatedness between prime and target. They also introduced a quantitative

Accessing Different Types of Lexical Semantic Information: Evidence From Priming

The types of semantic information that are automatically retrieved from the mental lexicon on hearing a word were investigated in 3 semantic priming experiments. The authors probed for activation of

Temporal associations and prior-list intrusions in free recall.

Although repetition enhanced recall of list items, subjects were significantly more likely to make PLIs following the recall of repeated items, suggesting that temporal associations formed in earlier lists can induce recall errors.