Semantics derived automatically from language corpora contain human-like biases
@article{Caliskan2016SemanticsDA, title={Semantics derived automatically from language corpora contain human-like biases}, author={Aylin Caliskan and Joanna J. Bryson and Arvind Narayanan}, journal={Science}, year={2016}, volume={356}, pages={183 - 186} }
Machines learn what people know implicitly AlphaGo has demonstrated that a machine can learn how to do things that people spend many years of concentrated study learning, and it can rapidly learn how to do them better than any human can. Caliskan et al. now show that machines can learn word associations from written texts and that these associations mirror those learned by humans, as measured by the Implicit Association Test (IAT) (see the Perspective by Greenwald). Why does this matter…
1,574 Citations
Context Matters: Recovering Human Semantic Structure from Machine Learning Analysis of Large‐Scale Text Corpora
- Computer ScienceCogn. Sci.
- 2022
This work trained state‐of‐the‐art machine learning algorithms using contextually‐constrained text corpora (domain‐specific subsets of Wikipedia articles, 50+ million words each) and showed that this procedure greatly improved predictions of empirical similarity judgments and feature ratings of contextually relevant concepts.
Large pre-trained language models contain human-like biases of what is right and wrong to do
- Computer ScienceNat. Mach. Intell.
- 2022
The capabilities of the “moral direction” for guiding (even other) LMs towards producing normative text and showcase it on RealToxicityPrompts testbed, preventing the neural toxic degeneration in GPT-2.
Corpus-based Comparison of Distributional Models of Language and Knowledge Graphs
- Computer Science
- 2020
A corpus-based comparative model is introduced that allows us to compare representations of different sources generated under the distributional semantic theory, and it is shown that it can also deal with entity names and not just words, overcoming some problems that follow from the ambiguity of natural language.
Beyond Big Data: What Can We Learn from AI Models?: Invited Keynote
- Computer ScienceAISec@CCS
- 2017
It is shown that humans have unique linguistic fingerprints, which can be de-anonymize humans that have written certain text, source code, or even executable binaries of compiled code, which is a serious privacy threat for individuals that would like to remain anonymous.
Semantic projection: recovering human knowledge of multiple, distinct object features from word embeddings
- Computer Science, PsychologyArXiv
- 2018
A powerful, domain-general solution: "semantic projection" of word-vectors onto lines that represent various object features, like size, intelligence, and danger, which recovers human judgments across a range of object categories and properties.
Word Embeddings Reveal How Fundamental Sentiments Structure Natural Language
- PsychologyAmerican Behavioral Scientist
- 2019
Central to affect control theory are culturally shared meanings of concepts. That these sentiments overlap among members of a culture presumably reflects their roots in the language use that members…
What are the Biases in My Word Embedding?
- Computer ScienceAIES
- 2019
An algorithm for enumerating biases in word embeddings that outputs a number of Word Embedding Association Tests (WEATs) that capture various biases present in the data, which makes it easier to identify biases against intersectional groups, which depend on combinations of sensitive features.
The Generative Nature of Commonsense Knowledge: Insights from Machine Learning
- Computer Science
The main finding of this paper is that the knowledge base that directly facilitates both human agreement and the model’s measure of fit is by its very nature generative, and only truly exists in representation as it is applied.
Semantic projection recovers rich human knowledge of multiple object features from word embeddings.
- Computer ScienceNature human behaviour
- 2022
A domain-general method is used to extract context-dependent relationships from word embeddings: 'semantic projection' of word-vectors onto lines that represent features such as size or danger ('safe' to 'dangerous'), analogous to 'mental scales'.
Predicting High-Level Human Judgment Across Diverse Behavioral Domains
- Psychology, Computer ScienceCollabra: Psychology
- 2019
It is shown that word embeddings can be used to predict complex theoretically- and practically- relevant human perceptions and evaluations in domains as diverse as social cognition, health behavior, risk perception, organizational behavior, and marketing.
References
SHOWING 1-10 OF 75 REFERENCES
Extracting Semantics from the Enron Corpus
- Psychology
- 2013
Indirect measures must be used when analysing attitudes, as individuals are unlikely to voluntarily express beliefs that are opposed by social norms. The IAT indirectly assesses attitudes through the…
The Distributional Hypothesis
- Linguistics
- 2008
There is a correlation between distributional similarity and meaning similarity, which allows us to utilize the former in order to estimate the latter, and one can pose two very basic questions concerning the distributional hypothesis: what kind of distributional properties the authors should look for, and what — if any — the differences are between different kinds of Distributional properties.
Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings
- Computer ScienceNIPS
- 2016
This work empirically demonstrates that its algorithms significantly reduce gender bias in embeddings while preserving the its useful properties such as the ability to cluster related concepts and to solve analogy tasks.
Extracting semantic representations from word co-occurrence statistics: A computational study
- PsychologyBehavior research methods
- 2007
This article presents a systematic exploration of the principal computational possibilities for formulating and validating representations of word meanings from word co-occurrence statistics and finds that, once the best procedures are identified, a very simple approach is surprisingly successful and robust over a range of psychologically relevant evaluation measures.
A Neural Probabilistic Language Model
- Computer ScienceJ. Mach. Learn. Res.
- 2000
This work proposes to fight the curse of dimensionality by learning a distributed representation for words which allows each training sentence to inform the model about an exponential number of semantically neighboring sentences.
From Frequency to Meaning: Vector Space Models of Semantics
- Computer ScienceJ. Artif. Intell. Res.
- 2010
The goal in this survey is to show the breadth of applications of VSMs for semantics, to provide a new perspective on VSMs, and to provide pointers into the literature for those who are less familiar with the field.
Word and Object
- Philosophy
- 1960
This edition offers a new preface by Quine's student and colleague Dagfinn Follesdal that describes the never-realized plans for a second edition of Word and Object, in which Quine would offer a more unified treatment of the public nature of meaning, modalities, and propositional attitudes.
The Direct Route: Mediated Priming in Semantic Space
- Psychology
- 2000
McKoon and Ratcliff (1992) presented a theory of mediated priming where the priming effect is due to a direct but weak relatedness between prime and target. They also introduced a quantitative…
Accessing Different Types of Lexical Semantic Information: Evidence From Priming
- Psychology
- 1995
The types of semantic information that are automatically retrieved from the mental lexicon on hearing a word were investigated in 3 semantic priming experiments. The authors probed for activation of…
Temporal associations and prior-list intrusions in free recall.
- PsychologyJournal of experimental psychology. Learning, memory, and cognition
- 2006
Although repetition enhanced recall of list items, subjects were significantly more likely to make PLIs following the recall of repeated items, suggesting that temporal associations formed in earlier lists can induce recall errors.