Prompting Contrastive Explanations for Commonsense Reasoning Tasks

  title={Prompting Contrastive Explanations for Commonsense Reasoning Tasks},
  author={Bhargavi Paranjape and Julian Michael and Marjan Ghazvininejad and Luke Zettlemoyer and Hanna Hajishirzi},
Many commonsense reasoning NLP tasks involve choosing between one or more possible answers to a question or prompt based on knowledge that is often implicit. Large pretrained language models (PLMs) can achieve near-human performance on such tasks, while providing little human-interpretable evidence of the underlying reasoning they use. In this work, we show how to use these same models to generate such evidence: inspired by the contrastive nature of human explanations, we use PLMs to complete… Expand


Explain Yourself! Leveraging Language Models for Commonsense Reasoning
This work collects human explanations for commonsense reasoning in the form of natural language sequences and highlighted annotations in a new dataset called Common Sense Explanations to train language models to automatically generate explanations that can be used during training and inference in a novel Commonsense Auto-Generated Explanation framework. Expand
PIQA: Reasoning about Physical Commonsense in Natural Language
The task of physical commonsense reasoning and a corresponding benchmark dataset Physical Interaction: Question Answering or PIQA are introduced and analysis about the dimensions of knowledge that existing models lack are provided, which offers significant opportunities for future research. Expand
A Simple Method for Commonsense Reasoning
Key to this method is the use of language models, trained on a massive amount of unlabled data, to score multiple choice questions posed by commonsense reasoning tests, which outperform previous state-of-the-art methods by a large margin. Expand
Social IQA: Commonsense Reasoning about Social Interactions
Social IQa is introduced, the first largescale benchmark for commonsense reasoning about social situations, using a new framework that mitigates stylistic artifacts in incorrect answers by asking workers to provide the right answer to a different but related question. Expand
Explaining Question Answering Models through Text Generation
A model for multi-choice question answering, where a LM-based generator generates a textual hypothesis that is later used by a classifier to answer the question, and produces hypotheses that elucidate the knowledge used by the LM for answering the question. Expand
e-SNLI: Natural Language Inference with Natural Language Explanations
The Stanford Natural Language Inference dataset is extended with an additional layer of human-annotated natural language explanations of the entailment relations, which can be used for various goals, such as obtaining full sentence justifications of a model’s decisions, improving universal sentence representations and transferring to out-of-domain NLI datasets. Expand
CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge
This work presents CommonsenseQA: a challenging new dataset for commonsense question answering, which extracts from ConceptNet multiple target concepts that have the same semantic relation to a single source concept. Expand
WinoWhy: A Deep Diagnosis of Essential Commonsense Knowledge for Answering Winograd Schema Challenge
This paper presents the first comprehensive categorization of essential commonsense knowledge for answering the Winograd Schema Challenge (WSC), and leverages the collected reasons to develop a new task called WinoWhy, which requires models to distinguish plausible reasons from very similar but wrong reasons for all WSC questions. Expand
Contrastive Explanation: A Structural-Model Approach
The structural causal model approach is extended to define two complementary notions of contrastive explanation, and these notions are demonstrated on two classical AI problems: classification and planning. Expand
Eliciting Knowledge from Language Models Using Automatically Generated Prompts
The remarkable success of pretrained language models has motivated the study of what kinds of knowledge these models learn during pretraining. Reformulating tasks as fill-in-the-blanks problemsExpand