Cosmos QA: Machine Reading Comprehension with Contextual Commonsense Reasoning

@inproceedings{Huang2019CosmosQM,
  title={Cosmos QA: Machine Reading Comprehension with Contextual Commonsense Reasoning},
  author={Lifu Huang and Ronan Le Bras and Chandra Bhagavatula and Yejin Choi},
  booktitle={EMNLP/IJCNLP},
  year={2019}
}
Understanding narratives requires reading between the lines, which in turn, requires interpreting the likely causes and effects of events, even when they are not mentioned explicitly. [...] Key Method To establish baseline performances on Cosmos QA, we experiment with several state-of-the-art neural architectures for reading comprehension, and also propose a new architecture that improves over the competitive baselines. Experimental results demonstrate a significant gap between machine (68.4%) and human…Expand
Commonsense Evidence Generation and Injection in Reading Comprehension
TLDR
A Commonsense Evidence Generation and Injection framework in reading comprehension, named CEGI, which injects two kinds of auxiliary commonsense evidence into comprehensive reading to equip the machine with the ability of rational thinking. Expand
On Making Reading Comprehension More Comprehensive
TLDR
This work justifies a question answering approach to reading comprehension and describes the various kinds of questions one might use to more fully test a system’s comprehension of a passage, moving beyond questions that only probe local predicate-argument structures. Expand
Reasoning in Dialog: Improving Response Generation by Context Reading Comprehension
TLDR
This paper proposes a joint framework that unifies these two tasks, sharing the same encoder to extract the common and task-invariant features with different decoders to learn task-specific features, and augment the Transformer architecture with a memory updater, designed to selectively store and update the history dialog information so as to support downstream tasks. Expand
COMMONGEN: Towards Generative Commonsense Reasoning via A Constrained Text Generation Challenge
Given a set of common concepts like “{apple (noun), pick (verb), tree (noun)}”, humans find it easy to write a sentence describing a grammatical and logically coherent scenario that covers theseExpand
CommonGen: A Constrained Text Generation Dataset Towards Generative Commonsense Reasoning
TLDR
This work presents CommonGen: a challenging dataset for testing generative commonsense reasoning with a constrained text generation task, and provides high-quality rationales behind the reasoning process for the development and test sets from the human annotators. Expand
ESTER: A Machine Reading Comprehension Dataset for Event Semantic Relation Reasoning
TLDR
ESTER, a comprehensive machine reading comprehension (MRC) dataset for Event Semantic Relation Reasoning is introduced, which study five most commonly used event semantic relations and formulate them as questionanswering tasks. Expand
Natural Language Inference in Context - Investigating Contextual Reasoning over Long Texts
TLDR
ConTRoL is a new dataset for ConTextual Reasoning over Long texts, a passage-level NLI dataset with a focus on complex contextual reasoning types such as logical reasoning, derived from competitive selection and recruitment test for police recruitment with expert level quality. Expand
Explaining Question Answering Models through Text Generation
TLDR
A model for multi-choice question answering, where a LM-based generator generates a textual hypothesis that is later used by a classifier to answer the question, and produces hypotheses that elucidate the knowledge used by the LM for answering the question. Expand
Reasoning Paths Generation for Commonsense Question Answering
  • Peifeng Wang
  • 2019
Commonsense question answering (QA) requires a model to acquire some necessary background knowledge about how the world operates and people interact with each others. A large number of works haveExpand
Seeing the World through Text: Evaluating Image Descriptions for Commonsense Reasoning in Machine Reading Comprehension
Despite recent achievements in natural language understanding, reasoning over commonsense knowledge still represents a big challenge to AI systems. As the name suggests, common sense is related toExpand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 49 REFERENCES
The NarrativeQA Reading Comprehension Challenge
TLDR
A new dataset and set of tasks in which the reader must answer questions about stories by reading entire books or movie scripts are presented, designed so that successfully answering their questions requires understanding the underlying narrative rather than relying on shallow pattern matching or salience. Expand
MCTest: A Challenge Dataset for the Open-Domain Machine Comprehension of Text
TLDR
MCTest is presented, a freely available set of stories and associated questions intended for research on the machine comprehension of text that requires machines to answer multiple-choice reading comprehension questions about fictional stories, directly tackling the high-level goal of open-domain machine comprehension. Expand
CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge
TLDR
This work presents CommonsenseQA: a challenging new dataset for commonsense question answering, which extracts from ConceptNet multiple target concepts that have the same semantic relation to a single source concept. Expand
DREAM: A Challenge Data Set and Models for Dialogue-Based Reading Comprehension
TLDR
Experimental results on the DREAM data set show the effectiveness of dialogue structure and general world knowledge, the first dialogue-based multiple-choice reading comprehension data set to focus on in-depth multi-turn multi-party dialogue understanding. Expand
MCScript: A Novel Dataset for Assessing Machine Comprehension Using Script Knowledge
TLDR
A large dataset of narrative texts and questions about these texts, intended to be used in a machine comprehension task that requires reasoning using commonsense knowledge, and shows that the mode of data collection via crowdsourcing results in a substantial amount of inference questions. Expand
Annotation Artifacts in Natural Language Inference Data
TLDR
It is shown that a simple text categorization model can correctly classify the hypothesis alone in about 67% of SNLI and 53% of MultiNLI, and that specific linguistic phenomena such as negation and vagueness are highly correlated with certain inference classes. Expand
NewsQA: A Machine Comprehension Dataset
TLDR
NewsQA, a challenging machine comprehension dataset of over 100,000 human-generated question-answer pairs, is presented and analysis confirms that NewsQA demands abilities beyond simple word matching and recognizing textual entailment. Expand
SQuAD: 100,000+ Questions for Machine Comprehension of Text
TLDR
A strong logistic regression model is built, which achieves an F1 score of 51.0%, a significant improvement over a simple baseline (20%). Expand
CoQA: A Conversational Question Answering Challenge
TLDR
CoQA is introduced, a novel dataset for building Conversational Question Answering systems and it is shown that conversational questions have challenging phenomena not present in existing reading comprehension datasets (e.g., coreference and pragmatic reasoning). Expand
Know What You Don’t Know: Unanswerable Questions for SQuAD
TLDR
SQuadRUn is a new dataset that combines the existing Stanford Question Answering Dataset (SQuAD) with over 50,000 unanswerable questions written adversarially by crowdworkers to look similar to answerable ones. Expand
...
1
2
3
4
5
...