Corpus ID: 24462950

SciTaiL: A Textual Entailment Dataset from Science Question Answering

@inproceedings{Khot2018SciTaiLAT,
  title={SciTaiL: A Textual Entailment Dataset from Science Question Answering},
  author={Tushar Khot and Ashish Sabharwal and Peter Clark},
  booktitle={AAAI},
  year={2018}
}
We present a new dataset and model for textual entailment, derived from treating multiple-choice question-answering as an entailment problem. [...] Key Result As a step forward, we demonstrate that one can improve accuracy on SCITAIL by 5% using a new neural model that exploits linguistic structure.Expand
End-Task Oriented Textual Entailment via Deep Exploring Inter-Sentence Interactions
TLDR
DEISTE (deep explorations of inter-sentence interactions for textual entailment) is proposed for SciTail, a natural entailment challenge derived from a multi-choice question answering problem that is more challenging than other entailment data sets and more directly useful to the end-task -- question answering. Expand
End-Task Oriented Textual Entailment via Deep Explorations of Inter-Sentence Interactions.
TLDR
DEISTE (deep explorations of inter-sentence interactions for textual entailment) is proposed for SciTail, a natural entailment challenge derived from a multi-choice question answering problem that is more challenging than other entailment data sets and more directly useful to the end-task -- question answering. Expand
AWE: Asymmetric Word Embedding for Textual Entailment
TLDR
Experimental results on SciTail and SNLI datasets show that the learned asymmetric word embeddings could significantly improve the word-word interaction based textual entailment models. Expand
Repurposing Entailment for Multi-Hop Question Answering Tasks
TLDR
Multee is introduced, a general architecture that can effectively use entailment models for multi-hop QA tasks and outperforms QA models trained only on the target QA datasets and the OpenAI transformer models when using an entailment function pre-trained on NLI datasets. Expand
Answering Science Exam Questions Using Query Rewriting with Background Knowledge
TLDR
A system that rewrites a given question into queries that are used to retrieve supporting text from a large corpus of science-related text is presented and is able to outperform several strong baselines on the ARC dataset. Expand
Answering Science Exam Questions Using Query Reformulation with Background Knowledge
TLDR
This paper presents a system that reformulates a given question into queries that are used to retrieve supporting text from a large corpus of science-related text and outperforms several strong baselines on the ARC dataset. Expand
Universal Natural Language Processing with Limited Annotations: Try Few-shot Textual Entailment as a Start
TLDR
This work demonstrates that this framework enables a pretrained entailment model to work well on new entailment domains in a few-shot setting, and shows its effectiveness as a unified solver for several downstream NLP tasks such as question answering and coreference resolution when the end-task annotations are limited. Expand
Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge
TLDR
A new question set, text corpus, and baselines assembled to encourage AI research in advanced question answering constitute the AI2 Reasoning Challenge (ARC), which requires far more powerful knowledge and reasoning than previous challenges such as SQuAD or SNLI. Expand
Knowledge-aware Textual Entailment with Graph Attention Network
TLDR
A Knowledge-Context Interactive Textual Entailment Network (KCI-TEN) that learns graph level sentence representations by harnessing external knowledge graph with graph attention network and a text-graph interaction mechanism for neural based entailment matching learning that endows the redundancy and noise with less importance and put emphasis on the informative representations. Expand
Medical Knowledge-enriched Textual Entailment Framework
TLDR
A novel Medical Knowledge-Enriched Textual Entailment framework that allows the model to acquire a semantic and global representation of the input medical text with the help of a relevant domain-specific knowledge graph is presented. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 32 REFERENCES
Textual Entailment with Structured Attentions and Composition
TLDR
This work shows that it is beneficial to extend the attention model to tree nodes between premise and hypothesis, and studies the recursive composition of this subtree-level entailment relation, which can be viewed as a soft version of the Natural Logic framework. Expand
Question Answering via Integer Programming over Semi-Structured Knowledge
TLDR
This work proposes a structured inference system for this task, formulated as an Integer Linear Program (ILP), that answers natural language questions using a semi-structured knowledge base derived from text, including questions requiring multi-step inference and a combination of multiple facts. Expand
A SICK cure for the evaluation of compositional distributional semantic models
TLDR
This work aims to help the research community working on compositional distributional semantic models (CDSMs) by providing SICK (Sentences Involving Compositional Knowldedge), a large size English benchmark tailored for them. Expand
TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension
TLDR
It is shown that, in comparison to other recently introduced large-scale datasets, TriviaQA has relatively complex, compositional questions, has considerable syntactic and lexical variability between questions and corresponding answer-evidence sentences, and requires more cross sentence reasoning to find answers. Expand
A large annotated corpus for learning natural language inference
TLDR
The Stanford Natural Language Inference corpus is introduced, a new, freely available collection of labeled sentence pairs, written by humans doing a novel grounded task based on image captioning, which allows a neural network-based model to perform competitively on natural language inference benchmarks for the first time. Expand
Exploring Markov Logic Networks for Question Answering
TLDR
A system that reasons with knowledge derived from textbooks, represented in a subset of firstorder logic, called Praline, which demonstrates a 15% accuracy boost and a 10x reduction in runtime as compared to other MLNbased methods, and comparable accuracy to word-based baseline approaches. Expand
MCTest: A Challenge Dataset for the Open-Domain Machine Comprehension of Text
TLDR
MCTest is presented, a freely available set of stories and associated questions intended for research on the machine comprehension of text that requires machines to answer multiple-choice reading comprehension questions about fictional stories, directly tackling the high-level goal of open-domain machine comprehension. Expand
Learning What is Essential in Questions
TLDR
This paper develops a classifier that reliably identifies and ranks essential terms in questions and demonstrates that the notion of question term essentiality allows state-of-the-art QA solver for elementary-level science questions to make better and more informed decisions, improving performance by up to 5%. Expand
SQuAD: 100,000+ Questions for Machine Comprehension of Text
TLDR
A strong logistic regression model is built, which achieves an F1 score of 51.0%, a significant improvement over a simple baseline (20%). Expand
The Sixth PASCAL Recognizing Textual Entailment Challenge
TLDR
This paper presents the Sixth Recognizing Textual Entailment (RTE-6) challenge, as the traditional Main Task was replaced by a new task, similar to the RTE-5 Search Pilot, in which TextualEntailment is performed on a real corpus in the Update Summarization scenario. Expand
...
1
2
3
4
...