What’s Missing: A Knowledge Gap Guided Approach for Multi-hop Question Answering

@article{Khot2019WhatsMA,
  title={What’s Missing: A Knowledge Gap Guided Approach for Multi-hop Question Answering},
  author={Tushar Khot and Ashish Sabharwal and Peter Clark},
  journal={ArXiv},
  year={2019},
  volume={abs/1909.09253}
}
Multi-hop textual question answering requires combining information from multiple sentences. [...] Key Method The model, GapQA, learns to fill this gap by determining the relationship between the span and an answer choice, based on retrieved knowledge targeting this gap. We propose jointly training a model to simultaneously fill this knowledge gap and compose it with the provided partial knowledge. On the OpenBookQA dataset, given partial knowledge, explicitly identifying what's missing substantially…Expand
QASC: A Dataset for Question Answering via Sentence Composition
TLDR
This work presents a multi-hop reasoning dataset, Question Answering via Sentence Composition (QASC), that requires retrieving facts from a large corpus and composing them to answer a multiple-choice question, and presents a two-step approach to mitigate the retrieval challenges. Expand
How Additional Knowledge can Improve Natural Language Commonsense Question Answering
TLDR
This work first categorizes external knowledge sources, and shows performance does improve on using such sources, then explores three different strategies for knowledge incorporation and four different models for question-answering using external commonsense knowledge. Expand
Knowledge Fusion and Semantic Knowledge Ranking for Open Domain Question Answering
TLDR
This work learns a semantic knowledge ranking model to re-rank knowledge retrieved through Lucene based information retrieval systems and proposes a "knowledge fusion model" which leverages knowledge in BERT-based language models with externally retrieved knowledge and improves the knowledge understanding of the BERT based language models. Expand
Unsupervised Alignment-based Iterative Evidence Retrieval for Multi-hop Question Answering
TLDR
This work introduces a simple, fast, and unsupervised iterative evidence retrieval method that outperforms all the previous methods on the evidence selection task on two datasets: MultiRC and QASC. Expand
IIRC: A Dataset of Incomplete Information Reading Comprehension Questions
TLDR
A dataset with more than 13K questions over paragraphs from English Wikipedia that provide only partial information to answer them, with the missing information occurring in one or more linked documents, finding that it achieves 31.1% F1 on this task, while estimated human performance is 88.4%. Expand
If You Want to Go Far Go Together: Unsupervised Joint Candidate Evidence Retrieval for Multi-hop Question Answering
TLDR
This work proposes a simple approach that retrieves and reranks set of evidence facts jointly and shows that jointly retrieving candidate evidence leads to substantially higher evidence retrieval performance when fed to the same supervised reranker. Expand
Towards Question Format Independent Numerical Reasoning: A Set of Prerequisite Tasks
TLDR
This work introduces NUMBERGAME, a multifaceted benchmark to evaluate model performance across numerical reasoning tasks of eight diverse formats, and takes forward the recent progress in generic system development, demonstrating the scope of under-explored tasks. Expand
A Neural Model for Joint Document and Snippet Ranking in Question Answering for Large Document Collections
TLDR
This work presents an architecture for joint document and snippet ranking, the two middle stages, which leverages the intuition that relevant documents have good snippets and good snippets come from relevant documents and can be used with any neural text relevance ranker. Expand
Generating Commonsense Explanation by Extracting Bridge Concepts from Reasoning Paths
TLDR
This work proposes a method that first extracts the underlying concepts which are served as bridges in the reasoning chain and then integrates these concepts to generate the final explanation. Expand
A Survey on Explainability in Machine Reading Comprehension
TLDR
This paper presents a systematic review of benchmarks and approaches for explainability in Machine Reading Comprehension (MRC), and presents the evaluation methodologies to assess the performance of explainable systems. Expand
...
1
2
...

References

SHOWING 1-10 OF 52 REFERENCES
Careful Selection of Knowledge to Solve Open Book Question Answering
TLDR
This paper addresses QA with respect to the OpenBookQA dataset and combines state of the art language models with abductive information retrieval (IR), information gain based re-ranking, passage selection and weighted scoring to achieve 72.0% accuracy. Expand
Constructing Datasets for Multi-hop Reading Comprehension Across Documents
TLDR
A novel task to encourage the development of models for text understanding across multiple documents and to investigate the limits of existing methods, in which a model learns to seek and combine evidence — effectively performing multihop, alias multi-step, inference. Expand
Improving Question Answering with External Knowledge
TLDR
This work explores simple yet effective methods for exploiting two sources of externalknowledge for exploiting unstructured external knowledge for subject-area QA on multiple-choice question answering tasks in subject areas such as science. Expand
HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering
TLDR
It is shown that HotpotQA is challenging for the latest QA systems, and the supporting facts enable models to improve performance and make explainable predictions. Expand
Understanding Dataset Design Choices for Multi-hop Reasoning
TLDR
This paper investigates two recently proposed datasets, WikiHop and HotpotQA, and explores sentence-factored models for these tasks; by design, these models cannot do multi-hop reasoning, but they are still able to solve a large number of examples in both datasets. Expand
Combining Natural Logic and Shallow Reasoning for Question Answering
TLDR
This work extends the breadth of inferences afforded by natural logic to include relational entailment and meronymy and trains an evaluation function – akin to gameplaying – to evaluate the expected truth of candidate premises on the fly. Expand
Semantic Parsing on Freebase from Question-Answer Pairs
TLDR
This paper trains a semantic parser that scales up to Freebase and outperforms their state-of-the-art parser on the dataset of Cai and Yates (2013), despite not having annotated logical forms. Expand
Question Answering via Integer Programming over Semi-Structured Knowledge
TLDR
This work proposes a structured inference system for this task, formulated as an Integer Linear Program (ILP), that answers natural language questions using a semi-structured knowledge base derived from text, including questions requiring multi-step inference and a combination of multiple facts. Expand
Looking Beyond the Surface: A Challenge Set for Reading Comprehension over Multiple Sentences
TLDR
The dataset is the first to study multi-sentence inference at scale, with an open-ended set of question types that requires reasoning skills, and finds human solvers to achieve an F1-score of 88.1%. Expand
Compositional Semantic Parsing on Semi-Structured Tables
TLDR
This paper proposes a logical-form driven parsing algorithm guided by strong typing constraints and shows that it obtains significant improvements over natural baselines and is made publicly available. Expand
...
1
2
3
4
5
...