DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs

@inproceedings{Dua2019DROPAR,
  title={DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs},
  author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner},
  booktitle={NAACL},
  year={2019}
}
Reading comprehension has recently seen rapid progress, with systems matching humans on the most popular datasets for the task. However, a large body of work has highlighted the brittleness of these systems, showing that there is much work left to be done. We introduce a new reading comprehension benchmark, DROP, which requires Discrete Reasoning Over the content of Paragraphs. In this crowdsourced, adversarially-created, 55k-question benchmark, a system must resolve references in a question… Expand
On Making Reading Comprehension More Comprehensive
Comprehensive Multi-Dataset Evaluation of Reading Comprehension
ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning
Tag-based Multi-Span Extraction in Reading Comprehension
MOCHA: A Dataset for Training and Evaluating Generative Reading Comprehension Metrics
Getting Closer to AI Complete Question Answering: A Set of Prerequisite Real Tasks
Quoref: A Reading Comprehension Dataset with Questions Requiring Coreferential Reasoning
A Multi-Type Multi-Span Network for Reading Comprehension that Requires Discrete Reasoning
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 58 REFERENCES
The NarrativeQA Reading Comprehension Challenge
A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task
Constructing Datasets for Multi-hop Reading Comprehension Across Documents
SQuAD: 100, 000+ Questions for Machine Comprehension of Text
Know What You Don't Know: Unanswerable Questions for SQuAD
...
1
2
3
4
5
...