HellaSwag: Can a Machine Really Finish Your Sentence?

@inproceedings{Zellers2019HellaSwagCA,
  title={HellaSwag: Can a Machine Really Finish Your Sentence?},
  author={Rowan Zellers and Ari Holtzman and Yonatan Bisk and Ali Farhadi and Yejin Choi},
  booktitle={ACL},
  year={2019}
}
  • Rowan Zellers, Ari Holtzman, +2 authors Yejin Choi
  • Published in ACL 2019
  • Computer Science
  • Recent work by Zellers et al. (2018) introduced a new task of commonsense natural language inference: given an event description such as "A woman sits at a piano," a machine must select the most likely followup: "She sets her fingers on the keys. [...] Key Result More broadly, it suggests a new path forward for NLP research, in which benchmarks co-evolve with the evolving state-of-the-art in an adversarial way, so as to present ever-harder challenges.Expand Abstract
    84 Citations
    Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension
    • 2
    • PDF
    G-DAUG: Generative Data Augmentation for Commonsense Reasoning
    • 8
    Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning
    • 4
    • Highly Influenced
    • PDF
    CommonGen: A Constrained Text Generation Dataset Towards Generative Commonsense Reasoning
    • 5
    Adversarial NLI: A New Benchmark for Natural Language Understanding
    • 69
    • PDF
    Generative Data Augmentation for Commonsense Reasoning.
    • 1
    • PDF
    CommonGen: A Constrained Text Generation Challenge for Generative Commonsense Reasoning
    • 6
    • PDF
    Generating Adversarial Examples for Topic-Dependent Argument Classification
    Evaluating NLP Models via Contrast Sets
    • 42
    • PDF

    References

    SHOWING 1-10 OF 23 REFERENCES
    SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference
    • 236
    • PDF
    Annotation Artifacts in Natural Language Inference Data
    • 339
    • PDF
    BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
    • 11,729
    • Highly Influential
    • PDF
    Language Models are Unsupervised Multitask Learners
    • 1,986
    • PDF
    Improving Language Understanding by Generative Pre-Training
    • 1,531
    • Highly Influential
    • PDF
    The Curious Case of Neural Text Degeneration
    • 251
    • PDF
    Enhanced LSTM for Natural Language Inference
    • 510
    • PDF
    Hypothesis Only Baselines in Natural Language Inference
    • 189
    • PDF
    Synthetic and Natural Noise Both Break Neural Machine Translation
    • 260
    • PDF