GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding

@inproceedings{Wang2018GLUEAM,
  title={GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},
  author={Alex Wang and Amanpreet Singh and Julian Michael and Felix Hill and Omer Levy and Samuel R. Bowman},
  booktitle={BlackboxNLP@EMNLP},
  year={2018}
}
For natural language understanding (NLU) technology to be maximally useful, it must be able to process language in a way that is not exclusive to a single task, genre, or dataset. [...] Key ResultWe evaluate baselines based on current methods for transfer and representation learning and find that multi-task training on all tasks performs better than training a separate model per task. However, the low absolute performance of our best model indicates the need for improved general NLU systems. Expand Abstract

Citations

Publications citing this paper.
SHOWING 1-10 OF 201 CITATIONS

BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions

VIEW 9 EXCERPTS
CITES METHODS & BACKGROUND
HIGHLY INFLUENCED

ERNIE 2.0: A Continual Pre-training Framework for Language Understanding

  • ArXiv
  • 2019
VIEW 11 EXCERPTS
CITES METHODS
HIGHLY INFLUENCED

Parameter-Efficient Transfer Learning for NLP

VIEW 4 EXCERPTS
CITES BACKGROUND
HIGHLY INFLUENCED

Improving Language Understanding by Generative Pre-Training

VIEW 16 EXCERPTS
CITES BACKGROUND & METHODS
HIGHLY INFLUENCED

FILTER CITATIONS BY YEAR

2017
2019

CITATION STATISTICS

  • 54 Highly Influenced Citations

  • Averaged 67 Citations per year from 2017 through 2019

  • 541% Increase in citations per year in 2019 over 2018

References

Publications referenced by this paper.
SHOWING 1-10 OF 48 REFERENCES

Skip-Thought Vectors

  • NIPS
  • 2015
VIEW 7 EXCERPTS
HIGHLY INFLUENTIAL

Neural Network Acceptability Judgments

  • Transactions of the Association for Computational Linguistics
  • 2018
VIEW 1 EXCERPT