Corpus ID: 220128348

Evaluation of Text Generation: A Survey

@article{elikyilmaz2020EvaluationOT,
  title={Evaluation of Text Generation: A Survey},
  author={A. Çelikyilmaz and Elizabeth Clark and Jianfeng Gao},
  journal={ArXiv},
  year={2020},
  volume={abs/2006.14799}
}
  • A. Çelikyilmaz, Elizabeth Clark, Jianfeng Gao
  • Published 2020
  • Computer Science
  • ArXiv
  • The paper surveys evaluation methods of natural language generation (NLG) systems that have been developed in the last few years. We group NLG evaluation methods into three categories: (1) human-centric evaluation metrics, (2) automatic metrics that require no training, and (3) machine-learned metrics. For each category, we discuss the progress that has been made and the challenges still being faced, with a focus on the evaluation of recently proposed NLG tasks and neural NLG models. We then… CONTINUE READING

    References

    Publications referenced by this paper.
    SHOWING 1-10 OF 253 REFERENCES
    A Study of Translation Edit Rate with Targeted Human Annotation
    1787
    BLEURT: Learning Robust Metrics for Text Generation
    9
    Generating Semantically Precise Scene Graphs from Textual Descriptions for Improved Image Retrieval
    123
    ROUGE: A Package for Automatic Evaluation of Summaries
    4305
    SPICE: Semantic Propositional Image Caption Evaluation
    406
    The NIST 1999 Speaker Recognition Evaluation - An Overview
    164
    The Significance of Recall in Automatic Metrics for MT Evaluation
    114
    A Quantitative Method for Machine Translation Evaluation
    20
    BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
    8573