Sequence-Level Knowledge Distillation

@inproceedings{Kim2016SequenceLevelKD,
  title={Sequence-Level Knowledge Distillation},
  author={Yoon Kim and Alexander M. Rush},
  booktitle={EMNLP},
  year={2016}
}
Neural machine translation (NMT) offers a novel alternative formulation of translation that is potentially simpler than statistical approaches. However to reach competitive performance, NMT models need to be exceedingly large. In this paper we consider applying knowledge distillation approaches (Bucila et al., 2006; Hinton et al., 2015) that have proven… CONTINUE READING

4 Figures & Tables

Topics

Statistics

0204060201620172018
Citations per Year

61 Citations

Semantic Scholar estimates that this publication has 61 citations based on the available data.

See our FAQ for additional information.