Incorporating Copying Mechanism in Sequence-to-Sequence Learning
- Jiatao Gu, Zhengdong Lu, Hang Li, V. Li
- Computer ScienceAnnual Meeting of the Association for…
- 21 March 2016
This paper incorporates copying into neural network-based Seq2Seq learning and proposes a new model called CopyNet with encoder-decoder structure which can nicely integrate the regular way of word generation in the decoder with the new copying mechanism which can choose sub-sequences in the input sequence and put them at proper places in the output sequence.
Non-Autoregressive Neural Machine Translation
- Jiatao Gu, James Bradbury, Caiming Xiong, V. Li, R. Socher
- Computer ScienceInternational Conference on Learning…
- 7 November 2017
A model is introduced that avoids this autoregressive property and produces its outputs in parallel, allowing an order of magnitude lower latency during inference, and achieves near-state-of-the-art performance on WMT 2016 English-Romanian.
Learning to Translate in Real-time with Neural Machine Translation
- Graham Neubig, Kyunghyun Cho, Jiatao Gu, V. Li
- Computer ScienceConference of the European Chapter of the…
- 3 October 2016
A neural machine translation (NMT) framework for simultaneous translation in which an agent learns to make decisions on when to translate from the interaction with a pre-trained NMT environment is proposed.
Universal Neural Machine Translation for Extremely Low Resource Languages
- Jiatao Gu, Hany Hassan, Jacob Devlin, V. Li
- Computer Science, LinguisticsNorth American Chapter of the Association for…
- 14 February 2018
The proposed approach utilizing a transfer-learning approach to share lexical and sentence level representations across multiple source languages into one target language is able to achieve 23 BLEU on Romanian-English WMT2016 using a tiny parallel corpus of 6k sentences.
Search Engine Guided Neural Machine Translation
- Jiatao Gu, Yong Wang, Kyunghyun Cho, V. Li
- Computer ScienceAAAI Conference on Artificial Intelligence
- 27 April 2018
An attention-based neural machine translation model is extended by allowing it to access an entire training set of parallel sentence pairs even after training, and significantly outperforms the baseline approach.
Improved Zero-shot Neural Machine Translation via Ignoring Spurious Correlations
- Jiatao Gu, Yong Wang, Kyunghyun Cho, V. Li
- Computer ScienceAnnual Meeting of the Association for…
- 4 June 2019
This work addresses the degeneracy problem due to capturing spurious correlations by quantitatively analyzing the mutual information between language IDs of the source and decoded sentences and proposes two simple but effective approaches: decoder pre-training; back-translation.
Meta-Learning for Low-Resource Neural Machine Translation
- Jiatao Gu, Yong Wang, Yun Chen, Kyunghyun Cho, V. Li
- Computer ScienceConference on Empirical Methods in Natural…
- 25 August 2018
The proposed model-agnostic meta-learning algorithm for low-resource neural machine translation (NMT) is extended and significantly outperforms the multilingual, transfer learning based approach and enables us to train a competitive NMT system with only a fraction of training examples.
Online False Data Injection Attack Detection With Wavelet Transform and Deep Neural Networks
A new false data injection attack detection mechanism for ac state estimation that can effectively capture inconsistency by analyzing temporally consecutive estimated system states using wavelet transform and deep neural network techniques is proposed.
A Teacher-Student Framework for Zero-Resource Neural Machine Translation
- Yun Chen, Yang Liu, Yong Cheng, V. Li
- Computer ScienceAnnual Meeting of the Association for…
- 1 May 2017
This paper proposes a method for zero-resource NMT by assuming that parallel sentences have close probabilities of generating a sentence in a third language on a source-pivot parallel corpus.
Search Engine Guided Non-Parametric Neural Machine Translation
- Jiatao Gu, Yong Wang, Kyunghyun Cho, V. Li
- Computer ScienceArXiv
- 20 May 2017
Empirical evaluation of an attention-based neural machine translation model by allowing it to access an entire training set of parallel sentence pairs even after training shows that the proposed approach significantly outperforms the baseline approach and the improvement is more significant when more relevant sentence pairs were retrieved.
...
...