Offline bilingual word vectors, orthogonal transformations and the inverted softmax
@article{Smith2017OfflineBW, title={Offline bilingual word vectors, orthogonal transformations and the inverted softmax}, author={S. L. Smith and David H. P. Turban and S. Hamblin and N. Hammerla}, journal={ArXiv}, year={2017}, volume={abs/1702.03859} }
Usually bilingual word vectors are trained "online". Mikolov et al. showed they can also be found "offline", whereby two pre-trained embeddings are aligned with a linear transformation, using dictionaries compiled from expert knowledge. In this work, we prove that the linear transformation between two spaces should be orthogonal. This transformation can be obtained using the singular value decomposition. We introduce a novel "inverted softmax" for identifying translation pairs, with which we… Expand
362 Citations
Word and Phrase Dictionaries Generated with Multiple Translation Paths
- Mathematics, Computer Science
- ICIST
- 2018
Bilingual Dictionary Based Neural Machine Translation without Using Parallel Sentences
- Computer Science
- ACL
- 2020
- 3
- PDF
References
SHOWING 1-10 OF 22 REFERENCES
BilBOWA: Fast Bilingual Distributed Representations without Word Alignments
- Computer Science, Mathematics
- ICML
- 2015
- 334
- PDF
An Autoencoder Approach to Learning Bilingual Word Representations
- Computer Science, Mathematics
- NIPS
- 2014
- 291
- PDF
Learning principled bilingual mappings of word embeddings while preserving monolingual invariance
- Computer Science
- EMNLP
- 2016
- 237
- Highly Influential
- PDF
Normalized Word Embedding and Orthogonal Transform for Bilingual Word Translation
- Computer Science
- HLT-NAACL
- 2015
- 279
- PDF
Exploiting Similarities among Languages for Machine Translation
- Computer Science
- ArXiv
- 2013
- 1,143
- Highly Influential
- PDF
Improving Vector Space Word Representations Using Multilingual Correlation
- Computer Science
- EACL
- 2014
- 513
- Highly Influential
- PDF