HCCL at SemEval-2017 Task 2: Combining Multilingual Word Embeddings and Transliteration Model for Semantic Similarity

@inproceedings{He2017HCCLAS,
  title={HCCL at SemEval-2017 Task 2: Combining Multilingual Word Embeddings and Transliteration Model for Semantic Similarity},
  author={Junqing He and Long Wu and Xuemin Zhao and Yonghong Yan},
  booktitle={SemEval@ACL},
  year={2017}
}
In this paper, we introduce an approach to combining word embeddings and machine translation for multilingual semantic word similarity, the task2 of SemEval2017. Thanks to the unsupervised transliteration model, our cross-lingual word embeddings encounter decreased sums of OOVs. Our results are produced using only monolingual Wikipedia corpora and a limited amount of sentence-aligned data. Although relatively little resources are utilized, our system ranked 3rd in the monolingual subtask and… 

Figures and Tables from this paper

SemEval-2017 Task 2: Multilingual and Cross-lingual Semantic Word Similarity

TLDR
Results show that systems that combine statistical knowledge from text corpora, in the form of word embeddings, and external knowledge from lexical resources are best performers in both subtasks.

Uppsala University at SemEval-2022 Task 1: Can Foreign Entries Enhance an English Reverse Dictionary?

We present the Uppsala University system for SemEval-2022 Task 1: Comparing Dictionaries and Word Embeddings (CODWOE). We explore the performance of multilingual reverse dictionaries as well as the

Leveraging Orthographic Similarity for Multilingual Neural Transliteration

TLDR
This work proposes a modified neural encoder-decoder model that maximizes parameter sharing across language pairs in order to effectively leverage orthographic similarity in transliteration involving related tasks.

References

SHOWING 1-10 OF 32 REFERENCES

SemEval-2017 Task 2: Multilingual and Cross-lingual Semantic Word Similarity

TLDR
Results show that systems that combine statistical knowledge from text corpora, in the form of word embeddings, and external knowledge from lexical resources are best performers in both subtasks.

Trans-gram, Fast Cross-lingual Word-embeddings

TLDR
Trans-gram is introduced, a simple and computationally-efficient method to simultaneously learn and align wordembeddings for a variety of languages, using only monolingual data and a smaller set of sentence-aligned data, and shows that some linguistic features are aligned across languages for which the authors do not have aligned data.

Bilingual Word Embeddings from Parallel and Non-parallel Corpora for Cross-Language Text Classification

TLDR
Bilingual paRAgraph VEctors is introduced, a model to learn bilingual distributed representations of words without word alignments either from sentencealigned parallel or label-aligned non-parallel document corpora to support cross-language text classification.

Bilingual Word Embeddings for Phrase-Based Machine Translation

TLDR
A method to learn bilingual embeddings from a large unlabeled corpus, while utilizing MT word alignments to constrain translational equivalence is proposed, which significantly out-perform baselines in word semantic similarity.

Simple task-specific bilingual word embeddings

TLDR
A simple wrapper method that uses off-the-shelf word embedding algorithms to learn task-specific bilingual word embeddings that is independent of the choice of embedding algorithm, does not require parallel data, and can be adapted to specific tasks by re-defining the equivalence classes.

Multilingual Models for Compositional Distributed Semantics

TLDR
A novel technique for learning semantic representations, which extends the distributional hypothesis to multilingual data and joint-space embeddings and demonstrates that these representations are semantically plausible and can capture semantic relationships across languages without parallel data.

Integrating an Unsupervised Transliteration Model into Statistical Machine Translation

TLDR
This work induces a transliteration model from parallel data and uses it to translate OOV words and shows that the mined transliterations provide better rule coverage and translation quality compared to the gold standard transliterated corpora.

Improving Distributional Similarity with Lessons Learned from Word Embeddings

TLDR
It is revealed that much of the performance gains of word embeddings are due to certain system design choices and hyperparameter optimizations, rather than the embedding algorithms themselves, and these modifications can be transferred to traditional distributional models, yielding similar gains.

Exploiting Similarities among Languages for Machine Translation

TLDR
This method can translate missing word and phrase entries by learning language structures based on large monolingual data and mapping between languages from small bilingual data and uses distributed representation of words and learns a linear mapping between vector spaces of languages.

GloVe: Global Vectors for Word Representation

TLDR
A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.