Does Multimodality Help Human and Machine for Translation and Image Captioning?

  title={Does Multimodality Help Human and Machine for Translation and Image Captioning?},
  author={Ozan Caglayan and Walid Aransa and Yaxing Wang and Marc Masana and Mercedes Garc{\'i}a-Mart{\'i}nez and Fethi Bougares and Lo{\"i}c Barrault and Joost van de Weijer},
This paper presents the systems developed by LIUM and CVC for the WMT16 Multimodal Machine Translation challenge. We explored various comparative methods, namely phrase-based systems and attentional recurrent neural networks models trained using monomodal or multimodal data. We also performed a human evaluation in order to estimate the usefulness of multimodal data for human machine translation and image description generation. Our systems obtained the best results for both tasks according to… CONTINUE READING
Highly Cited
This paper has 51 citations. REVIEW CITATIONS
Related Discussions
This paper has been referenced on Twitter 24 times. VIEW TWEETS

From This Paper

Figures, tables, and topics from this paper.


Publications citing this paper.
Showing 1-10 of 34 extracted citations

51 Citations

Citations per Year
Semantic Scholar estimates that this publication has 51 citations based on the available data.

See our FAQ for additional information.


Publications referenced by this paper.
Showing 1-10 of 29 references

Continuous-Space Language Models for Statistical Machine Translation

Prague Bull. Math. Linguistics • 2006
View 6 Excerpts
Highly Influenced

Deep Residual Learning for Image Recognition

2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) • 2016
View 2 Excerpts

Similar Papers

Loading similar papers…