Multimodal Semantic Learning from Child-Directed Input

  title={Multimodal Semantic Learning from Child-Directed Input},
  author={Angeliki Lazaridou and Grzegorz Chrupala and Raquel Fern{\'a}ndez and Marco Baroni},
Children learn the meaning of words by being exposed to perceptually rich situations (linguistic discourse, visual scenes, etc). Current computational learning models typically simulate these rich situations through impoverished symbolic approximations. In this work, we present a distributed word learning model that operates on child-directed speech paired with realistic visual scenes. The model integrates linguistic and extra-linguistic information (visual and social cues), handles referential… CONTINUE READING

From This Paper

Figures, tables, and topics from this paper.

Explore Further: Topics Discussed in This Paper


Publications referenced by this paper.
Showing 1-10 of 27 references

ImageNet: A large-scale hierarchical image database

2009 IEEE Conference on Computer Vision and Pattern Recognition • 2009
View 5 Excerpts
Highly Influenced

Predicting the birth of a spoken word.

Proceedings of the National Academy of Sciences of the United States of America • 2015
View 1 Excerpt

Where are they looking? In Advances in Neural Information Processing Systems (NIPS)

Antonio Torralba.
∗ indicates equal contribution. • 2015

Similar Papers

Loading similar papers…