Describing images by feeding LSTM with structural words

@article{Ma2016DescribingIB,
  title={Describing images by feeding LSTM with structural words},
  author={Shubo Ma and Yahong Han},
  journal={2016 IEEE International Conference on Multimedia and Expo (ICME)},
  year={2016},
  pages={1-6}
}
Generating semantic description draws increasing attention recently. Describing objects with adaptive adjunct words make the sentence more informative. In this paper, we focus on the generation of descriptions for images according to the structural words we have generated such as a tetrad of <;object, attribute, activity, scene>. We propose to use deep machine translation method to generate semantically meaningful descriptions. In particular, the description is composed of objects with… CONTINUE READING

Citations

Publications citing this paper.

References

Publications referenced by this paper.
Showing 1-10 of 21 references

Deep Visual-Semantic Alignments for Generating Image Descriptions

IEEE Transactions on Pattern Analysis and Machine Intelligence • 2015
View 2 Excerpts

From captions to visual concepts and back

2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) • 2015
View 1 Excerpt

Long-Term Recurrent Convolutional Networks for Visual Recognition and Description

IEEE Transactions on Pattern Analysis and Machine Intelligence • 2015
View 1 Excerpt

Semisupervised Feature Selection via Spline Regression for Video Semantic Recognition

IEEE Transactions on Neural Networks and Learning Systems • 2015
View 1 Excerpt

Show and tell: A neural image caption generator

2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) • 2015
View 3 Excerpts

Similar Papers

Loading similar papers…