Corpus ID: 43920307

Generating Video Description using RNN with Semantic Attention

@inproceedings{Natsuda2017GeneratingVD,
  title={Generating Video Description using RNN with Semantic Attention},
  author={Laokulrat Natsuda and Okazaki Naoaki and Nakayama Hideki},
  year={2017}
}
Being able to understand videos can have a great impact and can be useful to many other applications. However, generated descriptions by computers often fail to mention correct objects appearing in videos. This work aims to alleviate this problem by including external fine-grained visual information detected from all video frames. In this paper, we propose an LSTM-based sequence-tosequence model with semantic attention for video description generation. The results show that using semantic… Expand
1 Citations
Describing human activities in video streams

References

SHOWING 1-10 OF 11 REFERENCES
Sequence to Sequence -- Video to Text
Image Captioning with Semantic Attention
From captions to visual concepts and back
Show and tell: A neural image caption generator
Very Deep Convolutional Networks for Large-Scale Image Recognition
Microsoft COCO Captions: Data Collection and Evaluation Server
Caffe: Convolutional Architecture for Fast Feature Embedding
Glove: Global Vectors for Word Representation
Collecting Highly Parallel Data for Paraphrase Evaluation
...
1
2
...