Spatio-temporal video autoencoder with differentiable memory

Abstract

We describe a new spatio-temporal video autoencoder, based on a classic spatial image autoencoder and a novel nested temporal autoencoder. The temporal encoder is represented by a differentiable visual memory composed of convolutional long short-term memory (LSTM) cells that integrate changes over time. Here we target motion changes and use as temporal decoder a robust optical flow prediction module together with an image sampler serving as built-in feedback loop. The architecture is end-to-end differentiable. At each time step, the system receives as input a video frame, predicts the optical flow based on the current observation and the LSTM memory state as a dense transformation map, and applies it to the current frame to generate the next frame. By minimising the reconstruction error between the predicted next frame and the corresponding ground truth next frame, we train the whole system to extract features useful for motion estimation without any supervision effort. We believe these features can in turn facilitate learning high-level tasks such as path planning, semantic segmentation, or action recognition, reducing the overall supervision effort.

Extracted Key Phrases

8 Figures and Tables

020406020162017
Citations per Year

63 Citations

Semantic Scholar estimates that this publication has 63 citations based on the available data.

See our FAQ for additional information.

Cite this paper

@article{Patraucean2015SpatiotemporalVA, title={Spatio-temporal video autoencoder with differentiable memory}, author={Viorica Patraucean and Ankur Handa and Roberto Cipolla}, journal={CoRR}, year={2015}, volume={abs/1511.06309} }