Attentive Semantic Video Generation Using Captions

Tanya Marwah, Gaurav Mittal, Vineeth N. Balasubramanian; Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 1426-1434

Abstract


This paper proposes a network architecture to perform variable length semantic video generation using captions. We adopt a new perspective towards video generation where we allow the captions to be combined with the long-term and short-term dependencies between video frames and thus generate a video in an incremental manner. Our experiments demonstrate our network architecture's ability to distinguish between objects, actions and interactions in a video and combine them to generate videos for unseen captions. The network also exhibits the capability to perform spatio-temporal style transfer when asked to generate videos for a sequence of captions. We also show that the network's ability to learn a latent representation allows it generate videos in an unsupervised manner and perform other tasks such as action recognition.

Related Material


[pdf] [arXiv]
[bibtex]
@InProceedings{Marwah_2017_ICCV,
author = {Marwah, Tanya and Mittal, Gaurav and Balasubramanian, Vineeth N.},
title = {Attentive Semantic Video Generation Using Captions},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV)},
month = {Oct},
year = {2017}
}