Video Autoencoder: Self-Supervised Disentanglement of Static 3D Structure and Motion

Zihang Lai, Sifei Liu, Alexei A. Efros, Xiaolong Wang; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 9730-9740

Abstract


We present Video Autoencoder for learning disentangled representations of 3D structure and camera pose from videos in a self-supervised manner. Relying on temporal continuity in videos, our work assumes that the 3D scene structure in nearby video frames remains static. Given a sequence of video frames as input, the Video Autoencoder extracts a disentangled representation of the scene including: (i) a temporally-consistent deep voxel feature to represent the 3D structure and (ii) a 3D trajectory of camera poses for each frame. These two representations will then be re-entangled for rendering the input video frames. Video Autoencoder can be trained directly using a pixel reconstruction loss, without any ground truth 3D or camera pose annotations. The disentangled representation can be applied to a range of tasks, including novel view synthesis, camera pose estimation, and video generation by motion following. We evaluate our method on several large-scale natural video datasets, and show generalization results on out-of-domain images.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Lai_2021_ICCV, author = {Lai, Zihang and Liu, Sifei and Efros, Alexei A. and Wang, Xiaolong}, title = {Video Autoencoder: Self-Supervised Disentanglement of Static 3D Structure and Motion}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2021}, pages = {9730-9740} }