Learning Spatial Common Sense With Geometry-Aware Recurrent Networks

Hsiao-Yu Fish Tung, Ricson Cheng, Katerina Fragkiadaki; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 2595-2603

Abstract


We integrate two powerful ideas, geometry and deep visual representation learning, into recurrent network architectures for mobile visual scene understanding. The proposed networks learn to "lift" 2D visual features and integrate them over time into latent 3D feature maps of the scene. They are equipped with differentiable geometric operations, such as projection, unprojection, egomotion estimation and stabilization, in order to compute a geometrically-consistent mapping between the world scene and their 3D latent feature space. We train the proposed architectures to predict novel image views given short frame sequences as input. Their predictions strongly generalize to scenes with a novel number of objects, appearances and configurations, and greatly outperform predictions of previous works that do not consider egomotion stabilization or a space-aware latent feature space. We train the proposed architectures to detect and segment objects in 3D, using the latent 3D feature map as input--as opposed to 2D feature maps computed from video frames. The resulting detections are permanent: they continue to exist even when an object gets occluded or leaves the field of view. Our experiments suggest the proposed space-aware latent feature arrangement and egomotion-stabilized convolutions are essential architectural choices for spatial common sense to emerge in artificial embodied visual agents.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Tung_2019_CVPR,
author = {Tung, Hsiao-Yu Fish and Cheng, Ricson and Fragkiadaki, Katerina},
title = {Learning Spatial Common Sense With Geometry-Aware Recurrent Networks},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2019}
}