SnapNet-R: Consistent 3D Multi-View Semantic Labeling for Robotics

Joris Guerry, Alexandre Boulch, Bertrand Le Saux, Julien Moras, Aurelien Plyer, David Filliat; Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 669-678

Abstract


In this paper we present a new approach for semantic recognition in the context of robotics. When a robot evolves in its environment, it gets 3D information given either by its sensors or by its own motion through 3D reconstruction. Our approach uses (i) 3D-coherent synthesis of scene observations and (ii) mix them in a multi-view framework for 3D labeling. (iii) This is efficient locally (for 2D semantic segmentation) and globally (for 3D structure labeling). This allows to add semantics to the observed scene that goes beyond simple image classification, as shown on challenging datasets such as SUNRGBD or the 3DRMS Reconstruction Challenge.

Related Material


[pdf]
[bibtex]
@InProceedings{Guerry_2017_ICCV,
author = {Guerry, Joris and Boulch, Alexandre and Le Saux, Bertrand and Moras, Julien and Plyer, Aurelien and Filliat, David},
title = {SnapNet-R: Consistent 3D Multi-View Semantic Labeling for Robotics},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV) Workshops},
month = {Oct},
year = {2017}
}