Fast Point R-CNN

Yilun Chen, Shu Liu, Xiaoyong Shen, Jiaya Jia; The IEEE International Conference on Computer Vision (ICCV), 2019, pp. 9775-9784

Abstract


We present a unified, efficient and effective framework for point-cloud based 3D object detection. Our two-stage approach utilizes both voxel representation and raw point cloud data to exploit respective advantages. The first stage network, with voxel representation as input, only consists of light convolutional operations, producing a small number of high-quality initial predictions. Coordinate and indexed convolutional feature of each point in initial prediction are effectively fused with the attention mechanism, preserving both accurate localization and context information. The second stage works on interior points with their fused feature for further refining the prediction. Our method is evaluated on KITTI dataset, in terms of both 3D and Bird's Eye View (BEV) detection, and achieves state-of-the-arts with a 15FPS detection rate.

Related Material


[pdf]
[bibtex]
@InProceedings{Chen_2019_ICCV,
author = {Chen, Yilun and Liu, Shu and Shen, Xiaoyong and Jia, Jiaya},
title = {Fast Point R-CNN},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {October},
year = {2019}
}