A Cylindrical Convolution Network for Dense Top-View Semantic Segmentation with LiDAR Point Clouds

Jiacheng Lu, Shuo Gu, Cheng-Zhong Xu, Hui Kong; Proceedings of the Asian Conference on Computer Vision (ACCV), 2022, pp. 4648-4664

Abstract


Accurate semantic scene understanding of the surrounding environment is a challenge for autonomous driving systems. Recent LiDAR-based semantic segmentation methods mainly focus on predicting point-wise semantic classes, which cannot be directly used before the further densification process. In this paper, we propose a cylindrical convolution network for dense semantic understanding in the top-view LiDAR data representation. 3D LiDAR point clouds are divided into cylindrical partitions before feeding to the network, where semantic segmentation is conducted in the cylindrical representation. Then a cylinder-to-BEV transformation module is introduced to obtain sparse semantic feature maps in the top view. In the end, we propose a modified encoder-decoder network to get the dense semantic estimations. Experimental results on the SemanticKITTI and nuScenes-LidarSeg datasets show that our method outperforms the state-of-the-art methods with a large margin.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Lu_2022_ACCV, author = {Lu, Jiacheng and Gu, Shuo and Xu, Cheng-Zhong and Kong, Hui}, title = {A Cylindrical Convolution Network for Dense Top-View Semantic Segmentation with LiDAR Point Clouds}, booktitle = {Proceedings of the Asian Conference on Computer Vision (ACCV)}, month = {December}, year = {2022}, pages = {4648-4664} }