Unsupervised Point Cloud Pre-Training via Occlusion Completion

Hanchen Wang, Qi Liu, Xiangyu Yue, Joan Lasenby, Matt J. Kusner; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 9782-9792

Abstract


We describe a simple pre-training approach for point clouds. It works in three steps: 1. Mask all points occluded in a camera view; 2. Learn an encoder-decoder model to reconstruct the occluded points; 3. Use the encoder weights as initialisation for downstream point cloud tasks. We find that even when we pre-train on a single dataset (ModelNet40), this method improves accuracy across different datasets and encoders, on a wide range of downstream tasks. Specifically, we show that our method outperforms previous pre-training methods in object classification, and both part-based and semantic segmentation tasks. We study the pre-trained features and find that they lead to wide downstream minima, have high transformation invariance, and have activations that are highly correlated with part labels. Code and data are available at https://github.com/hansen7/OcCo

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Wang_2021_ICCV, author = {Wang, Hanchen and Liu, Qi and Yue, Xiangyu and Lasenby, Joan and Kusner, Matt J.}, title = {Unsupervised Point Cloud Pre-Training via Occlusion Completion}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2021}, pages = {9782-9792} }