-
[pdf]
[supp]
[bibtex]@InProceedings{Wang_2024_ACCV, author = {Wang, Haocheng and Cao, Yanlong and Shou, Yejun and Shen, Lingfeng and Wei, Xiaoyao and Xu, Zhijie and Ren, Kai}, title = {iS-MAP: Neural Implicit Mapping and Positioning for Structural Environments}, booktitle = {Proceedings of the Asian Conference on Computer Vision (ACCV)}, month = {December}, year = {2024}, pages = {747-763} }
iS-MAP: Neural Implicit Mapping and Positioning for Structural Environments
Abstract
This work presents iS-MAP, a neural implicit RGB-D SLAM approach based on multi-scale hybrid representation in structural environments. iS-MAP encodes the scene using an efficient hybrid feature representation, which combines a 3D hash grid and multi-scale 2D feature planes. This hybrid representation is then decoded into TSDF and RGB values, leading to robust reconstruction and multilevel detail understanding. Additionally, we introduce Manhattan matching loss and structural consistency loss to fully incorporate the prior constraints of structured planes and lines. Compared with only color and depth losses, our structured losses are capable of guiding network optimization at the semantic level, resulting in more reasonable scene regularization. Experimental results on synthetic and real-world scene datasets demonstrate that our approach performs either better or competitive to existing neural implicit RGB-D SLAM methods in mapping and tracking accuracy, and predicts the most plausible reconstruction results for the unobserved structural regions. The source code will be released soon.
Related Material