Local to Global: Efficient Visual Localization for a Monocular Camera

Sang Jun Lee, Deokhwa Kim, Sung Soo Hwang, Donghwan Lee; Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2021, pp. 2231-2240

Abstract


Robust and accurate visual localization is one of the most fundamental elements in various technologies, such as autonomous driving and augmented reality. While recent visual localization algorithms demonstrate promising results in terms of accuracy and robustness, the associated high computational cost requires running these algorithms on server-sides rather than client devices. This paper proposes a real time monocular visual localization system that combines client-side visual odometry with server-side visual localization functionality. In particular, the proposed system utilizes handcrafted features for real time visual odometry while adopting learned features for robust visual localization. To link the two components, the proposed system employs a map alignment mechanism that transforms the local coordinates obtained using visual odometry to global coordinates. The system achieves comparable accuracy to that of the state-of-the-art structure-based methods and end-to-end methods for the visual localization on both indoor and outdoor datasets while operating in real time.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Lee_2021_WACV, author = {Lee, Sang Jun and Kim, Deokhwa and Hwang, Sung Soo and Lee, Donghwan}, title = {Local to Global: Efficient Visual Localization for a Monocular Camera}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2021}, pages = {2231-2240} }