DELTA: Dense Depth from Events and LiDAR using Transformer's Attention

Vincent Brebion, Julien Moreau, Franck Davoine; Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR) Workshops, 2025, pp. 4898-4907

Abstract


Event cameras and LiDARs provide complementary yet distinct data: respectively, asynchronous detections of changes in lighting versus sparse but accurate depth information at a fixed rate. To this day, few works have explored the combination of these two modalities. In this article, we propose a novel neural-network-based method for fusing event and LiDAR data in order to estimate dense depth maps. Our architecture, DELTA, exploits the concepts of self- and cross-attention to model the spatial and temporal relations within and between the event and LiDAR data. Following a thorough evaluation, we demonstrate that DELTA sets a new state of the art in the event-based depth estimation problem, and that it is able to reduce the errors up to four times for close ranges compared to the previous SOTA.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Brebion_2025_CVPR, author = {Brebion, Vincent and Moreau, Julien and Davoine, Franck}, title = {DELTA: Dense Depth from Events and LiDAR using Transformer's Attention}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR) Workshops}, month = {June}, year = {2025}, pages = {4898-4907} }