-
[pdf]
[supp]
[arXiv]
[bibtex]@InProceedings{Franchi_2024_WACV, author = {Franchi, Gianni and Hariat, Marwane and Yu, Xuanlong and Belkhir, Nacim and Manzanera, Antoine and Filliat, David}, title = {InfraParis: A Multi-Modal and Multi-Task Autonomous Driving Dataset}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2024}, pages = {2973-2983} }
InfraParis: A Multi-Modal and Multi-Task Autonomous Driving Dataset
Abstract
Current deep neural networks (DNNs) for autonomous driving computer vision are typically trained on specific datasets that only involve a single type of data and urban scenes. Consequently, these models struggle to handle new objects, noise, nighttime conditions, and diverse scenarios, which is essential for safety-critical applications. Despite ongoing efforts to enhance the resilience of computer vision DNNs, progress has been sluggish, partly due to the absence of benchmarks featuring multiple modalities. We introduce a novel and versatile dataset named InfraParis that supports multiple tasks across three modalities: RGB, depth, and infrared. We assess various state-of-the-art baseline techniques, encompassing models for the tasks of semantic segmentation, object detection, and depth estimation.
Related Material