Source-free Domain Adaptation for Video Object Detection Under Adverse Image Conditions

Xingguang Zhang, Chih-Hsien Chou; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2024, pp. 5010-5019

Abstract


When deploying pre-trained video object detectors in real-world scenarios the domain gap between training and testing data caused by adverse image conditions often leads to performance degradation. Addressing this issue becomes particularly challenging when only the pre-trained model and degraded videos are available. Although various source-free domain adaptation (SFDA) methods have been proposed for single-frame object detectors SFDA for video object detection (VOD) remains unexplored. Moreover most unsupervised domain adaptation works for object detection rely on two-stage detectors while SFDA for one-stage detectors which are more vulnerable to fine-tuning is not well addressed in the literature. In this paper we propose Spatial-Temporal Alternate Refinement with Mean Teacher (STAR-MT) a simple yet effective SFDA method for VOD. Specifically we aim to improve the performance of the one-stage VOD method YOLOV under adverse image conditions including noise air turbulence and haze. Extensive experiments on the ImageNetVOD dataset and its degraded versions demonstrate that our method consistently improves video object detection performance in challenging imaging conditions showcasing its potential for real-world applications.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Zhang_2024_CVPR, author = {Zhang, Xingguang and Chou, Chih-Hsien}, title = {Source-free Domain Adaptation for Video Object Detection Under Adverse Image Conditions}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2024}, pages = {5010-5019} }