Match me if you can: Semi-Supervised Semantic Correspondence Learning with Unpaired Images

Jiwon Kim, Byeongho Heo, Sangdoo Yun, Seungryong Kim, Dongyoon Han; Proceedings of the Asian Conference on Computer Vision (ACCV), 2024, pp. 3154-3171

Abstract


Semantic correspondence methods have advanced to obtaining high-quality correspondences employing complicated networks, aiming to maximize the model capacity. However, despite the performance improvements, they may remain constrained by the scarcity of training keypoint pairs, a consequence of the limited training images and the sparsity of keypoints. This paper builds on the hypothesis that there is an inherent data-hungry matter in learning semantic correspondences and uncovers the models can be more trained by employing densified training pairs. We demonstrate a simple machine annotator reliably enriches paired key points via machine supervision, requiring neither extra labeled key points nor trainable modules from unlabeled images. Consequently, our models surpass current state-of-the-art models on semantic correspondence learning benchmarks like SPair-71k, PF-PASCAL, and PF-WILLOW and enjoy further robustness on corruption benchmarks.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Kim_2024_ACCV, author = {Kim, Jiwon and Heo, Byeongho and Yun, Sangdoo and Kim, Seungryong and Han, Dongyoon}, title = {Match me if you can: Semi-Supervised Semantic Correspondence Learning with Unpaired Images}, booktitle = {Proceedings of the Asian Conference on Computer Vision (ACCV)}, month = {December}, year = {2024}, pages = {3154-3171} }