-
[pdf]
[supp]
[arXiv]
[bibtex]@InProceedings{Kim_2024_ACCV, author = {Kim, Jiwon and Heo, Byeongho and Yun, Sangdoo and Kim, Seungryong and Han, Dongyoon}, title = {Match me if you can: Semi-Supervised Semantic Correspondence Learning with Unpaired Images}, booktitle = {Proceedings of the Asian Conference on Computer Vision (ACCV)}, month = {December}, year = {2024}, pages = {3154-3171} }
Match me if you can: Semi-Supervised Semantic Correspondence Learning with Unpaired Images
Abstract
Semantic correspondence methods have advanced to obtaining high-quality correspondences employing complicated networks, aiming to maximize the model capacity. However, despite the performance improvements, they may remain constrained by the scarcity of training keypoint pairs, a consequence of the limited training images and the sparsity of keypoints. This paper builds on the hypothesis that there is an inherent data-hungry matter in learning semantic correspondences and uncovers the models can be more trained by employing densified training pairs. We demonstrate a simple machine annotator reliably enriches paired key points via machine supervision, requiring neither extra labeled key points nor trainable modules from unlabeled images. Consequently, our models surpass current state-of-the-art models on semantic correspondence learning benchmarks like SPair-71k, PF-PASCAL, and PF-WILLOW and enjoy further robustness on corruption benchmarks.
Related Material