DCTM: Discrete-Continuous Transformation Matching for Semantic Flow
Seungryong Kim, Dongbo Min, Stephen Lin, Kwanghoon Sohn; Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 4529-4538
Abstract
Techniques for dense semantic correspondence have provided limited ability to deal with the geometric variations that commonly exist between semantically similar images. While variations due to scale and rotation have been examined, there is a lack of practical solutions for more complex deformations such as affine transformations because of the tremendous size of the associated solution space. To address this problem, we present a discrete-continuous transformation matching (DCTM) framework where dense affine transformation fields are inferred through a discrete label optimization in which the labels are iteratively updated via continuous regularization. In this way, our approach draws solutions from the continuous space of affine transformations in a manner that can be computed efficiently through constant-time edge-aware filtering and a proposed affine-varying CNN-based descriptor. Experimental results show that this model outperforms the state-of-the-art methods for dense semantic correspondence on various benchmarks.
Related Material
[pdf]
[arXiv]
[video]
[
bibtex]
@InProceedings{Kim_2017_ICCV,
author = {Kim, Seungryong and Min, Dongbo and Lin, Stephen and Sohn, Kwanghoon},
title = {DCTM: Discrete-Continuous Transformation Matching for Semantic Flow},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV)},
month = {Oct},
year = {2017}
}