Occlusion Boundary Detection via Deep Exploration of Context

Huan Fu, Chaohui Wang, Dacheng Tao, Michael J. Black; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 241-250

Abstract


Occlusion boundaries contain rich perceptual information about the underlying scene structure. They also provide important cues in many visual perception tasks such as scene understanding, object recognition, and segmentation. In this paper, we improve occlusion boundary detection via enhanced exploration of contextual information (e.g., local structural boundary patterns, observations from surrounding regions, and temporal context), and in doing so develop a novel approach based on convolutional neural networks (CNNs) and conditional random fields (CRFs). Experimental results demonstrate that our detector significantly outperforms the state-of-the-art (e.g., improving the F-measure from 0.62 to 0.71 on the commonly used CMU benchmark). Last but not least, we empirically assess the roles of several important components of the proposed detector, so as to validate the rationale behind this approach.

Related Material


[pdf]
[bibtex]
@InProceedings{Fu_2016_CVPR,
author = {Fu, Huan and Wang, Chaohui and Tao, Dacheng and Black, Michael J.},
title = {Occlusion Boundary Detection via Deep Exploration of Context},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2016}
}