Co-Occurrent Features in Semantic Segmentation

Hang Zhang, Han Zhang, Chenguang Wang, Junyuan Xie; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 548-557

Abstract


Recent work has achieved great success in utilizing global contextual information for semantic segmentation, including increasing the receptive field and aggregating pyramid feature representations. In this paper, we go beyond global context and explore the fine-grained representation using co-occurrent features by introducing Co-occurrent Feature Model, which predicts the distribution of co-occurrent features for a given target. To leverage the semantic context in the co-occurrent features, we build an Aggregated Co-occurrent Feature (ACF) Module by aggregating the probability of the co-occurrent feature with the co-occurrent context. ACF Module learns a fine-grained spatial invariant representation to capture co-occurrent context information across the scene. Our approach significantly improves the segmentation results using FCN and achieves superior performance 54.0% mIoU on Pascal Context, 87.2% mIoU on Pascal VOC 2012 and 44.89% mIoU on ADE20K datasets. The source code and complete system will be publicly available upon publication.

Related Material


[pdf]
[bibtex]
@InProceedings{Zhang_2019_CVPR,
author = {Zhang, Hang and Zhang, Han and Wang, Chenguang and Xie, Junyuan},
title = {Co-Occurrent Features in Semantic Segmentation},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2019}
}