Loss Max-Pooling for Semantic Image Segmentation

Samuel Rota Bulo, Gerhard Neuhold, Peter Kontschieder; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 2126-2135

Abstract


In this work we introduce a novel loss max-pooling concept for handling imbalanced training data distributions, applicable as alternative loss layer in the context of deep neural networks for semantic image segmentation tasks. Most real-world semantic segmentation datasets exhibit long tail distributions with few object categories comprising the majority of data and consequently biasing the classifiers towards them. Our method adaptively re-weights the contributions of each pixel based on their observed losses, targeting under-performing classification results as often encountered for under-represented object classes. Moreover, our approach goes beyond conventional cost-sensitive learning attempts through adaptive considerations that allow us to indirectly address both, inter- and intra-class imbalances. We provide a theoretical justification of our approach, complementary to experimental analyses on standard semantic segmentation datasets. In our experiments on the challenging Cityscapes and Pascal VOC 2012 segmentation benchmarks we find consistently improved results, demonstrating the efficacy of our approach.

Related Material


[pdf] [arXiv]
[bibtex]
@InProceedings{Bulo_2017_CVPR,
author = {Rota Bulo, Samuel and Neuhold, Gerhard and Kontschieder, Peter},
title = {Loss Max-Pooling for Semantic Image Segmentation},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {July},
year = {2017}
}