Deep Saliency With Encoded Low Level Distance Map and High Level Features

Gayoung Lee, Yu-Wing Tai, Junmo Kim; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 660-668

Abstract


Recent advances in saliency detection have utilized deep learning to obtain high level features to detect salient regions in a scene. They have demonstrated superior results over previous works that utilize hand-crafted low level features for saliency detection. In this paper, we demonstrate that the hand-crafted features can provide complementary effects to enhance performance of saliency detection that utilizes only high level features. Our method utilizes both high level and low level features for saliency detection under a unified deep learning framework. The high level features are extracted using the VGG-net, and the low level features are compared with other parts of an image to form a low level distance map. The low level distance map is then encoded using a CNN with multiple 1*1 convolutional and ReLU layers. We concatenate the encoded low level distance map and the high level features, and connect them to a fully connected neural network classifier to evaluate the saliency of a query region. Our experiments show that our method can further improve performance of the state-of-the-art deep learning based saliency detection methods.

Related Material


[pdf]
[bibtex]
@InProceedings{Lee_2016_CVPR,
author = {Lee, Gayoung and Tai, Yu-Wing and Kim, Junmo},
title = {Deep Saliency With Encoded Low Level Distance Map and High Level Features},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2016}
}