Group Sampling for Scale Invariant Face Detection

Xiang Ming, Fangyun Wei, Ting Zhang, Dong Chen, Fang Wen; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 3446-3456

Abstract


Detectors based on deep learning tend to detect multi-scale faces on a single input image for efficiency. Recent works, such as FPN and SSD, generally use feature maps from multiple layers with different spatial resolutions to detect objects at different scales, e.g., high-resolution feature maps for small objects. However, we find that such multi-layer prediction is not necessary. Faces at all scales can be well detected with features from a single layer of the network. In this paper, we carefully examine the factors affecting face detection across a large range of scales, and conclude that the balance of training samples, including both positive and negative ones, at different scales is the key. We propose a group sampling method which divides the anchors into several groups according to the scale, and ensure that the number of samples for each group is the same during training. Our approach using only the last layer of FPN as features is able to advance the state-of-the-arts. Comprehensive analysis and extensive experiments have been conducted to show the effectiveness of the proposed method. Our approach, evaluated on face detection benchmarks including FDDB and WIDER FACE datasets, achieves state-of-the-art results without bells and whistles.

Related Material


[pdf] [supp]
[bibtex]
@InProceedings{Ming_2019_CVPR,
author = {Ming, Xiang and Wei, Fangyun and Zhang, Ting and Chen, Dong and Wen, Fang},
title = {Group Sampling for Scale Invariant Face Detection},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2019}
}