ResSaNet: A Hybrid Backbone of Residual Block and Self-Attention Module for Masked Face Recognition

Wei-Yi Chang, Ming-Ying Tsai, Shih-Chieh Lo; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, 2021, pp. 1468-1476

Abstract


In recent years, the performances of face recognition have been improved significantly by using convolution neural networks (CNN) as the feature extractors. On the other hands, in order to avoid spreading COVID-19 virus, people would wear mask even when they want to pass the face recognition system. Thus, it is necessary to improve the performance of masked face recognition so that users could utilize face recognition methods more easily. In this paper, we propose a feature extraction backbone named ResSaNet that integrates CNN (especially Residual block) and Self-attention module into the same network. By capturing the local and global information of face area simultaneously, our proposed ResSaNet could achieve promising results on both masked and non-masked testing data.

Related Material


[pdf]
[bibtex]
@InProceedings{Chang_2021_ICCV, author = {Chang, Wei-Yi and Tsai, Ming-Ying and Lo, Shih-Chieh}, title = {ResSaNet: A Hybrid Backbone of Residual Block and Self-Attention Module for Masked Face Recognition}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops}, month = {October}, year = {2021}, pages = {1468-1476} }