Multiple Transfer Learning and Multi-Label Balanced Training Strategies for Facial AU Detection in the Wild

Sijie Ji, Kai Wang, Xiaojiang Peng, Jianfei Yang, Zhaoyang Zeng, Yu Qiao; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2020, pp. 414-415

Abstract


This paper presents SIAT-NTU solution and results of facial action unit (AU) detection in the EmotiNet Challenge 2020. The task aims to detect 23 AUs from facial images in the wild, and its main difficulties lie in the imbalanced AU distribution and discriminative feature learning. We tackle these difficulties from the following aspects. First, to address the unconstrained heterogeneity of in-the-wild images, we detect and align faces with multi-task convolutional neural networks (MTCNN). Second, by using multiple transfer strategies, we pre-train large CNNs on multiple related datasets, e.g. face recognition datasets and facial expression datasets, and fine-tune them on the EmotiNetdataset. Third, we employ a multi-label balanced sampling strategy and a weighted loss to mitigate the imbalance problem. Last but not the least, to further improve performance, we ensemble multiple models and optimize the thresholds for each AU. Our proposed solution achieves an accuracy of 90.13% and F1 of 44.10% in the final test phase. Our Code is available at:https://github.com/kaiwang960112/ENC2020_AU_Detection

Related Material


[pdf] [video]
[bibtex]
@InProceedings{Ji_2020_CVPR_Workshops,
author = {Ji, Sijie and Wang, Kai and Peng, Xiaojiang and Yang, Jianfei and Zeng, Zhaoyang and Qiao, Yu},
title = {Multiple Transfer Learning and Multi-Label Balanced Training Strategies for Facial AU Detection in the Wild},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}