Generalizing Gaze Estimation With Outlier-Guided Collaborative Adaptation

Yunfei Liu, Ruicong Liu, Haofei Wang, Feng Lu; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021, pp. 3835-3844

Abstract


Deep neural networks have significantly improved appearance-based gaze estimation accuracy. However, it still suffers from unsatisfactory performance when generalizing the trained model to new domains, e.g., unseen environments or persons. In this paper, we propose a plug-and-play gaze adaptation framework (PnP-GA), which is an ensemble of networks that learn collaboratively with the guidance of outliers. Since our proposed framework does not require ground-truth labels in the target domain, the existing gaze estimation networks can be directly plugged into PnP-GA and generalize the algorithms to new domains. We test PnP-GA on four gaze domain adaptation tasks, ETH-to-MPII, ETH-to-EyeDiap, Gaze360-to-MPII, and Gaze360-to-EyeDiap. The experimental results demonstrate that the PnP-GA framework achieves considerable performance improvements of 36.9%, 31.6%, 19.4%, and 11.8% over the baseline system. The proposed framework also outperforms the state-of-the-art domain adaptation approaches on gaze domain adaptation tasks.

Related Material


[pdf] [arXiv]
[bibtex]
@InProceedings{Liu_2021_ICCV, author = {Liu, Yunfei and Liu, Ruicong and Wang, Haofei and Lu, Feng}, title = {Generalizing Gaze Estimation With Outlier-Guided Collaborative Adaptation}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2021}, pages = {3835-3844} }