Reactive Model Correction: Mitigating Harm to Task-Relevant Features via Conditional Bias Suppression

Dilyara Bareeva, Maximilian Dreyer, Frederik Pahde, Wojciech Samek, Sebastian Lapuschkin; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2024, pp. 3532-3541

Abstract


Deep Neural Networks are prone to learning and relying on spurious correlations in the training data which for high-risk applications can have fatal consequences. Various approaches to suppress model reliance on harmful features have been proposed that can be applied post-hoc without additional training. Whereas those methods can be applied with efficiency they also tend to harm model performance by globally shifting the distribution of latent features. To mitigate unintended overcorrection of model behavior we propose a reactive approach conditioned on model-derived knowledge and eXplainable Artificial Intelligence (XAI) insights. While the reactive approach can be applied to many post-hoc methods we demonstrate the incorporation of reactivity in particular for P-ClArC (Projective Class Artifact Compensation) introducing a new method called R-ClArC (Reactive Class Artifact Compensation). Through rigorous experiments in controlled settings (FunnyBirds) and with a real-world dataset (ISIC2019) we show that introducing reactivity can minimize the detrimental effect of the applied correction while simultaneously ensuring low reliance on spurious features.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Bareeva_2024_CVPR, author = {Bareeva, Dilyara and Dreyer, Maximilian and Pahde, Frederik and Samek, Wojciech and Lapuschkin, Sebastian}, title = {Reactive Model Correction: Mitigating Harm to Task-Relevant Features via Conditional Bias Suppression}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2024}, pages = {3532-3541} }