Pre-Trained Multiple Latent Variable Generative Models are Good Defenders Against Adversarial Attacks

Dario Serez, Marco Cristani, Alessio Del Bue, Vittorio Murino, Pietro Morerio; Proceedings of the Winter Conference on Applications of Computer Vision (WACV), 2025, pp. 6506-6516

Abstract


Attackers can deliberately perturb classifiers' input with subtle noise altering final predictions. Among proposed countermeasures adversarial purification employs generative networks to preprocess input images filtering out adversarial noise. In this study we propose specific generators defined Multiple Latent Variable Generative Models (MLVGMs) for adversarial purification. These models possess multiple latent variables that naturally disentangle coarse from fine features. Taking advantage of these properties we autoencode images to maintain class-relevant information while discarding and re-sampling any detail including adversarial noise. The procedure is completely training-free exploring the generalization abilities of pre-trained MLVGMs on the adversarial purification downstream task. Despite the lack of large models trained on billions of samples we show that smaller MLVGMs are already competitive with traditional methods and can be used as foundation models. Official code released at https://github.com/SerezD/gen_adversarial.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Serez_2025_WACV, author = {Serez, Dario and Cristani, Marco and Del Bue, Alessio and Murino, Vittorio and Morerio, Pietro}, title = {Pre-Trained Multiple Latent Variable Generative Models are Good Defenders Against Adversarial Attacks}, booktitle = {Proceedings of the Winter Conference on Applications of Computer Vision (WACV)}, month = {February}, year = {2025}, pages = {6506-6516} }