How to Choose your Best Allies for a Transferable Attack?

Thibault Maho, Seyed-Mohsen Moosavi-Dezfooli, Teddy Furon; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 4542-4551

Abstract


The transferability of adversarial examples is a key issue in the security of deep neural networks. The possibility of an adversarial example crafted for a source model fooling another targeted model makes the threat of adversarial attacks more realistic. Measuring transferability is a crucial problem, but the Attack Success Rate alone does not provide a sound evaluation. This paper proposes a new methodology for evaluating transferability by putting distortion in a central position. This new tool shows that transferable attacks may perform far worse than a black box attack if the attacker randomly picks the source model. To address this issue, we propose a new selection mechanism, called FiT, which aims at choosing the best source model with only a few preliminary queries to the target. Our experimental results show that FiT is highly effective at selecting the best source model for multiple scenarios such as single-model attacks, ensemble-model attacks and multiple attacks.

Related Material


[pdf] [supp] [arXiv]
[bibtex]
@InProceedings{Maho_2023_ICCV, author = {Maho, Thibault and Moosavi-Dezfooli, Seyed-Mohsen and Furon, Teddy}, title = {How to Choose your Best Allies for a Transferable Attack?}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)}, month = {October}, year = {2023}, pages = {4542-4551} }