Assistive Signals for Deep Neural Network Classifiers

Camilo Pestana, Wei Liu, David Glance, Robyn Owens, Ajmal Mian; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2021, pp. 1221-1225


Deep Neural Networks are brittle in that small changes in the input can drastically affect their prediction outcome and confidence. Consequently, research in this area mainly focus on adversarial attacks and defenses. In this paper, we take an alternative stance and introduce the concept of Assistive Signals, which are perturbations optimized to improve a model's confidence score regardless if it's under attack or not. We analyze some interesting properties of these assistive perturbations and extend the idea to optimize them in the 3D space simulating different lighting conditions and viewing angles. Experimental evaluations show that the assistive signals generated by our optimization method increase the accuracy and confidence of deep models more than those generated by conventional methods that work in the 2D space. `Assistive Signals' also illustrate bias of ML models towards certain patterns in real-life objects.

Related Material

[pdf] [arXiv]
@InProceedings{Pestana_2021_CVPR, author = {Pestana, Camilo and Liu, Wei and Glance, David and Owens, Robyn and Mian, Ajmal}, title = {Assistive Signals for Deep Neural Network Classifiers}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2021}, pages = {1221-1225} }