Towards Comprehensive Testing on the Robustness of Cooperative Multi-Agent Reinforcement Learning

Jun Guo, Yonghong Chen, Yihang Hao, Zixin Yin, Yin Yu, Simin Li; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2022, pp. 115-122

Abstract


While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL algorithm before it was deployed in reality. Existing adversarial attacks for MARL could be used for testing, but is limited to one robustness aspects (e.g., reward, state, action), while c-MARL model could be attacked from any aspect. To overcome the challenge, we propose MARLSafe, the first robustness testing framework for c-MARL algorithms. First, motivated by Markov Decision Process (MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively from three aspects, namely state robustness, action robustness and reward robustness. Any c-MARL algorithm must simultaneously satisfy these robustness aspects to be considered secure. Second, due to the scarceness of c-MARL attack, we propose several c-MARL attack from multi-aspect as testing algorithms for c-MARL robustness. Experiments on SMAC environment reveals that all state-of-the-art c-MARL algorithm are of low robustness in all aspect.

Related Material


[pdf] [arXiv]
[bibtex]
@InProceedings{Guo_2022_CVPR, author = {Guo, Jun and Chen, Yonghong and Hao, Yihang and Yin, Zixin and Yu, Yin and Li, Simin}, title = {Towards Comprehensive Testing on the Robustness of Cooperative Multi-Agent Reinforcement Learning}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2022}, pages = {115-122} }