2022
DOI: 10.48550/arxiv.2206.10158
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Certifiably Robust Policy Learning against Adversarial Communication in Multi-agent Systems

Abstract: Communication is important in many multi-agent reinforcement learning (MARL) problems for agents to share information and make good decisions. However, when deploying trained communicative agents in a real-world application where noise and potential attackers exist, the safety of communication-based policies becomes a severe issue that is underexplored. Specifically, if communication messages are manipulated by malicious attackers, agents relying on untrustworthy communication may take unsafe actions that lead… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 19 publications
0
2
0
Order By: Relevance
“…These defenses mainly focus on the adversarial perturbations directly applied to the agent's inputs. [61] propose a certifiable defense against adversarial communication in MARL systems. To the best of our knowledge, our paper is the first to provide provable convergence guarantees for adversarial training against adversarial attacks on behaviors of other agents in the environment.…”
Section: Attacks and Defenses On Communication Inmentioning
confidence: 99%
“…These defenses mainly focus on the adversarial perturbations directly applied to the agent's inputs. [61] propose a certifiable defense against adversarial communication in MARL systems. To the best of our knowledge, our paper is the first to provide provable convergence guarantees for adversarial training against adversarial attacks on behaviors of other agents in the environment.…”
Section: Attacks and Defenses On Communication Inmentioning
confidence: 99%
“…RL has also been shown to be susceptible to perturbation in the observations of an RL agent (Huang et al 2017;Behzadan and Munir 2017) or in environments (Gleave et al 2019). Some adversarial defence works for RL are proposed (Donti et al 2020;Eysenbach and Levine 2021;Shen et al 2020;Sun et al 2022) and then towards these defences, stronger attacks are proposed (Salman et al 2019;Russo and Proutiere 2019). To end this repeated game, Wu et al (2021) and Kumar, Levine, and Feizi (2021) proposed to use probabilistic approaches to provide robustness certification for RLs.…”
Section: Introductionmentioning
confidence: 99%