Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning

While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and u...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Guo, Jun, Chen, Yonghong, Hao, Yihang, Yin, Zixin, Yu, Yin, Li, Simin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Guo, Jun
Chen, Yonghong
Hao, Yihang
Yin, Zixin
Yu, Yin
Li, Simin
description While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL algorithm before it was deployed in reality. Existing adversarial attacks for MARL could be used for testing, but is limited to one robustness aspects (e.g., reward, state, action), while c-MARL model could be attacked from any aspect. To overcome the challenge, we propose MARLSafe, the first robustness testing framework for c-MARL algorithms. First, motivated by Markov Decision Process (MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively from three aspects, namely state robustness, action robustness and reward robustness. Any c-MARL algorithm must simultaneously satisfy these robustness aspects to be considered secure. Second, due to the scarceness of c-MARL attack, we propose c-MARL attacks as robustness testing algorithms from multiple aspects. Experiments on \textit{SMAC} environment reveals that many state-of-the-art c-MARL algorithms are of low robustness in all aspect, pointing out the urgent need to test and enhance robustness of c-MARL algorithms.
doi_str_mv 10.48550/arxiv.2204.07932
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2204_07932</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2204_07932</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-a9cb2be6cf015bccf5747ddab27a9c0332efe8043f9301ae8a2626db5e4570de3</originalsourceid><addsrcrecordid>eNotj8tqwzAURLXpoqT9gK6qH7ArS5ZlL4vpC1ICwXtzZV8lglgykpK2f1877WoY5jBwCHkoWF7WUrInCN_2knPOypypRvBbAp3_gjBG2vppDnhEF-0FaYcxWXeg3tF0RLr3-hyTwxipNwvqZwyQVvDzfEo2gwO6RPdonfFhwGltW4Tglo87cmPgFPH-Pzeke33p2vdsu3v7aJ-3GVSKZ9AMmmusBsMKqYfBSFWqcQTN1TIxITgarFkpTCNYAVgDr3g1aomlVGxEsSGPf7dXx34OdoLw06-u_dVV_AIy-1GR</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</title><source>arXiv.org</source><creator>Guo, Jun ; Chen, Yonghong ; Hao, Yihang ; Yin, Zixin ; Yu, Yin ; Li, Simin</creator><creatorcontrib>Guo, Jun ; Chen, Yonghong ; Hao, Yihang ; Yin, Zixin ; Yu, Yin ; Li, Simin</creatorcontrib><description>While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL algorithm before it was deployed in reality. Existing adversarial attacks for MARL could be used for testing, but is limited to one robustness aspects (e.g., reward, state, action), while c-MARL model could be attacked from any aspect. To overcome the challenge, we propose MARLSafe, the first robustness testing framework for c-MARL algorithms. First, motivated by Markov Decision Process (MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively from three aspects, namely state robustness, action robustness and reward robustness. Any c-MARL algorithm must simultaneously satisfy these robustness aspects to be considered secure. Second, due to the scarceness of c-MARL attack, we propose c-MARL attacks as robustness testing algorithms from multiple aspects. Experiments on \textit{SMAC} environment reveals that many state-of-the-art c-MARL algorithms are of low robustness in all aspect, pointing out the urgent need to test and enhance robustness of c-MARL algorithms.</description><identifier>DOI: 10.48550/arxiv.2204.07932</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Computer Science - Multiagent Systems</subject><creationdate>2022-04</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2204.07932$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2204.07932$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Guo, Jun</creatorcontrib><creatorcontrib>Chen, Yonghong</creatorcontrib><creatorcontrib>Hao, Yihang</creatorcontrib><creatorcontrib>Yin, Zixin</creatorcontrib><creatorcontrib>Yu, Yin</creatorcontrib><creatorcontrib>Li, Simin</creatorcontrib><title>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</title><description>While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL algorithm before it was deployed in reality. Existing adversarial attacks for MARL could be used for testing, but is limited to one robustness aspects (e.g., reward, state, action), while c-MARL model could be attacked from any aspect. To overcome the challenge, we propose MARLSafe, the first robustness testing framework for c-MARL algorithms. First, motivated by Markov Decision Process (MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively from three aspects, namely state robustness, action robustness and reward robustness. Any c-MARL algorithm must simultaneously satisfy these robustness aspects to be considered secure. Second, due to the scarceness of c-MARL attack, we propose c-MARL attacks as robustness testing algorithms from multiple aspects. Experiments on \textit{SMAC} environment reveals that many state-of-the-art c-MARL algorithms are of low robustness in all aspect, pointing out the urgent need to test and enhance robustness of c-MARL algorithms.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Multiagent Systems</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8tqwzAURLXpoqT9gK6qH7ArS5ZlL4vpC1ICwXtzZV8lglgykpK2f1877WoY5jBwCHkoWF7WUrInCN_2knPOypypRvBbAp3_gjBG2vppDnhEF-0FaYcxWXeg3tF0RLr3-hyTwxipNwvqZwyQVvDzfEo2gwO6RPdonfFhwGltW4Tglo87cmPgFPH-Pzeke33p2vdsu3v7aJ-3GVSKZ9AMmmusBsMKqYfBSFWqcQTN1TIxITgarFkpTCNYAVgDr3g1aomlVGxEsSGPf7dXx34OdoLw06-u_dVV_AIy-1GR</recordid><startdate>20220417</startdate><enddate>20220417</enddate><creator>Guo, Jun</creator><creator>Chen, Yonghong</creator><creator>Hao, Yihang</creator><creator>Yin, Zixin</creator><creator>Yu, Yin</creator><creator>Li, Simin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220417</creationdate><title>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</title><author>Guo, Jun ; Chen, Yonghong ; Hao, Yihang ; Yin, Zixin ; Yu, Yin ; Li, Simin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-a9cb2be6cf015bccf5747ddab27a9c0332efe8043f9301ae8a2626db5e4570de3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Multiagent Systems</topic><toplevel>online_resources</toplevel><creatorcontrib>Guo, Jun</creatorcontrib><creatorcontrib>Chen, Yonghong</creatorcontrib><creatorcontrib>Hao, Yihang</creatorcontrib><creatorcontrib>Yin, Zixin</creatorcontrib><creatorcontrib>Yu, Yin</creatorcontrib><creatorcontrib>Li, Simin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Guo, Jun</au><au>Chen, Yonghong</au><au>Hao, Yihang</au><au>Yin, Zixin</au><au>Yu, Yin</au><au>Li, Simin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</atitle><date>2022-04-17</date><risdate>2022</risdate><abstract>While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL algorithm before it was deployed in reality. Existing adversarial attacks for MARL could be used for testing, but is limited to one robustness aspects (e.g., reward, state, action), while c-MARL model could be attacked from any aspect. To overcome the challenge, we propose MARLSafe, the first robustness testing framework for c-MARL algorithms. First, motivated by Markov Decision Process (MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively from three aspects, namely state robustness, action robustness and reward robustness. Any c-MARL algorithm must simultaneously satisfy these robustness aspects to be considered secure. Second, due to the scarceness of c-MARL attack, we propose c-MARL attacks as robustness testing algorithms from multiple aspects. Experiments on \textit{SMAC} environment reveals that many state-of-the-art c-MARL algorithms are of low robustness in all aspect, pointing out the urgent need to test and enhance robustness of c-MARL algorithms.</abstract><doi>10.48550/arxiv.2204.07932</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2204.07932
ispartof
issn
language eng
recordid cdi_arxiv_primary_2204_07932
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
Computer Science - Multiagent Systems
title Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T22%3A26%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Towards%20Comprehensive%20Testing%20on%20the%20Robustness%20of%20Cooperative%20Multi-agent%20Reinforcement%20Learning&rft.au=Guo,%20Jun&rft.date=2022-04-17&rft_id=info:doi/10.48550/arxiv.2204.07932&rft_dat=%3Carxiv_GOX%3E2204_07932%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true