Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning
While deep neural networks (DNNs) have strengthened the performance of cooperative multi-agent reinforcement learning (c-MARL), the agent policy can be easily perturbed by adversarial examples. Considering the safety critical applications of c-MARL, such as traffic management, power management and u...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Guo, Jun Chen, Yonghong Hao, Yihang Yin, Zixin Yu, Yin Li, Simin |
description | While deep neural networks (DNNs) have strengthened the performance of
cooperative multi-agent reinforcement learning (c-MARL), the agent policy can
be easily perturbed by adversarial examples. Considering the safety critical
applications of c-MARL, such as traffic management, power management and
unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL
algorithm before it was deployed in reality. Existing adversarial attacks for
MARL could be used for testing, but is limited to one robustness aspects (e.g.,
reward, state, action), while c-MARL model could be attacked from any aspect.
To overcome the challenge, we propose MARLSafe, the first robustness testing
framework for c-MARL algorithms. First, motivated by Markov Decision Process
(MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively
from three aspects, namely state robustness, action robustness and reward
robustness. Any c-MARL algorithm must simultaneously satisfy these robustness
aspects to be considered secure. Second, due to the scarceness of c-MARL
attack, we propose c-MARL attacks as robustness testing algorithms from
multiple aspects. Experiments on \textit{SMAC} environment reveals that many
state-of-the-art c-MARL algorithms are of low robustness in all aspect,
pointing out the urgent need to test and enhance robustness of c-MARL
algorithms. |
doi_str_mv | 10.48550/arxiv.2204.07932 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2204_07932</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2204_07932</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-a9cb2be6cf015bccf5747ddab27a9c0332efe8043f9301ae8a2626db5e4570de3</originalsourceid><addsrcrecordid>eNotj8tqwzAURLXpoqT9gK6qH7ArS5ZlL4vpC1ICwXtzZV8lglgykpK2f1877WoY5jBwCHkoWF7WUrInCN_2knPOypypRvBbAp3_gjBG2vppDnhEF-0FaYcxWXeg3tF0RLr3-hyTwxipNwvqZwyQVvDzfEo2gwO6RPdonfFhwGltW4Tglo87cmPgFPH-Pzeke33p2vdsu3v7aJ-3GVSKZ9AMmmusBsMKqYfBSFWqcQTN1TIxITgarFkpTCNYAVgDr3g1aomlVGxEsSGPf7dXx34OdoLw06-u_dVV_AIy-1GR</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</title><source>arXiv.org</source><creator>Guo, Jun ; Chen, Yonghong ; Hao, Yihang ; Yin, Zixin ; Yu, Yin ; Li, Simin</creator><creatorcontrib>Guo, Jun ; Chen, Yonghong ; Hao, Yihang ; Yin, Zixin ; Yu, Yin ; Li, Simin</creatorcontrib><description>While deep neural networks (DNNs) have strengthened the performance of
cooperative multi-agent reinforcement learning (c-MARL), the agent policy can
be easily perturbed by adversarial examples. Considering the safety critical
applications of c-MARL, such as traffic management, power management and
unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL
algorithm before it was deployed in reality. Existing adversarial attacks for
MARL could be used for testing, but is limited to one robustness aspects (e.g.,
reward, state, action), while c-MARL model could be attacked from any aspect.
To overcome the challenge, we propose MARLSafe, the first robustness testing
framework for c-MARL algorithms. First, motivated by Markov Decision Process
(MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively
from three aspects, namely state robustness, action robustness and reward
robustness. Any c-MARL algorithm must simultaneously satisfy these robustness
aspects to be considered secure. Second, due to the scarceness of c-MARL
attack, we propose c-MARL attacks as robustness testing algorithms from
multiple aspects. Experiments on \textit{SMAC} environment reveals that many
state-of-the-art c-MARL algorithms are of low robustness in all aspect,
pointing out the urgent need to test and enhance robustness of c-MARL
algorithms.</description><identifier>DOI: 10.48550/arxiv.2204.07932</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Computer Science - Multiagent Systems</subject><creationdate>2022-04</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2204.07932$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2204.07932$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Guo, Jun</creatorcontrib><creatorcontrib>Chen, Yonghong</creatorcontrib><creatorcontrib>Hao, Yihang</creatorcontrib><creatorcontrib>Yin, Zixin</creatorcontrib><creatorcontrib>Yu, Yin</creatorcontrib><creatorcontrib>Li, Simin</creatorcontrib><title>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</title><description>While deep neural networks (DNNs) have strengthened the performance of
cooperative multi-agent reinforcement learning (c-MARL), the agent policy can
be easily perturbed by adversarial examples. Considering the safety critical
applications of c-MARL, such as traffic management, power management and
unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL
algorithm before it was deployed in reality. Existing adversarial attacks for
MARL could be used for testing, but is limited to one robustness aspects (e.g.,
reward, state, action), while c-MARL model could be attacked from any aspect.
To overcome the challenge, we propose MARLSafe, the first robustness testing
framework for c-MARL algorithms. First, motivated by Markov Decision Process
(MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively
from three aspects, namely state robustness, action robustness and reward
robustness. Any c-MARL algorithm must simultaneously satisfy these robustness
aspects to be considered secure. Second, due to the scarceness of c-MARL
attack, we propose c-MARL attacks as robustness testing algorithms from
multiple aspects. Experiments on \textit{SMAC} environment reveals that many
state-of-the-art c-MARL algorithms are of low robustness in all aspect,
pointing out the urgent need to test and enhance robustness of c-MARL
algorithms.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Multiagent Systems</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8tqwzAURLXpoqT9gK6qH7ArS5ZlL4vpC1ICwXtzZV8lglgykpK2f1877WoY5jBwCHkoWF7WUrInCN_2knPOypypRvBbAp3_gjBG2vppDnhEF-0FaYcxWXeg3tF0RLr3-hyTwxipNwvqZwyQVvDzfEo2gwO6RPdonfFhwGltW4Tglo87cmPgFPH-Pzeke33p2vdsu3v7aJ-3GVSKZ9AMmmusBsMKqYfBSFWqcQTN1TIxITgarFkpTCNYAVgDr3g1aomlVGxEsSGPf7dXx34OdoLw06-u_dVV_AIy-1GR</recordid><startdate>20220417</startdate><enddate>20220417</enddate><creator>Guo, Jun</creator><creator>Chen, Yonghong</creator><creator>Hao, Yihang</creator><creator>Yin, Zixin</creator><creator>Yu, Yin</creator><creator>Li, Simin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220417</creationdate><title>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</title><author>Guo, Jun ; Chen, Yonghong ; Hao, Yihang ; Yin, Zixin ; Yu, Yin ; Li, Simin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-a9cb2be6cf015bccf5747ddab27a9c0332efe8043f9301ae8a2626db5e4570de3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Multiagent Systems</topic><toplevel>online_resources</toplevel><creatorcontrib>Guo, Jun</creatorcontrib><creatorcontrib>Chen, Yonghong</creatorcontrib><creatorcontrib>Hao, Yihang</creatorcontrib><creatorcontrib>Yin, Zixin</creatorcontrib><creatorcontrib>Yu, Yin</creatorcontrib><creatorcontrib>Li, Simin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Guo, Jun</au><au>Chen, Yonghong</au><au>Hao, Yihang</au><au>Yin, Zixin</au><au>Yu, Yin</au><au>Li, Simin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning</atitle><date>2022-04-17</date><risdate>2022</risdate><abstract>While deep neural networks (DNNs) have strengthened the performance of
cooperative multi-agent reinforcement learning (c-MARL), the agent policy can
be easily perturbed by adversarial examples. Considering the safety critical
applications of c-MARL, such as traffic management, power management and
unmanned aerial vehicle control, it is crucial to test the robustness of c-MARL
algorithm before it was deployed in reality. Existing adversarial attacks for
MARL could be used for testing, but is limited to one robustness aspects (e.g.,
reward, state, action), while c-MARL model could be attacked from any aspect.
To overcome the challenge, we propose MARLSafe, the first robustness testing
framework for c-MARL algorithms. First, motivated by Markov Decision Process
(MDP), MARLSafe consider the robustness of c-MARL algorithms comprehensively
from three aspects, namely state robustness, action robustness and reward
robustness. Any c-MARL algorithm must simultaneously satisfy these robustness
aspects to be considered secure. Second, due to the scarceness of c-MARL
attack, we propose c-MARL attacks as robustness testing algorithms from
multiple aspects. Experiments on \textit{SMAC} environment reveals that many
state-of-the-art c-MARL algorithms are of low robustness in all aspect,
pointing out the urgent need to test and enhance robustness of c-MARL
algorithms.</abstract><doi>10.48550/arxiv.2204.07932</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2204.07932 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2204_07932 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning Computer Science - Multiagent Systems |
title | Towards Comprehensive Testing on the Robustness of Cooperative Multi-agent Reinforcement Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T22%3A26%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Towards%20Comprehensive%20Testing%20on%20the%20Robustness%20of%20Cooperative%20Multi-agent%20Reinforcement%20Learning&rft.au=Guo,%20Jun&rft.date=2022-04-17&rft_id=info:doi/10.48550/arxiv.2204.07932&rft_dat=%3Carxiv_GOX%3E2204_07932%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |