Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks

As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues f...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE internet of things journal 2023-02, Vol.10 (3), p.2667-2679
Hauptverfasser: Zhang, Hengxi, Lu, Chengyue, Tang, Huaze, Wei, Xiaoli, Liang, Le, Cheng, Ling, Ding, Wenbo, Han, Zhu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 2679
container_issue 3
container_start_page 2667
container_title IEEE internet of things journal
container_volume 10
creator Zhang, Hengxi
Lu, Chengyue
Tang, Huaze
Wei, Xiaoli
Liang, Le
Cheng, Ling
Ding, Wenbo
Han, Zhu
description As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.
doi_str_mv 10.1109/JIOT.2022.3214525
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2769391221</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2769391221</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</originalsourceid><addsrcrecordid>eNpNkF1PwjAUhhujiQT5Ad418XrYnm7dekmIKAYkMeht7boOi6PDdovx39sFLrw6H-97PvIgdEvJlFIi7p-Xm-0UCMCUAU0zyC7QCBjkSco5XP7Lr9EkhD0hJI5lVPAR-lgb5ZKFNU2VzGxlKrzum86qnXEdfjXW1a3X5jBUK6O8s26HYytKoe2jgmdN02rV2dZh6_C7-bS6b5THL6b7af1XuEFXtWqCmZzjGL0tHrbzp2S1eVzOZ6tEA6QsyQpWaapyk7GSEGZIKojSKRGV5lxwKERJypRrXue6LItUiQJyXRSFUHkNhLMxujvtPfr2uzehk_v4oIsnJeRcMEEBaHTRk0v7NgRvann09qD8r6REDizlwFIOLOWZJfsD62Fmaw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2769391221</pqid></control><display><type>article</type><title>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</title><source>IEEE Electronic Library (IEL)</source><creator>Zhang, Hengxi ; Lu, Chengyue ; Tang, Huaze ; Wei, Xiaoli ; Liang, Le ; Cheng, Ling ; Ding, Wenbo ; Han, Zhu</creator><creatorcontrib>Zhang, Hengxi ; Lu, Chengyue ; Tang, Huaze ; Wei, Xiaoli ; Liang, Le ; Cheng, Ling ; Ding, Wenbo ; Han, Zhu</creatorcontrib><description>As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.</description><identifier>ISSN: 2327-4662</identifier><identifier>EISSN: 2327-4662</identifier><identifier>DOI: 10.1109/JIOT.2022.3214525</identifier><language>eng</language><publisher>Piscataway: The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</publisher><subject>Communication networks ; Complexity ; Game theory ; Machine learning ; Multiagent systems ; Resource allocation ; Stability ; Vehicles</subject><ispartof>IEEE internet of things journal, 2023-02, Vol.10 (3), p.2667-2679</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</citedby><cites>FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</cites><orcidid>0000-0002-6606-5822 ; 0000-0002-8489-1933 ; 0000-0002-4463-1766 ; 0000-0001-7873-8206 ; 0000-0002-0597-4512</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,777,781,27905,27906</link.rule.ids></links><search><creatorcontrib>Zhang, Hengxi</creatorcontrib><creatorcontrib>Lu, Chengyue</creatorcontrib><creatorcontrib>Tang, Huaze</creatorcontrib><creatorcontrib>Wei, Xiaoli</creatorcontrib><creatorcontrib>Liang, Le</creatorcontrib><creatorcontrib>Cheng, Ling</creatorcontrib><creatorcontrib>Ding, Wenbo</creatorcontrib><creatorcontrib>Han, Zhu</creatorcontrib><title>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</title><title>IEEE internet of things journal</title><description>As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.</description><subject>Communication networks</subject><subject>Complexity</subject><subject>Game theory</subject><subject>Machine learning</subject><subject>Multiagent systems</subject><subject>Resource allocation</subject><subject>Stability</subject><subject>Vehicles</subject><issn>2327-4662</issn><issn>2327-4662</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpNkF1PwjAUhhujiQT5Ad418XrYnm7dekmIKAYkMeht7boOi6PDdovx39sFLrw6H-97PvIgdEvJlFIi7p-Xm-0UCMCUAU0zyC7QCBjkSco5XP7Lr9EkhD0hJI5lVPAR-lgb5ZKFNU2VzGxlKrzum86qnXEdfjXW1a3X5jBUK6O8s26HYytKoe2jgmdN02rV2dZh6_C7-bS6b5THL6b7af1XuEFXtWqCmZzjGL0tHrbzp2S1eVzOZ6tEA6QsyQpWaapyk7GSEGZIKojSKRGV5lxwKERJypRrXue6LItUiQJyXRSFUHkNhLMxujvtPfr2uzehk_v4oIsnJeRcMEEBaHTRk0v7NgRvann09qD8r6REDizlwFIOLOWZJfsD62Fmaw</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Zhang, Hengxi</creator><creator>Lu, Chengyue</creator><creator>Tang, Huaze</creator><creator>Wei, Xiaoli</creator><creator>Liang, Le</creator><creator>Cheng, Ling</creator><creator>Ding, Wenbo</creator><creator>Han, Zhu</creator><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-6606-5822</orcidid><orcidid>https://orcid.org/0000-0002-8489-1933</orcidid><orcidid>https://orcid.org/0000-0002-4463-1766</orcidid><orcidid>https://orcid.org/0000-0001-7873-8206</orcidid><orcidid>https://orcid.org/0000-0002-0597-4512</orcidid></search><sort><creationdate>20230201</creationdate><title>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</title><author>Zhang, Hengxi ; Lu, Chengyue ; Tang, Huaze ; Wei, Xiaoli ; Liang, Le ; Cheng, Ling ; Ding, Wenbo ; Han, Zhu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Communication networks</topic><topic>Complexity</topic><topic>Game theory</topic><topic>Machine learning</topic><topic>Multiagent systems</topic><topic>Resource allocation</topic><topic>Stability</topic><topic>Vehicles</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Hengxi</creatorcontrib><creatorcontrib>Lu, Chengyue</creatorcontrib><creatorcontrib>Tang, Huaze</creatorcontrib><creatorcontrib>Wei, Xiaoli</creatorcontrib><creatorcontrib>Liang, Le</creatorcontrib><creatorcontrib>Cheng, Ling</creatorcontrib><creatorcontrib>Ding, Wenbo</creatorcontrib><creatorcontrib>Han, Zhu</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE internet of things journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Hengxi</au><au>Lu, Chengyue</au><au>Tang, Huaze</au><au>Wei, Xiaoli</au><au>Liang, Le</au><au>Cheng, Ling</au><au>Ding, Wenbo</au><au>Han, Zhu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</atitle><jtitle>IEEE internet of things journal</jtitle><date>2023-02-01</date><risdate>2023</risdate><volume>10</volume><issue>3</issue><spage>2667</spage><epage>2679</epage><pages>2667-2679</pages><issn>2327-4662</issn><eissn>2327-4662</eissn><abstract>As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.</abstract><cop>Piscataway</cop><pub>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</pub><doi>10.1109/JIOT.2022.3214525</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0002-6606-5822</orcidid><orcidid>https://orcid.org/0000-0002-8489-1933</orcidid><orcidid>https://orcid.org/0000-0002-4463-1766</orcidid><orcidid>https://orcid.org/0000-0001-7873-8206</orcidid><orcidid>https://orcid.org/0000-0002-0597-4512</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2327-4662
ispartof IEEE internet of things journal, 2023-02, Vol.10 (3), p.2667-2679
issn 2327-4662
2327-4662
language eng
recordid cdi_proquest_journals_2769391221
source IEEE Electronic Library (IEL)
subjects Communication networks
Complexity
Game theory
Machine learning
Multiagent systems
Resource allocation
Stability
Vehicles
title Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-18T07%3A00%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Mean-Field-Aided%20Multiagent%20Reinforcement%20Learning%20for%20Resource%20Allocation%20in%20Vehicular%20Networks&rft.jtitle=IEEE%20internet%20of%20things%20journal&rft.au=Zhang,%20Hengxi&rft.date=2023-02-01&rft.volume=10&rft.issue=3&rft.spage=2667&rft.epage=2679&rft.pages=2667-2679&rft.issn=2327-4662&rft.eissn=2327-4662&rft_id=info:doi/10.1109/JIOT.2022.3214525&rft_dat=%3Cproquest_cross%3E2769391221%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2769391221&rft_id=info:pmid/&rfr_iscdi=true