Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks

Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE network 2021-03, Vol.35 (2), p.112-119
Hauptverfasser: Liu, Qingzhi, Cheng, Long, Jia, Adele Lu, Liu, Cong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 119
container_issue 2
container_start_page 112
container_title IEEE network
container_volume 35
creator Liu, Qingzhi
Cheng, Long
Jia, Adele Lu
Liu, Cong
description Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.
doi_str_mv 10.1109/MNET.011.2000303
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_ieee_primary_9351833</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9351833</ieee_id><sourcerecordid>2506473269</sourcerecordid><originalsourceid>FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</originalsourceid><addsrcrecordid>eNo9kEFLAzEQhYMoWKt3wUvA89ZJstnNHqW2KrQVpKK3kA2zmrpNarKl-O_d0uJp4PG9efARcs1gxBhUd_PFZDkCxkYcAASIEzJgUqqMyeLjlAxAVZApyPNzcpHSCoDlUvABWT4gbugrOt-EaHGNvqMzNNE7_0n7iI7Der31zprOBU-nbdj1ke9iaKnz9N1FbDElOsf0RRfY7UL8TpfkrDFtwqvjHZK36WQ5fspmL4_P4_tZZnnFuoxblFYpqKxV0kgr8sLUdS15aWxTFA3LjZLCVooZZk0NQtlGMoUlU1yVqhZDcnv4u4nhZ4up06uwjb6f1FxCkZeCF1VPwYGyMaQUsdGb6NYm_moGeu9O793p3p0-uusrN4eKQ8R_vBL9uhDiD64dar4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2506473269</pqid></control><display><type>article</type><title>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</title><source>IEEE Electronic Library (IEL)</source><creator>Liu, Qingzhi ; Cheng, Long ; Jia, Adele Lu ; Liu, Cong</creator><creatorcontrib>Liu, Qingzhi ; Cheng, Long ; Jia, Adele Lu ; Liu, Cong</creatorcontrib><description>Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.</description><identifier>ISSN: 0890-8044</identifier><identifier>EISSN: 1558-156X</identifier><identifier>DOI: 10.1109/MNET.011.2000303</identifier><identifier>CODEN: IENEET</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Adaptation models ; Benchmark testing ; Communication ; Data models ; Deep learning ; Finite element method ; Flow control ; Internet of Things ; Machine learning ; Network topology ; Properties (attributes) ; Reinforcement learning ; Throughput ; Training ; Wireless communications ; Wireless mesh networks ; Wireless networks</subject><ispartof>IEEE network, 2021-03, Vol.35 (2), p.112-119</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2021</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</citedby><cites>FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9351833$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9351833$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Liu, Qingzhi</creatorcontrib><creatorcontrib>Cheng, Long</creatorcontrib><creatorcontrib>Jia, Adele Lu</creatorcontrib><creatorcontrib>Liu, Cong</creatorcontrib><title>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</title><title>IEEE network</title><addtitle>NET-M</addtitle><description>Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.</description><subject>Adaptation models</subject><subject>Benchmark testing</subject><subject>Communication</subject><subject>Data models</subject><subject>Deep learning</subject><subject>Finite element method</subject><subject>Flow control</subject><subject>Internet of Things</subject><subject>Machine learning</subject><subject>Network topology</subject><subject>Properties (attributes)</subject><subject>Reinforcement learning</subject><subject>Throughput</subject><subject>Training</subject><subject>Wireless communications</subject><subject>Wireless mesh networks</subject><subject>Wireless networks</subject><issn>0890-8044</issn><issn>1558-156X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNo9kEFLAzEQhYMoWKt3wUvA89ZJstnNHqW2KrQVpKK3kA2zmrpNarKl-O_d0uJp4PG9efARcs1gxBhUd_PFZDkCxkYcAASIEzJgUqqMyeLjlAxAVZApyPNzcpHSCoDlUvABWT4gbugrOt-EaHGNvqMzNNE7_0n7iI7Der31zprOBU-nbdj1ke9iaKnz9N1FbDElOsf0RRfY7UL8TpfkrDFtwqvjHZK36WQ5fspmL4_P4_tZZnnFuoxblFYpqKxV0kgr8sLUdS15aWxTFA3LjZLCVooZZk0NQtlGMoUlU1yVqhZDcnv4u4nhZ4up06uwjb6f1FxCkZeCF1VPwYGyMaQUsdGb6NYm_moGeu9O793p3p0-uusrN4eKQ8R_vBL9uhDiD64dar4</recordid><startdate>20210301</startdate><enddate>20210301</enddate><creator>Liu, Qingzhi</creator><creator>Cheng, Long</creator><creator>Jia, Adele Lu</creator><creator>Liu, Cong</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20210301</creationdate><title>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</title><author>Liu, Qingzhi ; Cheng, Long ; Jia, Adele Lu ; Liu, Cong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Adaptation models</topic><topic>Benchmark testing</topic><topic>Communication</topic><topic>Data models</topic><topic>Deep learning</topic><topic>Finite element method</topic><topic>Flow control</topic><topic>Internet of Things</topic><topic>Machine learning</topic><topic>Network topology</topic><topic>Properties (attributes)</topic><topic>Reinforcement learning</topic><topic>Throughput</topic><topic>Training</topic><topic>Wireless communications</topic><topic>Wireless mesh networks</topic><topic>Wireless networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Liu, Qingzhi</creatorcontrib><creatorcontrib>Cheng, Long</creatorcontrib><creatorcontrib>Jia, Adele Lu</creatorcontrib><creatorcontrib>Liu, Cong</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE network</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Liu, Qingzhi</au><au>Cheng, Long</au><au>Jia, Adele Lu</au><au>Liu, Cong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</atitle><jtitle>IEEE network</jtitle><stitle>NET-M</stitle><date>2021-03-01</date><risdate>2021</risdate><volume>35</volume><issue>2</issue><spage>112</spage><epage>119</epage><pages>112-119</pages><issn>0890-8044</issn><eissn>1558-156X</eissn><coden>IENEET</coden><abstract>Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/MNET.011.2000303</doi><tpages>8</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 0890-8044
ispartof IEEE network, 2021-03, Vol.35 (2), p.112-119
issn 0890-8044
1558-156X
language eng
recordid cdi_ieee_primary_9351833
source IEEE Electronic Library (IEL)
subjects Adaptation models
Benchmark testing
Communication
Data models
Deep learning
Finite element method
Flow control
Internet of Things
Machine learning
Network topology
Properties (attributes)
Reinforcement learning
Throughput
Training
Wireless communications
Wireless mesh networks
Wireless networks
title Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T17%3A40%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Reinforcement%20Learning%20for%20Communication%20Flow%20Control%20in%20Wireless%20Mesh%20Networks&rft.jtitle=IEEE%20network&rft.au=Liu,%20Qingzhi&rft.date=2021-03-01&rft.volume=35&rft.issue=2&rft.spage=112&rft.epage=119&rft.pages=112-119&rft.issn=0890-8044&rft.eissn=1558-156X&rft.coden=IENEET&rft_id=info:doi/10.1109/MNET.011.2000303&rft_dat=%3Cproquest_RIE%3E2506473269%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2506473269&rft_id=info:pmid/&rft_ieee_id=9351833&rfr_iscdi=true