Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks
Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control...
Gespeichert in:
Veröffentlicht in: | IEEE network 2021-03, Vol.35 (2), p.112-119 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 119 |
---|---|
container_issue | 2 |
container_start_page | 112 |
container_title | IEEE network |
container_volume | 35 |
creator | Liu, Qingzhi Cheng, Long Jia, Adele Lu Liu, Cong |
description | Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution. |
doi_str_mv | 10.1109/MNET.011.2000303 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_ieee_primary_9351833</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9351833</ieee_id><sourcerecordid>2506473269</sourcerecordid><originalsourceid>FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</originalsourceid><addsrcrecordid>eNo9kEFLAzEQhYMoWKt3wUvA89ZJstnNHqW2KrQVpKK3kA2zmrpNarKl-O_d0uJp4PG9efARcs1gxBhUd_PFZDkCxkYcAASIEzJgUqqMyeLjlAxAVZApyPNzcpHSCoDlUvABWT4gbugrOt-EaHGNvqMzNNE7_0n7iI7Der31zprOBU-nbdj1ke9iaKnz9N1FbDElOsf0RRfY7UL8TpfkrDFtwqvjHZK36WQ5fspmL4_P4_tZZnnFuoxblFYpqKxV0kgr8sLUdS15aWxTFA3LjZLCVooZZk0NQtlGMoUlU1yVqhZDcnv4u4nhZ4up06uwjb6f1FxCkZeCF1VPwYGyMaQUsdGb6NYm_moGeu9O793p3p0-uusrN4eKQ8R_vBL9uhDiD64dar4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2506473269</pqid></control><display><type>article</type><title>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</title><source>IEEE Electronic Library (IEL)</source><creator>Liu, Qingzhi ; Cheng, Long ; Jia, Adele Lu ; Liu, Cong</creator><creatorcontrib>Liu, Qingzhi ; Cheng, Long ; Jia, Adele Lu ; Liu, Cong</creatorcontrib><description>Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.</description><identifier>ISSN: 0890-8044</identifier><identifier>EISSN: 1558-156X</identifier><identifier>DOI: 10.1109/MNET.011.2000303</identifier><identifier>CODEN: IENEET</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Adaptation models ; Benchmark testing ; Communication ; Data models ; Deep learning ; Finite element method ; Flow control ; Internet of Things ; Machine learning ; Network topology ; Properties (attributes) ; Reinforcement learning ; Throughput ; Training ; Wireless communications ; Wireless mesh networks ; Wireless networks</subject><ispartof>IEEE network, 2021-03, Vol.35 (2), p.112-119</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2021</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</citedby><cites>FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9351833$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9351833$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Liu, Qingzhi</creatorcontrib><creatorcontrib>Cheng, Long</creatorcontrib><creatorcontrib>Jia, Adele Lu</creatorcontrib><creatorcontrib>Liu, Cong</creatorcontrib><title>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</title><title>IEEE network</title><addtitle>NET-M</addtitle><description>Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.</description><subject>Adaptation models</subject><subject>Benchmark testing</subject><subject>Communication</subject><subject>Data models</subject><subject>Deep learning</subject><subject>Finite element method</subject><subject>Flow control</subject><subject>Internet of Things</subject><subject>Machine learning</subject><subject>Network topology</subject><subject>Properties (attributes)</subject><subject>Reinforcement learning</subject><subject>Throughput</subject><subject>Training</subject><subject>Wireless communications</subject><subject>Wireless mesh networks</subject><subject>Wireless networks</subject><issn>0890-8044</issn><issn>1558-156X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNo9kEFLAzEQhYMoWKt3wUvA89ZJstnNHqW2KrQVpKK3kA2zmrpNarKl-O_d0uJp4PG9efARcs1gxBhUd_PFZDkCxkYcAASIEzJgUqqMyeLjlAxAVZApyPNzcpHSCoDlUvABWT4gbugrOt-EaHGNvqMzNNE7_0n7iI7Der31zprOBU-nbdj1ke9iaKnz9N1FbDElOsf0RRfY7UL8TpfkrDFtwqvjHZK36WQ5fspmL4_P4_tZZnnFuoxblFYpqKxV0kgr8sLUdS15aWxTFA3LjZLCVooZZk0NQtlGMoUlU1yVqhZDcnv4u4nhZ4up06uwjb6f1FxCkZeCF1VPwYGyMaQUsdGb6NYm_moGeu9O793p3p0-uusrN4eKQ8R_vBL9uhDiD64dar4</recordid><startdate>20210301</startdate><enddate>20210301</enddate><creator>Liu, Qingzhi</creator><creator>Cheng, Long</creator><creator>Jia, Adele Lu</creator><creator>Liu, Cong</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20210301</creationdate><title>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</title><author>Liu, Qingzhi ; Cheng, Long ; Jia, Adele Lu ; Liu, Cong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c291t-2ce5c8809cc85a5c346abbb527acf66f14a853c981a1cab038cf518e7182878b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Adaptation models</topic><topic>Benchmark testing</topic><topic>Communication</topic><topic>Data models</topic><topic>Deep learning</topic><topic>Finite element method</topic><topic>Flow control</topic><topic>Internet of Things</topic><topic>Machine learning</topic><topic>Network topology</topic><topic>Properties (attributes)</topic><topic>Reinforcement learning</topic><topic>Throughput</topic><topic>Training</topic><topic>Wireless communications</topic><topic>Wireless mesh networks</topic><topic>Wireless networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Liu, Qingzhi</creatorcontrib><creatorcontrib>Cheng, Long</creatorcontrib><creatorcontrib>Jia, Adele Lu</creatorcontrib><creatorcontrib>Liu, Cong</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE network</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Liu, Qingzhi</au><au>Cheng, Long</au><au>Jia, Adele Lu</au><au>Liu, Cong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks</atitle><jtitle>IEEE network</jtitle><stitle>NET-M</stitle><date>2021-03-01</date><risdate>2021</risdate><volume>35</volume><issue>2</issue><spage>112</spage><epage>119</epage><pages>112-119</pages><issn>0890-8044</issn><eissn>1558-156X</eissn><coden>IENEET</coden><abstract>Wireless mesh network (WMN) is one of the most promising technologies for Internet of Things (IoT) applications because of its self-adaptive and self-organization nature. To meet different performance requirements on communications in WMNs, traditional approaches always have to program flow control strategies in an explicit way. In this case, the performance of WMNs will be significantly affected by the dynamic properties of underlying networks in real applications. With providing a more flexible solution in mind, in this article, for the first time, we present how we can apply emerging Deep Reinforcement Learning (DRL) on communication flow control in WMNs. Moreover, different from a general DRL based networking solution, in which the network properties are pre-defined, we leverage the adaptive nature of WMNs and propose a self-adaptive DRL approach. Specifically, our method can reconstruct a WMN during the training of a DRL model. In this way, the trained DRL model can capture more properties of WMNs and achieve better performance. As a proof of concept, we have implemented our method with a self-adap-tive Deep Q-learning Network (DQN) model. The evaluation results show that the presented solution can significantly improve the communication performance of data flows in WMNs, compared to a static benchmark solution.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/MNET.011.2000303</doi><tpages>8</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 0890-8044 |
ispartof | IEEE network, 2021-03, Vol.35 (2), p.112-119 |
issn | 0890-8044 1558-156X |
language | eng |
recordid | cdi_ieee_primary_9351833 |
source | IEEE Electronic Library (IEL) |
subjects | Adaptation models Benchmark testing Communication Data models Deep learning Finite element method Flow control Internet of Things Machine learning Network topology Properties (attributes) Reinforcement learning Throughput Training Wireless communications Wireless mesh networks Wireless networks |
title | Deep Reinforcement Learning for Communication Flow Control in Wireless Mesh Networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T17%3A40%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Reinforcement%20Learning%20for%20Communication%20Flow%20Control%20in%20Wireless%20Mesh%20Networks&rft.jtitle=IEEE%20network&rft.au=Liu,%20Qingzhi&rft.date=2021-03-01&rft.volume=35&rft.issue=2&rft.spage=112&rft.epage=119&rft.pages=112-119&rft.issn=0890-8044&rft.eissn=1558-156X&rft.coden=IENEET&rft_id=info:doi/10.1109/MNET.011.2000303&rft_dat=%3Cproquest_RIE%3E2506473269%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2506473269&rft_id=info:pmid/&rft_ieee_id=9351833&rfr_iscdi=true |