Multitask Transfer Deep Reinforcement Learning for Timely Data Collection in Rechargeable-UAV-Aided IoT Networks

Thanks to their high-flexibility and low-operational cost, unmanned aerial vehicles (UAVs) can be used to support mission-critical applications in the Internet of Things (IoT). However, due to the limited onboard energy, it is difficult for UAVs to provide continuous data collection. In this article...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE internet of things journal 2023-12, Vol.10 (23), p.20545-20559
Hauptverfasser: Yi, Mengjie, Wang, Xijun, Liu, Juan, Zhang, Yan, Hou, Ronghui
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 20559
container_issue 23
container_start_page 20545
container_title IEEE internet of things journal
container_volume 10
creator Yi, Mengjie
Wang, Xijun
Liu, Juan
Zhang, Yan
Hou, Ronghui
description Thanks to their high-flexibility and low-operational cost, unmanned aerial vehicles (UAVs) can be used to support mission-critical applications in the Internet of Things (IoT). However, due to the limited onboard energy, it is difficult for UAVs to provide continuous data collection. In this article, we study the problem of rechargeable-UAV-aided timely data collection in IoT networks, where the UAV collects status updates from multiple sensors and gets recharged from the charging stations (CSs) to keep its energy level above a threshold. To tradeoff the information freshness and energy consumption, we formulate a Markov decision process (MDP) with the objective of minimizing the weighted sum of the average total Age of Information and average recharging price. Under the dynamics and uncertainty of the environment, we propose a multitask transfer deep reinforcement learning method to jointly optimize the UAV ’ s flight trajectory, transmission scheduling, and battery recharging. To enable the application of the learned policy to new environments with similar settings and avoid starting from scratch, we develop a multitask network made up of common knowledge layers and task-specific knowledge layers. It specifically makes it possible for the transfer of common knowledge between environments with different network scales (e.g., different numbers of sensors/CSs) and/or topologies (e.g., different locations of sensors/CSs). Simulation results demonstrate that the proposed algorithm can adapt to new environments and achieve superior performance compared to the baseline algorithms.
doi_str_mv 10.1109/JIOT.2023.3300927
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2892375247</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2892375247</sourcerecordid><originalsourceid>FETCH-LOGICAL-c225t-b668a35de6c27dcea7d8a853d125bc3ff35dfa7113a9307d1e673f1dc33f304a3</originalsourceid><addsrcrecordid>eNpNkEtPAjEQxzdGEwnyAbw18bzYB7vdPRLwgUFJzOK1Ke0UC0uLbYnh27sEDp5mMvN_JL8suyd4SAiuH99mi2ZIMWVDxjCuKb_KepRRno_Kkl7_22-zQYwbjHFnK0hd9rL9-6FNNsm4RU2QLhoIaAqwR59gnfFBwQ5cQnOQwVm3Rt0JNXYH7RFNZZJo4tsWVLLeIes6k_qWYQ1y1UK-HH_lY6tBo5lv0AekXx-28S67MbKNMLjMfrZ8fmomr_l88TKbjOe5orRI-aosK8kKDaWiXCuQXFeyKpgmtFgpZkz3M5ITwmTNMNcESs4M0Yoxw_BIsn72cM7dB_9zgJjExh-C6yoFrWrKeEFHvFORs0oFH2MAI_bB7mQ4CoLFia04sRUntuLClv0B2QJs5g</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2892375247</pqid></control><display><type>article</type><title>Multitask Transfer Deep Reinforcement Learning for Timely Data Collection in Rechargeable-UAV-Aided IoT Networks</title><source>IEEE Electronic Library (IEL)</source><creator>Yi, Mengjie ; Wang, Xijun ; Liu, Juan ; Zhang, Yan ; Hou, Ronghui</creator><creatorcontrib>Yi, Mengjie ; Wang, Xijun ; Liu, Juan ; Zhang, Yan ; Hou, Ronghui</creatorcontrib><description>Thanks to their high-flexibility and low-operational cost, unmanned aerial vehicles (UAVs) can be used to support mission-critical applications in the Internet of Things (IoT). However, due to the limited onboard energy, it is difficult for UAVs to provide continuous data collection. In this article, we study the problem of rechargeable-UAV-aided timely data collection in IoT networks, where the UAV collects status updates from multiple sensors and gets recharged from the charging stations (CSs) to keep its energy level above a threshold. To tradeoff the information freshness and energy consumption, we formulate a Markov decision process (MDP) with the objective of minimizing the weighted sum of the average total Age of Information and average recharging price. Under the dynamics and uncertainty of the environment, we propose a multitask transfer deep reinforcement learning method to jointly optimize the UAV ’ s flight trajectory, transmission scheduling, and battery recharging. To enable the application of the learned policy to new environments with similar settings and avoid starting from scratch, we develop a multitask network made up of common knowledge layers and task-specific knowledge layers. It specifically makes it possible for the transfer of common knowledge between environments with different network scales (e.g., different numbers of sensors/CSs) and/or topologies (e.g., different locations of sensors/CSs). Simulation results demonstrate that the proposed algorithm can adapt to new environments and achieve superior performance compared to the baseline algorithms.</description><identifier>ISSN: 2327-4662</identifier><identifier>EISSN: 2327-4662</identifier><identifier>DOI: 10.1109/JIOT.2023.3300927</identifier><language>eng</language><publisher>Piscataway: The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</publisher><subject>Algorithms ; Charging ; Data collection ; Deep learning ; Energy consumption ; Energy levels ; Internet of Things ; Knowledge management ; Markov processes ; Rechargeable batteries ; Recharging ; Sensors ; Topology ; Unmanned aerial vehicles</subject><ispartof>IEEE internet of things journal, 2023-12, Vol.10 (23), p.20545-20559</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c225t-b668a35de6c27dcea7d8a853d125bc3ff35dfa7113a9307d1e673f1dc33f304a3</cites><orcidid>0000-0003-3504-9763 ; 0000-0001-6343-6184 ; 0000-0002-4560-2350 ; 0000-0003-3402-3123</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Yi, Mengjie</creatorcontrib><creatorcontrib>Wang, Xijun</creatorcontrib><creatorcontrib>Liu, Juan</creatorcontrib><creatorcontrib>Zhang, Yan</creatorcontrib><creatorcontrib>Hou, Ronghui</creatorcontrib><title>Multitask Transfer Deep Reinforcement Learning for Timely Data Collection in Rechargeable-UAV-Aided IoT Networks</title><title>IEEE internet of things journal</title><description>Thanks to their high-flexibility and low-operational cost, unmanned aerial vehicles (UAVs) can be used to support mission-critical applications in the Internet of Things (IoT). However, due to the limited onboard energy, it is difficult for UAVs to provide continuous data collection. In this article, we study the problem of rechargeable-UAV-aided timely data collection in IoT networks, where the UAV collects status updates from multiple sensors and gets recharged from the charging stations (CSs) to keep its energy level above a threshold. To tradeoff the information freshness and energy consumption, we formulate a Markov decision process (MDP) with the objective of minimizing the weighted sum of the average total Age of Information and average recharging price. Under the dynamics and uncertainty of the environment, we propose a multitask transfer deep reinforcement learning method to jointly optimize the UAV ’ s flight trajectory, transmission scheduling, and battery recharging. To enable the application of the learned policy to new environments with similar settings and avoid starting from scratch, we develop a multitask network made up of common knowledge layers and task-specific knowledge layers. It specifically makes it possible for the transfer of common knowledge between environments with different network scales (e.g., different numbers of sensors/CSs) and/or topologies (e.g., different locations of sensors/CSs). Simulation results demonstrate that the proposed algorithm can adapt to new environments and achieve superior performance compared to the baseline algorithms.</description><subject>Algorithms</subject><subject>Charging</subject><subject>Data collection</subject><subject>Deep learning</subject><subject>Energy consumption</subject><subject>Energy levels</subject><subject>Internet of Things</subject><subject>Knowledge management</subject><subject>Markov processes</subject><subject>Rechargeable batteries</subject><subject>Recharging</subject><subject>Sensors</subject><subject>Topology</subject><subject>Unmanned aerial vehicles</subject><issn>2327-4662</issn><issn>2327-4662</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpNkEtPAjEQxzdGEwnyAbw18bzYB7vdPRLwgUFJzOK1Ke0UC0uLbYnh27sEDp5mMvN_JL8suyd4SAiuH99mi2ZIMWVDxjCuKb_KepRRno_Kkl7_22-zQYwbjHFnK0hd9rL9-6FNNsm4RU2QLhoIaAqwR59gnfFBwQ5cQnOQwVm3Rt0JNXYH7RFNZZJo4tsWVLLeIes6k_qWYQ1y1UK-HH_lY6tBo5lv0AekXx-28S67MbKNMLjMfrZ8fmomr_l88TKbjOe5orRI-aosK8kKDaWiXCuQXFeyKpgmtFgpZkz3M5ITwmTNMNcESs4M0Yoxw_BIsn72cM7dB_9zgJjExh-C6yoFrWrKeEFHvFORs0oFH2MAI_bB7mQ4CoLFia04sRUntuLClv0B2QJs5g</recordid><startdate>20231201</startdate><enddate>20231201</enddate><creator>Yi, Mengjie</creator><creator>Wang, Xijun</creator><creator>Liu, Juan</creator><creator>Zhang, Yan</creator><creator>Hou, Ronghui</creator><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0003-3504-9763</orcidid><orcidid>https://orcid.org/0000-0001-6343-6184</orcidid><orcidid>https://orcid.org/0000-0002-4560-2350</orcidid><orcidid>https://orcid.org/0000-0003-3402-3123</orcidid></search><sort><creationdate>20231201</creationdate><title>Multitask Transfer Deep Reinforcement Learning for Timely Data Collection in Rechargeable-UAV-Aided IoT Networks</title><author>Yi, Mengjie ; Wang, Xijun ; Liu, Juan ; Zhang, Yan ; Hou, Ronghui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c225t-b668a35de6c27dcea7d8a853d125bc3ff35dfa7113a9307d1e673f1dc33f304a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Algorithms</topic><topic>Charging</topic><topic>Data collection</topic><topic>Deep learning</topic><topic>Energy consumption</topic><topic>Energy levels</topic><topic>Internet of Things</topic><topic>Knowledge management</topic><topic>Markov processes</topic><topic>Rechargeable batteries</topic><topic>Recharging</topic><topic>Sensors</topic><topic>Topology</topic><topic>Unmanned aerial vehicles</topic><toplevel>online_resources</toplevel><creatorcontrib>Yi, Mengjie</creatorcontrib><creatorcontrib>Wang, Xijun</creatorcontrib><creatorcontrib>Liu, Juan</creatorcontrib><creatorcontrib>Zhang, Yan</creatorcontrib><creatorcontrib>Hou, Ronghui</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE internet of things journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yi, Mengjie</au><au>Wang, Xijun</au><au>Liu, Juan</au><au>Zhang, Yan</au><au>Hou, Ronghui</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multitask Transfer Deep Reinforcement Learning for Timely Data Collection in Rechargeable-UAV-Aided IoT Networks</atitle><jtitle>IEEE internet of things journal</jtitle><date>2023-12-01</date><risdate>2023</risdate><volume>10</volume><issue>23</issue><spage>20545</spage><epage>20559</epage><pages>20545-20559</pages><issn>2327-4662</issn><eissn>2327-4662</eissn><abstract>Thanks to their high-flexibility and low-operational cost, unmanned aerial vehicles (UAVs) can be used to support mission-critical applications in the Internet of Things (IoT). However, due to the limited onboard energy, it is difficult for UAVs to provide continuous data collection. In this article, we study the problem of rechargeable-UAV-aided timely data collection in IoT networks, where the UAV collects status updates from multiple sensors and gets recharged from the charging stations (CSs) to keep its energy level above a threshold. To tradeoff the information freshness and energy consumption, we formulate a Markov decision process (MDP) with the objective of minimizing the weighted sum of the average total Age of Information and average recharging price. Under the dynamics and uncertainty of the environment, we propose a multitask transfer deep reinforcement learning method to jointly optimize the UAV ’ s flight trajectory, transmission scheduling, and battery recharging. To enable the application of the learned policy to new environments with similar settings and avoid starting from scratch, we develop a multitask network made up of common knowledge layers and task-specific knowledge layers. It specifically makes it possible for the transfer of common knowledge between environments with different network scales (e.g., different numbers of sensors/CSs) and/or topologies (e.g., different locations of sensors/CSs). Simulation results demonstrate that the proposed algorithm can adapt to new environments and achieve superior performance compared to the baseline algorithms.</abstract><cop>Piscataway</cop><pub>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</pub><doi>10.1109/JIOT.2023.3300927</doi><tpages>15</tpages><orcidid>https://orcid.org/0000-0003-3504-9763</orcidid><orcidid>https://orcid.org/0000-0001-6343-6184</orcidid><orcidid>https://orcid.org/0000-0002-4560-2350</orcidid><orcidid>https://orcid.org/0000-0003-3402-3123</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2327-4662
ispartof IEEE internet of things journal, 2023-12, Vol.10 (23), p.20545-20559
issn 2327-4662
2327-4662
language eng
recordid cdi_proquest_journals_2892375247
source IEEE Electronic Library (IEL)
subjects Algorithms
Charging
Data collection
Deep learning
Energy consumption
Energy levels
Internet of Things
Knowledge management
Markov processes
Rechargeable batteries
Recharging
Sensors
Topology
Unmanned aerial vehicles
title Multitask Transfer Deep Reinforcement Learning for Timely Data Collection in Rechargeable-UAV-Aided IoT Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T22%3A27%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multitask%20Transfer%20Deep%20Reinforcement%20Learning%20for%20Timely%20Data%20Collection%20in%20Rechargeable-UAV-Aided%20IoT%20Networks&rft.jtitle=IEEE%20internet%20of%20things%20journal&rft.au=Yi,%20Mengjie&rft.date=2023-12-01&rft.volume=10&rft.issue=23&rft.spage=20545&rft.epage=20559&rft.pages=20545-20559&rft.issn=2327-4662&rft.eissn=2327-4662&rft_id=info:doi/10.1109/JIOT.2023.3300927&rft_dat=%3Cproquest_cross%3E2892375247%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2892375247&rft_id=info:pmid/&rfr_iscdi=true