AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning

The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of supercomputing 2023-02, Vol.79 (2), p.1592-1621
Hauptverfasser: Etefaghi, Amir, Sharifian, Saeed
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1621
container_issue 2
container_start_page 1592
container_title The Journal of supercomputing
container_volume 79
creator Etefaghi, Amir
Sharifian, Saeed
description The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.
doi_str_mv 10.1007/s11227-022-04728-5
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2765333767</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2765333767</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</originalsourceid><addsrcrecordid>eNp9kM1KAzEUhYMoWKsv4CrgejSTn8nUXSlaC8Vu6jqkMzcldZqMSUbxJXxmUyu4c3W4cL5zuAeh65LcloTIu1iWlMqCUFoQLmldiBM0KoVk-az5KRqRCSVFLTg9Rxcx7gghnEk2Ql_TVi_cM6R7rB3Wre6TfQdsnYEArgEMbmsdYOMDbm1MwW6GBC1uAXrsYAi6y5I-fHiN2BvTed1at80BeOHXxeNqjnXfd7bRyXoX8UbHTHuHA-QOHxrYg0u4Ax1c5i7RmdFdhKtfHaOXx4f17KlYruaL2XRZNKycpEJKw6AxE0OAs4roDW8MF0SUUk4MAOO1qDkYUoFuKRE10zUDbjaG8rqS-fExujnm9sG_DRCT2vkhuFypqKwEY0xWBxc9uprgYwxgVB_sXodPVRJ12F0dd1d5d_WzuxIZYkcoZrPbQviL_of6BkvAh8w</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2765333767</pqid></control><display><type>article</type><title>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</title><source>SpringerLink Journals - AutoHoldings</source><creator>Etefaghi, Amir ; Sharifian, Saeed</creator><creatorcontrib>Etefaghi, Amir ; Sharifian, Saeed</creatorcontrib><description>The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.</description><identifier>ISSN: 0920-8542</identifier><identifier>EISSN: 1573-0484</identifier><identifier>DOI: 10.1007/s11227-022-04728-5</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Accuracy ; Artificial neural networks ; Audio data ; Big Data ; Cloud computing ; Compilers ; Computational efficiency ; Computer Science ; Computing costs ; Deep learning ; Inference ; Interactive learning ; Internet of Things ; Interpreters ; Machine learning ; Mathematical models ; Network latency ; Neural networks ; Processor Architectures ; Programming Languages ; Recognition ; Teaching methods</subject><ispartof>The Journal of supercomputing, 2023-02, Vol.79 (2), p.1592-1621</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</citedby><cites>FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</cites><orcidid>0000-0001-7392-7892</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11227-022-04728-5$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11227-022-04728-5$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Etefaghi, Amir</creatorcontrib><creatorcontrib>Sharifian, Saeed</creatorcontrib><title>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</title><title>The Journal of supercomputing</title><addtitle>J Supercomput</addtitle><description>The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.</description><subject>Accuracy</subject><subject>Artificial neural networks</subject><subject>Audio data</subject><subject>Big Data</subject><subject>Cloud computing</subject><subject>Compilers</subject><subject>Computational efficiency</subject><subject>Computer Science</subject><subject>Computing costs</subject><subject>Deep learning</subject><subject>Inference</subject><subject>Interactive learning</subject><subject>Internet of Things</subject><subject>Interpreters</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>Network latency</subject><subject>Neural networks</subject><subject>Processor Architectures</subject><subject>Programming Languages</subject><subject>Recognition</subject><subject>Teaching methods</subject><issn>0920-8542</issn><issn>1573-0484</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kM1KAzEUhYMoWKsv4CrgejSTn8nUXSlaC8Vu6jqkMzcldZqMSUbxJXxmUyu4c3W4cL5zuAeh65LcloTIu1iWlMqCUFoQLmldiBM0KoVk-az5KRqRCSVFLTg9Rxcx7gghnEk2Ql_TVi_cM6R7rB3Wre6TfQdsnYEArgEMbmsdYOMDbm1MwW6GBC1uAXrsYAi6y5I-fHiN2BvTed1at80BeOHXxeNqjnXfd7bRyXoX8UbHTHuHA-QOHxrYg0u4Ax1c5i7RmdFdhKtfHaOXx4f17KlYruaL2XRZNKycpEJKw6AxE0OAs4roDW8MF0SUUk4MAOO1qDkYUoFuKRE10zUDbjaG8rqS-fExujnm9sG_DRCT2vkhuFypqKwEY0xWBxc9uprgYwxgVB_sXodPVRJ12F0dd1d5d_WzuxIZYkcoZrPbQviL_of6BkvAh8w</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Etefaghi, Amir</creator><creator>Sharifian, Saeed</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0001-7392-7892</orcidid></search><sort><creationdate>20230201</creationdate><title>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</title><author>Etefaghi, Amir ; Sharifian, Saeed</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Artificial neural networks</topic><topic>Audio data</topic><topic>Big Data</topic><topic>Cloud computing</topic><topic>Compilers</topic><topic>Computational efficiency</topic><topic>Computer Science</topic><topic>Computing costs</topic><topic>Deep learning</topic><topic>Inference</topic><topic>Interactive learning</topic><topic>Internet of Things</topic><topic>Interpreters</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>Network latency</topic><topic>Neural networks</topic><topic>Processor Architectures</topic><topic>Programming Languages</topic><topic>Recognition</topic><topic>Teaching methods</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Etefaghi, Amir</creatorcontrib><creatorcontrib>Sharifian, Saeed</creatorcontrib><collection>CrossRef</collection><jtitle>The Journal of supercomputing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Etefaghi, Amir</au><au>Sharifian, Saeed</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</atitle><jtitle>The Journal of supercomputing</jtitle><stitle>J Supercomput</stitle><date>2023-02-01</date><risdate>2023</risdate><volume>79</volume><issue>2</issue><spage>1592</spage><epage>1621</epage><pages>1592-1621</pages><issn>0920-8542</issn><eissn>1573-0484</eissn><abstract>The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11227-022-04728-5</doi><tpages>30</tpages><orcidid>https://orcid.org/0000-0001-7392-7892</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0920-8542
ispartof The Journal of supercomputing, 2023-02, Vol.79 (2), p.1592-1621
issn 0920-8542
1573-0484
language eng
recordid cdi_proquest_journals_2765333767
source SpringerLink Journals - AutoHoldings
subjects Accuracy
Artificial neural networks
Audio data
Big Data
Cloud computing
Compilers
Computational efficiency
Computer Science
Computing costs
Deep learning
Inference
Interactive learning
Internet of Things
Interpreters
Machine learning
Mathematical models
Network latency
Neural networks
Processor Architectures
Programming Languages
Recognition
Teaching methods
title AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T21%3A42%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=AdaInNet:%20an%20adaptive%20inference%20engine%20for%20distributed%20deep%20neural%20networks%20offloading%20in%20IoT-FOG%20applications%20based%20on%20reinforcement%20learning&rft.jtitle=The%20Journal%20of%20supercomputing&rft.au=Etefaghi,%20Amir&rft.date=2023-02-01&rft.volume=79&rft.issue=2&rft.spage=1592&rft.epage=1621&rft.pages=1592-1621&rft.issn=0920-8542&rft.eissn=1573-0484&rft_id=info:doi/10.1007/s11227-022-04728-5&rft_dat=%3Cproquest_cross%3E2765333767%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2765333767&rft_id=info:pmid/&rfr_iscdi=true