AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning
The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays...
Gespeichert in:
Veröffentlicht in: | The Journal of supercomputing 2023-02, Vol.79 (2), p.1592-1621 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1621 |
---|---|
container_issue | 2 |
container_start_page | 1592 |
container_title | The Journal of supercomputing |
container_volume | 79 |
creator | Etefaghi, Amir Sharifian, Saeed |
description | The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency. |
doi_str_mv | 10.1007/s11227-022-04728-5 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2765333767</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2765333767</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</originalsourceid><addsrcrecordid>eNp9kM1KAzEUhYMoWKsv4CrgejSTn8nUXSlaC8Vu6jqkMzcldZqMSUbxJXxmUyu4c3W4cL5zuAeh65LcloTIu1iWlMqCUFoQLmldiBM0KoVk-az5KRqRCSVFLTg9Rxcx7gghnEk2Ql_TVi_cM6R7rB3Wre6TfQdsnYEArgEMbmsdYOMDbm1MwW6GBC1uAXrsYAi6y5I-fHiN2BvTed1at80BeOHXxeNqjnXfd7bRyXoX8UbHTHuHA-QOHxrYg0u4Ax1c5i7RmdFdhKtfHaOXx4f17KlYruaL2XRZNKycpEJKw6AxE0OAs4roDW8MF0SUUk4MAOO1qDkYUoFuKRE10zUDbjaG8rqS-fExujnm9sG_DRCT2vkhuFypqKwEY0xWBxc9uprgYwxgVB_sXodPVRJ12F0dd1d5d_WzuxIZYkcoZrPbQviL_of6BkvAh8w</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2765333767</pqid></control><display><type>article</type><title>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</title><source>SpringerLink Journals - AutoHoldings</source><creator>Etefaghi, Amir ; Sharifian, Saeed</creator><creatorcontrib>Etefaghi, Amir ; Sharifian, Saeed</creatorcontrib><description>The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.</description><identifier>ISSN: 0920-8542</identifier><identifier>EISSN: 1573-0484</identifier><identifier>DOI: 10.1007/s11227-022-04728-5</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Accuracy ; Artificial neural networks ; Audio data ; Big Data ; Cloud computing ; Compilers ; Computational efficiency ; Computer Science ; Computing costs ; Deep learning ; Inference ; Interactive learning ; Internet of Things ; Interpreters ; Machine learning ; Mathematical models ; Network latency ; Neural networks ; Processor Architectures ; Programming Languages ; Recognition ; Teaching methods</subject><ispartof>The Journal of supercomputing, 2023-02, Vol.79 (2), p.1592-1621</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022. Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</citedby><cites>FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</cites><orcidid>0000-0001-7392-7892</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11227-022-04728-5$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11227-022-04728-5$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Etefaghi, Amir</creatorcontrib><creatorcontrib>Sharifian, Saeed</creatorcontrib><title>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</title><title>The Journal of supercomputing</title><addtitle>J Supercomput</addtitle><description>The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.</description><subject>Accuracy</subject><subject>Artificial neural networks</subject><subject>Audio data</subject><subject>Big Data</subject><subject>Cloud computing</subject><subject>Compilers</subject><subject>Computational efficiency</subject><subject>Computer Science</subject><subject>Computing costs</subject><subject>Deep learning</subject><subject>Inference</subject><subject>Interactive learning</subject><subject>Internet of Things</subject><subject>Interpreters</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>Network latency</subject><subject>Neural networks</subject><subject>Processor Architectures</subject><subject>Programming Languages</subject><subject>Recognition</subject><subject>Teaching methods</subject><issn>0920-8542</issn><issn>1573-0484</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kM1KAzEUhYMoWKsv4CrgejSTn8nUXSlaC8Vu6jqkMzcldZqMSUbxJXxmUyu4c3W4cL5zuAeh65LcloTIu1iWlMqCUFoQLmldiBM0KoVk-az5KRqRCSVFLTg9Rxcx7gghnEk2Ql_TVi_cM6R7rB3Wre6TfQdsnYEArgEMbmsdYOMDbm1MwW6GBC1uAXrsYAi6y5I-fHiN2BvTed1at80BeOHXxeNqjnXfd7bRyXoX8UbHTHuHA-QOHxrYg0u4Ax1c5i7RmdFdhKtfHaOXx4f17KlYruaL2XRZNKycpEJKw6AxE0OAs4roDW8MF0SUUk4MAOO1qDkYUoFuKRE10zUDbjaG8rqS-fExujnm9sG_DRCT2vkhuFypqKwEY0xWBxc9uprgYwxgVB_sXodPVRJ12F0dd1d5d_WzuxIZYkcoZrPbQviL_of6BkvAh8w</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Etefaghi, Amir</creator><creator>Sharifian, Saeed</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0001-7392-7892</orcidid></search><sort><creationdate>20230201</creationdate><title>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</title><author>Etefaghi, Amir ; Sharifian, Saeed</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-77f3ecf9f0e4360ab4cf45051779fee348584ef06ead20583a83e4fbf24867373</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Artificial neural networks</topic><topic>Audio data</topic><topic>Big Data</topic><topic>Cloud computing</topic><topic>Compilers</topic><topic>Computational efficiency</topic><topic>Computer Science</topic><topic>Computing costs</topic><topic>Deep learning</topic><topic>Inference</topic><topic>Interactive learning</topic><topic>Internet of Things</topic><topic>Interpreters</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>Network latency</topic><topic>Neural networks</topic><topic>Processor Architectures</topic><topic>Programming Languages</topic><topic>Recognition</topic><topic>Teaching methods</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Etefaghi, Amir</creatorcontrib><creatorcontrib>Sharifian, Saeed</creatorcontrib><collection>CrossRef</collection><jtitle>The Journal of supercomputing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Etefaghi, Amir</au><au>Sharifian, Saeed</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning</atitle><jtitle>The Journal of supercomputing</jtitle><stitle>J Supercomput</stitle><date>2023-02-01</date><risdate>2023</risdate><volume>79</volume><issue>2</issue><spage>1592</spage><epage>1621</epage><pages>1592-1621</pages><issn>0920-8542</issn><eissn>1573-0484</eissn><abstract>The increasing expansion of Internet-of-Things (IoT) in the world requires Big Data analytic infrastructures to produce valuable knowledge in IoT applications. IoT includes devices with limited resources, whereby it requires efficient platforms to process massive data obtained from sensors. Nowadays, many IoT applications such as audio and video recognition depend on state-of-the-art Deep Neural Networks (DNNs). Therefore, we need to execute DNNs on IoT devices. DNNs offer excellent recognition accuracy but they suffer from high computational and memory resource demands. Due to these constraints, currently, IoT applications that depend on deep learning are mostly offloaded to cloudlets and clouds. Offloading imposes extra network bandwidth consumption costs in addition to delayed response for IoT devices. In this paper, we propose a method that instead of using all layers of DNN for inference, only selects a subset of layers that provide sufficient accuracy for each task. We propose AdaInNet, a method to significantly reduce computational cost and network latency in DNN-based IoT applications while maintaining prediction accuracy based on Distributed DNNs (DDNNs). The method uses modified Distributed DNNs with early exits in order to minimize computation costs and network latency by selecting sub-layers or exit branches of DDNNs with early exits. We also proposed a hybrid Classifier-Wise (CW)—Interactive learning method for the training of DDNNs and Agent’s networks. Furthermore, we create a custom agent model for the Advantage Actor-Critic Deep Reinforcement Learning method in order to preserve recognition accuracy while utilizing a minimum number of layers. Finally, we execute the extensive numerical simulation, in order to evaluate and compare our proposed AdaInNet method with rival methods under standard CIFAR 100 and CIFAR 10 datasets and ResNet-110 and ResNet-32 DNNs which are used in IoT applications in previous works. The results provide strong quantitative evidence that the AdaInNet method not only accelerates inference but also reduces computational cost and latency.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11227-022-04728-5</doi><tpages>30</tpages><orcidid>https://orcid.org/0000-0001-7392-7892</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0920-8542 |
ispartof | The Journal of supercomputing, 2023-02, Vol.79 (2), p.1592-1621 |
issn | 0920-8542 1573-0484 |
language | eng |
recordid | cdi_proquest_journals_2765333767 |
source | SpringerLink Journals - AutoHoldings |
subjects | Accuracy Artificial neural networks Audio data Big Data Cloud computing Compilers Computational efficiency Computer Science Computing costs Deep learning Inference Interactive learning Internet of Things Interpreters Machine learning Mathematical models Network latency Neural networks Processor Architectures Programming Languages Recognition Teaching methods |
title | AdaInNet: an adaptive inference engine for distributed deep neural networks offloading in IoT-FOG applications based on reinforcement learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T21%3A42%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=AdaInNet:%20an%20adaptive%20inference%20engine%20for%20distributed%20deep%20neural%20networks%20offloading%20in%20IoT-FOG%20applications%20based%20on%20reinforcement%20learning&rft.jtitle=The%20Journal%20of%20supercomputing&rft.au=Etefaghi,%20Amir&rft.date=2023-02-01&rft.volume=79&rft.issue=2&rft.spage=1592&rft.epage=1621&rft.pages=1592-1621&rft.issn=0920-8542&rft.eissn=1573-0484&rft_id=info:doi/10.1007/s11227-022-04728-5&rft_dat=%3Cproquest_cross%3E2765333767%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2765333767&rft_id=info:pmid/&rfr_iscdi=true |