Tensor Completion Methods for Collaborative Intelligence

In the race to bring Artificial Intelligence (AI) to the edge, collaborative intelligence has emerged as a promising way to lighten the computation load on edge devices that run applications based on Deep Neural Networks (DNNs). Typically, a deep model is split at a given layer into edge and cloud s...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2020, Vol.8, p.41162-41174
Hauptverfasser: Bragilevsky, Lior, Bajic, Ivan V.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 41174
container_issue
container_start_page 41162
container_title IEEE access
container_volume 8
creator Bragilevsky, Lior
Bajic, Ivan V.
description In the race to bring Artificial Intelligence (AI) to the edge, collaborative intelligence has emerged as a promising way to lighten the computation load on edge devices that run applications based on Deep Neural Networks (DNNs). Typically, a deep model is split at a given layer into edge and cloud sub-models. The deep feature tensor produced by the edge sub-model is transmitted to the cloud, where the remaining computationally intensive workload is performed by the cloud sub-model. The communication channel between the edge and cloud is imperfect, which will result in missing data in the deep feature tensor received at the cloud side, an issue that has mostly been ignored by existing literature on the topic. In this paper we study four methods for recovering missing data in the deep feature tensor. Three of the studied methods are existing, generic tensor completion methods, and are adapted here to recover deep feature tensor data, while the fourth method is newly developed specifically for deep feature tensor completion. Simulation studies show that the new method is 3-18 times faster than the other three methods, which is an important consideration in collaborative intelligence. For VGG16's sparse tensors, all methods produce statistically equivalent classification results across all loss levels tested. For ResNet34's non-sparse tensors, the new method offers statistically better classification accuracy (by 0.25%-6.30%) compared to other methods for matched execution speeds, and second-best accuracy among the four methods when they are allowed to run until convergence.
doi_str_mv 10.1109/ACCESS.2020.2977050
format Article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2454756867</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9017944</ieee_id><doaj_id>oai_doaj_org_article_8fbc069cfa044497b38a94b6b41168b3</doaj_id><sourcerecordid>2454756867</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-bc6976178c53e14a96a290970fec00b14278ac9e605d579e4f9965f3d08fe00a3</originalsourceid><addsrcrecordid>eNpNkE9rwkAQxUNpodL6CbwEetbOZv8fJdhWsPSgPS-bzayNxKzdxEK_faMR6VxmeMx7M_ySZEJgRgjo53meL9brWQYZzDItJXC4SUYZEXpKORW3_-b7ZNy2O-hL9RKXo0RtsGlDTPOwP9TYVaFJ37H7CmWb-rNc17YI0XbVD6bLpsO6rrbYOHxM7rytWxxf-kPy-bLY5G_T1cfrMp-vpo6B6qaFE1oKIpXjFAmzWthMg5bg0QEUhGVSWadRAC-51Mi81oJ7WoLyCGDpQ7Iccstgd-YQq72NvybYypyFELfGxq5yNRrlCwdCO2-BMaZlQZXVrBAFI0SogvZZT0PWIYbvI7ad2YVjbPr3TcY4k1woIfstOmy5GNo2or9eJWBOxM1A3JyImwvx3jUZXBUiXh0aiNSM0T8iKnqw</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2454756867</pqid></control><display><type>article</type><title>Tensor Completion Methods for Collaborative Intelligence</title><source>DOAJ Directory of Open Access Journals</source><source>IEEE Xplore Open Access Journals</source><source>EZB Electronic Journals Library</source><creator>Bragilevsky, Lior ; Bajic, Ivan V.</creator><creatorcontrib>Bragilevsky, Lior ; Bajic, Ivan V.</creatorcontrib><description>In the race to bring Artificial Intelligence (AI) to the edge, collaborative intelligence has emerged as a promising way to lighten the computation load on edge devices that run applications based on Deep Neural Networks (DNNs). Typically, a deep model is split at a given layer into edge and cloud sub-models. The deep feature tensor produced by the edge sub-model is transmitted to the cloud, where the remaining computationally intensive workload is performed by the cloud sub-model. The communication channel between the edge and cloud is imperfect, which will result in missing data in the deep feature tensor received at the cloud side, an issue that has mostly been ignored by existing literature on the topic. In this paper we study four methods for recovering missing data in the deep feature tensor. Three of the studied methods are existing, generic tensor completion methods, and are adapted here to recover deep feature tensor data, while the fourth method is newly developed specifically for deep feature tensor completion. Simulation studies show that the new method is 3-18 times faster than the other three methods, which is an important consideration in collaborative intelligence. For VGG16's sparse tensors, all methods produce statistically equivalent classification results across all loss levels tested. For ResNet34's non-sparse tensors, the new method offers statistically better classification accuracy (by 0.25%-6.30%) compared to other methods for matched execution speeds, and second-best accuracy among the four methods when they are allowed to run until convergence.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2020.2977050</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Accuracy ; Artificial intelligence ; Artificial neural networks ; Classification ; Cloud computing ; Collaboration ; Collaborative intelligence ; Computational modeling ; deep feature transmission ; deep learning ; Mathematical analysis ; Matrix converters ; Matrix decomposition ; Missing data ; missing data imputation ; Production methods ; Statistical methods ; Tensile stress ; Tensor completion ; tensor decomposition ; tensor reconstruction ; Tensors ; Three-dimensional displays</subject><ispartof>IEEE access, 2020, Vol.8, p.41162-41174</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2020</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-bc6976178c53e14a96a290970fec00b14278ac9e605d579e4f9965f3d08fe00a3</citedby><cites>FETCH-LOGICAL-c408t-bc6976178c53e14a96a290970fec00b14278ac9e605d579e4f9965f3d08fe00a3</cites><orcidid>0000-0002-7617-4475</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9017944$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2102,4024,27633,27923,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Bragilevsky, Lior</creatorcontrib><creatorcontrib>Bajic, Ivan V.</creatorcontrib><title>Tensor Completion Methods for Collaborative Intelligence</title><title>IEEE access</title><addtitle>Access</addtitle><description>In the race to bring Artificial Intelligence (AI) to the edge, collaborative intelligence has emerged as a promising way to lighten the computation load on edge devices that run applications based on Deep Neural Networks (DNNs). Typically, a deep model is split at a given layer into edge and cloud sub-models. The deep feature tensor produced by the edge sub-model is transmitted to the cloud, where the remaining computationally intensive workload is performed by the cloud sub-model. The communication channel between the edge and cloud is imperfect, which will result in missing data in the deep feature tensor received at the cloud side, an issue that has mostly been ignored by existing literature on the topic. In this paper we study four methods for recovering missing data in the deep feature tensor. Three of the studied methods are existing, generic tensor completion methods, and are adapted here to recover deep feature tensor data, while the fourth method is newly developed specifically for deep feature tensor completion. Simulation studies show that the new method is 3-18 times faster than the other three methods, which is an important consideration in collaborative intelligence. For VGG16's sparse tensors, all methods produce statistically equivalent classification results across all loss levels tested. For ResNet34's non-sparse tensors, the new method offers statistically better classification accuracy (by 0.25%-6.30%) compared to other methods for matched execution speeds, and second-best accuracy among the four methods when they are allowed to run until convergence.</description><subject>Accuracy</subject><subject>Artificial intelligence</subject><subject>Artificial neural networks</subject><subject>Classification</subject><subject>Cloud computing</subject><subject>Collaboration</subject><subject>Collaborative intelligence</subject><subject>Computational modeling</subject><subject>deep feature transmission</subject><subject>deep learning</subject><subject>Mathematical analysis</subject><subject>Matrix converters</subject><subject>Matrix decomposition</subject><subject>Missing data</subject><subject>missing data imputation</subject><subject>Production methods</subject><subject>Statistical methods</subject><subject>Tensile stress</subject><subject>Tensor completion</subject><subject>tensor decomposition</subject><subject>tensor reconstruction</subject><subject>Tensors</subject><subject>Three-dimensional displays</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkE9rwkAQxUNpodL6CbwEetbOZv8fJdhWsPSgPS-bzayNxKzdxEK_faMR6VxmeMx7M_ySZEJgRgjo53meL9brWQYZzDItJXC4SUYZEXpKORW3_-b7ZNy2O-hL9RKXo0RtsGlDTPOwP9TYVaFJ37H7CmWb-rNc17YI0XbVD6bLpsO6rrbYOHxM7rytWxxf-kPy-bLY5G_T1cfrMp-vpo6B6qaFE1oKIpXjFAmzWthMg5bg0QEUhGVSWadRAC-51Mi81oJ7WoLyCGDpQ7Iccstgd-YQq72NvybYypyFELfGxq5yNRrlCwdCO2-BMaZlQZXVrBAFI0SogvZZT0PWIYbvI7ad2YVjbPr3TcY4k1woIfstOmy5GNo2or9eJWBOxM1A3JyImwvx3jUZXBUiXh0aiNSM0T8iKnqw</recordid><startdate>2020</startdate><enddate>2020</enddate><creator>Bragilevsky, Lior</creator><creator>Bajic, Ivan V.</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7617-4475</orcidid></search><sort><creationdate>2020</creationdate><title>Tensor Completion Methods for Collaborative Intelligence</title><author>Bragilevsky, Lior ; Bajic, Ivan V.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-bc6976178c53e14a96a290970fec00b14278ac9e605d579e4f9965f3d08fe00a3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Accuracy</topic><topic>Artificial intelligence</topic><topic>Artificial neural networks</topic><topic>Classification</topic><topic>Cloud computing</topic><topic>Collaboration</topic><topic>Collaborative intelligence</topic><topic>Computational modeling</topic><topic>deep feature transmission</topic><topic>deep learning</topic><topic>Mathematical analysis</topic><topic>Matrix converters</topic><topic>Matrix decomposition</topic><topic>Missing data</topic><topic>missing data imputation</topic><topic>Production methods</topic><topic>Statistical methods</topic><topic>Tensile stress</topic><topic>Tensor completion</topic><topic>tensor decomposition</topic><topic>tensor reconstruction</topic><topic>Tensors</topic><topic>Three-dimensional displays</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Bragilevsky, Lior</creatorcontrib><creatorcontrib>Bajic, Ivan V.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library Online</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Bragilevsky, Lior</au><au>Bajic, Ivan V.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Tensor Completion Methods for Collaborative Intelligence</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2020</date><risdate>2020</risdate><volume>8</volume><spage>41162</spage><epage>41174</epage><pages>41162-41174</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>In the race to bring Artificial Intelligence (AI) to the edge, collaborative intelligence has emerged as a promising way to lighten the computation load on edge devices that run applications based on Deep Neural Networks (DNNs). Typically, a deep model is split at a given layer into edge and cloud sub-models. The deep feature tensor produced by the edge sub-model is transmitted to the cloud, where the remaining computationally intensive workload is performed by the cloud sub-model. The communication channel between the edge and cloud is imperfect, which will result in missing data in the deep feature tensor received at the cloud side, an issue that has mostly been ignored by existing literature on the topic. In this paper we study four methods for recovering missing data in the deep feature tensor. Three of the studied methods are existing, generic tensor completion methods, and are adapted here to recover deep feature tensor data, while the fourth method is newly developed specifically for deep feature tensor completion. Simulation studies show that the new method is 3-18 times faster than the other three methods, which is an important consideration in collaborative intelligence. For VGG16's sparse tensors, all methods produce statistically equivalent classification results across all loss levels tested. For ResNet34's non-sparse tensors, the new method offers statistically better classification accuracy (by 0.25%-6.30%) compared to other methods for matched execution speeds, and second-best accuracy among the four methods when they are allowed to run until convergence.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2020.2977050</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0002-7617-4475</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2020, Vol.8, p.41162-41174
issn 2169-3536
2169-3536
language eng
recordid cdi_proquest_journals_2454756867
source DOAJ Directory of Open Access Journals; IEEE Xplore Open Access Journals; EZB Electronic Journals Library
subjects Accuracy
Artificial intelligence
Artificial neural networks
Classification
Cloud computing
Collaboration
Collaborative intelligence
Computational modeling
deep feature transmission
deep learning
Mathematical analysis
Matrix converters
Matrix decomposition
Missing data
missing data imputation
Production methods
Statistical methods
Tensile stress
Tensor completion
tensor decomposition
tensor reconstruction
Tensors
Three-dimensional displays
title Tensor Completion Methods for Collaborative Intelligence
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-20T21%3A17%3A57IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Tensor%20Completion%20Methods%20for%20Collaborative%20Intelligence&rft.jtitle=IEEE%20access&rft.au=Bragilevsky,%20Lior&rft.date=2020&rft.volume=8&rft.spage=41162&rft.epage=41174&rft.pages=41162-41174&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2020.2977050&rft_dat=%3Cproquest_ieee_%3E2454756867%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2454756867&rft_id=info:pmid/&rft_ieee_id=9017944&rft_doaj_id=oai_doaj_org_article_8fbc069cfa044497b38a94b6b41168b3&rfr_iscdi=true