DeepliteRT: Computer Vision at the Edge
The proliferation of edge devices has unlocked unprecedented opportunities for deep learning model deployment in computer vision applications. However, these complex models require considerable power, memory and compute resources that are typically not available on edge platforms. Ultra low-bit quan...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Ashfaq, Saad Hoffman, Alexander Mitra, Saptarshi Sah, Sudhakar AskariHemmat, MohammadHossein Saboori, Ehsan |
description | The proliferation of edge devices has unlocked unprecedented opportunities
for deep learning model deployment in computer vision applications. However,
these complex models require considerable power, memory and compute resources
that are typically not available on edge platforms. Ultra low-bit quantization
presents an attractive solution to this problem by scaling down the model
weights and activations from 32-bit to less than 8-bit. We implement highly
optimized ultra low-bit convolution operators for ARM-based targets that
outperform existing methods by up to 4.34x. Our operator is implemented within
Deeplite Runtime (DeepliteRT), an end-to-end solution for the compilation,
tuning, and inference of ultra low-bit models on ARM devices. Compiler passes
in DeepliteRT automatically convert a fake-quantized model in full precision to
a compact ultra low-bit representation, easing the process of quantized model
deployment on commodity hardware. We analyze the performance of DeepliteRT on
classification and detection models against optimized 32-bit floating-point,
8-bit integer, and 2-bit baselines, achieving significant speedups of up to
2.20x, 2.33x and 2.17x, respectively. |
doi_str_mv | 10.48550/arxiv.2309.10878 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2309_10878</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2309_10878</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-cb670473a802adc1759753e09bfec57917f7292ea9e166fd2b7eb43f59ae2c143</originalsourceid><addsrcrecordid>eNotzruKwkAUgOFpLBb1AbZyOqvEuWRyZuwkXlYQBAm24SQ5owNeQoyy-_aia_V3Px9j31LEiTVGTLD9DY9YaeFiKSzYLzaeEzWn0NEun_Lsem7uHbV8H27heuHY8e5IfFEfaMB6Hk83Gn7aZ_lykWc_0Wa7WmezTYQp2KgqUxAJaLRCYV1JMA6MJuFKT5UBJ8GDcorQkUxTX6sSqEy0Nw5JVTLRfTb6376lRdOGM7Z_xUtcvMX6CR1MOc8</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>DeepliteRT: Computer Vision at the Edge</title><source>arXiv.org</source><creator>Ashfaq, Saad ; Hoffman, Alexander ; Mitra, Saptarshi ; Sah, Sudhakar ; AskariHemmat, MohammadHossein ; Saboori, Ehsan</creator><creatorcontrib>Ashfaq, Saad ; Hoffman, Alexander ; Mitra, Saptarshi ; Sah, Sudhakar ; AskariHemmat, MohammadHossein ; Saboori, Ehsan</creatorcontrib><description>The proliferation of edge devices has unlocked unprecedented opportunities
for deep learning model deployment in computer vision applications. However,
these complex models require considerable power, memory and compute resources
that are typically not available on edge platforms. Ultra low-bit quantization
presents an attractive solution to this problem by scaling down the model
weights and activations from 32-bit to less than 8-bit. We implement highly
optimized ultra low-bit convolution operators for ARM-based targets that
outperform existing methods by up to 4.34x. Our operator is implemented within
Deeplite Runtime (DeepliteRT), an end-to-end solution for the compilation,
tuning, and inference of ultra low-bit models on ARM devices. Compiler passes
in DeepliteRT automatically convert a fake-quantized model in full precision to
a compact ultra low-bit representation, easing the process of quantized model
deployment on commodity hardware. We analyze the performance of DeepliteRT on
classification and detection models against optimized 32-bit floating-point,
8-bit integer, and 2-bit baselines, achieving significant speedups of up to
2.20x, 2.33x and 2.17x, respectively.</description><identifier>DOI: 10.48550/arxiv.2309.10878</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2023-09</creationdate><rights>http://creativecommons.org/licenses/by-nc-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2309.10878$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2309.10878$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Ashfaq, Saad</creatorcontrib><creatorcontrib>Hoffman, Alexander</creatorcontrib><creatorcontrib>Mitra, Saptarshi</creatorcontrib><creatorcontrib>Sah, Sudhakar</creatorcontrib><creatorcontrib>AskariHemmat, MohammadHossein</creatorcontrib><creatorcontrib>Saboori, Ehsan</creatorcontrib><title>DeepliteRT: Computer Vision at the Edge</title><description>The proliferation of edge devices has unlocked unprecedented opportunities
for deep learning model deployment in computer vision applications. However,
these complex models require considerable power, memory and compute resources
that are typically not available on edge platforms. Ultra low-bit quantization
presents an attractive solution to this problem by scaling down the model
weights and activations from 32-bit to less than 8-bit. We implement highly
optimized ultra low-bit convolution operators for ARM-based targets that
outperform existing methods by up to 4.34x. Our operator is implemented within
Deeplite Runtime (DeepliteRT), an end-to-end solution for the compilation,
tuning, and inference of ultra low-bit models on ARM devices. Compiler passes
in DeepliteRT automatically convert a fake-quantized model in full precision to
a compact ultra low-bit representation, easing the process of quantized model
deployment on commodity hardware. We analyze the performance of DeepliteRT on
classification and detection models against optimized 32-bit floating-point,
8-bit integer, and 2-bit baselines, achieving significant speedups of up to
2.20x, 2.33x and 2.17x, respectively.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzruKwkAUgOFpLBb1AbZyOqvEuWRyZuwkXlYQBAm24SQ5owNeQoyy-_aia_V3Px9j31LEiTVGTLD9DY9YaeFiKSzYLzaeEzWn0NEun_Lsem7uHbV8H27heuHY8e5IfFEfaMB6Hk83Gn7aZ_lykWc_0Wa7WmezTYQp2KgqUxAJaLRCYV1JMA6MJuFKT5UBJ8GDcorQkUxTX6sSqEy0Nw5JVTLRfTb6376lRdOGM7Z_xUtcvMX6CR1MOc8</recordid><startdate>20230919</startdate><enddate>20230919</enddate><creator>Ashfaq, Saad</creator><creator>Hoffman, Alexander</creator><creator>Mitra, Saptarshi</creator><creator>Sah, Sudhakar</creator><creator>AskariHemmat, MohammadHossein</creator><creator>Saboori, Ehsan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230919</creationdate><title>DeepliteRT: Computer Vision at the Edge</title><author>Ashfaq, Saad ; Hoffman, Alexander ; Mitra, Saptarshi ; Sah, Sudhakar ; AskariHemmat, MohammadHossein ; Saboori, Ehsan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-cb670473a802adc1759753e09bfec57917f7292ea9e166fd2b7eb43f59ae2c143</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Ashfaq, Saad</creatorcontrib><creatorcontrib>Hoffman, Alexander</creatorcontrib><creatorcontrib>Mitra, Saptarshi</creatorcontrib><creatorcontrib>Sah, Sudhakar</creatorcontrib><creatorcontrib>AskariHemmat, MohammadHossein</creatorcontrib><creatorcontrib>Saboori, Ehsan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Ashfaq, Saad</au><au>Hoffman, Alexander</au><au>Mitra, Saptarshi</au><au>Sah, Sudhakar</au><au>AskariHemmat, MohammadHossein</au><au>Saboori, Ehsan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DeepliteRT: Computer Vision at the Edge</atitle><date>2023-09-19</date><risdate>2023</risdate><abstract>The proliferation of edge devices has unlocked unprecedented opportunities
for deep learning model deployment in computer vision applications. However,
these complex models require considerable power, memory and compute resources
that are typically not available on edge platforms. Ultra low-bit quantization
presents an attractive solution to this problem by scaling down the model
weights and activations from 32-bit to less than 8-bit. We implement highly
optimized ultra low-bit convolution operators for ARM-based targets that
outperform existing methods by up to 4.34x. Our operator is implemented within
Deeplite Runtime (DeepliteRT), an end-to-end solution for the compilation,
tuning, and inference of ultra low-bit models on ARM devices. Compiler passes
in DeepliteRT automatically convert a fake-quantized model in full precision to
a compact ultra low-bit representation, easing the process of quantized model
deployment on commodity hardware. We analyze the performance of DeepliteRT on
classification and detection models against optimized 32-bit floating-point,
8-bit integer, and 2-bit baselines, achieving significant speedups of up to
2.20x, 2.33x and 2.17x, respectively.</abstract><doi>10.48550/arxiv.2309.10878</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2309.10878 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2309_10878 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | DeepliteRT: Computer Vision at the Edge |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-25T21%3A16%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DeepliteRT:%20Computer%20Vision%20at%20the%20Edge&rft.au=Ashfaq,%20Saad&rft.date=2023-09-19&rft_id=info:doi/10.48550/arxiv.2309.10878&rft_dat=%3Carxiv_GOX%3E2309_10878%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |