RISC-NN: Use RISC, NOT CISC as Neural Network Hardware Infrastructure
Neural Networks (NN) have been proven to be powerful tools to analyze Big Data. However, traditional CPUs cannot achieve the desired performance and/or energy efficiency for NN applications. Therefore, numerous NN accelerators have been used or designed to meet these goals. These accelerators all fa...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2021-03 |
---|---|
Hauptverfasser: | , , , , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Taoran Xiang Zhang, Lunkai An, Shuqian Ye, Xiaochun Zhang, Mingzhe Liu, Yanhuan Yan, Mingyu Wang, Da Zhang, Hao Li, Wenming Sun, Ninghui Fan, Dongrui |
description | Neural Networks (NN) have been proven to be powerful tools to analyze Big Data. However, traditional CPUs cannot achieve the desired performance and/or energy efficiency for NN applications. Therefore, numerous NN accelerators have been used or designed to meet these goals. These accelerators all fall into three categories: GPGPUs, ASIC NN Accelerators and CISC NN Accelerators. Though CISC NN Accelerators can achieve considerable smaller memory footprint than GPGPU thus improve energy efficiency; they still fail to provide same level of data reuse optimization achieved by ASIC NN Accelerators because of the inherited poor pragrammability of their CISC architecture. We argue that, for NN Accelerators, RISC is a better design choice than CISC, as is the case with general purpose processors. We propose RISC-NN, a novel many-core RISC-based NN accelerator that achieves high expressiveness and high parallelism and features strong programmability and low control-hardware costs. We show that, RISC-NN can implement all the necessary instructions of state-of-the-art CISC NN Accelerators; in the meantime, RISC-NN manages to achieve advanced optimization such as multiple-level data reuse and support for Sparse NN applications which previously only existed in ASIC NN Accelerators. Experiment results show that, RISC-NN achieves on average 11.88X performance efficiency compared with state-of-the-art Nvidia TITAN Xp GPGPU for various NN applications. RISC-NN also achieves on average 1.29X, 8.37X and 21.71X performance efficiency over CISC-based TPU in CNN, MLP and LSTM applications, respectively. Finally, RISC-NN can achieve additional 26.05% performance improvement and 33.13% energy reduction after applying pruning for Sparse NN applications. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2504621093</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2504621093</sourcerecordid><originalsourceid>FETCH-proquest_journals_25046210933</originalsourceid><addsrcrecordid>eNqNis0KgkAURocgSMp3uNA2Ybyj9rMVwzYTlK1lqHFhonXvDL5-Bj1Aq_MdvjMTASoVR7sEcSFC5lZKidkW01QForicrnmk9QFubOErG9DnCvJpgWHQ1pPpJrhxoCeUhh6jIQunviHDjvzdebIrMW9Mxzb8cSnWx6LKy-hFw9tbdnU7eOqnq8ZUJhnGcq_Uf9UH4Gw4Lw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2504621093</pqid></control><display><type>article</type><title>RISC-NN: Use RISC, NOT CISC as Neural Network Hardware Infrastructure</title><source>Free E- Journals</source><creator>Taoran Xiang ; Zhang, Lunkai ; An, Shuqian ; Ye, Xiaochun ; Zhang, Mingzhe ; Liu, Yanhuan ; Yan, Mingyu ; Wang, Da ; Zhang, Hao ; Li, Wenming ; Sun, Ninghui ; Fan, Dongrui</creator><creatorcontrib>Taoran Xiang ; Zhang, Lunkai ; An, Shuqian ; Ye, Xiaochun ; Zhang, Mingzhe ; Liu, Yanhuan ; Yan, Mingyu ; Wang, Da ; Zhang, Hao ; Li, Wenming ; Sun, Ninghui ; Fan, Dongrui</creatorcontrib><description>Neural Networks (NN) have been proven to be powerful tools to analyze Big Data. However, traditional CPUs cannot achieve the desired performance and/or energy efficiency for NN applications. Therefore, numerous NN accelerators have been used or designed to meet these goals. These accelerators all fall into three categories: GPGPUs, ASIC NN Accelerators and CISC NN Accelerators. Though CISC NN Accelerators can achieve considerable smaller memory footprint than GPGPU thus improve energy efficiency; they still fail to provide same level of data reuse optimization achieved by ASIC NN Accelerators because of the inherited poor pragrammability of their CISC architecture. We argue that, for NN Accelerators, RISC is a better design choice than CISC, as is the case with general purpose processors. We propose RISC-NN, a novel many-core RISC-based NN accelerator that achieves high expressiveness and high parallelism and features strong programmability and low control-hardware costs. We show that, RISC-NN can implement all the necessary instructions of state-of-the-art CISC NN Accelerators; in the meantime, RISC-NN manages to achieve advanced optimization such as multiple-level data reuse and support for Sparse NN applications which previously only existed in ASIC NN Accelerators. Experiment results show that, RISC-NN achieves on average 11.88X performance efficiency compared with state-of-the-art Nvidia TITAN Xp GPGPU for various NN applications. RISC-NN also achieves on average 1.29X, 8.37X and 21.71X performance efficiency over CISC-based TPU in CNN, MLP and LSTM applications, respectively. Finally, RISC-NN can achieve additional 26.05% performance improvement and 33.13% energy reduction after applying pruning for Sparse NN applications.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Accelerators ; Energy efficiency ; Hardware ; Neural networks ; Optimization</subject><ispartof>arXiv.org, 2021-03</ispartof><rights>2021. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Taoran Xiang</creatorcontrib><creatorcontrib>Zhang, Lunkai</creatorcontrib><creatorcontrib>An, Shuqian</creatorcontrib><creatorcontrib>Ye, Xiaochun</creatorcontrib><creatorcontrib>Zhang, Mingzhe</creatorcontrib><creatorcontrib>Liu, Yanhuan</creatorcontrib><creatorcontrib>Yan, Mingyu</creatorcontrib><creatorcontrib>Wang, Da</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Li, Wenming</creatorcontrib><creatorcontrib>Sun, Ninghui</creatorcontrib><creatorcontrib>Fan, Dongrui</creatorcontrib><title>RISC-NN: Use RISC, NOT CISC as Neural Network Hardware Infrastructure</title><title>arXiv.org</title><description>Neural Networks (NN) have been proven to be powerful tools to analyze Big Data. However, traditional CPUs cannot achieve the desired performance and/or energy efficiency for NN applications. Therefore, numerous NN accelerators have been used or designed to meet these goals. These accelerators all fall into three categories: GPGPUs, ASIC NN Accelerators and CISC NN Accelerators. Though CISC NN Accelerators can achieve considerable smaller memory footprint than GPGPU thus improve energy efficiency; they still fail to provide same level of data reuse optimization achieved by ASIC NN Accelerators because of the inherited poor pragrammability of their CISC architecture. We argue that, for NN Accelerators, RISC is a better design choice than CISC, as is the case with general purpose processors. We propose RISC-NN, a novel many-core RISC-based NN accelerator that achieves high expressiveness and high parallelism and features strong programmability and low control-hardware costs. We show that, RISC-NN can implement all the necessary instructions of state-of-the-art CISC NN Accelerators; in the meantime, RISC-NN manages to achieve advanced optimization such as multiple-level data reuse and support for Sparse NN applications which previously only existed in ASIC NN Accelerators. Experiment results show that, RISC-NN achieves on average 11.88X performance efficiency compared with state-of-the-art Nvidia TITAN Xp GPGPU for various NN applications. RISC-NN also achieves on average 1.29X, 8.37X and 21.71X performance efficiency over CISC-based TPU in CNN, MLP and LSTM applications, respectively. Finally, RISC-NN can achieve additional 26.05% performance improvement and 33.13% energy reduction after applying pruning for Sparse NN applications.</description><subject>Accelerators</subject><subject>Energy efficiency</subject><subject>Hardware</subject><subject>Neural networks</subject><subject>Optimization</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNqNis0KgkAURocgSMp3uNA2Ybyj9rMVwzYTlK1lqHFhonXvDL5-Bj1Aq_MdvjMTASoVR7sEcSFC5lZKidkW01QForicrnmk9QFubOErG9DnCvJpgWHQ1pPpJrhxoCeUhh6jIQunviHDjvzdebIrMW9Mxzb8cSnWx6LKy-hFw9tbdnU7eOqnq8ZUJhnGcq_Uf9UH4Gw4Lw</recordid><startdate>20210323</startdate><enddate>20210323</enddate><creator>Taoran Xiang</creator><creator>Zhang, Lunkai</creator><creator>An, Shuqian</creator><creator>Ye, Xiaochun</creator><creator>Zhang, Mingzhe</creator><creator>Liu, Yanhuan</creator><creator>Yan, Mingyu</creator><creator>Wang, Da</creator><creator>Zhang, Hao</creator><creator>Li, Wenming</creator><creator>Sun, Ninghui</creator><creator>Fan, Dongrui</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20210323</creationdate><title>RISC-NN: Use RISC, NOT CISC as Neural Network Hardware Infrastructure</title><author>Taoran Xiang ; Zhang, Lunkai ; An, Shuqian ; Ye, Xiaochun ; Zhang, Mingzhe ; Liu, Yanhuan ; Yan, Mingyu ; Wang, Da ; Zhang, Hao ; Li, Wenming ; Sun, Ninghui ; Fan, Dongrui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_25046210933</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Accelerators</topic><topic>Energy efficiency</topic><topic>Hardware</topic><topic>Neural networks</topic><topic>Optimization</topic><toplevel>online_resources</toplevel><creatorcontrib>Taoran Xiang</creatorcontrib><creatorcontrib>Zhang, Lunkai</creatorcontrib><creatorcontrib>An, Shuqian</creatorcontrib><creatorcontrib>Ye, Xiaochun</creatorcontrib><creatorcontrib>Zhang, Mingzhe</creatorcontrib><creatorcontrib>Liu, Yanhuan</creatorcontrib><creatorcontrib>Yan, Mingyu</creatorcontrib><creatorcontrib>Wang, Da</creatorcontrib><creatorcontrib>Zhang, Hao</creatorcontrib><creatorcontrib>Li, Wenming</creatorcontrib><creatorcontrib>Sun, Ninghui</creatorcontrib><creatorcontrib>Fan, Dongrui</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Taoran Xiang</au><au>Zhang, Lunkai</au><au>An, Shuqian</au><au>Ye, Xiaochun</au><au>Zhang, Mingzhe</au><au>Liu, Yanhuan</au><au>Yan, Mingyu</au><au>Wang, Da</au><au>Zhang, Hao</au><au>Li, Wenming</au><au>Sun, Ninghui</au><au>Fan, Dongrui</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>RISC-NN: Use RISC, NOT CISC as Neural Network Hardware Infrastructure</atitle><jtitle>arXiv.org</jtitle><date>2021-03-23</date><risdate>2021</risdate><eissn>2331-8422</eissn><abstract>Neural Networks (NN) have been proven to be powerful tools to analyze Big Data. However, traditional CPUs cannot achieve the desired performance and/or energy efficiency for NN applications. Therefore, numerous NN accelerators have been used or designed to meet these goals. These accelerators all fall into three categories: GPGPUs, ASIC NN Accelerators and CISC NN Accelerators. Though CISC NN Accelerators can achieve considerable smaller memory footprint than GPGPU thus improve energy efficiency; they still fail to provide same level of data reuse optimization achieved by ASIC NN Accelerators because of the inherited poor pragrammability of their CISC architecture. We argue that, for NN Accelerators, RISC is a better design choice than CISC, as is the case with general purpose processors. We propose RISC-NN, a novel many-core RISC-based NN accelerator that achieves high expressiveness and high parallelism and features strong programmability and low control-hardware costs. We show that, RISC-NN can implement all the necessary instructions of state-of-the-art CISC NN Accelerators; in the meantime, RISC-NN manages to achieve advanced optimization such as multiple-level data reuse and support for Sparse NN applications which previously only existed in ASIC NN Accelerators. Experiment results show that, RISC-NN achieves on average 11.88X performance efficiency compared with state-of-the-art Nvidia TITAN Xp GPGPU for various NN applications. RISC-NN also achieves on average 1.29X, 8.37X and 21.71X performance efficiency over CISC-based TPU in CNN, MLP and LSTM applications, respectively. Finally, RISC-NN can achieve additional 26.05% performance improvement and 33.13% energy reduction after applying pruning for Sparse NN applications.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2021-03 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2504621093 |
source | Free E- Journals |
subjects | Accelerators Energy efficiency Hardware Neural networks Optimization |
title | RISC-NN: Use RISC, NOT CISC as Neural Network Hardware Infrastructure |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T20%3A44%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=RISC-NN:%20Use%20RISC,%20NOT%20CISC%20as%20Neural%20Network%20Hardware%20Infrastructure&rft.jtitle=arXiv.org&rft.au=Taoran%20Xiang&rft.date=2021-03-23&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2504621093%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2504621093&rft_id=info:pmid/&rfr_iscdi=true |