OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks

Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structu...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2019-07
Hauptverfasser: Li, Jiashi, Qi, Qi, Wang, Jingyu, Ge, Ce, Li, Yujian, Zhangzhang Yue, Sun, Haifeng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Li, Jiashi
Qi, Qi
Wang, Jingyu
Ge, Ce
Li, Yujian
Zhangzhang Yue
Sun, Haifeng
description Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2232267162</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2232267162</sourcerecordid><originalsourceid>FETCH-proquest_journals_22322671623</originalsourceid><addsrcrecordid>eNqNjr0KwjAYAIMgWLTvEHAOtF9sFdeo2KWFtnsJkmprTGJ-EH16O_gATjfcDTdDEVCakt0GYIFi58YkSSDfQpbRCJVVwZp6j6vgSaEIu3GlhMSN4dYN_o1rcQ2S2-HD_aAV7rXFTD8Mv3h8EMLgUgTL5QT_0vbuVmjec-lE_OMSrU_Hlp2JsfoZhPPdqINVk-oAKEwbaQ70v-oLRIQ9UA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2232267162</pqid></control><display><type>article</type><title>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</title><source>Free E- Journals</source><creator>Li, Jiashi ; Qi, Qi ; Wang, Jingyu ; Ge, Ce ; Li, Yujian ; Zhangzhang Yue ; Sun, Haifeng</creator><creatorcontrib>Li, Jiashi ; Qi, Qi ; Wang, Jingyu ; Ge, Ce ; Li, Yujian ; Zhangzhang Yue ; Sun, Haifeng</creatorcontrib><description>Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Channels ; Greedy algorithms ; Iterative methods ; Model accuracy ; Neural networks ; Pruning ; Regularization ; Sparsity ; Training</subject><ispartof>arXiv.org, 2019-07</ispartof><rights>2019. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>781,785</link.rule.ids></links><search><creatorcontrib>Li, Jiashi</creatorcontrib><creatorcontrib>Qi, Qi</creatorcontrib><creatorcontrib>Wang, Jingyu</creatorcontrib><creatorcontrib>Ge, Ce</creatorcontrib><creatorcontrib>Li, Yujian</creatorcontrib><creatorcontrib>Zhangzhang Yue</creatorcontrib><creatorcontrib>Sun, Haifeng</creatorcontrib><title>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</title><title>arXiv.org</title><description>Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.</description><subject>Channels</subject><subject>Greedy algorithms</subject><subject>Iterative methods</subject><subject>Model accuracy</subject><subject>Neural networks</subject><subject>Pruning</subject><subject>Regularization</subject><subject>Sparsity</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNjr0KwjAYAIMgWLTvEHAOtF9sFdeo2KWFtnsJkmprTGJ-EH16O_gATjfcDTdDEVCakt0GYIFi58YkSSDfQpbRCJVVwZp6j6vgSaEIu3GlhMSN4dYN_o1rcQ2S2-HD_aAV7rXFTD8Mv3h8EMLgUgTL5QT_0vbuVmjec-lE_OMSrU_Hlp2JsfoZhPPdqINVk-oAKEwbaQ70v-oLRIQ9UA</recordid><startdate>20190701</startdate><enddate>20190701</enddate><creator>Li, Jiashi</creator><creator>Qi, Qi</creator><creator>Wang, Jingyu</creator><creator>Ge, Ce</creator><creator>Li, Yujian</creator><creator>Zhangzhang Yue</creator><creator>Sun, Haifeng</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20190701</creationdate><title>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</title><author>Li, Jiashi ; Qi, Qi ; Wang, Jingyu ; Ge, Ce ; Li, Yujian ; Zhangzhang Yue ; Sun, Haifeng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_22322671623</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Channels</topic><topic>Greedy algorithms</topic><topic>Iterative methods</topic><topic>Model accuracy</topic><topic>Neural networks</topic><topic>Pruning</topic><topic>Regularization</topic><topic>Sparsity</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Jiashi</creatorcontrib><creatorcontrib>Qi, Qi</creatorcontrib><creatorcontrib>Wang, Jingyu</creatorcontrib><creatorcontrib>Ge, Ce</creatorcontrib><creatorcontrib>Li, Yujian</creatorcontrib><creatorcontrib>Zhangzhang Yue</creatorcontrib><creatorcontrib>Sun, Haifeng</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Jiashi</au><au>Qi, Qi</au><au>Wang, Jingyu</au><au>Ge, Ce</au><au>Li, Yujian</au><au>Zhangzhang Yue</au><au>Sun, Haifeng</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</atitle><jtitle>arXiv.org</jtitle><date>2019-07-01</date><risdate>2019</risdate><eissn>2331-8422</eissn><abstract>Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2019-07
issn 2331-8422
language eng
recordid cdi_proquest_journals_2232267162
source Free E- Journals
subjects Channels
Greedy algorithms
Iterative methods
Model accuracy
Neural networks
Pruning
Regularization
Sparsity
Training
title OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T10%3A43%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=OICSR:%20Out-In-Channel%20Sparsity%20Regularization%20for%20Compact%20Deep%20Neural%20Networks&rft.jtitle=arXiv.org&rft.au=Li,%20Jiashi&rft.date=2019-07-01&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2232267162%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2232267162&rft_id=info:pmid/&rfr_iscdi=true