OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks
Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structu...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2019-07 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Li, Jiashi Qi, Qi Wang, Jingyu Ge, Ce Li, Yujian Zhangzhang Yue Sun, Haifeng |
description | Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2232267162</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2232267162</sourcerecordid><originalsourceid>FETCH-proquest_journals_22322671623</originalsourceid><addsrcrecordid>eNqNjr0KwjAYAIMgWLTvEHAOtF9sFdeo2KWFtnsJkmprTGJ-EH16O_gATjfcDTdDEVCakt0GYIFi58YkSSDfQpbRCJVVwZp6j6vgSaEIu3GlhMSN4dYN_o1rcQ2S2-HD_aAV7rXFTD8Mv3h8EMLgUgTL5QT_0vbuVmjec-lE_OMSrU_Hlp2JsfoZhPPdqINVk-oAKEwbaQ70v-oLRIQ9UA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2232267162</pqid></control><display><type>article</type><title>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</title><source>Free E- Journals</source><creator>Li, Jiashi ; Qi, Qi ; Wang, Jingyu ; Ge, Ce ; Li, Yujian ; Zhangzhang Yue ; Sun, Haifeng</creator><creatorcontrib>Li, Jiashi ; Qi, Qi ; Wang, Jingyu ; Ge, Ce ; Li, Yujian ; Zhangzhang Yue ; Sun, Haifeng</creatorcontrib><description>Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Channels ; Greedy algorithms ; Iterative methods ; Model accuracy ; Neural networks ; Pruning ; Regularization ; Sparsity ; Training</subject><ispartof>arXiv.org, 2019-07</ispartof><rights>2019. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>781,785</link.rule.ids></links><search><creatorcontrib>Li, Jiashi</creatorcontrib><creatorcontrib>Qi, Qi</creatorcontrib><creatorcontrib>Wang, Jingyu</creatorcontrib><creatorcontrib>Ge, Ce</creatorcontrib><creatorcontrib>Li, Yujian</creatorcontrib><creatorcontrib>Zhangzhang Yue</creatorcontrib><creatorcontrib>Sun, Haifeng</creatorcontrib><title>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</title><title>arXiv.org</title><description>Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.</description><subject>Channels</subject><subject>Greedy algorithms</subject><subject>Iterative methods</subject><subject>Model accuracy</subject><subject>Neural networks</subject><subject>Pruning</subject><subject>Regularization</subject><subject>Sparsity</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNjr0KwjAYAIMgWLTvEHAOtF9sFdeo2KWFtnsJkmprTGJ-EH16O_gATjfcDTdDEVCakt0GYIFi58YkSSDfQpbRCJVVwZp6j6vgSaEIu3GlhMSN4dYN_o1rcQ2S2-HD_aAV7rXFTD8Mv3h8EMLgUgTL5QT_0vbuVmjec-lE_OMSrU_Hlp2JsfoZhPPdqINVk-oAKEwbaQ70v-oLRIQ9UA</recordid><startdate>20190701</startdate><enddate>20190701</enddate><creator>Li, Jiashi</creator><creator>Qi, Qi</creator><creator>Wang, Jingyu</creator><creator>Ge, Ce</creator><creator>Li, Yujian</creator><creator>Zhangzhang Yue</creator><creator>Sun, Haifeng</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20190701</creationdate><title>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</title><author>Li, Jiashi ; Qi, Qi ; Wang, Jingyu ; Ge, Ce ; Li, Yujian ; Zhangzhang Yue ; Sun, Haifeng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_22322671623</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Channels</topic><topic>Greedy algorithms</topic><topic>Iterative methods</topic><topic>Model accuracy</topic><topic>Neural networks</topic><topic>Pruning</topic><topic>Regularization</topic><topic>Sparsity</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Jiashi</creatorcontrib><creatorcontrib>Qi, Qi</creatorcontrib><creatorcontrib>Wang, Jingyu</creatorcontrib><creatorcontrib>Ge, Ce</creatorcontrib><creatorcontrib>Li, Yujian</creatorcontrib><creatorcontrib>Zhangzhang Yue</creatorcontrib><creatorcontrib>Sun, Haifeng</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Jiashi</au><au>Qi, Qi</au><au>Wang, Jingyu</au><au>Ge, Ce</au><au>Li, Yujian</au><au>Zhangzhang Yue</au><au>Sun, Haifeng</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks</atitle><jtitle>arXiv.org</jtitle><date>2019-07-01</date><risdate>2019</risdate><eissn>2331-8422</eissn><abstract>Channel pruning can significantly accelerate and compress deep neural networks. Many channel pruning works utilize structured sparsity regularization to zero out all the weights in some channels and automatically obtain structure-sparse network in training stage. However, these methods apply structured sparsity regularization on each layer separately where the correlations between consecutive layers are omitted. In this paper, we first combine one out-channel in current layer and the corresponding in-channel in next layer as a regularization group, namely out-in-channel. Our proposed Out-In-Channel Sparsity Regularization (OICSR) considers correlations between successive layers to further retain predictive power of the compact network. Training with OICSR thoroughly transfers discriminative features into a fraction of out-in-channels. Correspondingly, OICSR measures channel importance based on statistics computed from two consecutive layers, not individual layer. Finally, a global greedy pruning algorithm is designed to remove redundant out-in-channels in an iterative way. Our method is comprehensively evaluated with various CNN architectures including CifarNet, AlexNet, ResNet, DenseNet and PreActSeNet on CIFAR-10, CIFAR-100 and ImageNet-1K datasets. Notably, on ImageNet-1K, we reduce 37.2% FLOPs on ResNet-50 while outperforming the original model by 0.22% top-1 accuracy.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2019-07 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2232267162 |
source | Free E- Journals |
subjects | Channels Greedy algorithms Iterative methods Model accuracy Neural networks Pruning Regularization Sparsity Training |
title | OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural Networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T10%3A43%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=OICSR:%20Out-In-Channel%20Sparsity%20Regularization%20for%20Compact%20Deep%20Neural%20Networks&rft.jtitle=arXiv.org&rft.au=Li,%20Jiashi&rft.date=2019-07-01&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2232267162%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2232267162&rft_id=info:pmid/&rfr_iscdi=true |