Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning

The success of convolutional neural networks (CNNs) in computer vision applications has been accompanied by a significant increase of computation and memory costs, which prohibits their usage on resource-limited environments, such as mobile systems or embedded devices. To this end, the research of C...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transaction on neural networks and learning systems 2020-02, Vol.31 (2), p.574-588
Hauptverfasser: Lin, Shaohui, Ji, Rongrong, Li, Yuchao, Deng, Cheng, Li, Xuelong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 588
container_issue 2
container_start_page 574
container_title IEEE transaction on neural networks and learning systems
container_volume 31
creator Lin, Shaohui
Ji, Rongrong
Li, Yuchao
Deng, Cheng
Li, Xuelong
description The success of convolutional neural networks (CNNs) in computer vision applications has been accompanied by a significant increase of computation and memory costs, which prohibits their usage on resource-limited environments, such as mobile systems or embedded devices. To this end, the research of CNN compression has recently become emerging. In this paper, we propose a novel filter pruning scheme, termed structured sparsity regularization (SSR), to simultaneously speed up the computation and reduce the memory overhead of CNNs, which can be well supported by various off-the-shelf deep learning libraries. Concretely, the proposed scheme incorporates two different regularizers of structured sparsity into the original objective function of filter pruning, which fully coordinates the global output and local pruning operations to adaptively prune filters. We further propose an alternative updating with Lagrange multipliers (AULM) scheme to efficiently solve its optimization. AULM follows the principle of alternating direction method of multipliers (ADMM) and alternates between promoting the structured sparsity of CNNs and optimizing the recognition loss, which leads to a very efficient solver ( 2.5\times to the most recent work that directly solves the group sparsity-based regularization). Moreover, by imposing the structured sparsity, the online inference is extremely memory-light since the number of filters and the output feature maps are simultaneously reduced. The proposed scheme has been deployed to a variety of state-of-the-art CNN structures, including LeNet, AlexNet, VGGNet, ResNet, and GoogLeNet, over different data sets. Quantitative results demonstrate that the proposed scheme achieves superior performance over the state-of-the-art methods. We further demonstrate the proposed compression scheme for the task of transfer learning, including domain adaptation and object detection, which also show exciting performance gains over the state-of-the-art filter pruning methods.
doi_str_mv 10.1109/TNNLS.2019.2906563
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_journals_2354615019</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>8689357</ieee_id><sourcerecordid>2210272646</sourcerecordid><originalsourceid>FETCH-LOGICAL-c466t-f4acb9357b1f4f6f5b088b17b748ea39247405bebfb33b3bf6a3a796d97c941f3</originalsourceid><addsrcrecordid>eNpdkEtLAzEQgIMoVmr_gIIsePGyNa_Nbo5SrBZKFVvBW0h2k7JlHzXJVuqvN7W1B-cyA_PNMPMBcIXgECHI7xez2XQ-xBDxIeaQJYycgAuMGI4xybLTY51-9MDAuRUMwWDCKD8HPQI5h5RmF2CyaL-kLaJRW69l7kNuNjPtXbQpZTT3tst9Z3U8X0vrSr-N3vSyq6Qtv3URjcvKaxu92q4pm-UlODOycnpwyH3wPn5cjJ7j6cvTZPQwjXPKmI8NlbniJEkVMtQwkyiYZQqlKqWZloRjmlKYKK2MIkQRZZgkMuWs4GnOKTKkD-72e9e2_ey086IuXa6rSja67ZzAGEGcYkZZQG__oau2s024TmCSUIaSoC9QeE_ltnXOaiPWtqyl3QoExc61-HUtdq7FwXUYujms7lSti-PIn9kAXO-BUmt9bGcs2_1OfgDAtoJy</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2354615019</pqid></control><display><type>article</type><title>Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning</title><source>IEEE Electronic Library (IEL)</source><creator>Lin, Shaohui ; Ji, Rongrong ; Li, Yuchao ; Deng, Cheng ; Li, Xuelong</creator><creatorcontrib>Lin, Shaohui ; Ji, Rongrong ; Li, Yuchao ; Deng, Cheng ; Li, Xuelong</creatorcontrib><description>The success of convolutional neural networks (CNNs) in computer vision applications has been accompanied by a significant increase of computation and memory costs, which prohibits their usage on resource-limited environments, such as mobile systems or embedded devices. To this end, the research of CNN compression has recently become emerging. In this paper, we propose a novel filter pruning scheme, termed structured sparsity regularization (SSR), to simultaneously speed up the computation and reduce the memory overhead of CNNs, which can be well supported by various off-the-shelf deep learning libraries. Concretely, the proposed scheme incorporates two different regularizers of structured sparsity into the original objective function of filter pruning, which fully coordinates the global output and local pruning operations to adaptively prune filters. We further propose an alternative updating with Lagrange multipliers (AULM) scheme to efficiently solve its optimization. AULM follows the principle of alternating direction method of multipliers (ADMM) and alternates between promoting the structured sparsity of CNNs and optimizing the recognition loss, which leads to a very efficient solver (&lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;2.5\times &lt;/tex-math&gt;&lt;/inline-formula&gt; to the most recent work that directly solves the group sparsity-based regularization). Moreover, by imposing the structured sparsity, the online inference is extremely memory-light since the number of filters and the output feature maps are simultaneously reduced. The proposed scheme has been deployed to a variety of state-of-the-art CNN structures, including LeNet, AlexNet, VGGNet, ResNet, and GoogLeNet, over different data sets. Quantitative results demonstrate that the proposed scheme achieves superior performance over the state-of-the-art methods. We further demonstrate the proposed compression scheme for the task of transfer learning, including domain adaptation and object detection, which also show exciting performance gains over the state-of-the-art filter pruning methods.</description><identifier>ISSN: 2162-237X</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2019.2906563</identifier><identifier>PMID: 30990448</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Adaptation models ; Artificial neural networks ; CNN acceleration ; CNN compression ; Compression ; Computation ; Computer vision ; Convergence ; Convolutional neural networks ; Convolutional neural networks (CNNs) ; Electronic devices ; Embedded systems ; Feature maps ; Filters ; Image coding ; Image reconstruction ; Lagrange multiplier ; Libraries ; Machine learning ; Mobile communication systems ; Neural networks ; Object detection ; Object recognition ; Objective function ; Optimization ; Pruning ; Regularization ; Sparsity ; structured sparsity ; Transfer learning</subject><ispartof>IEEE transaction on neural networks and learning systems, 2020-02, Vol.31 (2), p.574-588</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2020</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c466t-f4acb9357b1f4f6f5b088b17b748ea39247405bebfb33b3bf6a3a796d97c941f3</citedby><cites>FETCH-LOGICAL-c466t-f4acb9357b1f4f6f5b088b17b748ea39247405bebfb33b3bf6a3a796d97c941f3</cites><orcidid>0000-0003-2924-946X ; 0000-0001-9163-2932 ; 0000-0003-2620-3247</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/8689357$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/8689357$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/30990448$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Lin, Shaohui</creatorcontrib><creatorcontrib>Ji, Rongrong</creatorcontrib><creatorcontrib>Li, Yuchao</creatorcontrib><creatorcontrib>Deng, Cheng</creatorcontrib><creatorcontrib>Li, Xuelong</creatorcontrib><title>Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>The success of convolutional neural networks (CNNs) in computer vision applications has been accompanied by a significant increase of computation and memory costs, which prohibits their usage on resource-limited environments, such as mobile systems or embedded devices. To this end, the research of CNN compression has recently become emerging. In this paper, we propose a novel filter pruning scheme, termed structured sparsity regularization (SSR), to simultaneously speed up the computation and reduce the memory overhead of CNNs, which can be well supported by various off-the-shelf deep learning libraries. Concretely, the proposed scheme incorporates two different regularizers of structured sparsity into the original objective function of filter pruning, which fully coordinates the global output and local pruning operations to adaptively prune filters. We further propose an alternative updating with Lagrange multipliers (AULM) scheme to efficiently solve its optimization. AULM follows the principle of alternating direction method of multipliers (ADMM) and alternates between promoting the structured sparsity of CNNs and optimizing the recognition loss, which leads to a very efficient solver (&lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;2.5\times &lt;/tex-math&gt;&lt;/inline-formula&gt; to the most recent work that directly solves the group sparsity-based regularization). Moreover, by imposing the structured sparsity, the online inference is extremely memory-light since the number of filters and the output feature maps are simultaneously reduced. The proposed scheme has been deployed to a variety of state-of-the-art CNN structures, including LeNet, AlexNet, VGGNet, ResNet, and GoogLeNet, over different data sets. Quantitative results demonstrate that the proposed scheme achieves superior performance over the state-of-the-art methods. We further demonstrate the proposed compression scheme for the task of transfer learning, including domain adaptation and object detection, which also show exciting performance gains over the state-of-the-art filter pruning methods.</description><subject>Adaptation models</subject><subject>Artificial neural networks</subject><subject>CNN acceleration</subject><subject>CNN compression</subject><subject>Compression</subject><subject>Computation</subject><subject>Computer vision</subject><subject>Convergence</subject><subject>Convolutional neural networks</subject><subject>Convolutional neural networks (CNNs)</subject><subject>Electronic devices</subject><subject>Embedded systems</subject><subject>Feature maps</subject><subject>Filters</subject><subject>Image coding</subject><subject>Image reconstruction</subject><subject>Lagrange multiplier</subject><subject>Libraries</subject><subject>Machine learning</subject><subject>Mobile communication systems</subject><subject>Neural networks</subject><subject>Object detection</subject><subject>Object recognition</subject><subject>Objective function</subject><subject>Optimization</subject><subject>Pruning</subject><subject>Regularization</subject><subject>Sparsity</subject><subject>structured sparsity</subject><subject>Transfer learning</subject><issn>2162-237X</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpdkEtLAzEQgIMoVmr_gIIsePGyNa_Nbo5SrBZKFVvBW0h2k7JlHzXJVuqvN7W1B-cyA_PNMPMBcIXgECHI7xez2XQ-xBDxIeaQJYycgAuMGI4xybLTY51-9MDAuRUMwWDCKD8HPQI5h5RmF2CyaL-kLaJRW69l7kNuNjPtXbQpZTT3tst9Z3U8X0vrSr-N3vSyq6Qtv3URjcvKaxu92q4pm-UlODOycnpwyH3wPn5cjJ7j6cvTZPQwjXPKmI8NlbniJEkVMtQwkyiYZQqlKqWZloRjmlKYKK2MIkQRZZgkMuWs4GnOKTKkD-72e9e2_ey086IuXa6rSja67ZzAGEGcYkZZQG__oau2s024TmCSUIaSoC9QeE_ltnXOaiPWtqyl3QoExc61-HUtdq7FwXUYujms7lSti-PIn9kAXO-BUmt9bGcs2_1OfgDAtoJy</recordid><startdate>20200201</startdate><enddate>20200201</enddate><creator>Lin, Shaohui</creator><creator>Ji, Rongrong</creator><creator>Li, Yuchao</creator><creator>Deng, Cheng</creator><creator>Li, Xuelong</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QF</scope><scope>7QO</scope><scope>7QP</scope><scope>7QQ</scope><scope>7QR</scope><scope>7SC</scope><scope>7SE</scope><scope>7SP</scope><scope>7SR</scope><scope>7TA</scope><scope>7TB</scope><scope>7TK</scope><scope>7U5</scope><scope>8BQ</scope><scope>8FD</scope><scope>F28</scope><scope>FR3</scope><scope>H8D</scope><scope>JG9</scope><scope>JQ2</scope><scope>KR7</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>P64</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-2924-946X</orcidid><orcidid>https://orcid.org/0000-0001-9163-2932</orcidid><orcidid>https://orcid.org/0000-0003-2620-3247</orcidid></search><sort><creationdate>20200201</creationdate><title>Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning</title><author>Lin, Shaohui ; Ji, Rongrong ; Li, Yuchao ; Deng, Cheng ; Li, Xuelong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c466t-f4acb9357b1f4f6f5b088b17b748ea39247405bebfb33b3bf6a3a796d97c941f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Adaptation models</topic><topic>Artificial neural networks</topic><topic>CNN acceleration</topic><topic>CNN compression</topic><topic>Compression</topic><topic>Computation</topic><topic>Computer vision</topic><topic>Convergence</topic><topic>Convolutional neural networks</topic><topic>Convolutional neural networks (CNNs)</topic><topic>Electronic devices</topic><topic>Embedded systems</topic><topic>Feature maps</topic><topic>Filters</topic><topic>Image coding</topic><topic>Image reconstruction</topic><topic>Lagrange multiplier</topic><topic>Libraries</topic><topic>Machine learning</topic><topic>Mobile communication systems</topic><topic>Neural networks</topic><topic>Object detection</topic><topic>Object recognition</topic><topic>Objective function</topic><topic>Optimization</topic><topic>Pruning</topic><topic>Regularization</topic><topic>Sparsity</topic><topic>structured sparsity</topic><topic>Transfer learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Lin, Shaohui</creatorcontrib><creatorcontrib>Ji, Rongrong</creatorcontrib><creatorcontrib>Li, Yuchao</creatorcontrib><creatorcontrib>Deng, Cheng</creatorcontrib><creatorcontrib>Li, Xuelong</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Aluminium Industry Abstracts</collection><collection>Biotechnology Research Abstracts</collection><collection>Calcium &amp; Calcified Tissue Abstracts</collection><collection>Ceramic Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Computer and Information Systems Abstracts</collection><collection>Corrosion Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>Materials Business File</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Solid State and Superconductivity Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>ANTE: Abstracts in New Technology &amp; Engineering</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Civil Engineering Abstracts</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Lin, Shaohui</au><au>Ji, Rongrong</au><au>Li, Yuchao</au><au>Deng, Cheng</au><au>Li, Xuelong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2020-02-01</date><risdate>2020</risdate><volume>31</volume><issue>2</issue><spage>574</spage><epage>588</epage><pages>574-588</pages><issn>2162-237X</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>The success of convolutional neural networks (CNNs) in computer vision applications has been accompanied by a significant increase of computation and memory costs, which prohibits their usage on resource-limited environments, such as mobile systems or embedded devices. To this end, the research of CNN compression has recently become emerging. In this paper, we propose a novel filter pruning scheme, termed structured sparsity regularization (SSR), to simultaneously speed up the computation and reduce the memory overhead of CNNs, which can be well supported by various off-the-shelf deep learning libraries. Concretely, the proposed scheme incorporates two different regularizers of structured sparsity into the original objective function of filter pruning, which fully coordinates the global output and local pruning operations to adaptively prune filters. We further propose an alternative updating with Lagrange multipliers (AULM) scheme to efficiently solve its optimization. AULM follows the principle of alternating direction method of multipliers (ADMM) and alternates between promoting the structured sparsity of CNNs and optimizing the recognition loss, which leads to a very efficient solver (&lt;inline-formula&gt; &lt;tex-math notation="LaTeX"&gt;2.5\times &lt;/tex-math&gt;&lt;/inline-formula&gt; to the most recent work that directly solves the group sparsity-based regularization). Moreover, by imposing the structured sparsity, the online inference is extremely memory-light since the number of filters and the output feature maps are simultaneously reduced. The proposed scheme has been deployed to a variety of state-of-the-art CNN structures, including LeNet, AlexNet, VGGNet, ResNet, and GoogLeNet, over different data sets. Quantitative results demonstrate that the proposed scheme achieves superior performance over the state-of-the-art methods. We further demonstrate the proposed compression scheme for the task of transfer learning, including domain adaptation and object detection, which also show exciting performance gains over the state-of-the-art filter pruning methods.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>30990448</pmid><doi>10.1109/TNNLS.2019.2906563</doi><tpages>15</tpages><orcidid>https://orcid.org/0000-0003-2924-946X</orcidid><orcidid>https://orcid.org/0000-0001-9163-2932</orcidid><orcidid>https://orcid.org/0000-0003-2620-3247</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 2162-237X
ispartof IEEE transaction on neural networks and learning systems, 2020-02, Vol.31 (2), p.574-588
issn 2162-237X
2162-2388
language eng
recordid cdi_proquest_journals_2354615019
source IEEE Electronic Library (IEL)
subjects Adaptation models
Artificial neural networks
CNN acceleration
CNN compression
Compression
Computation
Computer vision
Convergence
Convolutional neural networks
Convolutional neural networks (CNNs)
Electronic devices
Embedded systems
Feature maps
Filters
Image coding
Image reconstruction
Lagrange multiplier
Libraries
Machine learning
Mobile communication systems
Neural networks
Object detection
Object recognition
Objective function
Optimization
Pruning
Regularization
Sparsity
structured sparsity
Transfer learning
title Toward Compact ConvNets via Structure-Sparsity Regularized Filter Pruning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-20T13%3A53%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Toward%20Compact%20ConvNets%20via%20Structure-Sparsity%20Regularized%20Filter%20Pruning&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Lin,%20Shaohui&rft.date=2020-02-01&rft.volume=31&rft.issue=2&rft.spage=574&rft.epage=588&rft.pages=574-588&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2019.2906563&rft_dat=%3Cproquest_RIE%3E2210272646%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2354615019&rft_id=info:pmid/30990448&rft_ieee_id=8689357&rfr_iscdi=true