Compression and acceleration of convolution neural network: a Genetic Algorithm based approach
Genetic Algorithm (GA) is a meta-heuristics search and optimization approach which we have utilized for fine-tuning the standard deep learning models for reducing the storage space with improvement in inference time. The pre-trained models have been widely acclaimed as the best choice of CNN for var...
Gespeichert in:
Veröffentlicht in: | Journal of ambient intelligence and humanized computing 2023-10, Vol.14 (10), p.13387-13397 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 13397 |
---|---|
container_issue | 10 |
container_start_page | 13387 |
container_title | Journal of ambient intelligence and humanized computing |
container_volume | 14 |
creator | Agarwal, Mohit Gupta, Suneet K. Biswas, Mainak Garg, Deepak |
description | Genetic Algorithm (GA) is a meta-heuristics search and optimization approach which we have utilized for fine-tuning the standard deep learning models for reducing the storage space with improvement in inference time. The pre-trained models have been widely acclaimed as the best choice of CNN for various image classification problems in different domains, but they require huge storage space and it causes a problem for deploying these models on mobile or edge devices. As these devices are constrained with limited memory and computational power. In this paper, a novel GA based method has been proposed which compresses and accelerates the CNN models so that these models can be easily deployed on edge devices. Extensive computer simulations have been conducted on widely used models such as AlexNet, VGG16, SqueezeNet, and ResNet50. We used benchmark datasets such as MNIST, CIFAR-10, and CIFAR-100 to determine the performance. Results reveal that storage space of the AlexNet model was reduced by 87.5%, 86.55% and 86.16% on MNIST, CIFAR-10 and CIFAR- 100 datasets respectively, while VGG16, ResNet50, and SqueezeNet have been compressed by nearly 91%, 78%, 38% respectively. From the results, it has been noticed that there is a significant improvement in inference time of around 35% in AlexNet, 9% in SqueezeNet, 73% in ResNet50, and 80% in VGG16. This improvement is noticed mainly because of the fine-tuning of the deep learning models using the GA. Overall, the proposed GA-based method showed outstanding performance and it motivates research and practitioners to explore it further. |
doi_str_mv | 10.1007/s12652-022-03793-1 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2919479918</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2919479918</sourcerecordid><originalsourceid>FETCH-LOGICAL-c385t-37608c1a5ef88ef1de8c473d1eac7b81002c4771f46acaa480f3cdb6d33bf3073</originalsourceid><addsrcrecordid>eNp9UMFKxDAQDaLgsu4PeCp4rmY6bZN6WxZdhQUvejWkabLbtW1q0ir-vdmt6M3AMPPCe2-YR8gl0GuglN14SPIsiWkSClmBMZyQGfCcxxmk2envjOycLLzf0_CwQACYkdeVbXunva9tF8muiqRSutFODocPayJluw_bjEfY6dHJJrTh07q320hGax1AraJls7WuHnZtVEqvg0vfOyvV7oKcGdl4vfjpc_Jyf_e8eog3T-vH1XITK-TZECPLKVcgM2041wYqzVXKsAItFSt5uDIJmIFJc6mkTDk1qKoyrxBLg5ThnFxNvmHt-6j9IPZ2dF1YKZICipQVBfDASiaWctZ7p43oXd1K9yWAikOUYopShCjFMUoBQYSTyAdyt9Xuz_of1Tdx6Hey</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2919479918</pqid></control><display><type>article</type><title>Compression and acceleration of convolution neural network: a Genetic Algorithm based approach</title><source>SpringerLink Journals - AutoHoldings</source><source>ProQuest Central</source><creator>Agarwal, Mohit ; Gupta, Suneet K. ; Biswas, Mainak ; Garg, Deepak</creator><creatorcontrib>Agarwal, Mohit ; Gupta, Suneet K. ; Biswas, Mainak ; Garg, Deepak</creatorcontrib><description>Genetic Algorithm (GA) is a meta-heuristics search and optimization approach which we have utilized for fine-tuning the standard deep learning models for reducing the storage space with improvement in inference time. The pre-trained models have been widely acclaimed as the best choice of CNN for various image classification problems in different domains, but they require huge storage space and it causes a problem for deploying these models on mobile or edge devices. As these devices are constrained with limited memory and computational power. In this paper, a novel GA based method has been proposed which compresses and accelerates the CNN models so that these models can be easily deployed on edge devices. Extensive computer simulations have been conducted on widely used models such as AlexNet, VGG16, SqueezeNet, and ResNet50. We used benchmark datasets such as MNIST, CIFAR-10, and CIFAR-100 to determine the performance. Results reveal that storage space of the AlexNet model was reduced by 87.5%, 86.55% and 86.16% on MNIST, CIFAR-10 and CIFAR- 100 datasets respectively, while VGG16, ResNet50, and SqueezeNet have been compressed by nearly 91%, 78%, 38% respectively. From the results, it has been noticed that there is a significant improvement in inference time of around 35% in AlexNet, 9% in SqueezeNet, 73% in ResNet50, and 80% in VGG16. This improvement is noticed mainly because of the fine-tuning of the deep learning models using the GA. Overall, the proposed GA-based method showed outstanding performance and it motivates research and practitioners to explore it further.</description><identifier>ISSN: 1868-5137</identifier><identifier>EISSN: 1868-5145</identifier><identifier>DOI: 10.1007/s12652-022-03793-1</identifier><language>eng</language><publisher>Berlin/Heidelberg: Springer Berlin Heidelberg</publisher><subject>Accuracy ; Artificial Intelligence ; Artificial neural networks ; Chromosomes ; Classification ; Computational Intelligence ; Datasets ; Deep learning ; Engineering ; Genetic algorithms ; Image classification ; Inference ; Machine learning ; Mutation ; Neural networks ; Original Research ; Robotics and Automation ; User Interfaces and Human Computer Interaction ; Wavelet transforms</subject><ispartof>Journal of ambient intelligence and humanized computing, 2023-10, Vol.14 (10), p.13387-13397</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c385t-37608c1a5ef88ef1de8c473d1eac7b81002c4771f46acaa480f3cdb6d33bf3073</citedby><cites>FETCH-LOGICAL-c385t-37608c1a5ef88ef1de8c473d1eac7b81002c4771f46acaa480f3cdb6d33bf3073</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s12652-022-03793-1$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2919479918?pq-origsite=primo$$EHTML$$P50$$Gproquest$$H</linktohtml><link.rule.ids>314,776,780,21367,27901,27902,33721,41464,42533,43781,51294</link.rule.ids></links><search><creatorcontrib>Agarwal, Mohit</creatorcontrib><creatorcontrib>Gupta, Suneet K.</creatorcontrib><creatorcontrib>Biswas, Mainak</creatorcontrib><creatorcontrib>Garg, Deepak</creatorcontrib><title>Compression and acceleration of convolution neural network: a Genetic Algorithm based approach</title><title>Journal of ambient intelligence and humanized computing</title><addtitle>J Ambient Intell Human Comput</addtitle><description>Genetic Algorithm (GA) is a meta-heuristics search and optimization approach which we have utilized for fine-tuning the standard deep learning models for reducing the storage space with improvement in inference time. The pre-trained models have been widely acclaimed as the best choice of CNN for various image classification problems in different domains, but they require huge storage space and it causes a problem for deploying these models on mobile or edge devices. As these devices are constrained with limited memory and computational power. In this paper, a novel GA based method has been proposed which compresses and accelerates the CNN models so that these models can be easily deployed on edge devices. Extensive computer simulations have been conducted on widely used models such as AlexNet, VGG16, SqueezeNet, and ResNet50. We used benchmark datasets such as MNIST, CIFAR-10, and CIFAR-100 to determine the performance. Results reveal that storage space of the AlexNet model was reduced by 87.5%, 86.55% and 86.16% on MNIST, CIFAR-10 and CIFAR- 100 datasets respectively, while VGG16, ResNet50, and SqueezeNet have been compressed by nearly 91%, 78%, 38% respectively. From the results, it has been noticed that there is a significant improvement in inference time of around 35% in AlexNet, 9% in SqueezeNet, 73% in ResNet50, and 80% in VGG16. This improvement is noticed mainly because of the fine-tuning of the deep learning models using the GA. Overall, the proposed GA-based method showed outstanding performance and it motivates research and practitioners to explore it further.</description><subject>Accuracy</subject><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Chromosomes</subject><subject>Classification</subject><subject>Computational Intelligence</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Engineering</subject><subject>Genetic algorithms</subject><subject>Image classification</subject><subject>Inference</subject><subject>Machine learning</subject><subject>Mutation</subject><subject>Neural networks</subject><subject>Original Research</subject><subject>Robotics and Automation</subject><subject>User Interfaces and Human Computer Interaction</subject><subject>Wavelet transforms</subject><issn>1868-5137</issn><issn>1868-5145</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9UMFKxDAQDaLgsu4PeCp4rmY6bZN6WxZdhQUvejWkabLbtW1q0ir-vdmt6M3AMPPCe2-YR8gl0GuglN14SPIsiWkSClmBMZyQGfCcxxmk2envjOycLLzf0_CwQACYkdeVbXunva9tF8muiqRSutFODocPayJluw_bjEfY6dHJJrTh07q320hGax1AraJls7WuHnZtVEqvg0vfOyvV7oKcGdl4vfjpc_Jyf_e8eog3T-vH1XITK-TZECPLKVcgM2041wYqzVXKsAItFSt5uDIJmIFJc6mkTDk1qKoyrxBLg5ThnFxNvmHt-6j9IPZ2dF1YKZICipQVBfDASiaWctZ7p43oXd1K9yWAikOUYopShCjFMUoBQYSTyAdyt9Xuz_of1Tdx6Hey</recordid><startdate>20231001</startdate><enddate>20231001</enddate><creator>Agarwal, Mohit</creator><creator>Gupta, Suneet K.</creator><creator>Biswas, Mainak</creator><creator>Garg, Deepak</creator><general>Springer Berlin Heidelberg</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>8FE</scope><scope>8FG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope></search><sort><creationdate>20231001</creationdate><title>Compression and acceleration of convolution neural network: a Genetic Algorithm based approach</title><author>Agarwal, Mohit ; Gupta, Suneet K. ; Biswas, Mainak ; Garg, Deepak</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c385t-37608c1a5ef88ef1de8c473d1eac7b81002c4771f46acaa480f3cdb6d33bf3073</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Chromosomes</topic><topic>Classification</topic><topic>Computational Intelligence</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Engineering</topic><topic>Genetic algorithms</topic><topic>Image classification</topic><topic>Inference</topic><topic>Machine learning</topic><topic>Mutation</topic><topic>Neural networks</topic><topic>Original Research</topic><topic>Robotics and Automation</topic><topic>User Interfaces and Human Computer Interaction</topic><topic>Wavelet transforms</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Agarwal, Mohit</creatorcontrib><creatorcontrib>Gupta, Suneet K.</creatorcontrib><creatorcontrib>Biswas, Mainak</creatorcontrib><creatorcontrib>Garg, Deepak</creatorcontrib><collection>CrossRef</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><jtitle>Journal of ambient intelligence and humanized computing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Agarwal, Mohit</au><au>Gupta, Suneet K.</au><au>Biswas, Mainak</au><au>Garg, Deepak</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Compression and acceleration of convolution neural network: a Genetic Algorithm based approach</atitle><jtitle>Journal of ambient intelligence and humanized computing</jtitle><stitle>J Ambient Intell Human Comput</stitle><date>2023-10-01</date><risdate>2023</risdate><volume>14</volume><issue>10</issue><spage>13387</spage><epage>13397</epage><pages>13387-13397</pages><issn>1868-5137</issn><eissn>1868-5145</eissn><abstract>Genetic Algorithm (GA) is a meta-heuristics search and optimization approach which we have utilized for fine-tuning the standard deep learning models for reducing the storage space with improvement in inference time. The pre-trained models have been widely acclaimed as the best choice of CNN for various image classification problems in different domains, but they require huge storage space and it causes a problem for deploying these models on mobile or edge devices. As these devices are constrained with limited memory and computational power. In this paper, a novel GA based method has been proposed which compresses and accelerates the CNN models so that these models can be easily deployed on edge devices. Extensive computer simulations have been conducted on widely used models such as AlexNet, VGG16, SqueezeNet, and ResNet50. We used benchmark datasets such as MNIST, CIFAR-10, and CIFAR-100 to determine the performance. Results reveal that storage space of the AlexNet model was reduced by 87.5%, 86.55% and 86.16% on MNIST, CIFAR-10 and CIFAR- 100 datasets respectively, while VGG16, ResNet50, and SqueezeNet have been compressed by nearly 91%, 78%, 38% respectively. From the results, it has been noticed that there is a significant improvement in inference time of around 35% in AlexNet, 9% in SqueezeNet, 73% in ResNet50, and 80% in VGG16. This improvement is noticed mainly because of the fine-tuning of the deep learning models using the GA. Overall, the proposed GA-based method showed outstanding performance and it motivates research and practitioners to explore it further.</abstract><cop>Berlin/Heidelberg</cop><pub>Springer Berlin Heidelberg</pub><doi>10.1007/s12652-022-03793-1</doi><tpages>11</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1868-5137 |
ispartof | Journal of ambient intelligence and humanized computing, 2023-10, Vol.14 (10), p.13387-13397 |
issn | 1868-5137 1868-5145 |
language | eng |
recordid | cdi_proquest_journals_2919479918 |
source | SpringerLink Journals - AutoHoldings; ProQuest Central |
subjects | Accuracy Artificial Intelligence Artificial neural networks Chromosomes Classification Computational Intelligence Datasets Deep learning Engineering Genetic algorithms Image classification Inference Machine learning Mutation Neural networks Original Research Robotics and Automation User Interfaces and Human Computer Interaction Wavelet transforms |
title | Compression and acceleration of convolution neural network: a Genetic Algorithm based approach |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T15%3A30%3A50IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Compression%20and%20acceleration%20of%20convolution%20neural%20network:%20a%20Genetic%20Algorithm%20based%20approach&rft.jtitle=Journal%20of%20ambient%20intelligence%20and%20humanized%20computing&rft.au=Agarwal,%20Mohit&rft.date=2023-10-01&rft.volume=14&rft.issue=10&rft.spage=13387&rft.epage=13397&rft.pages=13387-13397&rft.issn=1868-5137&rft.eissn=1868-5145&rft_id=info:doi/10.1007/s12652-022-03793-1&rft_dat=%3Cproquest_cross%3E2919479918%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2919479918&rft_id=info:pmid/&rfr_iscdi=true |