FTBME: feature transferring based multi-model ensemble

Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Multimedia tools and applications 2020-07, Vol.79 (25-26), p.18767-18799
Hauptverfasser: Yang, A. Yongquan, Lv, B. Haijun, Chen, C. Ning, Wu, D. Yang, Zheng, E. Zhongxi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 18799
container_issue 25-26
container_start_page 18767
container_title Multimedia tools and applications
container_volume 79
creator Yang, A. Yongquan
Lv, B. Haijun
Chen, C. Ning
Wu, D. Yang
Zheng, E. Zhongxi
description Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.
doi_str_mv 10.1007/s11042-020-08746-4
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2420901443</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2420901443</sourcerecordid><originalsourceid>FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</originalsourceid><addsrcrecordid>eNp9kLtOw0AQRVcIJELgB6gsUS_M7DOmgygBpCCaUK_s9SxK5EfYtQv-Hgcj0VHNFPfc0RzGrhFuEcDeJURQgoMADgurDFcnbIbaSm6twNNxlwvgVgOes4uU9gBotFAzZtbbx9fVfRao6IdIWR-LNgWKcdd-ZGWRqMqaoe53vOkqqjNqEzVlTZfsLBR1oqvfOWfv69V2-cw3b08vy4cN99KKnhfBYum99GghGAs5GK0g9xo9KmG8XlSIwZeVNFLkpLWxXqmgSw3e6FzIObuZeg-x-xwo9W7fDbEdTzqhxNiHSskxJaaUj11KkYI7xF1TxC-H4I5-3OTHjX7cjx-nRkhOUDocn6X4V_0P9Q22dWXm</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2420901443</pqid></control><display><type>article</type><title>FTBME: feature transferring based multi-model ensemble</title><source>SpringerLink Journals - AutoHoldings</source><creator>Yang, A. Yongquan ; Lv, B. Haijun ; Chen, C. Ning ; Wu, D. Yang ; Zheng, E. Zhongxi</creator><creatorcontrib>Yang, A. Yongquan ; Lv, B. Haijun ; Chen, C. Ning ; Wu, D. Yang ; Zheng, E. Zhongxi</creatorcontrib><description>Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-020-08746-4</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial intelligence ; Artificial neural networks ; Computer Communication Networks ; Computer Science ; Costs ; Data Structures and Information Theory ; Model testing ; Multimedia Information Systems ; Neural networks ; Special Purpose and Application-Based Systems ; Training ; Weight reduction</subject><ispartof>Multimedia tools and applications, 2020-07, Vol.79 (25-26), p.18767-18799</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020</rights><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</citedby><cites>FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-020-08746-4$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-020-08746-4$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Yang, A. Yongquan</creatorcontrib><creatorcontrib>Lv, B. Haijun</creatorcontrib><creatorcontrib>Chen, C. Ning</creatorcontrib><creatorcontrib>Wu, D. Yang</creatorcontrib><creatorcontrib>Zheng, E. Zhongxi</creatorcontrib><title>FTBME: feature transferring based multi-model ensemble</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.</description><subject>Artificial intelligence</subject><subject>Artificial neural networks</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Costs</subject><subject>Data Structures and Information Theory</subject><subject>Model testing</subject><subject>Multimedia Information Systems</subject><subject>Neural networks</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Training</subject><subject>Weight reduction</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>BENPR</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9kLtOw0AQRVcIJELgB6gsUS_M7DOmgygBpCCaUK_s9SxK5EfYtQv-Hgcj0VHNFPfc0RzGrhFuEcDeJURQgoMADgurDFcnbIbaSm6twNNxlwvgVgOes4uU9gBotFAzZtbbx9fVfRao6IdIWR-LNgWKcdd-ZGWRqMqaoe53vOkqqjNqEzVlTZfsLBR1oqvfOWfv69V2-cw3b08vy4cN99KKnhfBYum99GghGAs5GK0g9xo9KmG8XlSIwZeVNFLkpLWxXqmgSw3e6FzIObuZeg-x-xwo9W7fDbEdTzqhxNiHSskxJaaUj11KkYI7xF1TxC-H4I5-3OTHjX7cjx-nRkhOUDocn6X4V_0P9Q22dWXm</recordid><startdate>20200701</startdate><enddate>20200701</enddate><creator>Yang, A. Yongquan</creator><creator>Lv, B. Haijun</creator><creator>Chen, C. Ning</creator><creator>Wu, D. Yang</creator><creator>Zheng, E. Zhongxi</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope></search><sort><creationdate>20200701</creationdate><title>FTBME: feature transferring based multi-model ensemble</title><author>Yang, A. Yongquan ; Lv, B. Haijun ; Chen, C. Ning ; Wu, D. Yang ; Zheng, E. Zhongxi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Artificial intelligence</topic><topic>Artificial neural networks</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Costs</topic><topic>Data Structures and Information Theory</topic><topic>Model testing</topic><topic>Multimedia Information Systems</topic><topic>Neural networks</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Training</topic><topic>Weight reduction</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yang, A. Yongquan</creatorcontrib><creatorcontrib>Lv, B. Haijun</creatorcontrib><creatorcontrib>Chen, C. Ning</creatorcontrib><creatorcontrib>Wu, D. Yang</creatorcontrib><creatorcontrib>Zheng, E. Zhongxi</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI/INFORM Collection</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection (ProQuest)</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Research Library</collection><collection>Research Library (Corporate)</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, A. Yongquan</au><au>Lv, B. Haijun</au><au>Chen, C. Ning</au><au>Wu, D. Yang</au><au>Zheng, E. Zhongxi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>FTBME: feature transferring based multi-model ensemble</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2020-07-01</date><risdate>2020</risdate><volume>79</volume><issue>25-26</issue><spage>18767</spage><epage>18799</epage><pages>18767-18799</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-020-08746-4</doi><tpages>33</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1380-7501
ispartof Multimedia tools and applications, 2020-07, Vol.79 (25-26), p.18767-18799
issn 1380-7501
1573-7721
language eng
recordid cdi_proquest_journals_2420901443
source SpringerLink Journals - AutoHoldings
subjects Artificial intelligence
Artificial neural networks
Computer Communication Networks
Computer Science
Costs
Data Structures and Information Theory
Model testing
Multimedia Information Systems
Neural networks
Special Purpose and Application-Based Systems
Training
Weight reduction
title FTBME: feature transferring based multi-model ensemble
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-07T10%3A21%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=FTBME:%20feature%20transferring%20based%20multi-model%20ensemble&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Yang,%20A.%20Yongquan&rft.date=2020-07-01&rft.volume=79&rft.issue=25-26&rft.spage=18767&rft.epage=18799&rft.pages=18767-18799&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-020-08746-4&rft_dat=%3Cproquest_cross%3E2420901443%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2420901443&rft_id=info:pmid/&rfr_iscdi=true