FTBME: feature transferring based multi-model ensemble
Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time a...
Gespeichert in:
Veröffentlicht in: | Multimedia tools and applications 2020-07, Vol.79 (25-26), p.18767-18799 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 18799 |
---|---|
container_issue | 25-26 |
container_start_page | 18767 |
container_title | Multimedia tools and applications |
container_volume | 79 |
creator | Yang, A. Yongquan Lv, B. Haijun Chen, C. Ning Wu, D. Yang Zheng, E. Zhongxi |
description | Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks. |
doi_str_mv | 10.1007/s11042-020-08746-4 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2420901443</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2420901443</sourcerecordid><originalsourceid>FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</originalsourceid><addsrcrecordid>eNp9kLtOw0AQRVcIJELgB6gsUS_M7DOmgygBpCCaUK_s9SxK5EfYtQv-Hgcj0VHNFPfc0RzGrhFuEcDeJURQgoMADgurDFcnbIbaSm6twNNxlwvgVgOes4uU9gBotFAzZtbbx9fVfRao6IdIWR-LNgWKcdd-ZGWRqMqaoe53vOkqqjNqEzVlTZfsLBR1oqvfOWfv69V2-cw3b08vy4cN99KKnhfBYum99GghGAs5GK0g9xo9KmG8XlSIwZeVNFLkpLWxXqmgSw3e6FzIObuZeg-x-xwo9W7fDbEdTzqhxNiHSskxJaaUj11KkYI7xF1TxC-H4I5-3OTHjX7cjx-nRkhOUDocn6X4V_0P9Q22dWXm</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2420901443</pqid></control><display><type>article</type><title>FTBME: feature transferring based multi-model ensemble</title><source>SpringerLink Journals - AutoHoldings</source><creator>Yang, A. Yongquan ; Lv, B. Haijun ; Chen, C. Ning ; Wu, D. Yang ; Zheng, E. Zhongxi</creator><creatorcontrib>Yang, A. Yongquan ; Lv, B. Haijun ; Chen, C. Ning ; Wu, D. Yang ; Zheng, E. Zhongxi</creatorcontrib><description>Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-020-08746-4</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial intelligence ; Artificial neural networks ; Computer Communication Networks ; Computer Science ; Costs ; Data Structures and Information Theory ; Model testing ; Multimedia Information Systems ; Neural networks ; Special Purpose and Application-Based Systems ; Training ; Weight reduction</subject><ispartof>Multimedia tools and applications, 2020-07, Vol.79 (25-26), p.18767-18799</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020</rights><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</citedby><cites>FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-020-08746-4$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-020-08746-4$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Yang, A. Yongquan</creatorcontrib><creatorcontrib>Lv, B. Haijun</creatorcontrib><creatorcontrib>Chen, C. Ning</creatorcontrib><creatorcontrib>Wu, D. Yang</creatorcontrib><creatorcontrib>Zheng, E. Zhongxi</creatorcontrib><title>FTBME: feature transferring based multi-model ensemble</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.</description><subject>Artificial intelligence</subject><subject>Artificial neural networks</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Costs</subject><subject>Data Structures and Information Theory</subject><subject>Model testing</subject><subject>Multimedia Information Systems</subject><subject>Neural networks</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Training</subject><subject>Weight reduction</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>BENPR</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9kLtOw0AQRVcIJELgB6gsUS_M7DOmgygBpCCaUK_s9SxK5EfYtQv-Hgcj0VHNFPfc0RzGrhFuEcDeJURQgoMADgurDFcnbIbaSm6twNNxlwvgVgOes4uU9gBotFAzZtbbx9fVfRao6IdIWR-LNgWKcdd-ZGWRqMqaoe53vOkqqjNqEzVlTZfsLBR1oqvfOWfv69V2-cw3b08vy4cN99KKnhfBYum99GghGAs5GK0g9xo9KmG8XlSIwZeVNFLkpLWxXqmgSw3e6FzIObuZeg-x-xwo9W7fDbEdTzqhxNiHSskxJaaUj11KkYI7xF1TxC-H4I5-3OTHjX7cjx-nRkhOUDocn6X4V_0P9Q22dWXm</recordid><startdate>20200701</startdate><enddate>20200701</enddate><creator>Yang, A. Yongquan</creator><creator>Lv, B. Haijun</creator><creator>Chen, C. Ning</creator><creator>Wu, D. Yang</creator><creator>Zheng, E. Zhongxi</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope></search><sort><creationdate>20200701</creationdate><title>FTBME: feature transferring based multi-model ensemble</title><author>Yang, A. Yongquan ; Lv, B. Haijun ; Chen, C. Ning ; Wu, D. Yang ; Zheng, E. Zhongxi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c372t-af71bcc3c170f6709065409c51c1426c58d11fcbd36329e5567c44f5b50c65923</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Artificial intelligence</topic><topic>Artificial neural networks</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Costs</topic><topic>Data Structures and Information Theory</topic><topic>Model testing</topic><topic>Multimedia Information Systems</topic><topic>Neural networks</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Training</topic><topic>Weight reduction</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yang, A. Yongquan</creatorcontrib><creatorcontrib>Lv, B. Haijun</creatorcontrib><creatorcontrib>Chen, C. Ning</creatorcontrib><creatorcontrib>Wu, D. Yang</creatorcontrib><creatorcontrib>Zheng, E. Zhongxi</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI/INFORM Collection</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection (ProQuest)</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Research Library</collection><collection>Research Library (Corporate)</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, A. Yongquan</au><au>Lv, B. Haijun</au><au>Chen, C. Ning</au><au>Wu, D. Yang</au><au>Zheng, E. Zhongxi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>FTBME: feature transferring based multi-model ensemble</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2020-07-01</date><risdate>2020</risdate><volume>79</volume><issue>25-26</issue><spage>18767</spage><epage>18799</epage><pages>18767-18799</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>Multi-model ensemble is an important fundamental technique of practical value for many artificial intelligence applications. However, the usage for multi-model ensemble has been limited when it is combined with deep neural networks to construct ensemble of deep neural networks. Due to the big time and computing resources required to train and to integrate multiple deep neural networks for the achievement of multi-model ensemble, the engineering application field where developing time and computing resources are usually restricted, has not yet widespreadly benefited from ensemble of deep neural networks. To alleviate this situation, we present a new multi-model ensemble approach entitled feature transferring based multi-model ensemble (FTBME), for ensemble of deep neural networks. Primarily, we propose a feature transferring based multi-model training strategy to more affordably find multiple extra models based on a given previously optimized deep neural network model. Sequentially, to develop better ensemble solutions, we design a more effective random greedy based ensemble selection strategy to filter out models non-positive to ensemble generalization. Finally, inspired by the idea of averaging parameter points, we propose to fuse the obtained models in weight space which eventually reduces the expense of ensemble at the testing stage to a single deep neural network model while retaining the generalization. These three advances constitute the resulting technique FTBME. We conducted extensive experiments using deep neural networks, from light weight to complex, on ImageNet, CIFAR-10 and CIFAR-100. Results show that, given a deep neural network model which has been well-optimized and reaching its limit, FTBME can obtain better generalization with minor extra training expense while maintaining the expense to a single model at ensemble testing. This promising property of FTBME make us believe that it could be leveraged to broaden the usage for ensemble of deep neural networks, alleviating the situation that the engineering application field has not yet widespreadly benefited from ensemble of deep neural networks.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-020-08746-4</doi><tpages>33</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1380-7501 |
ispartof | Multimedia tools and applications, 2020-07, Vol.79 (25-26), p.18767-18799 |
issn | 1380-7501 1573-7721 |
language | eng |
recordid | cdi_proquest_journals_2420901443 |
source | SpringerLink Journals - AutoHoldings |
subjects | Artificial intelligence Artificial neural networks Computer Communication Networks Computer Science Costs Data Structures and Information Theory Model testing Multimedia Information Systems Neural networks Special Purpose and Application-Based Systems Training Weight reduction |
title | FTBME: feature transferring based multi-model ensemble |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-07T10%3A21%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=FTBME:%20feature%20transferring%20based%20multi-model%20ensemble&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Yang,%20A.%20Yongquan&rft.date=2020-07-01&rft.volume=79&rft.issue=25-26&rft.spage=18767&rft.epage=18799&rft.pages=18767-18799&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-020-08746-4&rft_dat=%3Cproquest_cross%3E2420901443%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2420901443&rft_id=info:pmid/&rfr_iscdi=true |