Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning

Machine Learning algorithms have been extensively researched throughout the last decade, leading to unprecedented advances in a broad range of applications, such as image classification and reconstruction, object recognition, and text categorization. Nonetheless, most Machine Learning algorithms are...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2022-12
Hauptverfasser: de Rosa, Gustavo H, Roder, Mateus, Papa, João Paulo, Claudio F G dos Santos
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator de Rosa, Gustavo H
Roder, Mateus
Papa, João Paulo
Claudio F G dos Santos
description Machine Learning algorithms have been extensively researched throughout the last decade, leading to unprecedented advances in a broad range of applications, such as image classification and reconstruction, object recognition, and text categorization. Nonetheless, most Machine Learning algorithms are trained via derivative-based optimizers, such as the Stochastic Gradient Descent, leading to possible local optimum entrapments and inhibiting them from achieving proper performances. A bio-inspired alternative to traditional optimization techniques, denoted as meta-heuristic, has received significant attention due to its simplicity and ability to avoid local optimums imprisonment. In this work, we propose to use meta-heuristic techniques to fine-tune pre-trained weights, exploring additional regions of the search space, and improving their effectiveness. The experimental evaluation comprises two classification tasks (image and text) and is assessed under four literature datasets. Experimental results show nature-inspired algorithms' capacity in exploring the neighborhood of pre-trained weights, achieving superior results than their counterpart pre-trained architectures. Additionally, a thorough analysis of distinct architectures, such as Multi-Layer Perceptron and Recurrent Neural Networks, attempts to visualize and provide more precise insights into the most critical weights to be fine-tuned in the learning process.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2755991733</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2755991733</sourcerecordid><originalsourceid>FETCH-proquest_journals_27559917333</originalsourceid><addsrcrecordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mRw9MwtKMovy8xLVwgoStUNKUrMzEtNUQhPzUzPKClWCMkoyi9Nz1DwTS1J1PVILS3KLC7JTC5WcAOq0g0pzQPq42FgTUvMKU7lhdLcDMpuriHOHrpAcwtLU4tL4rPyS4vygFLxRuamppaWhubGxsbEqQIAaeQ5WQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2755991733</pqid></control><display><type>article</type><title>Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning</title><source>Freely Accessible Journals</source><creator>de Rosa, Gustavo H ; Roder, Mateus ; Papa, João Paulo ; Claudio F G dos Santos</creator><creatorcontrib>de Rosa, Gustavo H ; Roder, Mateus ; Papa, João Paulo ; Claudio F G dos Santos</creatorcontrib><description>Machine Learning algorithms have been extensively researched throughout the last decade, leading to unprecedented advances in a broad range of applications, such as image classification and reconstruction, object recognition, and text categorization. Nonetheless, most Machine Learning algorithms are trained via derivative-based optimizers, such as the Stochastic Gradient Descent, leading to possible local optimum entrapments and inhibiting them from achieving proper performances. A bio-inspired alternative to traditional optimization techniques, denoted as meta-heuristic, has received significant attention due to its simplicity and ability to avoid local optimums imprisonment. In this work, we propose to use meta-heuristic techniques to fine-tune pre-trained weights, exploring additional regions of the search space, and improving their effectiveness. The experimental evaluation comprises two classification tasks (image and text) and is assessed under four literature datasets. Experimental results show nature-inspired algorithms' capacity in exploring the neighborhood of pre-trained weights, achieving superior results than their counterpart pre-trained architectures. Additionally, a thorough analysis of distinct architectures, such as Multi-Layer Perceptron and Recurrent Neural Networks, attempts to visualize and provide more precise insights into the most critical weights to be fine-tuned in the learning process.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Heuristic ; Heuristic methods ; Image classification ; Image reconstruction ; Machine learning ; Multilayer perceptrons ; Multilayers ; Object recognition ; Optimization ; Optimization techniques ; Recurrent neural networks</subject><ispartof>arXiv.org, 2022-12</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by-nc-sa/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>781,785</link.rule.ids></links><search><creatorcontrib>de Rosa, Gustavo H</creatorcontrib><creatorcontrib>Roder, Mateus</creatorcontrib><creatorcontrib>Papa, João Paulo</creatorcontrib><creatorcontrib>Claudio F G dos Santos</creatorcontrib><title>Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning</title><title>arXiv.org</title><description>Machine Learning algorithms have been extensively researched throughout the last decade, leading to unprecedented advances in a broad range of applications, such as image classification and reconstruction, object recognition, and text categorization. Nonetheless, most Machine Learning algorithms are trained via derivative-based optimizers, such as the Stochastic Gradient Descent, leading to possible local optimum entrapments and inhibiting them from achieving proper performances. A bio-inspired alternative to traditional optimization techniques, denoted as meta-heuristic, has received significant attention due to its simplicity and ability to avoid local optimums imprisonment. In this work, we propose to use meta-heuristic techniques to fine-tune pre-trained weights, exploring additional regions of the search space, and improving their effectiveness. The experimental evaluation comprises two classification tasks (image and text) and is assessed under four literature datasets. Experimental results show nature-inspired algorithms' capacity in exploring the neighborhood of pre-trained weights, achieving superior results than their counterpart pre-trained architectures. Additionally, a thorough analysis of distinct architectures, such as Multi-Layer Perceptron and Recurrent Neural Networks, attempts to visualize and provide more precise insights into the most critical weights to be fine-tuned in the learning process.</description><subject>Algorithms</subject><subject>Heuristic</subject><subject>Heuristic methods</subject><subject>Image classification</subject><subject>Image reconstruction</subject><subject>Machine learning</subject><subject>Multilayer perceptrons</subject><subject>Multilayers</subject><subject>Object recognition</subject><subject>Optimization</subject><subject>Optimization techniques</subject><subject>Recurrent neural networks</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mRw9MwtKMovy8xLVwgoStUNKUrMzEtNUQhPzUzPKClWCMkoyi9Nz1DwTS1J1PVILS3KLC7JTC5WcAOq0g0pzQPq42FgTUvMKU7lhdLcDMpuriHOHrpAcwtLU4tL4rPyS4vygFLxRuamppaWhubGxsbEqQIAaeQ5WQ</recordid><startdate>20221219</startdate><enddate>20221219</enddate><creator>de Rosa, Gustavo H</creator><creator>Roder, Mateus</creator><creator>Papa, João Paulo</creator><creator>Claudio F G dos Santos</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20221219</creationdate><title>Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning</title><author>de Rosa, Gustavo H ; Roder, Mateus ; Papa, João Paulo ; Claudio F G dos Santos</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_27559917333</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Heuristic</topic><topic>Heuristic methods</topic><topic>Image classification</topic><topic>Image reconstruction</topic><topic>Machine learning</topic><topic>Multilayer perceptrons</topic><topic>Multilayers</topic><topic>Object recognition</topic><topic>Optimization</topic><topic>Optimization techniques</topic><topic>Recurrent neural networks</topic><toplevel>online_resources</toplevel><creatorcontrib>de Rosa, Gustavo H</creatorcontrib><creatorcontrib>Roder, Mateus</creatorcontrib><creatorcontrib>Papa, João Paulo</creatorcontrib><creatorcontrib>Claudio F G dos Santos</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>de Rosa, Gustavo H</au><au>Roder, Mateus</au><au>Papa, João Paulo</au><au>Claudio F G dos Santos</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning</atitle><jtitle>arXiv.org</jtitle><date>2022-12-19</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>Machine Learning algorithms have been extensively researched throughout the last decade, leading to unprecedented advances in a broad range of applications, such as image classification and reconstruction, object recognition, and text categorization. Nonetheless, most Machine Learning algorithms are trained via derivative-based optimizers, such as the Stochastic Gradient Descent, leading to possible local optimum entrapments and inhibiting them from achieving proper performances. A bio-inspired alternative to traditional optimization techniques, denoted as meta-heuristic, has received significant attention due to its simplicity and ability to avoid local optimums imprisonment. In this work, we propose to use meta-heuristic techniques to fine-tune pre-trained weights, exploring additional regions of the search space, and improving their effectiveness. The experimental evaluation comprises two classification tasks (image and text) and is assessed under four literature datasets. Experimental results show nature-inspired algorithms' capacity in exploring the neighborhood of pre-trained weights, achieving superior results than their counterpart pre-trained architectures. Additionally, a thorough analysis of distinct architectures, such as Multi-Layer Perceptron and Recurrent Neural Networks, attempts to visualize and provide more precise insights into the most critical weights to be fine-tuned in the learning process.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-12
issn 2331-8422
language eng
recordid cdi_proquest_journals_2755991733
source Freely Accessible Journals
subjects Algorithms
Heuristic
Heuristic methods
Image classification
Image reconstruction
Machine learning
Multilayer perceptrons
Multilayers
Object recognition
Optimization
Optimization techniques
Recurrent neural networks
title Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-14T15%3A28%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Improving%20Pre-Trained%20Weights%20Through%20Meta-Heuristics%20Fine-Tuning&rft.jtitle=arXiv.org&rft.au=de%20Rosa,%20Gustavo%20H&rft.date=2022-12-19&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2755991733%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2755991733&rft_id=info:pmid/&rfr_iscdi=true