Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems

This paper introduces three hybrid algorithms that help in solving global optimization problems using reinforcement learning along with metaheuristic methods. Using the algorithms presented, the search agents try to find a global optimum avoiding the local optima trap. Compared to the classical meta...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Knowledge-based systems 2021-07, Vol.223, p.107044, Article 107044
Hauptverfasser: Seyyedabbasi, Amir, Aliyev, Royal, Kiani, Farzad, Gulle, Murat Ugur, Basyildiz, Hasan, Shah, Mohammed Ahmed
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 107044
container_title Knowledge-based systems
container_volume 223
creator Seyyedabbasi, Amir
Aliyev, Royal
Kiani, Farzad
Gulle, Murat Ugur
Basyildiz, Hasan
Shah, Mohammed Ahmed
description This paper introduces three hybrid algorithms that help in solving global optimization problems using reinforcement learning along with metaheuristic methods. Using the algorithms presented, the search agents try to find a global optimum avoiding the local optima trap. Compared to the classical metaheuristic approaches, the proposed algorithms display higher success in finding new areas as well as exhibiting a more balanced performance while in the exploration and exploitation phases. The algorithms employ reinforcement agents to select an environment based on predefined actions and tasks. A reward and penalty system is used by the agents to discover the environment, done dynamically without following a predetermined model or method. The study makes use of Q-Learning method in all three metaheuristic algorithms, so-called RLI−GWO, RLEx−GWO, and RLWOA algorithms, so as to check and control exploration and exploitation with Q-Table. The Q-Table values guide the search agents of the metaheuristic algorithms to select between the exploration and exploitation phases. A control mechanism is used to get the reward and penalty values for each action. The algorithms presented in this paper are simulated over 30 benchmark functions from CEC 2014, 2015 and the results obtained are compared with well-known metaheuristic and hybrid algorithms (GWO, RLGWO, I-GWO, Ex-GWO, and WOA). The proposed methods have also been applied to the inverse kinematics of the robot arms problem. The results of the used algorithms demonstrate that RLWOA provides better solutions for relevant problems. •Global optimization problems are generally NP-hard problem where metaheuristic algorithms find the optimal solution in the search space.•Reinforcement learning methods give high success rate in finding new global areas compared with metaheuristics and have a more balanced behavior.•Three metaheuristic-reinforcement learning hybrid algorithms are proposed switching between exploration and exploitation phases as and when needed making them more successful in finding better optimized solutions.•They have been applied over 30 benchmark functions and have been simulated to inverse kinematics of the robot arms problem.
doi_str_mv 10.1016/j.knosys.2021.107044
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2539313062</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0950705121003075</els_id><sourcerecordid>2539313062</sourcerecordid><originalsourceid>FETCH-LOGICAL-c334t-7fa652928af40127d40b2b977cb1cbd5b7fa30c3ce3248f8684c0d5e9148d7a63</originalsourceid><addsrcrecordid>eNp9kEtr3DAQx0VpoNs036AHQc_ejB627EuhhDYJBHJJz0KP8a62trSVtIFtv3y9dc89DTP8H8yPkI8MtgxYd3vY_oipnMuWA2fLSYGUb8iG9Yo3SsLwlmxgaKFR0LJ35H0pBwDgnPUb8vvhbHPw1Ey7lEPdz4VaU9DTFKlLsw0xxB3NGOKYssMZY6UTmvz3bKKnM1azx1MOpQZ32fbJF1oTLWl6RbqbkjUTTcca5vDL1LDkHnOyE87lA7kazVTw5t-8Jt-_fX25e2ienu8f7748NU4IWRs1mq7lA-_NKIFx5SVYbgelnGXO-tYuAgFOOBRc9mPf9dKBb3FgsvfKdOKafFpzl-KfJyxVH9Ipx6VS81YMggno-KKSq8rlVErGUR9zmE0-awb6glkf9IpZXzDrFfNi-7zacPngNWDWxQWMDn3I6Kr2Kfw_4A9NkYs9</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2539313062</pqid></control><display><type>article</type><title>Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems</title><source>Elsevier ScienceDirect Journals</source><creator>Seyyedabbasi, Amir ; Aliyev, Royal ; Kiani, Farzad ; Gulle, Murat Ugur ; Basyildiz, Hasan ; Shah, Mohammed Ahmed</creator><creatorcontrib>Seyyedabbasi, Amir ; Aliyev, Royal ; Kiani, Farzad ; Gulle, Murat Ugur ; Basyildiz, Hasan ; Shah, Mohammed Ahmed</creatorcontrib><description>This paper introduces three hybrid algorithms that help in solving global optimization problems using reinforcement learning along with metaheuristic methods. Using the algorithms presented, the search agents try to find a global optimum avoiding the local optima trap. Compared to the classical metaheuristic approaches, the proposed algorithms display higher success in finding new areas as well as exhibiting a more balanced performance while in the exploration and exploitation phases. The algorithms employ reinforcement agents to select an environment based on predefined actions and tasks. A reward and penalty system is used by the agents to discover the environment, done dynamically without following a predetermined model or method. The study makes use of Q-Learning method in all three metaheuristic algorithms, so-called RLI−GWO, RLEx−GWO, and RLWOA algorithms, so as to check and control exploration and exploitation with Q-Table. The Q-Table values guide the search agents of the metaheuristic algorithms to select between the exploration and exploitation phases. A control mechanism is used to get the reward and penalty values for each action. The algorithms presented in this paper are simulated over 30 benchmark functions from CEC 2014, 2015 and the results obtained are compared with well-known metaheuristic and hybrid algorithms (GWO, RLGWO, I-GWO, Ex-GWO, and WOA). The proposed methods have also been applied to the inverse kinematics of the robot arms problem. The results of the used algorithms demonstrate that RLWOA provides better solutions for relevant problems. •Global optimization problems are generally NP-hard problem where metaheuristic algorithms find the optimal solution in the search space.•Reinforcement learning methods give high success rate in finding new global areas compared with metaheuristics and have a more balanced behavior.•Three metaheuristic-reinforcement learning hybrid algorithms are proposed switching between exploration and exploitation phases as and when needed making them more successful in finding better optimized solutions.•They have been applied over 30 benchmark functions and have been simulated to inverse kinematics of the robot arms problem.</description><identifier>ISSN: 0950-7051</identifier><identifier>EISSN: 1872-7409</identifier><identifier>DOI: 10.1016/j.knosys.2021.107044</identifier><language>eng</language><publisher>Amsterdam: Elsevier B.V</publisher><subject>Algorithms ; Exploitation ; Exploration ; Global optimization ; Grey wolf optimization algorithm ; Heuristic methods ; Inverse kinematics ; Machine learning ; Metaheuristic algorithm ; Q-learning ; Reinforcement learning algorithm ; Robot arms ; Whale optimization algorithm</subject><ispartof>Knowledge-based systems, 2021-07, Vol.223, p.107044, Article 107044</ispartof><rights>2021 Elsevier B.V.</rights><rights>Copyright Elsevier Science Ltd. Jul 8, 2021</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c334t-7fa652928af40127d40b2b977cb1cbd5b7fa30c3ce3248f8684c0d5e9148d7a63</citedby><cites>FETCH-LOGICAL-c334t-7fa652928af40127d40b2b977cb1cbd5b7fa30c3ce3248f8684c0d5e9148d7a63</cites><orcidid>0000-0003-3847-4140 ; 0000-0001-8021-4639 ; 0000-0001-5186-4499 ; 0000-0001-8289-7840</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.knosys.2021.107044$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,777,781,3537,27905,27906,45976</link.rule.ids></links><search><creatorcontrib>Seyyedabbasi, Amir</creatorcontrib><creatorcontrib>Aliyev, Royal</creatorcontrib><creatorcontrib>Kiani, Farzad</creatorcontrib><creatorcontrib>Gulle, Murat Ugur</creatorcontrib><creatorcontrib>Basyildiz, Hasan</creatorcontrib><creatorcontrib>Shah, Mohammed Ahmed</creatorcontrib><title>Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems</title><title>Knowledge-based systems</title><description>This paper introduces three hybrid algorithms that help in solving global optimization problems using reinforcement learning along with metaheuristic methods. Using the algorithms presented, the search agents try to find a global optimum avoiding the local optima trap. Compared to the classical metaheuristic approaches, the proposed algorithms display higher success in finding new areas as well as exhibiting a more balanced performance while in the exploration and exploitation phases. The algorithms employ reinforcement agents to select an environment based on predefined actions and tasks. A reward and penalty system is used by the agents to discover the environment, done dynamically without following a predetermined model or method. The study makes use of Q-Learning method in all three metaheuristic algorithms, so-called RLI−GWO, RLEx−GWO, and RLWOA algorithms, so as to check and control exploration and exploitation with Q-Table. The Q-Table values guide the search agents of the metaheuristic algorithms to select between the exploration and exploitation phases. A control mechanism is used to get the reward and penalty values for each action. The algorithms presented in this paper are simulated over 30 benchmark functions from CEC 2014, 2015 and the results obtained are compared with well-known metaheuristic and hybrid algorithms (GWO, RLGWO, I-GWO, Ex-GWO, and WOA). The proposed methods have also been applied to the inverse kinematics of the robot arms problem. The results of the used algorithms demonstrate that RLWOA provides better solutions for relevant problems. •Global optimization problems are generally NP-hard problem where metaheuristic algorithms find the optimal solution in the search space.•Reinforcement learning methods give high success rate in finding new global areas compared with metaheuristics and have a more balanced behavior.•Three metaheuristic-reinforcement learning hybrid algorithms are proposed switching between exploration and exploitation phases as and when needed making them more successful in finding better optimized solutions.•They have been applied over 30 benchmark functions and have been simulated to inverse kinematics of the robot arms problem.</description><subject>Algorithms</subject><subject>Exploitation</subject><subject>Exploration</subject><subject>Global optimization</subject><subject>Grey wolf optimization algorithm</subject><subject>Heuristic methods</subject><subject>Inverse kinematics</subject><subject>Machine learning</subject><subject>Metaheuristic algorithm</subject><subject>Q-learning</subject><subject>Reinforcement learning algorithm</subject><subject>Robot arms</subject><subject>Whale optimization algorithm</subject><issn>0950-7051</issn><issn>1872-7409</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kEtr3DAQx0VpoNs036AHQc_ejB627EuhhDYJBHJJz0KP8a62trSVtIFtv3y9dc89DTP8H8yPkI8MtgxYd3vY_oipnMuWA2fLSYGUb8iG9Yo3SsLwlmxgaKFR0LJ35H0pBwDgnPUb8vvhbHPw1Ey7lEPdz4VaU9DTFKlLsw0xxB3NGOKYssMZY6UTmvz3bKKnM1azx1MOpQZ32fbJF1oTLWl6RbqbkjUTTcca5vDL1LDkHnOyE87lA7kazVTw5t-8Jt-_fX25e2ienu8f7748NU4IWRs1mq7lA-_NKIFx5SVYbgelnGXO-tYuAgFOOBRc9mPf9dKBb3FgsvfKdOKafFpzl-KfJyxVH9Ipx6VS81YMggno-KKSq8rlVErGUR9zmE0-awb6glkf9IpZXzDrFfNi-7zacPngNWDWxQWMDn3I6Kr2Kfw_4A9NkYs9</recordid><startdate>20210708</startdate><enddate>20210708</enddate><creator>Seyyedabbasi, Amir</creator><creator>Aliyev, Royal</creator><creator>Kiani, Farzad</creator><creator>Gulle, Murat Ugur</creator><creator>Basyildiz, Hasan</creator><creator>Shah, Mohammed Ahmed</creator><general>Elsevier B.V</general><general>Elsevier Science Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>E3H</scope><scope>F2A</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0003-3847-4140</orcidid><orcidid>https://orcid.org/0000-0001-8021-4639</orcidid><orcidid>https://orcid.org/0000-0001-5186-4499</orcidid><orcidid>https://orcid.org/0000-0001-8289-7840</orcidid></search><sort><creationdate>20210708</creationdate><title>Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems</title><author>Seyyedabbasi, Amir ; Aliyev, Royal ; Kiani, Farzad ; Gulle, Murat Ugur ; Basyildiz, Hasan ; Shah, Mohammed Ahmed</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c334t-7fa652928af40127d40b2b977cb1cbd5b7fa30c3ce3248f8684c0d5e9148d7a63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Algorithms</topic><topic>Exploitation</topic><topic>Exploration</topic><topic>Global optimization</topic><topic>Grey wolf optimization algorithm</topic><topic>Heuristic methods</topic><topic>Inverse kinematics</topic><topic>Machine learning</topic><topic>Metaheuristic algorithm</topic><topic>Q-learning</topic><topic>Reinforcement learning algorithm</topic><topic>Robot arms</topic><topic>Whale optimization algorithm</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Seyyedabbasi, Amir</creatorcontrib><creatorcontrib>Aliyev, Royal</creatorcontrib><creatorcontrib>Kiani, Farzad</creatorcontrib><creatorcontrib>Gulle, Murat Ugur</creatorcontrib><creatorcontrib>Basyildiz, Hasan</creatorcontrib><creatorcontrib>Shah, Mohammed Ahmed</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>Library &amp; Information Sciences Abstracts (LISA)</collection><collection>Library &amp; Information Science Abstracts (LISA)</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Knowledge-based systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Seyyedabbasi, Amir</au><au>Aliyev, Royal</au><au>Kiani, Farzad</au><au>Gulle, Murat Ugur</au><au>Basyildiz, Hasan</au><au>Shah, Mohammed Ahmed</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems</atitle><jtitle>Knowledge-based systems</jtitle><date>2021-07-08</date><risdate>2021</risdate><volume>223</volume><spage>107044</spage><pages>107044-</pages><artnum>107044</artnum><issn>0950-7051</issn><eissn>1872-7409</eissn><abstract>This paper introduces three hybrid algorithms that help in solving global optimization problems using reinforcement learning along with metaheuristic methods. Using the algorithms presented, the search agents try to find a global optimum avoiding the local optima trap. Compared to the classical metaheuristic approaches, the proposed algorithms display higher success in finding new areas as well as exhibiting a more balanced performance while in the exploration and exploitation phases. The algorithms employ reinforcement agents to select an environment based on predefined actions and tasks. A reward and penalty system is used by the agents to discover the environment, done dynamically without following a predetermined model or method. The study makes use of Q-Learning method in all three metaheuristic algorithms, so-called RLI−GWO, RLEx−GWO, and RLWOA algorithms, so as to check and control exploration and exploitation with Q-Table. The Q-Table values guide the search agents of the metaheuristic algorithms to select between the exploration and exploitation phases. A control mechanism is used to get the reward and penalty values for each action. The algorithms presented in this paper are simulated over 30 benchmark functions from CEC 2014, 2015 and the results obtained are compared with well-known metaheuristic and hybrid algorithms (GWO, RLGWO, I-GWO, Ex-GWO, and WOA). The proposed methods have also been applied to the inverse kinematics of the robot arms problem. The results of the used algorithms demonstrate that RLWOA provides better solutions for relevant problems. •Global optimization problems are generally NP-hard problem where metaheuristic algorithms find the optimal solution in the search space.•Reinforcement learning methods give high success rate in finding new global areas compared with metaheuristics and have a more balanced behavior.•Three metaheuristic-reinforcement learning hybrid algorithms are proposed switching between exploration and exploitation phases as and when needed making them more successful in finding better optimized solutions.•They have been applied over 30 benchmark functions and have been simulated to inverse kinematics of the robot arms problem.</abstract><cop>Amsterdam</cop><pub>Elsevier B.V</pub><doi>10.1016/j.knosys.2021.107044</doi><orcidid>https://orcid.org/0000-0003-3847-4140</orcidid><orcidid>https://orcid.org/0000-0001-8021-4639</orcidid><orcidid>https://orcid.org/0000-0001-5186-4499</orcidid><orcidid>https://orcid.org/0000-0001-8289-7840</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0950-7051
ispartof Knowledge-based systems, 2021-07, Vol.223, p.107044, Article 107044
issn 0950-7051
1872-7409
language eng
recordid cdi_proquest_journals_2539313062
source Elsevier ScienceDirect Journals
subjects Algorithms
Exploitation
Exploration
Global optimization
Grey wolf optimization algorithm
Heuristic methods
Inverse kinematics
Machine learning
Metaheuristic algorithm
Q-learning
Reinforcement learning algorithm
Robot arms
Whale optimization algorithm
title Hybrid algorithms based on combining reinforcement learning and metaheuristic methods to solve global optimization problems
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-19T15%3A12%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Hybrid%20algorithms%20based%20on%20combining%20reinforcement%20learning%20and%20metaheuristic%20methods%20to%20solve%20global%20optimization%20problems&rft.jtitle=Knowledge-based%20systems&rft.au=Seyyedabbasi,%20Amir&rft.date=2021-07-08&rft.volume=223&rft.spage=107044&rft.pages=107044-&rft.artnum=107044&rft.issn=0950-7051&rft.eissn=1872-7409&rft_id=info:doi/10.1016/j.knosys.2021.107044&rft_dat=%3Cproquest_cross%3E2539313062%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2539313062&rft_id=info:pmid/&rft_els_id=S0950705121003075&rfr_iscdi=true