Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment
Mobile robot path planning in an unknown environment is a fundamental and challenging problem in the field of robotics. Dynamic window approach (DWA) is an effective method of local path planning, however some of its evaluation functions are inadequate and the algorithm for choosing the weights of t...
Gespeichert in:
Veröffentlicht in: | Autonomous robots 2021, Vol.45 (1), p.51-76 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 76 |
---|---|
container_issue | 1 |
container_start_page | 51 |
container_title | Autonomous robots |
container_volume | 45 |
creator | Chang, Lu Shan, Liang Jiang, Chao Dai, Yuewei |
description | Mobile robot path planning in an unknown environment is a fundamental and challenging problem in the field of robotics. Dynamic window approach (DWA) is an effective method of local path planning, however some of its evaluation functions are inadequate and the algorithm for choosing the weights of these functions is lacking, which makes it highly dependent on the global reference and prone to fail in an unknown environment. In this paper, an improved DWA based on Q-learning is proposed. First, the original evaluation functions are modified and extended by adding two new evaluation functions to enhance the performance of global navigation. Then, considering the balance of effectiveness and speed, we define the state space, action space and reward function of the adopted Q-learning algorithm for the robot motion planning. After that, the parameters of the proposed DWA are adaptively learned by Q-learning and a trained agent is obtained to adapt to the unknown environment. At last, by a series of comparative simulations, the proposed method shows higher navigation efficiency and successful rate in the complex unknown environment. The proposed method is also validated in experiments based on XQ-4 Pro robot to verify its navigation capability in both static and dynamic environment. |
doi_str_mv | 10.1007/s10514-020-09947-4 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2484057494</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2484057494</sourcerecordid><originalsourceid>FETCH-LOGICAL-c358t-bf6718ffb605d0841d8a72e23d31353568fe63b96e971f22b5dd12281d9e10563</originalsourceid><addsrcrecordid>eNp9kM1KAzEURoMoWKsv4CrgOnqTTCaTpRT_oCCIrkNmkqnTdpIxmbb07U2t4M7VJZfzfTcchK4p3FIAeZcoCFoQYEBAqUKS4gRNqJCcSMHkKZqAYooIofg5ukhpCQBKAkzQ6s11vg2xcb3zI65Nchb3oe7WDsdQhxEPZvzEw9p43_kF3nX51fVDDNsM2r03fdfkrbdhh82Q96bJgMcbv_Jh57Hz2y4Gf2i_RGetWSd39Tun6OPx4X32TOavTy-z-zlpuKhGUrelpFXb1iUIC1VBbWUkc4xbTrngoqxaV_JalU5J2jJWC2spYxW1ymUNJZ-im2Nv_s3XxqVRL8Mm-nxSs6IqQMhCFZliR6qJIaXoWj3Erjdxrynog1R9lKqzVP0jVR9C_BhKGfYLF_-q_0l9A9GLe4Y</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2484057494</pqid></control><display><type>article</type><title>Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment</title><source>SpringerLink (Online service)</source><creator>Chang, Lu ; Shan, Liang ; Jiang, Chao ; Dai, Yuewei</creator><creatorcontrib>Chang, Lu ; Shan, Liang ; Jiang, Chao ; Dai, Yuewei</creatorcontrib><description>Mobile robot path planning in an unknown environment is a fundamental and challenging problem in the field of robotics. Dynamic window approach (DWA) is an effective method of local path planning, however some of its evaluation functions are inadequate and the algorithm for choosing the weights of these functions is lacking, which makes it highly dependent on the global reference and prone to fail in an unknown environment. In this paper, an improved DWA based on Q-learning is proposed. First, the original evaluation functions are modified and extended by adding two new evaluation functions to enhance the performance of global navigation. Then, considering the balance of effectiveness and speed, we define the state space, action space and reward function of the adopted Q-learning algorithm for the robot motion planning. After that, the parameters of the proposed DWA are adaptively learned by Q-learning and a trained agent is obtained to adapt to the unknown environment. At last, by a series of comparative simulations, the proposed method shows higher navigation efficiency and successful rate in the complex unknown environment. The proposed method is also validated in experiments based on XQ-4 Pro robot to verify its navigation capability in both static and dynamic environment.</description><identifier>ISSN: 0929-5593</identifier><identifier>EISSN: 1573-7527</identifier><identifier>DOI: 10.1007/s10514-020-09947-4</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Algorithms ; Artificial Intelligence ; Computer Imaging ; Control ; Engineering ; Machine learning ; Mechatronics ; Motion planning ; Navigation ; Path planning ; Pattern Recognition and Graphics ; Robot dynamics ; Robotics ; Robotics and Automation ; Robots ; Unknown environments ; Vision</subject><ispartof>Autonomous robots, 2021, Vol.45 (1), p.51-76</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020</rights><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c358t-bf6718ffb605d0841d8a72e23d31353568fe63b96e971f22b5dd12281d9e10563</citedby><cites>FETCH-LOGICAL-c358t-bf6718ffb605d0841d8a72e23d31353568fe63b96e971f22b5dd12281d9e10563</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10514-020-09947-4$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10514-020-09947-4$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Chang, Lu</creatorcontrib><creatorcontrib>Shan, Liang</creatorcontrib><creatorcontrib>Jiang, Chao</creatorcontrib><creatorcontrib>Dai, Yuewei</creatorcontrib><title>Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment</title><title>Autonomous robots</title><addtitle>Auton Robot</addtitle><description>Mobile robot path planning in an unknown environment is a fundamental and challenging problem in the field of robotics. Dynamic window approach (DWA) is an effective method of local path planning, however some of its evaluation functions are inadequate and the algorithm for choosing the weights of these functions is lacking, which makes it highly dependent on the global reference and prone to fail in an unknown environment. In this paper, an improved DWA based on Q-learning is proposed. First, the original evaluation functions are modified and extended by adding two new evaluation functions to enhance the performance of global navigation. Then, considering the balance of effectiveness and speed, we define the state space, action space and reward function of the adopted Q-learning algorithm for the robot motion planning. After that, the parameters of the proposed DWA are adaptively learned by Q-learning and a trained agent is obtained to adapt to the unknown environment. At last, by a series of comparative simulations, the proposed method shows higher navigation efficiency and successful rate in the complex unknown environment. The proposed method is also validated in experiments based on XQ-4 Pro robot to verify its navigation capability in both static and dynamic environment.</description><subject>Algorithms</subject><subject>Artificial Intelligence</subject><subject>Computer Imaging</subject><subject>Control</subject><subject>Engineering</subject><subject>Machine learning</subject><subject>Mechatronics</subject><subject>Motion planning</subject><subject>Navigation</subject><subject>Path planning</subject><subject>Pattern Recognition and Graphics</subject><subject>Robot dynamics</subject><subject>Robotics</subject><subject>Robotics and Automation</subject><subject>Robots</subject><subject>Unknown environments</subject><subject>Vision</subject><issn>0929-5593</issn><issn>1573-7527</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9kM1KAzEURoMoWKsv4CrgOnqTTCaTpRT_oCCIrkNmkqnTdpIxmbb07U2t4M7VJZfzfTcchK4p3FIAeZcoCFoQYEBAqUKS4gRNqJCcSMHkKZqAYooIofg5ukhpCQBKAkzQ6s11vg2xcb3zI65Nchb3oe7WDsdQhxEPZvzEw9p43_kF3nX51fVDDNsM2r03fdfkrbdhh82Q96bJgMcbv_Jh57Hz2y4Gf2i_RGetWSd39Tun6OPx4X32TOavTy-z-zlpuKhGUrelpFXb1iUIC1VBbWUkc4xbTrngoqxaV_JalU5J2jJWC2spYxW1ymUNJZ-im2Nv_s3XxqVRL8Mm-nxSs6IqQMhCFZliR6qJIaXoWj3Erjdxrynog1R9lKqzVP0jVR9C_BhKGfYLF_-q_0l9A9GLe4Y</recordid><startdate>2021</startdate><enddate>2021</enddate><creator>Chang, Lu</creator><creator>Shan, Liang</creator><creator>Jiang, Chao</creator><creator>Dai, Yuewei</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>F28</scope><scope>FR3</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>L6V</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>S0W</scope></search><sort><creationdate>2021</creationdate><title>Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment</title><author>Chang, Lu ; Shan, Liang ; Jiang, Chao ; Dai, Yuewei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c358t-bf6718ffb605d0841d8a72e23d31353568fe63b96e971f22b5dd12281d9e10563</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Algorithms</topic><topic>Artificial Intelligence</topic><topic>Computer Imaging</topic><topic>Control</topic><topic>Engineering</topic><topic>Machine learning</topic><topic>Mechatronics</topic><topic>Motion planning</topic><topic>Navigation</topic><topic>Path planning</topic><topic>Pattern Recognition and Graphics</topic><topic>Robot dynamics</topic><topic>Robotics</topic><topic>Robotics and Automation</topic><topic>Robots</topic><topic>Unknown environments</topic><topic>Vision</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Chang, Lu</creatorcontrib><creatorcontrib>Shan, Liang</creatorcontrib><creatorcontrib>Jiang, Chao</creatorcontrib><creatorcontrib>Dai, Yuewei</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Database (1962 - current)</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>ANTE: Abstracts in New Technology & Engineering</collection><collection>Engineering Research Database</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Engineering Database</collection><collection>ProQuest advanced technologies & aerospace journals</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection><collection>DELNET Engineering & Technology Collection</collection><jtitle>Autonomous robots</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chang, Lu</au><au>Shan, Liang</au><au>Jiang, Chao</au><au>Dai, Yuewei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment</atitle><jtitle>Autonomous robots</jtitle><stitle>Auton Robot</stitle><date>2021</date><risdate>2021</risdate><volume>45</volume><issue>1</issue><spage>51</spage><epage>76</epage><pages>51-76</pages><issn>0929-5593</issn><eissn>1573-7527</eissn><abstract>Mobile robot path planning in an unknown environment is a fundamental and challenging problem in the field of robotics. Dynamic window approach (DWA) is an effective method of local path planning, however some of its evaluation functions are inadequate and the algorithm for choosing the weights of these functions is lacking, which makes it highly dependent on the global reference and prone to fail in an unknown environment. In this paper, an improved DWA based on Q-learning is proposed. First, the original evaluation functions are modified and extended by adding two new evaluation functions to enhance the performance of global navigation. Then, considering the balance of effectiveness and speed, we define the state space, action space and reward function of the adopted Q-learning algorithm for the robot motion planning. After that, the parameters of the proposed DWA are adaptively learned by Q-learning and a trained agent is obtained to adapt to the unknown environment. At last, by a series of comparative simulations, the proposed method shows higher navigation efficiency and successful rate in the complex unknown environment. The proposed method is also validated in experiments based on XQ-4 Pro robot to verify its navigation capability in both static and dynamic environment.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10514-020-09947-4</doi><tpages>26</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0929-5593 |
ispartof | Autonomous robots, 2021, Vol.45 (1), p.51-76 |
issn | 0929-5593 1573-7527 |
language | eng |
recordid | cdi_proquest_journals_2484057494 |
source | SpringerLink (Online service) |
subjects | Algorithms Artificial Intelligence Computer Imaging Control Engineering Machine learning Mechatronics Motion planning Navigation Path planning Pattern Recognition and Graphics Robot dynamics Robotics Robotics and Automation Robots Unknown environments Vision |
title | Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-07T16%3A49%3A35IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Reinforcement%20based%20mobile%20robot%20path%20planning%20with%20improved%20dynamic%20window%20approach%20in%20unknown%20environment&rft.jtitle=Autonomous%20robots&rft.au=Chang,%20Lu&rft.date=2021&rft.volume=45&rft.issue=1&rft.spage=51&rft.epage=76&rft.pages=51-76&rft.issn=0929-5593&rft.eissn=1573-7527&rft_id=info:doi/10.1007/s10514-020-09947-4&rft_dat=%3Cproquest_cross%3E2484057494%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2484057494&rft_id=info:pmid/&rfr_iscdi=true |