Path planning of mobile robot in unknown dynamic continuous environment using reward‐modified deep Q‐network

The path planning problem of mobile robot in unknown dynamic environment (UDE) is discussed in this article by building a continuous dynamic simulation environment. To achieve a collision‐free path in UDE, the reinforcement learning theory with deep Q‐network (DQN) is applied for the mobile robot to...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Optimal control applications & methods 2023-05, Vol.44 (3), p.1570-1587
Hauptverfasser: Huang, Runnan, Qin, Chengxuan, Li, Jian Ling, Lan, Xuejing
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The path planning problem of mobile robot in unknown dynamic environment (UDE) is discussed in this article by building a continuous dynamic simulation environment. To achieve a collision‐free path in UDE, the reinforcement learning theory with deep Q‐network (DQN) is applied for the mobile robot to learn optimal decisions. A reward function is designed with weight to balance the obstacle avoidance and the approach to the goal. Moreover, it is found that the relative motion between moving obstacles and robots may cause abnormal rewards and further lead to a collision between robot and obstacle. To address this problem, two reward thresholds are set to modify the abnormal rewards, and the experiments shows that the robot can avoid all obstacles and reach the goal successfully. Finally, double DQN (DDQN) and dueling DQN are applied in this article. This article compares the results of reward‐modified DQN (RMDQN), reward‐modified DDQN (RMDDQN), dueling RMDQN, and dueling RMDDQN and concludes that the result of RMDDQN is the best.
ISSN:0143-2087
1099-1514
DOI:10.1002/oca.2781