UAV Collision Avoidance in Unknown Scenarios with Causal Representation Disentanglement

Deep reinforcement learning (DRL) has significantly advanced online path planning for unmanned aerial vehicles (UAVs). Nonetheless, DRL-based methods often encounter reduced performance when dealing with unfamiliar scenarios. This decline is mainly caused by the presence of non-causal and domain-spe...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Drones (Basel) 2024-12, Vol.9 (1), p.10
Hauptverfasser: Fan, Zhun, Xia, Zihao, Lin, Che, Han, Gaofei, Li, Wenji, Wang, Dongliang, Chen, Yindong, Hao, Zhifeng, Cai, Ruichu, Zhuang, Jiafan
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Deep reinforcement learning (DRL) has significantly advanced online path planning for unmanned aerial vehicles (UAVs). Nonetheless, DRL-based methods often encounter reduced performance when dealing with unfamiliar scenarios. This decline is mainly caused by the presence of non-causal and domain-specific elements within visual representations, which negatively impact the learning of policies. Present techniques generally depend on predefined augmentation or regularization methods intended to direct the model toward identifying causal and domain-invariant components, thereby enhancing the model’s ability to generalize. However, these manually crafted approaches are intrinsically constrained in their coverage and do not consider the entire spectrum of possible scenarios, resulting in less effective performance in new environments. Unlike prior studies, this work prioritizes representation learning and presents a novel method for causal representation disentanglement. The approach successfully distinguishes between causal and non-causal elements in visual data. By concentrating on causal aspects during the policy learning phase, the impact of non-causal factors is minimized, thereby improving the generalizability of DRL models. Experimental results demonstrate that our technique achieves reliable navigation and effective collision avoidance in unseen scenarios, surpassing state-of-the-art deep reinforcement learning algorithms.
ISSN:2504-446X
2504-446X
DOI:10.3390/drones9010010