Learning Multipursuit Evasion for Safe Targeted Navigation of Drones

Safe navigation of drones in the presence of adversarial physical attacks from multiple pursuers is a challenging task. This article proposes a novel approach, asynchronous multistage deep reinforcement learning (AMS-DRL), to train adversarial neural networks that can learn from the actions of multi...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on artificial intelligence 2024-12, Vol.5 (12), p.6210-6224
Hauptverfasser: Xiao, Jiaping, Feroskhan, Mir
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Safe navigation of drones in the presence of adversarial physical attacks from multiple pursuers is a challenging task. This article proposes a novel approach, asynchronous multistage deep reinforcement learning (AMS-DRL), to train adversarial neural networks that can learn from the actions of multiple evolved pursuers and adapt quickly to their behavior, enabling the drone to avoid attacks and reach its target. Specifically, AMS-DRL evolves adversarial agents in a pursuit-evasion game (PEG) where the pursuers and the evader are asynchronously trained in a bipartite graph way during multiple stages. Our approach guarantees convergence by ensuring Nash equilibrium (NE) among agents from the game-theory analysis. We evaluate our method in extensive simulations and show that it outperforms baselines with higher navigation success rates (SRs). We also analyze how parameters such as the relative maximum speed affect navigation performance. Furthermore, we have conducted physical experiments and validated the effectiveness of the trained policies in real-time flights. A SR heatmap is introduced to elucidate how spatial geometry influences navigation outcomes.
ISSN:2691-4581
2691-4581
DOI:10.1109/TAI.2024.3366871