Heterogeneous multi-player imitation learning

This paper studies imitation learning in nonlinear multi-player game systems with heterogeneous control input dynamics. We propose a model-free data-driven inverse reinforcement learning (RL) algorithm for a leaner to find the cost functions of a N -player Nash expert system given the expert’s state...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Control theory and technology 2023-08, Vol.21 (3), p.281-291
Hauptverfasser: Lian, Bosen, Xue, Wenqian, Lewis, Frank L.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:This paper studies imitation learning in nonlinear multi-player game systems with heterogeneous control input dynamics. We propose a model-free data-driven inverse reinforcement learning (RL) algorithm for a leaner to find the cost functions of a N -player Nash expert system given the expert’s states and control inputs. This allows us to address the imitation learning problem without prior knowledge of the expert’s system dynamics. To achieve this, we provide a basic model-based algorithm that is built upon RL and inverse optimal control. This serves as the foundation for our final model-free inverse RL algorithm which is implemented via neural network-based value function approximators. Theoretical analysis and simulation examples verify the methods.
ISSN:2095-6983
2198-0942
DOI:10.1007/s11768-023-00171-w