Control Law Learning Based on LQR Reconstruction With Inverse Optimal Control

Designing controllers to generate various trajectories has been studied for years, while recently, recovering an optimal controller from trajectories receives increasing attention. In this paper, we reveal that the inherent linear quadratic regulator (LQR) problem of a moving agent can be reconstruc...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on automatic control 2024-09, p.1-8
Hauptverfasser: Qu, Chendi, He, Jianping, Duan, Xiaoming
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Designing controllers to generate various trajectories has been studied for years, while recently, recovering an optimal controller from trajectories receives increasing attention. In this paper, we reveal that the inherent linear quadratic regulator (LQR) problem of a moving agent can be reconstructed based on its trajectory observations only, which enables one to learn the control law of the target agent autonomously. Specifically, we propose a novel inverse optimal control (IOC) method to identify the weighting matrices of a discrete-time finite horizon LQR, and we also provide the corresponding identifiability conditions. Then we obtain the optimal estimate of the control horizon using binary search and finally reconstruct the LQR problem with above estimates. The strength of learning control law with optimization problem recovery lies in less computation consumption and strong generalization ability. We apply our algorithm to the future control input prediction and the discrepancy loss is further derived. Simulations and hardware experiments on a self-designed robot platform illustrate the effectiveness of our work.
ISSN:0018-9286
1558-2523
DOI:10.1109/TAC.2024.3469788