Evaluation of Constrained Reinforcement Learning Algorithms for Legged Locomotion
Shifting from traditional control strategies to Deep Reinforcement Learning (RL) for legged robots poses inherent challenges, especially when addressing real-world physical constraints during training. While high-fidelity simulations provide significant benefits, they often bypass these essential ph...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Shifting from traditional control strategies to Deep Reinforcement Learning
(RL) for legged robots poses inherent challenges, especially when addressing
real-world physical constraints during training. While high-fidelity
simulations provide significant benefits, they often bypass these essential
physical limitations. In this paper, we experiment with the Constrained Markov
Decision Process (CMDP) framework instead of the conventional unconstrained RL
for robotic applications. We perform a comparative study of different
constrained policy optimization algorithms to identify suitable methods for
practical implementation. Our robot experiments demonstrate the critical role
of incorporating physical constraints, yielding successful sim-to-real
transfers, and reducing operational errors on physical systems. The CMDP
formulation streamlines the training process by separately handling constraints
from rewards. Our findings underscore the potential of constrained RL for the
effective development and deployment of learned controllers in robotics. |
---|---|
DOI: | 10.48550/arxiv.2309.15430 |