Efficient Policy Evaluation with Safety Constraint for Reinforcement Learning
In reinforcement learning, classic on-policy evaluation methods often suffer from high variance and require massive online data to attain the desired accuracy. Previous studies attempt to reduce evaluation variance by searching for or designing proper behavior policies to collect data. However, thes...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | In reinforcement learning, classic on-policy evaluation methods often suffer
from high variance and require massive online data to attain the desired
accuracy. Previous studies attempt to reduce evaluation variance by searching
for or designing proper behavior policies to collect data. However, these
approaches ignore the safety of such behavior policies -- the designed behavior
policies have no safety guarantee and may lead to severe damage during online
executions. In this paper, to address the challenge of reducing variance while
ensuring safety simultaneously, we propose an optimal variance-minimizing
behavior policy under safety constraints. Theoretically, while ensuring safety
constraints, our evaluation method is unbiased and has lower variance than
on-policy evaluation. Empirically, our method is the only existing method to
achieve both substantial variance reduction and safety constraint satisfaction.
Furthermore, we show our method is even superior to previous methods in both
variance reduction and execution safety. |
---|---|
DOI: | 10.48550/arxiv.2410.05655 |