Navigating Noisy Feedback: Enhancing Reinforcement Learning with Error-Prone Language Models

The correct specification of reward models is a well-known challenge in reinforcement learning. Hand-crafted reward functions often lead to inefficient or suboptimal policies and may not be aligned with user values. Reinforcement learning from human feedback is a successful technique that can mitiga...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Lin, Muhan, Shi, Shuyang, Guo, Yue, Chalaki, Behdad, Tadiparthi, Vaishnav, Pari, Ehsan Moradi, Stepputtis, Simon, Campbell, Joseph, Sycara, Katia
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!