Robot motion adaptation through user intervention and reinforcement learning
•Three versions of an interactive framework for robot motion learning are proposed.•Each version requires a different level of user involvement in the learning process.•The experimental results show that user involvement speeds up the learning process.•Combination of human-guided and autonomous lear...
Gespeichert in:
Veröffentlicht in: | Pattern recognition letters 2018-04, Vol.105, p.67-75 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | •Three versions of an interactive framework for robot motion learning are proposed.•Each version requires a different level of user involvement in the learning process.•The experimental results show that user involvement speeds up the learning process.•Combination of human-guided and autonomous learning can improve the learning result.
Assistant robots are designed to perform specific tasks for the user, but their performance is rarely optimal, hence they are required to adapt to user preferences or new task requirements. In the previous work, the potential of an interactive learning framework based on user intervention and reinforcement learning (RL) was assessed. The framework allowed the user to correct an unfitted segment of the robot trajectory by using hand movements to guide the robot along a corrective path. So far, only the usability of the framework was evaluated through experiments with users. In the current work, the framework is described in detail and its ability to learn from a set of sample trajectories using an RL algorithm is analyzed. To evaluate the learning performance, three versions of the framework are proposed that differ in the method used to obtain the sample trajectories, which are: human-guided learning, autonomous learning, and combined human-guided with autonomous learning. The results show that the combination of the human-guided and autonomous learning achieved the best performance, and although it needed a higher number of sample trajectories than the human-guided learning, it required less user involvement. Autonomous learning alone obtained the lowest reward value and needed the highest number of sample trajectories. |
---|---|
ISSN: | 0167-8655 1872-7344 |
DOI: | 10.1016/j.patrec.2017.06.017 |