Learning Agile Locomotion on Risky Terrains

Quadruped robots have shown remarkable mobility on various terrains through reinforcement learning. Yet, in the presence of sparse footholds and risky terrains such as stepping stones and balance beams, which require precise foot placement to avoid falls, model-based approaches are often used. In th...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Zhang, Chong, Rudin, Nikita, Hoeller, David, Hutter, Marco
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Zhang, Chong
Rudin, Nikita
Hoeller, David
Hutter, Marco
description Quadruped robots have shown remarkable mobility on various terrains through reinforcement learning. Yet, in the presence of sparse footholds and risky terrains such as stepping stones and balance beams, which require precise foot placement to avoid falls, model-based approaches are often used. In this paper, we show that end-to-end reinforcement learning can also enable the robot to traverse risky terrains with dynamic motions. To this end, our approach involves training a generalist policy for agile locomotion on disorderly and sparse stepping stones before transferring its reusable knowledge to various more challenging terrains by finetuning specialist policies from it. Given that the robot needs to rapidly adapt its velocity on these terrains, we formulate the task as a navigation task instead of the commonly used velocity tracking which constrains the robot's behavior and propose an exploration strategy to overcome sparse rewards and achieve high robustness. We validate our proposed method through simulation and real-world experiments on an ANYmal-D robot achieving peak forward velocity of >= 2.5 m/s on sparse stepping stones and narrow balance beams. Video: youtu.be/Z5X0J8OH6z4
doi_str_mv 10.48550/arxiv.2311.10484
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_10484</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_10484</sourcerecordid><originalsourceid>FETCH-LOGICAL-a674-b24ab16a04311ca36b75ada90d84ad2b91d5749a5e333ed889aa3dd985773d0f3</originalsourceid><addsrcrecordid>eNotzr0KwjAUBeAsDlJ9ACe7S2vSmzTJWMQ_KAjSvdyaKEHbSiqib2-twoGznfMRMmM05koIukT_cs84AcZiRrniY7LILfrGNZcwu7ibDfP21Nbtw7VN2Ofouus7LKz36JpuQkZnvHV2-u-AFJt1sdpF-WG7X2V5hKnkUZVwrFiKlPc3J4S0kgINamoUR5NUmhkhuUZhAcAapTQiGKOVkBIMPUNA5r_ZQVvevavRv8uvuhzU8AGWazvW</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Learning Agile Locomotion on Risky Terrains</title><source>arXiv.org</source><creator>Zhang, Chong ; Rudin, Nikita ; Hoeller, David ; Hutter, Marco</creator><creatorcontrib>Zhang, Chong ; Rudin, Nikita ; Hoeller, David ; Hutter, Marco</creatorcontrib><description>Quadruped robots have shown remarkable mobility on various terrains through reinforcement learning. Yet, in the presence of sparse footholds and risky terrains such as stepping stones and balance beams, which require precise foot placement to avoid falls, model-based approaches are often used. In this paper, we show that end-to-end reinforcement learning can also enable the robot to traverse risky terrains with dynamic motions. To this end, our approach involves training a generalist policy for agile locomotion on disorderly and sparse stepping stones before transferring its reusable knowledge to various more challenging terrains by finetuning specialist policies from it. Given that the robot needs to rapidly adapt its velocity on these terrains, we formulate the task as a navigation task instead of the commonly used velocity tracking which constrains the robot's behavior and propose an exploration strategy to overcome sparse rewards and achieve high robustness. We validate our proposed method through simulation and real-world experiments on an ANYmal-D robot achieving peak forward velocity of &gt;= 2.5 m/s on sparse stepping stones and narrow balance beams. Video: youtu.be/Z5X0J8OH6z4</description><identifier>DOI: 10.48550/arxiv.2311.10484</identifier><language>eng</language><subject>Computer Science - Robotics</subject><creationdate>2023-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.10484$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.10484$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhang, Chong</creatorcontrib><creatorcontrib>Rudin, Nikita</creatorcontrib><creatorcontrib>Hoeller, David</creatorcontrib><creatorcontrib>Hutter, Marco</creatorcontrib><title>Learning Agile Locomotion on Risky Terrains</title><description>Quadruped robots have shown remarkable mobility on various terrains through reinforcement learning. Yet, in the presence of sparse footholds and risky terrains such as stepping stones and balance beams, which require precise foot placement to avoid falls, model-based approaches are often used. In this paper, we show that end-to-end reinforcement learning can also enable the robot to traverse risky terrains with dynamic motions. To this end, our approach involves training a generalist policy for agile locomotion on disorderly and sparse stepping stones before transferring its reusable knowledge to various more challenging terrains by finetuning specialist policies from it. Given that the robot needs to rapidly adapt its velocity on these terrains, we formulate the task as a navigation task instead of the commonly used velocity tracking which constrains the robot's behavior and propose an exploration strategy to overcome sparse rewards and achieve high robustness. We validate our proposed method through simulation and real-world experiments on an ANYmal-D robot achieving peak forward velocity of &gt;= 2.5 m/s on sparse stepping stones and narrow balance beams. Video: youtu.be/Z5X0J8OH6z4</description><subject>Computer Science - Robotics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzr0KwjAUBeAsDlJ9ACe7S2vSmzTJWMQ_KAjSvdyaKEHbSiqib2-twoGznfMRMmM05koIukT_cs84AcZiRrniY7LILfrGNZcwu7ibDfP21Nbtw7VN2Ofouus7LKz36JpuQkZnvHV2-u-AFJt1sdpF-WG7X2V5hKnkUZVwrFiKlPc3J4S0kgINamoUR5NUmhkhuUZhAcAapTQiGKOVkBIMPUNA5r_ZQVvevavRv8uvuhzU8AGWazvW</recordid><startdate>20231117</startdate><enddate>20231117</enddate><creator>Zhang, Chong</creator><creator>Rudin, Nikita</creator><creator>Hoeller, David</creator><creator>Hutter, Marco</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231117</creationdate><title>Learning Agile Locomotion on Risky Terrains</title><author>Zhang, Chong ; Rudin, Nikita ; Hoeller, David ; Hutter, Marco</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a674-b24ab16a04311ca36b75ada90d84ad2b91d5749a5e333ed889aa3dd985773d0f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Robotics</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Chong</creatorcontrib><creatorcontrib>Rudin, Nikita</creatorcontrib><creatorcontrib>Hoeller, David</creatorcontrib><creatorcontrib>Hutter, Marco</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhang, Chong</au><au>Rudin, Nikita</au><au>Hoeller, David</au><au>Hutter, Marco</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Learning Agile Locomotion on Risky Terrains</atitle><date>2023-11-17</date><risdate>2023</risdate><abstract>Quadruped robots have shown remarkable mobility on various terrains through reinforcement learning. Yet, in the presence of sparse footholds and risky terrains such as stepping stones and balance beams, which require precise foot placement to avoid falls, model-based approaches are often used. In this paper, we show that end-to-end reinforcement learning can also enable the robot to traverse risky terrains with dynamic motions. To this end, our approach involves training a generalist policy for agile locomotion on disorderly and sparse stepping stones before transferring its reusable knowledge to various more challenging terrains by finetuning specialist policies from it. Given that the robot needs to rapidly adapt its velocity on these terrains, we formulate the task as a navigation task instead of the commonly used velocity tracking which constrains the robot's behavior and propose an exploration strategy to overcome sparse rewards and achieve high robustness. We validate our proposed method through simulation and real-world experiments on an ANYmal-D robot achieving peak forward velocity of &gt;= 2.5 m/s on sparse stepping stones and narrow balance beams. Video: youtu.be/Z5X0J8OH6z4</abstract><doi>10.48550/arxiv.2311.10484</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2311.10484
ispartof
issn
language eng
recordid cdi_arxiv_primary_2311_10484
source arXiv.org
subjects Computer Science - Robotics
title Learning Agile Locomotion on Risky Terrains
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T07%3A32%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Learning%20Agile%20Locomotion%20on%20Risky%20Terrains&rft.au=Zhang,%20Chong&rft.date=2023-11-17&rft_id=info:doi/10.48550/arxiv.2311.10484&rft_dat=%3Carxiv_GOX%3E2311_10484%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true