Human–robot skills transfer interfaces for a flexible surgical robot

Abstract In minimally invasive surgery, tools go through narrow openings and manipulate soft organs to perform surgical tasks. There are limitations in current robot-assisted surgical systems due to the rigidity of robot tools. The aim of the STIFF-FLOP European project is to develop a soft robotic...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computer methods and programs in biomedicine 2014-09, Vol.116 (2), p.81-96
Hauptverfasser: Calinon, Sylvain, Bruno, Danilo, Malekzadeh, Milad S, Nanayakkara, Thrishantha, Caldwell, Darwin G
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 96
container_issue 2
container_start_page 81
container_title Computer methods and programs in biomedicine
container_volume 116
creator Calinon, Sylvain
Bruno, Danilo
Malekzadeh, Milad S
Nanayakkara, Thrishantha
Caldwell, Darwin G
description Abstract In minimally invasive surgery, tools go through narrow openings and manipulate soft organs to perform surgical tasks. There are limitations in current robot-assisted surgical systems due to the rigidity of robot tools. The aim of the STIFF-FLOP European project is to develop a soft robotic arm to perform surgical tasks. The flexibility of the robot allows the surgeon to move within organs to reach remote areas inside the body and perform challenging procedures in laparoscopy. This article addresses the problem of designing learning interfaces enabling the transfer of skills from human demonstration. Robot programming by demonstration encompasses a wide range of learning strategies, from simple mimicking of the demonstrator's actions to the higher level imitation of the underlying intent extracted from the demonstrations. By focusing on this last form, we study the problem of extracting an objective function explaining the demonstrations from an over-specified set of candidate reward functions, and using this information for self-refinement of the skill. In contrast to inverse reinforcement learning strategies that attempt to explain the observations with reward functions defined for the entire task (or a set of pre-defined reward profiles active for different parts of the task), the proposed approach is based on context-dependent reward-weighted learning, where the robot can learn the relevance of candidate objective functions with respect to the current phase of the task or encountered situation. The robot then exploits this information for skills refinement in the policy parameters space. The proposed approach is tested in simulation with a cutting task performed by the STIFF-FLOP flexible robot, using kinesthetic demonstrations from a Barrett WAM manipulator.
doi_str_mv 10.1016/j.cmpb.2013.12.015
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1551624518</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>1_s2_0_S0169260713004057</els_id><sourcerecordid>1541373799</sourcerecordid><originalsourceid>FETCH-LOGICAL-c444t-2c40e5ac546c7276d54a641e13e1efb306b78ba9c96f2616c920f3d471f3755d3</originalsourceid><addsrcrecordid>eNqNkb9O3TAUhy1UVG5pX6BDlbFLgo__3khVpQpBqYTEAJ0txzlGvjjJrZ0g2HiHviFP0qSXMjCgTmf5vt_wHUI-Aq2AgjraVK7bNhWjwCtgFQW5R1aw1qzUUsk3ZDVDdckU1QfkXc4bSimTUr0lB0yIGtharsjp2dTZ_vHhdxqaYSzyTYgxF2OyffaYitCPmLx1mAs_pMIWPuJdaCIWeUrXwdlY_BXfk31vY8YPT_eQ_Dw9uTo-K88vvv84_nZeOiHEWDInKErrpFBOM61aKawSgMAR0DecqkavG1u7WnmmQLmaUc9bocFzLWXLD8nn3e42Db8mzKPpQnYYo-1xmLIBKUExIWH9H6gArrmu6xllO9SlIeeE3mxT6Gy6N0DNktpszJLaLKkNMDOnnqVPT_tT02H7rPxrOwNfdgDOQW4DJpNdwN5hGxK60bRDeH3_6wvdxdAvyW_wHvNmmFI_pzZg8iyYy-XZy6-BUyqo1PwPhQqkNw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1541373799</pqid></control><display><type>article</type><title>Human–robot skills transfer interfaces for a flexible surgical robot</title><source>MEDLINE</source><source>ScienceDirect Journals (5 years ago - present)</source><creator>Calinon, Sylvain ; Bruno, Danilo ; Malekzadeh, Milad S ; Nanayakkara, Thrishantha ; Caldwell, Darwin G</creator><creatorcontrib>Calinon, Sylvain ; Bruno, Danilo ; Malekzadeh, Milad S ; Nanayakkara, Thrishantha ; Caldwell, Darwin G</creatorcontrib><description>Abstract In minimally invasive surgery, tools go through narrow openings and manipulate soft organs to perform surgical tasks. There are limitations in current robot-assisted surgical systems due to the rigidity of robot tools. The aim of the STIFF-FLOP European project is to develop a soft robotic arm to perform surgical tasks. The flexibility of the robot allows the surgeon to move within organs to reach remote areas inside the body and perform challenging procedures in laparoscopy. This article addresses the problem of designing learning interfaces enabling the transfer of skills from human demonstration. Robot programming by demonstration encompasses a wide range of learning strategies, from simple mimicking of the demonstrator's actions to the higher level imitation of the underlying intent extracted from the demonstrations. By focusing on this last form, we study the problem of extracting an objective function explaining the demonstrations from an over-specified set of candidate reward functions, and using this information for self-refinement of the skill. In contrast to inverse reinforcement learning strategies that attempt to explain the observations with reward functions defined for the entire task (or a set of pre-defined reward profiles active for different parts of the task), the proposed approach is based on context-dependent reward-weighted learning, where the robot can learn the relevance of candidate objective functions with respect to the current phase of the task or encountered situation. The robot then exploits this information for skills refinement in the policy parameters space. The proposed approach is tested in simulation with a cutting task performed by the STIFF-FLOP flexible robot, using kinesthetic demonstrations from a Barrett WAM manipulator.</description><identifier>ISSN: 0169-2607</identifier><identifier>EISSN: 1872-7565</identifier><identifier>DOI: 10.1016/j.cmpb.2013.12.015</identifier><identifier>PMID: 24491285</identifier><language>eng</language><publisher>Ireland: Elsevier Ireland Ltd</publisher><subject>Algorithms ; Artificial Intelligence ; Biomechanical Phenomena ; Computer Simulation ; Humans ; Internal Medicine ; Inverse reinforcement learning ; Learning from demonstration ; Motor Skills ; Other ; Phantoms, Imaging ; Robot-assisted surgery ; Robotic Surgical Procedures - instrumentation ; Robotics - instrumentation ; Skills transfer ; Soft robotics ; Stochastic optimization ; Task Performance and Analysis ; User-Computer Interface</subject><ispartof>Computer methods and programs in biomedicine, 2014-09, Vol.116 (2), p.81-96</ispartof><rights>Elsevier Ireland Ltd</rights><rights>2014 Elsevier Ireland Ltd</rights><rights>Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c444t-2c40e5ac546c7276d54a641e13e1efb306b78ba9c96f2616c920f3d471f3755d3</citedby><cites>FETCH-LOGICAL-c444t-2c40e5ac546c7276d54a641e13e1efb306b78ba9c96f2616c920f3d471f3755d3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.cmpb.2013.12.015$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>314,780,784,3550,27924,27925,45995</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/24491285$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Calinon, Sylvain</creatorcontrib><creatorcontrib>Bruno, Danilo</creatorcontrib><creatorcontrib>Malekzadeh, Milad S</creatorcontrib><creatorcontrib>Nanayakkara, Thrishantha</creatorcontrib><creatorcontrib>Caldwell, Darwin G</creatorcontrib><title>Human–robot skills transfer interfaces for a flexible surgical robot</title><title>Computer methods and programs in biomedicine</title><addtitle>Comput Methods Programs Biomed</addtitle><description>Abstract In minimally invasive surgery, tools go through narrow openings and manipulate soft organs to perform surgical tasks. There are limitations in current robot-assisted surgical systems due to the rigidity of robot tools. The aim of the STIFF-FLOP European project is to develop a soft robotic arm to perform surgical tasks. The flexibility of the robot allows the surgeon to move within organs to reach remote areas inside the body and perform challenging procedures in laparoscopy. This article addresses the problem of designing learning interfaces enabling the transfer of skills from human demonstration. Robot programming by demonstration encompasses a wide range of learning strategies, from simple mimicking of the demonstrator's actions to the higher level imitation of the underlying intent extracted from the demonstrations. By focusing on this last form, we study the problem of extracting an objective function explaining the demonstrations from an over-specified set of candidate reward functions, and using this information for self-refinement of the skill. In contrast to inverse reinforcement learning strategies that attempt to explain the observations with reward functions defined for the entire task (or a set of pre-defined reward profiles active for different parts of the task), the proposed approach is based on context-dependent reward-weighted learning, where the robot can learn the relevance of candidate objective functions with respect to the current phase of the task or encountered situation. The robot then exploits this information for skills refinement in the policy parameters space. The proposed approach is tested in simulation with a cutting task performed by the STIFF-FLOP flexible robot, using kinesthetic demonstrations from a Barrett WAM manipulator.</description><subject>Algorithms</subject><subject>Artificial Intelligence</subject><subject>Biomechanical Phenomena</subject><subject>Computer Simulation</subject><subject>Humans</subject><subject>Internal Medicine</subject><subject>Inverse reinforcement learning</subject><subject>Learning from demonstration</subject><subject>Motor Skills</subject><subject>Other</subject><subject>Phantoms, Imaging</subject><subject>Robot-assisted surgery</subject><subject>Robotic Surgical Procedures - instrumentation</subject><subject>Robotics - instrumentation</subject><subject>Skills transfer</subject><subject>Soft robotics</subject><subject>Stochastic optimization</subject><subject>Task Performance and Analysis</subject><subject>User-Computer Interface</subject><issn>0169-2607</issn><issn>1872-7565</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2014</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNqNkb9O3TAUhy1UVG5pX6BDlbFLgo__3khVpQpBqYTEAJ0txzlGvjjJrZ0g2HiHviFP0qSXMjCgTmf5vt_wHUI-Aq2AgjraVK7bNhWjwCtgFQW5R1aw1qzUUsk3ZDVDdckU1QfkXc4bSimTUr0lB0yIGtharsjp2dTZ_vHhdxqaYSzyTYgxF2OyffaYitCPmLx1mAs_pMIWPuJdaCIWeUrXwdlY_BXfk31vY8YPT_eQ_Dw9uTo-K88vvv84_nZeOiHEWDInKErrpFBOM61aKawSgMAR0DecqkavG1u7WnmmQLmaUc9bocFzLWXLD8nn3e42Db8mzKPpQnYYo-1xmLIBKUExIWH9H6gArrmu6xllO9SlIeeE3mxT6Gy6N0DNktpszJLaLKkNMDOnnqVPT_tT02H7rPxrOwNfdgDOQW4DJpNdwN5hGxK60bRDeH3_6wvdxdAvyW_wHvNmmFI_pzZg8iyYy-XZy6-BUyqo1PwPhQqkNw</recordid><startdate>20140901</startdate><enddate>20140901</enddate><creator>Calinon, Sylvain</creator><creator>Bruno, Danilo</creator><creator>Malekzadeh, Milad S</creator><creator>Nanayakkara, Thrishantha</creator><creator>Caldwell, Darwin G</creator><general>Elsevier Ireland Ltd</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>7QO</scope><scope>8FD</scope><scope>FR3</scope><scope>P64</scope></search><sort><creationdate>20140901</creationdate><title>Human–robot skills transfer interfaces for a flexible surgical robot</title><author>Calinon, Sylvain ; Bruno, Danilo ; Malekzadeh, Milad S ; Nanayakkara, Thrishantha ; Caldwell, Darwin G</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c444t-2c40e5ac546c7276d54a641e13e1efb306b78ba9c96f2616c920f3d471f3755d3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2014</creationdate><topic>Algorithms</topic><topic>Artificial Intelligence</topic><topic>Biomechanical Phenomena</topic><topic>Computer Simulation</topic><topic>Humans</topic><topic>Internal Medicine</topic><topic>Inverse reinforcement learning</topic><topic>Learning from demonstration</topic><topic>Motor Skills</topic><topic>Other</topic><topic>Phantoms, Imaging</topic><topic>Robot-assisted surgery</topic><topic>Robotic Surgical Procedures - instrumentation</topic><topic>Robotics - instrumentation</topic><topic>Skills transfer</topic><topic>Soft robotics</topic><topic>Stochastic optimization</topic><topic>Task Performance and Analysis</topic><topic>User-Computer Interface</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Calinon, Sylvain</creatorcontrib><creatorcontrib>Bruno, Danilo</creatorcontrib><creatorcontrib>Malekzadeh, Milad S</creatorcontrib><creatorcontrib>Nanayakkara, Thrishantha</creatorcontrib><creatorcontrib>Caldwell, Darwin G</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>Biotechnology Research Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>Biotechnology and BioEngineering Abstracts</collection><jtitle>Computer methods and programs in biomedicine</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Calinon, Sylvain</au><au>Bruno, Danilo</au><au>Malekzadeh, Milad S</au><au>Nanayakkara, Thrishantha</au><au>Caldwell, Darwin G</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Human–robot skills transfer interfaces for a flexible surgical robot</atitle><jtitle>Computer methods and programs in biomedicine</jtitle><addtitle>Comput Methods Programs Biomed</addtitle><date>2014-09-01</date><risdate>2014</risdate><volume>116</volume><issue>2</issue><spage>81</spage><epage>96</epage><pages>81-96</pages><issn>0169-2607</issn><eissn>1872-7565</eissn><abstract>Abstract In minimally invasive surgery, tools go through narrow openings and manipulate soft organs to perform surgical tasks. There are limitations in current robot-assisted surgical systems due to the rigidity of robot tools. The aim of the STIFF-FLOP European project is to develop a soft robotic arm to perform surgical tasks. The flexibility of the robot allows the surgeon to move within organs to reach remote areas inside the body and perform challenging procedures in laparoscopy. This article addresses the problem of designing learning interfaces enabling the transfer of skills from human demonstration. Robot programming by demonstration encompasses a wide range of learning strategies, from simple mimicking of the demonstrator's actions to the higher level imitation of the underlying intent extracted from the demonstrations. By focusing on this last form, we study the problem of extracting an objective function explaining the demonstrations from an over-specified set of candidate reward functions, and using this information for self-refinement of the skill. In contrast to inverse reinforcement learning strategies that attempt to explain the observations with reward functions defined for the entire task (or a set of pre-defined reward profiles active for different parts of the task), the proposed approach is based on context-dependent reward-weighted learning, where the robot can learn the relevance of candidate objective functions with respect to the current phase of the task or encountered situation. The robot then exploits this information for skills refinement in the policy parameters space. The proposed approach is tested in simulation with a cutting task performed by the STIFF-FLOP flexible robot, using kinesthetic demonstrations from a Barrett WAM manipulator.</abstract><cop>Ireland</cop><pub>Elsevier Ireland Ltd</pub><pmid>24491285</pmid><doi>10.1016/j.cmpb.2013.12.015</doi><tpages>16</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0169-2607
ispartof Computer methods and programs in biomedicine, 2014-09, Vol.116 (2), p.81-96
issn 0169-2607
1872-7565
language eng
recordid cdi_proquest_miscellaneous_1551624518
source MEDLINE; ScienceDirect Journals (5 years ago - present)
subjects Algorithms
Artificial Intelligence
Biomechanical Phenomena
Computer Simulation
Humans
Internal Medicine
Inverse reinforcement learning
Learning from demonstration
Motor Skills
Other
Phantoms, Imaging
Robot-assisted surgery
Robotic Surgical Procedures - instrumentation
Robotics - instrumentation
Skills transfer
Soft robotics
Stochastic optimization
Task Performance and Analysis
User-Computer Interface
title Human–robot skills transfer interfaces for a flexible surgical robot
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T06%3A41%3A39IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Human%E2%80%93robot%20skills%20transfer%20interfaces%20for%20a%20flexible%20surgical%20robot&rft.jtitle=Computer%20methods%20and%20programs%20in%20biomedicine&rft.au=Calinon,%20Sylvain&rft.date=2014-09-01&rft.volume=116&rft.issue=2&rft.spage=81&rft.epage=96&rft.pages=81-96&rft.issn=0169-2607&rft.eissn=1872-7565&rft_id=info:doi/10.1016/j.cmpb.2013.12.015&rft_dat=%3Cproquest_cross%3E1541373799%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1541373799&rft_id=info:pmid/24491285&rft_els_id=1_s2_0_S0169260713004057&rfr_iscdi=true