Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration

Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-base...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of intelligent & robotic systems 2020-11, Vol.100 (2), p.417-433
Hauptverfasser: Roveda, Loris, Maskani, Jeyhoon, Franceschi, Paolo, Abdi, Arash, Braghin, Francesco, Molinari Tosatti, Lorenzo, Pedrocchi, Nicola
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 433
container_issue 2
container_start_page 417
container_title Journal of intelligent & robotic systems
container_volume 100
creator Roveda, Loris
Maskani, Jeyhoon
Franceschi, Paolo
Abdi, Arash
Braghin, Francesco
Molinari Tosatti, Lorenzo
Pedrocchi, Nicola
description Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a Model-Based Reinforcement Learning (MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a Model Predictive Controller (MPC) with Cross-Entropy Method (CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping impedance control parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.
doi_str_mv 10.1007/s10846-020-01183-3
format Article
fullrecord <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_journals_2449447921</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A724793094</galeid><sourcerecordid>A724793094</sourcerecordid><originalsourceid>FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</originalsourceid><addsrcrecordid>eNp9kE1PxCAQhonRxHX1D3hq4hkdCv3gqBu_kjUmRr14IECHTU0LK3QP_nvRmngzHIa88z4D8xJyyuCcATQXiUEragolUGCs5ZTvkQWrGk5BgNwnC5Aly21ZH5KjlN4BQLaVXJC3h9DhQK90wq54wt67EC2O6KdijTr63m-KVx17bQYs7sctdtpbLFbBTzEMRXYXd7tRe_oUTJiyPgzahKinPvhjcuD0kPDkty7Jy8318-qOrh9v71eXa2pF3U7U1W1nnGG17bSRTDpeGWidwbquGDacZZ2JsjKOa-cElI1khtdNY43Vbb4uydk8dxvDxw7TpN7DLvr8pCqFkEI0efnsOp9dGz2g-l50itrm0-HY2-DR9Vm_bMps5yBFBsoZsDGkFNGpbexHHT8VA_WduppTVzl19ZO64hniM5Sy2W8w_v3lH-oL_9CFeA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2449447921</pqid></control><display><type>article</type><title>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</title><source>SpringerLink Journals - AutoHoldings</source><creator>Roveda, Loris ; Maskani, Jeyhoon ; Franceschi, Paolo ; Abdi, Arash ; Braghin, Francesco ; Molinari Tosatti, Lorenzo ; Pedrocchi, Nicola</creator><creatorcontrib>Roveda, Loris ; Maskani, Jeyhoon ; Franceschi, Paolo ; Abdi, Arash ; Braghin, Francesco ; Molinari Tosatti, Lorenzo ; Pedrocchi, Nicola</creatorcontrib><description>Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a Model-Based Reinforcement Learning (MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a Model Predictive Controller (MPC) with Cross-Entropy Method (CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping impedance control parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.</description><identifier>ISSN: 0921-0296</identifier><identifier>EISSN: 1573-0409</identifier><identifier>DOI: 10.1007/s10846-020-01183-3</identifier><language>eng</language><publisher>Dordrecht: Springer Netherlands</publisher><subject>Analysis ; Artificial Intelligence ; Artificial neural networks ; Collaboration ; Control ; Control methods ; Controllers ; Damping ; Dynamic models ; Electrical Engineering ; Engineering ; Fatigue ; Human engineering ; Impedance ; Industrial applications ; Injury prevention ; Learning ; Machine learning ; Mechanical Engineering ; Mechatronics ; Neural networks ; Predictive control ; Robot control ; Robotics ; Robotics industry ; Robots ; Stiffness ; Weight</subject><ispartof>Journal of intelligent &amp; robotic systems, 2020-11, Vol.100 (2), p.417-433</ispartof><rights>Springer Nature B.V. 2020</rights><rights>COPYRIGHT 2020 Springer</rights><rights>Springer Nature B.V. 2020.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</citedby><cites>FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</cites><orcidid>0000-0002-4427-536X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10846-020-01183-3$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10846-020-01183-3$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Roveda, Loris</creatorcontrib><creatorcontrib>Maskani, Jeyhoon</creatorcontrib><creatorcontrib>Franceschi, Paolo</creatorcontrib><creatorcontrib>Abdi, Arash</creatorcontrib><creatorcontrib>Braghin, Francesco</creatorcontrib><creatorcontrib>Molinari Tosatti, Lorenzo</creatorcontrib><creatorcontrib>Pedrocchi, Nicola</creatorcontrib><title>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</title><title>Journal of intelligent &amp; robotic systems</title><addtitle>J Intell Robot Syst</addtitle><description>Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a Model-Based Reinforcement Learning (MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a Model Predictive Controller (MPC) with Cross-Entropy Method (CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping impedance control parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.</description><subject>Analysis</subject><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Collaboration</subject><subject>Control</subject><subject>Control methods</subject><subject>Controllers</subject><subject>Damping</subject><subject>Dynamic models</subject><subject>Electrical Engineering</subject><subject>Engineering</subject><subject>Fatigue</subject><subject>Human engineering</subject><subject>Impedance</subject><subject>Industrial applications</subject><subject>Injury prevention</subject><subject>Learning</subject><subject>Machine learning</subject><subject>Mechanical Engineering</subject><subject>Mechatronics</subject><subject>Neural networks</subject><subject>Predictive control</subject><subject>Robot control</subject><subject>Robotics</subject><subject>Robotics industry</subject><subject>Robots</subject><subject>Stiffness</subject><subject>Weight</subject><issn>0921-0296</issn><issn>1573-0409</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNp9kE1PxCAQhonRxHX1D3hq4hkdCv3gqBu_kjUmRr14IECHTU0LK3QP_nvRmngzHIa88z4D8xJyyuCcATQXiUEragolUGCs5ZTvkQWrGk5BgNwnC5Aly21ZH5KjlN4BQLaVXJC3h9DhQK90wq54wt67EC2O6KdijTr63m-KVx17bQYs7sctdtpbLFbBTzEMRXYXd7tRe_oUTJiyPgzahKinPvhjcuD0kPDkty7Jy8318-qOrh9v71eXa2pF3U7U1W1nnGG17bSRTDpeGWidwbquGDacZZ2JsjKOa-cElI1khtdNY43Vbb4uydk8dxvDxw7TpN7DLvr8pCqFkEI0efnsOp9dGz2g-l50itrm0-HY2-DR9Vm_bMps5yBFBsoZsDGkFNGpbexHHT8VA_WduppTVzl19ZO64hniM5Sy2W8w_v3lH-oL_9CFeA</recordid><startdate>20201101</startdate><enddate>20201101</enddate><creator>Roveda, Loris</creator><creator>Maskani, Jeyhoon</creator><creator>Franceschi, Paolo</creator><creator>Abdi, Arash</creator><creator>Braghin, Francesco</creator><creator>Molinari Tosatti, Lorenzo</creator><creator>Pedrocchi, Nicola</creator><general>Springer Netherlands</general><general>Springer</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>7XB</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FR3</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L6V</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-4427-536X</orcidid></search><sort><creationdate>20201101</creationdate><title>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</title><author>Roveda, Loris ; Maskani, Jeyhoon ; Franceschi, Paolo ; Abdi, Arash ; Braghin, Francesco ; Molinari Tosatti, Lorenzo ; Pedrocchi, Nicola</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Analysis</topic><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Collaboration</topic><topic>Control</topic><topic>Control methods</topic><topic>Controllers</topic><topic>Damping</topic><topic>Dynamic models</topic><topic>Electrical Engineering</topic><topic>Engineering</topic><topic>Fatigue</topic><topic>Human engineering</topic><topic>Impedance</topic><topic>Industrial applications</topic><topic>Injury prevention</topic><topic>Learning</topic><topic>Machine learning</topic><topic>Mechanical Engineering</topic><topic>Mechatronics</topic><topic>Neural networks</topic><topic>Predictive control</topic><topic>Robot control</topic><topic>Robotics</topic><topic>Robotics industry</topic><topic>Robots</topic><topic>Stiffness</topic><topic>Weight</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Roveda, Loris</creatorcontrib><creatorcontrib>Maskani, Jeyhoon</creatorcontrib><creatorcontrib>Franceschi, Paolo</creatorcontrib><creatorcontrib>Abdi, Arash</creatorcontrib><creatorcontrib>Braghin, Francesco</creatorcontrib><creatorcontrib>Molinari Tosatti, Lorenzo</creatorcontrib><creatorcontrib>Pedrocchi, Nicola</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Engineering Research Database</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>ProQuest Engineering Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Engineering Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>ProQuest Central Basic</collection><jtitle>Journal of intelligent &amp; robotic systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Roveda, Loris</au><au>Maskani, Jeyhoon</au><au>Franceschi, Paolo</au><au>Abdi, Arash</au><au>Braghin, Francesco</au><au>Molinari Tosatti, Lorenzo</au><au>Pedrocchi, Nicola</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</atitle><jtitle>Journal of intelligent &amp; robotic systems</jtitle><stitle>J Intell Robot Syst</stitle><date>2020-11-01</date><risdate>2020</risdate><volume>100</volume><issue>2</issue><spage>417</spage><epage>433</epage><pages>417-433</pages><issn>0921-0296</issn><eissn>1573-0409</eissn><abstract>Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a Model-Based Reinforcement Learning (MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a Model Predictive Controller (MPC) with Cross-Entropy Method (CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping impedance control parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.</abstract><cop>Dordrecht</cop><pub>Springer Netherlands</pub><doi>10.1007/s10846-020-01183-3</doi><tpages>17</tpages><orcidid>https://orcid.org/0000-0002-4427-536X</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0921-0296
ispartof Journal of intelligent & robotic systems, 2020-11, Vol.100 (2), p.417-433
issn 0921-0296
1573-0409
language eng
recordid cdi_proquest_journals_2449447921
source SpringerLink Journals - AutoHoldings
subjects Analysis
Artificial Intelligence
Artificial neural networks
Collaboration
Control
Control methods
Controllers
Damping
Dynamic models
Electrical Engineering
Engineering
Fatigue
Human engineering
Impedance
Industrial applications
Injury prevention
Learning
Machine learning
Mechanical Engineering
Mechatronics
Neural networks
Predictive control
Robot control
Robotics
Robotics industry
Robots
Stiffness
Weight
title Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T06%3A01%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Model-Based%20Reinforcement%20Learning%20Variable%20Impedance%20Control%20for%20Human-Robot%20Collaboration&rft.jtitle=Journal%20of%20intelligent%20&%20robotic%20systems&rft.au=Roveda,%20Loris&rft.date=2020-11-01&rft.volume=100&rft.issue=2&rft.spage=417&rft.epage=433&rft.pages=417-433&rft.issn=0921-0296&rft.eissn=1573-0409&rft_id=info:doi/10.1007/s10846-020-01183-3&rft_dat=%3Cgale_proqu%3EA724793094%3C/gale_proqu%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2449447921&rft_id=info:pmid/&rft_galeid=A724793094&rfr_iscdi=true