Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration
Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-base...
Gespeichert in:
Veröffentlicht in: | Journal of intelligent & robotic systems 2020-11, Vol.100 (2), p.417-433 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 433 |
---|---|
container_issue | 2 |
container_start_page | 417 |
container_title | Journal of intelligent & robotic systems |
container_volume | 100 |
creator | Roveda, Loris Maskani, Jeyhoon Franceschi, Paolo Abdi, Arash Braghin, Francesco Molinari Tosatti, Lorenzo Pedrocchi, Nicola |
description | Industry 4.0
is taking
human-robot collaboration
at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a
Model-Based Reinforcement Learning
(MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a
Model Predictive Controller
(MPC) with
Cross-Entropy Method
(CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping
impedance control
parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system. |
doi_str_mv | 10.1007/s10846-020-01183-3 |
format | Article |
fullrecord | <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_journals_2449447921</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A724793094</galeid><sourcerecordid>A724793094</sourcerecordid><originalsourceid>FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</originalsourceid><addsrcrecordid>eNp9kE1PxCAQhonRxHX1D3hq4hkdCv3gqBu_kjUmRr14IECHTU0LK3QP_nvRmngzHIa88z4D8xJyyuCcATQXiUEragolUGCs5ZTvkQWrGk5BgNwnC5Aly21ZH5KjlN4BQLaVXJC3h9DhQK90wq54wt67EC2O6KdijTr63m-KVx17bQYs7sctdtpbLFbBTzEMRXYXd7tRe_oUTJiyPgzahKinPvhjcuD0kPDkty7Jy8318-qOrh9v71eXa2pF3U7U1W1nnGG17bSRTDpeGWidwbquGDacZZ2JsjKOa-cElI1khtdNY43Vbb4uydk8dxvDxw7TpN7DLvr8pCqFkEI0efnsOp9dGz2g-l50itrm0-HY2-DR9Vm_bMps5yBFBsoZsDGkFNGpbexHHT8VA_WduppTVzl19ZO64hniM5Sy2W8w_v3lH-oL_9CFeA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2449447921</pqid></control><display><type>article</type><title>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</title><source>SpringerLink Journals - AutoHoldings</source><creator>Roveda, Loris ; Maskani, Jeyhoon ; Franceschi, Paolo ; Abdi, Arash ; Braghin, Francesco ; Molinari Tosatti, Lorenzo ; Pedrocchi, Nicola</creator><creatorcontrib>Roveda, Loris ; Maskani, Jeyhoon ; Franceschi, Paolo ; Abdi, Arash ; Braghin, Francesco ; Molinari Tosatti, Lorenzo ; Pedrocchi, Nicola</creatorcontrib><description>Industry 4.0
is taking
human-robot collaboration
at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a
Model-Based Reinforcement Learning
(MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a
Model Predictive Controller
(MPC) with
Cross-Entropy Method
(CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping
impedance control
parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.</description><identifier>ISSN: 0921-0296</identifier><identifier>EISSN: 1573-0409</identifier><identifier>DOI: 10.1007/s10846-020-01183-3</identifier><language>eng</language><publisher>Dordrecht: Springer Netherlands</publisher><subject>Analysis ; Artificial Intelligence ; Artificial neural networks ; Collaboration ; Control ; Control methods ; Controllers ; Damping ; Dynamic models ; Electrical Engineering ; Engineering ; Fatigue ; Human engineering ; Impedance ; Industrial applications ; Injury prevention ; Learning ; Machine learning ; Mechanical Engineering ; Mechatronics ; Neural networks ; Predictive control ; Robot control ; Robotics ; Robotics industry ; Robots ; Stiffness ; Weight</subject><ispartof>Journal of intelligent & robotic systems, 2020-11, Vol.100 (2), p.417-433</ispartof><rights>Springer Nature B.V. 2020</rights><rights>COPYRIGHT 2020 Springer</rights><rights>Springer Nature B.V. 2020.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</citedby><cites>FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</cites><orcidid>0000-0002-4427-536X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10846-020-01183-3$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10846-020-01183-3$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Roveda, Loris</creatorcontrib><creatorcontrib>Maskani, Jeyhoon</creatorcontrib><creatorcontrib>Franceschi, Paolo</creatorcontrib><creatorcontrib>Abdi, Arash</creatorcontrib><creatorcontrib>Braghin, Francesco</creatorcontrib><creatorcontrib>Molinari Tosatti, Lorenzo</creatorcontrib><creatorcontrib>Pedrocchi, Nicola</creatorcontrib><title>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</title><title>Journal of intelligent & robotic systems</title><addtitle>J Intell Robot Syst</addtitle><description>Industry 4.0
is taking
human-robot collaboration
at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a
Model-Based Reinforcement Learning
(MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a
Model Predictive Controller
(MPC) with
Cross-Entropy Method
(CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping
impedance control
parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.</description><subject>Analysis</subject><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Collaboration</subject><subject>Control</subject><subject>Control methods</subject><subject>Controllers</subject><subject>Damping</subject><subject>Dynamic models</subject><subject>Electrical Engineering</subject><subject>Engineering</subject><subject>Fatigue</subject><subject>Human engineering</subject><subject>Impedance</subject><subject>Industrial applications</subject><subject>Injury prevention</subject><subject>Learning</subject><subject>Machine learning</subject><subject>Mechanical Engineering</subject><subject>Mechatronics</subject><subject>Neural networks</subject><subject>Predictive control</subject><subject>Robot control</subject><subject>Robotics</subject><subject>Robotics industry</subject><subject>Robots</subject><subject>Stiffness</subject><subject>Weight</subject><issn>0921-0296</issn><issn>1573-0409</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><recordid>eNp9kE1PxCAQhonRxHX1D3hq4hkdCv3gqBu_kjUmRr14IECHTU0LK3QP_nvRmngzHIa88z4D8xJyyuCcATQXiUEragolUGCs5ZTvkQWrGk5BgNwnC5Aly21ZH5KjlN4BQLaVXJC3h9DhQK90wq54wt67EC2O6KdijTr63m-KVx17bQYs7sctdtpbLFbBTzEMRXYXd7tRe_oUTJiyPgzahKinPvhjcuD0kPDkty7Jy8318-qOrh9v71eXa2pF3U7U1W1nnGG17bSRTDpeGWidwbquGDacZZ2JsjKOa-cElI1khtdNY43Vbb4uydk8dxvDxw7TpN7DLvr8pCqFkEI0efnsOp9dGz2g-l50itrm0-HY2-DR9Vm_bMps5yBFBsoZsDGkFNGpbexHHT8VA_WduppTVzl19ZO64hniM5Sy2W8w_v3lH-oL_9CFeA</recordid><startdate>20201101</startdate><enddate>20201101</enddate><creator>Roveda, Loris</creator><creator>Maskani, Jeyhoon</creator><creator>Franceschi, Paolo</creator><creator>Abdi, Arash</creator><creator>Braghin, Francesco</creator><creator>Molinari Tosatti, Lorenzo</creator><creator>Pedrocchi, Nicola</creator><general>Springer Netherlands</general><general>Springer</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>7XB</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FR3</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K7-</scope><scope>L6V</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0N</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-4427-536X</orcidid></search><sort><creationdate>20201101</creationdate><title>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</title><author>Roveda, Loris ; Maskani, Jeyhoon ; Franceschi, Paolo ; Abdi, Arash ; Braghin, Francesco ; Molinari Tosatti, Lorenzo ; Pedrocchi, Nicola</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c468t-f68dbfb16cdab919f35b08fbe6651e731cda1425bf3aff402791b3677cbca81b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Analysis</topic><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Collaboration</topic><topic>Control</topic><topic>Control methods</topic><topic>Controllers</topic><topic>Damping</topic><topic>Dynamic models</topic><topic>Electrical Engineering</topic><topic>Engineering</topic><topic>Fatigue</topic><topic>Human engineering</topic><topic>Impedance</topic><topic>Industrial applications</topic><topic>Injury prevention</topic><topic>Learning</topic><topic>Machine learning</topic><topic>Mechanical Engineering</topic><topic>Mechatronics</topic><topic>Neural networks</topic><topic>Predictive control</topic><topic>Robot control</topic><topic>Robotics</topic><topic>Robotics industry</topic><topic>Robots</topic><topic>Stiffness</topic><topic>Weight</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Roveda, Loris</creatorcontrib><creatorcontrib>Maskani, Jeyhoon</creatorcontrib><creatorcontrib>Franceschi, Paolo</creatorcontrib><creatorcontrib>Abdi, Arash</creatorcontrib><creatorcontrib>Braghin, Francesco</creatorcontrib><creatorcontrib>Molinari Tosatti, Lorenzo</creatorcontrib><creatorcontrib>Pedrocchi, Nicola</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Engineering Research Database</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>Computer Science Database</collection><collection>ProQuest Engineering Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Computing Database</collection><collection>Engineering Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>ProQuest Central Basic</collection><jtitle>Journal of intelligent & robotic systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Roveda, Loris</au><au>Maskani, Jeyhoon</au><au>Franceschi, Paolo</au><au>Abdi, Arash</au><au>Braghin, Francesco</au><au>Molinari Tosatti, Lorenzo</au><au>Pedrocchi, Nicola</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration</atitle><jtitle>Journal of intelligent & robotic systems</jtitle><stitle>J Intell Robot Syst</stitle><date>2020-11-01</date><risdate>2020</risdate><volume>100</volume><issue>2</issue><spage>417</spage><epage>433</epage><pages>417-433</pages><issn>0921-0296</issn><eissn>1573-0409</eissn><abstract>Industry 4.0
is taking
human-robot collaboration
at the center of the production environment. Collaborative robots enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot interaction dynamics. With this aim, this paper proposes a
Model-Based Reinforcement Learning
(MBRL) variable impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a
Model Predictive Controller
(MPC) with
Cross-Entropy Method
(CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping
impedance control
parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.</abstract><cop>Dordrecht</cop><pub>Springer Netherlands</pub><doi>10.1007/s10846-020-01183-3</doi><tpages>17</tpages><orcidid>https://orcid.org/0000-0002-4427-536X</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0921-0296 |
ispartof | Journal of intelligent & robotic systems, 2020-11, Vol.100 (2), p.417-433 |
issn | 0921-0296 1573-0409 |
language | eng |
recordid | cdi_proquest_journals_2449447921 |
source | SpringerLink Journals - AutoHoldings |
subjects | Analysis Artificial Intelligence Artificial neural networks Collaboration Control Control methods Controllers Damping Dynamic models Electrical Engineering Engineering Fatigue Human engineering Impedance Industrial applications Injury prevention Learning Machine learning Mechanical Engineering Mechatronics Neural networks Predictive control Robot control Robotics Robotics industry Robots Stiffness Weight |
title | Model-Based Reinforcement Learning Variable Impedance Control for Human-Robot Collaboration |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T06%3A01%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Model-Based%20Reinforcement%20Learning%20Variable%20Impedance%20Control%20for%20Human-Robot%20Collaboration&rft.jtitle=Journal%20of%20intelligent%20&%20robotic%20systems&rft.au=Roveda,%20Loris&rft.date=2020-11-01&rft.volume=100&rft.issue=2&rft.spage=417&rft.epage=433&rft.pages=417-433&rft.issn=0921-0296&rft.eissn=1573-0409&rft_id=info:doi/10.1007/s10846-020-01183-3&rft_dat=%3Cgale_proqu%3EA724793094%3C/gale_proqu%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2449447921&rft_id=info:pmid/&rft_galeid=A724793094&rfr_iscdi=true |