Asynchronous Event-Based Visual Shape Tracking for Stable Haptic Feedback in Microrobotics

Micromanipulation systems have recently been receiving increased attention. Teleoperated or automated micromanipulation is a challenging task due to the need for high-frequency position or force feedback to guarantee stability. In addition, the integration of sensors within micromanipulation platfor...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on robotics 2012-10, Vol.28 (5), p.1081-1089
Hauptverfasser: Zhenjiang Ni, Bolopion, A., Agnus, J., Benosman, R., Regnier, S.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1089
container_issue 5
container_start_page 1081
container_title IEEE transactions on robotics
container_volume 28
creator Zhenjiang Ni
Bolopion, A.
Agnus, J.
Benosman, R.
Regnier, S.
description Micromanipulation systems have recently been receiving increased attention. Teleoperated or automated micromanipulation is a challenging task due to the need for high-frequency position or force feedback to guarantee stability. In addition, the integration of sensors within micromanipulation platforms is complex. Vision is a commonly used solution for sensing; unfortunately, the update rate of the frame-based acquisition process of current available cameras cannot ensure-at reasonable costs-stable automated or teleoperated control at the microscale level, where low inertia produces highly unreachable dynamic phenomena. This paper presents a novel vision-based microrobotic system combining both an asynchronous address event representation silicon retina and a conventional frame-based camera. Unlike frame-based cameras, recent artificial retinas transmit their outputs as a continuous stream of asynchronous temporal events in a manner similar to the output cells of a biological retina, enabling high update rates. This paper introduces an event-based iterative closest point algorithm to track a microgripper's position at a frequency of 4 kHz. The temporal precision of the asynchronous silicon retina is used to provide a haptic feedback to assist users during manipulation tasks, whereas the frame-based camera is used to retrieve the position of the object that must be manipulated. This paper presents the results of an experiment on teleoperating a sphere of diameter around 50 μm using a piezoelectric gripper in a pick-and-place task.
doi_str_mv 10.1109/TRO.2012.2198930
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_00767638v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>6204348</ieee_id><sourcerecordid>2777348321</sourcerecordid><originalsourceid>FETCH-LOGICAL-c444t-df82aa54247a42caf957c390e63a51235e70960f3281482e06cd428205bb15ba3</originalsourceid><addsrcrecordid>eNo9kM1rGzEQxZfSQNM090IvgpJDDuuOvnalo2OcOuBgiJ0eehGzsrZWul050trg_74yNj7NMO83j5lXFF8pjCgF_WP1shgxoGzEqFaaw4fimmpBSxCV-ph7KVnJQatPxeeU3gCY0MCvi9_jdOjtJoY-7BKZ7l0_lA-Y3Jr88mmHHVlucOvIKqL96_s_pA2RLAdsOkdmuB28JY_OrZusEt-TZ29jiKEJWUhfiqsWu-Ruz_WmeH2criazcr74-TQZz0srhBjKdasYohRM1CiYxVbL2nINruIoKePS1aAraDlTVCjmoLJrwRQD2TRUNshvivuT7wY7s43-H8aDCejNbDw3xxlAXdUVV3ua2e8ndhvD-86lwbyFXezzeYZC9gShgGUKTlT-JqXo2ostBXNM2-S0zTFtc047r9ydjTFZ7NqIvfXpspfvrRSXOnPfTpx3zl3kioHgQvH_bxCFvg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1082004802</pqid></control><display><type>article</type><title>Asynchronous Event-Based Visual Shape Tracking for Stable Haptic Feedback in Microrobotics</title><source>IEEE Electronic Library (IEL)</source><creator>Zhenjiang Ni ; Bolopion, A. ; Agnus, J. ; Benosman, R. ; Regnier, S.</creator><creatorcontrib>Zhenjiang Ni ; Bolopion, A. ; Agnus, J. ; Benosman, R. ; Regnier, S.</creatorcontrib><description>Micromanipulation systems have recently been receiving increased attention. Teleoperated or automated micromanipulation is a challenging task due to the need for high-frequency position or force feedback to guarantee stability. In addition, the integration of sensors within micromanipulation platforms is complex. Vision is a commonly used solution for sensing; unfortunately, the update rate of the frame-based acquisition process of current available cameras cannot ensure-at reasonable costs-stable automated or teleoperated control at the microscale level, where low inertia produces highly unreachable dynamic phenomena. This paper presents a novel vision-based microrobotic system combining both an asynchronous address event representation silicon retina and a conventional frame-based camera. Unlike frame-based cameras, recent artificial retinas transmit their outputs as a continuous stream of asynchronous temporal events in a manner similar to the output cells of a biological retina, enabling high update rates. This paper introduces an event-based iterative closest point algorithm to track a microgripper's position at a frequency of 4 kHz. The temporal precision of the asynchronous silicon retina is used to provide a haptic feedback to assist users during manipulation tasks, whereas the frame-based camera is used to retrieve the position of the object that must be manipulated. This paper presents the results of an experiment on teleoperating a sphere of diameter around 50 μm using a piezoelectric gripper in a pick-and-place task.</description><identifier>ISSN: 1552-3098</identifier><identifier>EISSN: 1941-0468</identifier><identifier>DOI: 10.1109/TRO.2012.2198930</identifier><identifier>CODEN: ITREAE</identifier><language>eng</language><publisher>New York, NY: IEEE</publisher><subject>Access methods and protocols, osi model ; Applied sciences ; Asynchronous ; Automation ; Cameras ; Computer science; control theory; systems ; Computer systems and distributed systems. User interface ; Control theory. Systems ; Dynamic vision sensor ; Engineering Sciences ; Exact sciences and technology ; Feedback ; Grippers ; haptic feedback ; Haptic interfaces ; Mechanical engineering. Machine design ; Micro and nanotechnologies ; Microelectronics ; micromanipulation ; microrobotics ; Precision engineering, watch making ; Retina ; Robotics ; Sensors ; Shape ; Software ; Telecommunications ; Telecommunications and information theory ; Teleprocessing networks. Isdn ; tracking algorithm ; Voltage control</subject><ispartof>IEEE transactions on robotics, 2012-10, Vol.28 (5), p.1081-1089</ispartof><rights>2015 INIST-CNRS</rights><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) Oct 2012</rights><rights>Distributed under a Creative Commons Attribution 4.0 International License</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c444t-df82aa54247a42caf957c390e63a51235e70960f3281482e06cd428205bb15ba3</citedby><cites>FETCH-LOGICAL-c444t-df82aa54247a42caf957c390e63a51235e70960f3281482e06cd428205bb15ba3</cites><orcidid>0009-0000-5867-687X ; 0000-0002-4201-1141</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/6204348$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>230,314,780,784,796,885,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/6204348$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=28268359$$DView record in Pascal Francis$$Hfree_for_read</backlink><backlink>$$Uhttps://hal.science/hal-00767638$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhenjiang Ni</creatorcontrib><creatorcontrib>Bolopion, A.</creatorcontrib><creatorcontrib>Agnus, J.</creatorcontrib><creatorcontrib>Benosman, R.</creatorcontrib><creatorcontrib>Regnier, S.</creatorcontrib><title>Asynchronous Event-Based Visual Shape Tracking for Stable Haptic Feedback in Microrobotics</title><title>IEEE transactions on robotics</title><addtitle>TRO</addtitle><description>Micromanipulation systems have recently been receiving increased attention. Teleoperated or automated micromanipulation is a challenging task due to the need for high-frequency position or force feedback to guarantee stability. In addition, the integration of sensors within micromanipulation platforms is complex. Vision is a commonly used solution for sensing; unfortunately, the update rate of the frame-based acquisition process of current available cameras cannot ensure-at reasonable costs-stable automated or teleoperated control at the microscale level, where low inertia produces highly unreachable dynamic phenomena. This paper presents a novel vision-based microrobotic system combining both an asynchronous address event representation silicon retina and a conventional frame-based camera. Unlike frame-based cameras, recent artificial retinas transmit their outputs as a continuous stream of asynchronous temporal events in a manner similar to the output cells of a biological retina, enabling high update rates. This paper introduces an event-based iterative closest point algorithm to track a microgripper's position at a frequency of 4 kHz. The temporal precision of the asynchronous silicon retina is used to provide a haptic feedback to assist users during manipulation tasks, whereas the frame-based camera is used to retrieve the position of the object that must be manipulated. This paper presents the results of an experiment on teleoperating a sphere of diameter around 50 μm using a piezoelectric gripper in a pick-and-place task.</description><subject>Access methods and protocols, osi model</subject><subject>Applied sciences</subject><subject>Asynchronous</subject><subject>Automation</subject><subject>Cameras</subject><subject>Computer science; control theory; systems</subject><subject>Computer systems and distributed systems. User interface</subject><subject>Control theory. Systems</subject><subject>Dynamic vision sensor</subject><subject>Engineering Sciences</subject><subject>Exact sciences and technology</subject><subject>Feedback</subject><subject>Grippers</subject><subject>haptic feedback</subject><subject>Haptic interfaces</subject><subject>Mechanical engineering. Machine design</subject><subject>Micro and nanotechnologies</subject><subject>Microelectronics</subject><subject>micromanipulation</subject><subject>microrobotics</subject><subject>Precision engineering, watch making</subject><subject>Retina</subject><subject>Robotics</subject><subject>Sensors</subject><subject>Shape</subject><subject>Software</subject><subject>Telecommunications</subject><subject>Telecommunications and information theory</subject><subject>Teleprocessing networks. Isdn</subject><subject>tracking algorithm</subject><subject>Voltage control</subject><issn>1552-3098</issn><issn>1941-0468</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2012</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNo9kM1rGzEQxZfSQNM090IvgpJDDuuOvnalo2OcOuBgiJ0eehGzsrZWul050trg_74yNj7NMO83j5lXFF8pjCgF_WP1shgxoGzEqFaaw4fimmpBSxCV-ph7KVnJQatPxeeU3gCY0MCvi9_jdOjtJoY-7BKZ7l0_lA-Y3Jr88mmHHVlucOvIKqL96_s_pA2RLAdsOkdmuB28JY_OrZusEt-TZ29jiKEJWUhfiqsWu-Ruz_WmeH2criazcr74-TQZz0srhBjKdasYohRM1CiYxVbL2nINruIoKePS1aAraDlTVCjmoLJrwRQD2TRUNshvivuT7wY7s43-H8aDCejNbDw3xxlAXdUVV3ua2e8ndhvD-86lwbyFXezzeYZC9gShgGUKTlT-JqXo2ostBXNM2-S0zTFtc047r9ydjTFZ7NqIvfXpspfvrRSXOnPfTpx3zl3kioHgQvH_bxCFvg</recordid><startdate>20121001</startdate><enddate>20121001</enddate><creator>Zhenjiang Ni</creator><creator>Bolopion, A.</creator><creator>Agnus, J.</creator><creator>Benosman, R.</creator><creator>Regnier, S.</creator><general>IEEE</general><general>Institute of Electrical and Electronics Engineers</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>1XC</scope><scope>VOOES</scope><orcidid>https://orcid.org/0009-0000-5867-687X</orcidid><orcidid>https://orcid.org/0000-0002-4201-1141</orcidid></search><sort><creationdate>20121001</creationdate><title>Asynchronous Event-Based Visual Shape Tracking for Stable Haptic Feedback in Microrobotics</title><author>Zhenjiang Ni ; Bolopion, A. ; Agnus, J. ; Benosman, R. ; Regnier, S.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c444t-df82aa54247a42caf957c390e63a51235e70960f3281482e06cd428205bb15ba3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2012</creationdate><topic>Access methods and protocols, osi model</topic><topic>Applied sciences</topic><topic>Asynchronous</topic><topic>Automation</topic><topic>Cameras</topic><topic>Computer science; control theory; systems</topic><topic>Computer systems and distributed systems. User interface</topic><topic>Control theory. Systems</topic><topic>Dynamic vision sensor</topic><topic>Engineering Sciences</topic><topic>Exact sciences and technology</topic><topic>Feedback</topic><topic>Grippers</topic><topic>haptic feedback</topic><topic>Haptic interfaces</topic><topic>Mechanical engineering. Machine design</topic><topic>Micro and nanotechnologies</topic><topic>Microelectronics</topic><topic>micromanipulation</topic><topic>microrobotics</topic><topic>Precision engineering, watch making</topic><topic>Retina</topic><topic>Robotics</topic><topic>Sensors</topic><topic>Shape</topic><topic>Software</topic><topic>Telecommunications</topic><topic>Telecommunications and information theory</topic><topic>Teleprocessing networks. Isdn</topic><topic>tracking algorithm</topic><topic>Voltage control</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhenjiang Ni</creatorcontrib><creatorcontrib>Bolopion, A.</creatorcontrib><creatorcontrib>Agnus, J.</creatorcontrib><creatorcontrib>Benosman, R.</creatorcontrib><creatorcontrib>Regnier, S.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Hyper Article en Ligne (HAL)</collection><collection>Hyper Article en Ligne (HAL) (Open Access)</collection><jtitle>IEEE transactions on robotics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhenjiang Ni</au><au>Bolopion, A.</au><au>Agnus, J.</au><au>Benosman, R.</au><au>Regnier, S.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Asynchronous Event-Based Visual Shape Tracking for Stable Haptic Feedback in Microrobotics</atitle><jtitle>IEEE transactions on robotics</jtitle><stitle>TRO</stitle><date>2012-10-01</date><risdate>2012</risdate><volume>28</volume><issue>5</issue><spage>1081</spage><epage>1089</epage><pages>1081-1089</pages><issn>1552-3098</issn><eissn>1941-0468</eissn><coden>ITREAE</coden><abstract>Micromanipulation systems have recently been receiving increased attention. Teleoperated or automated micromanipulation is a challenging task due to the need for high-frequency position or force feedback to guarantee stability. In addition, the integration of sensors within micromanipulation platforms is complex. Vision is a commonly used solution for sensing; unfortunately, the update rate of the frame-based acquisition process of current available cameras cannot ensure-at reasonable costs-stable automated or teleoperated control at the microscale level, where low inertia produces highly unreachable dynamic phenomena. This paper presents a novel vision-based microrobotic system combining both an asynchronous address event representation silicon retina and a conventional frame-based camera. Unlike frame-based cameras, recent artificial retinas transmit their outputs as a continuous stream of asynchronous temporal events in a manner similar to the output cells of a biological retina, enabling high update rates. This paper introduces an event-based iterative closest point algorithm to track a microgripper's position at a frequency of 4 kHz. The temporal precision of the asynchronous silicon retina is used to provide a haptic feedback to assist users during manipulation tasks, whereas the frame-based camera is used to retrieve the position of the object that must be manipulated. This paper presents the results of an experiment on teleoperating a sphere of diameter around 50 μm using a piezoelectric gripper in a pick-and-place task.</abstract><cop>New York, NY</cop><pub>IEEE</pub><doi>10.1109/TRO.2012.2198930</doi><tpages>9</tpages><orcidid>https://orcid.org/0009-0000-5867-687X</orcidid><orcidid>https://orcid.org/0000-0002-4201-1141</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1552-3098
ispartof IEEE transactions on robotics, 2012-10, Vol.28 (5), p.1081-1089
issn 1552-3098
1941-0468
language eng
recordid cdi_hal_primary_oai_HAL_hal_00767638v1
source IEEE Electronic Library (IEL)
subjects Access methods and protocols, osi model
Applied sciences
Asynchronous
Automation
Cameras
Computer science
control theory
systems
Computer systems and distributed systems. User interface
Control theory. Systems
Dynamic vision sensor
Engineering Sciences
Exact sciences and technology
Feedback
Grippers
haptic feedback
Haptic interfaces
Mechanical engineering. Machine design
Micro and nanotechnologies
Microelectronics
micromanipulation
microrobotics
Precision engineering, watch making
Retina
Robotics
Sensors
Shape
Software
Telecommunications
Telecommunications and information theory
Teleprocessing networks. Isdn
tracking algorithm
Voltage control
title Asynchronous Event-Based Visual Shape Tracking for Stable Haptic Feedback in Microrobotics
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-29T02%3A17%3A31IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Asynchronous%20Event-Based%20Visual%20Shape%20Tracking%20for%20Stable%20Haptic%20Feedback%20in%20Microrobotics&rft.jtitle=IEEE%20transactions%20on%20robotics&rft.au=Zhenjiang%20Ni&rft.date=2012-10-01&rft.volume=28&rft.issue=5&rft.spage=1081&rft.epage=1089&rft.pages=1081-1089&rft.issn=1552-3098&rft.eissn=1941-0468&rft.coden=ITREAE&rft_id=info:doi/10.1109/TRO.2012.2198930&rft_dat=%3Cproquest_RIE%3E2777348321%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1082004802&rft_id=info:pmid/&rft_ieee_id=6204348&rfr_iscdi=true