COMPUTERIZED SYSTEM AND METHOD USING DIFFERENT IMAGE VIEWS TO FIND GRASP LOCATIONS AND TRAJECTORIES FOR ROBOTIC PICK UP

Computerized system and method are provided. A robotic manipulator (12) is arranged to grasp objects (20). A gripper (16) is attached to robotic manipulator (12), which includes an imaging sensor (14). During motion of robotic manipulator (12), imaging sensor (14) is arranged to capture images provi...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: SOLOWJOW, Eugen, WEN, Changtao, SEHR, Martin, APARICIO OJEA, Juan L, CLAUSSEN, Heiko
Format: Patent
Sprache:eng ; fre ; ger
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator SOLOWJOW, Eugen
WEN, Changtao
SEHR, Martin
APARICIO OJEA, Juan L
CLAUSSEN, Heiko
description Computerized system and method are provided. A robotic manipulator (12) is arranged to grasp objects (20). A gripper (16) is attached to robotic manipulator (12), which includes an imaging sensor (14). During motion of robotic manipulator (12), imaging sensor (14) is arranged to capture images providing different views of objects in the environment of the robotic manipulator. A processor (18) is configured to find, based on the different views, candidate grasp locations and trajectories to perform a grasp of a respective object in the environment of the robotic manipulator. Processor (18) is configured to calculate respective values indicative of grasp quality for the candidate grasp locations, and, based on the calculated respective values indicative of grasp quality for the candidate grasp locations, processor (18) is configured to select a grasp location likely to result in a successful grasp of the respective object.
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_EP3695941B1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>EP3695941B1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_EP3695941B13</originalsourceid><addsrcrecordid>eNqNy7EKwjAQgOEuDqK-w72AQ6kKHdPk0p7aXkiuii6lSJxECxV8fUV8AKd_-f5p8tJcu1bQ0xkNhFMQrEE1BmqUig20gZoSDFmLHhsBqlWJcCA8BhAGSx9aehUc7FkrIW7CdxevtqiFPWEAyx48FyykwZHeQevmyeTa38a4-HWWgEXR1TIOjy6OQ3-J9_js0GWbfJ2v0iLN_iBvAPM5aA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>COMPUTERIZED SYSTEM AND METHOD USING DIFFERENT IMAGE VIEWS TO FIND GRASP LOCATIONS AND TRAJECTORIES FOR ROBOTIC PICK UP</title><source>esp@cenet</source><creator>SOLOWJOW, Eugen ; WEN, Changtao ; SEHR, Martin ; APARICIO OJEA, Juan L ; CLAUSSEN, Heiko</creator><creatorcontrib>SOLOWJOW, Eugen ; WEN, Changtao ; SEHR, Martin ; APARICIO OJEA, Juan L ; CLAUSSEN, Heiko</creatorcontrib><description>Computerized system and method are provided. A robotic manipulator (12) is arranged to grasp objects (20). A gripper (16) is attached to robotic manipulator (12), which includes an imaging sensor (14). During motion of robotic manipulator (12), imaging sensor (14) is arranged to capture images providing different views of objects in the environment of the robotic manipulator. A processor (18) is configured to find, based on the different views, candidate grasp locations and trajectories to perform a grasp of a respective object in the environment of the robotic manipulator. Processor (18) is configured to calculate respective values indicative of grasp quality for the candidate grasp locations, and, based on the calculated respective values indicative of grasp quality for the candidate grasp locations, processor (18) is configured to select a grasp location likely to result in a successful grasp of the respective object.</description><language>eng ; fre ; ger</language><subject>CHAMBERS PROVIDED WITH MANIPULATION DEVICES ; HAND TOOLS ; MANIPULATORS ; PERFORMING OPERATIONS ; PORTABLE POWER-DRIVEN TOOLS ; TRANSPORTING</subject><creationdate>2022</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20220413&amp;DB=EPODOC&amp;CC=EP&amp;NR=3695941B1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76289</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20220413&amp;DB=EPODOC&amp;CC=EP&amp;NR=3695941B1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>SOLOWJOW, Eugen</creatorcontrib><creatorcontrib>WEN, Changtao</creatorcontrib><creatorcontrib>SEHR, Martin</creatorcontrib><creatorcontrib>APARICIO OJEA, Juan L</creatorcontrib><creatorcontrib>CLAUSSEN, Heiko</creatorcontrib><title>COMPUTERIZED SYSTEM AND METHOD USING DIFFERENT IMAGE VIEWS TO FIND GRASP LOCATIONS AND TRAJECTORIES FOR ROBOTIC PICK UP</title><description>Computerized system and method are provided. A robotic manipulator (12) is arranged to grasp objects (20). A gripper (16) is attached to robotic manipulator (12), which includes an imaging sensor (14). During motion of robotic manipulator (12), imaging sensor (14) is arranged to capture images providing different views of objects in the environment of the robotic manipulator. A processor (18) is configured to find, based on the different views, candidate grasp locations and trajectories to perform a grasp of a respective object in the environment of the robotic manipulator. Processor (18) is configured to calculate respective values indicative of grasp quality for the candidate grasp locations, and, based on the calculated respective values indicative of grasp quality for the candidate grasp locations, processor (18) is configured to select a grasp location likely to result in a successful grasp of the respective object.</description><subject>CHAMBERS PROVIDED WITH MANIPULATION DEVICES</subject><subject>HAND TOOLS</subject><subject>MANIPULATORS</subject><subject>PERFORMING OPERATIONS</subject><subject>PORTABLE POWER-DRIVEN TOOLS</subject><subject>TRANSPORTING</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2022</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNy7EKwjAQgOEuDqK-w72AQ6kKHdPk0p7aXkiuii6lSJxECxV8fUV8AKd_-f5p8tJcu1bQ0xkNhFMQrEE1BmqUig20gZoSDFmLHhsBqlWJcCA8BhAGSx9aehUc7FkrIW7CdxevtqiFPWEAyx48FyykwZHeQevmyeTa38a4-HWWgEXR1TIOjy6OQ3-J9_js0GWbfJ2v0iLN_iBvAPM5aA</recordid><startdate>20220413</startdate><enddate>20220413</enddate><creator>SOLOWJOW, Eugen</creator><creator>WEN, Changtao</creator><creator>SEHR, Martin</creator><creator>APARICIO OJEA, Juan L</creator><creator>CLAUSSEN, Heiko</creator><scope>EVB</scope></search><sort><creationdate>20220413</creationdate><title>COMPUTERIZED SYSTEM AND METHOD USING DIFFERENT IMAGE VIEWS TO FIND GRASP LOCATIONS AND TRAJECTORIES FOR ROBOTIC PICK UP</title><author>SOLOWJOW, Eugen ; WEN, Changtao ; SEHR, Martin ; APARICIO OJEA, Juan L ; CLAUSSEN, Heiko</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_EP3695941B13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng ; fre ; ger</language><creationdate>2022</creationdate><topic>CHAMBERS PROVIDED WITH MANIPULATION DEVICES</topic><topic>HAND TOOLS</topic><topic>MANIPULATORS</topic><topic>PERFORMING OPERATIONS</topic><topic>PORTABLE POWER-DRIVEN TOOLS</topic><topic>TRANSPORTING</topic><toplevel>online_resources</toplevel><creatorcontrib>SOLOWJOW, Eugen</creatorcontrib><creatorcontrib>WEN, Changtao</creatorcontrib><creatorcontrib>SEHR, Martin</creatorcontrib><creatorcontrib>APARICIO OJEA, Juan L</creatorcontrib><creatorcontrib>CLAUSSEN, Heiko</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>SOLOWJOW, Eugen</au><au>WEN, Changtao</au><au>SEHR, Martin</au><au>APARICIO OJEA, Juan L</au><au>CLAUSSEN, Heiko</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>COMPUTERIZED SYSTEM AND METHOD USING DIFFERENT IMAGE VIEWS TO FIND GRASP LOCATIONS AND TRAJECTORIES FOR ROBOTIC PICK UP</title><date>2022-04-13</date><risdate>2022</risdate><abstract>Computerized system and method are provided. A robotic manipulator (12) is arranged to grasp objects (20). A gripper (16) is attached to robotic manipulator (12), which includes an imaging sensor (14). During motion of robotic manipulator (12), imaging sensor (14) is arranged to capture images providing different views of objects in the environment of the robotic manipulator. A processor (18) is configured to find, based on the different views, candidate grasp locations and trajectories to perform a grasp of a respective object in the environment of the robotic manipulator. Processor (18) is configured to calculate respective values indicative of grasp quality for the candidate grasp locations, and, based on the calculated respective values indicative of grasp quality for the candidate grasp locations, processor (18) is configured to select a grasp location likely to result in a successful grasp of the respective object.</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng ; fre ; ger
recordid cdi_epo_espacenet_EP3695941B1
source esp@cenet
subjects CHAMBERS PROVIDED WITH MANIPULATION DEVICES
HAND TOOLS
MANIPULATORS
PERFORMING OPERATIONS
PORTABLE POWER-DRIVEN TOOLS
TRANSPORTING
title COMPUTERIZED SYSTEM AND METHOD USING DIFFERENT IMAGE VIEWS TO FIND GRASP LOCATIONS AND TRAJECTORIES FOR ROBOTIC PICK UP
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T10%3A18%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=SOLOWJOW,%20Eugen&rft.date=2022-04-13&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EEP3695941B1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true