Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects
Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks. To mitigate this problem, we propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We fu...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Dai, Qiyu Zhang, Jiyao Li, Qiwei Wu, Tianhao Dong, Hao Liu, Ziyuan Tan, Ping Wang, He |
description | Commercial depth sensors usually generate noisy and missing depths,
especially on specular and transparent objects, which poses critical issues to
downstream depth or point cloud-based tasks. To mitigate this problem, we
propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We
further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach
to simulate an active stereo depth system using physically based rendering and
generate a large-scale synthetic dataset that contains 130K photorealistic RGB
images along with their simulated depths carrying realistic sensor noises. To
evaluate depth restoration methods, we also curate a real-world dataset, namely
STD, that captures 30 cluttered scenes composed of 50 objects with different
materials from specular, transparent, to diffuse. Experiments demonstrate that
the proposed DREDS dataset bridges the sim-to-real domain gap such that,
trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth
datasets, e.g. ClearGrasp, and outperform the competing methods on depth
restoration with a real-time speed. We further show that our depth restoration
effectively boosts the performance of downstream tasks, including
category-level pose estimation and grasping tasks. Our data and code are
available at https://github.com/PKU-EPIC/DREDS |
doi_str_mv | 10.48550/arxiv.2208.03792 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2208_03792</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2208_03792</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-ff96cdfe7a2e9ff4b51dd09933053442f10009d45fc2027a5021bc666d6bbb913</originalsourceid><addsrcrecordid>eNotkM1uwjAQhH3poaJ9gJ7qFwh17NjBxwoorYREBblHa3tdXBEnclLUn5cvBE6rb2d2pRlCHnI2LWZSsidI3-E45ZzNpkyUmt-Sv0XbQIh0C9G1TfiFIbQxW8Y9RIuOLrAb9nQXmq_DqNCTjW6xH9p0Yd8m-o7JYjiG-DHKqwR9d4Zdh_Z0l8ZtlSD2HSSMA92YT7RDf0duPBx6vL_OCaleltX8NVtvVm_z53UGquSZ91pZ57EEjtr7wsjcOaa1EEyKouA-Z4xpV0hvOeMlSMZzY5VSThljdC4m5PHydkxfdyk0kH7qcwv12IL4B9CgWgo</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</title><source>arXiv.org</source><creator>Dai, Qiyu ; Zhang, Jiyao ; Li, Qiwei ; Wu, Tianhao ; Dong, Hao ; Liu, Ziyuan ; Tan, Ping ; Wang, He</creator><creatorcontrib>Dai, Qiyu ; Zhang, Jiyao ; Li, Qiwei ; Wu, Tianhao ; Dong, Hao ; Liu, Ziyuan ; Tan, Ping ; Wang, He</creatorcontrib><description>Commercial depth sensors usually generate noisy and missing depths,
especially on specular and transparent objects, which poses critical issues to
downstream depth or point cloud-based tasks. To mitigate this problem, we
propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We
further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach
to simulate an active stereo depth system using physically based rendering and
generate a large-scale synthetic dataset that contains 130K photorealistic RGB
images along with their simulated depths carrying realistic sensor noises. To
evaluate depth restoration methods, we also curate a real-world dataset, namely
STD, that captures 30 cluttered scenes composed of 50 objects with different
materials from specular, transparent, to diffuse. Experiments demonstrate that
the proposed DREDS dataset bridges the sim-to-real domain gap such that,
trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth
datasets, e.g. ClearGrasp, and outperform the competing methods on depth
restoration with a real-time speed. We further show that our depth restoration
effectively boosts the performance of downstream tasks, including
category-level pose estimation and grasping tasks. Our data and code are
available at https://github.com/PKU-EPIC/DREDS</description><identifier>DOI: 10.48550/arxiv.2208.03792</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2022-08</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2208.03792$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2208.03792$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dai, Qiyu</creatorcontrib><creatorcontrib>Zhang, Jiyao</creatorcontrib><creatorcontrib>Li, Qiwei</creatorcontrib><creatorcontrib>Wu, Tianhao</creatorcontrib><creatorcontrib>Dong, Hao</creatorcontrib><creatorcontrib>Liu, Ziyuan</creatorcontrib><creatorcontrib>Tan, Ping</creatorcontrib><creatorcontrib>Wang, He</creatorcontrib><title>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</title><description>Commercial depth sensors usually generate noisy and missing depths,
especially on specular and transparent objects, which poses critical issues to
downstream depth or point cloud-based tasks. To mitigate this problem, we
propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We
further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach
to simulate an active stereo depth system using physically based rendering and
generate a large-scale synthetic dataset that contains 130K photorealistic RGB
images along with their simulated depths carrying realistic sensor noises. To
evaluate depth restoration methods, we also curate a real-world dataset, namely
STD, that captures 30 cluttered scenes composed of 50 objects with different
materials from specular, transparent, to diffuse. Experiments demonstrate that
the proposed DREDS dataset bridges the sim-to-real domain gap such that,
trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth
datasets, e.g. ClearGrasp, and outperform the competing methods on depth
restoration with a real-time speed. We further show that our depth restoration
effectively boosts the performance of downstream tasks, including
category-level pose estimation and grasping tasks. Our data and code are
available at https://github.com/PKU-EPIC/DREDS</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotkM1uwjAQhH3poaJ9gJ7qFwh17NjBxwoorYREBblHa3tdXBEnclLUn5cvBE6rb2d2pRlCHnI2LWZSsidI3-E45ZzNpkyUmt-Sv0XbQIh0C9G1TfiFIbQxW8Y9RIuOLrAb9nQXmq_DqNCTjW6xH9p0Yd8m-o7JYjiG-DHKqwR9d4Zdh_Z0l8ZtlSD2HSSMA92YT7RDf0duPBx6vL_OCaleltX8NVtvVm_z53UGquSZ91pZ57EEjtr7wsjcOaa1EEyKouA-Z4xpV0hvOeMlSMZzY5VSThljdC4m5PHydkxfdyk0kH7qcwv12IL4B9CgWgo</recordid><startdate>20220807</startdate><enddate>20220807</enddate><creator>Dai, Qiyu</creator><creator>Zhang, Jiyao</creator><creator>Li, Qiwei</creator><creator>Wu, Tianhao</creator><creator>Dong, Hao</creator><creator>Liu, Ziyuan</creator><creator>Tan, Ping</creator><creator>Wang, He</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220807</creationdate><title>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</title><author>Dai, Qiyu ; Zhang, Jiyao ; Li, Qiwei ; Wu, Tianhao ; Dong, Hao ; Liu, Ziyuan ; Tan, Ping ; Wang, He</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-ff96cdfe7a2e9ff4b51dd09933053442f10009d45fc2027a5021bc666d6bbb913</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Dai, Qiyu</creatorcontrib><creatorcontrib>Zhang, Jiyao</creatorcontrib><creatorcontrib>Li, Qiwei</creatorcontrib><creatorcontrib>Wu, Tianhao</creatorcontrib><creatorcontrib>Dong, Hao</creatorcontrib><creatorcontrib>Liu, Ziyuan</creatorcontrib><creatorcontrib>Tan, Ping</creatorcontrib><creatorcontrib>Wang, He</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dai, Qiyu</au><au>Zhang, Jiyao</au><au>Li, Qiwei</au><au>Wu, Tianhao</au><au>Dong, Hao</au><au>Liu, Ziyuan</au><au>Tan, Ping</au><au>Wang, He</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</atitle><date>2022-08-07</date><risdate>2022</risdate><abstract>Commercial depth sensors usually generate noisy and missing depths,
especially on specular and transparent objects, which poses critical issues to
downstream depth or point cloud-based tasks. To mitigate this problem, we
propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We
further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach
to simulate an active stereo depth system using physically based rendering and
generate a large-scale synthetic dataset that contains 130K photorealistic RGB
images along with their simulated depths carrying realistic sensor noises. To
evaluate depth restoration methods, we also curate a real-world dataset, namely
STD, that captures 30 cluttered scenes composed of 50 objects with different
materials from specular, transparent, to diffuse. Experiments demonstrate that
the proposed DREDS dataset bridges the sim-to-real domain gap such that,
trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth
datasets, e.g. ClearGrasp, and outperform the competing methods on depth
restoration with a real-time speed. We further show that our depth restoration
effectively boosts the performance of downstream tasks, including
category-level pose estimation and grasping tasks. Our data and code are
available at https://github.com/PKU-EPIC/DREDS</abstract><doi>10.48550/arxiv.2208.03792</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2208.03792 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2208_03792 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-30T18%3A51%3A34IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Domain%20Randomization-Enhanced%20Depth%20Simulation%20and%20Restoration%20for%20Perceiving%20and%20Grasping%20Specular%20and%20Transparent%20Objects&rft.au=Dai,%20Qiyu&rft.date=2022-08-07&rft_id=info:doi/10.48550/arxiv.2208.03792&rft_dat=%3Carxiv_GOX%3E2208_03792%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |