Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects

Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks. To mitigate this problem, we propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We fu...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Dai, Qiyu, Zhang, Jiyao, Li, Qiwei, Wu, Tianhao, Dong, Hao, Liu, Ziyuan, Tan, Ping, Wang, He
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Dai, Qiyu
Zhang, Jiyao
Li, Qiwei
Wu, Tianhao
Dong, Hao
Liu, Ziyuan
Tan, Ping
Wang, He
description Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks. To mitigate this problem, we propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach to simulate an active stereo depth system using physically based rendering and generate a large-scale synthetic dataset that contains 130K photorealistic RGB images along with their simulated depths carrying realistic sensor noises. To evaluate depth restoration methods, we also curate a real-world dataset, namely STD, that captures 30 cluttered scenes composed of 50 objects with different materials from specular, transparent, to diffuse. Experiments demonstrate that the proposed DREDS dataset bridges the sim-to-real domain gap such that, trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth datasets, e.g. ClearGrasp, and outperform the competing methods on depth restoration with a real-time speed. We further show that our depth restoration effectively boosts the performance of downstream tasks, including category-level pose estimation and grasping tasks. Our data and code are available at https://github.com/PKU-EPIC/DREDS
doi_str_mv 10.48550/arxiv.2208.03792
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2208_03792</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2208_03792</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-ff96cdfe7a2e9ff4b51dd09933053442f10009d45fc2027a5021bc666d6bbb913</originalsourceid><addsrcrecordid>eNotkM1uwjAQhH3poaJ9gJ7qFwh17NjBxwoorYREBblHa3tdXBEnclLUn5cvBE6rb2d2pRlCHnI2LWZSsidI3-E45ZzNpkyUmt-Sv0XbQIh0C9G1TfiFIbQxW8Y9RIuOLrAb9nQXmq_DqNCTjW6xH9p0Yd8m-o7JYjiG-DHKqwR9d4Zdh_Z0l8ZtlSD2HSSMA92YT7RDf0duPBx6vL_OCaleltX8NVtvVm_z53UGquSZ91pZ57EEjtr7wsjcOaa1EEyKouA-Z4xpV0hvOeMlSMZzY5VSThljdC4m5PHydkxfdyk0kH7qcwv12IL4B9CgWgo</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</title><source>arXiv.org</source><creator>Dai, Qiyu ; Zhang, Jiyao ; Li, Qiwei ; Wu, Tianhao ; Dong, Hao ; Liu, Ziyuan ; Tan, Ping ; Wang, He</creator><creatorcontrib>Dai, Qiyu ; Zhang, Jiyao ; Li, Qiwei ; Wu, Tianhao ; Dong, Hao ; Liu, Ziyuan ; Tan, Ping ; Wang, He</creatorcontrib><description>Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks. To mitigate this problem, we propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach to simulate an active stereo depth system using physically based rendering and generate a large-scale synthetic dataset that contains 130K photorealistic RGB images along with their simulated depths carrying realistic sensor noises. To evaluate depth restoration methods, we also curate a real-world dataset, namely STD, that captures 30 cluttered scenes composed of 50 objects with different materials from specular, transparent, to diffuse. Experiments demonstrate that the proposed DREDS dataset bridges the sim-to-real domain gap such that, trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth datasets, e.g. ClearGrasp, and outperform the competing methods on depth restoration with a real-time speed. We further show that our depth restoration effectively boosts the performance of downstream tasks, including category-level pose estimation and grasping tasks. Our data and code are available at https://github.com/PKU-EPIC/DREDS</description><identifier>DOI: 10.48550/arxiv.2208.03792</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2022-08</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2208.03792$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2208.03792$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dai, Qiyu</creatorcontrib><creatorcontrib>Zhang, Jiyao</creatorcontrib><creatorcontrib>Li, Qiwei</creatorcontrib><creatorcontrib>Wu, Tianhao</creatorcontrib><creatorcontrib>Dong, Hao</creatorcontrib><creatorcontrib>Liu, Ziyuan</creatorcontrib><creatorcontrib>Tan, Ping</creatorcontrib><creatorcontrib>Wang, He</creatorcontrib><title>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</title><description>Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks. To mitigate this problem, we propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach to simulate an active stereo depth system using physically based rendering and generate a large-scale synthetic dataset that contains 130K photorealistic RGB images along with their simulated depths carrying realistic sensor noises. To evaluate depth restoration methods, we also curate a real-world dataset, namely STD, that captures 30 cluttered scenes composed of 50 objects with different materials from specular, transparent, to diffuse. Experiments demonstrate that the proposed DREDS dataset bridges the sim-to-real domain gap such that, trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth datasets, e.g. ClearGrasp, and outperform the competing methods on depth restoration with a real-time speed. We further show that our depth restoration effectively boosts the performance of downstream tasks, including category-level pose estimation and grasping tasks. Our data and code are available at https://github.com/PKU-EPIC/DREDS</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotkM1uwjAQhH3poaJ9gJ7qFwh17NjBxwoorYREBblHa3tdXBEnclLUn5cvBE6rb2d2pRlCHnI2LWZSsidI3-E45ZzNpkyUmt-Sv0XbQIh0C9G1TfiFIbQxW8Y9RIuOLrAb9nQXmq_DqNCTjW6xH9p0Yd8m-o7JYjiG-DHKqwR9d4Zdh_Z0l8ZtlSD2HSSMA92YT7RDf0duPBx6vL_OCaleltX8NVtvVm_z53UGquSZ91pZ57EEjtr7wsjcOaa1EEyKouA-Z4xpV0hvOeMlSMZzY5VSThljdC4m5PHydkxfdyk0kH7qcwv12IL4B9CgWgo</recordid><startdate>20220807</startdate><enddate>20220807</enddate><creator>Dai, Qiyu</creator><creator>Zhang, Jiyao</creator><creator>Li, Qiwei</creator><creator>Wu, Tianhao</creator><creator>Dong, Hao</creator><creator>Liu, Ziyuan</creator><creator>Tan, Ping</creator><creator>Wang, He</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220807</creationdate><title>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</title><author>Dai, Qiyu ; Zhang, Jiyao ; Li, Qiwei ; Wu, Tianhao ; Dong, Hao ; Liu, Ziyuan ; Tan, Ping ; Wang, He</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-ff96cdfe7a2e9ff4b51dd09933053442f10009d45fc2027a5021bc666d6bbb913</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Dai, Qiyu</creatorcontrib><creatorcontrib>Zhang, Jiyao</creatorcontrib><creatorcontrib>Li, Qiwei</creatorcontrib><creatorcontrib>Wu, Tianhao</creatorcontrib><creatorcontrib>Dong, Hao</creatorcontrib><creatorcontrib>Liu, Ziyuan</creatorcontrib><creatorcontrib>Tan, Ping</creatorcontrib><creatorcontrib>Wang, He</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dai, Qiyu</au><au>Zhang, Jiyao</au><au>Li, Qiwei</au><au>Wu, Tianhao</au><au>Dong, Hao</au><au>Liu, Ziyuan</au><au>Tan, Ping</au><au>Wang, He</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects</atitle><date>2022-08-07</date><risdate>2022</risdate><abstract>Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks. To mitigate this problem, we propose a powerful RGBD fusion network, SwinDRNet, for depth restoration. We further propose Domain Randomization-Enhanced Depth Simulation (DREDS) approach to simulate an active stereo depth system using physically based rendering and generate a large-scale synthetic dataset that contains 130K photorealistic RGB images along with their simulated depths carrying realistic sensor noises. To evaluate depth restoration methods, we also curate a real-world dataset, namely STD, that captures 30 cluttered scenes composed of 50 objects with different materials from specular, transparent, to diffuse. Experiments demonstrate that the proposed DREDS dataset bridges the sim-to-real domain gap such that, trained on DREDS, our SwinDRNet can seamlessly generalize to other real depth datasets, e.g. ClearGrasp, and outperform the competing methods on depth restoration with a real-time speed. We further show that our depth restoration effectively boosts the performance of downstream tasks, including category-level pose estimation and grasping tasks. Our data and code are available at https://github.com/PKU-EPIC/DREDS</abstract><doi>10.48550/arxiv.2208.03792</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2208.03792
ispartof
issn
language eng
recordid cdi_arxiv_primary_2208_03792
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-30T18%3A51%3A34IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Domain%20Randomization-Enhanced%20Depth%20Simulation%20and%20Restoration%20for%20Perceiving%20and%20Grasping%20Specular%20and%20Transparent%20Objects&rft.au=Dai,%20Qiyu&rft.date=2022-08-07&rft_id=info:doi/10.48550/arxiv.2208.03792&rft_dat=%3Carxiv_GOX%3E2208_03792%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true