Robust Depth Enhancement via Polarization Prompt Fusion Tuning

Existing depth sensors are imperfect and may provide inaccurate depth values in challenging scenarios, such as in the presence of transparent or reflective objects. In this work, we present a general framework that leverages polarization imaging to improve inaccurate depth measurements from various...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Ikemura, Kei, Huang, Yiming, Heide, Felix, Zhang, Zhaoxiang, Chen, Qifeng, Lei, Chenyang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Ikemura, Kei
Huang, Yiming
Heide, Felix
Zhang, Zhaoxiang
Chen, Qifeng
Lei, Chenyang
description Existing depth sensors are imperfect and may provide inaccurate depth values in challenging scenarios, such as in the presence of transparent or reflective objects. In this work, we present a general framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors. Previous polarization-based depth enhancement methods focus on utilizing pure physics-based formulas for a single sensor. In contrast, our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors. To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets, as the size of the polarization dataset is limited to train a strong model from scratch. We conducted extensive experiments on a public dataset, and the results demonstrate that the proposed method performs favorably compared to existing depth enhancement baselines. Code and demos are available at https://lastbasket.github.io/PPFT/.
doi_str_mv 10.48550/arxiv.2404.04318
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2404_04318</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2404_04318</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-f946a6d2db1497358956c1321ab4068048c5a414460c203e2a17b282616df7be3</originalsourceid><addsrcrecordid>eNotz09LwzAcxvFcdpDNF-DJvIHW_PklTS-CzM0JA4f0Xn5pUxdY05KmQ331Y9PTw_fywIeQB85yMEqxJ4zf_pwLYJAzkNzckefPwc5Toq9uTEe6CUcMjetdSPTskR6GE0b_i8kPgR7i0I-JbufpWtUcfPhakUWHp8nd_--SVNtNtd5l-4-39_XLPkNdmKwrQaNuRWs5lIVUplS64VJwtMC0YWAahcABNGsEk04gL6wwQnPddoV1ckke_25vgHqMvsf4U18h9Q0iLy9jQfM</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Robust Depth Enhancement via Polarization Prompt Fusion Tuning</title><source>arXiv.org</source><creator>Ikemura, Kei ; Huang, Yiming ; Heide, Felix ; Zhang, Zhaoxiang ; Chen, Qifeng ; Lei, Chenyang</creator><creatorcontrib>Ikemura, Kei ; Huang, Yiming ; Heide, Felix ; Zhang, Zhaoxiang ; Chen, Qifeng ; Lei, Chenyang</creatorcontrib><description>Existing depth sensors are imperfect and may provide inaccurate depth values in challenging scenarios, such as in the presence of transparent or reflective objects. In this work, we present a general framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors. Previous polarization-based depth enhancement methods focus on utilizing pure physics-based formulas for a single sensor. In contrast, our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors. To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets, as the size of the polarization dataset is limited to train a strong model from scratch. We conducted extensive experiments on a public dataset, and the results demonstrate that the proposed method performs favorably compared to existing depth enhancement baselines. Code and demos are available at https://lastbasket.github.io/PPFT/.</description><identifier>DOI: 10.48550/arxiv.2404.04318</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2024-04</creationdate><rights>http://creativecommons.org/licenses/by-nc-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2404.04318$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2404.04318$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Ikemura, Kei</creatorcontrib><creatorcontrib>Huang, Yiming</creatorcontrib><creatorcontrib>Heide, Felix</creatorcontrib><creatorcontrib>Zhang, Zhaoxiang</creatorcontrib><creatorcontrib>Chen, Qifeng</creatorcontrib><creatorcontrib>Lei, Chenyang</creatorcontrib><title>Robust Depth Enhancement via Polarization Prompt Fusion Tuning</title><description>Existing depth sensors are imperfect and may provide inaccurate depth values in challenging scenarios, such as in the presence of transparent or reflective objects. In this work, we present a general framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors. Previous polarization-based depth enhancement methods focus on utilizing pure physics-based formulas for a single sensor. In contrast, our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors. To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets, as the size of the polarization dataset is limited to train a strong model from scratch. We conducted extensive experiments on a public dataset, and the results demonstrate that the proposed method performs favorably compared to existing depth enhancement baselines. Code and demos are available at https://lastbasket.github.io/PPFT/.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz09LwzAcxvFcdpDNF-DJvIHW_PklTS-CzM0JA4f0Xn5pUxdY05KmQ331Y9PTw_fywIeQB85yMEqxJ4zf_pwLYJAzkNzckefPwc5Toq9uTEe6CUcMjetdSPTskR6GE0b_i8kPgR7i0I-JbufpWtUcfPhakUWHp8nd_--SVNtNtd5l-4-39_XLPkNdmKwrQaNuRWs5lIVUplS64VJwtMC0YWAahcABNGsEk04gL6wwQnPddoV1ckke_25vgHqMvsf4U18h9Q0iLy9jQfM</recordid><startdate>20240405</startdate><enddate>20240405</enddate><creator>Ikemura, Kei</creator><creator>Huang, Yiming</creator><creator>Heide, Felix</creator><creator>Zhang, Zhaoxiang</creator><creator>Chen, Qifeng</creator><creator>Lei, Chenyang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240405</creationdate><title>Robust Depth Enhancement via Polarization Prompt Fusion Tuning</title><author>Ikemura, Kei ; Huang, Yiming ; Heide, Felix ; Zhang, Zhaoxiang ; Chen, Qifeng ; Lei, Chenyang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-f946a6d2db1497358956c1321ab4068048c5a414460c203e2a17b282616df7be3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Ikemura, Kei</creatorcontrib><creatorcontrib>Huang, Yiming</creatorcontrib><creatorcontrib>Heide, Felix</creatorcontrib><creatorcontrib>Zhang, Zhaoxiang</creatorcontrib><creatorcontrib>Chen, Qifeng</creatorcontrib><creatorcontrib>Lei, Chenyang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Ikemura, Kei</au><au>Huang, Yiming</au><au>Heide, Felix</au><au>Zhang, Zhaoxiang</au><au>Chen, Qifeng</au><au>Lei, Chenyang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Robust Depth Enhancement via Polarization Prompt Fusion Tuning</atitle><date>2024-04-05</date><risdate>2024</risdate><abstract>Existing depth sensors are imperfect and may provide inaccurate depth values in challenging scenarios, such as in the presence of transparent or reflective objects. In this work, we present a general framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors. Previous polarization-based depth enhancement methods focus on utilizing pure physics-based formulas for a single sensor. In contrast, our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors. To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets, as the size of the polarization dataset is limited to train a strong model from scratch. We conducted extensive experiments on a public dataset, and the results demonstrate that the proposed method performs favorably compared to existing depth enhancement baselines. Code and demos are available at https://lastbasket.github.io/PPFT/.</abstract><doi>10.48550/arxiv.2404.04318</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2404.04318
ispartof
issn
language eng
recordid cdi_arxiv_primary_2404_04318
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
title Robust Depth Enhancement via Polarization Prompt Fusion Tuning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-29T13%3A00%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Robust%20Depth%20Enhancement%20via%20Polarization%20Prompt%20Fusion%20Tuning&rft.au=Ikemura,%20Kei&rft.date=2024-04-05&rft_id=info:doi/10.48550/arxiv.2404.04318&rft_dat=%3Carxiv_GOX%3E2404_04318%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true