Dynamic Reconstruction of Hand-Object Interaction with Distributed Force-aware Contact Representation

We present ViTaM-D, a novel visual-tactile framework for dynamic hand-object interaction reconstruction, integrating distributed tactile sensing for more accurate contact modeling. While existing methods focus primarily on visual inputs, they struggle with capturing detailed contact interactions suc...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-11
Hauptverfasser: Yu, Zhenjun, Xu, Wenqiang, Xie, Pengfei, Li, Yutong, Lu, Cewu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Yu, Zhenjun
Xu, Wenqiang
Xie, Pengfei
Li, Yutong
Lu, Cewu
description We present ViTaM-D, a novel visual-tactile framework for dynamic hand-object interaction reconstruction, integrating distributed tactile sensing for more accurate contact modeling. While existing methods focus primarily on visual inputs, they struggle with capturing detailed contact interactions such as object deformation. Our approach leverages distributed tactile sensors to address this limitation by introducing DF-Field. This distributed force-aware contact representation models both kinetic and potential energy in hand-object interaction. ViTaM-D first reconstructs hand-object interactions using a visual-only network, VDT-Net, and then refines contact details through a force-aware optimization (FO) process, enhancing object deformation modeling. To benchmark our approach, we introduce the HOT dataset, which features 600 sequences of hand-object interactions, including deformable objects, built in a high-precision simulation environment. Extensive experiments on both the DexYCB and HOT datasets demonstrate significant improvements in accuracy over previous state-of-the-art methods such as gSDF and HOTrack. Our results highlight the superior performance of ViTaM-D in both rigid and deformable object reconstruction, as well as the effectiveness of DF-Field in refining hand poses. This work offers a comprehensive solution to dynamic hand-object interaction reconstruction by seamlessly integrating visual and tactile data. Codes, models, and datasets will be available.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3128887285</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3128887285</sourcerecordid><originalsourceid>FETCH-proquest_journals_31288872853</originalsourceid><addsrcrecordid>eNqNjE0KwjAYRIMgWLR3CLgOtIm12beWuhLEvaTpV0zRpOaH4u2N6AFcDcN7MwuUUMZywneUrlDq3JhlGd2XtChYgqB-afFQEp9BGu28DdIro7EZcCt0T07dCNLjo_ZgxRfNyt9wraKruuChx42xEoiYhQVcGe2jF-8mCw5i-Ww2aDmIu4P0l2u0bQ6XqiWTNc8Azl9HE6yO6MpyyjkvKS_Yf9YbvHtHdw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3128887285</pqid></control><display><type>article</type><title>Dynamic Reconstruction of Hand-Object Interaction with Distributed Force-aware Contact Representation</title><source>Free E- Journals</source><creator>Yu, Zhenjun ; Xu, Wenqiang ; Xie, Pengfei ; Li, Yutong ; Lu, Cewu</creator><creatorcontrib>Yu, Zhenjun ; Xu, Wenqiang ; Xie, Pengfei ; Li, Yutong ; Lu, Cewu</creatorcontrib><description>We present ViTaM-D, a novel visual-tactile framework for dynamic hand-object interaction reconstruction, integrating distributed tactile sensing for more accurate contact modeling. While existing methods focus primarily on visual inputs, they struggle with capturing detailed contact interactions such as object deformation. Our approach leverages distributed tactile sensors to address this limitation by introducing DF-Field. This distributed force-aware contact representation models both kinetic and potential energy in hand-object interaction. ViTaM-D first reconstructs hand-object interactions using a visual-only network, VDT-Net, and then refines contact details through a force-aware optimization (FO) process, enhancing object deformation modeling. To benchmark our approach, we introduce the HOT dataset, which features 600 sequences of hand-object interactions, including deformable objects, built in a high-precision simulation environment. Extensive experiments on both the DexYCB and HOT datasets demonstrate significant improvements in accuracy over previous state-of-the-art methods such as gSDF and HOTrack. Our results highlight the superior performance of ViTaM-D in both rigid and deformable object reconstruction, as well as the effectiveness of DF-Field in refining hand poses. This work offers a comprehensive solution to dynamic hand-object interaction reconstruction by seamlessly integrating visual and tactile data. Codes, models, and datasets will be available.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Datasets ; Deformation effects ; Force distribution ; Formability ; Modelling ; Potential energy ; Reconstruction ; Representations ; Tactile sensors (robotics) ; Visual fields</subject><ispartof>arXiv.org, 2024-11</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Yu, Zhenjun</creatorcontrib><creatorcontrib>Xu, Wenqiang</creatorcontrib><creatorcontrib>Xie, Pengfei</creatorcontrib><creatorcontrib>Li, Yutong</creatorcontrib><creatorcontrib>Lu, Cewu</creatorcontrib><title>Dynamic Reconstruction of Hand-Object Interaction with Distributed Force-aware Contact Representation</title><title>arXiv.org</title><description>We present ViTaM-D, a novel visual-tactile framework for dynamic hand-object interaction reconstruction, integrating distributed tactile sensing for more accurate contact modeling. While existing methods focus primarily on visual inputs, they struggle with capturing detailed contact interactions such as object deformation. Our approach leverages distributed tactile sensors to address this limitation by introducing DF-Field. This distributed force-aware contact representation models both kinetic and potential energy in hand-object interaction. ViTaM-D first reconstructs hand-object interactions using a visual-only network, VDT-Net, and then refines contact details through a force-aware optimization (FO) process, enhancing object deformation modeling. To benchmark our approach, we introduce the HOT dataset, which features 600 sequences of hand-object interactions, including deformable objects, built in a high-precision simulation environment. Extensive experiments on both the DexYCB and HOT datasets demonstrate significant improvements in accuracy over previous state-of-the-art methods such as gSDF and HOTrack. Our results highlight the superior performance of ViTaM-D in both rigid and deformable object reconstruction, as well as the effectiveness of DF-Field in refining hand poses. This work offers a comprehensive solution to dynamic hand-object interaction reconstruction by seamlessly integrating visual and tactile data. Codes, models, and datasets will be available.</description><subject>Datasets</subject><subject>Deformation effects</subject><subject>Force distribution</subject><subject>Formability</subject><subject>Modelling</subject><subject>Potential energy</subject><subject>Reconstruction</subject><subject>Representations</subject><subject>Tactile sensors (robotics)</subject><subject>Visual fields</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNjE0KwjAYRIMgWLR3CLgOtIm12beWuhLEvaTpV0zRpOaH4u2N6AFcDcN7MwuUUMZywneUrlDq3JhlGd2XtChYgqB-afFQEp9BGu28DdIro7EZcCt0T07dCNLjo_ZgxRfNyt9wraKruuChx42xEoiYhQVcGe2jF-8mCw5i-Ww2aDmIu4P0l2u0bQ6XqiWTNc8Azl9HE6yO6MpyyjkvKS_Yf9YbvHtHdw</recordid><startdate>20241114</startdate><enddate>20241114</enddate><creator>Yu, Zhenjun</creator><creator>Xu, Wenqiang</creator><creator>Xie, Pengfei</creator><creator>Li, Yutong</creator><creator>Lu, Cewu</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241114</creationdate><title>Dynamic Reconstruction of Hand-Object Interaction with Distributed Force-aware Contact Representation</title><author>Yu, Zhenjun ; Xu, Wenqiang ; Xie, Pengfei ; Li, Yutong ; Lu, Cewu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31288872853</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Datasets</topic><topic>Deformation effects</topic><topic>Force distribution</topic><topic>Formability</topic><topic>Modelling</topic><topic>Potential energy</topic><topic>Reconstruction</topic><topic>Representations</topic><topic>Tactile sensors (robotics)</topic><topic>Visual fields</topic><toplevel>online_resources</toplevel><creatorcontrib>Yu, Zhenjun</creatorcontrib><creatorcontrib>Xu, Wenqiang</creatorcontrib><creatorcontrib>Xie, Pengfei</creatorcontrib><creatorcontrib>Li, Yutong</creatorcontrib><creatorcontrib>Lu, Cewu</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yu, Zhenjun</au><au>Xu, Wenqiang</au><au>Xie, Pengfei</au><au>Li, Yutong</au><au>Lu, Cewu</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Dynamic Reconstruction of Hand-Object Interaction with Distributed Force-aware Contact Representation</atitle><jtitle>arXiv.org</jtitle><date>2024-11-14</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>We present ViTaM-D, a novel visual-tactile framework for dynamic hand-object interaction reconstruction, integrating distributed tactile sensing for more accurate contact modeling. While existing methods focus primarily on visual inputs, they struggle with capturing detailed contact interactions such as object deformation. Our approach leverages distributed tactile sensors to address this limitation by introducing DF-Field. This distributed force-aware contact representation models both kinetic and potential energy in hand-object interaction. ViTaM-D first reconstructs hand-object interactions using a visual-only network, VDT-Net, and then refines contact details through a force-aware optimization (FO) process, enhancing object deformation modeling. To benchmark our approach, we introduce the HOT dataset, which features 600 sequences of hand-object interactions, including deformable objects, built in a high-precision simulation environment. Extensive experiments on both the DexYCB and HOT datasets demonstrate significant improvements in accuracy over previous state-of-the-art methods such as gSDF and HOTrack. Our results highlight the superior performance of ViTaM-D in both rigid and deformable object reconstruction, as well as the effectiveness of DF-Field in refining hand poses. This work offers a comprehensive solution to dynamic hand-object interaction reconstruction by seamlessly integrating visual and tactile data. Codes, models, and datasets will be available.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-11
issn 2331-8422
language eng
recordid cdi_proquest_journals_3128887285
source Free E- Journals
subjects Datasets
Deformation effects
Force distribution
Formability
Modelling
Potential energy
Reconstruction
Representations
Tactile sensors (robotics)
Visual fields
title Dynamic Reconstruction of Hand-Object Interaction with Distributed Force-aware Contact Representation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T08%3A19%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Dynamic%20Reconstruction%20of%20Hand-Object%20Interaction%20with%20Distributed%20Force-aware%20Contact%20Representation&rft.jtitle=arXiv.org&rft.au=Yu,%20Zhenjun&rft.date=2024-11-14&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3128887285%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3128887285&rft_id=info:pmid/&rfr_iscdi=true