Front2Back: Single View 3D Shape Reconstruction via Front to Back Prediction

Reconstruction of a 3D shape from a single 2D image is a classical computer vision problem, whose difficulty stems from the inherent ambiguity of recovering occluded or only partially observed surfaces. Recent methods address this challenge through the use of largely unstructured neural networks tha...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Yao, Yuan, Schertler, Nico, Rosales, Enrique, Rhodin, Helge, Sigal, Leonid, Sheffer, Alla
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Yao, Yuan
Schertler, Nico
Rosales, Enrique
Rhodin, Helge
Sigal, Leonid
Sheffer, Alla
description Reconstruction of a 3D shape from a single 2D image is a classical computer vision problem, whose difficulty stems from the inherent ambiguity of recovering occluded or only partially observed surfaces. Recent methods address this challenge through the use of largely unstructured neural networks that effectively distill conditional mapping and priors over 3D shape. In this work, we induce structure and geometric constraints by leveraging three core observations: (1) the surface of most everyday objects is often almost entirely exposed from pairs of typical opposite views; (2) everyday objects often exhibit global reflective symmetries which can be accurately predicted from single views; (3) opposite orthographic views of a 3D shape share consistent silhouettes. Following these observations, we first predict orthographic 2.5D visible surface maps (depth, normal and silhouette) from perspective 2D images, and detect global reflective symmetries in this data; second, we predict the back facing depth and normal maps using as input the front maps and, when available, the symmetric reflections of these maps; and finally, we reconstruct a 3D mesh from the union of these maps using a surface reconstruction method best suited for this data. Our experiments demonstrate that our framework outperforms state-of-the art approaches for 3D shape reconstructions from 2D and 2.5D data in terms of input fidelity and details preservation. Specifically, we achieve 12% better performance on average in ShapeNet benchmark dataset, and up to 19% for certain classes of objects (e.g., chairs and vessels).
doi_str_mv 10.48550/arxiv.1912.10589
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1912_10589</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1912_10589</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-4224961017da7b21b8988a87aa8e46f70a291694c80ea6e28784dd176823221f3</originalsourceid><addsrcrecordid>eNotj0FOwzAQRb1hgQoHYMVcIMEzcewxOygUkCKBaMU2miYOWJSkckOB26MGVn_x9Z70lDpDnRsuS30h6Tvuc_RIOeqS_bGqFmnoR7qW5v0SlrF_3QR4ieELihtYvsk2wHNohn43ps9mjEMP-ygwMTAOcMDgKYU2TueJOupkswun_ztTq8Xtan6fVY93D_OrKhPrfGaIjLeo0bXi1oRr9szCToSDsZ3TQh6tNw3rIDYQOzZti84yFUTYFTN1_qedcuptih-SfupDVj1lFb--cUW-</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Front2Back: Single View 3D Shape Reconstruction via Front to Back Prediction</title><source>arXiv.org</source><creator>Yao, Yuan ; Schertler, Nico ; Rosales, Enrique ; Rhodin, Helge ; Sigal, Leonid ; Sheffer, Alla</creator><creatorcontrib>Yao, Yuan ; Schertler, Nico ; Rosales, Enrique ; Rhodin, Helge ; Sigal, Leonid ; Sheffer, Alla</creatorcontrib><description>Reconstruction of a 3D shape from a single 2D image is a classical computer vision problem, whose difficulty stems from the inherent ambiguity of recovering occluded or only partially observed surfaces. Recent methods address this challenge through the use of largely unstructured neural networks that effectively distill conditional mapping and priors over 3D shape. In this work, we induce structure and geometric constraints by leveraging three core observations: (1) the surface of most everyday objects is often almost entirely exposed from pairs of typical opposite views; (2) everyday objects often exhibit global reflective symmetries which can be accurately predicted from single views; (3) opposite orthographic views of a 3D shape share consistent silhouettes. Following these observations, we first predict orthographic 2.5D visible surface maps (depth, normal and silhouette) from perspective 2D images, and detect global reflective symmetries in this data; second, we predict the back facing depth and normal maps using as input the front maps and, when available, the symmetric reflections of these maps; and finally, we reconstruct a 3D mesh from the union of these maps using a surface reconstruction method best suited for this data. Our experiments demonstrate that our framework outperforms state-of-the art approaches for 3D shape reconstructions from 2D and 2.5D data in terms of input fidelity and details preservation. Specifically, we achieve 12% better performance on average in ShapeNet benchmark dataset, and up to 19% for certain classes of objects (e.g., chairs and vessels).</description><identifier>DOI: 10.48550/arxiv.1912.10589</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Graphics</subject><creationdate>2019-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1912.10589$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1912.10589$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Yao, Yuan</creatorcontrib><creatorcontrib>Schertler, Nico</creatorcontrib><creatorcontrib>Rosales, Enrique</creatorcontrib><creatorcontrib>Rhodin, Helge</creatorcontrib><creatorcontrib>Sigal, Leonid</creatorcontrib><creatorcontrib>Sheffer, Alla</creatorcontrib><title>Front2Back: Single View 3D Shape Reconstruction via Front to Back Prediction</title><description>Reconstruction of a 3D shape from a single 2D image is a classical computer vision problem, whose difficulty stems from the inherent ambiguity of recovering occluded or only partially observed surfaces. Recent methods address this challenge through the use of largely unstructured neural networks that effectively distill conditional mapping and priors over 3D shape. In this work, we induce structure and geometric constraints by leveraging three core observations: (1) the surface of most everyday objects is often almost entirely exposed from pairs of typical opposite views; (2) everyday objects often exhibit global reflective symmetries which can be accurately predicted from single views; (3) opposite orthographic views of a 3D shape share consistent silhouettes. Following these observations, we first predict orthographic 2.5D visible surface maps (depth, normal and silhouette) from perspective 2D images, and detect global reflective symmetries in this data; second, we predict the back facing depth and normal maps using as input the front maps and, when available, the symmetric reflections of these maps; and finally, we reconstruct a 3D mesh from the union of these maps using a surface reconstruction method best suited for this data. Our experiments demonstrate that our framework outperforms state-of-the art approaches for 3D shape reconstructions from 2D and 2.5D data in terms of input fidelity and details preservation. Specifically, we achieve 12% better performance on average in ShapeNet benchmark dataset, and up to 19% for certain classes of objects (e.g., chairs and vessels).</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Graphics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj0FOwzAQRb1hgQoHYMVcIMEzcewxOygUkCKBaMU2miYOWJSkckOB26MGVn_x9Z70lDpDnRsuS30h6Tvuc_RIOeqS_bGqFmnoR7qW5v0SlrF_3QR4ieELihtYvsk2wHNohn43ps9mjEMP-ygwMTAOcMDgKYU2TueJOupkswun_ztTq8Xtan6fVY93D_OrKhPrfGaIjLeo0bXi1oRr9szCToSDsZ3TQh6tNw3rIDYQOzZti84yFUTYFTN1_qedcuptih-SfupDVj1lFb--cUW-</recordid><startdate>20191222</startdate><enddate>20191222</enddate><creator>Yao, Yuan</creator><creator>Schertler, Nico</creator><creator>Rosales, Enrique</creator><creator>Rhodin, Helge</creator><creator>Sigal, Leonid</creator><creator>Sheffer, Alla</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20191222</creationdate><title>Front2Back: Single View 3D Shape Reconstruction via Front to Back Prediction</title><author>Yao, Yuan ; Schertler, Nico ; Rosales, Enrique ; Rhodin, Helge ; Sigal, Leonid ; Sheffer, Alla</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-4224961017da7b21b8988a87aa8e46f70a291694c80ea6e28784dd176823221f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Graphics</topic><toplevel>online_resources</toplevel><creatorcontrib>Yao, Yuan</creatorcontrib><creatorcontrib>Schertler, Nico</creatorcontrib><creatorcontrib>Rosales, Enrique</creatorcontrib><creatorcontrib>Rhodin, Helge</creatorcontrib><creatorcontrib>Sigal, Leonid</creatorcontrib><creatorcontrib>Sheffer, Alla</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Yao, Yuan</au><au>Schertler, Nico</au><au>Rosales, Enrique</au><au>Rhodin, Helge</au><au>Sigal, Leonid</au><au>Sheffer, Alla</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Front2Back: Single View 3D Shape Reconstruction via Front to Back Prediction</atitle><date>2019-12-22</date><risdate>2019</risdate><abstract>Reconstruction of a 3D shape from a single 2D image is a classical computer vision problem, whose difficulty stems from the inherent ambiguity of recovering occluded or only partially observed surfaces. Recent methods address this challenge through the use of largely unstructured neural networks that effectively distill conditional mapping and priors over 3D shape. In this work, we induce structure and geometric constraints by leveraging three core observations: (1) the surface of most everyday objects is often almost entirely exposed from pairs of typical opposite views; (2) everyday objects often exhibit global reflective symmetries which can be accurately predicted from single views; (3) opposite orthographic views of a 3D shape share consistent silhouettes. Following these observations, we first predict orthographic 2.5D visible surface maps (depth, normal and silhouette) from perspective 2D images, and detect global reflective symmetries in this data; second, we predict the back facing depth and normal maps using as input the front maps and, when available, the symmetric reflections of these maps; and finally, we reconstruct a 3D mesh from the union of these maps using a surface reconstruction method best suited for this data. Our experiments demonstrate that our framework outperforms state-of-the art approaches for 3D shape reconstructions from 2D and 2.5D data in terms of input fidelity and details preservation. Specifically, we achieve 12% better performance on average in ShapeNet benchmark dataset, and up to 19% for certain classes of objects (e.g., chairs and vessels).</abstract><doi>10.48550/arxiv.1912.10589</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1912.10589
ispartof
issn
language eng
recordid cdi_arxiv_primary_1912_10589
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
Computer Science - Graphics
title Front2Back: Single View 3D Shape Reconstruction via Front to Back Prediction
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T04%3A40%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Front2Back:%20Single%20View%203D%20Shape%20Reconstruction%20via%20Front%20to%20Back%20Prediction&rft.au=Yao,%20Yuan&rft.date=2019-12-22&rft_id=info:doi/10.48550/arxiv.1912.10589&rft_dat=%3Carxiv_GOX%3E1912_10589%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true