High-fidelity facial reflectance and geometry inference from an unconstrained image

We present a deep learning-based technique to infer high-quality facial reflectance and geometry given a single unconstrained image of the subject, which may contain partial occlusions and arbitrary illumination conditions. The reconstructed high-resolution textures, which are generated in only a fe...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:ACM transactions on graphics 2018-08, Vol.37 (4), p.1-14
Hauptverfasser: Yamaguchi, Shugo, Saito, Shunsuke, Nagano, Koki, Zhao, Yajie, Chen, Weikai, Olszewski, Kyle, Morishima, Shigeo, Li, Hao
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 14
container_issue 4
container_start_page 1
container_title ACM transactions on graphics
container_volume 37
creator Yamaguchi, Shugo
Saito, Shunsuke
Nagano, Koki
Zhao, Yajie
Chen, Weikai
Olszewski, Kyle
Morishima, Shigeo
Li, Hao
description We present a deep learning-based technique to infer high-quality facial reflectance and geometry given a single unconstrained image of the subject, which may contain partial occlusions and arbitrary illumination conditions. The reconstructed high-resolution textures, which are generated in only a few seconds, include high-resolution skin surface reflectance maps, representing both the diffuse and specular albedo, and medium- and high-frequency displacement maps, thereby allowing us to render compelling digital avatars under novel lighting conditions. To extract this data, we train our deep neural networks with a high-quality skin reflectance and geometry database created with a state-of-the-art multi-view photometric stereo system using polarized gradient illumination. Given the raw facial texture map extracted from the input image, our neural networks synthesize complete reflectance and displacement maps, as well as complete missing regions caused by occlusions. The completed textures exhibit consistent quality throughout the face due to our network architecture, which propagates texture features from the visible region, resulting in high-fidelity details that are consistent with those seen in visible regions. We describe how this highly underconstrained problem is made tractable by dividing the full inference into smaller tasks, which are addressed by dedicated neural networks. We demonstrate the effectiveness of our network design with robust texture completion from images of faces that are largely occluded. With the inferred reflectance and geometry data, we demonstrate the rendering of high-fidelity 3D avatars from a variety of subjects captured under different lighting conditions. In addition, we perform evaluations demonstrating that our method can infer plausible facial reflectance and geometric details comparable to those obtained from high-end capture devices, and outperform alternative approaches that require only a single unconstrained input image.
doi_str_mv 10.1145/3197517.3201364
format Article
fullrecord <record><control><sourceid>crossref</sourceid><recordid>TN_cdi_crossref_primary_10_1145_3197517_3201364</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_1145_3197517_3201364</sourcerecordid><originalsourceid>FETCH-LOGICAL-c307t-abc5d71d5de4dac380fcac8f1dc904ff78e04248d44322e66f9d5bfe80ace5703</originalsourceid><addsrcrecordid>eNotkMFKxDAURYMoWEfXbvMDmXlpkiZdyqDOwIALdV0yyUuNtKmkddG_t4NdXbgXLodDyCOHLedS7QSvteJ6K0rgopJXpOBKaaZFZa5JAVoAAwH8ltyN4zcAVFJWBXk_xPaLheixi9NMg3XRdjRj6NBNNjmkNnna4tDjlGcaU8CMlzrkoV82-pvckMYp25jQ09jbFu_JTbDdiA9rbsjny_PH_sBOb6_H_dOJOQF6YvbslNfcK4_SWycMBGedCdy7GmQI2iDIUhovpShLrKpQe3UOaMA6VBrEhuz-f10exnFhbn7yApDnhkNzcdKsTprVifgD6gtV3g</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>High-fidelity facial reflectance and geometry inference from an unconstrained image</title><source>ACM Digital Library Complete</source><creator>Yamaguchi, Shugo ; Saito, Shunsuke ; Nagano, Koki ; Zhao, Yajie ; Chen, Weikai ; Olszewski, Kyle ; Morishima, Shigeo ; Li, Hao</creator><creatorcontrib>Yamaguchi, Shugo ; Saito, Shunsuke ; Nagano, Koki ; Zhao, Yajie ; Chen, Weikai ; Olszewski, Kyle ; Morishima, Shigeo ; Li, Hao</creatorcontrib><description>We present a deep learning-based technique to infer high-quality facial reflectance and geometry given a single unconstrained image of the subject, which may contain partial occlusions and arbitrary illumination conditions. The reconstructed high-resolution textures, which are generated in only a few seconds, include high-resolution skin surface reflectance maps, representing both the diffuse and specular albedo, and medium- and high-frequency displacement maps, thereby allowing us to render compelling digital avatars under novel lighting conditions. To extract this data, we train our deep neural networks with a high-quality skin reflectance and geometry database created with a state-of-the-art multi-view photometric stereo system using polarized gradient illumination. Given the raw facial texture map extracted from the input image, our neural networks synthesize complete reflectance and displacement maps, as well as complete missing regions caused by occlusions. The completed textures exhibit consistent quality throughout the face due to our network architecture, which propagates texture features from the visible region, resulting in high-fidelity details that are consistent with those seen in visible regions. We describe how this highly underconstrained problem is made tractable by dividing the full inference into smaller tasks, which are addressed by dedicated neural networks. We demonstrate the effectiveness of our network design with robust texture completion from images of faces that are largely occluded. With the inferred reflectance and geometry data, we demonstrate the rendering of high-fidelity 3D avatars from a variety of subjects captured under different lighting conditions. In addition, we perform evaluations demonstrating that our method can infer plausible facial reflectance and geometric details comparable to those obtained from high-end capture devices, and outperform alternative approaches that require only a single unconstrained input image.</description><identifier>ISSN: 0730-0301</identifier><identifier>EISSN: 1557-7368</identifier><identifier>DOI: 10.1145/3197517.3201364</identifier><language>eng</language><ispartof>ACM transactions on graphics, 2018-08, Vol.37 (4), p.1-14</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c307t-abc5d71d5de4dac380fcac8f1dc904ff78e04248d44322e66f9d5bfe80ace5703</citedby><cites>FETCH-LOGICAL-c307t-abc5d71d5de4dac380fcac8f1dc904ff78e04248d44322e66f9d5bfe80ace5703</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Yamaguchi, Shugo</creatorcontrib><creatorcontrib>Saito, Shunsuke</creatorcontrib><creatorcontrib>Nagano, Koki</creatorcontrib><creatorcontrib>Zhao, Yajie</creatorcontrib><creatorcontrib>Chen, Weikai</creatorcontrib><creatorcontrib>Olszewski, Kyle</creatorcontrib><creatorcontrib>Morishima, Shigeo</creatorcontrib><creatorcontrib>Li, Hao</creatorcontrib><title>High-fidelity facial reflectance and geometry inference from an unconstrained image</title><title>ACM transactions on graphics</title><description>We present a deep learning-based technique to infer high-quality facial reflectance and geometry given a single unconstrained image of the subject, which may contain partial occlusions and arbitrary illumination conditions. The reconstructed high-resolution textures, which are generated in only a few seconds, include high-resolution skin surface reflectance maps, representing both the diffuse and specular albedo, and medium- and high-frequency displacement maps, thereby allowing us to render compelling digital avatars under novel lighting conditions. To extract this data, we train our deep neural networks with a high-quality skin reflectance and geometry database created with a state-of-the-art multi-view photometric stereo system using polarized gradient illumination. Given the raw facial texture map extracted from the input image, our neural networks synthesize complete reflectance and displacement maps, as well as complete missing regions caused by occlusions. The completed textures exhibit consistent quality throughout the face due to our network architecture, which propagates texture features from the visible region, resulting in high-fidelity details that are consistent with those seen in visible regions. We describe how this highly underconstrained problem is made tractable by dividing the full inference into smaller tasks, which are addressed by dedicated neural networks. We demonstrate the effectiveness of our network design with robust texture completion from images of faces that are largely occluded. With the inferred reflectance and geometry data, we demonstrate the rendering of high-fidelity 3D avatars from a variety of subjects captured under different lighting conditions. In addition, we perform evaluations demonstrating that our method can infer plausible facial reflectance and geometric details comparable to those obtained from high-end capture devices, and outperform alternative approaches that require only a single unconstrained input image.</description><issn>0730-0301</issn><issn>1557-7368</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><recordid>eNotkMFKxDAURYMoWEfXbvMDmXlpkiZdyqDOwIALdV0yyUuNtKmkddG_t4NdXbgXLodDyCOHLedS7QSvteJ6K0rgopJXpOBKaaZFZa5JAVoAAwH8ltyN4zcAVFJWBXk_xPaLheixi9NMg3XRdjRj6NBNNjmkNnna4tDjlGcaU8CMlzrkoV82-pvckMYp25jQ09jbFu_JTbDdiA9rbsjny_PH_sBOb6_H_dOJOQF6YvbslNfcK4_SWycMBGedCdy7GmQI2iDIUhovpShLrKpQe3UOaMA6VBrEhuz-f10exnFhbn7yApDnhkNzcdKsTprVifgD6gtV3g</recordid><startdate>20180831</startdate><enddate>20180831</enddate><creator>Yamaguchi, Shugo</creator><creator>Saito, Shunsuke</creator><creator>Nagano, Koki</creator><creator>Zhao, Yajie</creator><creator>Chen, Weikai</creator><creator>Olszewski, Kyle</creator><creator>Morishima, Shigeo</creator><creator>Li, Hao</creator><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20180831</creationdate><title>High-fidelity facial reflectance and geometry inference from an unconstrained image</title><author>Yamaguchi, Shugo ; Saito, Shunsuke ; Nagano, Koki ; Zhao, Yajie ; Chen, Weikai ; Olszewski, Kyle ; Morishima, Shigeo ; Li, Hao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c307t-abc5d71d5de4dac380fcac8f1dc904ff78e04248d44322e66f9d5bfe80ace5703</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yamaguchi, Shugo</creatorcontrib><creatorcontrib>Saito, Shunsuke</creatorcontrib><creatorcontrib>Nagano, Koki</creatorcontrib><creatorcontrib>Zhao, Yajie</creatorcontrib><creatorcontrib>Chen, Weikai</creatorcontrib><creatorcontrib>Olszewski, Kyle</creatorcontrib><creatorcontrib>Morishima, Shigeo</creatorcontrib><creatorcontrib>Li, Hao</creatorcontrib><collection>CrossRef</collection><jtitle>ACM transactions on graphics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yamaguchi, Shugo</au><au>Saito, Shunsuke</au><au>Nagano, Koki</au><au>Zhao, Yajie</au><au>Chen, Weikai</au><au>Olszewski, Kyle</au><au>Morishima, Shigeo</au><au>Li, Hao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>High-fidelity facial reflectance and geometry inference from an unconstrained image</atitle><jtitle>ACM transactions on graphics</jtitle><date>2018-08-31</date><risdate>2018</risdate><volume>37</volume><issue>4</issue><spage>1</spage><epage>14</epage><pages>1-14</pages><issn>0730-0301</issn><eissn>1557-7368</eissn><abstract>We present a deep learning-based technique to infer high-quality facial reflectance and geometry given a single unconstrained image of the subject, which may contain partial occlusions and arbitrary illumination conditions. The reconstructed high-resolution textures, which are generated in only a few seconds, include high-resolution skin surface reflectance maps, representing both the diffuse and specular albedo, and medium- and high-frequency displacement maps, thereby allowing us to render compelling digital avatars under novel lighting conditions. To extract this data, we train our deep neural networks with a high-quality skin reflectance and geometry database created with a state-of-the-art multi-view photometric stereo system using polarized gradient illumination. Given the raw facial texture map extracted from the input image, our neural networks synthesize complete reflectance and displacement maps, as well as complete missing regions caused by occlusions. The completed textures exhibit consistent quality throughout the face due to our network architecture, which propagates texture features from the visible region, resulting in high-fidelity details that are consistent with those seen in visible regions. We describe how this highly underconstrained problem is made tractable by dividing the full inference into smaller tasks, which are addressed by dedicated neural networks. We demonstrate the effectiveness of our network design with robust texture completion from images of faces that are largely occluded. With the inferred reflectance and geometry data, we demonstrate the rendering of high-fidelity 3D avatars from a variety of subjects captured under different lighting conditions. In addition, we perform evaluations demonstrating that our method can infer plausible facial reflectance and geometric details comparable to those obtained from high-end capture devices, and outperform alternative approaches that require only a single unconstrained input image.</abstract><doi>10.1145/3197517.3201364</doi><tpages>14</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0730-0301
ispartof ACM transactions on graphics, 2018-08, Vol.37 (4), p.1-14
issn 0730-0301
1557-7368
language eng
recordid cdi_crossref_primary_10_1145_3197517_3201364
source ACM Digital Library Complete
title High-fidelity facial reflectance and geometry inference from an unconstrained image
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-21T16%3A28%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=High-fidelity%20facial%20reflectance%20and%20geometry%20inference%20from%20an%20unconstrained%20image&rft.jtitle=ACM%20transactions%20on%20graphics&rft.au=Yamaguchi,%20Shugo&rft.date=2018-08-31&rft.volume=37&rft.issue=4&rft.spage=1&rft.epage=14&rft.pages=1-14&rft.issn=0730-0301&rft.eissn=1557-7368&rft_id=info:doi/10.1145/3197517.3201364&rft_dat=%3Ccrossref%3E10_1145_3197517_3201364%3C/crossref%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true