End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On
The 2D virtual try-on task has recently attracted a lot of interest from the research community, for its direct potential applications in online shopping as well as for its inherent and non-addressed scientific challenges. This task requires to fit an in-shop cloth image on the image of a person. It...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Issenhuth, Thibaut Mary, Jérémie Calauzènes, Clément |
description | The 2D virtual try-on task has recently attracted a lot of interest from the
research community, for its direct potential applications in online shopping as
well as for its inherent and non-addressed scientific challenges. This task
requires to fit an in-shop cloth image on the image of a person. It is highly
challenging because it requires to warp the cloth on the target person while
preserving its patterns and characteristics, and to compose the item with the
person in a realistic manner. Current state-of-the-art models generate images
with visible artifacts, due either to a pixel-level composition step or to the
geometric transformation. In this paper, we propose WUTON: a Warping U-net for
a Virtual Try-On system. It is a siamese U-net generator whose skip connections
are geometrically transformed by a convolutional geometric matcher. The whole
architecture is trained end-to-end with a multi-task loss including an
adversarial one. This enables our network to generate and use realistic spatial
transformations of the cloth to synthesize images of high visual quality. The
proposed architecture can be trained end-to-end and allows us to advance
towards a detail-preserving and photo-realistic 2D virtual try-on system. Our
method outperforms the current state-of-the-art with visual results as well as
with the Learned Perceptual Image Similarity (LPIPS) metric. |
doi_str_mv | 10.48550/arxiv.1906.01347 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1906_01347</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1906_01347</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-79472c005282086a2857a37d55ab741dce11ac8ccb0f4108ab5ba75ab961146c3</originalsourceid><addsrcrecordid>eNotj71OwzAYAL0woMIDMOEXcLAT_2VEpS1IQWWIWKMvjoMsNXb1xUX07UkL0w0nnXSEPAheSKsUfwL8Cd-FqLkuuKikuSUfmziwnNgC2njAGOIXTSPd-TT5jMHRFz8mnCCHFOeL2XrIJ_T0HY4zXRT9DJhPcKAtntk-3pGbEQ6zv__nirTbTbt-Zc1-97Z-bhhoY5ippSkd56q0JbcaSqsMVGZQCnojxeC8EOCscz0fpeAWetWDWWSthZDaVSvy-Je9LnVHDBPgubusdde16hdggkgk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</title><source>arXiv.org</source><creator>Issenhuth, Thibaut ; Mary, Jérémie ; Calauzènes, Clément</creator><creatorcontrib>Issenhuth, Thibaut ; Mary, Jérémie ; Calauzènes, Clément</creatorcontrib><description>The 2D virtual try-on task has recently attracted a lot of interest from the
research community, for its direct potential applications in online shopping as
well as for its inherent and non-addressed scientific challenges. This task
requires to fit an in-shop cloth image on the image of a person. It is highly
challenging because it requires to warp the cloth on the target person while
preserving its patterns and characteristics, and to compose the item with the
person in a realistic manner. Current state-of-the-art models generate images
with visible artifacts, due either to a pixel-level composition step or to the
geometric transformation. In this paper, we propose WUTON: a Warping U-net for
a Virtual Try-On system. It is a siamese U-net generator whose skip connections
are geometrically transformed by a convolutional geometric matcher. The whole
architecture is trained end-to-end with a multi-task loss including an
adversarial one. This enables our network to generate and use realistic spatial
transformations of the cloth to synthesize images of high visual quality. The
proposed architecture can be trained end-to-end and allows us to advance
towards a detail-preserving and photo-realistic 2D virtual try-on system. Our
method outperforms the current state-of-the-art with visual results as well as
with the Learned Perceptual Image Similarity (LPIPS) metric.</description><identifier>DOI: 10.48550/arxiv.1906.01347</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2019-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1906.01347$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1906.01347$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Issenhuth, Thibaut</creatorcontrib><creatorcontrib>Mary, Jérémie</creatorcontrib><creatorcontrib>Calauzènes, Clément</creatorcontrib><title>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</title><description>The 2D virtual try-on task has recently attracted a lot of interest from the
research community, for its direct potential applications in online shopping as
well as for its inherent and non-addressed scientific challenges. This task
requires to fit an in-shop cloth image on the image of a person. It is highly
challenging because it requires to warp the cloth on the target person while
preserving its patterns and characteristics, and to compose the item with the
person in a realistic manner. Current state-of-the-art models generate images
with visible artifacts, due either to a pixel-level composition step or to the
geometric transformation. In this paper, we propose WUTON: a Warping U-net for
a Virtual Try-On system. It is a siamese U-net generator whose skip connections
are geometrically transformed by a convolutional geometric matcher. The whole
architecture is trained end-to-end with a multi-task loss including an
adversarial one. This enables our network to generate and use realistic spatial
transformations of the cloth to synthesize images of high visual quality. The
proposed architecture can be trained end-to-end and allows us to advance
towards a detail-preserving and photo-realistic 2D virtual try-on system. Our
method outperforms the current state-of-the-art with visual results as well as
with the Learned Perceptual Image Similarity (LPIPS) metric.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71OwzAYAL0woMIDMOEXcLAT_2VEpS1IQWWIWKMvjoMsNXb1xUX07UkL0w0nnXSEPAheSKsUfwL8Cd-FqLkuuKikuSUfmziwnNgC2njAGOIXTSPd-TT5jMHRFz8mnCCHFOeL2XrIJ_T0HY4zXRT9DJhPcKAtntk-3pGbEQ6zv__nirTbTbt-Zc1-97Z-bhhoY5ippSkd56q0JbcaSqsMVGZQCnojxeC8EOCscz0fpeAWetWDWWSthZDaVSvy-Je9LnVHDBPgubusdde16hdggkgk</recordid><startdate>20190604</startdate><enddate>20190604</enddate><creator>Issenhuth, Thibaut</creator><creator>Mary, Jérémie</creator><creator>Calauzènes, Clément</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20190604</creationdate><title>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</title><author>Issenhuth, Thibaut ; Mary, Jérémie ; Calauzènes, Clément</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-79472c005282086a2857a37d55ab741dce11ac8ccb0f4108ab5ba75ab961146c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Issenhuth, Thibaut</creatorcontrib><creatorcontrib>Mary, Jérémie</creatorcontrib><creatorcontrib>Calauzènes, Clément</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Issenhuth, Thibaut</au><au>Mary, Jérémie</au><au>Calauzènes, Clément</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</atitle><date>2019-06-04</date><risdate>2019</risdate><abstract>The 2D virtual try-on task has recently attracted a lot of interest from the
research community, for its direct potential applications in online shopping as
well as for its inherent and non-addressed scientific challenges. This task
requires to fit an in-shop cloth image on the image of a person. It is highly
challenging because it requires to warp the cloth on the target person while
preserving its patterns and characteristics, and to compose the item with the
person in a realistic manner. Current state-of-the-art models generate images
with visible artifacts, due either to a pixel-level composition step or to the
geometric transformation. In this paper, we propose WUTON: a Warping U-net for
a Virtual Try-On system. It is a siamese U-net generator whose skip connections
are geometrically transformed by a convolutional geometric matcher. The whole
architecture is trained end-to-end with a multi-task loss including an
adversarial one. This enables our network to generate and use realistic spatial
transformations of the cloth to synthesize images of high visual quality. The
proposed architecture can be trained end-to-end and allows us to advance
towards a detail-preserving and photo-realistic 2D virtual try-on system. Our
method outperforms the current state-of-the-art with visual results as well as
with the Learned Perceptual Image Similarity (LPIPS) metric.</abstract><doi>10.48550/arxiv.1906.01347</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1906.01347 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1906_01347 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T04%3A46%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=End-to-End%20Learning%20of%20Geometric%20Deformations%20of%20Feature%20Maps%20for%20Virtual%20Try-On&rft.au=Issenhuth,%20Thibaut&rft.date=2019-06-04&rft_id=info:doi/10.48550/arxiv.1906.01347&rft_dat=%3Carxiv_GOX%3E1906_01347%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |