End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On

The 2D virtual try-on task has recently attracted a lot of interest from the research community, for its direct potential applications in online shopping as well as for its inherent and non-addressed scientific challenges. This task requires to fit an in-shop cloth image on the image of a person. It...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Issenhuth, Thibaut, Mary, Jérémie, Calauzènes, Clément
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Issenhuth, Thibaut
Mary, Jérémie
Calauzènes, Clément
description The 2D virtual try-on task has recently attracted a lot of interest from the research community, for its direct potential applications in online shopping as well as for its inherent and non-addressed scientific challenges. This task requires to fit an in-shop cloth image on the image of a person. It is highly challenging because it requires to warp the cloth on the target person while preserving its patterns and characteristics, and to compose the item with the person in a realistic manner. Current state-of-the-art models generate images with visible artifacts, due either to a pixel-level composition step or to the geometric transformation. In this paper, we propose WUTON: a Warping U-net for a Virtual Try-On system. It is a siamese U-net generator whose skip connections are geometrically transformed by a convolutional geometric matcher. The whole architecture is trained end-to-end with a multi-task loss including an adversarial one. This enables our network to generate and use realistic spatial transformations of the cloth to synthesize images of high visual quality. The proposed architecture can be trained end-to-end and allows us to advance towards a detail-preserving and photo-realistic 2D virtual try-on system. Our method outperforms the current state-of-the-art with visual results as well as with the Learned Perceptual Image Similarity (LPIPS) metric.
doi_str_mv 10.48550/arxiv.1906.01347
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1906_01347</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1906_01347</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-79472c005282086a2857a37d55ab741dce11ac8ccb0f4108ab5ba75ab961146c3</originalsourceid><addsrcrecordid>eNotj71OwzAYAL0woMIDMOEXcLAT_2VEpS1IQWWIWKMvjoMsNXb1xUX07UkL0w0nnXSEPAheSKsUfwL8Cd-FqLkuuKikuSUfmziwnNgC2njAGOIXTSPd-TT5jMHRFz8mnCCHFOeL2XrIJ_T0HY4zXRT9DJhPcKAtntk-3pGbEQ6zv__nirTbTbt-Zc1-97Z-bhhoY5ippSkd56q0JbcaSqsMVGZQCnojxeC8EOCscz0fpeAWetWDWWSthZDaVSvy-Je9LnVHDBPgubusdde16hdggkgk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</title><source>arXiv.org</source><creator>Issenhuth, Thibaut ; Mary, Jérémie ; Calauzènes, Clément</creator><creatorcontrib>Issenhuth, Thibaut ; Mary, Jérémie ; Calauzènes, Clément</creatorcontrib><description>The 2D virtual try-on task has recently attracted a lot of interest from the research community, for its direct potential applications in online shopping as well as for its inherent and non-addressed scientific challenges. This task requires to fit an in-shop cloth image on the image of a person. It is highly challenging because it requires to warp the cloth on the target person while preserving its patterns and characteristics, and to compose the item with the person in a realistic manner. Current state-of-the-art models generate images with visible artifacts, due either to a pixel-level composition step or to the geometric transformation. In this paper, we propose WUTON: a Warping U-net for a Virtual Try-On system. It is a siamese U-net generator whose skip connections are geometrically transformed by a convolutional geometric matcher. The whole architecture is trained end-to-end with a multi-task loss including an adversarial one. This enables our network to generate and use realistic spatial transformations of the cloth to synthesize images of high visual quality. The proposed architecture can be trained end-to-end and allows us to advance towards a detail-preserving and photo-realistic 2D virtual try-on system. Our method outperforms the current state-of-the-art with visual results as well as with the Learned Perceptual Image Similarity (LPIPS) metric.</description><identifier>DOI: 10.48550/arxiv.1906.01347</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2019-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1906.01347$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1906.01347$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Issenhuth, Thibaut</creatorcontrib><creatorcontrib>Mary, Jérémie</creatorcontrib><creatorcontrib>Calauzènes, Clément</creatorcontrib><title>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</title><description>The 2D virtual try-on task has recently attracted a lot of interest from the research community, for its direct potential applications in online shopping as well as for its inherent and non-addressed scientific challenges. This task requires to fit an in-shop cloth image on the image of a person. It is highly challenging because it requires to warp the cloth on the target person while preserving its patterns and characteristics, and to compose the item with the person in a realistic manner. Current state-of-the-art models generate images with visible artifacts, due either to a pixel-level composition step or to the geometric transformation. In this paper, we propose WUTON: a Warping U-net for a Virtual Try-On system. It is a siamese U-net generator whose skip connections are geometrically transformed by a convolutional geometric matcher. The whole architecture is trained end-to-end with a multi-task loss including an adversarial one. This enables our network to generate and use realistic spatial transformations of the cloth to synthesize images of high visual quality. The proposed architecture can be trained end-to-end and allows us to advance towards a detail-preserving and photo-realistic 2D virtual try-on system. Our method outperforms the current state-of-the-art with visual results as well as with the Learned Perceptual Image Similarity (LPIPS) metric.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71OwzAYAL0woMIDMOEXcLAT_2VEpS1IQWWIWKMvjoMsNXb1xUX07UkL0w0nnXSEPAheSKsUfwL8Cd-FqLkuuKikuSUfmziwnNgC2njAGOIXTSPd-TT5jMHRFz8mnCCHFOeL2XrIJ_T0HY4zXRT9DJhPcKAtntk-3pGbEQ6zv__nirTbTbt-Zc1-97Z-bhhoY5ippSkd56q0JbcaSqsMVGZQCnojxeC8EOCscz0fpeAWetWDWWSthZDaVSvy-Je9LnVHDBPgubusdde16hdggkgk</recordid><startdate>20190604</startdate><enddate>20190604</enddate><creator>Issenhuth, Thibaut</creator><creator>Mary, Jérémie</creator><creator>Calauzènes, Clément</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20190604</creationdate><title>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</title><author>Issenhuth, Thibaut ; Mary, Jérémie ; Calauzènes, Clément</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-79472c005282086a2857a37d55ab741dce11ac8ccb0f4108ab5ba75ab961146c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Issenhuth, Thibaut</creatorcontrib><creatorcontrib>Mary, Jérémie</creatorcontrib><creatorcontrib>Calauzènes, Clément</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Issenhuth, Thibaut</au><au>Mary, Jérémie</au><au>Calauzènes, Clément</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On</atitle><date>2019-06-04</date><risdate>2019</risdate><abstract>The 2D virtual try-on task has recently attracted a lot of interest from the research community, for its direct potential applications in online shopping as well as for its inherent and non-addressed scientific challenges. This task requires to fit an in-shop cloth image on the image of a person. It is highly challenging because it requires to warp the cloth on the target person while preserving its patterns and characteristics, and to compose the item with the person in a realistic manner. Current state-of-the-art models generate images with visible artifacts, due either to a pixel-level composition step or to the geometric transformation. In this paper, we propose WUTON: a Warping U-net for a Virtual Try-On system. It is a siamese U-net generator whose skip connections are geometrically transformed by a convolutional geometric matcher. The whole architecture is trained end-to-end with a multi-task loss including an adversarial one. This enables our network to generate and use realistic spatial transformations of the cloth to synthesize images of high visual quality. The proposed architecture can be trained end-to-end and allows us to advance towards a detail-preserving and photo-realistic 2D virtual try-on system. Our method outperforms the current state-of-the-art with visual results as well as with the Learned Perceptual Image Similarity (LPIPS) metric.</abstract><doi>10.48550/arxiv.1906.01347</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1906.01347
ispartof
issn
language eng
recordid cdi_arxiv_primary_1906_01347
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title End-to-End Learning of Geometric Deformations of Feature Maps for Virtual Try-On
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T04%3A46%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=End-to-End%20Learning%20of%20Geometric%20Deformations%20of%20Feature%20Maps%20for%20Virtual%20Try-On&rft.au=Issenhuth,%20Thibaut&rft.date=2019-06-04&rft_id=info:doi/10.48550/arxiv.1906.01347&rft_dat=%3Carxiv_GOX%3E1906_01347%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true