Examining Performance of Sketch-to-Image Translation Models with Multiclass Automatically Generated Paired Training Data

Image translation is a computer vision task that involves translating one representation of the scene into another. Various approaches have been proposed and achieved highly desirable results. Nevertheless, its accomplishment requires abundant paired training data which are expensive to acquire. The...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
1. Verfasser: Hu, Dichao
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Hu, Dichao
description Image translation is a computer vision task that involves translating one representation of the scene into another. Various approaches have been proposed and achieved highly desirable results. Nevertheless, its accomplishment requires abundant paired training data which are expensive to acquire. Therefore, models for translation are usually trained on a set of paired training data which are carefully and laboriously designed. Our work is focused on learning through automatically generated paired data. We propose a method to generate fake sketches from images using an adversarial network and then pair the images with corresponding fake sketches to form large-scale multi-class paired training data for training a sketch-to-image translation model. Our model is an encoder-decoder architecture where the encoder generates fake sketches from images and the decoder performs sketch-to-image translation. Qualitative results show that the encoder can be used for generating large-scale multi-class paired data under low supervision. Our current dataset now contains 61255 image and (fake) sketch pairs from 256 different categories. These figures can be greatly increased in the future thanks to our weak reliance on manually labeled data.
doi_str_mv 10.48550/arxiv.1811.00249
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1811_00249</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1811_00249</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-e4911070e9122059923d616215f86a61543747f51d895da8751ba9469aab93793</originalsourceid><addsrcrecordid>eNotkEFOwzAURL1hgQoHYIUvkOCf2E68rEoplVpRieyj38RpLRwbOS6ktye0Xb3NaEbzCHkClvJSCPaCYTQ_KZQAKWMZV_dkXI7YG2fcge506Hzo0TWa-o5-funYHJPok3WPB02rgG6wGI13dOtbbQf6a-KRbk82msbiMND5Kfp-SjRo7ZmutNMBo27pDk2YMDVcl14x4gO569AO-vHGGaneltXiPdl8rNaL-SZBWahEcwXACqYVZBkTSmV5K0FmILpSogTB84IXnYC2VKLFshCwR8WlQtyrvFD5jDxfay_X6-9gegzn-l9BfVGQ_wHJ7Vbe</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Examining Performance of Sketch-to-Image Translation Models with Multiclass Automatically Generated Paired Training Data</title><source>arXiv.org</source><creator>Hu, Dichao</creator><creatorcontrib>Hu, Dichao</creatorcontrib><description>Image translation is a computer vision task that involves translating one representation of the scene into another. Various approaches have been proposed and achieved highly desirable results. Nevertheless, its accomplishment requires abundant paired training data which are expensive to acquire. Therefore, models for translation are usually trained on a set of paired training data which are carefully and laboriously designed. Our work is focused on learning through automatically generated paired data. We propose a method to generate fake sketches from images using an adversarial network and then pair the images with corresponding fake sketches to form large-scale multi-class paired training data for training a sketch-to-image translation model. Our model is an encoder-decoder architecture where the encoder generates fake sketches from images and the decoder performs sketch-to-image translation. Qualitative results show that the encoder can be used for generating large-scale multi-class paired data under low supervision. Our current dataset now contains 61255 image and (fake) sketch pairs from 256 different categories. These figures can be greatly increased in the future thanks to our weak reliance on manually labeled data.</description><identifier>DOI: 10.48550/arxiv.1811.00249</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2018-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1811.00249$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1811.00249$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Hu, Dichao</creatorcontrib><title>Examining Performance of Sketch-to-Image Translation Models with Multiclass Automatically Generated Paired Training Data</title><description>Image translation is a computer vision task that involves translating one representation of the scene into another. Various approaches have been proposed and achieved highly desirable results. Nevertheless, its accomplishment requires abundant paired training data which are expensive to acquire. Therefore, models for translation are usually trained on a set of paired training data which are carefully and laboriously designed. Our work is focused on learning through automatically generated paired data. We propose a method to generate fake sketches from images using an adversarial network and then pair the images with corresponding fake sketches to form large-scale multi-class paired training data for training a sketch-to-image translation model. Our model is an encoder-decoder architecture where the encoder generates fake sketches from images and the decoder performs sketch-to-image translation. Qualitative results show that the encoder can be used for generating large-scale multi-class paired data under low supervision. Our current dataset now contains 61255 image and (fake) sketch pairs from 256 different categories. These figures can be greatly increased in the future thanks to our weak reliance on manually labeled data.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotkEFOwzAURL1hgQoHYIUvkOCf2E68rEoplVpRieyj38RpLRwbOS6ktye0Xb3NaEbzCHkClvJSCPaCYTQ_KZQAKWMZV_dkXI7YG2fcge506Hzo0TWa-o5-funYHJPok3WPB02rgG6wGI13dOtbbQf6a-KRbk82msbiMND5Kfp-SjRo7ZmutNMBo27pDk2YMDVcl14x4gO569AO-vHGGaneltXiPdl8rNaL-SZBWahEcwXACqYVZBkTSmV5K0FmILpSogTB84IXnYC2VKLFshCwR8WlQtyrvFD5jDxfay_X6-9gegzn-l9BfVGQ_wHJ7Vbe</recordid><startdate>20181101</startdate><enddate>20181101</enddate><creator>Hu, Dichao</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20181101</creationdate><title>Examining Performance of Sketch-to-Image Translation Models with Multiclass Automatically Generated Paired Training Data</title><author>Hu, Dichao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-e4911070e9122059923d616215f86a61543747f51d895da8751ba9469aab93793</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Hu, Dichao</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Hu, Dichao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Examining Performance of Sketch-to-Image Translation Models with Multiclass Automatically Generated Paired Training Data</atitle><date>2018-11-01</date><risdate>2018</risdate><abstract>Image translation is a computer vision task that involves translating one representation of the scene into another. Various approaches have been proposed and achieved highly desirable results. Nevertheless, its accomplishment requires abundant paired training data which are expensive to acquire. Therefore, models for translation are usually trained on a set of paired training data which are carefully and laboriously designed. Our work is focused on learning through automatically generated paired data. We propose a method to generate fake sketches from images using an adversarial network and then pair the images with corresponding fake sketches to form large-scale multi-class paired training data for training a sketch-to-image translation model. Our model is an encoder-decoder architecture where the encoder generates fake sketches from images and the decoder performs sketch-to-image translation. Qualitative results show that the encoder can be used for generating large-scale multi-class paired data under low supervision. Our current dataset now contains 61255 image and (fake) sketch pairs from 256 different categories. These figures can be greatly increased in the future thanks to our weak reliance on manually labeled data.</abstract><doi>10.48550/arxiv.1811.00249</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1811.00249
ispartof
issn
language eng
recordid cdi_arxiv_primary_1811_00249
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title Examining Performance of Sketch-to-Image Translation Models with Multiclass Automatically Generated Paired Training Data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T02%3A28%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Examining%20Performance%20of%20Sketch-to-Image%20Translation%20Models%20with%20Multiclass%20Automatically%20Generated%20Paired%20Training%20Data&rft.au=Hu,%20Dichao&rft.date=2018-11-01&rft_id=info:doi/10.48550/arxiv.1811.00249&rft_dat=%3Carxiv_GOX%3E1811_00249%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true