Momentum Contrastive Autoencoder: Using Contrastive Learning for Latent Space Distribution Matching in WAE
Wasserstein autoencoder (WAE) shows that matching two distributions is equivalent to minimizing a simple autoencoder (AE) loss under the constraint that the latent space of this AE matches a pre-specified prior distribution. This latent space distribution matching is a core component of WAE, and a c...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Arpit, Devansh Bhatnagar, Aadyot Wang, Huan Xiong, Caiming |
description | Wasserstein autoencoder (WAE) shows that matching two distributions is
equivalent to minimizing a simple autoencoder (AE) loss under the constraint
that the latent space of this AE matches a pre-specified prior distribution.
This latent space distribution matching is a core component of WAE, and a
challenging task. In this paper, we propose to use the contrastive learning
framework that has been shown to be effective for self-supervised
representation learning, as a means to resolve this problem. We do so by
exploiting the fact that contrastive learning objectives optimize the latent
space distribution to be uniform over the unit hyper-sphere, which can be
easily sampled from. We show that using the contrastive learning framework to
optimize the WAE loss achieves faster convergence and more stable optimization
compared with existing popular algorithms for WAE. This is also reflected in
the FID scores on CelebA and CIFAR-10 datasets, and the realistic generated
image quality on the CelebA-HQ dataset. |
doi_str_mv | 10.48550/arxiv.2110.10303 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2110_10303</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2110_10303</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-dcecfe8059365d7d83e09c245d45ec427a121a682af4c34403f0f08416e3999c3</originalsourceid><addsrcrecordid>eNpVj71OwzAUhb0woMIDMNUvkGLHduKwRaH8SKkYaMUYXZxrMCJ25TgVvD1NYWE60neOjvQRcsXZSmql2DXEL3dY5fwIOBNMnJOPTRjQp2mgTfApwpjcAWk9pYDehB7jDd2Nzr_9q1uE6GdoQ6QtpOMBfd6DQXrrxhTd65Rc8HQDybzPM-fpS72-IGcWPke8_MsF2d6tt81D1j7dPzZ1m0FRiqw3aCxqpipRqL7stUBWmVyqXio0Mi-B5xwKnYOVRkjJhGWWackLFFVVGbEgy9_bk2y3j26A-N3N0t1JWvwAzC9Syg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Momentum Contrastive Autoencoder: Using Contrastive Learning for Latent Space Distribution Matching in WAE</title><source>arXiv.org</source><creator>Arpit, Devansh ; Bhatnagar, Aadyot ; Wang, Huan ; Xiong, Caiming</creator><creatorcontrib>Arpit, Devansh ; Bhatnagar, Aadyot ; Wang, Huan ; Xiong, Caiming</creatorcontrib><description>Wasserstein autoencoder (WAE) shows that matching two distributions is
equivalent to minimizing a simple autoencoder (AE) loss under the constraint
that the latent space of this AE matches a pre-specified prior distribution.
This latent space distribution matching is a core component of WAE, and a
challenging task. In this paper, we propose to use the contrastive learning
framework that has been shown to be effective for self-supervised
representation learning, as a means to resolve this problem. We do so by
exploiting the fact that contrastive learning objectives optimize the latent
space distribution to be uniform over the unit hyper-sphere, which can be
easily sampled from. We show that using the contrastive learning framework to
optimize the WAE loss achieves faster convergence and more stable optimization
compared with existing popular algorithms for WAE. This is also reflected in
the FID scores on CelebA and CIFAR-10 datasets, and the realistic generated
image quality on the CelebA-HQ dataset.</description><identifier>DOI: 10.48550/arxiv.2110.10303</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning</subject><creationdate>2021-10</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2110.10303$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2110.10303$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Arpit, Devansh</creatorcontrib><creatorcontrib>Bhatnagar, Aadyot</creatorcontrib><creatorcontrib>Wang, Huan</creatorcontrib><creatorcontrib>Xiong, Caiming</creatorcontrib><title>Momentum Contrastive Autoencoder: Using Contrastive Learning for Latent Space Distribution Matching in WAE</title><description>Wasserstein autoencoder (WAE) shows that matching two distributions is
equivalent to minimizing a simple autoencoder (AE) loss under the constraint
that the latent space of this AE matches a pre-specified prior distribution.
This latent space distribution matching is a core component of WAE, and a
challenging task. In this paper, we propose to use the contrastive learning
framework that has been shown to be effective for self-supervised
representation learning, as a means to resolve this problem. We do so by
exploiting the fact that contrastive learning objectives optimize the latent
space distribution to be uniform over the unit hyper-sphere, which can be
easily sampled from. We show that using the contrastive learning framework to
optimize the WAE loss achieves faster convergence and more stable optimization
compared with existing popular algorithms for WAE. This is also reflected in
the FID scores on CelebA and CIFAR-10 datasets, and the realistic generated
image quality on the CelebA-HQ dataset.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpVj71OwzAUhb0woMIDMNUvkGLHduKwRaH8SKkYaMUYXZxrMCJ25TgVvD1NYWE60neOjvQRcsXZSmql2DXEL3dY5fwIOBNMnJOPTRjQp2mgTfApwpjcAWk9pYDehB7jDd2Nzr_9q1uE6GdoQ6QtpOMBfd6DQXrrxhTd65Rc8HQDybzPM-fpS72-IGcWPke8_MsF2d6tt81D1j7dPzZ1m0FRiqw3aCxqpipRqL7stUBWmVyqXio0Mi-B5xwKnYOVRkjJhGWWackLFFVVGbEgy9_bk2y3j26A-N3N0t1JWvwAzC9Syg</recordid><startdate>20211019</startdate><enddate>20211019</enddate><creator>Arpit, Devansh</creator><creator>Bhatnagar, Aadyot</creator><creator>Wang, Huan</creator><creator>Xiong, Caiming</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20211019</creationdate><title>Momentum Contrastive Autoencoder: Using Contrastive Learning for Latent Space Distribution Matching in WAE</title><author>Arpit, Devansh ; Bhatnagar, Aadyot ; Wang, Huan ; Xiong, Caiming</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-dcecfe8059365d7d83e09c245d45ec427a121a682af4c34403f0f08416e3999c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Arpit, Devansh</creatorcontrib><creatorcontrib>Bhatnagar, Aadyot</creatorcontrib><creatorcontrib>Wang, Huan</creatorcontrib><creatorcontrib>Xiong, Caiming</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Arpit, Devansh</au><au>Bhatnagar, Aadyot</au><au>Wang, Huan</au><au>Xiong, Caiming</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Momentum Contrastive Autoencoder: Using Contrastive Learning for Latent Space Distribution Matching in WAE</atitle><date>2021-10-19</date><risdate>2021</risdate><abstract>Wasserstein autoencoder (WAE) shows that matching two distributions is
equivalent to minimizing a simple autoencoder (AE) loss under the constraint
that the latent space of this AE matches a pre-specified prior distribution.
This latent space distribution matching is a core component of WAE, and a
challenging task. In this paper, we propose to use the contrastive learning
framework that has been shown to be effective for self-supervised
representation learning, as a means to resolve this problem. We do so by
exploiting the fact that contrastive learning objectives optimize the latent
space distribution to be uniform over the unit hyper-sphere, which can be
easily sampled from. We show that using the contrastive learning framework to
optimize the WAE loss achieves faster convergence and more stable optimization
compared with existing popular algorithms for WAE. This is also reflected in
the FID scores on CelebA and CIFAR-10 datasets, and the realistic generated
image quality on the CelebA-HQ dataset.</abstract><doi>10.48550/arxiv.2110.10303</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2110.10303 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2110_10303 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
title | Momentum Contrastive Autoencoder: Using Contrastive Learning for Latent Space Distribution Matching in WAE |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T13%3A36%3A32IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Momentum%20Contrastive%20Autoencoder:%20Using%20Contrastive%20Learning%20for%20Latent%20Space%20Distribution%20Matching%20in%20WAE&rft.au=Arpit,%20Devansh&rft.date=2021-10-19&rft_id=info:doi/10.48550/arxiv.2110.10303&rft_dat=%3Carxiv_GOX%3E2110_10303%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |