Constraining Variational Inference with Geometric Jensen-Shannon Divergence

We examine the problem of controlling divergences for latent space regularisation in variational autoencoders. Specifically, when aiming to reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$ ($n\leq m$), while balancing this against the need for generalisable latent repre...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Deasy, Jacob, Simidjievski, Nikola, Liò, Pietro
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Deasy, Jacob
Simidjievski, Nikola
Liò, Pietro
description We examine the problem of controlling divergences for latent space regularisation in variational autoencoders. Specifically, when aiming to reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$ ($n\leq m$), while balancing this against the need for generalisable latent representations. We present a regularisation mechanism based on the skew-geometric Jensen-Shannon divergence $\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in $\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads to an intuitive interpolation between forward and reverse KL in the space of both distributions and divergences. We motivate its potential benefits for VAEs through low-dimensional examples, before presenting quantitative and qualitative results. Our experiments demonstrate that skewing our variant of $\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of $\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and generation when compared to several baseline VAEs. Our approach is entirely unsupervised and utilises only one hyperparameter which can be easily interpreted in latent space.
doi_str_mv 10.48550/arxiv.2006.10599
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2006_10599</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2006_10599</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-81f92acea6a68f1906b1c1a2696a46e0001808e3ce5304596aea41f6e280c9b63</originalsourceid><addsrcrecordid>eNotz8FOwzAQBFBfOKDSD-CEfyBhncSLfUQpLYVKHKi4Rltr3VpqN5UTFfh7aOE00mg00lPq1kDZOGvhnvJXOpUVAJYGrPfX6rXtZRgzJUmy1R-UE42pF9rrpUTOLIH1Zxp3esH9gcecgn5hGViK9x2J9KJn6cR5ex7eqKtI-4Gn_zlR6_nTun0uVm-LZfu4KggffOFM9BUFJiR00XjAjQmGKvRIDTIAGAeO68C2hsb-tkyNiciVg-A3WE_U3d_tRdMdczpQ_u7Oqu6iqn8A7RNH4w</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</title><source>arXiv.org</source><creator>Deasy, Jacob ; Simidjievski, Nikola ; Liò, Pietro</creator><creatorcontrib>Deasy, Jacob ; Simidjievski, Nikola ; Liò, Pietro</creatorcontrib><description>We examine the problem of controlling divergences for latent space regularisation in variational autoencoders. Specifically, when aiming to reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$ ($n\leq m$), while balancing this against the need for generalisable latent representations. We present a regularisation mechanism based on the skew-geometric Jensen-Shannon divergence $\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in $\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads to an intuitive interpolation between forward and reverse KL in the space of both distributions and divergences. We motivate its potential benefits for VAEs through low-dimensional examples, before presenting quantitative and qualitative results. Our experiments demonstrate that skewing our variant of $\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of $\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and generation when compared to several baseline VAEs. Our approach is entirely unsupervised and utilises only one hyperparameter which can be easily interpreted in latent space.</description><identifier>DOI: 10.48550/arxiv.2006.10599</identifier><language>eng</language><subject>Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2020-06</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2006.10599$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2006.10599$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Deasy, Jacob</creatorcontrib><creatorcontrib>Simidjievski, Nikola</creatorcontrib><creatorcontrib>Liò, Pietro</creatorcontrib><title>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</title><description>We examine the problem of controlling divergences for latent space regularisation in variational autoencoders. Specifically, when aiming to reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$ ($n\leq m$), while balancing this against the need for generalisable latent representations. We present a regularisation mechanism based on the skew-geometric Jensen-Shannon divergence $\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in $\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads to an intuitive interpolation between forward and reverse KL in the space of both distributions and divergences. We motivate its potential benefits for VAEs through low-dimensional examples, before presenting quantitative and qualitative results. Our experiments demonstrate that skewing our variant of $\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of $\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and generation when compared to several baseline VAEs. Our approach is entirely unsupervised and utilises only one hyperparameter which can be easily interpreted in latent space.</description><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz8FOwzAQBFBfOKDSD-CEfyBhncSLfUQpLYVKHKi4Rltr3VpqN5UTFfh7aOE00mg00lPq1kDZOGvhnvJXOpUVAJYGrPfX6rXtZRgzJUmy1R-UE42pF9rrpUTOLIH1Zxp3esH9gcecgn5hGViK9x2J9KJn6cR5ex7eqKtI-4Gn_zlR6_nTun0uVm-LZfu4KggffOFM9BUFJiR00XjAjQmGKvRIDTIAGAeO68C2hsb-tkyNiciVg-A3WE_U3d_tRdMdczpQ_u7Oqu6iqn8A7RNH4w</recordid><startdate>20200618</startdate><enddate>20200618</enddate><creator>Deasy, Jacob</creator><creator>Simidjievski, Nikola</creator><creator>Liò, Pietro</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20200618</creationdate><title>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</title><author>Deasy, Jacob ; Simidjievski, Nikola ; Liò, Pietro</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-81f92acea6a68f1906b1c1a2696a46e0001808e3ce5304596aea41f6e280c9b63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Deasy, Jacob</creatorcontrib><creatorcontrib>Simidjievski, Nikola</creatorcontrib><creatorcontrib>Liò, Pietro</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Deasy, Jacob</au><au>Simidjievski, Nikola</au><au>Liò, Pietro</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</atitle><date>2020-06-18</date><risdate>2020</risdate><abstract>We examine the problem of controlling divergences for latent space regularisation in variational autoencoders. Specifically, when aiming to reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$ ($n\leq m$), while balancing this against the need for generalisable latent representations. We present a regularisation mechanism based on the skew-geometric Jensen-Shannon divergence $\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in $\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads to an intuitive interpolation between forward and reverse KL in the space of both distributions and divergences. We motivate its potential benefits for VAEs through low-dimensional examples, before presenting quantitative and qualitative results. Our experiments demonstrate that skewing our variant of $\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of $\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and generation when compared to several baseline VAEs. Our approach is entirely unsupervised and utilises only one hyperparameter which can be easily interpreted in latent space.</abstract><doi>10.48550/arxiv.2006.10599</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2006.10599
ispartof
issn
language eng
recordid cdi_arxiv_primary_2006_10599
source arXiv.org
subjects Computer Science - Learning
Statistics - Machine Learning
title Constraining Variational Inference with Geometric Jensen-Shannon Divergence
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T06%3A38%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Constraining%20Variational%20Inference%20with%20Geometric%20Jensen-Shannon%20Divergence&rft.au=Deasy,%20Jacob&rft.date=2020-06-18&rft_id=info:doi/10.48550/arxiv.2006.10599&rft_dat=%3Carxiv_GOX%3E2006_10599%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true