Constraining Variational Inference with Geometric Jensen-Shannon Divergence
We examine the problem of controlling divergences for latent space regularisation in variational autoencoders. Specifically, when aiming to reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$ ($n\leq m$), while balancing this against the need for generalisable latent repre...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Deasy, Jacob Simidjievski, Nikola Liò, Pietro |
description | We examine the problem of controlling divergences for latent space
regularisation in variational autoencoders. Specifically, when aiming to
reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$
($n\leq m$), while balancing this against the need for generalisable latent
representations. We present a regularisation mechanism based on the
skew-geometric Jensen-Shannon divergence
$\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in
$\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads
to an intuitive interpolation between forward and reverse KL in the space of
both distributions and divergences. We motivate its potential benefits for VAEs
through low-dimensional examples, before presenting quantitative and
qualitative results. Our experiments demonstrate that skewing our variant of
$\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of
$\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and
generation when compared to several baseline VAEs. Our approach is entirely
unsupervised and utilises only one hyperparameter which can be easily
interpreted in latent space. |
doi_str_mv | 10.48550/arxiv.2006.10599 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2006_10599</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2006_10599</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-81f92acea6a68f1906b1c1a2696a46e0001808e3ce5304596aea41f6e280c9b63</originalsourceid><addsrcrecordid>eNotz8FOwzAQBFBfOKDSD-CEfyBhncSLfUQpLYVKHKi4Rltr3VpqN5UTFfh7aOE00mg00lPq1kDZOGvhnvJXOpUVAJYGrPfX6rXtZRgzJUmy1R-UE42pF9rrpUTOLIH1Zxp3esH9gcecgn5hGViK9x2J9KJn6cR5ex7eqKtI-4Gn_zlR6_nTun0uVm-LZfu4KggffOFM9BUFJiR00XjAjQmGKvRIDTIAGAeO68C2hsb-tkyNiciVg-A3WE_U3d_tRdMdczpQ_u7Oqu6iqn8A7RNH4w</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</title><source>arXiv.org</source><creator>Deasy, Jacob ; Simidjievski, Nikola ; Liò, Pietro</creator><creatorcontrib>Deasy, Jacob ; Simidjievski, Nikola ; Liò, Pietro</creatorcontrib><description>We examine the problem of controlling divergences for latent space
regularisation in variational autoencoders. Specifically, when aiming to
reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$
($n\leq m$), while balancing this against the need for generalisable latent
representations. We present a regularisation mechanism based on the
skew-geometric Jensen-Shannon divergence
$\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in
$\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads
to an intuitive interpolation between forward and reverse KL in the space of
both distributions and divergences. We motivate its potential benefits for VAEs
through low-dimensional examples, before presenting quantitative and
qualitative results. Our experiments demonstrate that skewing our variant of
$\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of
$\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and
generation when compared to several baseline VAEs. Our approach is entirely
unsupervised and utilises only one hyperparameter which can be easily
interpreted in latent space.</description><identifier>DOI: 10.48550/arxiv.2006.10599</identifier><language>eng</language><subject>Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2020-06</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2006.10599$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2006.10599$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Deasy, Jacob</creatorcontrib><creatorcontrib>Simidjievski, Nikola</creatorcontrib><creatorcontrib>Liò, Pietro</creatorcontrib><title>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</title><description>We examine the problem of controlling divergences for latent space
regularisation in variational autoencoders. Specifically, when aiming to
reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$
($n\leq m$), while balancing this against the need for generalisable latent
representations. We present a regularisation mechanism based on the
skew-geometric Jensen-Shannon divergence
$\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in
$\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads
to an intuitive interpolation between forward and reverse KL in the space of
both distributions and divergences. We motivate its potential benefits for VAEs
through low-dimensional examples, before presenting quantitative and
qualitative results. Our experiments demonstrate that skewing our variant of
$\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of
$\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and
generation when compared to several baseline VAEs. Our approach is entirely
unsupervised and utilises only one hyperparameter which can be easily
interpreted in latent space.</description><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz8FOwzAQBFBfOKDSD-CEfyBhncSLfUQpLYVKHKi4Rltr3VpqN5UTFfh7aOE00mg00lPq1kDZOGvhnvJXOpUVAJYGrPfX6rXtZRgzJUmy1R-UE42pF9rrpUTOLIH1Zxp3esH9gcecgn5hGViK9x2J9KJn6cR5ex7eqKtI-4Gn_zlR6_nTun0uVm-LZfu4KggffOFM9BUFJiR00XjAjQmGKvRIDTIAGAeO68C2hsb-tkyNiciVg-A3WE_U3d_tRdMdczpQ_u7Oqu6iqn8A7RNH4w</recordid><startdate>20200618</startdate><enddate>20200618</enddate><creator>Deasy, Jacob</creator><creator>Simidjievski, Nikola</creator><creator>Liò, Pietro</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20200618</creationdate><title>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</title><author>Deasy, Jacob ; Simidjievski, Nikola ; Liò, Pietro</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-81f92acea6a68f1906b1c1a2696a46e0001808e3ce5304596aea41f6e280c9b63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Deasy, Jacob</creatorcontrib><creatorcontrib>Simidjievski, Nikola</creatorcontrib><creatorcontrib>Liò, Pietro</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Deasy, Jacob</au><au>Simidjievski, Nikola</au><au>Liò, Pietro</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Constraining Variational Inference with Geometric Jensen-Shannon Divergence</atitle><date>2020-06-18</date><risdate>2020</risdate><abstract>We examine the problem of controlling divergences for latent space
regularisation in variational autoencoders. Specifically, when aiming to
reconstruct example $x\in\mathbb{R}^{m}$ via latent space $z\in\mathbb{R}^{n}$
($n\leq m$), while balancing this against the need for generalisable latent
representations. We present a regularisation mechanism based on the
skew-geometric Jensen-Shannon divergence
$\left(\textrm{JS}^{\textrm{G}_{\alpha}}\right)$. We find a variation in
$\textrm{JS}^{\textrm{G}_{\alpha}}$, motivated by limiting cases, which leads
to an intuitive interpolation between forward and reverse KL in the space of
both distributions and divergences. We motivate its potential benefits for VAEs
through low-dimensional examples, before presenting quantitative and
qualitative results. Our experiments demonstrate that skewing our variant of
$\textrm{JS}^{\textrm{G}_{\alpha}}$, in the context of
$\textrm{JS}^{\textrm{G}_{\alpha}}$-VAEs, leads to better reconstruction and
generation when compared to several baseline VAEs. Our approach is entirely
unsupervised and utilises only one hyperparameter which can be easily
interpreted in latent space.</abstract><doi>10.48550/arxiv.2006.10599</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2006.10599 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2006_10599 |
source | arXiv.org |
subjects | Computer Science - Learning Statistics - Machine Learning |
title | Constraining Variational Inference with Geometric Jensen-Shannon Divergence |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T06%3A38%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Constraining%20Variational%20Inference%20with%20Geometric%20Jensen-Shannon%20Divergence&rft.au=Deasy,%20Jacob&rft.date=2020-06-18&rft_id=info:doi/10.48550/arxiv.2006.10599&rft_dat=%3Carxiv_GOX%3E2006_10599%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |