Cumulant GAN
In this paper, we propose a novel loss function for training Generative Adversarial Networks (GANs) aiming towards deeper theoretical understanding as well as improved stability and performance for the underlying optimization problem. The new loss function is based on cumulant generating functions g...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Pantazis, Yannis Paul, Dipjyoti Fasoulakis, Michail Stylianou, Yannis Katsoulakis, Markos |
description | In this paper, we propose a novel loss function for training Generative
Adversarial Networks (GANs) aiming towards deeper theoretical understanding as
well as improved stability and performance for the underlying optimization
problem. The new loss function is based on cumulant generating functions giving
rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula,
we show that the corresponding optimization problem is equivalent to R{\'e}nyi
divergence minimization, thus offering a (partially) unified perspective of GAN
losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD),
reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is
also a member of cumulant GAN. In terms of stability, we rigorously prove the
linear convergence of cumulant GAN to the Nash equilibrium for a linear
discriminator, Gaussian distributions and the standard gradient descent ascent
algorithm. Finally, we experimentally demonstrate that image generation is more
robust relative to Wasserstein GAN and it is substantially improved in terms of
both inception score and Fr\'echet inception distance when both weaker and
stronger discriminators are considered. |
doi_str_mv | 10.48550/arxiv.2006.06625 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2006_06625</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2006_06625</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-4268ee83e21a59611a76cd11511e0855600c577231fcfc3491a1e457a2a7ce6c3</originalsourceid><addsrcrecordid>eNotzrEOgjAUheEuDgZ8ACd9AbC37b2FkRBFE6ILO7mpJSEBY1CMvr2KTmf7zyfEEmRsEkS54eHZPmIlJcWSSOFchPnYjx1f7usiO4Zi1nB384v_BqLabat8H5Wn4pBnZcRkMTKKEu8T7RUwpgTAltwZAAG8_NyQlA6tVRoa1zhtUmDwBi0rts6T04FY_bITp74Obc_Dq_6y6oml3-D4Lvw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Cumulant GAN</title><source>arXiv.org</source><creator>Pantazis, Yannis ; Paul, Dipjyoti ; Fasoulakis, Michail ; Stylianou, Yannis ; Katsoulakis, Markos</creator><creatorcontrib>Pantazis, Yannis ; Paul, Dipjyoti ; Fasoulakis, Michail ; Stylianou, Yannis ; Katsoulakis, Markos</creatorcontrib><description>In this paper, we propose a novel loss function for training Generative
Adversarial Networks (GANs) aiming towards deeper theoretical understanding as
well as improved stability and performance for the underlying optimization
problem. The new loss function is based on cumulant generating functions giving
rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula,
we show that the corresponding optimization problem is equivalent to R{\'e}nyi
divergence minimization, thus offering a (partially) unified perspective of GAN
losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD),
reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is
also a member of cumulant GAN. In terms of stability, we rigorously prove the
linear convergence of cumulant GAN to the Nash equilibrium for a linear
discriminator, Gaussian distributions and the standard gradient descent ascent
algorithm. Finally, we experimentally demonstrate that image generation is more
robust relative to Wasserstein GAN and it is substantially improved in terms of
both inception score and Fr\'echet inception distance when both weaker and
stronger discriminators are considered.</description><identifier>DOI: 10.48550/arxiv.2006.06625</identifier><language>eng</language><subject>Computer Science - Information Theory ; Computer Science - Learning ; Mathematics - Information Theory ; Statistics - Machine Learning</subject><creationdate>2020-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2006.06625$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2006.06625$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Pantazis, Yannis</creatorcontrib><creatorcontrib>Paul, Dipjyoti</creatorcontrib><creatorcontrib>Fasoulakis, Michail</creatorcontrib><creatorcontrib>Stylianou, Yannis</creatorcontrib><creatorcontrib>Katsoulakis, Markos</creatorcontrib><title>Cumulant GAN</title><description>In this paper, we propose a novel loss function for training Generative
Adversarial Networks (GANs) aiming towards deeper theoretical understanding as
well as improved stability and performance for the underlying optimization
problem. The new loss function is based on cumulant generating functions giving
rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula,
we show that the corresponding optimization problem is equivalent to R{\'e}nyi
divergence minimization, thus offering a (partially) unified perspective of GAN
losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD),
reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is
also a member of cumulant GAN. In terms of stability, we rigorously prove the
linear convergence of cumulant GAN to the Nash equilibrium for a linear
discriminator, Gaussian distributions and the standard gradient descent ascent
algorithm. Finally, we experimentally demonstrate that image generation is more
robust relative to Wasserstein GAN and it is substantially improved in terms of
both inception score and Fr\'echet inception distance when both weaker and
stronger discriminators are considered.</description><subject>Computer Science - Information Theory</subject><subject>Computer Science - Learning</subject><subject>Mathematics - Information Theory</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzrEOgjAUheEuDgZ8ACd9AbC37b2FkRBFE6ILO7mpJSEBY1CMvr2KTmf7zyfEEmRsEkS54eHZPmIlJcWSSOFchPnYjx1f7usiO4Zi1nB384v_BqLabat8H5Wn4pBnZcRkMTKKEu8T7RUwpgTAltwZAAG8_NyQlA6tVRoa1zhtUmDwBi0rts6T04FY_bITp74Obc_Dq_6y6oml3-D4Lvw</recordid><startdate>20200611</startdate><enddate>20200611</enddate><creator>Pantazis, Yannis</creator><creator>Paul, Dipjyoti</creator><creator>Fasoulakis, Michail</creator><creator>Stylianou, Yannis</creator><creator>Katsoulakis, Markos</creator><scope>AKY</scope><scope>AKZ</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20200611</creationdate><title>Cumulant GAN</title><author>Pantazis, Yannis ; Paul, Dipjyoti ; Fasoulakis, Michail ; Stylianou, Yannis ; Katsoulakis, Markos</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-4268ee83e21a59611a76cd11511e0855600c577231fcfc3491a1e457a2a7ce6c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Information Theory</topic><topic>Computer Science - Learning</topic><topic>Mathematics - Information Theory</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Pantazis, Yannis</creatorcontrib><creatorcontrib>Paul, Dipjyoti</creatorcontrib><creatorcontrib>Fasoulakis, Michail</creatorcontrib><creatorcontrib>Stylianou, Yannis</creatorcontrib><creatorcontrib>Katsoulakis, Markos</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Mathematics</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Pantazis, Yannis</au><au>Paul, Dipjyoti</au><au>Fasoulakis, Michail</au><au>Stylianou, Yannis</au><au>Katsoulakis, Markos</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Cumulant GAN</atitle><date>2020-06-11</date><risdate>2020</risdate><abstract>In this paper, we propose a novel loss function for training Generative
Adversarial Networks (GANs) aiming towards deeper theoretical understanding as
well as improved stability and performance for the underlying optimization
problem. The new loss function is based on cumulant generating functions giving
rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula,
we show that the corresponding optimization problem is equivalent to R{\'e}nyi
divergence minimization, thus offering a (partially) unified perspective of GAN
losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD),
reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is
also a member of cumulant GAN. In terms of stability, we rigorously prove the
linear convergence of cumulant GAN to the Nash equilibrium for a linear
discriminator, Gaussian distributions and the standard gradient descent ascent
algorithm. Finally, we experimentally demonstrate that image generation is more
robust relative to Wasserstein GAN and it is substantially improved in terms of
both inception score and Fr\'echet inception distance when both weaker and
stronger discriminators are considered.</abstract><doi>10.48550/arxiv.2006.06625</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2006.06625 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2006_06625 |
source | arXiv.org |
subjects | Computer Science - Information Theory Computer Science - Learning Mathematics - Information Theory Statistics - Machine Learning |
title | Cumulant GAN |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T04%3A08%3A35IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Cumulant%20GAN&rft.au=Pantazis,%20Yannis&rft.date=2020-06-11&rft_id=info:doi/10.48550/arxiv.2006.06625&rft_dat=%3Carxiv_GOX%3E2006_06625%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |