Cumulant GAN

In this paper, we propose a novel loss function for training Generative Adversarial Networks (GANs) aiming towards deeper theoretical understanding as well as improved stability and performance for the underlying optimization problem. The new loss function is based on cumulant generating functions g...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Pantazis, Yannis, Paul, Dipjyoti, Fasoulakis, Michail, Stylianou, Yannis, Katsoulakis, Markos
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Pantazis, Yannis
Paul, Dipjyoti
Fasoulakis, Michail
Stylianou, Yannis
Katsoulakis, Markos
description In this paper, we propose a novel loss function for training Generative Adversarial Networks (GANs) aiming towards deeper theoretical understanding as well as improved stability and performance for the underlying optimization problem. The new loss function is based on cumulant generating functions giving rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula, we show that the corresponding optimization problem is equivalent to R{\'e}nyi divergence minimization, thus offering a (partially) unified perspective of GAN losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD), reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is also a member of cumulant GAN. In terms of stability, we rigorously prove the linear convergence of cumulant GAN to the Nash equilibrium for a linear discriminator, Gaussian distributions and the standard gradient descent ascent algorithm. Finally, we experimentally demonstrate that image generation is more robust relative to Wasserstein GAN and it is substantially improved in terms of both inception score and Fr\'echet inception distance when both weaker and stronger discriminators are considered.
doi_str_mv 10.48550/arxiv.2006.06625
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2006_06625</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2006_06625</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-4268ee83e21a59611a76cd11511e0855600c577231fcfc3491a1e457a2a7ce6c3</originalsourceid><addsrcrecordid>eNotzrEOgjAUheEuDgZ8ACd9AbC37b2FkRBFE6ILO7mpJSEBY1CMvr2KTmf7zyfEEmRsEkS54eHZPmIlJcWSSOFchPnYjx1f7usiO4Zi1nB384v_BqLabat8H5Wn4pBnZcRkMTKKEu8T7RUwpgTAltwZAAG8_NyQlA6tVRoa1zhtUmDwBi0rts6T04FY_bITp74Obc_Dq_6y6oml3-D4Lvw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Cumulant GAN</title><source>arXiv.org</source><creator>Pantazis, Yannis ; Paul, Dipjyoti ; Fasoulakis, Michail ; Stylianou, Yannis ; Katsoulakis, Markos</creator><creatorcontrib>Pantazis, Yannis ; Paul, Dipjyoti ; Fasoulakis, Michail ; Stylianou, Yannis ; Katsoulakis, Markos</creatorcontrib><description>In this paper, we propose a novel loss function for training Generative Adversarial Networks (GANs) aiming towards deeper theoretical understanding as well as improved stability and performance for the underlying optimization problem. The new loss function is based on cumulant generating functions giving rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula, we show that the corresponding optimization problem is equivalent to R{\'e}nyi divergence minimization, thus offering a (partially) unified perspective of GAN losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD), reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is also a member of cumulant GAN. In terms of stability, we rigorously prove the linear convergence of cumulant GAN to the Nash equilibrium for a linear discriminator, Gaussian distributions and the standard gradient descent ascent algorithm. Finally, we experimentally demonstrate that image generation is more robust relative to Wasserstein GAN and it is substantially improved in terms of both inception score and Fr\'echet inception distance when both weaker and stronger discriminators are considered.</description><identifier>DOI: 10.48550/arxiv.2006.06625</identifier><language>eng</language><subject>Computer Science - Information Theory ; Computer Science - Learning ; Mathematics - Information Theory ; Statistics - Machine Learning</subject><creationdate>2020-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2006.06625$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2006.06625$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Pantazis, Yannis</creatorcontrib><creatorcontrib>Paul, Dipjyoti</creatorcontrib><creatorcontrib>Fasoulakis, Michail</creatorcontrib><creatorcontrib>Stylianou, Yannis</creatorcontrib><creatorcontrib>Katsoulakis, Markos</creatorcontrib><title>Cumulant GAN</title><description>In this paper, we propose a novel loss function for training Generative Adversarial Networks (GANs) aiming towards deeper theoretical understanding as well as improved stability and performance for the underlying optimization problem. The new loss function is based on cumulant generating functions giving rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula, we show that the corresponding optimization problem is equivalent to R{\'e}nyi divergence minimization, thus offering a (partially) unified perspective of GAN losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD), reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is also a member of cumulant GAN. In terms of stability, we rigorously prove the linear convergence of cumulant GAN to the Nash equilibrium for a linear discriminator, Gaussian distributions and the standard gradient descent ascent algorithm. Finally, we experimentally demonstrate that image generation is more robust relative to Wasserstein GAN and it is substantially improved in terms of both inception score and Fr\'echet inception distance when both weaker and stronger discriminators are considered.</description><subject>Computer Science - Information Theory</subject><subject>Computer Science - Learning</subject><subject>Mathematics - Information Theory</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzrEOgjAUheEuDgZ8ACd9AbC37b2FkRBFE6ILO7mpJSEBY1CMvr2KTmf7zyfEEmRsEkS54eHZPmIlJcWSSOFchPnYjx1f7usiO4Zi1nB384v_BqLabat8H5Wn4pBnZcRkMTKKEu8T7RUwpgTAltwZAAG8_NyQlA6tVRoa1zhtUmDwBi0rts6T04FY_bITp74Obc_Dq_6y6oml3-D4Lvw</recordid><startdate>20200611</startdate><enddate>20200611</enddate><creator>Pantazis, Yannis</creator><creator>Paul, Dipjyoti</creator><creator>Fasoulakis, Michail</creator><creator>Stylianou, Yannis</creator><creator>Katsoulakis, Markos</creator><scope>AKY</scope><scope>AKZ</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20200611</creationdate><title>Cumulant GAN</title><author>Pantazis, Yannis ; Paul, Dipjyoti ; Fasoulakis, Michail ; Stylianou, Yannis ; Katsoulakis, Markos</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-4268ee83e21a59611a76cd11511e0855600c577231fcfc3491a1e457a2a7ce6c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Information Theory</topic><topic>Computer Science - Learning</topic><topic>Mathematics - Information Theory</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Pantazis, Yannis</creatorcontrib><creatorcontrib>Paul, Dipjyoti</creatorcontrib><creatorcontrib>Fasoulakis, Michail</creatorcontrib><creatorcontrib>Stylianou, Yannis</creatorcontrib><creatorcontrib>Katsoulakis, Markos</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Mathematics</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Pantazis, Yannis</au><au>Paul, Dipjyoti</au><au>Fasoulakis, Michail</au><au>Stylianou, Yannis</au><au>Katsoulakis, Markos</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Cumulant GAN</atitle><date>2020-06-11</date><risdate>2020</risdate><abstract>In this paper, we propose a novel loss function for training Generative Adversarial Networks (GANs) aiming towards deeper theoretical understanding as well as improved stability and performance for the underlying optimization problem. The new loss function is based on cumulant generating functions giving rise to \emph{Cumulant GAN}. Relying on a recently-derived variational formula, we show that the corresponding optimization problem is equivalent to R{\'e}nyi divergence minimization, thus offering a (partially) unified perspective of GAN losses: the R{\'e}nyi family encompasses Kullback-Leibler divergence (KLD), reverse KLD, Hellinger distance and $\chi^2$-divergence. Wasserstein GAN is also a member of cumulant GAN. In terms of stability, we rigorously prove the linear convergence of cumulant GAN to the Nash equilibrium for a linear discriminator, Gaussian distributions and the standard gradient descent ascent algorithm. Finally, we experimentally demonstrate that image generation is more robust relative to Wasserstein GAN and it is substantially improved in terms of both inception score and Fr\'echet inception distance when both weaker and stronger discriminators are considered.</abstract><doi>10.48550/arxiv.2006.06625</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2006.06625
ispartof
issn
language eng
recordid cdi_arxiv_primary_2006_06625
source arXiv.org
subjects Computer Science - Information Theory
Computer Science - Learning
Mathematics - Information Theory
Statistics - Machine Learning
title Cumulant GAN
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T04%3A08%3A35IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Cumulant%20GAN&rft.au=Pantazis,%20Yannis&rft.date=2020-06-11&rft_id=info:doi/10.48550/arxiv.2006.06625&rft_dat=%3Carxiv_GOX%3E2006_06625%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true