Stabilized GAN models training with kernel-histogram transformation and probability mass function distance
Image generation using generative adversarial networks (GANs) has been extensively researched in recent years. Despite active developments, the chronic issue of training instability in GANs remains unresolved. To alleviate this problem, this study proposes a model named probability mass function GAN...
Gespeichert in:
Veröffentlicht in: | Applied soft computing 2024-10, Vol.164, p.112003, Article 112003 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Image generation using generative adversarial networks (GANs) has been extensively researched in recent years. Despite active developments, the chronic issue of training instability in GANs remains unresolved. To alleviate this problem, this study proposes a model named probability mass function GANs (PMF-GAN), which handles the inherent limitation of GANs. The PMF-GAN framework employs kernels, histogram transformation, and probability mass function (PMF) distance for distribution learning. The configuration of PMF-GAN kernel and PMF distance offers flexibility, allowing for optimal settings tailored to datasets and experimental environments. In this study, experiments were conducted using the gaussian kernel across five different distances. The experiments demonstrated that PMF-GAN outperforms the baselines in terms of visual quality and evaluation metrics, such as Inception score and Frechet Inception distance (FID). For example, in the CIFAR-10 dataset, Euclidean-based PMF-GAN applying with 3 bins showed a 21.5 % and 32.8 % improvement in Inception score and FID, respectively, compared to conventional WGAN-GP. Similarly, in the AFHQ dataset with the same settings, the improvements were 56.9 % and 61.5 %. As a result, this study presents the potential to achieve stable training processes in GAN models with modified loss function structures. The flexibility of the proposed model allows for simultaneous application to various models, contributing to the overall improvement of generative model training processes in the future.
•Introduction of a novel GAN model using diverse kernels and distances for distribution comparison.•A new histogram transformation method in the discriminator improves distribution differentiation.•Evaluations on MNIST, CIFAR, CelebA, LSUN, and AFHQ show PMF-GANs superior image generation.•PMF-GANs integrate with latest GAN architectures, offering flexibility for diverse applications. |
---|---|
ISSN: | 1568-4946 |
DOI: | 10.1016/j.asoc.2024.112003 |