Abstract
Image generation using generative adversarial networks (GANs) has been extensively researched in recent years. Despite active developments, the chronic issue of training instability in GANs remains unresolved. To alleviate this problem, this study proposes a model named probability mass function GANs (PMF-GAN), which handles the inherent limitation of GANs. The PMF-GAN framework employs kernels, histogram transformation, and probability mass function (PMF) distance for distribution learning. The configuration of PMF-GAN kernel and PMF distance offers flexibility, allowing for optimal settings tailored to datasets and experimental environments. In this study, experiments were conducted using the gaussian kernel across five different distances. The experiments demonstrated that PMF-GAN outperforms the baselines in terms of visual quality and evaluation metrics, such as Inception score and Frechet Inception distance (FID). For example, in the CIFAR-10 dataset, Euclidean-based PMF-GAN applying with 3 bins showed a 21.5 % and 32.8 % improvement in Inception score and FID, respectively, compared to conventional WGAN-GP. Similarly, in the AFHQ dataset with the same settings, the improvements were 56.9 % and 61.5 %. As a result, this study presents the potential to achieve stable training processes in GAN models with modified loss function structures. The flexibility of the proposed model allows for simultaneous application to various models, contributing to the overall improvement of generative model training processes in the future.
Original language | English |
---|---|
Article number | 112003 |
Journal | Applied Soft Computing |
Volume | 164 |
DOIs | |
Publication status | Published - 2024 Oct |
Bibliographical note
Publisher Copyright:© 2024 Elsevier B.V.
Keywords
- Deep learning
- Generative adversarial network
- Histogram transformation
- Kernel density estimation
ASJC Scopus subject areas
- Software