Back to Search Start Over

Statistics Enhancement Generative Adversarial Networks for Diverse Conditional Image Synthesis

Authors :
Zuo, Zhiwen
Li, Ailin
Wang, Zhizhong
Zhao, Lei
Dong, Jianfeng
Wang, Xun
Wang, Meng
Source :
IEEE Transactions on Circuits and Systems for Video Technology; 2024, Vol. 34 Issue: 7 p6167-6180, 14p
Publication Year :
2024

Abstract

Conditional generative adversarial networks (cGANs) aim to synthesize diverse images given the input conditions and the latent codes, but they are prone to map an input to a single output regardless of the variations in latent code, which is also well known as the mode collapse problem of cGANs. To alleviate the problem, in this paper, we investigate explicitly enhancing the statistical dependency between the latent code and the synthesized image in cGANs by utilizing mutual information neural estimators to estimate and maximize the conditional mutual information (CMI) between them given the input condition. The method provides a new perspective from information theory to improve diversity for cGANs and can facilitate many existing conditional image synthesis frameworks with a simple neural estimator extension. Moreover, our studies show that several key designs, including the neural estimator choice, the neural estimator’s network design, and the sampling strategy, are crucial to the success of the method. Extensive experiments on four popular conditional image synthesis tasks, including class-conditioned image generation, paired and unpaired image-to-image translation, and text-to-image generation, demonstrate the effectiveness and superiority of the proposed method.

Details

Language :
English
ISSN :
10518215 and 15582205
Volume :
34
Issue :
7
Database :
Supplemental Index
Journal :
IEEE Transactions on Circuits and Systems for Video Technology
Publication Type :
Periodical
Accession number :
ejs66895136
Full Text :
https://doi.org/10.1109/TCSVT.2023.3348471