DivCo: Diverse Conditional Image Synthesis via Contrastive Generative Adversarial Network

被引:73
作者
Liu, Rui [1 ]
Ge, Yixiao [1 ]
Choi, Ching Lam [1 ,2 ]
Wang, Xiaogang [1 ]
Li, Hongsheng [1 ,3 ]
机构
[1] Chinese Univ Hong Kong, CUHK SenseTime Joint Lab, Hong Kong, Peoples R China
[2] NVIDIA, NVIDIA AI Technol Ctr, Hong Kong, Peoples R China
[3] Xidian Univ, Sch CST, Xian, Peoples R China
来源
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021 | 2021年
关键词
D O I
10.1109/CVPR46437.2021.01611
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Conditional generative adversarial networks (cGANs) target at synthesizing diverse images given the input conditions and latent codes, but unfortunately, they usually suffer from the issue of mode collapse. To solve this issue, previous works [47, 22] mainly focused on encouraging the correlation between the latent codes and their generated images, while ignoring the relations between images generated from various latent codes. The recent MSGAN [27] tried to encourage the diversity of the generated image but only considers "negative" relations between the image pairs. In this paper, we propose a novel DivCo framework to properly constrain both "positive" and "negative" relations between the generated images specified in the latent space. To the best of our knowledge, this is the first attempt to use contrastive learning for diverse conditional image synthesis. A novel latent-augmented contrastive loss is introduced, which encourages images generated from adjacent latent codes to be similar and those generated from distinct latent codes to be dissimilar. The proposed latent-augmented contrastive loss is well compatible with various cGAN architectures. Extensive experiments demonstrate that the proposed DivCo can produce more diverse images than state-of-the-art methods without sacrificing visual quality in multiple unpaired and paired image generation tasks. Training code and pretrained models are available at https://github.com/ruiliu-ai/DivCo.
引用
收藏
页码:16372 / 16381
页数:10
相关论文
共 47 条
[1]  
[Anonymous], 2019, ICML
[2]  
[Anonymous], 2018, ICML
[3]  
[Anonymous], 2017, NEURIPS
[4]  
[Anonymous], 2017, ICML
[5]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[6]  
Baek K., 2020, ARXIV200606500
[7]  
Brock A., 2019, INT C LEARNING REPRE
[8]  
Chen T, 2020, PR MACH LEARN RES, V119
[9]  
Chen Xi, 2016, Advances in Neural Information Processing Systems, V29
[10]  
ChoHaejin, 2018, Journal of Species Research, V7, P1, DOI 10.12651/JSR.2018.7.1.001