Triple Generative Adversarial Networks

被引:29
作者
Li, Chongxuan [1 ,2 ]
Xu, Kun [1 ]
Zhu, Jun [1 ]
Liu, Jiashuo [1 ]
Zhang, Bo [1 ]
机构
[1] Tsinghua Univ, Ctr Bioinspired Comp Res, TNList Lab, Dept Comp Sci & Technol,Inst AI, Beijing 100084, Peoples R China
[2] Renmin Univ China, Gaoling Sch AI, Beijing 100872, Peoples R China
关键词
Generators; Generative adversarial networks; Task analysis; Semisupervised learning; Games; Entropy; Linear programming; Generative adversarial network; deep generative model; semi-supervised learning; extremely low data regime; conditional image generation;
D O I
10.1109/TPAMI.2021.3127558
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a unified game-theoretical framework to perform classification and conditional image generation given limited supervision. It is formulated as a three-player minimax game consisting of a generator, a classifier and a discriminator, and therefore is referred to as Triple Generative Adversarial Network (Triple-GAN). The generator and the classifier characterize the conditional distributions between images and labels to perform conditional generation and classification, respectively. The discriminator solely focuses on identifying fake image-label pairs. Theoretically, the three-player formulation guarantees consistency. Namely, under a nonparametric assumption, the unique equilibrium of the game is that the distributions characterized by the generator and the classifier converge to the data distribution. As a byproduct of the three-player formulation, Triple-GAN is flexible to incorporate different semi-supervised classifiers and GAN architectures. We evaluate Triple-GAN in two challenging settings, namely, semi-supervised learning and the extremely low data regime. In both settings, Triple-GAN can achieve excellent classification results and generate meaningful samples in a specific class simultaneously. In particular, using a commonly adopted 13-layer CNN classifier, Triple-GAN outperforms extensive semi-supervised learning methods substantially on several benchmarks no matter data augmentation is applied or not.
引用
收藏
页码:9629 / 9640
页数:12
相关论文
共 70 条
  • [1] [Anonymous], 2017, P ADV NEUR INF PROC
  • [2] [Anonymous], 2017, PROC INT C LEARN REP
  • [3] [Anonymous], 2018, P ICLR
  • [4] Arjovsky M, 2017, PR MACH LEARN RES, V70
  • [5] Athiwaratkun B, 2019, Arxiv, DOI arXiv:1806.05594
  • [6] Berthelot D, 2019, PROC INT C LEARN REP
  • [7] Berthelot D, 2019, ADV NEUR IN, V32
  • [8] Brock A., 2018, PROC INT C LEARN REP
  • [9] Chen X, 2016, ADV NEUR IN, V29
  • [10] Coates A., 2011, JMLR WORKSHOP C P, P215