DualGAN: Unsupervised Dual Learning for Image-to-Image Translation

被引:1514
作者
Yi, Zili [1 ,2 ]
Zhang, Hao [2 ]
Tan, Ping [2 ]
Gong, Minglun [1 ]
机构
[1] Mem Univ Newfoundland, St John, NF, Canada
[2] Simon Fraser Univ, Burnaby, BC, Canada
来源
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) | 2017年
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
10.1109/ICCV.2017.310
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Conditional Generative Adversarial Networks( GANs) for cross-domain image-to-image translation have made much progress recently [7, 8, 21, 12, 4, 18]. Depending on the task complexity, thousands to millions of labeled image pairs are needed to train a conditional GAN. However, human labeling is expensive, even impractical, and large quantities of data may not always be available. Inspired by dual learning from natural language translation [23], we develop a novel dual-GAN mechanism, which enables image translators to be trained from two sets of unlabeled images from two domains. In our architecture, the primal GAN learns to translate images from domain U to those in domain V, while the dual GAN learns to invert the task. The closed loop made by the primal and dual tasks allows images from either domain to be translated and then reconstructed. Hence a loss function that accounts for the reconstruction error of images can be used to train the translators. Experiments on multiple image translation tasks with unlabeled data show considerable performance gain of DualGAN over a single GAN. For some tasks, DualGAN can even achieve comparable or slightly better results than conditional GAN trained on fully labeled data.
引用
收藏
页码:2868 / 2876
页数:9
相关论文
共 26 条
[1]  
[Anonymous], 2016, P NIPS WORKSHOP ADVE
[2]  
Arjovsky M., 2017, ARXIV170107875
[3]  
Aytar Y, 2016, ADV NEUR IN, V29
[4]   Generative Adversarial Networks [J].
Goodfellow, Ian ;
Pouget-Abadie, Jean ;
Mirza, Mehdi ;
Xu, Bing ;
Warde-Farley, David ;
Ozair, Sherjil ;
Courville, Aaron ;
Bengio, Yoshua .
COMMUNICATIONS OF THE ACM, 2020, 63 (11) :139-144
[5]  
Isola P., 2017, P IEEE C COMPUTER VI, P1125, DOI [DOI 10.1109/CVPR.2017.632, 10.1109/CVPR.2017.632]
[6]   Transient Attributes for High-Level Understanding and Editing of Outdoor Scenes [J].
Laffont, Pierre-Yves ;
Ren, Zhile ;
Tao, Xiaofeng ;
Qian, Chao ;
Hays, James .
ACM TRANSACTIONS ON GRAPHICS, 2014, 33 (04)
[7]  
Larsen S. K., 2016, INT C MACHINE LEARNI, P1558
[8]   Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network [J].
Ledig, Christian ;
Theis, Lucas ;
Huszar, Ferenc ;
Caballero, Jose ;
Cunningham, Andrew ;
Acosta, Alejandro ;
Aitken, Andrew ;
Tejani, Alykhan ;
Totz, Johannes ;
Wang, Zehan ;
Shi, Wenzhe .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :105-114
[9]  
Li C, 2016, MATH PROBL ENG, V2016, P1
[10]  
Liu M.-.Y., 2016, ADV NEURAL INFORM PR, V29, P469, DOI DOI 10.5555/3157096.3157149