ComboGAN: Unrestrained Scalability for Image Domain Translation

被引:124
作者
Anoosheh, Asha [1 ]
Agustsson, Eirikur [1 ]
Timofte, Radu [2 ,3 ]
Van Gool, Luc [2 ,4 ]
机构
[1] Swiss Fed Inst Technol, D ITET, Comp Vis Lab, Zurich, Switzerland
[2] Swiss Fed Inst Technol, D ITET, Zurich, Switzerland
[3] Merantix, Berlin, Germany
[4] Katholieke Univ Leuven, ESAT, Leuven, Belgium
来源
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW) | 2018年
关键词
D O I
10.1109/CVPRW.2018.00122
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The past year alone has seen unprecedented leaps in the area of learning-based image translation, namely CycleGAN, by Zhu et al. But experiments so far have been tailored to merely two domains at a time, and scaling them to more would require an quadratic number of models to be trained. And with two-domain models taking days to train on current hardware, the number of domains quickly becomes limited by the time and resources required to process them. In this paper, we propose a multi-component image translation model and training scheme which scales linearly - both in resource consumption and time required - with the number of domains. We demonstrate its capabilities on a dataset of paintings by 14 different artists and on images of the four different seasons in the Alps. Note that 14 data groups would need ( 14 choose 2) = 91 different CycleGAN models: a total of 182 generator/discriminator pairs; whereas our model requires only 14 generator/discriminator pairs.
引用
收藏
页码:896 / 903
页数:8
相关论文
共 12 条
[1]  
[Anonymous], ARXIV E PRINTS
[2]  
[Anonymous], ARXIV E PRINTS
[3]  
[Anonymous], 2016, ARXIV E PRINTS
[4]  
[Anonymous], 2017, CORR
[5]  
[Anonymous], 2016, ARXIV E PRINTS
[6]  
[Anonymous], 2017, CORR
[7]  
[Anonymous], ARXIV E PRINTS
[8]  
[Anonymous], ARXIV E PRINTS
[9]  
[Anonymous], CORR
[10]  
Gatys L. A, 2015, ARXIV E PRINTS