Zero-shot unsupervised image-to-image translation via exploiting semantic attributes

被引:2
作者
Chen, Yuanqi [1 ,2 ]
Yu, Xiaoming [1 ,2 ]
Liu, Shan [3 ]
Gao, Wei [1 ,2 ]
Li, Ge [1 ]
机构
[1] Peking Univ, Sch Elect & Comp Engn, Shenzhen Grad Sch, Shenzhen 518055, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Tencent Inc, Shenzhen 518000, Peoples R China
基金
国家重点研发计划;
关键词
Image -to-image translation; Image synthesis; Zero-shot learning; Generative adversarial networks; GENERATIVE ADVERSARIAL NETWORKS; GAN; CLASSIFICATION;
D O I
10.1016/j.imavis.2022.104489
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent studies have shown remarkable success in unsupervised image-to-image translation. However, if there is no access to enough images in target classes, learning a mapping from source classes to the target classes always suffers from mode collapse, especially the zero shot case, which limits the application of the existing methods. In this work, we propose a zero-shot unsupervised image-to-image translation framework to address this limita-tion, by effectively associating categories with their side information like attributes. To generalize the translator to previously unseen classes, we introduce two strategies for exploiting the semantic attribute space. First, we propose to preserve semantic relations to the visual space for effective guidance on where to map the input image. Second, expanding attribute space is introduced by utilizing attribute vectors of unseen classes, which al-leviates the mapping bias for unseen classes. Both of these strategies encourage the translator to explore the modes of unseen classes. Quantitative and qualitative results on different datasets validate the effectiveness of our proposed approach. Moreover, we demonstrate that our framework can be applied to fashion design task. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 56 条
[1]   MFC-GAN: Class-imbalanced dataset classification using Multiple Fake Class Generative Adversarial Network [J].
Ali-Gombe, Adamu ;
Elyan, Eyad .
NEUROCOMPUTING, 2019, 361 :212-221
[2]   Preserving Semantic Relations for Zero-Shot Learning [J].
Annadani, Yashas ;
Biswas, Soma .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7603-7612
[3]  
Benaim L., P 31 INT C NEUR INF, P752
[4]  
Benaim S, 2018, ADV NEUR IN, V31
[5]   An Empirical Study and Analysis of Generalized Zero-Shot Learning for Object Recognition in the Wild [J].
Chao, Wei-Lun ;
Changpinyo, Soravit ;
Gong, Boqing ;
Sha, Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :52-68
[6]  
Chen Xi, 2016, ADV NEURAL INFORM PR, V29
[7]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[8]  
Frome A., 2013, Advances in neural information processing systems, P2121
[9]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[10]  
Guillaume L., 2017, ADV NEURAL INFORM PR, P5967