Annealing Genetic GAN for Imbalanced Web Data Learning

被引:18
作者
Hao, Jingyu [1 ]
Wang, Chengjia [2 ]
Yang, Guang [3 ]
Gao, Zhifan [1 ]
Zhang, Jinglin [4 ]
Zhang, Heye [1 ]
机构
[1] Sun Yat Sen Univ, Sch Biomed Engn, Shenzhen 518107, Peoples R China
[2] Univ Edinburgh, Ctr Cardiovasc Sci, Edinburgh EH16 4TJ, Midlothian, Scotland
[3] Imperial Coll London, Natl Heart & Lung Inst, Fac Med, London SW7 2AZ, England
[4] Hebei Univ Technol, Sch Artificial Intelligence, Tianjin 300401, Peoples R China
基金
欧洲研究理事会; 欧盟地平线“2020”;
关键词
Training; Generators; Genetic algorithms; Annealing; Simulated annealing; Generative adversarial networks; Optimization; Class imbalance problem; evolutionary computation; data augmentation; CLASSIFICATION; ALGORITHMS; SMOTE;
D O I
10.1109/TMM.2021.3120642
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Class imbalance is one of the most basic and important problems of web data. The key to overcoming the class imbalance problems is to increase the effective instances of the minority, that is, data augmentation. Generative Adversarial Networks (GANs), which have recently been successfully applied in the field of image generation, can be used for data augmentation because they can learn the data distribution given ample training data instances and generate more data. However, learning the distributions from the imbalanced data can make GANs easily get stuck in a local optimum. In this work, we propose a new training strategy called Annealing Genetic GAN (AGGAN), which incorporates simulated annealing genetic algorithm into the training process of GANs. And this can help GANs avoid the local optimum trapping problem, which easily occurs when the training set is imbalanced. Unlike existing GANs, which use a fixed adversarial learning objective alternately training a generator, we use multiple adversarial learning objectives to train a set of generators and use the Metropolis criterion in simulated annealing to decide whether the generator should update. More specifically, the Metropolis criterion accepts worse solutions with a certain probability, so it can make our AGGAN escape from the local optimum and find a better solution. Theory and mathematical analysis provide strong theoretical support for the proposed training strategy. And experiments on several datasets demonstrate that AGGAN achieves convincing ability to solve the class imbalanced problem and reduces the training problems inherent in existing GANs.
引用
收藏
页码:1164 / 1174
页数:11
相关论文
共 46 条
[41]  
Rotemberg V, 2021, SCI DATA, V8, DOI 10.1038/s41597-021-00815-z
[42]   Cost-sensitive boosting for classification of imbalanced data [J].
Sun, Yamnin ;
Kamel, Mohamed S. ;
Wong, Andrew K. C. ;
Wang, Yang .
PATTERN RECOGNITION, 2007, 40 (12) :3358-3378
[43]  
Tao Chenyang, 2019, INT C MACHINE LEARNI, P6176
[44]   Evolutionary Generative Adversarial Networks [J].
Wang, Chaoyue ;
Xu, Chang ;
Yao, Xin ;
Tao, Dacheng .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2019, 23 (06) :921-934
[45]  
Yu Fisher, 2015, Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop
[46]   StackGAN: Text to Photo-realistic Image Synthesis with Stacked Generative Adversarial Networks [J].
Zhang, Han ;
Xu, Tao ;
Li, Hongsheng ;
Zhang, Shaoting ;
Wang, Xiaogang ;
Huang, Xiaolei ;
Metaxas, Dimitris .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5908-5916