Evolving Deep Convolutional Variational Autoencoders for Image Classification

被引:32
作者
Chen, Xiangru [1 ]
Sun, Yanan [1 ,2 ]
Zhang, Mengjie [3 ]
Peng, Dezhong [4 ,5 ,6 ]
机构
[1] Sichuan Univ, Coll Comp Sci, Chengdu 610065, Peoples R China
[2] Sichuan Eface Technol Co Ltd, Chengdu 610093, Peoples R China
[3] Victoria Univ Wellington, Sch Engn & Comp Sci, Wellington 6140, New Zealand
[4] Sichuan Univ, Coll Comp Sci, Machine Intelligence Lab, Chengdu 610065, Peoples R China
[5] Shenzhen Peng Cheng Lab, Shenzhen 518052, Peoples R China
[6] Chengdu Sobey Digital Technol Co Ltd, Chengdu 610041, Peoples R China
基金
中国国家自然科学基金;
关键词
Computer architecture; Encoding; Task analysis; Approximation algorithms; Training; Genetic algorithms; Computer science; Convolutional variational autoencoder; evolving deep learning; genetic algorithm (GA); neural architecture search (NAS); NEURAL-NETWORKS; GENETIC ALGORITHM; REPRESENTATIONS;
D O I
10.1109/TEVC.2020.3047220
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have demonstrated their superiority in unsupervised learning for image processing in recent years. The performance of the VAEs highly depends on their architectures, which are often handcrafted by the human expertise in deep neural networks (DNNs). However, such expertise is not necessarily available to each of the end users interested. In this article, we propose a novel method to automatically design optimal architectures of VAEs for image classification, called evolving deep convolutional VAE (EvoVAE), based on a genetic algorithm (GA). In the proposed EvoVAE algorithm, the traditional VAEs are first generalized to a more generic and asymmetrical one with four different blocks, and then a variable-length gene encoding mechanism of the GA is presented to search for the optimal network depth. Furthermore, an effective genetic operator is designed to adapt to the proposed variable-length gene encoding strategy. To verify the performance of the proposed algorithm, nine variants of AEs and VAEs are chosen as the peer competitors to perform the comparisons on MNIST, street view house numbers, and CIFAR-10 benchmark datasets. The experiments reveal the superiority of the proposed EvoVAE algorithm, which wins 21 times out of the 24 comparisons and outperforms the best competitors by 1.39%, 14.21%, and 13.03% on the three benchmark datasets, respectively.
引用
收藏
页码:815 / 829
页数:15
相关论文
共 65 条
[1]  
[Anonymous], 2015, ICLR
[2]  
[Anonymous], 2016, CoRR
[3]  
[Anonymous], 2001, MULTIOBJECTIVE OPTIM
[4]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[5]  
Assuncao F., 2018, P IEEE C EV COMP CEC, P1
[6]  
Baker B., 2016, P 5 INT C LEARNING R
[7]  
Bengio Y, 2011, LECT NOTES ARTIF INT, V6925, P18, DOI 10.1007/978-3-642-24412-4_3
[8]   Modeling and Transforming Speech using Variational Autoencoders [J].
Blaauw, Merlijn ;
Bonada, Jordi .
17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, :1770-1774
[9]  
Cai H, 2018, AAAI CONF ARTIF INTE, P2787
[10]  
Chen X., 2019, PROC IEEE C COMPUT V, P46