Deep supervised learning with mixture of neural networks

被引:11
作者
Hu, Yaxian [1 ]
Luo, Senlin [1 ]
Han, Longfei [1 ]
Pan, Limin [1 ]
Zhang, Tiemei [2 ]
机构
[1] Beijing Inst Technol, Informat Syst & Secur & Countermeasures Expt Ctr, Beijing 100081, Peoples R China
[2] Minist Hlth, Beijing Hosp, Beijing Inst Geriatr, Beijing, Peoples R China
关键词
Deep neural network; Mixture model; Expectation maximization; Diabetes determination; MODELS; CLASSIFICATION;
D O I
10.1016/j.artmed.2019.101764
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Network (DNN), as a deep architectures, has shown excellent performance in classification tasks. However, when the data has different distributions or contains some latent non-observed factors, it is difficult for DNN to train a single model to perform well on the classification tasks. In this paper, we propose mixture model based on DNNs (MoNNs), a supervised approach to perform classification tasks with a gating network and multiple local expert models. We use a neural network as a gating function and use DNNs as local expert models. The gating network split the heterogeneous data into several homogeneous components. DNNs are combined to perform classification tasks in each component. Moreover, we use EM (Expectation Maximization) as an optimization algorithm. Experiments proved that our MoNNs outperformed the other compared methods on determination of diabetes, determination of benign or malignant breast cancer, and handwriting recognition. Therefore, the MoNNs can solve the problem of data heterogeneity and have a good effect on classification tasks.
引用
收藏
页数:6
相关论文
共 28 条
[1]   Positive vectors clustering using inverted Dirichlet finite mixture models [J].
Bdiri, Taoufik ;
Bouguila, Nizar .
EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (02) :1869-1882
[2]   Bayesian learning of finite generalized inverted Dirichlet mixtures: Application to object classification and forgery detection [J].
Bourouis, Sami ;
Al Mashrgy, Mohamed ;
Bouguila, Nizar .
EXPERT SYSTEMS WITH APPLICATIONS, 2014, 41 (05) :2329-2336
[3]   Model-Based Learning Using a Mixture of Mixtures of Gaussian and Uniform Distributions [J].
Browne, Ryan P. ;
McNicholas, Paul D. ;
Sparling, Matthew D. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (04) :814-817
[4]   Robust mixture of experts modeling using the t distribution [J].
Chamroukhi, F. .
NEURAL NETWORKS, 2016, 79 :20-36
[5]   Improving sentiment analysis via sentence type classification using BiLSTM-CRF and CNN [J].
Chen, Tao ;
Xu, Ruifeng ;
He, Yulan ;
Wang, Xuan .
EXPERT SYSTEMS WITH APPLICATIONS, 2017, 72 :221-230
[6]   Training mixture of weighted SVM for object detection using EM algorithm [J].
Cheng, De ;
Wang, Jinjun ;
Wei, Xing ;
Gong, Yihong .
NEUROCOMPUTING, 2015, 149 :473-482
[7]  
Ciresan D, 2012, PROC CVPR IEEE, P3642, DOI 10.1109/CVPR.2012.6248110
[8]   Mixture Class Recovery in GMM Under Varying Degrees of Class Separation: Frequentist Versus Bayesian Estimation [J].
Depaoli, Sarah .
PSYCHOLOGICAL METHODS, 2013, 18 (02) :186-219
[9]   Estimating model-error covariances in nonlinear state-space models using Kalman smoothing and the expectation-maximization algorithm [J].
Dreano, D. ;
Tandeo, P. ;
Pulido, M. ;
Ait-El-Fquih, B. ;
Chonavel, T. ;
Hoteit, I. .
QUARTERLY JOURNAL OF THE ROYAL METEOROLOGICAL SOCIETY, 2017, 143 (705) :1877-1885
[10]   Variational Learning for Finite Dirichlet Mixture Models and Applications [J].
Fan, Wentao ;
Bouguila, Nizar ;
Ziou, Djemel .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2012, 23 (05) :762-774