Mutual information-based dropout: Learning deep relevant feature representation architectures

被引:9
|
作者
Chen, Jie [1 ,2 ]
Wu, ZhongCheng [1 ,3 ]
Zhang, Jun [1 ]
Li, Fang [1 ]
机构
[1] Chinese Acad Sci, Hefei Inst Phys Sci, Hefei, Anhui, Peoples R China
[2] Univ Sci & Technol China, Grad Sch Comp Appl Technol, Hefei, Anhui, Peoples R China
[3] Univ Sci & Technol China, Hefei, Anhui, Peoples R China
关键词
Mutual information; Feature representation; Dropout; Overfitting; Autoencoder; Convolutional neural networks; NEURAL-NETWORKS; AUTOENCODERS; SELECTION;
D O I
10.1016/j.neucom.2019.04.090
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a new regularization strategy called DropMI, which is a generalization of Dropout for the regularization of networks that introduces mutual information (MI) dynamic analysis. The standard Dropout randomly drops a certain proportion of neural units, according to the Bernoulli distribution, thereby resulting in the loss of some important hidden feature information. In DropMI, we first evaluate the importance of each neural unit in the feature representation of the hidden layer based on the MI between it and the target. We then construct a new binary mask matrix based on the sorting distribution of MI, thus developing a dynamic DropMI strategy that highlights the important neural units that are beneficial to the feature representation. The results from the MNIST, NORB, CIFAR-10, CIFAR-100, SVHN, and Multi-PIE datasets indicate that, relative to other state-of-the-art regularization methods based on the benchmark autoencoder and convolutional neural networks, our method has better feature representation performance and effectively reduces the overfitting of the model. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:173 / 184
页数:12
相关论文
共 50 条
  • [1] Information-Based Learning of Deep Architectures for Feature Extraction
    Melacci, Stefano
    Lippi, Marco
    Gori, Marco
    Maggini, Marco
    IMAGE ANALYSIS AND PROCESSING (ICIAP 2013), PT II, 2013, 8157 : 101 - 110
  • [2] Mutual Information-Based Feature Selection and Ensemble Learning for Classification
    Qi, Chengming
    Zhou, Zhangbing
    Wang, Qun
    Hu, Lishuan
    2016 INTERNATIONAL CONFERENCE ON IDENTIFICATION, INFORMATION AND KNOWLEDGE IN THE INTERNET OF THINGS (IIKI), 2016, : 116 - 121
  • [3] Mutual information-based feature selection for radiomics
    Oubel, Estanislao
    Beaumont, Hubert
    Iannessi, Antoine
    MEDICAL IMAGING 2016: PACS AND IMAGING INFORMATICS: NEXT GENERATION AND INNOVATIONS, 2016, 9789
  • [4] Optimizing Multimodal Scene Recognition through Mutual Information-Based Feature Selection in Deep Learning Models
    Hammad, Mohamed
    Chelloug, Samia Allaoua
    Alayed, Walaa
    El-Latif, Ahmed A. Abd
    APPLIED SCIENCES-BASEL, 2023, 13 (21):
  • [5] Dynamic mutual information-based feature selection for multi-label learning
    Kim, Kyung-Jun
    Jun, Chi-Hyuck
    INTELLIGENT DATA ANALYSIS, 2023, 27 (04) : 891 - 909
  • [6] Stopping rules for mutual information-based feature selection
    Mielniczuk, Jan
    Teisseyre, Pawel
    NEUROCOMPUTING, 2019, 358 : 255 - 274
  • [7] Mutual information-based feature selection for multilabel classification
    Doquire, Gauthier
    Verleysen, Michel
    NEUROCOMPUTING, 2013, 122 : 148 - 155
  • [8] A Study on Mutual Information-Based Feature Selection in Classifiers
    Arundhathi, B.
    Athira, A.
    Rajan, Ranjidha
    ARTIFICIAL INTELLIGENCE AND EVOLUTIONARY COMPUTATIONS IN ENGINEERING SYSTEMS, ICAIECES 2016, 2017, 517 : 479 - 486
  • [9] CONDITIONAL DYNAMIC MUTUAL INFORMATION-BASED FEATURE SELECTION
    Liu, Huawen
    Mo, Yuchang
    Zhao, Jianmin
    COMPUTING AND INFORMATICS, 2012, 31 (06) : 1193 - 1216
  • [10] MUTUAL INFORMATION-BASED FAIR ACTIVE LEARNING
    Sonoda, Ryosuke
    Srinivasan, Ramya
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 4965 - 4969