Mixture of Experts with Entropic Regularization for Data Classification

被引:5
|
作者
Peralta, Billy [1 ]
Saavedra, Ariel [2 ]
Caro, Luis [2 ]
Soto, Alvaro [3 ]
机构
[1] Andres Bello Univ, Dept Engn Sci, Santiago 7500971, Chile
[2] Catholic Univ Temuco, Dept Informat Engn, Temuco 4781312, Chile
[3] Pontificia Univ Catolica Chile, Dept Comp Sci, Santiago 7820436, Chile
关键词
mixture-of-experts; regularization; entropy; classification;
D O I
10.3390/e21020190
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. Mixture-of-experts is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a winner-takes-all output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3-6% in some datasets. In future work, we plan to embed feature selection into this model.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] DeepME: Deep Mixture Experts for Large-scale Image Classification
    He, Ming
    Lv, Guangyi
    He, Weidong
    Fan, Jianping
    Zeng, Guihua
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 722 - 728
  • [32] Mixture of experts for classification of gender, ethnic origin, and pose of human faces
    Gutta, S
    Huang, JRJ
    Jonathon, P
    Wechsler, H
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2000, 11 (04): : 948 - 960
  • [33] A regularized root-quartic mixture of experts for complex classification problems
    Abbasi, Elham
    Ebrahim, Mohammad
    Ghatee, Mehdi
    KNOWLEDGE-BASED SYSTEMS, 2016, 110 : 98 - 109
  • [34] Classification in mixture of experts using hard clustering and a new gate function
    Bulut, Faruk
    Amasyali, M. Fatih
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2016, 31 (04): : 1017 - 1025
  • [35] A Hierarchical Mixture-Of-Experts Framework for Few Labeled Node Classification
    Wang, Yimeng
    Yang, Zhiyao
    Che, Xiangjiu
    NEURAL NETWORKS, 2025, 188
  • [36] Data dependent risk bounds for hierarchical mixture of experts classifiers
    Azran, A
    Meir, R
    LEARNING THEORY, PROCEEDINGS, 2004, 3120 : 427 - 441
  • [37] Steered Mixture-of-Experts Approximation of Spherical Image Data
    Verhack, Ruben
    Madhu, Nilesh
    Van Wallendael, Glenn
    Lambert, Peter
    Sikora, Thomas
    2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 256 - 260
  • [38] Relational Mixture of Experts: Explainable Demographics Prediction with Behavioral Data
    Oyamada, Masafumi
    Nakadai, Shinji
    2017 17TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2017, : 357 - 366
  • [39] A MIXTURE OF EXPERTS MODEL FOR RANK DATA WITH APPLICATIONS IN ELECTION STUDIES
    Gormley, Isobel Claire
    Murphy, Thomas Brendan
    ANNALS OF APPLIED STATISTICS, 2008, 2 (04): : 1452 - 1477
  • [40] Regularization of Unlabeled Data for Learning of Classifiers based on Mixture Models
    Iswanto, Bambang Heru
    ICICI-BME: 2009 INTERNATIONAL CONFERENCE ON INSTRUMENTATION, COMMUNICATION, INFORMATION TECHNOLOGY, AND BIOMEDICAL ENGINEERING, 2009, : 345 - 349