Mixture of Experts with Entropic Regularization for Data Classification

被引:5
|
作者
Peralta, Billy [1 ]
Saavedra, Ariel [2 ]
Caro, Luis [2 ]
Soto, Alvaro [3 ]
机构
[1] Andres Bello Univ, Dept Engn Sci, Santiago 7500971, Chile
[2] Catholic Univ Temuco, Dept Informat Engn, Temuco 4781312, Chile
[3] Pontificia Univ Catolica Chile, Dept Comp Sci, Santiago 7820436, Chile
关键词
mixture-of-experts; regularization; entropy; classification;
D O I
10.3390/e21020190
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. Mixture-of-experts is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a winner-takes-all output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3-6% in some datasets. In future work, we plan to embed feature selection into this model.
引用
收藏
页数:14
相关论文
共 50 条
  • [11] Advances in using hierarchical mixture of experts for signal classification
    Ramamurti, V
    Ghosh, J
    1996 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, CONFERENCE PROCEEDINGS, VOLS 1-6, 1996, : 3569 - 3572
  • [12] Semi-supervised mixture-of-experts classification
    Karakoulas, G
    Salakhutdinov, R
    FOURTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2004, : 138 - 145
  • [13] A mixture of experts network structure for EEG signals classification
    Guler, Inan
    Ubeyli, Elif Derya
    Guler, Nihal Fatma
    2005 27TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY, VOLS 1-7, 2005, : 2707 - 2710
  • [14] A Multilevel Mixture-of-Experts Framework for Pedestrian Classification
    Enzweiler, Markus
    Gavrila, Dariu M.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2011, 20 (10) : 2967 - 2979
  • [15] Toric geometry of entropic regularization
    Sturmfels, Bernd
    Telen, Simon
    Vialard, Francois-Xavier
    von Renesse, Max
    JOURNAL OF SYMBOLIC COMPUTATION, 2024, 120
  • [16] ENTROPIC REGULARIZATION OF NONGRADIENT SYSTEMS
    Adams, Daniel
    Manh Hong Duong
    dos Reis, Goncalo
    SIAM JOURNAL ON MATHEMATICAL ANALYSIS, 2022, 54 (04) : 4495 - 4535
  • [17] Classifying Incomplete Data with a Mixture of Subspace Experts
    Kizaric, Ben A.
    Pimentel-Alarcon, Daniel L.
    2022 58TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2022,
  • [18] Novel approaches to entropic regularization
    Ramos, FM
    Velho, HFC
    Carvalho, JC
    Ferreira, NJ
    INVERSE PROBLEMS, 1999, 15 (05) : 1139 - 1148
  • [19] Entropic Regularization in Hierarchical Games
    Mallozzi L.
    Pardalos P.M.
    Operations Research Forum, 3 (1)
  • [20] Understanding Entropic Regularization in GANs
    Reshetova, Daria
    Bai, Yikun
    Wu, Xiugang
    Ozgur, Ayfer
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25