LCBM: A Multi-View Probabilistic Model for Multi-Label Classification

被引:28
作者
Sun, Shiliang [1 ]
Zong, Daoming [1 ]
机构
[1] East China Normal Univ, Sch Comp Sci & Technol, 3663 North Zhongshan Rd, Shanghai 200062, Peoples R China
基金
中国国家自然科学基金;
关键词
Probabilistic logic; Task analysis; Prediction algorithms; Support vector machines; Kernel; Training; Semantics; Multi-view learning; multi-label classification; Bernoulli mixture; probabilistic model; variational autoencoder; INFERENCE;
D O I
10.1109/TPAMI.2020.2974203
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-label classification is an important research topic in machine learning, for which exploiting label dependencies is an effective modeling principle. Recently, probabilistic models have shown great potential in discovering dependencies among labels. In this paper, motivated by the recent success of multi-view learning to improve the generalization performance, we propose a novel multi-view probabilistic model named latent conditional Bernoulli mixture (LCBM) for multi-label classification. LCBM is a generative model taking features from different views as inputs, and conditional on the latent subspace shared by the views a Bernoulli mixture model is adopted to build label dependencies. Inside each component of the mixture, the labels have a weak correlation which facilitates computational convenience. The mean field variational inference framework is used to carry out approximate posterior inference in the probabilistic model, where we propose a Gaussian mixture variational autoencoder (GMVAE) for effective posterior approximation. We further develop a scalable stochastic training algorithm for efficiently optimizing the model parameters and variational parameters, and derive an efficient prediction procedure based on greedy search. Experimental results on multiple benchmark datasets show that our approach outperforms other state-of-the-art methods under various metrics.
引用
收藏
页码:2682 / 2696
页数:15
相关论文
共 49 条
  • [1] [Anonymous], 2016, P 33 INT C MACHINE L
  • [2] [Anonymous], 2006, Pattern Recognition and Machine Learning
  • [3] Bach F. R., 2005, 688 U CAL DEPT STAT
  • [4] HARAM: a Hierarchical ARAM neural network for large-scale text classification
    Benites, Fernando
    Sapozhnikova, Elena
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOP (ICDMW), 2015, : 847 - 854
  • [5] Blei D M, 2012, P 29 INT C MACH LEAR
  • [6] MLTSVM: A novel twin support vector machine to multi-label learning
    Chen, Wei-Jie
    Shao, Yuan -Hai
    Li, Chun-Na
    Deng, Nai-Yang
    [J]. PATTERN RECOGNITION, 2016, 52 : 61 - 74
  • [7] On label dependence and loss minimization in multi-label classification
    Dembczynski, Krzysztof
    Waegeman, Willem
    Cheng, Weiwei
    Huellermeier, Eyke
    [J]. MACHINE LEARNING, 2012, 88 (1-2) : 5 - 45
  • [8] Demsar J, 2006, J MACH LEARN RES, V7, P1
  • [9] Elisseeff A, 2002, ADV NEUR IN, V14, P681
  • [10] An ensemble-based approach for multi-view multi-label classification
    Gibaja E.L.
    Moyano J.M.
    Ventura S.
    [J]. Ventura, Sebastián (sventura@uco.es), 2016, Springer Verlag (05) : 251 - 259