MVQS: Robust multi-view instance-level cost-sensitive learning method for imbalanced data classification

被引:6
作者
Hou, Zhaojie [1 ,2 ]
Tang, Jingjing [1 ,2 ]
Li, Yan [1 ,2 ]
Fu, Saiji [3 ]
Tian, Yingjie [4 ,5 ,6 ,7 ]
机构
[1] Southwestern Univ Finance & Econ, Fac Business Adm, Sch Business Adm, Chengdu 611130, Peoples R China
[2] Southwestern Univ Finance & Econ, Inst Big Data, Chengdu 611130, Peoples R China
[3] Beijing Univ Posts & Telecommun, Sch Econ & Management, Beijing 100876, Peoples R China
[4] Univ Chinese Acad Sci, Sch Econ & Management, Beijing 100190, Peoples R China
[5] Chinese Acad Sci, Res Ctr Fictitious Econ & Data Sci, Beijing 100190, Peoples R China
[6] Chinese Acad Sci, Key Lab Big Data Min & Knowledge Management, Beijing 100190, Peoples R China
[7] UCAS, MOE Social Sci Lab Digital Econ Forecasts & Policy, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-view learning; Imbalanced classes; Noisy samples; Support vector machine; QTSE loss function; SUPPORT VECTOR MACHINE; KERNEL-METHOD; CONSENSUS; BLINEX; NOISE;
D O I
10.1016/j.ins.2024.120467
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-view imbalanced learning is to handle the datasets with multi-view representations and imbalanced classes. Existing multi-view imbalanced learning methods can be divided into two main categories: multi-view ensemble learning and multi-view cost-sensitive learning. However, these methods suffer from the following problems: 1) neglecting either consensus or complementary information, 2) complex preprocessing and information fusion in multiview ensemble learning and manual assignment of misclassification costs in multi-view costsensitive learning, and 3) limited ability to handle noisy samples. Therefore, we aim to design a concise and unified framework to grapple with the multi-view representations, imbalanced classes and noisy samples simultaneously. Inspired by the merits of support vector machine (SVM) and quadratic type squared error (QTSE) loss function, we propose a robust multi-view instance-level cost-sensitive SVM with QTSE loss (MVQS) for imbalanced data classification. The consensus regularization term and combination weight strategy are employed to fully exploit multi-view information. The QTSE loss can adaptively impose instance-level penalties to the misclassification of samples, and make MVQS be robust to noisy samples. We solve MVQS with the alternating direction method of multipliers (ADMM) and the gradient descent (GD) algorithm. Comprehensive experiments validate that MVQS is more competitive and robust than other benchmark approaches.
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Benielli D, 2022, J MACH LEARN RES, V23, P1
  • [2] Robust twin bounded support vector machines for outliers and imbalanced data
    Borah, Parashjyoti
    Gupta, Deepak
    [J]. APPLIED INTELLIGENCE, 2021, 51 (08) : 5314 - 5343
  • [3] An ensemble approach to multi-view multi-instance learning
    Cano, Alberto
    [J]. KNOWLEDGE-BASED SYSTEMS, 2017, 136 : 46 - 57
  • [4] Consensus and complementarity based maximum entropy discrimination for multi-view classification
    Chao, Guoqing
    Sun, Shiliang
    [J]. INFORMATION SCIENCES, 2016, 367 : 296 - 310
  • [5] Collobert R., 2006, P 23 INT C MACH LEAR, P201
  • [6] Dinuzzo F., 2012, Advances in Neural Information Processing Systems, P189
  • [7] An emotion role mining approach based on multiview ensemble learning in social networks
    Du, Yajun
    Wang, Yakun
    Hu, Jinrong
    Li, Xianyong
    Chen, Xiaoliang
    [J]. INFORMATION FUSION, 2022, 88 : 100 - 114
  • [8] Ensemble sparse representation-based cyber threat hunting for security of smart cities
    Fard, Seyed Mehdi Hazrati
    Karimipour, Hadis
    Dehghantanha, Ali
    Jahromi, Amir Namavar
    Srivastava, Gautam
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2020, 88
  • [9] Fernndez A., 2018, Learning from Imbalanced Data Sets, DOI [DOI 10.1007/978-3-319-98074-4, 10.1007/978-3-319-98074-4]
  • [10] Classification in the Presence of Label Noise: a Survey
    Frenay, Benoit
    Verleysen, Michel
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (05) : 845 - 869