Gate-Attention and Dual-End Enhancement Mechanism for Multi-Label Text Classification

被引:0
作者
Cheng, Jieren [1 ,2 ]
Chen, Xiaolong [1 ]
Xu, Wenghang [3 ]
Hua, Shuai [3 ]
Tang, Zhu [1 ]
Sheng, Victor S. [4 ]
机构
[1] Hainan Univ, Sch Comp Sci & Technol, Haikou 570228, Peoples R China
[2] Hainan Univ, Hainan Blockchain Technol Engn Res Ctr, Haikou 570228, Peoples R China
[3] Hainan Univ, Sch Cyberspace Secur, Haikou 570228, Peoples R China
[4] Texas Tech Univ, Dept Comp Sci, Lubbock, TX 79409 USA
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2023年 / 77卷 / 02期
基金
中国国家自然科学基金;
关键词
Multi-label text classification; feature extraction; label distribution information; sequence generation;
D O I
10.32604/cmc.2023.042980
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the realm of Multi-Label Text Classification (MLTC), the dual challenges of extracting rich semantic features from text and discerning inter-label relationships have spurred innovative approaches. Many studies in semantic feature extraction have turned to external knowledge to augment the model's grasp of textual content, often overlooking intrinsic textual cues such as label statistical features. In contrast, these endogenous insights naturally align with the classification task. In our paper, to complement this focus on intrinsic knowledge, we introduce a novel Gate-Attention mechanism. This mechanism adeptly integrates statistical features from the text itself into the semantic fabric, enhancing the model's capacity to understand and represent the data. Additionally, to address the intricate task of mining label correlations, we propose a Dual-end enhancement mechanism. This mechanism effectively mitigates the challenges of information loss and erroneous transmission inherent in traditional long short term memory propagation. We conducted an extensive battery of experiments on the AAPD and RCV1-2 datasets. These experiments serve the dual purpose of confirming the efficacy of both the Gate-Attention mechanism and the Dual-end enhancement mechanism. Our final model unequivocally outperforms the baseline model, attesting to its robustness. These findings emphatically underscore the imperativeness of taking into account not just external knowledge but also the inherent intricacies of textual data when crafting potent MLTC models.
引用
收藏
页码:1779 / 1793
页数:15
相关论文
共 34 条
[1]   Learning multi-label scene classification [J].
Boutell, MR ;
Luo, JB ;
Shen, XP ;
Brown, CM .
PATTERN RECOGNITION, 2004, 37 (09) :1757-1771
[2]  
Chen GB, 2017, IEEE IJCNN, P2377, DOI 10.1109/IJCNN.2017.7966144
[3]  
Chen JD, 2019, AAAI CONF ARTIF INTE, P6252
[4]   Improving Emotion Analysis for Speech-Induced EEGs Through EEMD-HHT-Based Feature Extraction and Electrode Selection [J].
Chen, Jing ;
Li, Haifeng ;
Ma, Lin ;
Bo, Hongjian .
INTERNATIONAL JOURNAL OF MULTIMEDIA DATA ENGINEERING & MANAGEMENT, 2021, 12 (02) :1-18
[5]   A Survey of Multi-label Text Classification Based on Deep Learning [J].
Chen, Xiaolong ;
Cheng, Jieren ;
Liu, Jingxin ;
Xu, Wenghang ;
Hua, Shuai ;
Tang, Zhu ;
Sheng, Victor S. .
ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT I, 2022, 13338 :443-456
[6]  
Chen YB, 2020, J CONTEMP CHINA, V29, P1, DOI [10.1007/s12652-020-02066-z, 10.1080/10670564.2019.1621526, 10.1080/01932691.2020.1791172]
[7]   Multi-label text classification with latent word-wise label information [J].
Chen, Ziheng ;
Ren, Jiangtao .
APPLIED INTELLIGENCE, 2021, 51 (02) :966-979
[8]   Accurate use of label dependency in multi-label text classification through the lens of causality [J].
Fan, Caoyun ;
Chen, Wenqing ;
Tian, Jidong ;
Li, Yitian ;
He, Hao ;
Jin, Yaohui .
APPLIED INTELLIGENCE, 2023, 53 (19) :21841-21857
[9]   A Hereditary Attentive Template-based Approach for Complex Knowledge Base Question Answering Systems [J].
Gomes, Jorao ;
de Mello, Romulo Chrispim ;
Stroele, Victor ;
de Souza, Jairo Francisco .
EXPERT SYSTEMS WITH APPLICATIONS, 2022, 205
[10]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]