Research on News Text Classification Based on BERT-BiLSTM-TextCNN-Attention

被引:2
作者
Wang, Jia [1 ]
Li, Zongting [2 ]
Ma, Chenyang [2 ]
机构
[1] Dalian Polytech Univ, Dalian 116034, Liaoning, Peoples R China
[2] Dalian Polytech Univ, Sch Informat Sci & Engn, Dalian 116034, Liaoning, Peoples R China
来源
PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024 | 2024年
关键词
Deep learning; text classification; natural language processing; neural network;
D O I
10.1145/3672919.3672973
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditional machine learning models are difficult to capture complex features and contextual relationships. While a singular deep learning architecture surpasses machine learning in text processing, it falls short of encompassing the entirety of textual information [5]. Enter a novel approach: a news text classifier built upon BERT-BiLSTM-TextCNN-Attention. This model employs BERT's pre-trained language models to delve into text content. It then channels this data into a BiLSTM layer, capturing sequence nuances and long-term dependencies for comprehensive semantic insight. Following this, the output moves through a TextCNN layer, effectively capturing local semantic cues through convolution. The model culminates with attention mechanisms that highlight pivotal text attributes, refining feature vectors for the Softmax layer's classification. The experimentation utilized a subset of the THUCNews Chinese news text dataset. Results indicate that the BERT BiLSTM TextCNN Attention model achieved 96.48% accuracy, outperforming other benchmarks. This underscores its superiority in handling Chinese news text classification and validating its prowess in extracting deep semantic nuances and crucial local features from the text.
引用
收藏
页码:295 / 298
页数:4
相关论文
共 10 条
[1]  
[陈可嘉 Chen Kejia], 2022, [计算机工程, Computer Engineering], V48, P59
[2]  
Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
[3]  
Han Hongyong, 2020, Science and Technology Wind, P223, DOI [10.19392/j.cnki.1671-7341.202014191, DOI 10.19392/J.CNKI.1671-7341.202014191]
[4]  
Jawahar G, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P3651
[5]  
Lin D.P., 2021, J. Beijing Inst. Print, V29, P156
[6]  
Wang Hui, 2023, Journal of Xi'an University of Arts and Sciences (Natural Science Edition), V26, P49
[7]  
Xiang Zhihua, 2019, J. Software, V40, P94
[8]  
Ye R., 2023, Electron Tech App, V49, P8
[9]  
Zhang Xiaowei, 2021, Television Technology, V45, P146, DOI [10.16280/j.videoe.2021.07.040, DOI 10.16280/J.VIDEOE.2021.07.040]
[10]  
Zhang Xin, 2023, Computer and Digital Engineering, V51