An effective multi-modal adaptive contextual feature information fusion method for Chinese long text classification

被引:0
作者
Xu, Yangshuyi [1 ]
Liu, Guangzhong [1 ]
Zhang, Lin [1 ]
Shen, Xiang [1 ]
Luo, Sizhe [1 ]
机构
[1] Shanghai Maritime Univ, Coll Informat Engn, Shanghai 201306, Peoples R China
关键词
Chinese long text classification; Graph convolutional network; Modal interaction; Noise information filtering; Adaptive modal feature fusion;
D O I
10.1007/s10462-024-10835-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Chinese long text classification plays a vital role in Natural Language Processing. Compared to Chinese short texts, Chinese long texts contain more complex semantic feature information. Furthermore, the distribution of these semantic features is uneven due to the varying lengths of the texts. Current research on Chinese long text classification models primarily focuses on enhancing text semantic features and representing Chinese long texts as graph-structured data. Nonetheless, these methods are still susceptible to noise information and tend to overlook the deep semantic information in long texts. To address the above challenges, this study proposes a novel and effective method called MACFM, which introduces a deep feature information mining method and an adaptive modal feature information fusion strategy to learn the semantic features of Chinese long texts thoroughly. First, we present the DCAM module to capture complex semantic features in Chinese long texts, allowing the model to learn detailed high-level representation features. Then, we explore the relationships between word vectors and text graphs, enabling the model to capture abundant semantic information and text positional information from the graph. Finally, we develop the AMFM module to effectively combine different modal feature representations and eliminate the unrelated noise information. The experimental results on five Chinese long text datasets show that our method significantly improves the accuracy of Chinese long text classification tasks. Furthermore, the generalization experiments on five English datasets and the visualized results demonstrate the effectiveness and interpretability of the MACFM model.
引用
收藏
页数:29
相关论文
共 66 条
  • [1] Graph Receptive Transformer Encoder for Text Classification
    Aras, Arda Can
    Alikasifoglu, Tuna
    Koc, Aykut
    [J]. IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2024, 10 : 347 - 359
  • [2] Gated multimodal networks
    Arevalo, John
    Solorio, Thamar
    Montes-y-Gomez, Manuel
    Gonzalez, Fabio A.
    [J]. NEURAL COMPUTING & APPLICATIONS, 2020, 32 (14) : 10209 - 10228
  • [3] Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
  • [4] Deep Learning with Graph Convolutional Networks: An Overview and Latest Applications in Computational Intelligence
    Bhatti, Uzair Aslam
    Tang, Hao
    Wu, Guilu
    Marjan, Shah
    Hussain, Aamir
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2023, 2023
  • [5] Brown Tom B., 2020, ADV NEURAL INFORM PR, V2005, P14165, DOI DOI 10.48550/ARXIV.2005.14165
  • [6] A Long-Text Classification Method of Chinese News Based on BERT and CNN
    Chen, Xinying
    Cong, Peimin
    Lv, Shuo
    [J]. IEEE ACCESS, 2022, 10 : 34046 - 34057
  • [7] HSAN-capsule: A novel text classification model
    Cheng, Yan
    Zou, Haifeng
    Sun, Huan
    Chen, Haomai
    Cai, Yingying
    Li, Meng
    Du, Qinyi
    [J]. NEUROCOMPUTING, 2022, 489 : 521 - 533
  • [8] Self-training method based on GCN for semi-supervised short text classification
    Cui, Hongyan
    Wang, Gangkun
    Li, Yuanxin
    Welsch, Roy E.
    [J]. INFORMATION SCIENCES, 2022, 611 : 18 - 29
  • [9] Pre-Training With Whole Word Masking for Chinese BERT
    Cui, Yiming
    Che, Wanxiang
    Liu, Ting
    Qin, Bing
    Yang, Ziqing
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 3504 - 3514
  • [10] Dai JQ, 2021, Arxiv, DOI [arXiv:2104.04986, DOI 10.48550/ARXIV.2104.04986]