HBert: A Long Text Processing Method Based on BERT and Hierarchical Attention Mechanisms

被引:3
作者
Lv, Xueqiang [1 ]
Liu, Zhaonan [1 ]
Zhao, Ying [1 ]
Xu, Ge [2 ]
You, Xindong [1 ]
机构
[1] Beijing Informat Sci & Technol Univ, Beijing, Peoples R China
[2] Minjiang Univ, Fuzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
BERT; Hierarchical Attention; Long Text Processing;
D O I
10.4018/IJSWIS.322769
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the emergence of a large-scale pre-training model based on the transformer model, the effect of all-natural language processing tasks has been pushed to a new level. However, due to the high complexity of the transformer's self-attention mechanism, these models have poor processing ability for long text. Aiming at solving this problem, a long text processing method named HBert based on Bert and hierarchical attention neural network is proposed. Firstly, the long text is divided into multiple sentences whose vectors are obtained through the word encoder composed of Bert and the word attention layer. And the article vector is obtained through the sentence encoder that is composed of transformer and sentence attention. Then the article vector is used to complete the subsequent tasks. The experimental results show that the proposed HBert method achieves good results in text classification and QA tasks. The F1 value is 95.7% in longer text classification tasks and 75.2% in QA tasks, which are better than the state-of-the-art model longformer.
引用
收藏
页数:14
相关论文
共 24 条
  • [1] Sentiment analysis method of consumer comment text based on BERT and hierarchical attention in e-commerce big data environment
    Chang, Wanjun
    Zhu, Mingdong
    JOURNAL OF INTELLIGENT SYSTEMS, 2023, 32 (01)
  • [2] A Multiscale Interactive Attention Short Text Classification Model Based on BERT
    Zhou, Lu
    Wang, Peng
    Zhang, Huijun
    Wu, Shengbo
    Zhang, Tao
    IEEE ACCESS, 2024, 12 : 160992 - 161001
  • [3] Language Models for Hierarchical Classification of Radiology Reports With Attention Mechanisms, BERT, and GPT-4
    Olivato, Matteo
    Putelli, Luca
    Arici, Nicola
    Emilio Gerevini, Alfonso
    Lavelli, Alberto
    Serina, Ivan
    IEEE ACCESS, 2024, 12 : 69710 - 69727
  • [4] BVMHA: Text classification model with variable multihead hybrid attention based on BERT
    Peng, Bo
    Zhang, Tao
    Han, Kundong
    Zhang, Zhe
    Ma, Yuquan
    Ma, Mengnan
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2024, 46 (01) : 1443 - 1454
  • [5] The Automatic Text Classification Method Based on BERT and Feature Union
    Li, Wenting
    Gao, Shangbing
    Zhou, Hong
    Huang, Zihe
    Zhang, Kewen
    Li, Wei
    2019 IEEE 25TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2019, : 774 - 777
  • [6] Chinese Text Classification Method Based on BERT Word Embedding
    Wang, Ziniu
    Huang, Zhilin
    Gao, Jianling
    2020 5TH INTERNATIONAL CONFERENCE ON MATHEMATICS AND ARTIFICIAL INTELLIGENCE (ICMAI 2020), 2020, : 66 - 71
  • [7] An Efficient Long Chinese Text Sentiment Analysis Method Using BERT-Based Models with BiGRU
    Sheng, Deming
    Yuan, Jingling
    PROCEEDINGS OF THE 2021 IEEE 24TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN (CSCWD), 2021, : 192 - 197
  • [8] Long text semantic matching model based on BERT and densecomposite network
    Chen Y.-L.
    Gao Z.-C.
    Cai X.-D.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2024, 54 (01): : 232 - 239
  • [9] Bi2-BERT-Based Long Sequence Text Vectorization Method for Power Grid Work Orders
    Cheng, Xiangzhao
    Li, Ting
    Zhang, Zhuo
    Zheng, Yue
    Hu, Guanglei
    Li, Hongxin
    Zhang, Tongqing
    INTERNATIONAL JOURNAL OF MOBILE COMPUTING AND MULTIMEDIA COMMUNICATIONS, 2025, 16 (01)
  • [10] BERT-based response selection in dialogue systems using utterance attention mechanisms
    Park, Yeongjoon
    Ko, Youngjoong
    Seo, Jungyun
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 209