Transformer-based Hierarchical Encoder for Document Classification

被引:1
|
作者
Sakhrani, Harsh [1 ]
Parekh, Saloni [1 ]
Ratadiya, Pratik [2 ]
机构
[1] Pune Inst Comp Technol, Pune, Maharashtra, India
[2] vCreaTek Consulting Serv Pvt Ltd, Pune, Maharashtra, India
关键词
Transformer; Self-attention; Document Classification;
D O I
10.1109/ICDMW53433.2021.00109
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Document Classification has a wide range of applications in various domains like Ontology Mapping, Sentiment Analysis, Topic Categorization and Document Clustering, to mention a few. Unlike Text Classification, Document Classification works with longer sequences that typically contain multiple paragraphs. Previous approaches for this task have achieved promising results, but have often relied on complex recurrence mechanisms that are expensive and time-consuming in nature. Recently, self-attention based models like Transformers and BERT have achieved state-of-the-art performance on several Natural Language Understanding (NLU) tasks, but owing to the quadratic computational complexity of the self-attention mechanism with respect to the input sequence length, these approaches are generally applied to shorter text sequences. In this paper, we address this issue, by proposing a new Transformer-based Hierarchical Encoder approach for the Document Classification task. The hierarchical framework we adopt helps us extend the self-attention mechanism to long-form text modelling thereby reducing the complexity considerably. We use the Bidirectional Transformer Encoder (BTE) at the sentence-level to generate a fixed-size sentence embedding for each sentence in the document. A document-level Transformer Encoder is then used to model the global document context and learn the inter-sentence dependencies. We also carry out experiments with the BTE in a feature-extraction and a fine-tuning setup, allowing us to evaluate the trade-off between computation power and accuracy. Furthermore, we also conduct ablation experiments, and evaluate the impact of different pre-training strategies on the overall performance. Experimental results demonstrate that our proposed model achieves state-of-the-art performance on two standard benchmark datasets.
引用
收藏
页码:852 / 858
页数:7
相关论文
共 50 条
  • [21] Transformer-Based Hierarchical Model for Non-Small Cell Lung Cancer Detection and Classification
    Imran, Muhammad
    Haq, Bushra
    Elbasi, Ersin
    Topcu, Ahmet E.
    Shao, Wei
    IEEE ACCESS, 2024, 12 : 145920 - 145933
  • [22] TRANSFORMER-BASED HIERARCHICAL CLUSTERING FOR BRAIN NETWORK ANALYSIS
    Dai, Wei
    Cui, Hejie
    Kan, Xuan
    Guo, Ying
    Van Rooij, Sanne
    Yang, Carl
    2023 IEEE 20TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING, ISBI, 2023,
  • [23] Transformer-Based Bidirectional Encoder Representations for Emotion Detection from Text
    Kumar, Ashok J.
    Cambria, Erik
    Trueman, Tina Esther
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [24] Transformer-based Encoder-Decoder Model for Surface Defect Detection
    Lu, Xiaofeng
    Fan, Wentao
    6TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE, ICIAI2022, 2022, : 125 - 130
  • [25] Transformer-based Sparse Encoder and Answer Decoder for Visual Question Answering
    Peng, Longkun
    An, Gaoyun
    Ruan, Qiuqi
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 120 - 123
  • [26] Transformer-Based Approach for Automatic Semantic Financial Document Verification
    Toprak, Ahmet
    Turan, Metin
    IEEE ACCESS, 2024, 12 : 184327 - 184349
  • [27] A 3-D-Swin Transformer-Based Hierarchical Contrastive Learning Method for Hyperspectral Image Classification
    Huang, Xin
    Dong, Mengjie
    Li, Jiayi
    Guo, Xian
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [28] A Transformer-Based Framework for Payload Malware Detection and Classification
    Stein, Kyle
    Mahyari, Arash
    Francia, Guillermo, III
    El-Sheikh, Eman
    2024 IEEE 5TH ANNUAL WORLD AI IOT CONGRESS, AIIOT 2024, 2024, : 0105 - 0111
  • [29] An improved transformer-based concrete crack classification method
    Guanting Ye
    Wei Dai
    Jintai Tao
    Jinsheng Qu
    Lin Zhu
    Qiang Jin
    Scientific Reports, 14
  • [30] TRANSFORMER-BASED DOMAIN ADAPTATION FOR EVENT DATA CLASSIFICATION
    Zhao, Junwei
    Zhang, Shiliang
    Huang, Tiejun
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4673 - 4677