A Self-attention Based Model for Offline Handwritten Text Recognition

被引:2
|
作者
Nam Tuan Ly [1 ]
Trung Tan Ngo [1 ]
Nakagawa, Masaki [1 ]
机构
[1] Tokyo Univ Agr & Technol, Tokyo, Japan
来源
PATTERN RECOGNITION, ACPR 2021, PT II | 2022年 / 13189卷
关键词
Self-attention; Multi-head; Handwritten text recognition; CNN; BLSTM; CTC; SEQUENCE;
D O I
10.1007/978-3-031-02444-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Offline handwritten text recognition is an important part of document analysis and it has been receiving a lot of attention from numerous researchers for decades. In this paper, we present a self-attention-based model for offline handwritten textline recognition. The proposed model consists of three main components: a feature extractor by CNN; an encoder by a BLSTM network and a self-attention module; and a decoder by CTC. The self-attention module is complementary to RNN in the encoder and helps the encoder to capture long-range and multi-level dependencies across an input sequence. According to the extensive experiments on the two datasets of IAM Handwriting and Kuzushiji, the proposed model achieves better accuracy than the state-of-the-art models. The self-attention map visualization shows that the self-attention mechanism helps the encoder capture long-range and multi-level dependencies across an input sequence.
引用
收藏
页码:356 / 369
页数:14
相关论文
共 50 条
  • [31] Chinese Text Sentiment Analysis Based on BI-GRU and Self-attention
    Pan, Yaxing
    Liang, Mingfeng
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 1983 - 1988
  • [32] A self-attention model for viewport prediction based on distance constraint
    Lan, ChengDong
    Qiu, Xu
    Miao, Chenqi
    Zheng, MengTing
    VISUAL COMPUTER, 2024, 40 (09) : 5997 - 6014
  • [33] Offline handwritten signature recognition based on generative adversarial networks
    Jiang, Xiaoguang
    INTERNATIONAL JOURNAL OF BIOMETRICS, 2024, 16 (3-4) : 236 - 255
  • [34] ON THE USEFULNESS OF SELF-ATTENTION FOR AUTOMATIC SPEECH RECOGNITION WITH TRANSFORMERS
    Zhang, Shucong
    Loweimi, Erfan
    Bell, Peter
    Renals, Steve
    2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, : 89 - 96
  • [35] A Study on Self-attention Mechanism for AMR-to-text Generation
    Vu Trong Sinh
    Nguyen Le Minh
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2019), 2019, 11608 : 321 - 328
  • [36] Self-Attention Networks for Text-Independent Speaker Verification
    Bian, Tengyue
    Chen, Fangzhou
    Xu, Li
    PROCEEDINGS OF THE 2019 31ST CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2019), 2019, : 3955 - 3960
  • [37] UniFormer: Unifying Convolution and Self-Attention for Visual Recognition
    Li, Kunchang
    Wang, Yali
    Zhang, Junhao
    Gao, Peng
    Song, Guanglu
    Liu, Yu
    Li, Hongsheng
    Qiao, Yu
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12581 - 12600
  • [38] Self-attention based speaker recognition using Cluster-Range Loss
    Bian, Tengyue
    Chen, Fangzhou
    Xu, Li
    NEUROCOMPUTING, 2019, 368 : 59 - 68
  • [39] A self-attention based neural architecture for Chinese medical named entity recognition
    Wan, Qian
    Liu, Jie
    Wei, Luona
    Ji, Bin
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2020, 17 (04) : 3498 - 3511
  • [40] BiSMSM: A Hybrid MLP-Based Model of Global Self-Attention Processes for EEG-Based Emotion Recognition
    Li, Wei
    Tian, Ye
    Hou, Bowen
    Dong, Jianzhang
    Shao, Shitong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT I, 2022, 13529 : 37 - 48