A Self-attention Based Model for Offline Handwritten Text Recognition

被引:2
|
作者
Nam Tuan Ly [1 ]
Trung Tan Ngo [1 ]
Nakagawa, Masaki [1 ]
机构
[1] Tokyo Univ Agr & Technol, Tokyo, Japan
来源
PATTERN RECOGNITION, ACPR 2021, PT II | 2022年 / 13189卷
关键词
Self-attention; Multi-head; Handwritten text recognition; CNN; BLSTM; CTC; SEQUENCE;
D O I
10.1007/978-3-031-02444-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Offline handwritten text recognition is an important part of document analysis and it has been receiving a lot of attention from numerous researchers for decades. In this paper, we present a self-attention-based model for offline handwritten textline recognition. The proposed model consists of three main components: a feature extractor by CNN; an encoder by a BLSTM network and a self-attention module; and a decoder by CTC. The self-attention module is complementary to RNN in the encoder and helps the encoder to capture long-range and multi-level dependencies across an input sequence. According to the extensive experiments on the two datasets of IAM Handwriting and Kuzushiji, the proposed model achieves better accuracy than the state-of-the-art models. The self-attention map visualization shows that the self-attention mechanism helps the encoder capture long-range and multi-level dependencies across an input sequence.
引用
收藏
页码:356 / 369
页数:14
相关论文
共 50 条
  • [21] Self-attention for Speech Emotion Recognition
    Tarantino, Lorenzo
    Garner, Philip N.
    Lazaridis, Alexandros
    INTERSPEECH 2019, 2019, : 2578 - 2582
  • [22] Data Augmentation for Offline Handwritten Text Recognition: A Systematic Literature Review
    de Sousa Neto A.F.
    Bezerra B.L.D.
    de Moura G.C.D.
    Toselli A.H.
    SN Computer Science, 5 (2)
  • [23] Efficient Self-Attention Model for Speech Recognition-Based Assistive Robots Control
    Poirier, Samuel
    Cote-Allard, Ulysse
    Routhier, Francois
    Campeau-Lecours, Alexandre
    SENSORS, 2023, 23 (13)
  • [24] An Aerial Target Recognition Algorithm Based on Self-Attention and LSTM
    Liang, Futai
    Chen, Xin
    He, Song
    Song, Zihao
    Lu, Hao
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 81 (01): : 1101 - 1121
  • [25] Joint Model of Entity Recognition and Relation Extraction with Self-attention Mechanism
    Liu, Maofu
    Zhang, Yukun
    Li, Wenjie
    Ji, Donghong
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (04)
  • [26] A Computationally Efficient Pipeline Approach to Full Page Offline Handwritten Text Recognition
    Chung, Jonathan
    Delteil, Thomas
    2019 INTERNATIONAL CONFERENCE ON DOCUMENT ANALYSIS AND RECOGNITION WORKSHOPS (ICDARW), VOL 5, 2019, : 35 - 40
  • [27] An Effective Personality-Based Model for Short Text Sentiment Classification Using BiLSTM and Self-Attention
    Liu, Kejian
    Feng, Yuanyuan
    Zhang, Liying
    Wang, Rongju
    Wang, Wei
    Yuan, Xianzhi
    Cui, Xuran
    Li, Xianyong
    Li, Hailing
    ELECTRONICS, 2023, 12 (15)
  • [28] Cyclic Self-attention for Point Cloud Recognition
    Zhu, Guanyu
    Zhou, Yong
    Yao, Rui
    Zhu, Hancheng
    Zhao, Jiaqi
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (01)
  • [29] Quantum self-attention neural networks for text classification
    Li, Guangxi
    Zhao, Xuanqiang
    Wang, Xin
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (04)
  • [30] SELF-ATTENTION BASED DARKNET NAMED ENTITY RECOGNITION WITH BERT METHODS
    Chen, Yuxuan
    Guo, Yubin
    Jiang, Hong
    Ding, Jianwei
    Chen, Zhouguo
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2021, 17 (06): : 1973 - 1988