A Self-attention Based Model for Offline Handwritten Text Recognition

被引:2
|
作者
Nam Tuan Ly [1 ]
Trung Tan Ngo [1 ]
Nakagawa, Masaki [1 ]
机构
[1] Tokyo Univ Agr & Technol, Tokyo, Japan
来源
PATTERN RECOGNITION, ACPR 2021, PT II | 2022年 / 13189卷
关键词
Self-attention; Multi-head; Handwritten text recognition; CNN; BLSTM; CTC; SEQUENCE;
D O I
10.1007/978-3-031-02444-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Offline handwritten text recognition is an important part of document analysis and it has been receiving a lot of attention from numerous researchers for decades. In this paper, we present a self-attention-based model for offline handwritten textline recognition. The proposed model consists of three main components: a feature extractor by CNN; an encoder by a BLSTM network and a self-attention module; and a decoder by CTC. The self-attention module is complementary to RNN in the encoder and helps the encoder to capture long-range and multi-level dependencies across an input sequence. According to the extensive experiments on the two datasets of IAM Handwriting and Kuzushiji, the proposed model achieves better accuracy than the state-of-the-art models. The self-attention map visualization shows that the self-attention mechanism helps the encoder capture long-range and multi-level dependencies across an input sequence.
引用
收藏
页码:356 / 369
页数:14
相关论文
共 50 条
  • [1] 2D Self-attention Convolutional Recurrent Network for Offline Handwritten Text Recognition
    Ly, Nam Tuan
    Nguyen, Hung Tuan
    Nakagawa, Masaki
    DOCUMENT ANALYSIS AND RECOGNITION - ICDAR 2021, PT I, 2021, 12821 : 191 - 204
  • [2] Self-attention Networks for Non-recurrent Handwritten Text Recognition
    d'Arce, Rafael
    Norton, Terence
    Hannuna, Sion
    Cristianini, Nello
    FRONTIERS IN HANDWRITING RECOGNITION, ICFHR 2022, 2022, 13639 : 389 - 403
  • [3] An attention based method for offline handwritten L rdu text recognition
    Anjum, Tayaba
    Khan, Nazar
    2020 17TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION (ICFHR 2020), 2020, : 169 - 174
  • [4] Attention-Based Deep Learning Model for Arabic Handwritten Text Recognition
    Gader T.B.A.
    Echi A.K.
    Machine Graphics and Vision, 2022, 31 (1-4): : 49 - 73
  • [5] Attention Augmented Convolutional Recurrent Network for Handwritten Japanese Text Recognition
    Ly, Nam Tuan
    Nguyen, Cuong Tuan
    Nakagawa, Masaki
    2020 17TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION (ICFHR 2020), 2020, : 163 - 168
  • [6] Self-Attention based Siamese Neural Network recognition Model
    Liu, Yuxing
    Chang, Geng
    Fu, Guofeng
    Wei, Yingchao
    Lan, Jie
    Liu, Jiarui
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 721 - 724
  • [7] A Residual-Attention Offline Handwritten Chinese Text Recognition Based on Fully Convolutional Neural Networks
    Wang, Yintong
    Yang, Yingjie
    Ding, Weiping
    Li, Shuo
    IEEE ACCESS, 2021, 9 : 132301 - 132310
  • [8] CALText: Contextual Attention Localization for Offline Handwritten Text
    Tayaba Anjum
    Nazar Khan
    Neural Processing Letters, 2023, 55 : 7227 - 7257
  • [9] CALText: Contextual Attention Localization for Offline Handwritten Text
    Anjum, Tayaba
    Khan, Nazar
    NEURAL PROCESSING LETTERS, 2023, 55 (06) : 7227 - 7257
  • [10] Rejection strategies for offline handwritten text line recognition
    Bertolami, Roman
    Zimmermann, Matthias
    Bunke, Horst
    PATTERN RECOGNITION LETTERS, 2006, 27 (16) : 2005 - 2012