End-to-End Transformer-Based Models in Textual-Based NLP

被引:32
|
作者
Rahali, Abir [1 ]
Akhloufi, Moulay A. [1 ]
机构
[1] Univ Moncton, Dept Comp Sci, Percept Robot & Intelligent Machines Res Grp PRIME, Moncton, NB E1A 3E9, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Transformers; deep learning; natural language processing; transfer learning; PRE-TRAINED BERT; PREDICTION; SYSTEMS;
D O I
10.3390/ai4010004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer architectures are highly expressive because they use self-attention mechanisms to encode long-range dependencies in the input sequences. In this paper, we present a literature review on Transformer-based (TB) models, providing a detailed overview of each model in comparison to the Transformer's standard architecture. This survey focuses on TB models used in the field of Natural Language Processing (NLP) for textual-based tasks. We begin with an overview of the fundamental concepts at the heart of the success of these models. Then, we classify them based on their architecture and training mode. We compare the advantages and disadvantages of popular techniques in terms of architectural design and experimental value. Finally, we discuss open research, directions, and potential future work to help solve current TB application challenges in NLP.
引用
收藏
页码:54 / 110
页数:57
相关论文
共 50 条
  • [1] Transformer-Based End-to-End Anatomical and Functional Image Fusion
    Zhang, Jing
    Liu, Aiping
    Wang, Dan
    Liu, Yu
    Wang, Z. Jane
    Chen, Xun
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [2] SymFormer: End-to-End Symbolic Regression Using Transformer-Based Architecture
    Vastl, Martin
    Kulhanek, Jonas
    Kubalik, Jiri
    Derner, Erik
    Babuska, Robert
    IEEE ACCESS, 2024, 12 : 37840 - 37849
  • [3] Transformer-Based End-to-End Classification of Variable-Length Volumetric Data
    Oghbaie, Marzieh
    Araujo, Teresa
    Emre, Taha
    Schmidt-Erfurth, Ursula
    Bogunovic, Hrvoje
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT VI, 2023, 14225 : 358 - 367
  • [4] HyperSFormer: A Transformer-Based End-to-End Hyperspectral Image Classification Method for Crop Classification
    Xie, Jiaxing
    Hua, Jiajun
    Chen, Shaonan
    Wu, Peiwen
    Gao, Peng
    Sun, Daozong
    Lyu, Zhendong
    Lyu, Shilei
    Xue, Xiuyun
    Lu, Jianqiang
    REMOTE SENSING, 2023, 15 (14)
  • [5] A Comparison of Transformer-Based Language Models on NLP Benchmarks
    Greco, Candida Maria
    Tagarelli, Andrea
    Zumpano, Ester
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 490 - 501
  • [6] End-to-End Asbestos Roof Detection on Orthophotos Using Transformer-Based YOLO Deep Neural Network
    Pace, Cesare Davide
    Bria, Alessandro
    Focareta, Mariano
    Lozupone, Gabriele
    Marrocco, Claudio
    Meoli, Giuseppe
    Molinara, Mario
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2023, PT I, 2023, 14233 : 232 - 244
  • [7] A transformer-based end-to-end data-driven model for multisensor time series monitoring of machine tool condition
    Agung', Oroko Joanes
    James, Kimotho
    Samuel, Kabini
    Evan, Murimi
    ENGINEERING REPORTS, 2023, 5 (05)
  • [8] End-to-End lightweight Transformer-Based neural network for grasp detection towards fruit robotic handling
    Guo, Congmin
    Zhu, Chenhao
    Liu, Yuchen
    Huang, Renjun
    Cao, Boyuan
    Zhu, Qingzhen
    Zhang, Ranxin
    Zhang, Baohua
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 221
  • [9] Measurement of Semantic Textual Similarity in Clinical Texts: Comparison of Transformer-Based Models
    Yang, Xi
    He, Xing
    Zhang, Hansi
    Ma, Yinghan
    Bian, Jiang
    Wu, Yonghui
    JMIR MEDICAL INFORMATICS, 2020, 8 (11)
  • [10] E2EET: from pipeline to end-to-end entity typing via transformer-based embeddings
    Stewart, Michael
    Liu, Wei
    KNOWLEDGE AND INFORMATION SYSTEMS, 2022, 64 (01) : 95 - 113