End-to-End Transformer-Based Models in Textual-Based NLP

被引:32
|
作者
Rahali, Abir [1 ]
Akhloufi, Moulay A. [1 ]
机构
[1] Univ Moncton, Dept Comp Sci, Percept Robot & Intelligent Machines Res Grp PRIME, Moncton, NB E1A 3E9, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Transformers; deep learning; natural language processing; transfer learning; PRE-TRAINED BERT; PREDICTION; SYSTEMS;
D O I
10.3390/ai4010004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer architectures are highly expressive because they use self-attention mechanisms to encode long-range dependencies in the input sequences. In this paper, we present a literature review on Transformer-based (TB) models, providing a detailed overview of each model in comparison to the Transformer's standard architecture. This survey focuses on TB models used in the field of Natural Language Processing (NLP) for textual-based tasks. We begin with an overview of the fundamental concepts at the heart of the success of these models. Then, we classify them based on their architecture and training mode. We compare the advantages and disadvantages of popular techniques in terms of architectural design and experimental value. Finally, we discuss open research, directions, and potential future work to help solve current TB application challenges in NLP.
引用
收藏
页码:54 / 110
页数:57
相关论文
共 50 条
  • [31] End-to-End Structural analysis in civil engineering based on deep learning
    Wang, Chen
    Song, Ling-han
    Fan, Jian-sheng
    AUTOMATION IN CONSTRUCTION, 2022, 138
  • [32] Deep Learning Assisted End-to-End Synthesis of mm-Wave Passive Networks with 3D EM Structures: A Study on A Transformer-Based Matching Network
    Er, Siawpeng
    Liu, Edward
    Chen, Minshuo
    Li, Yan
    Liu, Yuqi
    Zhao, Tuo
    Wang, Hua
    2021 IEEE MTT-S INTERNATIONAL MICROWAVE SYMPOSIUM (IMS), 2021, : 66 - 69
  • [33] End-to-End Session-Based Recommendation on GPU
    Pereira Moreira, Gabriel de Souza
    Rabhi, Sara
    Ak, Ronay
    Schifferer, Benedikt
    15TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS 2021), 2021, : 831 - 833
  • [34] End-to-End Computational Imaging Based on Multispectral Fusion
    Nan Yilan
    Shen Junfei
    Zhang Qican
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (20)
  • [35] End-to-End Deblending of Simultaneous Source Data Using Transformer
    Zu, Shaohuan
    Ke, Chaofan
    Hou, Chengzhi
    Cao, Junxing
    Zhang, Hongjing
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [36] CTsynther: Contrastive Transformer Model for End-to-End Retrosynthesis Prediction
    Lu, Hao
    Wei, Zhiqiang
    Zhang, Kun
    Wang, Xuze
    Ali, Liaqat
    Liu, Hao
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2024, 21 (06) : 2235 - 2245
  • [37] Tweets Topic Classification and Sentiment Analysis Based on Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2023, 10 (02) : 117 - 134
  • [38] SGTR plus : End-to-End Scene Graph Generation With Transformer
    Li, Rongjie
    Zhang, Songyang
    He, Xuming
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (04) : 2191 - 2205
  • [39] eval-rationales: An End-to-End Toolkit to Explain and Evaluate Transformers-Based Models
    Maachou, Khalil
    Lovon-Melgarejo, Jesus
    Moreno, Jose G.
    Tamine, Lynda
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT V, 2024, 14612 : 212 - 217
  • [40] End-to-end tire defect detection model based on transfer learning techniques
    Saleh R.A.A.
    Konyar M.Z.
    Kaplan K.
    Ertunç H.M.
    Neural Computing and Applications, 2024, 36 (20) : 12483 - 12503