Natural language processing with transformers: a review

被引:1
作者
Tucudean, Georgiana [1 ]
Bucos, Marian [1 ]
Dragulescu, Bogdan [1 ]
Caleanu, Catalin Daniel [2 ]
机构
[1] Politehn Univ Timisoara, Commun Dept, Timisoara, Timis, Romania
[2] Politehn Univ Timisoara, Appl Elect Dept, Timisoara, Timis, Romania
关键词
Transfomers; Natural language processing; Deep neural network architectures; Review; Trends; TEXT;
D O I
10.7717/peerj-cs.2222
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural language processing (NLP) tasks can be addressed with several deep learning architectures, and many different approaches have proven to be efficient. fi cient. This study aims to briefly fl y summarize the use cases for NLP tasks along with the main architectures. This research presents transformer-based solutions for NLP tasks such as Bidirectional Encoder Representations from Transformers (BERT), and Generative Pre-Training (GPT) architectures. To achieve that, we conducted a stepby-step process in the review strategy: identify the recent studies that include Transformers, apply fi lters to extract the most consistent studies, identify and define fi ne inclusion and exclusion criteria, assess the strategy proposed in each study, and fi nally discuss the methods and architectures presented in the resulting articles. These steps facilitated the systematic summarization and comparative analysis of NLP applications based on Transformer architectures. The primary focus is the current state of the NLP domain, particularly regarding its applications, language models, and data set types. The results provide insights into the challenges encountered in this research domain.
引用
收藏
页数:22
相关论文
共 50 条
[1]   Transformer models for text-based emotion detection: a review of BERT-based approaches [J].
Acheampong, Francisca Adoma ;
Nunoo-Mensah, Henry ;
Chen, Wenyu .
ARTIFICIAL INTELLIGENCE REVIEW, 2021, 54 (08) :5789-5829
[2]   Arabic Fake News Detection: Comparative Study of Neural Networks and Transformer-Based Approaches [J].
Al-Yahya, Maha ;
Al-Khalifa, Hend ;
Al-Baity, Heyam ;
AlSaeed, Duaa ;
Essam, Amr .
COMPLEXITY, 2021, 2021
[3]   Combat COVID-19 infodemic using explainable natural language processing models [J].
Ayoub, Jackie ;
Yang, X. Jessie ;
Zhou, Feng .
INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (04)
[4]   MolGPT: Molecular Generation Using a Transformer-Decoder Model [J].
Bagal, Viraj ;
Aggarwal, Rishal ;
Vinod, P. K. ;
Priyakumar, U. Deva .
JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2022, 62 (09) :2064-2076
[5]  
Bakker C., 2023, Hypothesis: Research Journal for Health Information Professionals, V35, P26528, DOI [10.18060/26528, DOI 10.18060/26528]
[6]   To BERT or Not To BERT: Comparing Speech and Language-based Approaches for Alzheimer's Disease Detection [J].
Balagopalan, Aparna ;
Eyre, Benjamin ;
Rudzicz, Frank ;
Novikova, Jekaterina .
INTERSPEECH 2020, 2020, :2167-2171
[7]   Taming Pretrained Transformers for Extreme Multi-label Text Classification [J].
Chang, Wei-Cheng ;
Yu, Hsiang-Fu ;
Zhong, Kai ;
Yang, Yiming ;
Dhillon, Inderjit S. .
KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, :3163-3171
[8]   Comparing deep learning architectures for sentiment analysis on drug reviews [J].
Colon-Ruiz, Cristobal ;
Segura-Bedmar, Isabel .
JOURNAL OF BIOMEDICAL INFORMATICS, 2020, 110
[9]  
Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
[10]   Evaluation of the benchmark datasets for testing the efficacy of deep convolutional neural networks [J].
Dhar, Sanchari ;
Shamir, Lior .
VISUAL INFORMATICS, 2021, 5 (03) :92-101