A data-centric review of deep transfer learning with applications to text data

被引:45
作者
Bashath, Samar [1 ]
Perera, Nadeesha [1 ]
Tripathi, Shailesh [1 ]
Manjang, Kalifa [1 ]
Dehmer, Matthias [2 ,3 ,4 ,5 ]
Streib, Frank Emmert [1 ]
机构
[1] Tampere Univ, Predict Soc & Data Analyt Lab, Korkeakoulunkatu 10, Tampere 33720, Finland
[2] Swiss Distance Univ Appl Sci, Dept Comp Sci, Brig, Switzerland
[3] Xian Technol Univ, Sch Sci, Xian, Peoples R China
[4] Nankai Univ, Coll Artificial Intelligence, Tianjin, Peoples R China
[5] Hlth & Life Sci Univ, Dept Biomed Comp Sci & Mechatron, UMIT, Hall In Tirol, Austria
关键词
Transfer learning; Deep learning; Natural language processing; Machine learning; Domain adaptation; SENTIMENT ANALYSIS; NETWORK;
D O I
10.1016/j.ins.2021.11.061
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, many applications are using various forms of deep learning models. Such methods are usually based on traditional learning paradigms requiring the consistency of properties among the feature spaces of the training and test data and also the availability of large amounts of training data, e.g., for performing supervised learning tasks. However, many real-world data do not adhere to such assumptions. In such situations transfer learning can provide feasible solutions, e.g., by simultaneously learning from data-rich source data and data-sparse target data to transfer information for learning a target task. In this paper, we survey deep transfer learning models with a focus on applications to text data. First, we review the terminology used in the literature and introduce a new nomenclature allowing the unequivocal description of a transfer learning model. Second, we introduce a visual taxonomy of deep learning approaches that provides a systematic structure to the many diverse models introduced until now. Furthermore, we provide comprehensive information about text data that have been used for studying such models because only by the application of methods to data, performance measures can be estimated and models assessed. (c) 2021 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页码:498 / 528
页数:31
相关论文
共 156 条
[1]   EmoNet: Fine-Grained Emotion Detection with Gated Recurrent Neural Networks [J].
Abdul-Mageed, Muhammad ;
Ungar, Lyle .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :718-728
[2]  
Ajakan Hana, DOMAIN ADVERSARIAL N
[3]   Approaches to Cross-Domain Sentiment Analysis: A Systematic Literature Review [J].
Al-Moslmi, Tareq ;
Omar, Nazlia ;
Abdullah, Salwani ;
Albared, Mohammed .
IEEE ACCESS, 2017, 5 :16173-16192
[4]  
Alam F, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P1077
[5]   A State-of-the-Art Survey on Deep Learning Theory and Architectures [J].
Alom, Md Zahangir ;
Taha, Tarek M. ;
Yakopcic, Chris ;
Westberg, Stefan ;
Sidike, Paheding ;
Nasrin, Mst Shamima ;
Hasan, Mahmudul ;
Van Essen, Brian C. ;
Awwal, Abdul A. S. ;
Asari, Vijayan K. .
ELECTRONICS, 2019, 8 (03)
[6]   A UNIVERSAL THEOREM ON LEARNING-CURVES [J].
AMARI, SI .
NEURAL NETWORKS, 1993, 6 (02) :161-166
[7]  
Amini M.-R., 2009, NIPS, P28
[8]  
Andrew G., 2013, ICML
[9]  
[Anonymous], 2015, INT C LEARNING REPRE
[10]  
[Anonymous], 2012, P 29 INT C MACH LEAR