Accommodating Multiple Tasks' Disparities With Distributed Knowledge-Sharing Mechanism

被引:5
作者
Chen, Xinqi [1 ,2 ]
Zhou, Guoxu [1 ,3 ]
Wang, Yanjiao [1 ,4 ]
Hou, Ming [5 ]
Zhao, Qibin [1 ,5 ,6 ]
Xie, Shengli [1 ,4 ]
机构
[1] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[2] Guangdong Univ Technol, Guangdong Key Lab IoT Informat Technol, Guangzhou 510006, Peoples R China
[3] Minist Educ, Key Lab Intelligent Detect & Internet Things Mfg, Guangzhou 510006, Peoples R China
[4] Guangdong Univ Technol, Guangdong Hong Kong Macao Joint Lab Smart Discret, Guangzhou 510006, Peoples R China
[5] RIKEN, Tensor Learning Team, Ctr Adv Intelligence Project, Tokyo 1030027, Japan
[6] Minist Educ, Joint Int Res Lab Intelligent Informat Proc & Sys, Guangzhou 510006, Peoples R China
基金
日本学术振兴会;
关键词
Task analysis; Tensors; Knowledge engineering; Feature extraction; Complexity theory; Automation; Network architecture; Heterogenous network architectures; multitask learning; tensor decomposition; MULTITASK; DECOMPOSITION; REDUCTION; IMAGE; RANK;
D O I
10.1109/TCYB.2020.3002911
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep multitask learning (MTL) shares beneficial knowledge across participating tasks, alleviating the impacts of extreme learning conditions on their performances such as the data scarcity problem. In practice, participators stemming from different domain sources often have varied complexities and input sizes, for example, in the joint learning of computer vision tasks with RGB and grayscale images. For adapting to these differences, it is appropriate to design networks with proper representational capacities and construct neural layers with corresponding widths. Nevertheless, most of the state-of-the-art methods pay little attention to such situations, and actually fail to handle the disparities. To work with the dissimilitude of tasks' network designs, this article presents a distributed knowledge-sharing framework called tensor ring multitask learning (TRMTL), in which the relationship between knowledge sharing and original weight matrices is cut up. The framework of TRMTL is flexible, which is not only capable of sharing knowledge across heterogenous networks but also able to jointly learn tasks with varied input sizes, significantly improving performances of data-insufficient tasks. Comprehensive experiments on challenging datasets are conducted to empirically validate the effectiveness, efficiency, and flexibility of TRMTL in dealing with the disparities in MTL.
引用
收藏
页码:2440 / 2452
页数:13
相关论文
共 52 条
[1]  
[Anonymous], 2009, CIFAR-100 Dataset
[2]  
[Anonymous], 2018, LEARNING TASK GROUPI
[3]  
[Anonymous], 2011, Proc. Adv. Neural Inf. Process. Syst.
[4]  
[Anonymous], 2016, arXiv
[5]   Convex multi-task feature learning [J].
Argyriou, Andreas ;
Evgeniou, Theodoros ;
Pontil, Massimiliano .
MACHINE LEARNING, 2008, 73 (03) :243-272
[6]   Efficient Tensor Completion for Color Image and Video Recovery: Low-Rank Tensor Train [J].
Bengua, Johann A. ;
Phien, Ho N. ;
Hoang Duong Tuan ;
Do, Minh N. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (05) :2466-2479
[7]   A Convex Formulation for Learning a Shared Predictive Structure from Multiple Tasks [J].
Chen, Jianhui ;
Tang, Lei ;
Liu, Jun ;
Ye, Jieping .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (05) :1025-1038
[8]   On the best rank-1 and rank-(R1,R2,...,RN) approximation of higher-order tensors [J].
De Lathauwer, L ;
De Moor, B ;
Vandewalle, J .
SIAM JOURNAL ON MATRIX ANALYSIS AND APPLICATIONS, 2000, 21 (04) :1324-1342
[9]   A link between the canonical decomposition in multilinear algebra and simultaneous matrix diagonalization [J].
De Lathauwer, Lieven .
SIAM JOURNAL ON MATRIX ANALYSIS AND APPLICATIONS, 2006, 28 (03) :642-666
[10]  
Denil M., 2013, ADV NEURAL INFORM PR, P2148