Multi-task Supervised Learning via Cross-learning

被引:0
作者
Cervino, Juan [1 ]
Andres Bazerque, Juan [2 ]
Calvo-Fullana, Miguel [3 ]
Ribeiro, Alejandro [1 ]
机构
[1] Univ Penn, Philadelphia, PA 19104 USA
[2] Univ Republica, Montevideo, Uruguay
[3] MIT, Cambridge, MA 02139 USA
来源
29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021) | 2021年
关键词
Supervised learning; multi-task learning; optimization;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper we consider a problem known as multi-task learning, consisting of fitting a set of classifier or regression functions intended for solving different tasks. In our novel formulation, we couple the parameters of these functions, so that they learn in their task specific domains while staying close to each other. This facilitates cross-fertilization in which data collected across different domains help improving the learning performance at each other task. First, we present a simplified case in which the goal is to estimate the means of two Gaussian variables, for the purpose of gaining some insights on the advantage of the proposed cross-learning strategy. Then we provide a stochastic projected gradient algorithm to perform cross-learning over a generic loss function. If the number of parameters is large, then the projection step becomes computationally expensive. To avoid this situation, we derive a primal-dual algorithm that exploits the structure of the dual problem, achieving a formulation whose complexity only depends on the number of tasks. Preliminary numerical experiments for image classification by neural networks trained on a dataset divided in different domains corroborate that the cross-learned function outperforms both the task-specific and the consensus approaches.
引用
收藏
页码:1381 / 1385
页数:5
相关论文
共 24 条
[1]  
Agarwal Arvind., 2010, NIPS. Ed. by, P46
[2]   Convex multi-task feature learning [J].
Argyriou, Andreas ;
Evgeniou, Theodoros ;
Pontil, Massimiliano .
MACHINE LEARNING, 2008, 73 (03) :243-272
[3]   Exploiting task relatedness for multiple task learning [J].
Ben-David, S ;
Schuller, R .
LEARNING THEORY AND KERNEL MACHINES, 2003, 2777 :567-580
[4]  
Bertsekas D., 1999, Athena Scientific Optimization and Computation Series, V2nd
[5]  
Boyd S., 2009, CONVEX OPTIMIZATION
[6]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[7]  
Cervino J., 2020, ARXIV PREPRINT ARXIV
[8]   Meta-Learning through Coupled Optimization in Reproducing Kernel Hilbert Spaces [J].
Cervino, Juan ;
Bazerque, Juan Andres ;
Calvo-Fullana, Miguel ;
Ribeiro, Alejandro .
2019 AMERICAN CONTROL CONFERENCE (ACC), 2019, :4840-4846
[9]  
Chamon LFO, 2020, INT CONF ACOUST SPEE, P8374, DOI [10.1109/icassp40776.2020.9054128, 10.1109/ICASSP40776.2020.9054128]
[10]  
Dong DX, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P1723