Robust multitask learning in high dimensions under memory constraint

被引:0
作者
Chen, Canyi [1 ]
Chen, Bingzhen [1 ]
Kong, Lingchen [2 ]
Zhu, Liping [1 ]
机构
[1] Renmin Univ China, Inst Stat & Big Data, Beijing, Peoples R China
[2] Beijing Jiaotong Univ, Sch Sci, Beijing, Peoples R China
关键词
distributed estimation; high dimension; multitask learning; oracle; robustness; QUANTILE REGRESSION; FRAMEWORK; SELECTION; RECOVERY; SPARSITY;
D O I
10.1002/sam.11700
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate multitask learning in the context of multivariate linear regression with high dimensional covariates and heavy-tailed noise, while under the constraint of limited memory. To tackle the computational complexity arising from the non-smoothness of the quantile loss, we reformulate it as an equivalent least squares loss, which yields robust solutions even in the presence of heavy-tailed noise. We incorporate a group lasso penalty into the quantile loss to produce sparse solutions, and an accelerated proximal sub-gradient descent algorithm to speed up the computation while ensuring explicit forms for penalized solutions at each iteration. The proposed algorithm is general and can be applied to similar optimization problems. Moreover, we introduce a communication-efficient distributed algorithm that guarantees optimal convergence rates after finite communication rounds in cases where computing resources such as memory are insufficient. We also study the theoretical properties of the resultant estimate and relax the widely used model selection consistency assumption on the initial estimate. We demonstrate the effectiveness of our proposal through extensive numerical studies.
引用
收藏
页数:16
相关论文
共 47 条
[1]  
[Anonymous], 2008, P 25 INT C MACH LEAR
[2]   DIVIDE AND CONQUER IN NONSTANDARD PROBLEMS AND THE SUPER-EFFICIENCY PHENOMENON [J].
Banerjee, Moulinath ;
Durot, Cecile ;
Sen, Bodhisattva .
ANNALS OF STATISTICS, 2019, 47 (02) :720-757
[3]   DISTRIBUTED TESTING AND ESTIMATION UNDER SPARSE HIGH DIMENSIONAL MODELS [J].
Battey, Heather ;
Fan, Jianqing ;
Liu, Han ;
Lu, Junwei ;
Zhu, Ziwei .
ANNALS OF STATISTICS, 2018, 46 (03) :1352-1382
[4]   l1-PENALIZED QUANTILE REGRESSION IN HIGH-DIMENSIONAL SPARSE MODELS [J].
Belloni, Alexandre ;
Chernozhukov, Victor .
ANNALS OF STATISTICS, 2011, 39 (01) :82-130
[5]  
Bennett James, 2007, P KDD CUP WORKSH NEW
[6]  
Boyd S., 2004, Convex Optimization
[7]   Communication Lower Bounds for Statistical Estimation Problems via a Distributed Data Processing Inequality [J].
Braverman, Mark ;
Garg, Ankit ;
Ma, Tengyu ;
Nguyen, Huy L. ;
Woodruff, David P. .
STOC'16: PROCEEDINGS OF THE 48TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING, 2016, :1011-1020
[8]   Matrix Completion With Noise [J].
Candes, Emmanuel J. ;
Plan, Yaniv .
PROCEEDINGS OF THE IEEE, 2010, 98 (06) :925-936
[9]   Robust reduced rank regression in a distributed setting [J].
Chen, Xi ;
Liu, Weidong ;
Mao, Xiaojun .
SCIENCE CHINA-MATHEMATICS, 2022, 65 (08) :1707-1730
[10]  
Chen X, 2020, J MACH LEARN RES, V21