Distributed Learning with Regularized Least Squares

被引:0
作者
Lin, Shao-Bo [1 ]
Guo, Xin [2 ]
Zhou, Ding-Xuan [1 ]
机构
[1] City Univ Hong Kong, Dept Math, Tat Chee Ave, Kowloon, Hong Kong, Peoples R China
[2] Hong Kong Polytech Univ, Dept Appl Math, Kowloon, Hong Kong, Peoples R China
关键词
Distributed learning; divide-and-conquer; error analysis; integral operator; second order decomposition; KERNEL; ALGORITHMS; REGRESSION; RATES; OPERATORS; NETWORKS; GRADIENT; THEOREM;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We study distributed learning with the least squares regularization scheme in a reproducing kernel Hilbert space (RKHS). By a divide-and-conquer approach, the algorithm partitions a data set into disjoint data subsets, applies the least squares regularization scheme to each data subset to produce an output function, and then takes an average of the individual output functions as a final global estimator or predictor. We show with error bounds and learning rates in expectation in both the L-2-metric and RKHS-metric that the global output function of this distributed learning is a good approximation to the algorithm processing the whole data in one single machine. Our derived learning rates in expectation are optimal and stated in a general setting without any eigenfunction assumption. The analysis is achieved by a novel second order decomposition of operator differences in our integral operator approach. Even for the classical least squares regularization scheme in the RKHS associated with a general kernel, we give the best learning rate in expectation in the literature.
引用
收藏
页数:31
相关论文
共 47 条
[1]  
Bach F., 2013, PMLR, P185
[2]   On regularization algorithms in learning theory [J].
Bauer, Frank ;
Pereverzev, Sergei ;
Rosasco, Lorenzo .
JOURNAL OF COMPLEXITY, 2007, 23 (01) :52-72
[3]  
Blanchard G., 2010, NIPS, V23, P226
[4]   Convergence rates of Kernel Conjugate Gradient for random design regression [J].
Blanchard, Gilles ;
Kraemer, Nicole .
ANALYSIS AND APPLICATIONS, 2016, 14 (06) :763-794
[5]  
Caponnetto A, 2007, FOUND COMPUT MATH, V7, P331, DOI 10.1007/S10208-006-0196-8
[6]   CROSS-VALIDATION BASED ADAPTATION FOR REGULARIZATION OPERATORS IN LEARNING THEORY [J].
Caponnetto, Andrea ;
Yao, Yuan .
ANALYSIS AND APPLICATIONS, 2010, 8 (02) :161-183
[7]  
Chen DR, 2004, J MACH LEARN RES, V5, P1143
[8]   Model selection for regularized least-squares algorithm in learning theory [J].
De Vito, E ;
Caponnetto, A ;
Rosasco, L .
FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2005, 5 (01) :59-85
[9]   Adaptive Kernel Methods Using the Balancing Principle [J].
De Vito, E. ;
Pereverzyev, S. ;
Rosasco, L. .
FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2010, 10 (04) :455-479
[10]   An extension of Mercer theorem to matrix-valued measurable kernels [J].
De Vito, Ernesto ;
Umanita, Veronica ;
Villa, Silvia .
APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2013, 34 (03) :339-351