ASPDC: Accelerated SPDC Regularized Empirical Risk Minimization for Ill-Conditioned Problems in Large-Scale Machine Learning

被引:0
作者
Liang, Haobang [1 ]
Cai, Hao [2 ]
Wu, Hejun [3 ]
Shang, Fanhua [4 ]
Cheng, James [5 ]
Li, Xiying [6 ]
机构
[1] Sun Yat Sen Univ, Sch Biomed Engn, Guangzhou 510006, Peoples R China
[2] Shantou Univ, Coll Engn, Shantou 515041, Peoples R China
[3] Sun Yat Sen Univ, Dept Comp Sci, Guangzhou 510006, Peoples R China
[4] Xidian Univ, Sch Artificial Intelligence, Xian 710071, Peoples R China
[5] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Hong Kong, Peoples R China
[6] Sun Yat Sen Univ, Sch Intelligent Syst Engn, Guangzhou 510006, Peoples R China
关键词
stochastic optimization; machine learning; empirical risk minimization; coordinate ascent algorithm; primal-dual algorithm; strongly convex and smooth;
D O I
10.3390/electronics11152382
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper aims to improve the response speed of SPDC (stochastic primal-dual coordinate ascent) in large-scale machine learning, as the complexity of per-iteration of SPDC is not satisfactory. We propose an accelerated stochastic primal-dual coordinate ascent called ASPDC and its further accelerated variant, ASPDC-i. Our proposed ASPDC methods achieve a good balance between low per-iteration computation complexity and fast convergence speed, even when the condition number becomes very large. The large condition number causes ill-conditioned problems, which usually requires many more iterations before convergence and longer per-iteration times in data training for machine learning. We performed experiments on various machine learning problems. The experimental results demonstrate that ASPDC and ASPDC-i converge faster than their counterparts, and enjoy low per-iteration complexity as well.
引用
收藏
页数:18
相关论文
共 25 条
[1]   Katyusha: The First Direct Acceleration of Stochastic Gradient Methods [J].
Allen-Zhu, Zeyuan .
STOC'17: PROCEEDINGS OF THE 49TH ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING, 2017, :1200-1205
[2]   A Computational Model for Path Loss in Wireless Sensor Networks in Orchard Environments [J].
Anastassiu, Hristos T. ;
Vougioukas, Stavros ;
Fronimos, Theodoros ;
Regen, Christian ;
Petrou, Loukas ;
Zude, Manuela ;
Kaethner, Jana .
SENSORS, 2014, 14 (03) :5118-5135
[3]  
Bauschke HH, 2011, CMS BOOKS MATH, P1, DOI 10.1007/978-1-4419-9467-7
[4]  
Bertsekas D.P., 2009, Convex Optimization Theory
[5]  
Chang KW, 2008, J MACH LEARN RES, V9, P1369
[6]   Parallel Dual Coordinate Descent Method for Large-scale Linear Classification in Multi-core Environments [J].
Chiang, Wei-Lin ;
Lee, Mu-Chu ;
Lin, Chih-Jen .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :1485-1494
[7]   An iterative algorithm for solving ill-conditioned linear least squares problems [J].
Deng Xingsheng ;
Yin Liangbo ;
Peng Sichun ;
Ding Meiqing .
GEODESY AND GEODYNAMICS, 2015, 6 (06) :453-459
[8]   First-order methods of smooth convex optimization with inexact oracle [J].
Devolder, Olivier ;
Glineur, Francois ;
Nesterov, Yurii .
MATHEMATICAL PROGRAMMING, 2014, 146 (1-2) :37-75
[9]  
Frostig R, 2015, PR MACH LEARN RES, V37, P2540
[10]   NEW PROXIMAL POINT ALGORITHMS FOR CONVEX MINIMIZATION [J].
Gueler, Osman .
SIAM JOURNAL ON OPTIMIZATION, 1992, 2 (04) :649-664