A Comparison of Optimization Methods and Software for Large-scale L1-regularized Linear Classification

被引:0
作者
Yuan, Guo-Xun [1 ]
Chang, Kai-Wei [1 ]
Hsieh, Cho-Jui [1 ]
Lin, Chih-Jen [1 ]
机构
[1] Natl Taiwan Univ, Dept Comp Sci, Taipei 106, Taiwan
关键词
L1; regularization; linear classification; optimization methods; logistic regression; support vector machines; document classification; NEWTON METHOD; LOGISTIC-REGRESSION; FEATURE-SELECTION; ALGORITHM; DESCENT; GRADIENT; PROJECTION; LASSO;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large-scale linear classification is widely used in many areas. The L1-regularized form can be applied for feature selection; however, its non-differentiability causes more difficulties in training. Although various optimization methods have been proposed in recent years, these have not yet been compared suitably. In this paper, we first broadly review existing methods. Then, we discuss state-of-the-art software packages in detail and propose two efficient implementations. Extensive comparisons indicate that carefully implemented coordinate descent methods are very suitable for training large document data.
引用
收藏
页码:3183 / 3234
页数:52
相关论文
共 70 条
[1]  
Andrew G A G J, 2007, P 24 INT C MACH LEAR
[2]  
[Anonymous], COMPUTATION IN PRESS
[3]  
[Anonymous], ALGORITHMS SPARSE LI
[4]  
[Anonymous], 2011, ACM T INTEL SYST TEC
[5]  
[Anonymous], MCSP9090901 ARG NAT
[6]  
[Anonymous], IEEE J SEL TOP SIGNA
[7]  
[Anonymous], P 26 INT C MACH LEAR
[8]  
[Anonymous], WALK 2 NORM SVM 1 NO
[9]  
[Anonymous], 2003, INTRO LECT CONVEX OP
[10]  
[Anonymous], P ANN M ASS COMP LIN