A Sparse L2-Regularized Support Vector Machines for Large-Scale Natural Language Learning

被引:0
作者
Wu, Yu-Chieh [1 ,2 ]
Lee, Yue-Shi [4 ]
Yang, Jie-Chi [3 ]
Yen, Show-Jane [4 ]
机构
[1] Ming Chuan Univ, Dept Finance, 250 Zhong Shan N Rd,Sec 5, Taipei 111, Taiwan
[2] Ming Chuan Univ, Sch Commun, Taipei 111, Taiwan
[3] Natl Cent Univ, Grad Inst Network Learning Technol, Taoyuan 32001, Taiwan
[4] Ming Chuan Univ, Dept Comp Sci & Informat Engn, Taoyuan 333, Taiwan
来源
INFORMATION RETRIEVAL TECHNOLOGY | 2010年 / 6458卷
关键词
L-2-regularization; part-of-speech tagging; support vector machines; machine learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Linear support vector machines (SVMs) have become one of the most prominent classification algorithms for many natural language learning problems such as sequential labeling tasks. Even though the L-2-regularized SVMs yields slightly more superior accuracy than L-1-SVM, it produces too much near but non zero feature weights. In this paper, we present a cutting-weight algorithm to guide the optimization process of L-2-SVM into sparse solution. To verify the proposed method, we conduct the experiments with three well-known sequential labeling tasks and one dependency parsing task. The result shows that our method achieved at least 400% feature parameter reduction rates in comparison to the original L-2-SVM, with almost no change in accuracy and training times. In terms of run time efficiency, our method is faster than the original L-2-regularized SVMs at least 20% in all tasks.
引用
收藏
页码:340 / +
页数:3
相关论文
共 18 条
[1]  
[Anonymous], 2001, P 18 INT C MACH LEAR
[2]  
[Anonymous], 2006, P ACMSIGKDD INT C KN
[3]  
[Anonymous], 2008, Introduction to information retrieval
[4]  
[Anonymous], 2008, P 14 ACM SIGKDD INT
[5]  
Collins M, 2002, PROCEEDINGS OF THE 2002 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, P1
[6]   Fast cg-based methods for Tikhonov-Phillips regularization [J].
Frommer, A ;
Maass, P .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 1999, 20 (05) :1831-1850
[7]  
GAO J, 2007, 45 ANN M ASS COMP LI, P824
[8]  
HSIEH CJ, 2008, 15 INT C MACH LEARN, P408
[9]  
Keerthi SS, 2005, J MACH LEARN RES, V6, P341
[10]  
Kudo T, 2003, 41ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, P24