Semantic concept space based progressive transductive learning for text classification

被引:0
作者
Zhang, Xiaobin [1 ]
Yin, Yingshun [1 ]
Gao, Lili [1 ]
Zheng, Jing
Niu, Yanzhan
机构
[1] Xian Polytechn Univ, Sch Comp Sci, Xian 710048, Peoples R China
来源
RECENT ADVANCE OF CHINESE COMPUTING TECHNOLOGIES | 2007年
关键词
text classification; progressive tranductive support vector machines; kernelprinciple component analysis; kernelized Hebbian algorithm;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Though the learning ability of Progressive Tranductive Support Vector Machines may be independent of the dimensionality of feature space, dimension reduction is an essential issue to efficiently handle large scale terms in practical applications of text classification. A novel approach is proposed to reduce dimensions of documents and extract semantic concepts of words using Kernel Principle Component Analysis. The better performance are achieved via indirectly compute and store covariance matrix at each iteration. The experiment results of Reuters -21578 data show substantial improvements over progressive inductive methods in semantic concept space. It is very promising for a mixed training set of small labeled examples and a large number of unlabeled examples.
引用
收藏
页码:324 / 328
页数:5
相关论文
共 7 条
  • [1] [Anonymous], 2005, EUR C MACH LEARN
  • [2] CHEN YS, 2002, P IEEE ICDM
  • [3] CRISTIANINI N, 2001, P 18 INT C MACH LEAR
  • [4] Joachims Thorsten, 1999, INT C MACH LEARN
  • [5] JOHN ST, 2004, KERNEL METHODS PATTE, P327
  • [6] KIM KI, 2003, K HEBBIAN ALGORITHM
  • [7] Yang YM, 1999, SIGIR'99: PROCEEDINGS OF 22ND INTERNATIONAL CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, P42, DOI 10.1145/312624.312647