SRDA: An efficient algorithm for large-scale discriminant analysis

被引:328
作者
Cai, Deng
He, Xiaofei
Han, Jiawei
机构
[1] Univ Illinois, Dept Comp Sci, Urbana, IL 61801 USA
[2] Yahoo, Burbank, CA 91504 USA
基金
美国国家科学基金会;
关键词
linear discriminant analysis; spectral regression; dimensionality reduction;
D O I
10.1109/TKDE.2007.190669
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Linear Discriminant Analysis (LDA) has been a popular method for extracting features that preserves class separability. The projection functions of LDA are commonly obtained by maximizing the between-class covariance and simultaneously minimizing the within-class covariance. It has been widely used in many fields of information processing, such as machine learning, data mining, information retrieval, and pattern recognition. However, the computation of LDA involves dense matrices eigendecomposition, which can be computationally expensive in both time and memory. Specifically, LDA has O(mnt + t(3) )time complexity and requires O(mn + mt + nt) memory, where m is the number of samples, n is the number of features, and t = min(m,n). When both m and n are large, it is infeasible to apply LDA. In this paper, we propose a novel algorithm for discriminant analysis, called Spectral Regression Discriminant Analysis (SRDA). By using spectral graph analysis, SRDA casts discriminant analysis into a regression framework that facilitates both efficient computation and the use of regularization techniques. Specifically, SRDA only needs to solve a set of regularized least squares problems, and there is no eigenvector computation involved, which is a huge save of both time and memory. Our theoretical analysis shows that SRDA can be computed with O(ms) time and O(ms) memory, where s(<= n) n is the average number of nonzero features in each sample. Extensive experimental results on four real-world data sets demonstrate the effectiveness and efficiency of our algorithm.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 27 条
  • [1] [Anonymous], 2006, REGULARIZED DISCRIMI
  • [2] [Anonymous], 2007, P 24 INT C MACH LEAR
  • [3] [Anonymous], 1997, CBMS REG C SER MATH
  • [4] Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection
    Belhumeur, PN
    Hespanha, JP
    Kriegman, DJ
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (07) : 711 - 720
  • [5] Cai D., 2007, P 11 INT C COMP VIS
  • [6] CAI D, 2007, P 16 ACM INT C INF K
  • [7] Cai D., 2007, P INT C DAT MIN ICDM
  • [8] Cai D., 2007, P ACM C MULT
  • [9] Duda R. O., 1973, Pattern Classification
  • [10] Friedman J, 2001, The elements of statistical learning, V1, DOI DOI 10.1007/978-0-387-21606-5