Fast and accurate pseudoinverse with sparse matrix reordering and incremental approach

被引:0
作者
Jinhong Jung
Lee Sael
机构
[1] Jeonbuk National University,
[2] Ajou University,undefined
来源
Machine Learning | 2020年 / 109卷
关键词
Pseudoinverse; Sparse matrix reordering; Incremental SVD; Multi-label linear regression;
D O I
暂无
中图分类号
学科分类号
摘要
How can we compute the pseudoinverse of a sparse feature matrix efficiently and accurately for solving optimization problems? A pseudoinverse is a generalization of a matrix inverse, which has been extensively utilized as a fundamental building block for solving linear systems in machine learning. However, an approximate computation, let alone an exact computation, of pseudoinverse is very time-consuming due to its demanding time complexity, which limits it from being applied to large data. In this paper, we propose FastPI (Fast PseudoInverse), a novel incremental singular value decomposition (SVD) based pseudoinverse method for sparse matrices. Based on the observation that many real-world feature matrices are sparse and highly skewed, FastPI reorders and divides the feature matrix and incrementally computes low-rank SVD from the divided components. To show the efficacy of proposed FastPI, we apply them in real-world multi-label linear regression problems. Through extensive experiments, we demonstrate that FastPI computes the pseudoinverse faster than other approximate methods without loss of accuracy. Results imply that our method efficiently computes the low-rank pseudoinverse of a large and sparse matrix that other existing methods cannot handle with limited time and space.
引用
收藏
页码:2333 / 2347
页数:14
相关论文
共 32 条
[1]  
Baglama J(2005)Augmented implicitly restarted lanczos bidiagonalization methods SIAM Journal on Scientific Computing 27 19-42
[2]  
Reichel L(1996)Efficient algorithms for computing a strong rank-revealing QR factorization SIAM Journal on Scientific Computing 17 848-869
[3]  
Gu M(2011)Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions SIAM Review 53 217-288
[4]  
Eisenstat SC(2016)Novel applications of multitask learning and multiple output regression to multiple genetic trait prediction Bioinformatics 32 i37-i43
[5]  
Halko N(2013)Robust extreme learning machine Neurocomputing 102 31-44
[6]  
Martinsson PG(2016)Random walk with restart on large graphs using block elimination ACM Transactions on Database Systems (TODS) 41 1-43
[7]  
Tropp JA(2004)Rcv1: A new benchmark collection for text categorization research Journal of Machine Learning Research 5 361-397
[8]  
He D(2014)Slashburn: Graph compression and mining beyond caveman communities IEEE Transactions on Knowledge and Data Engineering 26 3077-3089
[9]  
Kuhn D(2008)Incremental learning for robust visual tracking International Journal of Computer Vision 77 125-141
[10]  
Parida L(2016)Multi-target regression via input space expansion: Treating targets as inputs Machine Learning 104 55-98