Tri-regularized nonnegative matrix tri-factorization for co-clustering

被引:29
作者
Deng, Ping [1 ,2 ]
Li, Tianrui [1 ,2 ]
Wang, Hongjun [1 ,2 ]
Horng, Shi-Jinn [3 ]
Yu, Zeng [1 ,2 ]
Wang, Xiaomin [1 ,2 ]
机构
[1] Southwest Jiaotong Univ, Sch Informat Sci & Technol, Chengdu 611756, Peoples R China
[2] Southwest Jiaotong Univ, Inst Artificial Intelligence, Chengdu 611756, Peoples R China
[3] Natl Taiwan Univ Sci & Technol, Dept Comp Sci & Informat Engn, Taipei 106, Taiwan
基金
国家重点研发计划;
关键词
Nonnegative matrix tri-factorization; Graph regularization; Entrywise norm; Sparsity; Co-clustering; SPARSITY;
D O I
10.1016/j.knosys.2021.107101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The objective of co-clustering is to simultaneously identify blocks of similarity between the sample set and feature set. Co-clustering has become a widely used technique in data mining, machine learning, and other research areas. The nonnegative matrix tri-factorization (NMTF) algorithm, which aims to decompose an objective matrix into three low-dimensional matrices, is an important tool to achieve coclustering. However, noise is usually introduced during objective matrix factorization, and the method of square loss is very sensitive to noise, which significantly reduces the performance of the model. To solve this issue, this paper proposes a tri-regularized NMTF (TRNMTF) model for co-clustering, which combines graph regularization, Frobenius norm, and l(1) norm to simultaneously optimize the objective function. TRNMTF can execute feature selection well, enhance the sparseness of the model, adjust the eigenvalues in the low-dimensional matrix, eliminate noise in the model, and obtain cleaner data matrices to approximate the objective matrix, which significantly improves the performance of the model and its generalization ability. Furthermore, to solve the iterative optimization schemes of TRNMTF, this study converts the objective function into elemental form to infer and provide detailed iterative update rules. Experimental results on 8 data sets show that the proposed model displays superior performance. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:12
相关论文
共 60 条
[1]  
[Anonymous], 2003, P 9 ACM SIGKDD INT C
[2]  
[Anonymous], 2011, P 20 ACM INT C INF K
[3]  
Asuncion Arthur, 2007, Uci machine learning repository
[4]  
Banerjee A, 2007, J MACH LEARN RES, V8, P1919
[5]  
Boyd S. P., 2004, Convex Optimization
[6]   Graph Regularized Nonnegative Matrix Factorization for Data Representation [J].
Cai, Deng ;
He, Xiaofei ;
Han, Jiawei ;
Huang, Thomas S. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (08) :1548-1560
[7]   Non-negative Matrix Factorization on Manifold [J].
Cai, Deng ;
He, Xiaofei ;
Wu, Xiaoyun ;
Han, Jiawei .
ICDM 2008: EIGHTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2008, :63-+
[8]   Subspace Weighting Co-Clustering of Gene Expression Data [J].
Chen, Xiaojun ;
Huang, Joshua Z. ;
Wu, Qingyao ;
Yang, Min .
IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2019, 16 (02) :352-364
[9]   Signal recovery by proximal forward-backward splitting [J].
Combettes, PL ;
Wajs, VR .
MULTISCALE MODELING & SIMULATION, 2005, 4 (04) :1168-1200
[10]   Linear discriminant analysis guided by unsupervised ensemble learning [J].
Deng, Ping ;
Wang, Hongjun ;
Li, Tianrui ;
Horng, Shi-Jinn ;
Zhu, Xinwen .
INFORMATION SCIENCES, 2019, 480 :211-221