Efficient Low-Rank Semidefinite Programming With Robust Loss Functions

被引:2
作者
Yao, Quanming [1 ,2 ]
Yang, Hansi [2 ]
Hu, En-Liang [3 ]
Kwok, James T. [4 ]
机构
[1] 4Paradigm Inc, Beijing 100089, Peoples R China
[2] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[3] Yunnan Normal Univ, Dept Math, Kunming 650092, Yunnan, Peoples R China
[4] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Kowloon, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Semi-definite programming; robustness; majorization-minimization; alternating direction method of multipliers; MATRIX FACTORIZATION; OPTIMIZATION; CONVERGENCE; COMPLETION; ALGORITHM;
D O I
10.1109/TPAMI.2021.3085858
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In real-world applications, it is important for machine learning algorithms to be robust against data outliers or corruptions. In this paper, we focus on improving the robustness of a large class of learning algorithms that are formulated as low-rank semi-definite programming (SDP) problems. Traditional formulations use the square loss, which is notorious for being sensitive to outliers. We propose to replace this with more robust noise models, including the l(1)-loss and other nonconvex losses. However, the resultant optimization problem becomes difficult as the objective is no longer convex or smooth. To alleviate this problem, we design an efficient algorithm based on majorization-minimization. The crux is on constructing a good optimization surrogate, and we show that this surrogate can be efficiently obtained by the alternating direction method of multipliers (ADMM). By properly monitoring ADMM's convergence, the proposed algorithm is empirically efficient and also theoretically guaranteed to converge to a critical point. Extensive experiments are performed on four machine learning applications using both synthetic and real-world data sets. Results show that the proposed algorithm is not only fast but also has better performance than the state-of-the-arts.
引用
收藏
页码:6153 / 6168
页数:16
相关论文
共 73 条
[1]  
[Anonymous], 2016, C LEARN THEOR
[2]  
[Anonymous], 2015, J GLOBAL OPTIM, DOI [DOI 10.1007/s10898-014-0247-2, 10.1007/s10898-014-0247-2]
[3]  
[Anonymous], 1990, Classics Appl. Math., DOI DOI 10.1137/1.9781611971309
[4]  
[Anonymous], 2013, INT C MACH LEARN
[5]   Photometric stereo with general, unknown lighting [J].
Basri, Ronen ;
Jacobs, David ;
Kemelmacher, Ira .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2007, 72 (03) :239-257
[6]  
Bhargava A, 2017, PR MACH LEARN RES, V54, P1349
[7]  
Bishop W. E., 2014, Advances in Neural Information Processing Systems, P2762
[8]  
Boumal N, 2016, ADV NEUR IN, V29
[9]   Deterministic Guarantees for Burer-Monteiro Factorizations of Smooth Semidefinite Programs [J].
Boumal, Nicolas ;
Voroninski, Vladislav ;
Bandeira, Afonso S. .
COMMUNICATIONS ON PURE AND APPLIED MATHEMATICS, 2020, 73 (03) :581-608
[10]   Distributed optimization and statistical learning via the alternating direction method of multipliers [J].
Boyd S. ;
Parikh N. ;
Chu E. ;
Peleato B. ;
Eckstein J. .
Foundations and Trends in Machine Learning, 2010, 3 (01) :1-122