Scalable and Efficient Pairwise Learning to Achieve Statistical Accuracy

被引:0
作者
Gu, Bin [1 ]
Huo, Zhouyuan [2 ]
Huang, Heng [1 ,2 ]
机构
[1] JDDGlobal Com, Westlake Village, CA USA
[2] Univ Pittsburgh, Dept Elect & Comp Engn, Pittsburgh, PA 15260 USA
来源
THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2019年
关键词
GENERALIZATION BOUNDS; ALGORITHMS; RANKING;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pairwise learning is an important learning topic in the machine learning community, where the loss function involves pairs of samples (e.g., AUC maximization and metric learning). Existing pairwise learning algorithms do not perform well in the generality, scalability and efficiency simultaneously. To address these challenging problems, in this paper, we first analyze the relationship between the statistical accuracy and the regularized empire risk for pairwise loss. Based on the relationship, we propose a scalable and efficient adaptive doubly stochastic gradient algorithm (AdaDSG) for generalized regularized pairwise learning problems. More importantly, we prove that the overall computational cost of AdaDSG is O(n) to achieve the statistical accuracy on the full training set with the size of n, which is the best theoretical result for pairwise learning to the best of our knowledge. The experimental results on a variety of real-world datasets not only confirm the effectiveness of our AdaDSG algorithm, but also show that AdaDSG has significantly better scalability and efficiency than the existing pairwise learning algorithms.
引用
收藏
页码:3697 / 3704
页数:8
相关论文
共 30 条
  • [1] Agarwal S, 2005, J MACH LEARN RES, V6, P393
  • [2] Agarwal S, 2009, J MACH LEARN RES, V10, P441
  • [3] [Anonymous], 2011, PROC ICML
  • [4] [Anonymous], 2014, Advances in Neural Information Processing Systems
  • [5] [Anonymous], 2013, ADV NEURAL INFORM PR
  • [6] [Anonymous], 2016, PROC 33 INT C MACH L
  • [7] Boissier M, 2016, JMLR WORKSH CONF PRO, V51, P204
  • [8] Boyd Stephen P., 2014, Convex Optimization
  • [9] Generalization bounds for metric and similarity learning
    Cao, Qiong
    Guo, Zheng-Chu
    Ying, Yiming
    [J]. MACHINE LEARNING, 2016, 102 (01) : 115 - 132
  • [10] On the generalization ability of on-line learning algorithms
    Cesa-Bianchi, N
    Conconi, A
    Gentile, C
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2004, 50 (09) : 2050 - 2057