Regret Bounds for Online Pairwise Learning with Non-Convex Loss Functions Using Stability Analysis

被引:0
|
作者
Lang X. [1 ]
Li C. [1 ,2 ]
Liu Y. [3 ,4 ]
Wang M. [1 ,2 ]
机构
[1] College of Computer and Information Technology, Northeastern Petroleum University, Heilongjiang, Daqing
[2] Heilongjiang Provincial Key Laboratory of Petroleum Big Data and Intelligent Analysis, Northeastern Petroleum University, Heilongjiang, Daqing
[3] Gaoling School of Artificial Intelligence, Renmin University of China, Beijing
[4] Beijing Key Laboratory of Big Data Management and Analysis Methods, Renmin University of China, Beijing
来源
Jisuanji Yanjiu yu Fazhan/Computer Research and Development | 2023年 / 60卷 / 12期
基金
中国国家自然科学基金;
关键词
non-convex; offline optimization oracle; online pairwise learning; regret bounds; stability;
D O I
10.7544/issn1000-1239.202220221
中图分类号
学科分类号
摘要
Pairwise learning refers to a learning task which involves a loss function depending on pairs of instances. Recently, there is a growing interest in studying pairwise learning since it includes many important machine learning tasks as specific examples, e.g., metric learning, AUC maximization and ranking. Regret bounds are particularly important for generalization analysis of online pairwise learning. The existing online pairwise learning analysis provides regret bounds only with convex loss functions. To fill the gap in the theoretical study of online pairwise learning with non-convex loss functions, we present a systematic study on the generalization analysis for online pairwise learning and propose regret bounds for non-convex online pairwise learning in this paper. We consider online learning in an adversarial, non-convex setting under the assumption that the learner has access to an offline optimization oracle and the learner’s prediction with expert advice. We first propose a general online pairwise learning framework and establish the stability of online pairwise learning with non-convex loss functions. Then, the regret bounds can be derived naturally from stability. Finally, we show that the general online pairwise learning framework with non-convex loss functions achieves optimal regret bounds of O(T−1/2) when the learner has access to an offline optimization oracle. © 2023 Science Press. All rights reserved.
引用
收藏
页码:2806 / 2813
页数:7
相关论文
共 26 条
  • [11] Xiaming Chen, Lei Yunwen, Refined bounds for online pairwise learning algorithms, Neurocomputing, 275, pp. 2656-2665, (2018)
  • [12] Zhengchu Guo, Yiming Ying, Dingxuan Zhou, Online regularized learning with pairwise loss functions[J], Advances in Computational Mathematics, 43, 1, pp. 127-150, (2017)
  • [13] Cheng Wang, Ting Hu, Online regularized pairwise learning with least squares loss[J], Analysis and Applications, 18, 1, pp. 49-78, (2020)
  • [14] Ting Hu, Jun Fan, Daohong Xiang, Convergence analysis of distributed multi-penalty regularized pairwise learning[J], Analysis and Applications, 18, 1, pp. 109-127, (2020)
  • [15] Bousquet O, Elisseeff A., Stability and generalization[J], The Journal of Machine Learning Research, 2, pp. 499-526, (2002)
  • [16] Elisseeff A, Evgeniou T, Pontil M, Et al., Stability of randomized learning algorithms[J], Journal of Machine Learning Research, 6, 1, pp. 55-79, (2005)
  • [17] Wei Shen, Zhenhuan Yang, Yiming Ying, Et al., Stability and optimization error of stochastic gradient descent for pairwise learning[J], Analysis and Applications, 18, 5, pp. 887-927, (2020)
  • [18] Yunwen Lei, Ledent A, Kloft M., Sharper generalization bounds for pairwise learning, Advances in Neural Information Processing Systems, 33, pp. 21236-21246, (2020)
  • [19] McMahan B., Follow-the-regularized-leader and mirror descent: Equivalence theorems and L1 regularization[C], Proc of the 14th Int Conf on Artificial Intelligence and Statistics, pp. 525-533, (2011)
  • [20] Agarwal N, Gonen A, Hazan E., Learning in non-convex games with an optimization oracle[C], Proc of the 32nd Conf on Learning Theory, pp. 18-29, (2019)