Value-Function-Based Sequential Minimization for Bi-Level Optimization

被引:3
作者
Liu, Risheng [1 ,2 ]
Liu, Xuan [1 ,2 ]
Zeng, Shangzhi [3 ]
Zhang, Jin [4 ,5 ]
Zhang, Yixuan [6 ]
机构
[1] Dalian Univ Technol, DUT RU Int Sch Informat Sci & Engn, Dalian 116024, Peoples R China
[2] Key Lab Ubiquitous Network & Serv Software Liaoni, Dalian 116024, Peoples R China
[3] Univ Victoria, Dept Math & Stat, Victoria, BC V8P 5C2, Canada
[4] Southern Univ Sci & Technol, Natl Ctr Appl Math Shenzhen, Dept Math, SUSTech Int Ctr Math, Shenzhen 518055, Peoples R China
[5] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[6] Hong Kong Polytech Univ, Dept Appl Math, Hong Kong, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Bi-level optimization; gradient-based method; hyper-parameter optimization; sequential minimization; value-function; MODEL;
D O I
10.1109/TPAMI.2023.3303227
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gradient-based Bi-Level Optimization (BLO) methods have been widely applied to handle modern learning tasks. However, most existing strategies are theoretically designed based on restrictive assumptions (e.g., convexity of the lower-level sub-problem), and computationally not applicable for high-dimensional tasks. Moreover, there are almost no gradient-based methods able to solve BLO in those challenging scenarios, such as BLO with functional constraints and pessimistic BLO. In this work, by reformulating BLO into approximated single-level problems, we provide a new algorithm, named Bi-level Value-Function-based Sequential Minimization (BVFSM), to address the above issues. Specifically, BVFSM constructs a series of value-function-based approximations, and thus avoids repeated calculations of recurrent gradient and Hessian inverse required by existing approaches, time-consuming especially for high-dimensional tasks. We also extend BVFSM to address BLO with additional functional constraints. More importantly, BVFSM can be used for the challenging pessimistic BLO, which has never been properly solved before. In theory, we prove the asymptotic convergence of BVFSM on these types of BLO, in which the restrictive lower-level convexity assumption is discarded. To our best knowledge, this is the first gradient-based algorithm that can solve different kinds of BLO (e.g., optimistic, pessimistic, and with constraints) with solid convergence guarantees. Extensive experiments verify the theoretical investigations and demonstrate our superiority on various real-world applications.
引用
收藏
页码:15930 / 15948
页数:19
相关论文
共 52 条