Learning to steer nonlinear interior-point methods

被引:0
|
作者
Kuhlmann, Renke [1 ]
机构
[1] Ctr Ind Math ZeTeM, Optimizat & Optimal Control, Bibliothekstr 5, D-28359 Bremen, Germany
关键词
Nonlinear programming; Constrained optimization; Interior-point algorithm; Reinforcement learning; Deep Q-learning; PRIMAL-DUAL METHODS; LINE-SEARCH; ALGORITHM; IMPLEMENTATION; OPTIMIZATION; SOFTWARE;
D O I
10.1007/s13675-019-00118-4
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
Interior-point or barrier methods handle nonlinear programs by sequentially solving barrier subprograms with a decreasing sequence of barrier parameters. The specific barrier update rule strongly influences the theoretical convergence properties as well as the practical efficiency. While many global and local convergence analyses consider a monotone update that decreases the barrier parameter for every approximately solved subprogram, computational studies show a superior performance of more adaptive strategies. In this paper we interpret the adaptive barrier update as a reinforcement learning task. A deep Q-learning agent is trained by both imitation and random action selection. Numerical results based on an implementation within the nonlinear programming solver WORHP show that the agent successfully learns to steer the barrier parameter and additionally improves WORHP's performance on the CUTEst test set.
引用
收藏
页码:381 / 419
页数:39
相关论文
共 50 条