Efficient hyperparameters optimization through model-based reinforcement learning with experience exploiting and meta-learning

被引:5
|
作者
Liu, Xiyuan [1 ]
Wu, Jia [1 ]
Chen, Senpeng [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
基金
美国国家科学基金会;
关键词
Hyperparameters optimization; Reinforcement learning; Meta-learning; Deep learning; CLASSIFIERS;
D O I
10.1007/s00500-023-08050-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hyperparameter optimization plays a significant role in the overall performance of machine learning algorithms. However, the computational cost of algorithm evaluation can be extremely high for complex algorithm or large dataset. In this paper, we propose a model-based reinforcement learning with experience variable and meta-learning optimization method to speed up the training process of hyperparameter optimization. Specifically, an RL agent is employed to select hyperparameters and treat the k-fold cross-validation result as a reward signal to update the agent. To guide the agent's policy update, we design an embedding representation called "experience variable" and dynamically update it during the training process. Besides, we employ a predictive model to predict the performance of machine learning algorithm with the selected hyperparameters and limit the model rollout in short horizon to reduce the impact of the inaccuracy of the model. Finally, we use the meta-learning technique to pre-train the model for fast adapting to a new task. To prove the advantages of our method, we conduct experiments on 25 real HPO tasks and the experimental results show that with the limited computational resources, the proposed method outperforms the state-of-the-art Bayesian methods and evolution method.
引用
收藏
页码:8661 / 8678
页数:18
相关论文
共 50 条
  • [31] Multi-Task Reinforcement Meta-Learning in Neural Networks
    Shakah, Ghazi
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (07) : 263 - 269
  • [32] Model gradient: unified model and policy learning in model-based reinforcement learning
    Chengxing Jia
    Fuxiang Zhang
    Tian Xu
    Jing-Cheng Pang
    Zongzhang Zhang
    Yang Yu
    Frontiers of Computer Science, 2024, 18
  • [33] Model gradient: unified model and policy learning in model-based reinforcement learning
    Jia, Chengxing
    Zhang, Fuxiang
    Xu, Tian
    Pang, Jing-Cheng
    Zhang, Zongzhang
    Yu, Yang
    FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (04)
  • [34] Model-Based OPC With Adaptive PID Control Through Reinforcement Learning
    Kim, Taeyoung
    Zhang, Shilong
    Shin, Youngsoo
    IEEE TRANSACTIONS ON SEMICONDUCTOR MANUFACTURING, 2025, 38 (01) : 48 - 56
  • [35] Optimization of the model predictive control meta-parameters through reinforcement learning
    Bohn, Eivind
    Gros, Sebastien
    Moe, Signe
    Johansen, Tor Arne
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [36] Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning
    Wang, Bokun
    Yuan, Zhuoning
    Ying, Yiming
    Yang, Tianbao
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [37] Model-based average reward reinforcement learning
    Tadepalli, P
    Ok, D
    ARTIFICIAL INTELLIGENCE, 1998, 100 (1-2) : 177 - 224
  • [38] Model-Based Reinforcement Learning in Robotics: A Survey
    Sun S.
    Lan X.
    Zhang H.
    Zheng N.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (01): : 1 - 16
  • [39] Pre-training with Augmentations for Efficient Transfer in Model-Based Reinforcement Learning
    Esteves, Bernardo
    Vasco, Miguel
    Melo, Francisco S.
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2023, PT I, 2023, 14115 : 133 - 145
  • [40] Case-Based Task Generalization in Model-Based Reinforcement Learning
    Zholus, Artem
    Panov, Aleksandr, I
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2021, 2022, 13154 : 344 - 354