FASTune: Towards Fast and Stable Database Tuning System with Reinforcement Learning

被引:0
作者
Shi, Lei [1 ,2 ,3 ]
Li, Tian [2 ]
Wei, Lin [1 ]
Tao, Yongcai [2 ]
Li, Cuixia [1 ]
Gao, Yufei [1 ,3 ]
机构
[1] Zhengzhou Univ, Sch Cyber Sci & Engn, Zhengzhou 450002, Peoples R China
[2] Zhengzhou Univ, Sch Comp & Artificial Intelligence, Zhengzhou 450001, Peoples R China
[3] Songshan Lab, Zhengzhou 450046, Peoples R China
关键词
database tuning; reinforcement learning; decision making; deep learning;
D O I
10.3390/electronics12102168
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Configuration tuning is vital to achieving high performance for a database management system (DBMS). Recently, automatic tuning methods using Reinforcement Learning (RL) have been explored to find better configurations compared with database administrators (DBAs) and heuristics. However, existing RL-based methods still have several limitations: (1) Excessive overhead due to reliance on cloned databases; (2) trial-and-error strategy may produce dangerous configurations that lead to database failure; (3) lack the ability to handle dynamic workload. To address the above challenges, a fast and stable RL-based database tuning system, FASTune, is proposed. A virtual environment is proposed to evaluate configurations which is an equivalent yet more efficient scheme than the cloned database. To ensure stability during tuning, FASTune adopts an environment proxy to avoid dangerous configurations. In addition, a Multi-State Soft Actor-Critic (MS-SAC) model is proposed to handle dynamic workloads, which utilizes the soft actor-critic network to tune the database according to workload and database states. The experimental results indicate that, compared with the state-of-the-art methods, FASTune can achieve improvements in performance while maintaining stability in the tuning.
引用
收藏
页数:22
相关论文
共 50 条
[21]   Reinforcement learning approach to autonomous PID tuning [J].
Dogru, Oguzhan ;
Velswamy, Kirubakaran ;
Ibrahim, Fadi ;
Wu, Yuqi ;
Sundaramoorthy, Arun Senthil ;
Huang, Biao ;
Xu, Shu ;
Nixon, Mark ;
Bell, Noel .
COMPUTERS & CHEMICAL ENGINEERING, 2022, 161
[22]   Intelligent Index Tuning Using Reinforcement Learning [J].
Matczak, Michal ;
Czochanski, Tomasz .
NEW TRENDS IN DATABASE AND INFORMATION SYSTEMS, ADBIS 2023, 2023, 1850 :523-534
[23]   Recipe tuning by Reinforcement Learning in the SandS ecosystem [J].
Fernandez-Gauna, Borja ;
Grana, Manuel .
2014 6TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL ASPECTS OF SOCIAL NETWORKS (CASON), 2014, :55-60
[24]   A Lyapunov approach for stable reinforcement learning [J].
Julio B. Clempner .
Computational and Applied Mathematics, 2022, 41
[25]   Meta-reinforcement learning for the tuning of PI controllers: An offline approach [J].
McClement, Daniel G. ;
Lawrence, Nathan P. ;
Backstroem, Johan U. ;
Loewen, Philip D. ;
Forbes, Michael G. ;
Gopaluni, R. Bhushan .
JOURNAL OF PROCESS CONTROL, 2022, 118 :139-152
[26]   Deep reinforcement learning with shallow controllers: An experimental application to PID tuning [J].
Lawrence, Nathan P. ;
Forbes, Michael G. ;
Loewen, Philip D. ;
McClement, Daniel G. ;
Backstrom, Johan U. ;
Gopaluni, R. Bhushan .
CONTROL ENGINEERING PRACTICE, 2022, 121
[27]   Learning to Score: Tuning Cluster Schedulers through Reinforcement Learning [J].
Asenov, Martin ;
Deng, Qiwen ;
Yeung, Gingfung ;
Barker, Adam .
2023 IEEE INTERNATIONAL CONFERENCE ON CLOUD ENGINEERING, IC2E, 2023, :113-120
[28]   Reinforcement Learning and Inverse Reinforcement Learning with System 1 and System 2 [J].
Peysakhovich, Alexander .
AIES '19: PROCEEDINGS OF THE 2019 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2019, :409-415
[29]   Fast Reinforcement Learning By Mirror Images [J].
Kitao, Takehiro ;
Miura, Takao .
2016 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE (WI 2016), 2016, :660-663
[30]   Federated Reinforcement Learning For Fast Personalization [J].
Nadiger, Chetan ;
Kumar, Anil ;
Abdelhak, Sherine .
2019 IEEE SECOND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE), 2019, :123-127