Testing the Plasticity of Reinforcement Learning-based Systems

被引:9
|
作者
Biagiola, Matteo [1 ]
Tonella, Paolo [1 ]
机构
[1] Univ Svizzera Italiana, CH-6900 Lugano, Switzerland
基金
欧洲研究理事会;
关键词
Software testing; reinforcement learning; empirical software engineering; NEURAL-NETWORKS;
D O I
10.1145/3511701
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The dataset available for pre-release training of a machine-learning based system is often not representative of all possible execution contexts that the system will encounter in the field. Reinforcement Learning (RL) is a prominent approach among those that support continual learning, i.e., learning continually in the field, in the post-release phase. No study has so far investigated any method to test the plasticity of RL-based systems, i.e., their capability to adapt to an execution context that may deviate from the training one. We propose an approach to test the plasticity of RL-based systems. The output of our approach is a quantification of the adaptation and anti-regression capabilities of the system, obtained by computing the adaptation frontier of the system in a changed environment. We visualize such frontier as an adaptation/anti-regression heatmap in two dimensions, or as a clustered projection when more than two dimensions are involved. In this way, we provide developers with information on the amount of changes that can be accommodated by the continual learning component of the system, which is key to decide if online, in-the-field learning can be safely enabled or not.
引用
收藏
页数:46
相关论文
共 50 条
  • [31] Graph learning-based generation of abstractions for reinforcement learning
    Xue, Yuan
    Kudenko, Daniel
    Khosla, Megha
    NEURAL COMPUTING & APPLICATIONS, 2023,
  • [32] Deep Reinforcement Learning-based Quantization for Federated Learning
    Zheng, Sihui
    Dong, Yuhan
    Chen, Xiang
    2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,
  • [33] Reinforcement learning-based detection method for malware behavior in industrial control systems
    Gao Y.
    Wang L.-W.
    Ren W.
    Xie F.
    Mo X.-F.
    Luo X.
    Wang W.-P.
    Yang X.
    Gongcheng Kexue Xuebao/Chinese Journal of Engineering, 2020, 42 (04): : 455 - 462
  • [34] Reinforcement Learning-Based Control for a Class of Nonlinear Systems with unknown control directions
    Song, Xiaoling
    Huang, Miao
    Wen, Gang
    Ma, Longhua
    Yao, Jiaqing
    Lu, Zheming
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 2519 - 2524
  • [35] Optimal Reinforcement Learning-Based Control Algorithm for a Class of Nonlinear Macroeconomic Systems
    Ding, Qing
    Jahanshahi, Hadi
    Wang, Ye
    Bekiros, Stelios
    Alassafi, Madini O.
    MATHEMATICS, 2022, 10 (03)
  • [36] A reinforcement learning-based scheme for adaptive optimal control of linear stochastic systems
    Wong, Wee Chin
    Lee, Jay H.
    2008 AMERICAN CONTROL CONFERENCE, VOLS 1-12, 2008, : 57 - 62
  • [37] AlphaSOC: Reinforcement Learning-based Cybersecurity Automation for Cyber-Physical Systems
    Silva, Ryan
    Hickert, Cameron
    Sarfaraz, Nicolas
    Brush, Jeff
    Silbermann, Josh
    Sookoor, Tamim
    2022 13TH ACM/IEEE INTERNATIONAL CONFERENCE ON CYBER-PHYSICAL SYSTEMS (ICCPS 2022), 2022, : 290 - 291
  • [38] Reinforcement Learning-Based Demand Response Management in Smart Grid Systems With Prosumers
    Sangoleye, Fisayo
    Jao, Jenilee
    Faris, Kimberly
    Tsiropoulou, Eirini Eleni
    Papavassiliou, Symeon
    IEEE SYSTEMS JOURNAL, 2023, 17 (02): : 1797 - 1807
  • [39] Reinforcement Learning-Based Tracking Control for Networked Control Systems With DoS Attacks
    Liu, Jinliang
    Dong, Yanhui
    Zha, Lijuan
    Xie, Xiangpeng
    Tian, Engang
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 4188 - 4197
  • [40] Reinforcement learning-based robust optimal tracking control for disturbed nonlinear systems
    Zhong-Xin Fan
    Lintao Tang
    Shihua Li
    Rongjie Liu
    Neural Computing and Applications, 2023, 35 : 23987 - 23996