Clustering-based attack detection for adversarial reinforcement learning

被引:0
作者
Rubén Majadas
Javier García
Fernando Fernández
机构
[1] Universidad Carlos III de Madrid,Departamento de Informática
[2] Universidad Santiago de Compostela,undefined
来源
Applied Intelligence | 2024年 / 54卷
关键词
Adversarial reinforcement learning; Adversarial attacks; Change-point detection; Clustering applications;
D O I
暂无
中图分类号
学科分类号
摘要
Detecting malicious attacks presents a major challenge in the field of reinforcement learning (RL), as such attacks can force the victim to perform abnormal actions, with potentially severe consequences. To mitigate these risks, current research focuses on the enhancement of RL algorithms with efficient detection mechanisms, especially for real-world applications. Adversarial attacks have the potential to alter the environmental dynamics of a Markov Decision Process (MDP) perceived by an RL agent. Leveraging these changes in dynamics, we propose a novel approach to detect attacks. Our contribution can be summarized in two main aspects. Firstly, we propose a novel formalization of the attack detection problem that entails analyzing modifications made by attacks to the transition and reward dynamics within the environment. This problem can be framed as a context change detection problem, where the goal is to identify the transition from a “free-of-attack” situation to an “under-attack” scenario. To solve this problem, we propose a groundbreaking “model-free” clustering-based countermeasure. This approach consists of two essential steps: first, partitioning the transition space into clusters, and then using this partitioning to identify changes in environmental dynamics caused by adversarial attacks. To assess the efficiency of our detection method, we performed experiments on four established RL domains (grid-world, mountain car, carpole, and acrobot) and subjected them to four advanced attack types. Uniform, Strategically-timed, Q-value, and Multi-objective. Our study proves that our technique has a high potential for perturbation detection, even in scenarios where attackers employ more sophisticated strategies.
引用
收藏
页码:2631 / 2647
页数:16
相关论文
共 10 条
[1]  
Chen T(2019)Adversarial attack and defense in reinforcement learning-from AI security view Cybersecur. 2 11-549
[2]  
Liu J(2006)A multivariate change-point model for statistical process control Technometrics 48 539-1054
[3]  
Xiang Y(1998)Reinforcement learning: an introduction IEEE Trans Neural Networks 9 1054-undefined
[4]  
Niu W(undefined)undefined undefined undefined undefined-undefined
[5]  
Tong E(undefined)undefined undefined undefined undefined-undefined
[6]  
Han Z(undefined)undefined undefined undefined undefined-undefined
[7]  
Zamba K(undefined)undefined undefined undefined undefined-undefined
[8]  
Hawkins DM(undefined)undefined undefined undefined undefined-undefined
[9]  
Sutton RS(undefined)undefined undefined undefined undefined-undefined
[10]  
Barto AG(undefined)undefined undefined undefined undefined-undefined