Machine Learning Cyberattack and Defense Strategies

被引:32
作者
Bland, John A. [1 ]
Petty, Mikel D. [1 ]
Whitaker, Tymaine S. [1 ]
Maxwell, Katia P. [2 ]
Cantrell, Walter Alan [3 ]
机构
[1] Univ Alabama Huntsville, 301 Sparkman Dr,OKT N353, Huntsville, AL 35899 USA
[2] Athens State Univ, 300 N Beaty St,Waters Hall S103C, Athens, AL 35611 USA
[3] Lipscomb Univ, Coll Comp & Technol, 1 Univ Pk Dr, Nashville, TN 37204 USA
关键词
Cybersecurity; Modeling; Petri Net; Machine Learning; CAPEC; Reinforcement Learning;
D O I
10.1016/j.cose.2020.101738
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cybersecurity is an increasingly important challenge for computer systems. In this work, cyberattacks were modeled using an extension of the well-known Petri net formalism. That formalism, designated Petri nets with players, strategies, and costs, models the states of the cyberattack and events during the attack as markings and transition firings in the net respectively. The formalism models the attacker and defender as competing players who may observe the marking of a subset of the net and based on the observed marking act by changing the stochastic firing rates of a subset of the transitions in order to achieve their competing goals. Rate changes by the players incur a cost. Using the formalism, nets were constructed to model specific cyberattack patterns (cross-site scripting and spear phishing) documented in the Common Attack Pattern Enumeration and Classification database. The models were validated by a panel of cybersecurity experts in a structured face validation process. Given those validated nets, a reinforcement learning algorithm using an-Greedy policy was implemented and set to the task of learning which actions to take, i.e., which transition rates to change for the different observable markings, so as to accomplish the goals of the attacker or defender. Experiments were conducted with a dynamic (learning) attacker against a static (fixed) defender, a static attacker against a dynamic defender, and a dynamic attacker against a dynamic defender. In all cases, the reinforcement learning algorithm was able to improve its performance, in terms of achieving the player's objective and reducing the cost of doing so, over time. These results demonstrate the potential of formally modeling cyberattacks and of applying reinforcement learning to improving cybersecurity. (C) 2020 The Authors. Published by Elsevier Ltd.
引用
收藏
页数:23
相关论文
共 22 条
[1]  
[Anonymous], [No title captured]
[2]  
[Anonymous], [No title captured]
[3]  
[Anonymous], 2018, REINFORCEMENT LEARNI
[4]  
Bland J.A., 2018, P 2018 ALASIM INT C
[5]  
Cantrell Walter Alan, 2018, P 2018 ALASIM INT C
[6]   Cyberphysical Security and Dependability Analysis of Digital Control Systems in Nuclear Power Plants [J].
Cho, Chi-Shiang ;
Chung, Wei-Ho ;
Kuo, Sy-Yen .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2016, 46 (03) :356-369
[7]  
El Bouchti A, 2016, INT CONF FUTURE GEN, P42, DOI 10.1109/FGCT.2016.7605070
[8]  
Henry Matthew H., 2009, 2009 IEEE Conference on Technologies for Homeland Security (HST), P607, DOI 10.1109/THS.2009.5168093
[9]  
Lin C., 2008, 29th International Conference on Application and Theory of Petri Nets and other Models of Concurrency, P21
[10]  
Mayfield K.P., 2018, P 2018 ALASIM INT C