Unveiling bitcoin network attack using deep reinforcement learning with Boltzmann exploration

被引:0
|
作者
Shetty, Monali [1 ]
Tamane, Sharvari [2 ]
机构
[1] MGM Univ, Jawaharlal Nehru Engn Coll, CSE Dept, Aurangabad 431001, Maharashtra, India
[2] MGM Univ, Dept Informat & Commun Technol, Aurangabad 431001, Maharashtra, India
关键词
Blockchain; Bitcoin; Ransomware; Cryptocurrency; Boltzmann exploration; Attack; Reinforcement learning; RANSOMWARE;
D O I
10.1007/s12083-024-01829-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This study tackles the critical issue of identifying ransomware transactions within the Bitcoin network. These transactions threaten the stability and security of the cryptocurrency world. Traditional machine learning methods struggle to adapt to the evolving tactics employed by ransomware attackers. They rely on predefined features and metrics, limiting their ability to replicate the adaptability of human analysts. To address this challenge and to address the dynamic nature of fraudulent Bitcoin transactions, we propose a novel approach that incorporates Deep Q-Network (DQN) with Boltzmann exploration model that can autonomously learn and identify evolving attack patterns. The proposed Deep Reinforcement Learning (DRL) offers a more flexible approach by mimicking how security experts learn and adjust their strategies. DQN is a type of reinforcement learning that allows the agent to learn through trial-and-error interactions with the environment. Boltzmann exploration is a technique used to balance exploration (trying new actions) and exploitation (taking actions with the highest expected reward) during the learning process. Proposed DQN model with Boltzmann exploration was evaluated in a simulated environment. This strategy emphasizes the importance of dynamic decision-making for achieving convergence and stability during the learning process, ultimately leading to optimized results. The model achieved a promising validation accuracy of 91% and a strong F1 score demonstrating its ability to generalize effectively to unseen data. This is crucial for real-world applications where encountering entirely new attack scenarios is likely. Compared to alternative exploration techniques like Epsilon-Greedy and Random Exploration, Boltzmann exploration led to superior performance on unseen data. This suggests that the Boltzmann temperature parameter effectively guided the agent's exploration-exploitation trade-off, allowing it to discover valuable patterns applicable to new datasets. In conclusion, our findings demonstrate the potential of DQN with Boltzmann exploration for unsupervised ransomware transaction detection in the Bitcoin network. This approach offers a promising solution for improving the security and resilience of Bitcoin networks against evolving ransomware threats.
引用
收藏
页码:20 / 20
页数:1
相关论文
共 50 条
  • [41] Goal- Driven Autonomous Exploration Through Deep Reinforcement Learning
    Cimurs, Reinis
    Suh, Il Hong
    Lee, Jin Han
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 730 - 737
  • [42] A Hierarchical SLAM Framework Based on Deep Reinforcement Learning for Active Exploration
    Xue, Yuntao
    Chen, Weisheng
    Zhang, Liangbin
    PROCEEDINGS OF 2022 INTERNATIONAL CONFERENCE ON AUTONOMOUS UNMANNED SYSTEMS, ICAUS 2022, 2023, 1010 : 957 - 966
  • [43] Understanding via Exploration: Discovery of Interpretable Features With Deep Reinforcement Learning
    Wei, Jiawen
    Qiu, Zhifeng
    Wang, Fangyuan
    Lin, Wenwei
    Gui, Ning
    Gui, Weihua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 1696 - 1707
  • [44] Reinforcement learning using chaotic exploration in maze world
    Morihiro, K
    Matsui, N
    Nishimura, H
    SICE 2004 ANNUAL CONFERENCE, VOLS 1-3, 2004, : 1368 - 1371
  • [45] Adaptive deep Q learning network with reinforcement learning for crime prediction
    J. Vimala Devi
    K. S. Kavitha
    Evolutionary Intelligence, 2023, 16 : 685 - 696
  • [46] Deep Belief Network Using Reinforcement Learning and Its Applications to Time Series Forecasting
    Hirata, Takaomi
    Kuremoto, Takashi
    Obayashi, Masanao
    Mabu, Shingo
    Kobayashi, Kunikazu
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT III, 2016, 9949 : 30 - 37
  • [47] Adaptive deep Q learning network with reinforcement learning for crime prediction
    Devi, J. Vimala
    Kavitha, K. S.
    EVOLUTIONARY INTELLIGENCE, 2023, 16 (02) : 685 - 696
  • [48] Variational Dynamic for Self-Supervised Exploration in Deep Reinforcement Learning
    Bai, Chenjia
    Liu, Peng
    Liu, Kaiyu
    Wang, Lingxiao
    Zhao, Yingnan
    Han, Lei
    Wang, Zhaoran
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4776 - 4790
  • [49] Multiobjective Reinforcement Learning for Cognitive Satellite Communications Using Deep Neural Network Ensembles
    Rodrigues Ferreira, Paulo Victor
    Paffenroth, Randy
    Wyglinski, Alexander M.
    Hackett, Timothy M.
    Bilen, Sven G.
    Reinhart, Richard C.
    Mortensen, Dale J.
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2018, 36 (05) : 1030 - 1041
  • [50] DynPen: Automated Penetration Testing in Dynamic Network Scenarios Using Deep Reinforcement Learning
    Li, Qianyu
    Wang, Ruipeng
    Li, Dong
    Shi, Fan
    Zhang, Min
    Chattopadhyay, Anupam
    Shen, Yi
    Li, Yang
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 8966 - 8981