Neutrons Sensitivity of Deep Reinforcement Learning Policies on EdgeAI Accelerators

被引:1
|
作者
Bodmann, Pablo R. [1 ]
Saveriano, Matteo [2 ]
Kritikakou, Angeliki [3 ]
Rech, Paolo [2 ]
机构
[1] Univ Fed Rio Grande do Sul, Informat Inst, BR-91501970 Porto Alegre, Brazil
[2] Univ Trento, Dept Ind Engn, I-38123 Trento, Italy
[3] INRIA, F-35042 Rennes, France
关键词
Robots; Reliability; Neutrons; Particle beams; Internet; Transient analysis; Task analysis; Artificial intelligence; EdgeAI; reinforcement learning (RL); reliability; ROBOT; SAFETY;
D O I
10.1109/TNS.2024.3387087
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Autonomous robots and their applications are becoming popular in several different fields, including tasks where robots closely interact with humans. Therefore, the reliability of computation must be paramount. In this work, we measure the reliability of Google's Coral Edge tensor processing unit (TPU) executing three deep reinforcement learning (DRL) models through an accelerated neutrons beam. We experimentally collect data that, when scaled to the natural neutron flux, account for more than 5 million years. Based on our extensive evaluation, we quantify and qualify the radiation-induced corruption on the correctness of DRL. Crucially, our data show that the Edge TPU executing DRL has an error rate that is up to 18 times higher the limit imposed by international reliability standards. We found that despite the feedback and intrinsic redundancy of DRL, the propagation of the fault induces the model to fail in the vast majority of cases or the model manages to finish but reports wrong metrics (i.e., speed, final position, and reward). We provide insights on how radiation corrupts the model, on how the fault propagates in the computation, and about the failure characteristic of the controlled robot.
引用
收藏
页码:1480 / 1486
页数:7
相关论文
共 50 条
  • [31] Understanding Error Propagation in Deep Learning Neural Network (DNN) Accelerators and Applications
    Li, Guanpeng
    Hari, Siva Kumar Sastry
    Sullivan, Michael
    Tsai, Timothy
    Pattabiraman, Karthik
    Emer, Joel
    Keckler, Stephen W.
    SC'17: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2017,
  • [32] An Improved Anti-Jamming Method Based on Deep Reinforcement Learning and Feature Engineering
    Chang, Xin
    Li, Yanbin
    Zhao, Yan
    Du, Yufeng
    Liu, Donghui
    IEEE ACCESS, 2022, 10 : 69992 - 70000
  • [33] Empowering Security and Trust in 5G and Beyond: A Deep Reinforcement Learning Approach
    Moudoud, Hajar
    Cherkaoui, Soumaya
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2023, 4 : 2410 - 2420
  • [34] Reliable Computation Offloading of DAG Applications in Internet of Vehicles Based on Deep Reinforcement Learning
    Su, Shengchao
    Yuan, Pengtao
    Dai, Yufeng
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2025, 74 (02) : 2116 - 2128
  • [35] Intelligent Trainer for Dyna-Style Model-Based Deep Reinforcement Learning
    Dong, Linsen
    Li, Yuanlong
    Zhou, Xin
    Wen, Yonggang
    Guan, Kyle
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (06) : 2758 - 2771
  • [36] A review of emerging trends in photonic deep learning accelerators
    Atwany, Mohammad
    Pardo, Sarah
    Serunjogi, Solomon
    Rasras, Mahmoud
    FRONTIERS IN PHYSICS, 2024, 12
  • [37] State Representation Learning With Adjacent State Consistency Loss for Deep Reinforcement Learning
    Zhao, Tianyu
    Zhao, Jian
    Zhou, Wengang
    Zhou, Yun
    Li, Houqiang
    IEEE MULTIMEDIA, 2021, 28 (03) : 117 - 127
  • [38] Bayesian Reinforcement Learning and Bayesian Deep Learning for Blockchains With Mobile Edge Computing
    Asheralieva, Alia
    Niyato, Dusit
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2021, 7 (01) : 319 - 335
  • [39] Deep Reinforcement Learning for Cyber Security
    Thanh Thi Nguyen
    Reddi, Vijay Janapa
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 3779 - 3795
  • [40] Deep reinforcement learning for conservation decisions
    Lapeyrolerie, Marcus
    Chapman, Melissa S.
    Norman, Kari E. A.
    Boettiger, Carl
    METHODS IN ECOLOGY AND EVOLUTION, 2022, 13 (11): : 2649 - 2662