Toward Energy-Efficient Spike-Based Deep Reinforcement Learning With Temporal Coding

被引:0
作者
Zhang, Malu [1 ]
Wang, Shuai [1 ]
Wu, Jibin [2 ]
Wei, Wenjie [1 ]
Zhang, Dehao [1 ]
Zhou, Zijian [1 ]
Wang, Siying [1 ]
Zhang, Fan [1 ]
Yang, Yang [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 610054, Peoples R China
[2] Hong Kong Polytech Univ, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Computational modeling; Biological system modeling; Decision making; Memory management; Deep reinforcement learning; Energy efficiency; Encoding; Real-time systems; Timing; Computational complexity; POWER;
D O I
10.1109/MCI.2025.3541572
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning (DRL) facilitates efficient interaction with complex environments by enabling continuous optimization strategies and providing agents with autonomous learning abilities. However, traditional DRL methods often require large-scale neural networks and extensive computational resources, which limits their applicability in power-sensitive and resource-constrained edge environments, such as mobile robots and drones. To overcome these limitations, we leverage the energy-efficient properties of brain-inspired spiking neural networks (SNNs) to develop a novel spike-based DRL framework, referred to as Spike-DRL. Unlike traditional SNN-based reinforcement learning methods, Spike-DRL incorporates the energy-efficient time-to-first-spike (TTFS) encoding scheme, where information is encoded through the precise timing of a single spike. This TTFS-based method allows Spike-DRL to work in a sparse, event-driven manner, significantly reducing energy consumption. In addition, to improve the deployment capability of Spike-DRL in resource-constrained environments, a lightweight strategy for quantizing synaptic weights into low-bit representations is introduced, significantly reducing memory usage and computational complexity. Extensive experiments have been conducted to evaluate the performance of the proposed Spike-DRL, and the results show that our method achieves competitive performance with higher energy efficiency and lower memory requirements. This work presents a biologically inspired model that is well suited for real-time decision-making and autonomous learning in power-sensitive and resource-limited edge environments.
引用
收藏
页码:45 / 57
页数:13
相关论文
共 84 条
[1]   Toward robust and scalable deep spiking reinforcement learning [J].
Akl, Mahmoud ;
Ergene, Deniz ;
Walter, Florian ;
Knoll, Alois .
FRONTIERS IN NEUROROBOTICS, 2023, 16
[2]   Continuous Action Reinforcement Learning From a Mixture of Interpretable Experts [J].
Akrour, Riad ;
Tateo, Davide ;
Peters, Jan .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) :6795-6806
[3]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[4]   A Survey of Encoding Techniques for Signal Processing in Spiking Neural Networks [J].
Auge, Daniel ;
Hille, Julian ;
Mueller, Etienne ;
Knoll, Alois .
NEURAL PROCESSING LETTERS, 2021, 53 (06) :4693-4710
[5]  
Bengio Y, 2013, Arxiv, DOI arXiv:1308.3432
[6]   Fully Spiking Actor Network With Intralayer Connections for Reinforcement Learning [J].
Chen, Ding ;
Peng, Peixi ;
Huang, Tiejun ;
Tian, Yonghong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (02) :2881-2893
[7]  
Chen D, 2024, Arxiv, DOI [arXiv:2201.09754, 10.48550/arXiv.2201.09754]
[8]   Temporal Coding in Spiking Neural Networks With Alpha Synaptic Function: Learning With Backpropagation [J].
Comsa, Iulia-Maria ;
Potempa, Krzysztof ;
Versari, Luca ;
Fischbacher, Thomas ;
Gesmundo, Andrea ;
Alakuijala, Jyrki .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) :5939-5952
[9]  
Deng S., 2022, arXiv, DOI 10.48550/arXiv.2202.11946
[10]  
Diehl PU, 2016, 2016 IEEE INTERNATIONAL CONFERENCE ON REBOOTING COMPUTING (ICRC)