Power Management of Wireless Sensor Nodes with Coordinated Distributed Reinforcement Learning

被引:7
作者
Shresthamali, Shaswot [1 ]
Kondo, Masaaki [1 ]
Nakamura, Hiroshi [1 ]
机构
[1] Univ Tokyo, Tokyo, Japan
来源
2019 IEEE 37TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2019) | 2019年
基金
日本科学技术振兴机构;
关键词
Distributed Reinforcement Learning; Deep Reinforcement Learning; Energy Harvesting Wireless Sensor Nodes; Energy Neutral Operation; Internet of Things; epsilon-greedy exploration;
D O I
10.1109/ICCD46524.2019.00092
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Energy Harvesting Wireless Sensor Nodes (EHWSNs) require adaptive energy management policies for uninterrupted perpetual operation in their physical environments. Contemporary online Reinforcement Learning (RL) solutions take an unrealistically long time exploring the environment to converge on working policies. Our work accelerates learning by partitioning the state-space for simultaneous exploration by multiple agents. We achieve this by using a novel coordinated epsilon-greedy method and implement it via Distributed RL (DiRL) in an EHWSN network. Our simulation results show a four-fold increase in state-space penetration and reduction in time to achieve optimal operation by an order of magnitude (50x). Moreover, we also propose methods to reduce instances of disastrous outcomes associated with learning and exploration. This translates to reducing the downtimes of the nodes in simulations corresponding to a real-world scenario by one thirds.
引用
收藏
页码:638 / 647
页数:10
相关论文
共 36 条
[1]  
Achiam J., 2017, CoRR
[2]   A Learning Theoretic Approach to Energy Harvesting Communication System Optimization [J].
Blasco, Pol ;
Guenduez, Deniz ;
Dohler, Mischa .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2013, 12 (04) :1872-1882
[3]   Towards Cognitive IoT: Autonomous Prediction Model Selection for Solar-Powered Nodes [J].
Braten, Anders Eivind ;
Kraemer, Frank Alexander .
2018 IEEE INTERNATIONAL CONGRESS ON INTERNET OF THINGS (ICIOT), 2018, :118-125
[4]  
Fortunato M., 2017, Noisy networks for exploration
[5]   Optimal resource allocation using reinforcement learning for IoT content-centric services [J].
Gai, Keke ;
Qiu, Meikang .
APPLIED SOFT COMPUTING, 2018, 70 :12-21
[6]   Getting More Out of Energy-harvesting Systems: Energy Management under Time-varying Utility with PREACT [J].
Geissdoerfer, Kai ;
Jurdak, Raja ;
Kusy, Brano ;
Zimmerling, Marco .
IPSN '19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON INFORMATION PROCESSING IN SENSOR NETWORKS, 2019, :109-120
[7]  
Glorot X., 2010, P 13 INT C ART INT S, V9, P249
[8]   Optimising SME Potential in Modern Healthcare Systems: Challenges, Opportunities and Policy Recommendations [J].
Horgan, Denis ;
van Kranen, Henk J. ;
Morre, Servaas A. .
PUBLIC HEALTH GENOMICS, 2018, 21 (1-2) :1-17
[9]  
Houthooft Rein, 2016, Advances in Neural Information Processing Systems, V29
[10]   A Reinforcement Learning-Based ToD Provisioning Dynamic Power Management for Sustainable Operation of Energy Harvesting Wireless Sensor Node [J].
Hsu, Roy Chaoming ;
Liu, Cheng-Ting ;
Wang, Hao-Li .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2014, 2 (02) :181-191