Application of reinforcement learning to wireless sensor networks: models and algorithms

被引:0
|
作者
Kok-Lim Alvin Yau
Hock Guan Goh
David Chieng
Kae Hsiang Kwong
机构
[1] Sunway University,Faculty of Science and Technology
[2] Universiti Tunku Abdul Rahman,Faculty of Information and Communication Technology
[3] MIMOS Technology Park Malaysia,Wireless Communication Cluster
[4] Recovision R&D,undefined
来源
Computing | 2015年 / 97卷
关键词
Wireless sensor networks; Reinforcement learning; Q-learning; Artificial intelligence; Context awareness; 68T05;
D O I
暂无
中图分类号
学科分类号
摘要
Wireless sensor network (WSN) consists of a large number of sensors and sink nodes which are used to monitor events or environmental parameters, such as movement, temperature, humidity, etc. Reinforcement learning (RL) has been applied in a wide range of schemes in WSNs, such as cooperative communication, routing and rate control, so that the sensors and sink nodes are able to observe and carry out optimal actions on their respective operating environment for network and application performance enhancements. This article provides an extensive review on the application of RL to WSNs. This covers many components and features of RL, such as state, action and reward. This article presents how most schemes in WSNs have been approached using the traditional and enhanced RL models and algorithms. It also presents performance enhancements brought about by the RL algorithms, and open issues associated with the application of RL in WSNs. This article aims to establish a foundation in order to spark new research interests in this area. Our discussion has been presented in a tutorial manner so that it is comprehensive and applicable to readers outside the specialty of both RL and WSNs.
引用
收藏
页码:1045 / 1075
页数:30
相关论文
共 50 条
  • [31] An Innovative Routing Algorithm with Reinforcement Learning and Pattern Tree Adjustment for Wireless Sensor Networks
    Fan, Chia-Yu
    Hsu, Chien-Chang
    Wang, Wei-Yi
    COMPUTATIONAL COLLECTIVE INTELLIGENCE: TECHNOLOGIES AND APPLICATIONS, PT III, 2010, 6423 : 398 - 405
  • [32] Optimizing the lifetime of wireless sensor networks via reinforcement-learning-based routing
    Guo, Wenjing
    Yan, Cairong
    Lu, Ting
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2019, 15 (02)
  • [33] Reinforcement Learning Models and Algorithms for Diabetes Management
    Yau, Kok-Lim Alvin
    Chong, Yung-Wey
    Fan, Xiumei
    Wu, Celimuge
    Saleem, Yasir
    Lim, Phei-Ching
    IEEE ACCESS, 2023, 11 : 28391 - 28415
  • [34] Aggregator Election in Wireless Sensor Networks: A Distributed Reinforcement Learning Approach
    Hajishabani, Maryam
    Kordafshari, Mohammad Sadegh
    Meybodi, Mohammad Reza
    2013 13TH IRANIAN CONFERENCE ON FUZZY SYSTEMS (IFSC), 2013,
  • [35] Machine Learning in Wireless Sensor Networks
    Kundaliya, Brijesh L.
    Patel, Josh
    Hadia, S. K.
    Patel, Upesh
    WIRELESS PERSONAL COMMUNICATIONS, 2024, 137 (01) : 367 - 397
  • [36] Reinforcement Learning-Based Routing Protocol for Underwater Wireless Sensor Networks: A Comparative Survey
    Rodoshi, Rehenuma Tasnim
    Song, Yujae
    Choi, Wooyeol
    IEEE ACCESS, 2021, 9 : 154578 - 154599
  • [37] Reinforcement Learning for Multiple Access Control in Wireless Sensor Networks: Review, Model, and Open Issues
    Fathi, Mohammad
    Maihami, Vafa
    Moradi, Parham
    WIRELESS PERSONAL COMMUNICATIONS, 2013, 72 (01) : 535 - 547
  • [38] Energy-aware Task Scheduling in Wireless Sensor Networks based on Cooperative Reinforcement Learning
    Khan, Muhidul Islam
    Rinner, Bernhard
    2014 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC), 2014, : 871 - 877
  • [39] Reinforcement Learning-Based Adaptive Stateless Routing for Ambient Backscatter Wireless Sensor Networks
    Guo, Huanyu
    Yang, Donghua
    Gao, Hong
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2024, 72 (07) : 4206 - 4225
  • [40] Energy balancing in multi-hop Wireless Sensor Networks: an approach based on reinforcement learning
    Oddi, Guido
    Pietrabissa, Antonio
    Liberati, Francesco
    2014 NASA/ESA CONFERENCE ON ADAPTIVE HARDWARE AND SYSTEMS (AHS), 2014, : 262 - 269