Heuristically Accelerated Reinforcement Learning for Dynamic Secondary Spectrum Sharing

被引:24
作者
Morozs, Nils [1 ]
Clarke, Tim [1 ]
Grace, David [1 ]
机构
[1] Univ York, Dept Elect, York YO10 5DD, N Yorkshire, England
关键词
Heuristically accelerated reinforcement learning; spectrum sharing; dynamic spectrum access; COGNITIVE RADIO NETWORKS; COEXISTENCE; CHALLENGES;
D O I
10.1109/ACCESS.2015.2507158
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper examines how flexible cellular system architectures and efficient spectrum management techniques can be used to play a key role in accommodating the exponentially increasing demand for mobile data capacity in the near future. The efficiency of the use of radio spectrum for wireless communications can be dramatically increased by dynamic secondary spectrum sharing; an intelligent approach that allows unlicensed devices access to those parts of the spectrum that are otherwise underutilized by the incumbent users. In this paper, we propose a heuristically accelerated reinforcement learning (HARL)-based framework, designed for dynamic secondary spectrum sharing in Long Term Evolution cellular systems. It utilizes a radio environment map as external information for guiding the learning process of cognitive cellular systems. System level simulations of a stadium temporary event scenario show that the schemes based on the proposed HARL framework achieve high controllability of spectrum sharing patterns in a fully autonomous way. This results in a significant decrease in the primary system quality of service degradation due to interference from the secondary cognitive systems, compared with a state-of-the-art reinforcement learning solution and a purely heuristic typical LTE solution. The spectrum sharing patterns that emerge by using the proposed schemes also result in remarkable reliability of the cognitive eNodeB on the aerial platform. Furthermore, the novel principle and the general structure of heuristic functions proposed in the context of HARL are applicable to a wide range of self-organization problems beyond the wireless communications domain.
引用
收藏
页码:2771 / 2783
页数:13
相关论文
共 36 条
[1]  
3GPP, 2013, TS36213 3GPP
[2]  
Alnwaimi G., 2013, P IEEE 78 VTC FALL S, P1
[3]  
[Anonymous], 2012, TR36952 3GPP
[4]  
[Anonymous], 2010, Tech. Specification TS 36.814 V9.0.0
[5]  
[Anonymous], 2014, TS36321 3GPP
[6]   Self-Organization in Small Cell Networks: A Reinforcement Learning Approach [J].
Bennis, Mehdi ;
Perlaza, Samir M. ;
Blasco, Pol ;
Han, Zhu ;
Poor, H. Vincent .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2013, 12 (07) :3202-3212
[7]   Heuristically-Accelerated Multiagent Reinforcement Learning [J].
Bianchi, Reinaldo A. C. ;
Martins, Murilo F. ;
Ribeiro, Carlos H. C. ;
Costa, Anna H. R. .
IEEE TRANSACTIONS ON CYBERNETICS, 2014, 44 (02) :252-265
[8]   Case-Based Multiagent Reinforcement Learning: Cases as Heuristics for Selection of Actions [J].
Bianchi, Reinaldo A. C. ;
Lopez de Mantaras, Ramon .
ECAI 2010 - 19TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2010, 215 :355-360
[9]   Multiagent learning using a variable learning rate [J].
Bowling, M ;
Veloso, M .
ARTIFICIAL INTELLIGENCE, 2002, 136 (02) :215-250
[10]   Stochastic Power Adaptation with Multiagent Reinforcement Learning for Cognitive Wireless Mesh Networks [J].
Chen, Xianfu ;
Zhao, Zhifeng ;
Zhang, Honggang .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2013, 12 (11) :2155-2166