Reinforcement learning-based resource allocation for dynamic aggregated WiFi/VLC HetNet

被引:2
作者
Luo, Liujun [1 ]
Bai, Bo [1 ]
Zhang, Xiaowei [1 ]
Han, Guoqing [1 ]
机构
[1] Xidian Univ, Sch Commun Engn, Xian 710071, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Heterogeneous network; Visible light communication; Resource allocation; Reinforcement learning; Handover overhead; VISIBLE-LIGHT COMMUNICATION; POWER ALLOCATION; SYSTEMS; VLC; NETWORKS; LIFI;
D O I
10.1016/j.optcom.2024.130450
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
High transmission rates and low power consumption make visible light communication (VLC) a highly promising supplementary technology for the next generation of mobile communication. Taking into account the limited coverage area, VLC can be combined with WiFi as a heterogeneous network (HetNet), thanks to their non-overlapping spectrum. In order to fully utilize the advantages of the WiFi and VLC technologies, a new aggregated WiFi/VLC HetNet consisting of a single WiFi AP and multiple VLC APs, is designed, where the user equipment (UE), for the first time, can access multiple VLC APs and one WiFi AP simultaneously, and be allowed to acquire multiple resource blocks (RB) in each AP at the same time. To optimize the performance of the above designed HetNet, a multi-objective optimization problem (MOOP) is formulated, which aims to maximize system throughput while reducing the handover rate. Since the above MOOP is nonconvex and nonlinear, the traditional resource allocation (RA) algorithm has a complex calculation process and poor timeliness performance to deal with this problem. To solve the above MOOP, a reinforcement learning (RL)-based RA algorithm is proposed, and considering the RB handover overhead in the aggregated WiFi/VLC HetNet, a reward function related to the system throughput and the UE handover rate is carefully designed. System throughput, system handover rate, user satisfaction, as well as user fairness performance are analyzed under three typical indoor illumination layouts. Finally, compared with the greedy algorithm and the hypergraph-based carrier aggregation algorithm, numerical results show that the proposed RL-based RA algorithm could improve the system throughput over the former two algorithms by 30.26% and 19.71%, while reducing the system handover rate by 0.15% and 0.02%, respectively.
引用
收藏
页数:9
相关论文
共 50 条
[21]   Deep Reinforcement Learning for Dynamic Uplink/Downlink Resource Allocation in High Mobility 5G HetNet [J].
Tang, Fengxiao ;
Zhou, Yibo ;
Kato, Nei .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2020, 38 (12) :2773-2782
[22]   Deep Reinforcement Learning-Based Service-Oriented Resource Allocation in Smart Grids [J].
Xi, Linhan ;
Wang, Ying ;
Wang, Yang ;
Wang, Zhihui ;
Wang, Xue ;
Chen, Yuanbin .
IEEE ACCESS, 2021, 9 (09) :77637-77648
[23]   Reinforcement Learning-Based UAVs Resource Allocation for Integrated Sensing and Communication (ISAC) System [J].
Wang, Min ;
Chen, Peng ;
Cao, Zhenxin ;
Chen, Yun .
ELECTRONICS, 2022, 11 (03)
[24]   Reinforcement learning-based online resource allocation for edge computing network [J].
Li Y.-J. ;
Jiang H.-T. ;
Gao M.-H. .
Kongzhi yu Juece/Control and Decision, 2022, 37 (11) :2880-2886
[25]   Deep Q-Network Learning Based Downlink Resource Allocation for Hybrid RF/VLC Systems [J].
Shrivastava, Shivanshu ;
Chen, Bin ;
Chen, Chen ;
Wang, Hui ;
Dai, Mingjun .
IEEE ACCESS, 2020, 8 :149412-149434
[26]   An Off-Policy Reinforcement Learning-Based Adaptive Optimization Method for Dynamic Resource Allocation Problem [J].
He, Baiyang ;
Meng, Ying ;
Tang, Lixin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (02) :3504-3518
[27]   LEARNING-BASED RESOURCE ALLOCATION WITH DYNAMIC DATA RATE CONSTRAINTS [J].
Behmandpoor, Pourya ;
Patrinos, Panagiotis ;
Moonen, Marc .
2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, :4088-4092
[28]   Reinforcement Learning Enabled Dynamic Resource Allocation in the Internet of Vehicles [J].
Liang, Hongbin ;
Zhang, Xiaohui ;
Hong, Xintao ;
Zhang, Zongyuan ;
Li, Mushu ;
Hu, Guangdi ;
Hou, Fen .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (07) :4957-4967
[29]   Reinforcement Learning-Based NOMA Power Allocation in the Presence of Smart Jamming [J].
Xiao, Liang ;
Li, Yanda ;
Dai, Canhuang ;
Dai, Huaiyu ;
Poor, H. Vincent .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2018, 67 (04) :3377-3389
[30]   Reinforcement Learning-Based User Scheduling and Resource Allocation for Massive MU-MIMO System [J].
Bu, Gaojing ;
Jiang, Jing .
2019 IEEE/CIC INTERNATIONAL CONFERENCE ON COMMUNICATIONS IN CHINA (ICCC), 2019,