ADAPTIVE CONTENTION WINDOW DESIGN USING DEEP Q-LEARNING

被引:28
作者
Kumar, Abhishek [1 ]
Verma, Gunjan [2 ]
Rao, Chirag [2 ]
Swami, Ananthram [2 ]
Segarra, Santiago [1 ]
机构
[1] Rice Univ, Houston, TX 77251 USA
[2] US Armys CCDC Army Res Lab, Adelphi, MD USA
来源
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021) | 2021年
关键词
Wireless network; random access; contention window; reinforcement learning; deep Q-learning; ACCESS-CONTROL;
D O I
10.1109/ICASSP39728.2021.9414805
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We study the problem of adaptive contention window (CW) design for random-access wireless networks. More precisely, our goal is to design an intelligent node that can dynamically adapt its minimum CW (MCW) parameter to maximize a network-level utility knowing neither the MCWs of other nodes nor how these change over time. To achieve this goal, we adopt a reinforcement learning (RL) framework where we circumvent the lack of system knowledge with local channel observations and we reward actions that lead to high utilities. To efficiently learn these preferred actions, we follow a deep Q-learning approach, where the Q-value function is parametrized using a multi-layer perceptron. In particular, we implement a rainbow agent, which incorporates several empirical improvements over the basic deep Q-network. Numerical experiments based on the NS3 simulator reveal that the proposed RL agent performs close to optimal and markedly improves upon existing learning and non-learning based alternatives.
引用
收藏
页码:4950 / 4954
页数:5
相关论文
共 29 条
[11]   Contention Window Optimization for IEEE 802.11 DCF Access Control [J].
Deng, Der-Jiunn ;
Ke, Chih-Heng ;
Chen, Hslao-Hwa ;
Huang, Yueh-Min .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2008, 7 (12) :5129-5135
[12]   Dynamically Tuning IEEE 802.11's Contention Window Using Machine Learning [J].
Edalat, Yalda ;
Obraczka, Katia .
MSWIM'19: PROCEEDINGS OF THE 22ND INTERNATIONAL ACM CONFERENCE ON MODELING, ANALYSIS AND SIMULATION OF WIRELESS AND MOBILE SYSTEMS, 2019, :19-26
[13]   Channel Condition Based Contention Window Adaptation in IEEE 802.11 WLANs [J].
Hong, Kunho ;
Lee, SuKyoung ;
Kim, Kyungsoo ;
Kim, YoonHyuk .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2012, 60 (02) :469-478
[14]  
Ksentini A., 2005, 2005 IEEE 16th International Symposium on Personal, Indoor and Mobile Radio Communications (IEEE Cat. No. 05TH8889), P2712
[15]   Silencing of TMEM158 Inhibits Tumorigenesis and Multidrug Resistance in Colorectal Cancer [J].
Liu, Lihua ;
Zhang, Jiantao ;
Li, Shiquan ;
Yin, Libin ;
Tai, Jiandong .
NUTRITION AND CANCER-AN INTERNATIONAL JOURNAL, 2020, 72 (04) :662-671
[16]   Applications of Deep Reinforcement Learning in Communications and Networking: A Survey [J].
Luong, Nguyen Cong ;
Hoang, Dinh Thai ;
Gong, Shimin ;
Niyato, Dusit ;
Wang, Ping ;
Liang, Ying-Chang ;
Kim, Dong In .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2019, 21 (04) :3133-3174
[17]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[18]   Multi-Agent Deep Reinforcement Learning for Dynamic Power Allocation in Wireless Networks [J].
Nasir, Yasar Sinan ;
Guo, Dongning .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (10) :2239-2250
[19]  
Paszke A, 2019, ADV NEUR IN, V32
[20]   A Q-Learning Approach With Collective Contention Estimation for Bandwidth-Efficient and Fair Access Control in IEEE 802.11p Vehicular Networks [J].
Pressas, Andreas ;
Sheng, Zhengguo ;
Ali, Falah ;
Tian, Daxin .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (09) :9136-9150