Feature Extraction in Q-Learning using Neural Networks

被引:0
作者
Zhu, Henghui [1 ]
Paschalidis, Ioannis Ch. [2 ,3 ,4 ]
Hasselmo, Michael E. [5 ]
机构
[1] Boston Univ, Ctr Informat & Syst Engn, Boston, MA 02215 USA
[2] Boston Univ, Dept Elect & Comp Engn, 8 St Marys St, Boston, MA 02215 USA
[3] Boston Univ, Div Syst Engn, 8 St Marys St, Boston, MA 02215 USA
[4] Boston Univ, Dept Biomed Engn, 8 St Marys St, Boston, MA 02215 USA
[5] Boston Univ, Ctr Syst Neurosci, Boston, MA 02215 USA
来源
2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC) | 2017年
关键词
Q-learning; reinforcement learning; Markov decision processes; neural networks;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Integrating deep neural networks with reinforcement learning has exhibited excellent performance in the literature, highlighting the ability of neural networks to extract features. This paper begins with a simple Markov decision process inspired from a cognitive task. We show that Q-learning, and approximate Q-learning using a linear function approximation fail in this task. Instead, we show that Q-learning combined with a neural network-based function approximator can learn the optimal policy. Motivated by this finding, we outline procedures that allow the use of a neural network to extract appropriate features, which can then be used in a Q-learning framework with a linear function approximation, obtaining performance similar to that observed using Q-learning with neural networks. Our work suggests that neural networks can be used as feature extractors in the context of Q-learning.
引用
收藏
页数:6
相关论文
共 50 条
[41]   Model based path planning using Q-Learning [J].
Sharma, Avinash ;
Gupta, Kanika ;
Kumar, Anirudha ;
Sharma, Aishwarya ;
Kumar, Rajesh .
2017 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2017, :837-842
[42]   BEAM MANAGEMENT SOLUTION USING Q-LEARNING FRAMEWORK [J].
Araujo, Daniel C. ;
de Almeida, Andre L. F. .
2019 IEEE 8TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP 2019), 2019, :594-598
[43]   Solving Twisty Puzzles Using Parallel Q-learning [J].
Hukmani, Kavish ;
Kolekar, Sucheta ;
Vobugari, Sreekumar .
ENGINEERING LETTERS, 2021, 29 (04) :1535-1543
[44]   An Autonomous Path Finding Robot Using Q-Learning [J].
Babu, Madhu ;
Krishna, Vamshi U. ;
Shahensha, S. K. .
PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND CONTROL (ISCO'16), 2016,
[45]   A novel contour extraction approach based on Q-learning [J].
Liang, Jun-Bin ;
Xu, Jian-Min .
PROCEEDINGS OF 2006 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2006, :3807-+
[46]   Q-learning for risk-sensitive control [J].
Borkar, VS .
MATHEMATICS OF OPERATIONS RESEARCH, 2002, 27 (02) :294-311
[47]   CHROMOSOME LOCATION AND FEATURE-EXTRACTION USING NEURAL NETWORKS [J].
TURNER, M ;
AUSTIN, J ;
ALLINSON, NM ;
THOMPSON, P .
IMAGE AND VISION COMPUTING, 1993, 11 (04) :235-239
[48]   Feature Extraction Using Neural Networks for Vietnamese Text Classification [J].
To Nguyen Phuoc Vinh ;
Ha Hoang Kha .
2021 INTERNATIONAL SYMPOSIUM ON ELECTRICAL AND ELECTRONICS ENGINEERING (ISEE 2021), 2021, :120-124
[49]   Autonomous Driving in Roundabout Maneuvers Using Reinforcement Learning with Q-Learning [J].
Garcia Cuenca, Laura ;
Puertas, Enrique ;
Fernandez Andres, Javier ;
Aliane, Nourdine .
ELECTRONICS, 2019, 8 (12)
[50]   Q-Learning Driven Routing for Aeronautical Ad-Hoc Networks [J].
Bilen, Tugce ;
Canberk, Berk .
PERVASIVE AND MOBILE COMPUTING, 2022, 87