Influence on Learning of Various Conditions in Deep Q-Network

被引:0
|
作者
Niitsuma, Jun [1 ]
Osana, Yuko [1 ]
机构
[1] Tokyo Univ Technol, Sch Comp Sci, 1404-1 Katakura, Tokyo 1920982, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a learning method to acquire an appropriate action sequence by interaction with the environment without using a teacher signal, various researches on reinforcement learning have been carried out. On the other hand, recently, deep learning attracts attention as a method which has performance superior to conventional methods in the field of image recognition and speech recognition. Furthermore, the Deep Q-Network, which is a method can learn the action value in Q-Learning using the convolutional neural network, has been proposed. The Deep Q-Network is applied for many games without adjusting for each game, and it gains higher scores than humans in some games. In this paper, experiments in Deep Q-Network are carried out in the case when the time to be considered as an input is different, and the case when another action selection method is used and so on. As a result, we confirmed as follows: (1) There is a possibility that the performance may be improved by increasing the time to consider as an input to the Deep Q-Network, and (2) There is a high possibility that the probability that an action whose value is maximum in the action selection is chosen influences on learning.
引用
收藏
页码:1932 / 1935
页数:4
相关论文
共 50 条
  • [41] Decomposed Deep Q-Network for Coherent Task-Oriented Dialogue Policy Learning
    Zhao, Yangyang
    Yin, Kai
    Wang, Zhenyu
    Dastani, Mehdi
    Wang, Shihan
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 1380 - 1391
  • [42] An Improved Deep Q-Network with Convolution Block Attention
    Li, Shilin
    Qu, Junsuo
    Yang, Dan
    PROCEEDINGS OF 2022 INTERNATIONAL CONFERENCE ON AUTONOMOUS UNMANNED SYSTEMS, ICAUS 2022, 2023, 1010 : 2921 - 2929
  • [43] Manufacturing Resource Scheduling Based on Deep Q-Network
    ZHANG Yufei
    ZOU Yuanhao
    ZHAO Xiaodong
    Wuhan University Journal of Natural Sciences, 2022, 27 (06) : 531 - 538
  • [44] Inhomogeneous deep Q-network for time sensitive applications
    Chen, Xu
    Wang, Jun
    ARTIFICIAL INTELLIGENCE, 2022, 312
  • [45] Dynamic Parallel Machine Scheduling With Deep Q-Network
    Liu, Chien-Liang
    Tseng, Chun-Jan
    Huang, Tzu-Hsuan
    Wang, Jhih-Wun
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (11): : 6792 - 6804
  • [46] Deep Q-Network with Reinforcement Learning for Fault Detection in Cyber-Physical Systems
    Jayaprakash, J. Stanly
    Priyadarsini, M. Jasmine Pemeena
    Parameshachari, B. D.
    Karimi, Hamid Reza
    Gurumoorthy, Sasikumar
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (09)
  • [47] Meta Q-network: a combination of reinforcement learning and meta learning
    Lu, Min
    Wang, Yi
    Wang, Wenfeng
    INTERNATIONAL JOURNAL OF APPLIED NONLINEAR SCIENCE, 2022, 3 (03) : 179 - 188
  • [48] Application of Video Game Algorithm Based on Deep Q-Network Learning in Music Rhythm Teaching
    Zhang, Shilian
    Huang, Zheng
    Lang, Yalin
    INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2025, 22 (01) : 124 - 138
  • [49] Deep Reinforcement Learning Approach for X-rudder AUVs Fault Diagnosis Based on Deep Q-network
    Chen, Chuanfa
    Gao, Xiang
    Li, Yueming
    Chen, Xuezhi
    Cao, Jian
    Zhang, Yinghao
    JOURNAL OF MARINE SCIENCE AND APPLICATION, 2025,
  • [50] A Multitask and Kernel approach for Learning to Push Objects with a Target-Parameterized Deep Q-Network
    Ewerton, Marco
    Villamizar, Michael
    Jankowski, Julius
    Calinon, Sylvain
    Odobez, Jean-Marc
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 1878 - 1884