Influence on Learning of Various Conditions in Deep Q-Network

被引:0
|
作者
Niitsuma, Jun [1 ]
Osana, Yuko [1 ]
机构
[1] Tokyo Univ Technol, Sch Comp Sci, 1404-1 Katakura, Tokyo 1920982, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a learning method to acquire an appropriate action sequence by interaction with the environment without using a teacher signal, various researches on reinforcement learning have been carried out. On the other hand, recently, deep learning attracts attention as a method which has performance superior to conventional methods in the field of image recognition and speech recognition. Furthermore, the Deep Q-Network, which is a method can learn the action value in Q-Learning using the convolutional neural network, has been proposed. The Deep Q-Network is applied for many games without adjusting for each game, and it gains higher scores than humans in some games. In this paper, experiments in Deep Q-Network are carried out in the case when the time to be considered as an input is different, and the case when another action selection method is used and so on. As a result, we confirmed as follows: (1) There is a possibility that the performance may be improved by increasing the time to consider as an input to the Deep Q-Network, and (2) There is a high possibility that the probability that an action whose value is maximum in the action selection is chosen influences on learning.
引用
收藏
页码:1932 / 1935
页数:4
相关论文
共 50 条
  • [31] Tuning Apex DQN: A Reinforcement Learning based Deep Q-Network Algorithm
    Ruhela, Dhani
    Ruhela, Amit
    PRACTICE AND EXPERIENCE IN ADVANCED RESEARCH COMPUTING 2024, PEARC 2024, 2024,
  • [32] Learning Q-network for Active Information Acquisition
    Jeong, Heejin
    Schlotfeldt, Brent
    Hassani, Hamed
    Morari, Manfred
    Lee, Daniel D.
    Pappas, George J.
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 6822 - 6827
  • [33] Deep Attention Q-Network for Personalized Treatment Recommendation
    Ma, Simin
    Lee, Junghwan
    Serban, Nicoleta
    Yang, Shihao
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 329 - 337
  • [34] Knowledge Induced Deep Q-Network for Robot Push and Grasp Manipulation Skills Learning
    Gui, Boxing
    Qian, Kun
    Chen, Shenghao
    Jing, Xingshuo
    Ma, Xudong
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 4078 - 4083
  • [35] Accurate Price Prediction by Double Deep Q-Network
    Feizi-Derakhshi, Mohammad-Reza
    Lotfimanesh, Bahram
    Amani, Omid
    INTELIGENCIA ARTIFICIAL-IBEROAMERICAN JOURNAL OF ARTIFICIAL INTELLIGENCE, 2024, 27 (74): : 12 - 21
  • [36] Train Scheduling with Deep Q-Network: A Feasibility Test
    Gong, Intaek
    Oh, Sukmun
    Min, Yunhong
    APPLIED SCIENCES-BASEL, 2020, 10 (23): : 1 - 14
  • [37] Autonomous Robot Navigation System with Learning Based on Deep Q-Network and Topological Maps
    Kato, Yuki
    Kamiyama, Koji
    Morioka, Kazuyuki
    2017 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), 2017, : 1040 - 1046
  • [38] A Framework of Hierarchical Deep Q-Network for Portfolio Management
    Gao, Yuan
    Gao, Ziming
    Hu, Yi
    Song, Sifan
    Jiang, Zhengyong
    Su, Jionglong
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, 2021, : 132 - 140
  • [39] Classification of Alzheimer's disease: application of a transfer learning deep Q-network method
    Ma, Huibin
    Wang, Yadan
    Hao, Zeqi
    Yu, Yang
    Jia, Xize
    Li, Mengting
    Chen, Lanfen
    EUROPEAN JOURNAL OF NEUROSCIENCE, 2024, 59 (08) : 2118 - 2127
  • [40] Learning to schedule dynamic distributed reconfigurable workshops using expected deep Q-network
    Yang, Shengluo
    Wang, Junyi
    Xu, Zhigang
    ADVANCED ENGINEERING INFORMATICS, 2024, 59