Influence on Learning of Various Conditions in Deep Q-Network

被引:0
|
作者
Niitsuma, Jun [1 ]
Osana, Yuko [1 ]
机构
[1] Tokyo Univ Technol, Sch Comp Sci, 1404-1 Katakura, Tokyo 1920982, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a learning method to acquire an appropriate action sequence by interaction with the environment without using a teacher signal, various researches on reinforcement learning have been carried out. On the other hand, recently, deep learning attracts attention as a method which has performance superior to conventional methods in the field of image recognition and speech recognition. Furthermore, the Deep Q-Network, which is a method can learn the action value in Q-Learning using the convolutional neural network, has been proposed. The Deep Q-Network is applied for many games without adjusting for each game, and it gains higher scores than humans in some games. In this paper, experiments in Deep Q-Network are carried out in the case when the time to be considered as an input is different, and the case when another action selection method is used and so on. As a result, we confirmed as follows: (1) There is a possibility that the performance may be improved by increasing the time to consider as an input to the Deep Q-Network, and (2) There is a high possibility that the probability that an action whose value is maximum in the action selection is chosen influences on learning.
引用
收藏
页码:1932 / 1935
页数:4
相关论文
共 50 条
  • [1] Deep Deformable Q-Network: An Extension of Deep Q-Network
    Jin, Beibei
    Yang, Jianing
    Huang, Xiangsheng
    Khan, Dawar
    2017 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE (WI 2017), 2017, : 963 - 966
  • [2] Multiagent Learning and Coordination with Clustered Deep Q-Network
    Pageaud, Simon
    Deslandres, Veronique
    Lehoux, Vassilissa
    Hassas, Salima
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2156 - 2158
  • [3] Deep Reinforcement Learning. Case Study: Deep Q-Network
    Vrejoiu, Mihnea Horia
    ROMANIAN JOURNAL OF INFORMATION TECHNOLOGY AND AUTOMATIC CONTROL-REVISTA ROMANA DE INFORMATICA SI AUTOMATICA, 2019, 29 (03): : 65 - 78
  • [4] Deep Reinforcement Learning Pairs Trading with a Double Deep Q-Network
    Brim, Andrew
    2020 10TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2020, : 222 - 227
  • [5] A Deep Q-Network for the Beer Game: Deep Reinforcement Learning for Inventory Optimization
    Oroojiooyjadid, Afshin
    Nazari, MohammadReza
    Snyder, Lawrence, V
    Takac, Martin
    M&SOM-MANUFACTURING & SERVICE OPERATIONS MANAGEMENT, 2022, 24 (01) : 285 - 304
  • [6] Visual Analysis of Deep Q-network
    Seng, Dewen
    Zhang, Jiaming
    Shi, Xiaoying
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2021, 15 (03): : 853 - 873
  • [7] A Deep Q-Network for the Beer Game: Deep Reinforcement Learning for Inventory Optimization
    Oroojlooyjadid A.
    Nazari M.
    Snyder L.V.
    Takáč M.
    Manufacturing and Service Operations Management, 2022, 24 (01): : 285 - 304
  • [8] Stochastic Double Deep Q-Network
    Lv, Pingli
    Wang, Xuesong
    Cheng, Yuhu
    Duan, Ziming
    IEEE ACCESS, 2019, 7 : 79446 - 79454
  • [9] Active one-shot learning by a deep Q-network strategy
    Chen, Li
    Huang, Honglan
    Feng, Yanghe
    Cheng, Guangquan
    Huang, Jincai
    Liu, Zhong
    Neurocomputing, 2022, 383 : 324 - 335
  • [10] Timeslot Scheduling with Reinforcement Learning Using a Double Deep Q-Network
    Ryu, Jihye
    Kwon, Juhyeok
    Ryoo, Jeong-Dong
    Cheung, Taesik
    Joung, Jinoo
    ELECTRONICS, 2023, 12 (04)