Research and Application of Dynamic Neural Network Based on Reinforcement Learning

被引:0
|
作者
Yadav, Anil Kumar [1 ]
Sachan, Ajay Kumar [2 ]
机构
[1] IFTM Univ, Dept CSE, Moradabad, UP, India
[2] RITS, Bhopal, India
关键词
Dynamic neural network; Machine learning; Reinforcement learning; Neural network classifier; Agent; State Action;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dynamic neural network is became one of the most important approaches to eliminate Q table (look-up-table) of the machine intelligence. On the basis of comparison between general Artificial neural network and dynamic neural network, the development of dynamic neural network will be discussed. After the introduction of the theory and algorithms of reinforcement learning (RL), dynamic neural network will be applied as a basic decision taking unit (classifier neural network) in the form of a new technology. This will develop the application of reinforcement learning and provides a new idea for agent learning during real time operation. Use neural network for supervised learning, state as input/action as label. Reinforcement learning is widely use by different research field as intelligent control, robotics and neuroscience. It provides us possible solution within unknown environment. But at the same time we have to take care of its decision because RL can independently learn without prior knowledge or training and it take decision by learning experience through trail-and-error interaction with its environment. In this paper, we discussed a new dynamic neural network model and its algorithms in detail, together with the issues that arise in Q table (look-up-table). Additionally, the benefit and challenges of reinforcement learning are described along with some of the problem domains where the dynamic neural network techniques have been applied. In order to access dynamic neural network is to eliminate Q table (look-up-table) and agent should learn during real time operation.
引用
收藏
页码:931 / +
页数:2
相关论文
共 50 条
  • [21] Prediction of Effective Reinforcement Depth of Dynamic Compaction Based on Neural Network
    An Huize
    Gao Xing
    Feng Ruiling
    CIVIL ENGINEERING IN CHINA - CURRENT PRACTICE AND RESEARCH REPORT: PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON CIVIL ENGINEERING, 2011, : 3 - 7
  • [22] Application research of network learning algorithm based on neural network disturbance compensation in satellite attitude control
    John F.L.
    Dogra D.
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 (12) : 16513 - 16520
  • [23] Reinforcement Learning Based Neural Controllers for Dynamic Processes without Exploration
    Steege, Frank-Florian
    Hartmann, Andre
    Schaffernicht, Erik
    Gross, Horst-Michael
    ARTIFICIAL NEURAL NETWORKS-ICANN 2010, PT II, 2010, 6353 : 222 - +
  • [24] Category learning in a recurrent neural network with reinforcement learning
    Zhang, Ying
    Pan, Xiaochuan
    Wang, Yihong
    FRONTIERS IN PSYCHIATRY, 2022, 13
  • [25] A CGRA based Neural Network Inference Engine for Deep Reinforcement Learning
    Liang, Minglan
    Chen, Mingsong
    Wang, Zheng
    Sun, Jingwei
    2018 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2018), 2018, : 540 - 543
  • [26] Controlling chaos by GA-based reinforcement learning neural network
    IEEE
    不详
    IEEE Trans Neural Networks, 4 (846-859):
  • [27] Process Industry Scheduling Based on Graph Neural Network and Reinforcement Learning
    Wu, Zhenyu
    Wang, Yin
    39TH YOUTH ACADEMIC ANNUAL CONFERENCE OF CHINESE ASSOCIATION OF AUTOMATION, YAC 2024, 2024, : 1598 - 1603
  • [28] Memristor-based spiking neural network with online reinforcement learning
    Vlasov, Danila
    Minnekhanov, Anton
    Rybka, Roman
    Davydov, Yury
    Sboev, Alexander
    Serenko, Alexey
    Ilyasov, Alexander
    Demin, Vyacheslav
    NEURAL NETWORKS, 2023, 166 : 512 - 523
  • [30] Reinforcement learning control for nonlinear systems based on Elman neural network
    Wang, Xue-Song
    Cheng, Yu-Hu
    Yi, Jian-Qiang
    Wang, Wei-Qiang
    Zhongguo Kuangye Daxue Xuebao/Journal of China University of Mining and Technology, 2006, 35 (05): : 653 - 657