共 13 条
A Huber reward function-driven deep reinforcement learning solution for cart-pole balancing problem
被引:3
|作者:
Mishra, Shaili
[1
]
Arora, Anuja
[1
]
机构:
[1] Jaypee Inst Informat Technol, Dept Comp Sci Engn & Informat Technol, Noida, India
关键词:
Reinforcement learning;
Cart-pole problem;
Q-learning;
Deep Q learning;
DQN;
Double DQN;
INVERTED PENDULUM;
DESIGN;
D O I:
10.1007/s00521-022-07606-6
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Lots of learning tasks require experience learning based on activities performed in real scenarios which are affected by environmental factors. Therefore, real-time systems demand a model to learn from working experience-such as physical object properties-driven system models, trajectory prediction, and Atari games. This experience-driven learning model uses reinforcement learning which is considered as an important research topic and needs problem-specific reasoning model simulation. In this research paper, cart-pole balancing problem is selected as a problem where the system learns using Q-learning and Deep Q network reinforcement learning approaches. Pragmatic foundation of cart-pole problem and its solution with the help of Q learning and DQN reinforcement learning model are validated, and a comparison of achieved outcome in the form of accuracy and fast convergence is presented. An unexperienced Huber loss function is applied on cart-pole balancing problem, and results are in favor of Huber loss function in comparison with mean-squared error loss function. Hence, experimental study suggests the use of DQN with Huber loss reward function for fast learning and convergence of cart pole in balanced condition.
引用
收藏
页码:16705 / 16722
页数:18
相关论文