Improving Reinforcement Learning Performance through a Behavioral Psychology-Inspired Variable Reward Scheme

被引:0
作者
Rathore, Heena [1 ]
Griffith, Henry [2 ]
机构
[1] Texas State Univ, Dept Comp Sci, San Marcos, TX 78666 USA
[2] San Antonio Coll, Dept Engn, San Antonio, TX USA
来源
2023 IEEE INTERNATIONAL CONFERENCE ON SMART COMPUTING, SMARTCOMP | 2023年
关键词
variable reward; reinforcement learning; psychology; q-learning;
D O I
10.1109/SMARTCOMP58114.2023.00050
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) algorithms employ a fixed-ratio schedule which can lead to overfitting, where the agent learns to optimize for the specific rewards it receives, rather than learning the underlying task. Further, the agent can simply repeat the same actions that have worked in the past and do not explore different actions and strategies to see what works best. This leads to generalization issue, where the agent struggles to apply what it has learned to new, unseen situations. This can be particularly problematic in complex environments where the agent needs to learn to generalize from limited data. Introducing variable reward schedules in RL inspired from behavioral psychology can be more effective than traditional reward schemes because they can mimic real-world environments where rewards are not always consistent or predictable. This can also encourage an RL agent to explore more and become more adaptable to changes in the environment. The simulation results showed that variable reward scheme has faster learning rate as compared to fixed rewards.
引用
收藏
页码:210 / 212
页数:3
相关论文
共 49 条
[21]   Effect of immediate reward function on the performance of reinforcement learning-based energy management system [J].
Biswas, Atriya ;
Wang, Yue ;
Emadi, Ali .
2022 IEEE/AIAA TRANSPORTATION ELECTRIFICATION CONFERENCE AND ELECTRIC AIRCRAFT TECHNOLOGIES SYMPOSIUM (ITEC+EATS 2022), 2022, :1021-1026
[22]   Simulation Study of Reward Function to Improve the Performance of Chemical Process Control based on Reinforcement Learning [J].
Park J. ;
Shim J.H. ;
Lee J.M. .
Journal of Institute of Control, Robotics and Systems, 2022, 28 (12) :1185-1190
[23]   Improving Sample Efficiency Through Stability Enhancement in Deep-Reinforcement Learning [J].
Wang, Ziru ;
Jiang, Wanli ;
Peng, Ru ;
Kou, Qian ;
Wan, Lipeng ;
Lan, Xuguang .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025,
[24]   Maneuver Decision-Making through Automatic Curriculum Reinforcement Learning without Handcrafted Reward Functions [J].
Wei, Yujie ;
Zhang, Hongpeng ;
Wang, Yuan ;
Huang, Changqiang .
APPLIED SCIENCES-BASEL, 2023, 13 (16)
[25]   Population-based exploration in reinforcement learning through repulsive reward shaping using eligibility traces [J].
Bal, Melis Ilayda ;
Iyigun, Cem ;
Polat, Faruk ;
Aydin, Huseyin .
ANNALS OF OPERATIONS RESEARCH, 2025, 347 (02) :1059-1091
[26]   Improving unbalanced image classification through fine-tuning method of reinforcement learning [J].
Wang, Jin-Qiang ;
Guo, Lan ;
Jiang, Yuanbo ;
Zhang, Shengjie ;
Zhou, Qingguo .
APPLIED SOFT COMPUTING, 2024, 163
[27]   Improving Radiology Report Generation Quality and Diversity through Reinforcement Learning and Text Augmentation [J].
Parres, Daniel ;
Albiol, Alberto ;
Paredes, Roberto .
BIOENGINEERING-BASEL, 2024, 11 (04)
[28]   Optimal feature selection through reinforcement learning and fuzzy signature for improving classification accuracy [J].
Mansouri N. ;
Zandvakili A. ;
Javidi M.M. .
Multimedia Tools and Applications, 2025, 84 (10) :6931-6965
[29]   Improving Convolutional Neural Network-Based Webshell Detection Through Reinforcement Learning [J].
Wu, Yalun ;
Song, Minglu ;
Li, Yike ;
Tian, Yunzhe ;
Tong, Endong ;
Niu, Wenjia ;
Jia, Bowei ;
Huang, Haixiang ;
Li, Qiong ;
Liu, Jiqiang .
INFORMATION AND COMMUNICATIONS SECURITY (ICICS 2021), PT I, 2021, 12918 :368-383
[30]   DiffSkill: Improving Reinforcement Learning through diffusion-based skill denoiser for robotic [J].
Liu, Siao ;
Liu, Yang ;
Hu, Linqiang ;
Zhou, Ziqing ;
Xie, Yi ;
Zhao, Zhile ;
Li, Wei ;
Gan, Zhongxue .
KNOWLEDGE-BASED SYSTEMS, 2024, 300