Extending the Capabilities of Reinforcement Learning Through Curriculum: A Review of Methods and Applications

被引:0
作者
Kashish Gupta
Debasmita Mukherjee
Homayoun Najjaran
机构
[1] School of Engineering, University of British Columbia, Kelowna, BC
[2] School of Engineering, University of British Columbia, 1137 Alumni Ave., Kelowna, V1V 1V7, BC
基金
加拿大自然科学与工程研究理事会;
关键词
Curriculum learning; Curriculum-based reinforcement learning; Reinforcement learning; Reinforcement learning applications;
D O I
10.1007/s42979-021-00934-9
中图分类号
学科分类号
摘要
Reinforcement learning has long been advertised as the one with the capability to intelligently mimic and understand human learning and behavior. While the upshot of the field’s advances is not underrated, its applicability and extension to large, complex and highly dynamic environments remain inefficient, inaccurate or unsolved. Curriculum learning presents an intuitive yet elegant solution to these problems and when incorporated into the solution, provides a structured approach to alleviate some of the core challenges. As reinforcement learning framework proceeds to tackle harder challenges, it necessitates the study of essential support frameworks including curriculum learning. Through this paper, we review the current state-of-the-art in the field of curriculum-based reinforcement learning. We analyze and classify numerous scientific articles and present a summary of their methodologies and applications. In addition to the detailed review and analysis of the targeted algorithms, we summarise the current progress in the field by tabulating distinct identifying features of reviewed works with respect to their curriculum design methodology and applications. © The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd 2021.
引用
收藏
相关论文
共 138 条
[1]  
Abdulhai B., Kattan L., Reinforcement learning: introduction to theory and potential for transport applications, Can J Civ Eng, 30, 6, pp. 981-991, (2003)
[2]  
Abdulhai B., Pringle R., Karakoulas G.J., Reinforcement learning for True adaptive traffic signal control, J Transport Eng, 129, 3, pp. 278-285, (2003)
[3]  
Achiam J., Sastry S., Surprise-based intrinsic motivation for deep reinforcement, learning, (2017)
[4]  
Allgower E.L., Georg K., Numerical continuation methods, Springer series in computational mathematics, 13, (1990)
[5]  
Andrychowicz M., Wolski F., Ray A., Schneider J., Fong R., Welinder P., McGrew B., Tobin J., Abbeel P., Zaremba W., Hindsight experience replay, Tech. rep, (2017)
[6]  
Barto A.G., Intrinsic motivation and reinforcement learning, Intrinsically motivated learning in natural and artificial systems, vol 9783642323751, pp. 17-47, (2013)
[7]  
Barto A.G., Mahadevan S., Recent advances in hierarchical reinforcement learning, (2003)
[8]  
Bellemare M.G., Srinivasan S., Ostrovski G., Schaul T., Saxton D., Deepmind G., Munos R., Unifying count-based exploration and intrinsic motivation, Tech. rep, (2016)
[9]  
Bengio Y., Louradour J., Collobert R., Weston J., Curriculum learning, ACM international conference proceeding series, 382, pp. 1-8, (2009)
[10]  
Berseth G., Xie C., Cernek P., van De Panne M., Progressive reinforcement learning with distillation for multi-skilled motion control, 6Th International Conference on Learning Representations, ICLR 2018—conference Track Proceedings., 1802, (2018)