Real-time Motion Generation for Imaginary Creatures Using Hierarchical Reinforcement Learning

被引:0
|
作者
Ogaki, Keisuke [1 ]
Nakamura, Masayoshi [1 ]
机构
[1] DWANGO Co Ltd, Tokyo, Japan
来源
SIGGRAPH'18: ACM SIGGRAPH 2018 STUDIO | 2018年
关键词
Reinforcement Learning; Q-Learning; Neural Network;
D O I
10.1145/3214822.3214826
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Describing the motions of imaginary original creatures is an essential part of animations and computer games. One approach to generate such motions involves finding an optimal motion for approaching a goal by using the creatures' body and motor skills. Currently, researchers are employing deep reinforcement learning (DeepRL) to find such optimal motions. Some end-to-end DeepRL approaches learn the policy function, which outputs target pose for each joint according to the environment. In our study, we employed a hierarchical approach with a separate DeepRL decision maker and simple exploration-based sequence maker, and an action token, through which these two layers can communicate. By optimizing these two functions independently, we can achieve a light, fast-learning system available on mobile devices. In addition, we propose another technique to learn the policy at a faster pace with the help of a heuristic rule. By treating the heuristic rule as an additional action token, we can naturally incorporate it via Q-learning. The experimental results show that creatures can achieve better performance with the use of both heuristics and DeepRL than by using them independently.
引用
收藏
页数:2
相关论文
共 50 条
  • [41] Toward Real-Time Decentralized Reinforcement Learning Using Finite Support Basis Functions
    Lobos-Tsunekawa, Kenzo
    Leottau, David L.
    Ruiz-del-Solar, Javier
    ROBOCUP 2017: ROBOT WORLD CUP XXI, 2018, 11175 : 95 - 107
  • [42] Applying Reinforcement Learning Method for Real-time Energy Management
    Dayani, Aida Borhan
    Fazlollahtabar, Hamed
    Ahmadiahangar, Roya
    Rosin, Argo
    Naderi, Mohammad Salay
    Bagheri, Mehdi
    2019 IEEE INTERNATIONAL CONFERENCE ON ENVIRONMENT AND ELECTRICAL ENGINEERING AND 2019 IEEE INDUSTRIAL AND COMMERCIAL POWER SYSTEMS EUROPE (EEEIC / I&CPS EUROPE), 2019,
  • [43] A Real-time Anonymous Traffic Detection based on Reinforcement Learning
    Liu, Dazhou
    Park, Younghee
    2024 IEEE 21ST CONSUMER COMMUNICATIONS & NETWORKING CONFERENCE, CCNC, 2024, : 574 - 577
  • [44] Stability Constrained Reinforcement Learning for Real-Time Voltage Control
    Shi, Yuanyuan
    Qu, Guannan
    Low, Steven
    Anandkumar, Anima
    Wierman, Adam
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 2715 - 2721
  • [45] RLStereo: Real-Time Stereo Matching Based on Reinforcement Learning
    Yang, Menglong
    Wu, Fangrui
    Li, Wei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 9442 - 9455
  • [46] Real Time Path Planning of Robot using Deep Reinforcement Learning
    Raajan, Jeevan
    Srihari, P., V
    Satya, Jayadev P.
    Bhikkaji, B.
    Pasumarthy, Ramkrishna
    IFAC PAPERSONLINE, 2020, 53 (02): : 15602 - 15607
  • [47] Hierarchical Gait Generation for Modular Robots Using Deep Reinforcement Learning
    Wang, Jiayu
    Hu, Chuxiong
    Zhu, Yu
    2021 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS (ICM), 2021,
  • [48] A Real-Time Intelligent Energy Management Strategy for Hybrid Electric Vehicles Using Reinforcement Learning
    Lee, Woong
    Jeoung, Haeseong
    Park, Dohyun
    Kim, Tacksu
    Lee, Heeyun
    Kim, Namwook
    IEEE ACCESS, 2021, 9 : 72759 - 72768
  • [49] Phased learning with hierarchical reinforcement learning in nonholonomic motion control
    Goto, Takaknuni
    Homma, Noriyasu
    Yoshizawa, Makoto
    Abe, Kenichi
    2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 2200 - +
  • [50] Hierarchical Memory System With STT-MRAM and SRAM to Support Transfer and Real-Time Reinforcement Learning in Autonomous Drones
    Yoon, Insik
    Anwar, Malik Aqeel
    Joshi, Rajiv V.
    Rakshit, Titash
    Raychowdhury, Arijit
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (03) : 485 - 497