On the use of Deep Autoencoders for Efficient Embedded Reinforcement Learning

被引:7
作者
Prakash, Bharat [1 ]
Horton, Mark [1 ]
Waytowich, Nicholas R. [2 ]
Hairston, William David [2 ]
Oates, Tim [1 ]
Mohsenin, Tinoosh [1 ]
机构
[1] Univ Maryland Baltimore Cty, Baltimore, MD 21228 USA
[2] US Army, Res Lab, Washington, DC 20310 USA
来源
GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI | 2019年
关键词
Autoencoders; Neural networks; Reinforcement learning; Embedded devices; Deep learning;
D O I
10.1145/3299874.3319493
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In autonomous embedded systems, it is often vital to reduce the amount of actions taken in the real world and energy required to learn a policy. Training reinforcement learning agents from high dimensional image representations can be very expensive and time consuming. Autoencoders are deep neural network used to compress high dimensional data such as pixelated images into small latent representations. This compression model is vital to efficiently learn policies, especially when learning on embedded systems. We have implemented this model on the NVIDIA Jetson TX2 embedded GPU, and evaluated the power consumption, throughput, and energy consumption of the autoencoders for various CPU/GPU core combinations, frequencies, and model parameters. Additionally, we have shown the reconstructions generated by the autoencoder to analyze the quality of the generated compressed representation and also the performance of the reinforcement learning agent. Finally, we have presented an assessment of the viability of training these models on embedded systems and their usefulness in developing autonomous policies. Using autoencoders, we were able to achieve 4-5 x improved performance compared to a baseline RL agent with a convolutional feature extractor, while using less than 2W of power.
引用
收藏
页码:507 / 512
页数:6
相关论文
共 12 条
  • [1] [Anonymous], 2017, ARXIV170705173
  • [2] Goecks VG, 2019, AAAI CONF ARTIF INTE, P2462
  • [3] Ha D., 2018, ARXIV180310122, DOI DOI 10.5281/ZENODO.1207631
  • [4] Haarnoja T, 2018, PR MACH LEARN RES, V80
  • [5] JAFARI A, 2018, IEEE T CIRCUITS-I, V96, P1, DOI DOI 10.1016/J.FORPOL.2018.08.001
  • [6] Mnih V, 2016, PR MACH LEARN RES, V48
  • [7] Human-level control through deep reinforcement learning
    Mnih, Volodymyr
    Kavukcuoglu, Koray
    Silver, David
    Rusu, Andrei A.
    Veness, Joel
    Bellemare, Marc G.
    Graves, Alex
    Riedmiller, Martin
    Fidjeland, Andreas K.
    Ostrovski, Georg
    Petersen, Stig
    Beattie, Charles
    Sadik, Amir
    Antonoglou, Ioannis
    King, Helen
    Kumaran, Dharshan
    Wierstra, Daan
    Legg, Shane
    Hassabis, Demis
    [J]. NATURE, 2015, 518 (7540) : 529 - 533
  • [8] Raffin A., 2019, Learning to drive smoothly in minutes
  • [9] Schulman J, 2017, ARXIV170706347
  • [10] Sutton Richard S., Advances in Neural Information Processing Systems, V12, P1057