State representation learning for control: An overview

被引:160
作者
Lesort, Timothee [1 ,2 ]
Diaz-Rodriguez, Natalia [2 ]
Goudou, Jean-Franois [1 ]
Filliat, David [2 ]
机构
[1] Thales, Vis Lab, Palaiseau, France
[2] Univ Paris Saclay, Inria FLOWERS Team, ENSTA ParisTech, U2IS, Palaiseau, France
基金
欧盟地平线“2020”;
关键词
State representation learning; Low dimensional embedding learning; Learning disentangled representations; Disentanglement of control factors; Robotics; Reinforcement learning; NETWORK;
D O I
10.1016/j.neunet.2018.07.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Representation learning algorithms are designed to learn abstract features that characterize data. State representation learning (SRL) focuses on a particular kind of representation learning where learned features are in low dimension, evolve through time, and are influenced by actions of an agent. The representation is learned to capture the variation in the environment generated by the agent's actions; this kind of representation is particularly suitable for robotics and control scenarios. In particular, the low dimension characteristic of the representation helps to overcome the curse of dimensionality, provides easier interpretation and utilization by humans and can help improve performance and speed in policy learning algorithms such as reinforcement learning. This survey aims at covering the state-of-the-art on state representation learning in the most recent years. It reviews different SRL methods that involve interaction with the environment, their implementations and their applications in robotics control tasks (simulated or real). In particular, it highlights how generic learning objectives are differently exploited in the reviewed algorithms. Finally, it discusses evaluation methods to assess the representation learned and summarizes current and future lines of research. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:379 / 392
页数:14
相关论文
共 75 条
  • [1] Achille A., 2017, ARXIV171103321
  • [2] [Anonymous], 2017, ARXIV171206560
  • [3] [Anonymous], 2016, CORR
  • [4] [Anonymous], NIPS DEEP REINF LEAR
  • [5] [Anonymous], 2013, Paladyn, Journal of Behavioral Robotics
  • [6] [Anonymous], 2002, TECH REP
  • [7] [Anonymous], ARXIV170907857
  • [8] [Anonymous], CORR
  • [9] [Anonymous], 2014, PROC 2 INT C LEARN R
  • [10] [Anonymous], 2016, ARXIV160604695