Unsupervised Model-Free Representation Learning

被引:0
|
作者
Ryabko, Daniil [1 ]
机构
[1] INRIA Lille, Lille, France
来源
关键词
PATTERN;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Numerous control and learning problems face the situation where sequences of high-dimensional highly dependent data are available, but no or little feedback is provided to the learner. In such situations it may be useful to find a concise representation of the input signal, that would preserve as much as possible of the relevant information. In this work we are interested in the problems where the relevant information is in the time-series dependence. Thus, the problem can be formalized as follows. Given a series of observations X-0 ,..., X-n coming from a large (high-dimensional) space chi, find a representation function f mapping chi to a finite space Y such that the series f(X-0) ,..., f(X-n) preserve as much information as possible about the original time-series dependence in X-0 ,..., X-n. For stationary time series, the function f can be selected as the one maximizing the time-series information I-infinity(f) = h(0)(f(X)) - h(infinity)(f(X)) where h0(f(X)) is the Shannon entropy of f(X-0) and h(infinity)(f(X)) is the entropy rate of the time series f(X-0) ,..., f(X-n),.... In this paper we study the functional I-infinity(f) from the learning-theoretic point of view. Specifically, we provide some uniform approximation results, and study the behaviour of I-infinity(f) in the problem of optimal control.
引用
收藏
页码:354 / 366
页数:13
相关论文
共 50 条
  • [31] Model-free stochastic learning in adaptive wireless networks
    Chandramouli, R.
    2007 IEEE SARNOFF SYMPOSIUM, 2007, : 462 - 466
  • [32] Model-Free Preference-Based Reinforcement Learning
    Wirth, Christian
    Fuernkranz, Johannes
    Neumann, Gerhard
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2222 - 2228
  • [33] Improving Optimistic Exploration in Model-Free Reinforcement Learning
    Grzes, Marek
    Kudenko, Daniel
    ADAPTIVE AND NATURAL COMPUTING ALGORITHMS, 2009, 5495 : 360 - 369
  • [34] Constrained model-free reinforcement learning for process optimization
    Pan, Elton
    Petsagkourakis, Panagiotis
    Mowbray, Max
    Zhang, Dongda
    del Rio-Chanona, Ehecatl Antonio
    COMPUTERS & CHEMICAL ENGINEERING, 2021, 154
  • [35] Model-free incremental learning of the semantics of manipulation actions
    Aksoy, Eren Erdal
    Tamosiunaite, Minija
    Woergoetter, Florentin
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2015, 71 : 118 - 133
  • [36] Model-Free μ Synthesis via Adversarial Reinforcement Learning
    Keivan, Darioush
    Havens, Aaron
    Seiler, Peter
    Dullerud, Geir
    Hu, Bin
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 3335 - 3341
  • [37] An adaptive clustering method for model-free reinforcement learning
    Matt, A
    Regensburger, G
    INMIC 2004: 8TH INTERNATIONAL MULTITOPIC CONFERENCE, PROCEEDINGS, 2004, : 362 - 367
  • [38] Model-Free Reinforcement Learning for Mean Field Games
    Mishra, Rajesh
    Vasal, Deepanshu
    Vishwanath, Sriram
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2023, 10 (04): : 2141 - 2151
  • [39] Counterfactual Credit Assignment in Model-Free Reinforcement Learning
    Mesnard, Thomas
    Weber, Theophane
    Viola, Fabio
    Thakoor, Shantanu
    Saade, Alaa
    Harutyunyan, Anna
    Dabney, Will
    Stepleton, Tom
    Heess, Nicolas
    Guez, Arthur
    Moulines, Eric
    Hutter, Marcus
    Buesing, Lars
    Munos, Remi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [40] Model-Free Learning of Safe yet Effective Controllers
    Bozkurt, Alper Kamil
    Wang, Yu
    Pajic, Miroslav
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 6560 - 6565