Learning task-specific sensing, control and memory policies

被引:0
|
作者
Rajendran, S [1 ]
Huber, M [1 ]
机构
[1] Univ Texas, Dept Comp & Engn, Arlington, TX 76019 USA
关键词
focus of attention; event memory; reinforcement learning;
D O I
10.1142/S0218213005002119
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
AI agents and robots that can adapt and handle multiple tasks in real time promise to be a powerful tool. To address the control challenges involved in such systems, the underlying control approach has to take into account the important sensory information. Modern sensors, however, can generate huge amounts of data, rendering the processing and representation of all sensor data in real time computationally intractable. This issue can be addressed by developing task-specific focus of attention strategies that limit the sensory data that is processed at any point in time to the data relevant for the given task. Alone, however, this mechanism is not adequate for solving complex tasks since the robot also has to maintain selected pieces of past information. This necessitates AI agents and robots to have the capability to remember significant past events that are required for task completion. This paper presents an approach that considers focus of attention as a problem of selecting controller and feature pairs to be processed at any given point in time to optimize system performance. This approach is further extended by incorporating short term memory and a learned memory management policy. The result is a system that learns control, sensing, and memory policies that are task-specific and adaptable to real world situations using feedback from the world in a reinforcement learning framework. The approach is illustrated using table cleaning, sorting, star-king, and copying tasks in the blocks world domain.
引用
收藏
页码:303 / 327
页数:25
相关论文
共 50 条
  • [21] Direct learning of improved control policies from historical plant data
    Alhazmi, Khalid
    Sarathy, S. Mani
    COMPUTERS & CHEMICAL ENGINEERING, 2024, 185
  • [22] Stable Inverse Reinforcement Learning: Policies From Control Lyapunov Landscapes
    Tesfazgi, Samuel
    Sprandl, Leonhard
    Lederer, Armin
    Hirche, Sandra
    IEEE OPEN JOURNAL OF CONTROL SYSTEMS, 2024, 3 : 358 - 374
  • [23] Evolving hierarchical memory-prediction machines in multi-task reinforcement learning
    Kelly, Stephen
    Voegerl, Tatiana
    Banzhaf, Wolfgang
    Gondro, Cedric
    GENETIC PROGRAMMING AND EVOLVABLE MACHINES, 2021, 22 (04) : 573 - 605
  • [24] Evolving hierarchical memory-prediction machines in multi-task reinforcement learning
    Stephen Kelly
    Tatiana Voegerl
    Wolfgang Banzhaf
    Cedric Gondro
    Genetic Programming and Evolvable Machines, 2021, 22 : 573 - 605
  • [25] Multi-Timescale Memory Dynamics Extend Task Repertoire in a Reinforcement Learning Network With Attention-Gated Memory
    Martinolli, Marco
    Gerstner, Wulfram
    Gilra, Aditya
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2018, 12
  • [26] Machine learning for industrial sensing and control: A survey and practical perspective
    Lawrence, Nathan P.
    Damarla, Seshu Kumar
    Kim, Jong Woo
    Tulsyan, Aditya
    Amjad, Faraz
    Wang, Kai
    Chachuat, Benoit
    Lee, Jong Min
    Huang, Biao
    Gopaluni, R. Bhushan
    CONTROL ENGINEERING PRACTICE, 2024, 145
  • [27] Reinforcement learning of a simple control task using the spike response model
    de Queiroz, Murilo Saraiva
    de Berredo, Roberto Coelho
    Braga, Antonio de Padua
    NEUROCOMPUTING, 2006, 70 (1-3) : 14 - 20
  • [28] Task Space Control of Hydraulic Construction Machines Using Reinforcement Learning
    Lee, Hyung Joo
    Brell-Cokcan, Sigrid
    HUMAN-FRIENDLY ROBOTICS 2023, HFR 2023, 2024, 29 : 181 - 195
  • [29] Decentralized Multi-Agent Control of a Manipulator in Continuous Task Learning
    Shahid, Asad Ali
    Sesin, Jorge Said Vidal
    Pecioski, Damjan
    Braghin, Francesco
    Piga, Dario
    Roveda, Loris
    APPLIED SCIENCES-BASEL, 2021, 11 (21):
  • [30] Impedance control of robot manipulator in contact task using machine learning
    Kim, Byungchan
    Park, Shinsuk
    2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 233 - +