Actor-Q based active perception learning system

被引:0
|
作者
Shibata, K [1 ]
Nishino, T [1 ]
Okabe, Y [1 ]
机构
[1] Oita Univ, Dept Elect & Elect Engn, Oita 8701192, Japan
来源
2001 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS I-IV, PROCEEDINGS | 2001年
关键词
Actor-Q architecture; reinforcement learning; neural network; active perception; visual sensor;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
An active perception learning system based on reinforcement learning is proposed. A novel reinforcement architecture called Actor-Q is employed in which Q-learning and Actor-Critic are combined. The system decides its actions according to Q-values. One of the actions is to move its sensor and the others are to make all answer of its recognition result, each of which corresponds to each pattern. When the sensor motion is selected the sensor moves according to thc actor's output signals. The Q-value for the sensor motion is trained by Q-learning. and the Actor is trained hy the Q-value for the sensor motion on behalf of the critic When one of the other actions is selected the system outputs the recognition result. When the recognition answer is correct, the Q-value is trained to be the upper limit of the Q-value, and when the answer is not correct, it is trained to be 0.0. The module to compute Q-value and the actor module are both consisted of a neural network and are trained by Error Back Propagation. The training signals are generated based on the above reinforcement learning. It was confirmed by some simulations using a visual sensor with non-uniform visual cells that the system moves its sensor to the place where it can recognize the presented pattern correctly. Even though the Q-value surface as a function of the sensor location has some local peaks. the sensor was not trapped and moved to the appropriate direction because the Q-value for the sensor motion becomes larger.
引用
收藏
页码:1000 / 1005
页数:6
相关论文
共 50 条
  • [1] Online Reinforcement Learning-Based Control of an Active Suspension System Using the Actor Critic Approach
    Fares, Ahmad
    Bani Younes, Ahmad
    APPLIED SCIENCES-BASEL, 2020, 10 (22): : 1 - 13
  • [2] Sequence Alignment with Q-Learning Based on the Actor-Critic Model
    Li, Yarong
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (05)
  • [3] Actor-critic learning based on fuzzy inference system
    Jouffe, L
    INFORMATION INTELLIGENCE AND SYSTEMS, VOLS 1-4, 1996, : 339 - 344
  • [4] Unsupervised Machine Learning Based Scalable Fusion for Active Perception
    Jayaratne, Madhura
    de Silva, Daswin
    Alahakoon, Damminda
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2019, 16 (04) : 1653 - 1663
  • [5] Active Perception and Reinforcement Learning
    Whitehead, Steven D.
    Ballard, Dana H.
    NEURAL COMPUTATION, 1990, 2 (04) : 409 - 419
  • [6] Q_learning based on active backup and memory mechanism
    Liu, Y
    Guo, MZ
    Yao, HX
    PROCEEDINGS OF THE 2004 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2004, : 271 - 275
  • [7] Study on Online εNN Q-Learning for Multi-Robot System Based on Community Perception Network
    Jin, Lu
    Yang, Yuequan
    Ni, Chunbo
    Kong, Yifei
    Cao, Zhiqiang
    PROCEEDINGS OF 2013 CHINESE INTELLIGENT AUTOMATION CONFERENCE: INTELLIGENT AUTOMATION, 2013, 254 : 861 - 868
  • [8] Neuro-Adaptive Learning Fuzzy-Based System for Actor Selection in Wireless Sensor and Actor Networks
    Kulla, Elis
    Elmazi, Donald
    Barolli, Leonard
    PROCEEDINGS OF 2016 10TH INTERNATIONAL CONFERENCE ON COMPLEX, INTELLIGENT, AND SOFTWARE INTENSIVE SYSTEMS (CISIS), 2016, : 488 - 493
  • [9] Student Perception of Active Learning Group in a Problem-Based Learning Curricular Environment
    Melovitz-Vasan C.
    Gentile M.
    Huff S.
    Vasan N.
    Medical Science Educator, 2018, 28 (1) : 195 - 201
  • [10] A deep learning-based binocular perception system
    SUN Zhao
    MA Chao
    WANG Liang
    MENG Ran
    PEI Shanshan
    Journal of Systems Engineering and Electronics, 2021, 32 (01) : 7 - 20