Split Deep Q-Learning for Robust Object Singulation

被引:0
|
作者
Sarantopoulos, Iason [1 ]
Kiatos, Marios [1 ,2 ]
Doulgeri, Zoe [1 ]
Malassiotis, Sotiris [2 ]
机构
[1] Aristotle Univ Thessaloniki, Dept Elect & Comp Engn, Thessaloniki 54124, Greece
[2] Informat Technol Inst ITI, Ctr Res & Technol Hellas CERTH, Thessaloniki 57001, Greece
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/icra40945.2020.9196647
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Extracting a known target object from a pile of other objects in a cluttered environment is a challenging robotic manipulation task encountered in many robotic applications. In such conditions, the target object touches or is covered by adjacent obstacle objects, thus rendering traditional grasping techniques ineffective. In this paper, we propose a pushing policy aiming at singulating the target object from its surrounding clutter, by means of lateral pushing movements of both the neighboring objects and the target object until sufficient 'grasping room' has been achieved. To achieve the above goal we employ reinforcement learning and particularly Deep Q-learning (DQN) to learn optimal push policies by trial and error. A novel Split DQN is proposed to improve the learning rate and increase the modularity of the algorithm. Experiments show that although learning is performed in a simulated environment the transfer of learned policies to a real environment is effective thanks to robust feature selection. Finally, we demonstrate that the modularity of the algorithm allows the addition of extra primitives without retraining the model from scratch.
引用
收藏
页码:6225 / 6231
页数:7
相关论文
共 50 条
  • [31] Deep Reinforcement Learning with Sarsa and Q-Learning: A Hybrid Approach
    Xu, Zhi-xiong
    Cao, Lei
    Chen, Xi-liang
    Li, Chen-xi
    Zhang, Yong-liang
    Lai, Jun
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2018, E101D (09) : 2315 - 2322
  • [32] Adaptive Learning Recommendation Strategy Based on Deep Q-learning
    Tan, Chunxi
    Han, Ruijian
    Ye, Rougang
    Chen, Kani
    APPLIED PSYCHOLOGICAL MEASUREMENT, 2020, 44 (04) : 251 - 266
  • [33] Q-LEARNING
    WATKINS, CJCH
    DAYAN, P
    MACHINE LEARNING, 1992, 8 (3-4) : 279 - 292
  • [34] Object Goal Navigation using Data Regularized Q-Learning
    Gireesh, Nandiraju
    Kiran, D. A. Sasi
    Banerjee, Snehasis
    Sridharan, Mohan
    Bhowmick, Brojeshwar
    Krishna, Madhava
    2022 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2022, : 1092 - 1097
  • [35] A Deep Q-Learning Dynamic Spectrum Sharing Experiment
    Shea, John M.
    Wong, Tan F.
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [36] Stabilizing deep Q-learning with Q-graph-based bounds
    Hoppe, Sabrina
    Giftthaler, Markus
    Krug, Robert
    Toussaint, Marc
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2023, 42 (09): : 633 - 654
  • [37] Hyperparameter Optimization for Tracking with Continuous Deep Q-Learning
    Dong, Xingping
    Shen, Jianbing
    Wang, Wenguan
    Liu, Yu
    Shao, Ling
    Porikli, Fatih
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 518 - 527
  • [38] Deep Spatial Q-Learning for Infectious Disease Control
    Zhishuai Liu
    Jesse Clifton
    Eric B. Laber
    John Drake
    Ethan X. Fang
    Journal of Agricultural, Biological and Environmental Statistics, 2023, 28 : 749 - 773
  • [39] A Deep Q-Learning Approach for GPU Task Scheduling
    Luley, Ryan S.
    Qiu, Qinru
    2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [40] Deep Q-Learning for Dry Stacking Irregular Objects
    Liu, Yifang
    Shamsi, Seyed Mahdi
    Fang, Le
    Chen, Changyou
    Napp, Nils
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 1569 - 1576