CLUE-AI: A Convolutional Three-Stream Anomaly Identification Framework for Robot Manipulation

被引:4
作者
Altan, Dogan [1 ,2 ]
Sariel, Sanem [1 ]
机构
[1] Istanbul Tech Univ, Fac Comp & Informat Engn, Artificial Intelligence & Robot Lab, TR-34469 Istanbul, Turkiye
[2] Dept Validat Intelligence Autonomous Software Syst, Simula Res Lab, Artificial Intelligence & Robot Lab, N-0164 Oslo, Norway
关键词
Cognitive robots; robot manipulation; robot safety; anomaly identification; robot learning;
D O I
10.1109/ACCESS.2023.3276297
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Despite the great promise of service robots in everyday tasks, many roboethics issues remain to be addressed before these robots can physically work in human environments. Robot safety is one of the essential concerns for roboethics which is not just a design-time issue. It is also crucial to devise the required onboard monitoring and control strategies to enable robots to be aware of and react to anomalies (i.e., unexpected deviations from intended outcomes) that arise during their operations in the real world. The detection and identification of these anomalies is an essential first step toward fulfilling these requirements. Although several architectures have been proposed for anomaly detection; identification has not yet been thoroughly investigated. This task is challenging since indicators may appear long before anomalies are detected. In this paper, we propose a ConvoLUtional threE-stream Anomaly Identification (CLUE-AI) framework to address this problem. The framework fuses visual, auditory and proprioceptive data streams to identify everyday object manipulation anomalies. A stream of 2D images gathered through an RGB-D camera placed on the head of the robot is processed within a self-attention-enabled visual stage to capture visual anomaly indicators. The auditory modality provided by the microphone placed on the robot's lower torso is processed within a designed convolutional neural network (CNN) in the auditory stage. Last, the force applied by the gripper and the gripper state is processed within a CNN to obtain proprioceptive features. These outputs are then combined with a late fusion scheme. Our novel three-stream framework design is analyzed on everyday object manipulation tasks with a Baxter humanoid robot in a semi-structured setting. The results indicate that CLUE-AI achieves an f-score of 94%, outperforming the other baselines in classifying anomalies.
引用
收藏
页码:48347 / 48357
页数:11
相关论文
共 52 条
[1]  
Abid A, 2015, INT CONF COMP SCI ED, P8, DOI 10.1109/ICCSE.2015.7250209
[2]  
Ak AC, 2019, AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, P1767
[3]  
Altan Dogan, 2014, 6th International Conference on Agents and Artificial Intelligence (ICAART 2014). Proceedings, P299
[4]  
Altan D., 2014, P 27 INT FLAIRS C, P1
[5]  
Altan D., 2016, PROC IEEERSJ INT C I, P1
[6]   What went wrong? Identification of everyday object manipulation anomalies [J].
Altan, Dogan ;
Sariel, Sanem .
INTELLIGENT SERVICE ROBOTICS, 2021, 14 (02) :215-234
[7]  
[Anonymous], 2016, ANN C PHM SOC, DOI DOI 10.36001/PHMCONF.2016.V8I1.2551
[8]  
[Anonymous], 2015, P 14 PYTH SCI C, DOI 10.25080/Majora-7b98e3ed-003
[9]  
Bowkett J., 2018, P REPR COMPL WORLD P, P1
[10]   Follow-up analysis of mobile robot failures [J].
Carlson, J ;
Murphy, RR ;
Nelson, A .
2004 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1- 5, PROCEEDINGS, 2004, :4987-4994