Event-Driven Stereo Visual Tracking Algorithm to Solve Object Occlusion

被引:38
作者
Camunas-Mesa, Luis A. [1 ,2 ]
Serrano-Gotarredona, Teresa [1 ,2 ]
Ieng, Sio-Hoi [3 ]
Benosman, Ryad [3 ]
Linares-Barranco, Bernabe [1 ,2 ]
机构
[1] CSIC, Inst Microelect Sevilla IMSE, CNM, Seville 41092, Spain
[2] Univ Seville, Seville 41092, Spain
[3] Univ Paris 06, UPMC, INSERM, UMR S968,CNRS 7210,Inst Vis, F-75005 Paris, France
关键词
Address event representation (AER); eventdriven processing; neuromorphic vision; object occlusion; object tracking; stereo vision; 3D RECONSTRUCTION; COMPUTER VISION; RECOGNITION; FEATURES;
D O I
10.1109/TNNLS.2017.2759326
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Object tracking is a major problem for many computer vision applications, but it continues to be computationally expensive. The use of bio-inspired neuromorphic event-driven dynamic vision sensors (DVSs) has heralded new methods for vision processing, exploiting reduced amount of data and very precise timing resolutions. Previous studies have shown these neural spiking sensors to be well suited to implementing single-sensor object tracking systems, although they experience difficulties when solving ambiguities caused by object occlusion. DVSs have also performed well in 3-D reconstruction in which event matching techniques are applied in stereo setups. In this paper, we propose a new event-driven stereo object tracking algorithm that simultaneously integrates 3-D reconstruction and cluster tracking, introducing feedback information in both tasks to improve their respective performances. This algorithm, inspired by human vision, identifies objects and learns their position and size in order to solve ambiguities. This strategy has been validated in four different experiments where the 3-D positions of two objects were tracked in a stereo setup even when occlusion occurred. The objects studied in the experiments were: 1) two swinging pens, the distance between which during movement was measured with an error of less than 0.5%; 2) a pen and a box, to confirm the correctness of the results obtained with a more complex object; 3) two straws attached to a fan and rotating at 6 revolutions per second, to demonstrate the high-speed capabilities of this approach; and 4) two people walking in a real-world environment.
引用
收藏
页码:4223 / 4237
页数:15
相关论文
共 28 条
[1]   Event-Based Visual Flow [J].
Benosman, Ryad ;
Clercq, Charles ;
Lagorce, Xavier ;
Ieng, Sio-Hoi ;
Bartolozzi, Chiara .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (02) :407-417
[2]  
Black J, 2002, IEEE WORKSHOP ON MOTION AND VIDEO COMPUTING (MOTION 2002), PROCEEDINGS, P169, DOI 10.1109/MOTION.2002.1182230
[3]   On the use of orientation filters for 3D reconstruction in event-driven stereo vision [J].
Camunas-Mesa, Luis A. ;
Serrano-Gotarredona, Teresa ;
Ieng, Sio H. ;
Benosman, Ryad B. ;
Linares-Barranco, Bernabe .
FRONTIERS IN NEUROSCIENCE, 2014, 8
[4]   Event-based 3D reconstruction from neuromorphic retinas [J].
Carneiro, Joao ;
Ieng, Sio-Hoi ;
Posch, Christoph ;
Benosman, Ryad .
NEURAL NETWORKS, 2013, 45 :27-38
[5]   MOTION-BASED RECOGNITION - A SURVEY [J].
CEDRAS, C ;
SHAH, M .
IMAGE AND VISION COMPUTING, 1995, 13 (02) :129-155
[6]  
Cohen I., 1999, P IEEE C COMP VIS PA, P325
[7]   A real-time computer vision system for vehicle tracking and traffic surveillance [J].
Coifman, B ;
Beymer, D ;
McLauchlan, P ;
Malik, J .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 1998, 6 (04) :271-288
[8]   Robotic goalie with 3 ms reaction time at 4% CPU load using event-based dynamic vision sensor [J].
Delbruck, Tobi ;
Lang, Manuel .
FRONTIERS IN NEUROSCIENCE, 2013, 7
[9]   Object recognition and tracking for remote video surveillance [J].
Foresti, GL .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 1999, 9 (07) :1045-1062
[10]  
Gómez-Rodríguez F, 2010, IEEE INT SYMP CIRC S, P1399, DOI 10.1109/ISCAS.2010.5537277