A Hybrid Compact Neural Architecture for Visual Place Recognition

被引:51
作者
Chancan, Marvin [1 ,2 ]
Hernandez-Nunez, Luis [2 ,3 ,4 ]
Narendra, Ajay [5 ]
Barron, Andrew B. [5 ]
Milford, Michael [1 ]
机构
[1] Queensland Univ Technol, Sch Elect Engn & Comp Sci, Brisbane, Qld 4000, Australia
[2] Univ Nacl Ingn, Sch Mechatron Engn, Lima 15333, Rimac, Peru
[3] Harvard Univ, Ctr Brain Sci, Cambridge, MA 02138 USA
[4] Harvard Univ, Dept Phys, Cambridge, MA 02138 USA
[5] Macquarie Univ, Dept Biol Sci, Sydney, NSW 2109, Australia
基金
澳大利亚研究理事会;
关键词
Biomimetics; localization; visual-based navigation; MODEL; LOCALIZATION; NAVIGATION;
D O I
10.1109/LRA.2020.2967324
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
State-of-the-art algorithms for visual place recognition, and related visual navigation systems, can be broadly split into two categories: computer-science-oriented models including deep learning or image retrieval-based techniques with minimal biological plausibility, and neuroscience-oriented dynamical networks that model temporal properties underlying spatial navigation in the brain. In this letter, we propose a new compact and high-performing place recognition model that bridges this divide for the first time. Our approach comprises two key neural models of these categories: (1) FlyNet, a compact, sparse two-layer neural network inspired by brain architectures of fruit flies, Drosophila melanogaster, and (2) a one-dimensional continuous attractor neural network (CANN). The resulting FlyNet+CANN network incorporates the compact pattern recognition capabilities of our FlyNet model with the powerful temporal filtering capabilities of an equally compact CANN, replicating entirely in a hybrid neural implementation the functionality that yields high performance in algorithmic localization approaches like SeqSLAM. We evaluate our model, and compare it to three state-of-the-art methods, on two benchmark real-world datasets with small viewpoint variations and extreme environmental changes - achieving 87% AUC results under day to night transitions compared to 60% for Multi-Process Fusion, 46% for LoST-X and 1% for SeqSLAM, while being 6.5, 310, and 1.5 times faster, respectively.
引用
收藏
页码:993 / 1000
页数:8
相关论文
共 71 条
[1]  
[Anonymous], 2014, ARXIV14082927
[2]  
[Anonymous], 2017, P INT C LEARNING REP
[3]  
[Anonymous], 2013, P WORKSH LONG TERM A
[4]  
[Anonymous], 2015, PROC CVPR IEEE
[5]  
[Anonymous], 2014, P AUSTR C ROB AUT
[6]  
[Anonymous], P ROB SCI SYST
[7]  
[Anonymous], P ROB SCI SYST
[8]  
[Anonymous], P INT C LEARN REPR
[9]  
Arandjelovic R, 2018, IEEE T PATTERN ANAL, V40, P1437, DOI [10.1109/TPAMI.2017.2711011, 10.1109/CVPR.2016.572]
[10]   Vector-based navigation using grid-like representations in artificial agents [J].
Banino, Andrea ;
Barry, Caswell ;
Uria, Benigno ;
Blundell, Charles ;
Lillicrap, Timothy ;
Mirowski, Piotr ;
Pritzel, Alexander ;
Chadwick, Martin J. ;
Degris, Thomas ;
Modayil, Joseph ;
Wayne, Greg ;
Soyer, Hubert ;
Viola, Fabio ;
Zhang, Brian ;
Goroshin, Ross ;
Rabinowitz, Neil ;
Pascanu, Razvan ;
Beattie, Charlie ;
Petersen, Stig ;
Sadik, Amir ;
Gaffney, Stephen ;
King, Helen ;
Kavukcuoglu, Koray ;
Hassabis, Demis ;
Hadsell, Raia ;
Kumaran, Dharshan .
NATURE, 2018, 557 (7705) :429-+