Multiple Binocular Cameras-Based Indoor Localization Technique Using Deep Learning and Multimodal Fusion

被引:6
作者
Yan, Jun [1 ]
Zhang, Yimei [2 ]
Kang, Bin [3 ]
Zhu, Wei-Ping [1 ,4 ]
Lun, Daniel Pak-Kong [5 ]
机构
[1] Nanjing Univ Posts & Telecommun, Coll Telecommun & Informat Engn, Nanjing 210003, Peoples R China
[2] Bank China, Hefei Branch, Software Ctr, Hefei 230000, Peoples R China
[3] Nanjing Univ Posts & Telecommun, Coll Internet Things, Nanjing 210003, Peoples R China
[4] Concordia Univ, Dept Elect & Comp Engn, Montreal, PQ H3G 1M8, Canada
[5] Hong Kong Polytech Univ, Dept Elect & Informat Engn, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Location awareness; Feature extraction; Cameras; Visualization; Estimation; Correlation; Indoor environment; Image-based indoor localization; feature extraction; convolutional neural network; multimodal fusion; COVID-19; CLASSIFICATION; IMAGE; CHALLENGES; DESCRIPTOR; EFFICIENT; NETWORK;
D O I
10.1109/JSEN.2021.3133488
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, an image based indoor localization technique using multiple binocular cameras is proposed by the deep learning and multimodal fusion. First, by taking advantage of the cross-model correlations between various multimodal images for localization purpose, the obtained images are concatenated to form two new modalities: three-channel gray image and three-channel depth image. Then, a two-stream convolutional neural network (CNN) is used for multimodal feature extraction which can ensure the independent of each image modality. Moreover, a decision-level fusion rule is proposed to fuse the extracted features with the linear weight sum method. At last, in order to make use of the feature correlation between each image modality, the fused feature is extracted once again by two convolutional max-pooling blocks. The shrinkage Loss based loss function is designed to obtain the position based regression function at last. Field tests show that the proposed algorithm can obtain more accurate position estimation than other existing image based localization approaches.
引用
收藏
页码:1597 / 1608
页数:12
相关论文
共 34 条
[1]  
Arroyo R, 2016, 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), P4656, DOI 10.1109/IROS.2016.7759685
[2]   New local difference binary image descriptor and algorithm for rapid and precise vehicle visual localisation [J].
Bibars, Ahmed ;
Mahroos, Mohsen .
IET COMPUTER VISION, 2019, 13 (05) :443-451
[3]   EmbraceNet: A robust deep learning architecture for multimodal classification [J].
Choi, Jun-Ho ;
Lee, Jong-Seok .
INFORMATION FUSION, 2019, 51 :259-270
[4]   Monocular Vision Aided Depth Map from RGB Images to Estimate of Localization and Support to Navigation of Mobile Robots [J].
da Silva, Suane Pires Pinheiro ;
Almeida, Jefferson Silva ;
Ohata, Elene Firmeza ;
Rodrigues, Joel J. P. C. ;
de Albuquerque, Victor Hugo C. ;
Reboucas Filho, Pedro Pedrosa .
IEEE SENSORS JOURNAL, 2020, 20 (20) :12040-12048
[5]  
Deng GQ, 2016, 2016 ASIA-PACIFIC CONFERENCE ON INTELLIGENT ROBOT SYSTEMS (ACIRS 2016), P187, DOI 10.1109/ACIRS.2016.7556210
[6]   SurfCNN: A Descriptor Accelerated Convolutional Neural Network for Image-Based Indoor Localization [J].
Elmoogy, Ahmed M. ;
Dong, Xiaodai ;
Lu, Tao ;
Westendorp, Robert ;
Tarimala, Kishore Reddy .
IEEE ACCESS, 2020, 8 :59750-59759
[7]   Visual simultaneous localization and mapping: a survey [J].
Fuentes-Pacheco, Jorge ;
Ruiz-Ascencio, Jose ;
Manuel Rendon-Mancha, Juan .
ARTIFICIAL INTELLIGENCE REVIEW, 2015, 43 (01) :55-81
[8]   CSI-Based Device-Free Wireless Localization and Activity Recognition Using Radio Image Features [J].
Gao, Qinhua ;
Wang, Jie ;
Ma, Xiaorui ;
Feng, Xueyan ;
Wang, Hongyu .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (11) :10346-10356
[9]  
Gribbon KT, 2004, INT SYM ELECT DES TE, P126
[10]   Energy-Efficient Indoor Localization of Smart Hand-Held Devices Using Bluetooth [J].
Gu, Yu ;
Ren, Fuji .
IEEE ACCESS, 2015, 3 :1450-1461