SDE2D: Semantic-Guided Discriminability Enhancement Feature Detector and Descriptor

被引:0
作者
Li, Jiapeng [1 ,2 ]
Zhang, Ruonan [3 ]
Li, Ge [4 ]
Li, Thomas H. [5 ]
机构
[1] Open Univ China, Beijing 100039, Peoples R China
[2] Peking Univ, Sch Elect & Comp Engn, Shenzhen Grad Sch, Shenzhen 518055, Peoples R China
[3] Ningxia Univ, Sch Adv Interdisciplinary Studies, Zhongwei 755000, Peoples R China
[4] Peking Univ, Sch Elect & Comp Engn SECE, Shenzhen Grad Sch, Guangdong Prov Key Lab Ultra High Definit Immers M, Shenzhen 518055, Peoples R China
[5] Peking Univ, Sch Elect & Comp Engn, Shenzhen Grad Sch, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Semantics; Detectors; Semantic segmentation; Neural networks; Feature detection; Vectors; Training; Decoding; Visualization; Keypoint detection; feature descriptor; feature discriminability; image matching; visual localization; ROBUST;
D O I
10.1109/TMM.2024.3521748
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Local feature detectors and descriptors serve various computer vision tasks, such as image matching, visual localization, and 3D reconstruction. To address the extreme variations of rotation and light in the real world, most detectors and descriptors capture as much invariance as possible. However, these methods ignore feature discriminability and perform poorly in indoor scenes. Indoor scenes have too many weak-textured and even repeatedly textured regions, so it is necessary for the extracted features to possess sufficient discriminability. Therefore, we propose a semantic-guided method (called SDE2D) enhancing feature discriminability to improve the performance of descriptors for indoor scenes. We develop a kind of semantic-guided discriminability enhancement (SDE) loss function that uses semantic information from indoor scenes. To the best of our knowledge, this is the first deep research that applies semantic segmentation to enhance discriminability. In addition, we design a novel framework that allows semantic segmentation network to be well embedded as a module in the overall framework and provides guidance information for training. Besides, we explore the impact of different semantic segmentation models on our method. The experimental results on indoor scenes datasets demonstrate that the proposed SDE2D performs well compared with the state-of-the-art models.
引用
收藏
页码:275 / 286
页数:12
相关论文
共 56 条
[1]   HPatches: A benchmark and evaluation of handcrafted and learned local descriptors [J].
Balntas, Vassileios ;
Lenc, Karel ;
Vedaldi, Andrea ;
Mikolajczyk, Krystian .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3852-3861
[2]   Key.Net: Keypoint Detection by Handcrafted and Learned CNN Filters Revisited [J].
Barroso-Laguna, Axel ;
Mikolajczyk, Krystian .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (01) :698-711
[3]   Speeded-Up Robust Features (SURF) [J].
Bay, Herbert ;
Ess, Andreas ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2008, 110 (03) :346-359
[4]   Reinforced Feature Points: Optimizing Feature Detection and Description for a High-Level Task [J].
Bhowmik, Aritra ;
Gumhold, Stefan ;
Rother, Carsten ;
Brachmann, Eric .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :4947-4956
[5]   Geometry-Aware Learning of Maps for Camera Localization [J].
Brahmbhatt, Samarth ;
Gu, Jinwei ;
Kim, Kihwan ;
Hays, James ;
Kautz, Jan .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2616-2625
[6]   Cascaded Parallel Filtering for Memory-Efficient Image-Based Localization [J].
Cheng, Wentao ;
Lin, Weisi ;
Chen, Kan ;
Zhang, Xinfeng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1032-1041
[7]  
Choy CB, 2016, ADV NEUR IN, V29
[8]   ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes [J].
Dai, Angela ;
Chang, Angel X. ;
Savva, Manolis ;
Halber, Maciej ;
Funkhouser, Thomas ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2432-2443
[9]   SuperPoint: Self-Supervised Interest Point Detection and Description [J].
DeTone, Daniel ;
Malisiewicz, Tomasz ;
Rabinovich, Andrew .
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, :337-349
[10]   D2-Net: A Trainable CNN for Joint Description and Detection of Local Features [J].
Dusmanu, Mihai ;
Rocco, Ignacio ;
Pajdla, Tomas ;
Pollefeys, Marc ;
Sivic, Josef ;
Torii, Akihiko ;
Sattler, Torsten .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :8084-8093