A Computer Vision Approach for Detecting Discrepancies in Map Textual Labels

被引:0
作者
Salama, Abdulrahman [1 ]
Elkamhawy, Mahmoud [1 ]
Hendawi, Abdeltawab [2 ]
Sabour, Adel [1 ]
Al-Masri, Eyhab [1 ]
Tan, Ming [3 ]
Agrawal, Vashutosh [3 ]
Prakash, Ravi [3 ]
Ali, Mohamed [1 ]
机构
[1] Univ Washington, Tacoma, WA 98402 USA
[2] Univ Rhode Isl, Kingston, RI USA
[3] Microsoft Corp, Redmond, WA 98052 USA
来源
35TH INTERNATIONAL CONFERENCE ON SCIENTIFIC AND STATISTICAL DATABASE MANAGEMENT, SSDBM 2023 | 2023年
关键词
detectron2; azure cognitive services; maps discrepancies; computer-vision; geospatial data; textual labels; faster-rcnn; neural networks;
D O I
10.1145/3603719.3603722
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Maps provide various sources of information. An important example of such information is textual labels such as cities, neighborhoods, and street names. Althoughwe treat this information as facts, and despite the massive effort done by providers to continuously improve their accuracy, this data is far from perfect. Discrepancies in textual labels rendered on the map are one of the major sources of inconsistencies across map providers. These discrepancies can have significant impacts on the reliability of the derived information and decision-making processes. Thus, it is important to validate the accuracy and consistency in such data. Most providers treat this data as their propriety data and it is not available to the public, thus we cannot compare the data directly. To address these challenges, we introduce a novel computer vision-based approach for automatically extracting and classifying labels based on the visual characteristics of the label, which indicates its category based on the format convention used by the specific map provider. Based on the extracted data, we detect the degree of discrepancies across map providers. We consider three map providers: Bing Maps, Google Maps, and OpenStreetMaps. The neural network we develop classifies the text labels with an accuracy up to 93% in all providers. We leverage our system to analyze randomly selected regions in different markets. The studied markets are USA, Germany, France, and Brazil. Experimental results and statistical analysis reveal the amount of discrepancies across map providers per region. We calculate the Jaccard distance between the extracted text sets for each pair of map providers, which represents the discrepancy percentage. Discrepancies percentages as high as 90% were found in some markets.
引用
收藏
页数:9
相关论文
共 20 条
  • [11] Jilani Musfira., 2013, P AG WORKSH ACT INT, P19
  • [12] Microsoft COCO: Common Objects in Context
    Lin, Tsung-Yi
    Maire, Michael
    Belongie, Serge
    Hays, James
    Perona, Pietro
    Ramanan, Deva
    Dollar, Piotr
    Zitnick, C. Lawrence
    [J]. COMPUTER VISION - ECCV 2014, PT V, 2014, 8693 : 740 - 755
  • [13] Feature Pyramid Networks for Object Detection
    Lin, Tsung-Yi
    Dollar, Piotr
    Girshick, Ross
    He, Kaiming
    Hariharan, Bharath
    Belongie, Serge
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 936 - 944
  • [14] Ranjan Ashish., 2021, OCR Using Computer Vision and Machine Learning, P83, DOI [DOI 10.1007/978-3-030-50641-4_6, 10.1007/978-3-030-50641-4_6]
  • [15] Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
    Ren, Shaoqing
    He, Kaiming
    Girshick, Ross
    Sun, Jian
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) : 1137 - 1149
  • [16] Schlegel I., 2021, AGILE GIScience Ser., V2, P1
  • [17] OSMRunner : A System for Exploring and Fixing OSM Connectivity
    Tabet, Fares
    Pentyala, Sikha
    Patel, Birva H.
    Hendawi, Abdeltawab
    Cao, Peiwei
    Song, Ashley
    Govind, Harsh
    Ali, Mohamed
    [J]. 2021 22ND IEEE INTERNATIONAL CONFERENCE ON MOBILE DATA MANAGEMENT (MDM 2021), 2021, : 193 - 200
  • [18] Wikipedia, 2023, Jaccard Index
  • [19] Wu Y., 2019, Detectron2
  • [20] Yao-Yi Chiang, 2010, Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR 2010), P3199, DOI 10.1109/ICPR.2010.783