Efficient and Scalable Object Localization in 3D on Mobile Device

被引:4
作者
Gupta, Neetika [1 ]
Khan, Naimul Mefraz [1 ]
机构
[1] Toronto Metropolitan Univ, Dept Elect Comp & Biomed Engn, Toronto, ON M5B 2K3, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
object localization; object detection; ARCore;
D O I
10.3390/jimaging8070188
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
Two-Dimensional (2D) object detection has been an intensely discussed and researched field of computer vision. With numerous advancements made in the field over the years, we still need to identify a robust approach to efficiently conduct classification and localization of objects in our environment by just using our mobile devices. Moreover, 2D object detection limits the overall understanding of the detected object and does not provide any additional information in terms of its size and position in the real world. This work proposes an object localization solution in Three-Dimension (3D) for mobile devices using a novel approach. The proposed method works by combining a 2D object detection Convolutional Neural Network (CNN) model with Augmented Reality (AR) technologies to recognize objects in the environment and determine their real-world coordinates. We leverage the in-built Simultaneous Localization and Mapping (SLAM) capability of Google's ARCore to detect planes and know the camera information for generating cuboid proposals from an object's 2D bounding box. The proposed method is fast and efficient for identifying everyday objects in real-world space and, unlike mobile offloading techniques, the method is well designed to work with limited resources of a mobile device.
引用
收藏
页数:12
相关论文
共 29 条
  • [1] [Anonymous], TENSORFLOW TENSORFLO
  • [2] [Anonymous], 2020, GOOGL BLOG REAL TIM
  • [3] Frugal Following: Power Thrifty Object Detection and Tracking for Mobile Augmented Reality
    Apicharttrisorn, Kittipat
    Ran, Xukan
    Chen, Jiasi
    Krishnamurthy, Srikanth, V
    Roy-Chowdhury, Amit K.
    [J]. PROCEEDINGS OF THE 17TH CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS (SENSYS '19), 2019, : 96 - 109
  • [4] ARCore, BUILD FUT
  • [5] Chen WL, 2015, PR MACH LEARN RES, V37, P2285
  • [6] Engineering at Meta, 2016, DEL REAL TIM PALM YO
  • [7] NestDNN: Resource-Aware Multi-Tenant On-Device Deep Learning for Continuous Mobile Vision
    Fang, Biyi
    Zeng, Xiao
    Zhang, Mi
    [J]. MOBICOM'18: PROCEEDINGS OF THE 24TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, 2018, : 115 - 127
  • [8] Han S, 2016, Harvard Yenching Ins, V101, P123
  • [9] Howard AG., 2017, ARXIV, DOI DOI 10.48550/ARXIV.1704.04861
  • [10] DeepMon: Mobile GPU-based Deep Learning Framework for Continuous Vision Applications
    Huynh, Loc N.
    Lee, Youngki
    Balan, Rajesh Krishna
    [J]. MOBISYS'17: PROCEEDINGS OF THE 15TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE SYSTEMS, APPLICATIONS, AND SERVICES, 2017, : 82 - 95