Interactions with 3D virtual objects in augmented reality using natural gestures

被引:1
|
作者
Dash, Ajaya Kumar [1 ,2 ]
Balaji, Koniki Venkata [2 ]
Dogra, Debi Prosad [2 ]
Kim, Byung-Gyu [3 ]
机构
[1] IIIT Bhubaneswar, Dept Comp Sci, Bhubaneswar 751003, India
[2] IIT Bhubaneswar, Sch Elect Sci, Bhubaneswar 752050, India
[3] Sookmyung Womens Univ, Seoul, South Korea
来源
VISUAL COMPUTER | 2024年 / 40卷 / 09期
关键词
Augmented reality; Deep learning; Interaction with virtual objects; HAND POSE ESTIMATION;
D O I
10.1007/s00371-023-03175-4
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Markers are the backbone of various cross-domain augmented reality (AR) applications available to the research community. However, the use of markers may limit anywhere augmentation. As smart sensors are being deployed across the large spectrum of consumer electronic (CE) products, it is becoming inevitable to rely upon natural gestures to render and interact with such CE products. It provides limitless options for augmented reality applications. This paper focuses on the use of the human palm as the natural target to render 3D virtual objects and interact with the virtual objects in a typical AR set-up. While printed markers are comparatively easier to detect for camera pose estimation, palm detection can be challenging as a replacement for physical markers. To mitigate this, we have used a two-stage palm detection model that helps to track multiple palms and the related key-points in real-time. The detected key-points help to calculate the camera pose before rendering the 3D objects. After successfully rendering the virtual objects, we use intuitive, one-handed (uni-manual) natural gestures to interact with them. A finite state machine (FSM) has been proposed to detect the change in gestures during interactions. We have validated the proposed interaction framework using a few well-known 3D virtual objects that are often used to demonstrate scientific concepts to students in various grades. Our framework has been found to perform better as compared to SOTA methods. Average precision of 96.5% (82.9% SSD+Mobilenet) and FPS of 58.27 (37.93 SSD+Mobilenet) have been achieved. Also, to widen the scope of the work, we have used a versatile gesture dataset and tested it with neural network-based models to detect gestures. The approach fits perfectly into the proposed AR pipeline at 46.83 FPS to work in real-time. This reveals that the proposed method has good potential to mitigate some of the challenges faced by the research community in the interactive AR space.
引用
收藏
页码:6449 / 6462
页数:14
相关论文
共 50 条
  • [31] Design of Collaborative 3D User Interfaces for Virtual and Augmented Reality
    Grandi, Jeronimo G.
    2017 IEEE VIRTUAL REALITY (VR), 2017, : 419 - 420
  • [32] CAPTURE, REPRESENTATION, AND RENDERING OF 3D AUDIO FOR VIRTUAL AND AUGMENTED REALITY
    Tashev, Ivan J.
    INTERNATIONAL JOURNAL ON INFORMATION TECHNOLOGIES AND SECURITY, 2019, 11 : 49 - 62
  • [33] Vision based 3D Gesture Tracking using Augmented Reality and Virtual Reality for Improved Learning Applications
    Mahayuddin, Zainal Rasyid
    Saif, A. F. M. Saifuddin
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (12) : 631 - 638
  • [34] Development of a customizable interactions questionnaire (CIQ) for evaluating interactions with objects in augmented/virtual reality
    Gao, Meiyuzi
    Boehm-Davis, Deborah A.
    VIRTUAL REALITY, 2023, 27 (02) : 699 - 716
  • [35] Development of a customizable interactions questionnaire (CIQ) for evaluating interactions with objects in augmented/virtual reality
    Meiyuzi Gao
    Deborah A. Boehm-Davis
    Virtual Reality, 2023, 27 : 699 - 716
  • [36] Virtual Borders in 3D: Defining a Drone's Movement Space Using Augmented Reality
    Riechmann, Malte
    Kirsch, Andre
    Koenig, Matthias
    Rexilius, Jan
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 15202 - 15208
  • [37] Nextmed: Automatic Imaging Segmentation, 3D Reconstruction, and 3D Model Visualization Platform Using Augmented and Virtual Reality
    Gonzalez Izard, Santiago
    Sanchez Torres, Ramiro
    Alonso Plaza, Oscar
    Juanes Mendez, Juan Antonio
    Jose Garcia-Penalvo, Francisco
    SENSORS, 2020, 20 (10)
  • [38] A 3D Interaction Technique for Selection and Manipulation Distant Objects in Augmented Reality
    Bellarbi, Abdelkader
    Zenati, Nadia
    Otmane, Samir
    Belghit, Hayet
    Benbelkacem, Samir
    Messaci, Assia
    Hamidia, Mahfoud
    2017 5TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING - BOUMERDES (ICEE-B), 2017,
  • [39] MoSART: Mobile Spatial Augmented Reality for 3D Interaction With Tangible Objects
    Cortes, Guillaume
    Marchand, Eric
    Brincin, Guillaume
    Lecuyer, Anatole
    FRONTIERS IN ROBOTICS AND AI, 2018, 5
  • [40] A Virtual Reality Interface for Interactions with Spatiotemporal 3D Data
    Quant, Hunter
    Banerjee, Sean
    Banerjee, Natasha Kholgade
    MULTIMEDIA MODELING, MMM 2018, PT II, 2018, 10705 : 343 - 347