Real-time multimodal interaction in virtual reality-a case study with a large virtual interface

被引:7
作者
Cao, Lizhou [1 ]
Zhang, Huadong [1 ]
Peng, Chao [1 ]
Hansberger, Jeffrey T. [2 ]
机构
[1] Rochester Inst Technol, Golisano Coll Comp & Informat Sci, Sch Interact Games & Media, Lomb Mem Dr, Rochester, NY 14623 USA
[2] Army Res Lab, Huntsville, AL USA
关键词
Multimodal interaction; Virtual reality; Virtual interface; Interaction design; GESTURE; SPEECH;
D O I
10.1007/s11042-023-14381-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The values of VR and multimodal interaction technologies offer creative, virtual alternatives to manipulate a large data set in a virtual environment. This work presents the design, implementation, and evaluation of a real-time multimodal interaction framework that enables users to navigate, select, and move data elements. The novel multimodal fusion method is able to recognize freehand gestures, voice commands, and head gaze pointer in real-time and fuse them to meaningful actions for interacting with the virtual environment. We worked with imagery analysts who were defense and security experts on designing and testing the interface and interaction modalities. The evaluation of the framework was conducted with a case study of photo management tasks based on a real-world scenario. Users are able to select photos in a large virtual interface and move them to the bins on the left and right sides of the main view. The evaluation focuses on performance, task completion time, and users' experience amongst several different combinations of input modalities. The evaluation result shows it is important to make multiple interaction modalities available to users, and the interaction design implications are concluded based on the evaluation.
引用
收藏
页码:25427 / 25448
页数:22
相关论文
共 43 条
[1]  
[Anonymous], 2022, META USE QUEST LINK
[2]   Multimodal Machine Learning: A Survey and Taxonomy [J].
Baltrusaitis, Tadas ;
Ahuja, Chaitanya ;
Morency, Louis-Philippe .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :423-443
[3]   An introduction to 3-D user interface design [J].
Bowman, DA ;
Kruijff, E ;
LaViola, JJ ;
Poupyrev, I .
PRESENCE-TELEOPERATORS AND VIRTUAL ENVIRONMENTS, 2001, 10 (01) :96-108
[4]   Ellic's Exercise Class: promoting physical activities during exergaming with immersive virtual reality [J].
Cao, Lizhou ;
Peng, Chao ;
Dong, Yangzi .
VIRTUAL REALITY, 2021, 25 (03) :597-612
[5]   A Large Curved Display System in Virtual Reality for Immersive Data Interaction [J].
Cao, Lizhou ;
Peng, Chao ;
Hansberger, Jeffery T. .
2019 IEEE GAMES, ENTERTAINMENT, MEDIA CONFERENCE (GEM), 2019,
[6]   Usability and Engagement Study for a Serious Virtual Reality Game of Lunar Exploration Missions [J].
Cao, Lizhou ;
Peng, Chao ;
Hansberger, Jeffrey T. .
INFORMATICS-BASEL, 2019, 6 (04)
[7]   Gaze plus Gesture: Expressive, Precise and Targeted Free-Space Interactions [J].
Chatterjee, Ishan ;
Xiao, Robert ;
Harrison, Chris .
ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, :131-138
[8]  
Chun LM, 2015, 5TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING AND INFORMATICS 2015, P59, DOI 10.1109/ICEEI.2015.7352470
[9]   Multimodal Gaze Interaction for Creative Design [J].
Creed, Chris ;
Frutos-Pascual, Maite ;
Williams, Ian .
PROCEEDINGS OF THE 2020 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'20), 2020,
[10]  
Ferracani A., 2014, P 2014 ACM INT WORKS, P27, DOI [10.1145/2656719.2656725, DOI 10.1145/2656719.2656725]