Multiparty Interaction Understanding Using Smart Multimodal Digital Signage

被引:13
|
作者
Tung, Tony [1 ,2 ]
Gomez, Randy [1 ,3 ]
Kawahara, Tatsuya
Matsuyama, Takashi [2 ]
机构
[1] Kyoto Univ, Acad Ctr Media Studies, Kyoto 6068501, Japan
[2] Kyoto Univ, Grad Sch Informat, Kyoto 6068501, Japan
[3] Honda Res Inst Japan, Wako, Saitama 6068501, Japan
关键词
Human-machine system; multimodal interaction dynamics; multiparty interaction; smart digital signage; RECOGNITION;
D O I
10.1109/THMS.2014.2326873
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a novel multimodal system designed for multi-party human-human interaction analysis. The design of human-machine interfaces for multiple users is challenging because simultaneous processing of actions and reactions have to be consistent. The proposed system consists of a large display equipped with multiple sensing devices: microphone array, HD video cameras, and depth sensors. Multiple users positioned in front of the panel freely interact using voice or gesture while looking at the displayed content, without wearing any particular devices (such as motion capture sensors or head mounted devices). Acoustic and visual information is captured and processed jointly using established and state-of-the-art techniques to obtain individual speech and gaze direction. Furthermore, a new framework is proposed to model A/V multimodal interaction between verbal and nonverbal communication events. Dynamics of audio signals obtained from speaker diarization and head poses extracted from video images are modeled using hybrid dynamical systems (HDS). We show that HDS temporal structure characteristics can be used for multimodal interaction level estimation, which is useful feedback that can help to improve multi-party communication experience. Experimental results using synthetic and real-world datasets of group communication such as poster presentations show the feasibility of the proposed multimodal system.
引用
收藏
页码:625 / 637
页数:13
相关论文
共 16 条
  • [1] Multiparty interaction: a multimodal perspective on relevance
    Norris, S
    DISCOURSE STUDIES, 2006, 8 (03) : 401 - 421
  • [2] Multimodal Multiparty Social Interaction with the Furhat Head
    Al Moubayed, Samer
    Skantze, Gabriel
    Beskow, Jonas
    Stefanov, Kalin
    Gustafson, Joakim
    ICMI '12: PROCEEDINGS OF THE ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2012, : 293 - 294
  • [3] A Design of Smart Beaker Structure and Interaction Paradigm Based on Multimodal Fusion Understanding
    Dong, Di
    Feng, Zhiquan
    Yuan, Jie
    Meng, Xin
    Meng, Junhong
    Kong, Dan
    IEEE ACCESS, 2020, 8 : 173766 - 173778
  • [4] Interacting with Digital Signage Using Hand Gestures
    Chen, Qing
    Malric, Francois
    Zhang, Yi
    Abid, Muhammad
    Cordeiro, Albino
    Petriu, Emil M.
    Georganas, Nicolas D.
    IMAGE ANALYSIS AND RECOGNITION, PROCEEDINGS, 2009, 5627 : 347 - +
  • [5] Toward Better Understanding of Engagement in Multiparty Spoken Interaction with Children
    Al Moubayed, Samer
    Lehman, Jill Fain
    ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, : 211 - 218
  • [6] Who Speaks Next? Turn Change and Next Speaker Prediction in Multimodal Multiparty Interaction
    Malik, Usman
    Saunier, Julien
    Funakoshi, Kotaro
    Pauchet, Alexandre
    2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2020, : 349 - 354
  • [7] Human-Robot Collaborative Tutoring using Multiparty Multimodal Spoken Dialogue
    Al Moubayed, Samer
    Beskow, Jonas
    Bollepalli, Bajibabu
    Gustafson, Joakim
    Hussen-Abdelaziz, Ahmed
    Johansson, Martin
    Koutsombogera, Maria
    Lopes, Jose David
    Novikova, Jekaterina
    Oertel, Catharine
    Skantze, Gabriel
    Stefanov, Kalin
    Varol, Gul
    HRI'14: PROCEEDINGS OF THE 2014 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2014, : 112 - 113
  • [8] A Multimodal Human-Computer Interaction for Smart Learning System
    Alzubi, Tareq Mahmod
    Alzubi, Jafar A.
    Singh, Ashish
    Alzubi, Omar A.
    Subramanian, Murali
    INTERNATIONAL JOURNAL OF HUMAN-COMPUTER INTERACTION, 2025, 41 (03) : 1718 - 1728
  • [9] Smile Intensity Detection in Multiparty Interaction using Deep Learning
    Witzig, Philine
    Kennedy, James
    Segalin, Cristina
    2019 8TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS (ACIIW), 2019, : 168 - 174
  • [10] Performance analysis of smart digital signage system based on software-defined IoT and invisible image sensor communication
    Hossain, Mohammad Arif
    Islam, Amirul
    Le, Nam Tuan
    Lee, Yong Tae
    Lee, Hyun Woo
    Jang, Yeong Min
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2016, 12 (07): : 1 - 14