A Scoping Review of the Literature On Prosodic Elements Related to Emotional Speech in Human-Robot Interaction

被引:7
作者
Gasteiger, Norina [1 ,2 ]
Lim, JongYoon [1 ]
Hellou, Mehdi [1 ,3 ]
MacDonald, Bruce A. [1 ]
Ahn, Ho Seok [1 ]
机构
[1] Univ Auckland, Dept Elect Comp & Software Engn, Auckland 1142, New Zealand
[2] Univ Manchester, Sch Hlth Sci, Manchester, Lancs, England
[3] Sorbonne Univ, Facultes Sci & Ingenieries, Dept Informat, Paris, France
关键词
affective computing; speech; HRI; robotics; social robots; sentiment; EXPRESSION; FUTURE; WORDS; MODEL; TEXT;
D O I
10.1007/s12369-022-00913-x
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Background Sentiment expression and detection are crucial for effective and empathetic human-robot interaction. Previous work in this field often focuses on non-verbal emotion expression, such as facial expressions and gestures. Less is known about which specific prosodic speech elements are required in human-robot interaction. Our research question was: what prosodic elements are related to emotional speech in human-computer/robot interaction? Methods The scoping review was conducted in alignment with the Arksey and O'Malley methods. Literature was identified from the SCOPUS, IEEE Xplore, ACM Digital Library and PsycINFO databases in May 2021. After screening and de-duplication, data were extracted into an Excel coding sheet and summarised. Results Thirteen papers, published from 2012 to 2020 were included in the review. The most commonly used prosodic elements were tone/pitch (n = 8), loudness/volume (n = 6) speech speed (n = 4) and pauses (n = 3). Non-linguistic vocalisations (n = 1) were less frequently used. The prosodic elements were generally effective in helping to convey or detect emotion, but were less effective for negative sentiment (e.g., anger, fear, frustration, sadness and disgust). Discussion Future research should explore the effectiveness of commonly used prosodic elements (tone, loudness, speed and pauses) in emotional speech, using larger sample sizes and real-life interaction scenarios. The success of prosody in conveying negative sentiment to humans may be improved with additional non-verbal cues (e.g., coloured light or motion). More research is needed to determine how these may be combined with prosody and which combination is most effective in human-robot affective interaction.
引用
收藏
页码:659 / 670
页数:12
相关论文
共 50 条
[11]   Can social robots help children in healthcare contexts? A scoping review [J].
Dawe, Julia ;
Sutherland, Craig ;
Barco, Alex ;
Broadbent, Elizabeth .
BMJ PAEDIATRICS OPEN, 2019, 3 (01)
[12]   AN ARGUMENT FOR BASIC EMOTIONS [J].
EKMAN, P .
COGNITION & EMOTION, 1992, 6 (3-4) :169-200
[13]   A Multitask Approach to Continuous Five-Dimensional Affect Sensing in Natural Speech [J].
Eyben, Florian ;
Woellmer, Martin ;
Schuller, Bjoern .
ACM TRANSACTIONS ON INTERACTIVE INTELLIGENT SYSTEMS, 2012, 2 (01) :1-29
[14]   What Color Is Your Anger? Assessing Color-Emotion Pairings in English Speakers [J].
Fugate, Jennifer Marie Binzak ;
Franco, Courtny L. .
FRONTIERS IN PSYCHOLOGY, 2019, 10
[15]  
Gasteiger N., 2021, ROUTLEDGE SOCIAL SCI
[16]   Friends from the Future: A Scoping Review of Research into Robots and Computer Agents to Combat Loneliness in Older People [J].
Gasteiger, Norina ;
Loveys, Kate ;
Law, Mikaela ;
Broadbent, Elizabeth .
CLINICAL INTERVENTIONS IN AGING, 2021, 16 :941-971
[17]   Older adults' experiences and perceptions of living with Bomy, an assistive dailycare robot: a qualitative study [J].
Gasteiger, Norina ;
Ahn, Ho Seok ;
Fok, Christine ;
Lim, JongYoon ;
Lee, Christopher ;
MacDonald, Bruce A. ;
Kim, Geon Ha ;
Broadbent, Elizabeth .
ASSISTIVE TECHNOLOGY, 2022, 34 (04) :487-497
[18]  
Ghazi D., 2010, Proceedings of the NAACL HLT 2010 Workshop on Computational Approaches to Analysis and Generation of Emotion in Text, Los Angeles, California, P140
[19]  
Guizzo Eric, 2014, How Aldebaran Robotics Built its Friendly Humanoid Robot, Pepper
[20]  
Hsieh WF, 2020, J ROBOT MECHATRON, V32, P224