Gender and accent stereotypes in communication with an intelligent virtual assistant

被引:0
|
作者
Piercy, Cameron W. [1 ]
Montgomery-Vestecka, Gretchen [2 ]
Lee, Sun Kyong [3 ]
机构
[1] Univ Kansas, Dept Commun Studies, Lawrence, KS USA
[2] Univ Oklahoma, Dept Commun, Norman, OK USA
[3] Korea Univ, Sch Media & Commun, 145 Anamro, Seoul 02841, South Korea
关键词
Intelligent virtual assistant; Trust; Accents; Stereotype; Gender; Experiment; RESPONSES; PERSONALITY; MACHINES; ROBOTS; TRUST;
D O I
10.1016/j.ijhcs.2024.103407
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
People are using intelligent virtual assistants (IVAs) more than ever before. Today's IVAs can be customized with unique voices including both gender and accent cues. Following evidence that people treat others differently based on their gender and accent, we ask: How do gender and accent of Siri, an IVA, affect users' trust? Students from two institutions (N = 270) participated in a two (Siri's voice gender: male or female) by two (Siri's voice accent: American or Indian) by two (task type: social or functional) fully crossed experiment, including a supplemental quasi-experimental condition for gender match between participants' and Siri's voice. Results show little effect for gender or accent alone, but the functional tasks condition received higher ratings in reliability, understandability, and faith dimensions of trust. Interactions reveal nuanced effects regarding gender match and varying across accent types. Implications for human-machine communication, in particular differences between human-human and human-machine interaction scripts are presented.
引用
收藏
页数:12
相关论文
共 50 条