Fully Embodied Conversational Avatars: Making Communicative Behaviors Autonomous

被引:56
作者
Cassell J. [1 ]
Vilhjálmsson H. [1 ]
机构
[1] MIT Media Laboratory, Cambridge, MA 02139
基金
美国国家科学基金会;
关键词
Avatars; Communicative behaviors; Embodied conversational agents; Lifelike;
D O I
10.1023/A:1010027123541
中图分类号
学科分类号
摘要
Although avatars may resemble communicative interface agents, they have for the most part not profited from recent research into autonomous embodied conversational systems. In particular, even though avatars function within conversational environments (for example, chat or games), and even though they often resemble humans (with a head, hands, and a body) they are incapable of representing the kinds of knowledge that humans have about how to use the body during communication. Humans, however, do make extensive use of the visual channel for interaction management where many subtle and even involuntary cues are read from stance, gaze, and gesture. We argue that the modeling and animation of such fundamental behavior is crucial for the credibility and effectiveness of the virtual interaction in chat. By treating the avatar as a communicative agent, we propose a method to automate the animation of important communicative behavior, deriving from work in conversation and discourse theory. BodyChat is a system that allows users to communicate via text while their avatars automatically animate attention, salutations, turn taking, back-channel feedback, and facial expression. An evaluation shows that users found an avatar with autonomous conversational behaviors to be more natural than avatars whose behaviors they controlled, and to increase the perceived expressiveness of the conversation. Interestingly, users also felt that avatars with autonomous communicative behaviors provided a greater sense of user control.
引用
收藏
页码:45 / 64
页数:19
相关论文
共 27 条
[1]  
Anderson D.B., Barrus J.W., Brogan D., Casey M., McKeown S., Sterns I., Waters R., Yerazunis W., Diamond Park and Spline: A Social Virtual Reality System with 3D Animation, Spoken Interaction, and Runtime Modifiability, Technical Report at MERL, (1996)
[2]  
Argyle M., Cook M., Gaze and Mutual Gaze, (1976)
[3]  
Argyle M., Ingham R., Alkema F., McCallin M., The Different Functions of Gaze, Semiotica, (1973)
[4]  
Bates J., Loyall A.B., Reilley W.S., Broad Agents, SIGART Bulletin, 4, 2, (1991)
[5]  
Benford S., Bowers J., Fahlen L.E., Greenhalgh C., Snowdon D., User Embodiment in Collaborative Virtual Environments, Proceedings of CHI'95, pp. 242-249, (1995)
[6]  
Blumberg B.M., Galyean T.A., Multi-Level Direction of Autonomous Creatures for Real-Time Virtual Environments, Proceedings of SIGGRAPH '95
[7]  
Cary M.S., The Role of Gaze in the Initiation of Conversation, Social Psychology, 41, 3, (1978)
[8]  
Cassell J., Embodied Conversation: Integrating Face and Gesture into Automatic Spoken Dialogue Systems, Spoken Dialogue Systems, (1999)
[9]  
Cassell J., Pelachaud C., Badler N., Steedman M., Achorn B., Becket T., Douville B., Prevost S., Stone M., Animated Conversation: Rule-based Generation of Facial Expression, Gesture & Spoken Intonation for Multiple Conversational Agents, Proceedings of SIGGRAPH '94, (1994)
[10]  
Cassell J., Thorisson K., The Power of a Nod and a Glance: Envelope vs. Emotional Feedback in Animated Conversational Agents, Journal of Applied Artificial Intelligence