A Multi-Gesture Interaction System Using a 3-D Iris Disk Model for Gaze Estimation and an Active Appearance Model for 3-D Hand Pointing

被引:61
作者
Reale, Michael J. [1 ]
Canavan, Shaun [1 ]
Yin, Lijun [1 ]
Hu, Kaoning [1 ]
Hung, Terry [2 ]
机构
[1] SUNY Binghamton, Dept Comp Sci, Binghamton, NY 13902 USA
[2] Corning Corp, Taichung 40763, Taiwan
基金
美国国家科学基金会;
关键词
Gaze estimation; hand tracking; human-computer interaction (HCI); TRACKING;
D O I
10.1109/TMM.2011.2120600
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we present a vision-based human-computer interaction system, which integrates control components using multiple gestures, including eye gaze, head pose, hand pointing, and mouth motions. To track head, eye, and mouth movements, we present a two-camera system that detects the face from a fixed, wide-angle camera, estimates a rough location for the eye region using an eye detector based on topographic features, and directs another active pan-tilt-zoom camera to focus in on this eye region. We also propose a novel eye gaze estimation approach for point-of-regard (POR) tracking on a viewing screen. To allow for greater head pose freedom, we developed a new calibration approach to find the 3-D eyeball location, eyeball radius, and fovea position. Moreover, in order to get the optical axis, we create a 3-D iris disk by mapping both the iris center and iris contour points to the eyeball sphere. We then rotate the fovea accordingly and compute the final, visual axis gaze direction. This part of the system permits natural, non-intrusive, pose-invariant POR estimation from a distance without resorting to infrared or complex hardware setups. We also propose and integrate a two-camera hand pointing estimation algorithm for hand gesture tracking in 3-D from a distance. The algorithms of gaze pointing and hand finger pointing are evaluated individually, and the feasibility of the entire system is validated through two interactive information visualization applications.
引用
收藏
页码:474 / 486
页数:13
相关论文
共 49 条
[31]   Visual interpretation of hand gestures for human-computer interaction: A review [J].
Pavlovic, VI ;
Sharma, R ;
Huang, TS .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (07) :677-695
[32]  
Pogalin E, 2007, THIRD INTERNATIONAL SYMPOSIUM ON 3D DATA PROCESSING, VISUALIZATION, AND TRANSMISSION, PROCEEDINGS, P57
[33]  
Press W. H., 2002, NUMERICAL RECIPES C
[34]   Reconstruction of Display and Eyes from a Single Image [J].
Schnieders, Dirk ;
Fu, Xingdou ;
Wong, Kwan-Yee K. .
2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, :1442-1449
[35]   Hand detection and tracking using pixel value distribution model for multiple-camera-based gesture interactions [J].
Utsumi, A ;
Tetsutani, N ;
Igi, S .
IEEE WORKSHOP ON KNOWLEDGE MEDIA NETWORKING, PROCEEDINGS, 2002, :31-36
[36]  
Vezhnevets V., 2003, P GRAPHICON, P81
[37]   Robust real-time face detection [J].
Viola, P ;
Jones, MJ .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2004, 57 (02) :137-154
[38]   Vision-Based Hand-Gesture Applications [J].
Wachs, Juan Pablo ;
Koelsch, Mathias ;
Stern, Helman ;
Edan, Yael .
COMMUNICATIONS OF THE ACM, 2011, 54 (02) :60-71
[39]  
Wang HG, 2001, IMAGE VISION COMPUT, V19, P891, DOI 10.1016/S0262-8856(01)00051-8
[40]   Using geometric properties of topographic manifold to detect and track eyes for human-computer interaction [J].
Wang, Jun ;
Yin, Lijun ;
Moore, Jason .
ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2007, 3 (04)