Learning Person-Person Interaction in Collective Activity Recognition

被引:33
作者
Chang, Xiaobin [1 ,2 ]
Zheng, Wei-Shi [1 ,3 ]
Zhang, Jianguo [4 ]
机构
[1] Sun Yat Sen Univ, Sch Informat Sci & Technol, Guangzhou 510275, Guangdong, Peoples R China
[2] SYSU CMU Shunde Int Joint Res Inst, Shunde 510006, Peoples R China
[3] Guangdong Prov Key Lab Computat Sci, Guangzhou 510275, Guangdong, Peoples R China
[4] Univ Dundee, Sch Comp, Dundee DD1 4HN, Scotland
关键词
Collective activity recognition; interaction modeling; action analysis; MODEL;
D O I
10.1109/TIP.2015.2409564
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Collective activity is a collection of atomic activities (individual person's activity) and can hardly be distinguished by an atomic activity in isolation. The interactions among people are important cues for recognizing collective activity. In this paper, we concentrate on modeling the person-person interactions for collective activity recognition. Rather than relying on hand-craft description of the person-person interaction, we propose a novel learning-based approach that is capable of computing the class-specific person-person interaction patterns. In particular, we model each class of collective activity by an interaction matrix, which is designed to measure the connection between any pair of atomic activities in a collective activity instance. We then formulate an interaction response (IR) model by assembling all these measurements and make the IR class specific and distinct from each other. A multitask IR is further proposed to jointly learn different person-person interaction patterns simultaneously in order to learn the relation between different person-person interactions and keep more distinct activity-specific factor for each interaction at the same time. Our model is able to exploit discriminative low-rank representation of person-person interaction. Experimental results on two challenging data sets demonstrate our proposed model is comparable with the state-of-the-art models and show that learning person-person interactions plays a critical role in collective activity recognition.
引用
收藏
页码:1905 / 1918
页数:14
相关论文
共 41 条
[21]   Approximate kernel competitive learning [J].
Wu, Jian-Sheng ;
Zheng, Wei-Shi ;
Lai, Jian-Huang .
NEURAL NETWORKS, 2015, 63 :117-132
[22]   Interactive Phrases: Semantic Descriptions for Human Interaction Recognition [J].
Kong, Yu ;
Jia, Yunde ;
Fu, Yun .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2014, 36 (09) :1775-1788
[23]   Discriminative Latent Models for Recognizing Contextual Group Activities [J].
Lan, Tian ;
Wang, Yang ;
Yang, Weilong ;
Robinovitch, Stephen N. ;
Mori, Greg .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (08) :1549-1562
[24]  
Leskovec J, 2010, J MACH LEARN RES, V11, P985
[25]   Recognizing Interactive Group Activities Using Temporal Interaction Matrices and Their Riemannian Statistics [J].
Li, Ruonan ;
Chellappa, Rama ;
Zhou, Shaohua Kevin .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2013, 101 (02) :305-328
[26]   A Heat-Map-Based Algorithm for Recognizing Group Activities in Videos [J].
Lin, Weiyao ;
Chu, Hang ;
Wu, Jianxin ;
Sheng, Bin ;
Chen, Zhenzhong .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2013, 23 (11) :1980-1992
[27]   Realistic action recognition via sparsely-constructed Gaussian processes [J].
Liu, Li ;
Shao, Ling ;
Zheng, Feng ;
Li, Xuelong .
PATTERN RECOGNITION, 2014, 47 (12) :3819-3827
[28]  
Rodriguez M, 2011, IEEE I CONF COMP VIS, P1235, DOI 10.1109/ICCV.2011.6126374
[29]   Spatio-Temporal Laplacian Pyramid Coding for Action Recognition [J].
Shao, Ling ;
Zhen, Xiantong ;
Tao, Dacheng ;
Li, Xuelong .
IEEE TRANSACTIONS ON CYBERNETICS, 2014, 44 (06) :817-827
[30]   Efficient Search and Localization of Human Actions in Video Databases [J].
Shao, Ling ;
Jones, Simon ;
Li, Xuelong .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2014, 24 (03) :504-512