Attention with structure regularization for action recognition

被引:13
作者
Quan, Yuhui [1 ,3 ]
Chen, Yixin [1 ]
Xu, Ruotao [1 ]
Ji, Hui [2 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510006, Guangdong, Peoples R China
[2] Natl Univ Singapore, Dept Math, Singapore 119076, Singapore
[3] Guangdong Prov Key Lab Computat Intelligence & Cy, Guangzhou, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Action recognition; Attention; Block-wise sparsity; Deep recurrent network; VIDEOS;
D O I
10.1016/j.cviu.2019.102794
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recognizing human action in video is an important task with a wide range of applications. Recently, motivated by the findings in human visual perception, there have been numerous attempts on introducing attention mechanisms to action recognition systems. However, it is empirically observed that an implementation of attention mechanism using attention mask of free form often generates ineffective distracted attention regions caused by overfitting, which limits the benefit of attention mechanisms for action recognition. By exploiting block-structured sparsity prior on attention regions, this paper proposed an l(2,1)-norm group sparsity regularization for learning structured attention masks. Built upon such a regularized attention module, an attention-based recurrent network is developed for action recognition. The experimental results on two benchmark datasets showed that, the proposed method can noticeably improve the accuracy of attention masks, which results in performance gain in action recognition.
引用
收藏
页数:11
相关论文
共 61 条
[1]  
[Anonymous], 2014, ADV NEURAL INFORM PR
[2]  
[Anonymous], ARXIV170503146
[3]  
[Anonymous], 2013, IEEE T PATTERN ANAL, DOI DOI 10.1109/TPAMI.2012.59
[4]  
[Anonymous], 2007, SINGLE VIEW HUMAN AC
[5]  
[Anonymous], MULTIMEDIA TOOLS APP
[6]  
[Anonymous], P C COMP VIS PATT RE
[7]  
[Anonymous], P BRIT MACH VIS C BR
[8]  
[Anonymous], 2018, IEEE C COMP VIS PATT
[9]  
[Anonymous], 2015, NEURAL INFORM PROCES
[10]  
[Anonymous], P CVPR