Dual-referenced assistive network for action quality assessment

被引:0
作者
Huang, Keyi [1 ]
Tian, Yi [1 ]
Yu, Chen [1 ]
Huang, Yaping [1 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp Sci & Technol, Beijing Key Lab Traff Data Anal & Min, Beijing 100044, Peoples R China
基金
中国国家自然科学基金;
关键词
Action quality assessment; Human action understanding; VIDEO;
D O I
10.1016/j.neucom.2024.128786
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Action quality assessment (AQA) aims to evaluate the performing quality of a specific action. It is a challenging task as it requires to identify the subtle differences between the videos containing the same action. Most of existing AQA methods directly adopt a pretrained network designed for other tasks to extract video features, which are too coarse to describe fine-grained details of action quality. In this paper, we propose a novel Dual-Referenced Assistive (DuRA) network to polish original coarse-grained features into fine-grained quality-oriented representations. Specifically, we introduce two levels of referenced assistants to highlight the discriminative quality-related contents by comparing a target video and the referenced objects, instead of obtrusively estimating the quality score from an individual video. Firstly, we design a Rating-guided Attention module, which takes advantage of a series of semantic-level referenced assistants to acquire implicit hierarchical semantic knowledge and progressively emphasize quality-focused features embedded in original inherent information. Subsequently, we further design a couple of Consistency Preserving constraints, which introduce a set of individual-level referenced assistants to further eliminate score-unrelated information through more detailed comparisons of differences between actions. The experiments show that our proposed method achieves promising performance on the AQA-7 and MTL-AQA datasets.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Temporal attention learning for action quality assessment in sports video
    Qing Lei
    Hongbo Zhang
    Jixiang Du
    Signal, Image and Video Processing, 2021, 15 : 1575 - 1583
  • [22] Temporal attention learning for action quality assessment in sports video
    Lei, Qing
    Zhang, Hongbo
    Du, Jixiang
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (07) : 1575 - 1583
  • [23] MULTI-STAGE CONTRASTIVE REGRESSION FOR ACTION QUALITY ASSESSMENT
    An, Qi
    Qi, Mengshi
    Ma, Huadong
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 4110 - 4114
  • [24] Learning and fusing multiple hidden substages for action quality assessment
    Dong, Li-Jia
    Zhang, Hong-Bo
    Shi, Qinghongya
    Lei, Qing
    Du, Ji-Xiang
    Gao, Shangce
    KNOWLEDGE-BASED SYSTEMS, 2021, 229
  • [25] Auto-encoding score distribution regression for action quality assessment
    Zhang, Boyu
    Chen, Jiayuan
    Xu, Yinfei
    Zhang, Hui
    Yang, Xu
    Geng, Xin
    NEURAL COMPUTING & APPLICATIONS, 2023, 36 (2) : 929 - 942
  • [26] Procedure-Aware Action Quality Assessment: Datasets and Performance Evaluation
    Xu, Jinglin
    Rao, Yongming
    Zhou, Jie
    Lu, Jiwen
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 6069 - 6090
  • [27] Vision-based human action quality assessment: A systematic review
    Liu, Jiang
    Wang, Huasheng
    Stawarz, Katarzyna
    Li, Shiyin
    Fu, Yao
    Li, Hantao
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 263
  • [28] Auto-encoding score distribution regression for action quality assessment
    Boyu Zhang
    Jiayuan Chen
    Yinfei Xu
    Hui Zhang
    Xu Yang
    Xin Geng
    Neural Computing and Applications, 2024, 36 : 929 - 942
  • [29] Uncertainty-aware Score Distribution Learning for Action Quality Assessment
    Tang, Yansong
    Ni, Zanlin
    Zhou, Jiahuan
    Zhang, Danyang
    Lu, Jiwen
    Wu, Ying
    Zhou, Jie
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 9836 - 9845
  • [30] IRIS: Interpretable Rubric-Informed Segmentation for Action Quality Assessment
    Matsuyama, Hitoshi
    Kawaguchi, Nobuo
    Lim, Brian Y.
    PROCEEDINGS OF 2023 28TH ANNUAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2023, 2023, : 368 - 378