Weakly-supervised Temporal Action Localization with Adaptive Clustering and Refining Network

被引:0
|
作者
Ren, Hao [1 ]
Ran, Wu [1 ]
Liu, Xingson [1 ]
Ren, Haoran [1 ]
Lu, Hong [1 ]
Zhang, Rui [1 ]
Jin, Cheng [1 ,2 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Intelligent Informat Proc, Shanghai, Peoples R China
[2] Peng Cheng Lab, Shenzhen, Peoples R China
关键词
Temporal Action Localization; Weakly-supervised Learning; Adaptive Clustering;
D O I
10.1109/ICME55011.2023.00177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Weakly-supervised temporal action localization task aims to localize temporal boundaries of action instances by using only video-level labels. Existing methods primarily adopt Multi-Instance-Learning (MIL) scheme to handle this task. The effectiveness of MIL scheme depends heavily on the selection of top-k action snippets, which is unstable and requires manual tuning. To address these deficiencies, we propose an Adaptive Clustering and Refining Network (ACRNet). Specifically, we present an action-aware clustering strategy that is adaptable and requires no manual tuning to separate action and background snippets of diverse videos based on intra-class activation distribution. And a cluster refining step is included to eliminate false action snippets by considering inter-class activation distribution, which greatly improves robustness and localization accuracy. Extensive experiments on THUMOS14, ActivityNet 1.2&1.3 benchmarks show that our method achieves state-of-the-art performance.
引用
收藏
页码:1008 / 1013
页数:6
相关论文
共 50 条
  • [31] Context Sensitive Network for weakly-supervised fine-grained temporal action localization
    Dong, Cerui
    Liu, Qinying
    Wang, Zilei
    Zhang, Yixin
    Zhao, Feng
    NEURAL NETWORKS, 2025, 185
  • [32] Entropy guided attention network for weakly-supervised action localization
    Cheng, Yi
    Sun, Ying
    Fan, Hehe
    Zhuo, Tao
    Lim, Joo-Hwee
    Kankanhalli, Mohan
    PATTERN RECOGNITION, 2022, 129
  • [33] Semantic and Temporal Contextual Correlation Learning for Weakly-Supervised Temporal Action Localization
    Fu, Jie
    Gao, Junyu
    Xu, Changsheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12427 - 12443
  • [34] Learning Action Completeness from Points for Weakly-supervised Temporal Action Localization
    Lee, Pilhyeon
    Byun, Hyeran
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 13628 - 13637
  • [35] Enhancing action discrimination via category-specific frame clustering for weakly-supervised temporal action localization
    Xia, Huifen
    Zhan, Yongzhao
    Liu, Honglin
    Ren, Xiaopeng
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2024, 25 (06) : 809 - 823
  • [36] Action-Aware Network with Upper and Lower Limit Loss for Weakly-Supervised Temporal Action Localization
    Bi, Mingwen
    Li, Jiaqi
    Liu, Xinliang
    Zhang, Qingchuan
    Yang, Zhenghong
    NEURAL PROCESSING LETTERS, 2023, 55 (04) : 4307 - 4324
  • [37] Action-Aware Network with Upper and Lower Limit Loss for Weakly-Supervised Temporal Action Localization
    Mingwen Bi
    Jiaqi Li
    Xinliang Liu
    Qingchuan Zhang
    Zhenghong Yang
    Neural Processing Letters, 2023, 55 : 4307 - 4324
  • [38] CoLA: Weakly-Supervised Temporal Action Localization with Snippet Contrastive Learning
    Zhang, Can
    Cao, Meng
    Yang, Dongming
    Chen, Jie
    Zou, Yuexian
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16005 - 16014
  • [39] Diffusion-based framework for weakly-supervised temporal action localization
    Zou, Yuanbing
    Zhao, Qingjie
    Sarker, Prodip Kumar
    Li, Shanshan
    Wang, Lei
    Liu, Wangwang
    Pattern Recognition, 2025, 160
  • [40] Adversarial Seeded Sequence Growing for Weakly-Supervised Temporal Action Localization
    Zhang, Chengwei
    Xu, Yunlu
    Cheng, Zhanzhan
    Niu, Yi
    Pu, Shiliang
    Wu, Fei
    Zou, Futai
    PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 738 - 746