Accurate and Robust Video Saliency Detection via Self-Paced Diffusion

被引:43
|
作者
Li, Yunxiao [1 ]
Li, Shuai [1 ]
Chen, Chenglizhao [1 ,2 ]
Hao, Aimin [1 ]
Qin, Hong [3 ]
机构
[1] Beihang Univ, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[2] Qingdao Univ, Qingdao 266071, Peoples R China
[3] SUNY Stony Brook, Stony Brook, NY 11794 USA
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Saliency detection; Proposals; Video sequences; Spatial coherence; Computational modeling; Optical imaging; Optical sensors; Video saliency detection; long-term saliency revealing; key frame strategy; self-paced saliency diffusion; OBJECT DETECTION; SEGMENTATION; OPTIMIZATION; FUSION;
D O I
10.1109/TMM.2019.2940851
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Conventional video saliency detection methods frequently follow the common bottom-up thread to estimate video saliency within the short-term fashion. As a result, such methods can not avoid the obstinate accumulation of errors when the collected low-level clues are constantly ill-detected. Also, being noticed that a portion of video frames, which are not nearby the current video frame over the time axis, may potentially benefit the saliency detection in the current video frame. Thus, we propose to solve the aforementioned problem using our newly-designed key frame strategy (KFS), whose core rationale is to utilize both the spatial-temporal coherency of the salient foregrounds and the objectness prior (i.e., how likely it is for an object proposal to contain an object of any class) to reveal the valuable long-term information. We could utilize all this newly-revealed long-term information to guide our subsequent "self-paced" saliency diffusion, which enables each key frame itself to determine its diffusion range and diffusion strength to correct those ill-detected video frames. At the algorithmic level, we first divide a video sequence into short-term frame batches, and the object proposals are obtained in a frame-wise manner. Then, for each object proposal, we utilize a pre-trained deep saliency model to obtain high-dimensional features in order to represent the spatial contrast. Since the contrast computation within multiple neighbored video frames (i.e., the non-local manner) is relatively insensitive to the appearance variation, those object proposals with high-quality low-level saliency estimation frequently exhibit strong similarity over the temporal scale. Next, the long-term common consistency (e.g., appearance models/movement patterns) of the salient foregrounds could be explicitly revealed via similarity analysis accordingly. We further boost the detection accuracy via long-term information guided saliency diffusion in a self-paced manner. We have conducted extensive experiments to compare our method with 16 state-of-the-art methods over 4 largest public available benchmarks, and all results demonstrate the superiority of our method in terms of both accuracy and robustness.
引用
收藏
页码:1153 / 1167
页数:15
相关论文
共 50 条
  • [21] Robust pixelwise saliency detection via progressive graph rankings
    Wang, Lihua
    Jiang, Bo
    Tu, Zhengzheng
    Hussain, Amir
    Tang, Jin
    NEUROCOMPUTING, 2019, 329 : 433 - 446
  • [22] ROBUST AND ACCURATE OBJECT DETECTION VIA SELF-KNOWLEDGE DISTILLATION
    Xu, Weipeng
    Chu, Pengzhi
    Xie, Renhao
    Xiao, Xiongziyan
    Huang, Hongcheng
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 91 - 95
  • [23] Video saliency detection via combining temporal difference and pixel gradient
    Lu, Xiangwei
    Jian, Muwei
    Wang, Rui
    Liu, Xiangyu
    Lin, Peiguang
    Yu, Hui
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (13) : 37589 - 37602
  • [24] A self-paced learning algorithm for change detection in synthetic aperture radar images
    Shang, Ronghua
    Yuan, Yijing
    Jiao, Licheng
    Meng, Yang
    Ghalamzan, Amir Masoud
    SIGNAL PROCESSING, 2018, 142 : 375 - 387
  • [25] Automatic Foreground Seeds Discovery for Robust Video Saliency Detection
    Zhang, Lin
    Lu, Yao
    Zhou, Tianfei
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II, 2018, 10736 : 89 - 97
  • [26] Foregroundness-Aware Task Disentanglement and Self-Paced Curriculum Learning for Domain Adaptive Object Detection
    Liu, Yabo
    Wang, Jinghua
    Xiao, Linhui
    Liu, Chengliang
    Wu, Zhihao
    Xu, Yong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 369 - 380
  • [27] Self-Paced Feature Attention Fusion Network for Concealed Object Detection in Millimeter-Wave Image
    Wang, Xinlin
    Gou, Shuiping
    Li, Jichao
    Zhao, Yinghai
    Liu, Zhen
    Jiao, Changzhe
    Mao, Shasha
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) : 224 - 239
  • [28] Study of visual saliency detection via nonlocal anisotropic diffusion equation
    Zhang, Xiujun
    Xu, Chen
    Li, Min
    Teng, Robert K. F.
    PATTERN RECOGNITION, 2015, 48 (04) : 1315 - 1327
  • [29] Saliency Detection via A Graph Based Diffusion Model
    He, Zhouqin
    Jiang, Bo
    Xiao, Yun
    Ding, Chris
    Luo, Bin
    GRAPH-BASED REPRESENTATIONS IN PATTERN RECOGNITION (GBRPR 2017), 2017, 10310 : 3 - 12
  • [30] Video Saliency Detection via Graph Clustering With Motion Energy and Spatiotemporal Objectness
    Xu, Mingzhu
    Liu, Bing
    Fu, Ping
    Li, Junbao
    Hu, Yu Hen
    IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (11) : 2790 - 2805