Exploiting recollection effects for memory-based video object segmentation

被引:0
|
作者
Cho E. [1 ]
Kim M. [1 ]
Kim H.-I. [2 ]
Moon J. [2 ]
Kim S.T. [1 ]
机构
[1] Department of Computer Science and Engineering, Kyung Hee University, Gyeonggi-do, Yongin-si
[2] Electronics and Telecommunications Research Institute (ETRI), Daejeon
基金
新加坡国家研究基金会;
关键词
Deep learning; Memory networks; Video object segmentation;
D O I
10.1016/j.imavis.2023.104866
中图分类号
学科分类号
摘要
Recent advances in deep learning have led to numerous studies on video object segmentation (VOS). Memory-based models, in particular, have demonstrated superior performance by leveraging the ability to store and recall information from previous frames. While extensive research efforts have been devoted to developing memory networks for effective VOS, only a few studies have investigated the quality of memory in terms of determining which information should be stored. In fact, in most recent memory-based VOS studies, the frame information is regularly stored in the memory without specific consideration. In other words, there is a lack of explicit criteria or guidelines for determining the essential information that should be retained in memory. In this study, we introduce a new method for evaluating the effect of storing the features, which can be used for various memory-based networks to improve performance in a plug-and-play manner. For this purpose, we introduce the concept of recollection effects, which refers to the stability of predictions based on the presence or absence of specific features in memory. By explicitly measuring the recollection effects, we establish a criterion for evaluating the relevance of information and determining whether features from a particular frame should be stored. This approach effectively encourages memory-based networks to construct memory that contains valuable cues. To validate the effectiveness of our method, we conduct comparative experiments. Experimental results demonstrate the effectiveness of our method to enhance the selection and retention of useful cues within the memory, leading to improving segmentation results. © 2023 Elsevier B.V.
引用
收藏
相关论文
共 50 条
  • [41] Video object segmentation based on edge tracking
    Zaletelj, J
    Tasic, JF
    2001 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOL II, PROCEEDINGS, 2001, : 813 - 816
  • [42] Mosaic-based video object segmentation
    Chen, LH
    Lai, YC
    Su, CW
    Liao, HY
    PROCEEDINGS OF THE EIGHTH IASTED INTERNATIONAL CONFERENCE ON INTERNET AND MULTIMEDIA SYSTEMS AND APPLICATIONS, 2004, : 335 - 340
  • [43] Video object segmentation for content-based video coding
    Zhou, JY
    Ong, EP
    Ko, CC
    PROCEEDINGS OF THE FIFTH JOINT CONFERENCE ON INFORMATION SCIENCES, VOLS 1 AND 2, 2000, : A334 - A337
  • [44] Mesh based segmentation and update for object based video
    Gökçetekin, MH
    Harmanci, MD
    Celasun, I
    Tekalp, AM
    2000 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOL I, PROCEEDINGS, 2000, : 343 - 346
  • [46] Design of A Memory-Based VLC Decoder for Portable Video Applications
    Lee, Wei-Chin
    Li, Yao
    Lee, Chen-Yi
    2008 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2008), VOLS 1-4, 2008, : 1340 - 1343
  • [47] Object Oriented Memory Recollection and Retrieval based on Spatial Log in the iSpace
    Saito, Kazufumi
    Yoshimura, Akira
    Lee, Joo-Ho
    2012 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), 2012, : 271 - 276
  • [48] Distinct neural mechanisms underlie subjective and objective recollection and guide memory-based decision making
    Fandakova, Yana
    Johnson, Elliott G.
    Ghetti, Simona
    ELIFE, 2021, 10
  • [49] Breaking the "Object" in Video Object Segmentation
    Tokmakov, Pavel
    Li, Jie
    Gaidon, Adrien
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 22836 - 22845
  • [50] Accurate object detection using memory-based models in surveillance scenes
    Li, Xudong
    Ye, Mao
    Liu, Yiguang
    Zhang, Feng
    Liu, Dan
    Tang, Song
    PATTERN RECOGNITION, 2017, 67 : 73 - 84