Deblurring Videos Using Spatial-Temporal Contextual Transformer With Feature Propagation

被引:1
|
作者
Zhang, Liyan [1 ]
Xu, Boming [2 ]
Yang, Zhongbao [2 ]
Pan, Jinshan [2 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 210016, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
基金
中国国家自然科学基金;
关键词
Videos; Transformers; Feature extraction; Three-dimensional displays; Image restoration; Convolutional neural networks; Optical imaging; Context modeling; Computational modeling; Optical propagation; Video deblurring; self-attention; contextual transformer; non-local temporal information; feature propagation;
D O I
10.1109/TIP.2024.3482176
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a simple and effective approach to explore both local spatial-temporal contexts and non-local temporal information for video deblurring. First, we develop an effective spatial-temporal contextual transformer to explore local spatial-temporal contexts from videos. As the features extracted by the spatial-temporal contextual transformer does not model the non-local temporal information of video well, we then develop a feature propagation method to aggregate useful features from the long-range frames so that both local spatial-temporal contexts and non-local temporal information can be better utilized for video deblurring. Finally, we formulate the spatial-temporal contextual transformer with the feature propagation into a unified deep convolutional neural network (CNN) and train it in an end-to-end manner. We show that using the spatial-temporal contextual transformer with the feature propagation is able to generate useful features and makes the deep CNN model more compact and effective for video deblurring. Extensive experimental results show that the proposed method performs favorably against state-of-the-art ones on the benchmark datasets in terms of accuracy and model parameters.
引用
收藏
页码:6354 / 6366
页数:13
相关论文
共 50 条
  • [1] Spatial-Temporal Transformer for Crime Recognition in Surveillance Videos
    Boekhoudt, Kayleigh
    Talavera, Estefania
    2022 18TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS 2022), 2022,
  • [2] LightViD: Efficient Video Deblurring With Spatial-Temporal Feature Fusion
    Lin, Liqun
    Wei, Guangpeng
    Liu, Kanglin
    Feng, Wanjian
    Zhao, Tiesong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (08) : 7430 - 7439
  • [3] Hierarchy Spatial-Temporal Transformer for Action Recognition in Short Videos
    Cai, Guoyong
    Cai, Yumeng
    FUZZY SYSTEMS AND DATA MINING VI, 2020, 331 : 760 - 774
  • [4] Spatial-Temporal Contextual Feature Fusion Network for Movie Description
    Liao, Yihui
    Fan, Lu
    Ding, Huiming
    Xie, Zhifeng
    ARTIFICIAL INTELLIGENCE, CICAI 2022, PT I, 2022, 13604 : 490 - 501
  • [5] Spatial-temporal Graph Transformer Network for Spatial-temporal Forecasting
    Dao, Minh-Son
    Zetsu, Koji
    Hoang, Duy-Tang
    Proceedings - 2024 IEEE International Conference on Big Data, BigData 2024, 2024, : 1276 - 1281
  • [6] Behavior Recognition in Mouse Videos using Contextual Features Encoded by Spatial-temporal Stacked Fisher Vectors
    Jiang, Zheheng
    Crookes, Danny
    Green, Brian Desmond
    Zhang, Shengping
    Zhou, Huiyu
    ICPRAM: PROCEEDINGS OF THE 6TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS, 2017, : 259 - 269
  • [7] Scene Retrieval in Soccer Videos by Spatial-temporal Attention with Video Vision Transformer
    Gan, Yaozong
    Togo, Ren
    Ogawa, Takahiro
    Haseyama, Mild
    2022 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN, IEEE ICCE-TW 2022, 2022, : 453 - 454
  • [8] Spatial-temporal feature guided adaptive rate control for screen content videos
    Lin, Qi
    Chen, Jing
    JOURNAL OF REAL-TIME IMAGE PROCESSING, 2024, 21 (06)
  • [9] Emotion Classification Based on Transformer and CNN for EEG Spatial-Temporal Feature Learning
    Yao, Xiuzhen
    Li, Tianwen
    Ding, Peng
    Wang, Fan
    Zhao, Lei
    Gong, Anmin
    Nan, Wenya
    Fu, Yunfa
    BRAIN SCIENCES, 2024, 14 (03)
  • [10] Fast Spatial-Temporal Transformer Network
    Escher, Rafael Molossi
    de Bem, Rodrigo Andrade
    Jorge Drews Jr, Paulo Lilles
    2021 34TH SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI 2021), 2021, : 65 - 72