Deeply feature fused video super-resolution network using temporal grouping

被引:0
作者
Zhensen Chen
Wenyuan Yang
Jingmin Yang
机构
[1] Minnan Normal University,School of Computer Science
[2] Fujian Province University,Key Laboratory of Data Science and Intelligence Application
[3] Minnan Normal University,Fujian Key Laboratory of Granular Computing and Application
来源
The Journal of Supercomputing | 2022年 / 78卷
关键词
Video super-resolution; Deep learning; Deformable convolution; Feature fusion; Temporal grouping;
D O I
暂无
中图分类号
学科分类号
摘要
The video super-resolution (VSR) task refers to the use of corresponding low-resolution frames and multiple neighboring frames to generate high-resolution (HR) frames. An important step in VSR is to fuse the features of the reference frame with the features of the supporting frame. Existing VSR methods do not take full advantage of the information provided by distant neighboring frames and usually fuse the information in a one-stage manner. In this paper, we propose a deep fusion video super-resolution network based on temporal grouping. We divide the input sequence into groups according to different frame rates to provide more accurate supplementary information. Our method aggregates temporal-spatial information at different fusion stages. Firstly, we group the input sequence. Then the temporal-spatial information is extracted and fused hierarchically, and these groups are used to recover the information lost in the reference frame. Secondly, integrate information within each group to generate group-wise features, and then perform multi-stage fusion. The information of the reference frame is fully utilized, resulting in a better recovery of video details. Finally, the upsampling module is used to generate HR frames. We conduct a comprehensive comparative experiment on Vid4, SPMC-11 and Vimeo-90K-T datasets. The results show that the proposed method achieves good performance compared with state-of-the-art methods.
引用
收藏
页码:8999 / 9016
页数:17
相关论文
共 51 条
  • [1] Huang Yan(2015)Bidirectional recurrent convolutional networks for multi-frame super-resolution Adv Neural Inf Process Syst 28 235-243
  • [2] Wang Wei(2021)Interpretable detail-fidelity attention network for single image super-resolution IEEE Trans Image Process 30 2325-2339
  • [3] Wang Liang(2016)Video super-resolution with convolutional neural networks IEEE Trans Comput Imag 2 109-122
  • [4] Huang Yuanfei(2013)On Bayesian adaptive video super resolution IEEE Trans Pattern Anal Mach Intell 36 346-360
  • [5] Li Jie(2020)Pulmonary nodule image super-resolution using multi-scale deep residual channel attention network with joint optimization J Supercomput 76 1005-1019
  • [6] Gao Xinbo(2021)Simultaneous fidelity and regularization learning for image restoration IEEE Trans Pattern Anal Mach Intell 43 284-299
  • [7] Yanting Hu(2021)Multi-stage feature fusion network for video super-resolution IEEE Trans Image Process 30 2923-2934
  • [8] Wen Lu(2021)Dnetunet: a semi-supervised CNN of medical image segmentation for super-computing ai service J Supercomput 77 3594-3615
  • [9] Armin Kappeler(2020)Deep video super-resolution using hr optical flow estimation IEEE Trans Image Process 29 4323-4336
  • [10] Seunghwan Yoo(2021)Structure-aware motion deblurring using multi-adversarial optimized cyclegan IEEE Trans Image Process 30 6142-6155