Learning Temporal Dynamics for Video Super-Resolution: A Deep Learning Approach

被引:68
|
作者
Liu, Ding [1 ,2 ]
Wang, Zhaowen [3 ]
Fan, Yuchen [1 ,2 ]
Liu, Xianming [1 ,2 ,4 ]
Wang, Zhangyang [5 ]
Chang, Shiyu [6 ]
Wang, Xinchao
Huang, Thomas S. [1 ,2 ]
机构
[1] Univ Illinois, Dept Elect & Comp Engn, Urbana, IL 61801 USA
[2] Univ Illinois, Beckman Inst Adv Sci & Technol, Urbana, IL 61801 USA
[3] Adobe Syst Inc, San Jose, CA 95110 USA
[4] Facebook Inc, San Francisco, CA 94025 USA
[5] Texas A&M Univ, Dept Comp Sci & Engn, College Stn, TX 77843 USA
[6] IBM Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
关键词
Super-resolution; deep learning; deep neural networks; QUALITY ASSESSMENT;
D O I
10.1109/TIP.2018.2820807
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video super-resolution (SR) aims at estimating a high-resolution video sequence from a low-resolution (LR) one. Given that the deep learning has been successfully applied to the task of single image SR, which demonstrates the strong capability of neural networks for modeling spatial relation within one single image, the key challenge to conduct video SR is how to efficiently and effectively exploit the temporal dependence among consecutive LR frames other than the spatial relation. However, this remains challenging because the complex motion is difficult to model and can bring detrimental effects if not handled properly. We tackle the problem of learning temporal dynamics from two aspects. First, we propose a temporal adaptive neural network that can adaptively determine the optimal scale of temporal dependence. Inspired by the inception module in GoogLeNet [1], filters of various temporal scales are applied to the input LR sequence before their responses are adaptively aggregated, in order to fully exploit the temporal relation among the consecutive LR frames. Second, we decrease the complexity of motion among neighboring frames using a spatial alignment network that can be end-to-end trained with the temporal adaptive network and has the merit of increasing the robustness to complex motion and the efficiency compared with the competing image alignment methods. We provide a comprehensive evaluation of the temporal adaptation and the spatial alignment modules. We show that the temporal adaptive design considerably improves the SR quality over its plain counterparts, and the spatial alignment network is able to attain comparable SR performance with the sophisticated optical flow-based approach, but requires a much less running time. Overall, our proposed model with learned temporal dynamics is shown to achieve the state-of-the-art SR results in terms of not only spatial consistency but also the temporal coherence on public video data sets. More information can be found in http://www.ifp.illinois.edu/similar to dingliu2/videoSR/.
引用
收藏
页码:3432 / 3445
页数:14
相关论文
共 50 条
  • [1] A Survey of Deep Learning Video Super-Resolution
    Baniya, Arbind Agrahari
    Lee, Tsz-Kwan
    Eklund, Peter W.
    Aryal, Sunil
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (04): : 2655 - 2676
  • [2] Omnidirectional Video Super-Resolution Using Deep Learning
    Baniya, Arbind Agrahari
    Lee, Tsz-Kwan
    Eklund, Peter W.
    Aryal, Sunil
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 540 - 554
  • [3] Deep Learning for Image/Video Restoration and Super-resolution
    Tekalp, A. Murat
    FOUNDATIONS AND TRENDS IN COMPUTER GRAPHICS AND VISION, 2022, 13 (01): : 1 - 110
  • [4] Video super-resolution based on deep learning: a comprehensive survey
    Liu, Hongying
    Ruan, Zhubo
    Zhao, Peng
    Dong, Chao
    Shang, Fanhua
    Liu, Yuanyuan
    Yang, Linlin
    Timofte, Radu
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (08) : 5981 - 6035
  • [5] Video super-resolution based on deep learning: a comprehensive survey
    Hongying Liu
    Zhubo Ruan
    Peng Zhao
    Chao Dong
    Fanhua Shang
    Yuanyuan Liu
    Linlin Yang
    Radu Timofte
    Artificial Intelligence Review, 2022, 55 : 5981 - 6035
  • [6] Learning a Deep Dual Attention Network for Video Super-Resolution
    Li, Feng
    Bai, Huihui
    Zhao, Yao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 4474 - 4488
  • [7] Learning Via Decision Trees Approach for Video Super-Resolution
    Zhang, Yu-Zhu
    Siu, Wan-Chi
    Liu, Zhi-Song
    Law, Ngai-Fong
    PROCEEDINGS 2017 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI), 2017, : 558 - 562
  • [8] Deep Learning Based Video Super-Resolution and its Application in Video Conferences
    Lin, Yinyan
    Zou, Chaoyang
    Feng, Ying
    Liang, Mingwei
    2021 6TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2021), 2021, : 288 - 293
  • [9] Temporal Consistency Learning of Inter-Frames for Video Super-Resolution
    Liu, Meiqin
    Jin, Shuo
    Yao, Chao
    Lin, Chunyu
    Zhao, Yao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (04) : 1507 - 1520
  • [10] Learning a spatial-temporal symmetry network for video super-resolution
    Xiaohang Wang
    Mingliang Liu
    Pengying Wei
    Applied Intelligence, 2023, 53 : 3530 - 3544