中国邮电高校学报(英文) ›› 2016, Vol. 23 ›› Issue (5): 68-81.doi: 10.1016/S1005-8885(16)60060-2

• Others • 上一篇    下一篇

Video super-resolution reconstruction based on deep convolutional neural network and spatio-temporal similarity

李玲慧1,杜军平2,梁美玉1,任楠1,范聃1   

  1. 1. 北京邮电大学计算机学院,智能通信软件与多媒体北京市重点实验室
    2. 北京邮电大学
  • 收稿日期:2016-03-21 修回日期:2016-09-29 出版日期:2016-10-30 发布日期:2016-10-26
  • 通讯作者: 杜军平 E-mail:junpingdu@126.com
  • 基金资助:
    the National Natural Science Foundation of China

Video super-resolution reconstruction based on deep convolutional neural network and spatio-temporal similarity

Li Linghui1, Du Junping 2, Liang Meiyu1, Ren Nan1, Fan Dan1   

  • Received:2016-03-21 Revised:2016-09-29 Online:2016-10-30 Published:2016-10-26
  • Supported by:
    the National Natural Science Foundation of China

摘要: Existing learning-based super-resolution (SR) reconstruction algorithms are mainly designed for single image, which ignore the spatio-temporal relationship between video frames. Aiming at applying the advantages of learning-based algorithms to video SR field, a novel video SR reconstruction algorithm based on deep convolutional neural network (CNN) and spatio-temporal similarity (STCNN-SR) was proposed in this paper. It is a deep learning method for video SR reconstruction, which considers not only the mapping relationship among associated low-resolution (LR) and high-resolution (HR) image blocks, but also the spatio-temporal non-local complementary and redundant information between adjacent low-resolution video frames. The reconstruction speed can be improved obviously with the pre-trained end-to-end reconstructed coefficients. Moreover, the performance of video SR will be further improved by the optimization process with spatio-temporal similarity. Experimental results demonstrated that the proposed algorithm achieves a competitive SR quality on both subjective and objective evaluations, when compared to other state-of-the-art algorithms.

关键词: video SR reconstruction, deep convolutional neural network, spatio-temporal similarity, Zernike moment feature

Abstract: Existing learning-based super-resolution (SR) reconstruction algorithms are mainly designed for single image, which ignore the spatio-temporal relationship between video frames. Aiming at applying the advantages of learning-based algorithms to video SR field, a novel video SR reconstruction algorithm based on deep convolutional neural network (CNN) and spatio-temporal similarity (STCNN-SR) was proposed in this paper. It is a deep learning method for video SR reconstruction, which considers not only the mapping relationship among associated low-resolution (LR) and high-resolution (HR) image blocks, but also the spatio-temporal non-local complementary and redundant information between adjacent low-resolution video frames. The reconstruction speed can be improved obviously with the pre-trained end-to-end reconstructed coefficients. Moreover, the performance of video SR will be further improved by the optimization process with spatio-temporal similarity. Experimental results demonstrated that the proposed algorithm achieves a competitive SR quality on both subjective and objective evaluations, when compared to other state-of-the-art algorithms.

Key words: video SR reconstruction, deep convolutional neural network, spatio-temporal similarity, Zernike moment feature

中图分类号: