[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3383812.3383836acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicigpConference Proceedingsconference-collections
research-article

A Video Frame Resolution and Frame Rate Amplification Method with Optical Flow Method and ESPCN Model

Published: 25 March 2020 Publication History

Abstract

Super-resolution reconstruction for video sequences includes two expansions: the video frame resolution and frame rate, which enhances the resolution of the video frame while increasing the frame rate (frame expansion). Three problems of superresolution to solve include: a) how to use adjacent frames to achieve high quality reconstruction; b) how to generate motion compensation frames to supplement the video; and c) how to improve the efficiency of reconstruction calculation and control the running time. Although many methods have been successfully applied to video super-resolution, these methods still face great challenges in balancing motion compensation accuracy, computational complexity, reconstruction quality and running time. In this paper, an optical flow approach combined with an efficient sub-pixel convolutional neural network (ESPCN) model is proposed for frame resolution and frame rate amplification. By adopting the technical strategy of hyper-splitting before frame insertion (enhancement before frame expansion), the motion compensation frame is generated through the image optical flow to improve the frame rate. The super resolution of video frame is realized by constructing the model combining the motion estimation (ME) between adjacent frames with ESPCN (named as ME+ESPCN). The strategy of first superresolution reconstruction and then inserting frame (enhancement before frame expansion) is adopted to generate motion compensation frame through image optical flow to improve the frame rate. The simulation results show that compared with Sparse Dictionary Learning (SDL) and Super-Resolution Convolutional Neural Network (SRCNN), the proposed method based on ME+ESPCN model accelerates the reconstruction operation significantly, takes about 18ms on average, has higher real-time performance, and the average PSNR value for evaluating the frame restoration quality is about 0.12dB higher than ESPCN (no motion estimation). In addition, compared with the SRCNN method, the PSNR of the amplification technology strategy designed in this paper is improved by about 0.32dB on average.

References

[1]
Takeda, H., Farsiu, S., and Milanfar, P. 2007. Kernel regression for image processing and reconstruction. IEEE Transactions on Image Processing, 16, 2, 349--366.
[2]
Xu, F., Xu, M., Jiang, D., Shi, J., and Shi, A. 2016. Temporal super-resolution based on pixel stream and featured prior model for motion blurred single video. Journal of Electronic Imaging, 25, 5, 053011.
[3]
Dai, Q., Yoo, S., Kappeler, A., and Katsaggelos, A. K. 2015. Dictionary-based multiple frame video super-resolution. IEEE International Conference on Image Processing, 83--87.
[4]
Guo, J., and Chao, H. 2017. Building an end-to-end spatial-temporal convolutional network for video super-resolution. Proceedings of the AAAI Conference on Artificial Intelligence.
[5]
Caballero, J., Ledig, C., Aitken, A., Acosta, A., Totz, J., Wang, Z, and Shi, W. 2017. Real-time video super-resolution with spatio-temporal networks and motion compensation. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 4778--4787.
[6]
Yang, W., Feng, J., Xie, G., Liu, J., Guo, Z., and Yan, S. 2018. Video super-resolution based spatial-temporal recurrent residual networks. Computer Vision and Image Understanding, 168, 79--92.
[7]
Zhong, W. 2018. Research on application of convolutional neural network in video super resolution. Master Thesis. University of Electronic Science and Technology of China.
[8]
Li, D. 2019. Research on deep learning based video super-resolution algorithm. Doctoral Thesis. University of Science and Technology of China.
[9]
Shi, W., Caballero, J., Huszar, F., Totz, J., Aitken, A. P., Bishop, R., Rueckert, D., Wang, Z., and Technology. M. P. 2016. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. the 29th IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, 1874--1883.
[10]
Du, X. 2019. Research on super-resolution reconstruction based on deep learning. Master Thesis. Hohai University.
[11]
Liu, C., and Sun, D. Q. 2011. A Bayesian approach to adaptive video super resolution. IEEE Conference on Computer Vision and Pattern Recognition, 42, 7, 209--216.
[12]
Bradski, G., Kaehler, A., Yu, S., Liu, R. translated. 2009. Learning OpenCV (Chinese version). Beijing: Tsinghua University Press (ISBN9787302209935).
[13]
Yang, J. C., Wright, J., Huang, T. S., and Ma, Y. 2010. Image superresolution via sparse representation. IEEE Transactions on Image Processing, 19, 11, 2861--2873.
[14]
Dong, C., Loy, C. C., He, K. M., and Tang, X. O. 2016. Image super-resolution using deep convolutional networks. IEEE Trans-actions on Pattern Analysis and Machine Intelligence, 38, 2, 295--307.
[15]
Wang., H., Gao, G., Xu, L., and Wen, C. 2018. A multi-region level set model based on texture feature for image segmentation. Acta Electronica Sinica, 46, 11, 2588--2596.
[16]
Lv, L., Fan, T., Sun, Z., Wang, J., and Xu, L. 2016. Object tracking with double-dictionary appearance model. Optical Engineering, 55, 8, 083106. Lv, L. Chen, Z., Zhang, Z., Fan, T., and Xu, L. 2018. Object tracking based on support vector dictionary learning. Signal Image and Video Processing, 12, 1189--1196

Cited By

View all
  • (2021)A Generalizable Sample Resolution Augmentation Method for Mechanical Fault Diagnosis Based on ESPCNJournal of Sensors10.1155/2021/74960072021:1Online publication date: 24-Nov-2021
  • (2021)Video Frame Rate Doubling Using Generative Adversarial NetworksComputer Communication, Networking and IoT10.1007/978-981-16-0980-0_43(463-474)Online publication date: 19-Jun-2021

Index Terms

  1. A Video Frame Resolution and Frame Rate Amplification Method with Optical Flow Method and ESPCN Model

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICIGP '20: Proceedings of the 2020 3rd International Conference on Image and Graphics Processing
    February 2020
    172 pages
    ISBN:9781450377201
    DOI:10.1145/3383812
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    In-Cooperation

    • Nanyang Technological University
    • UNIBO: University of Bologna

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 March 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Video super-resolution
    2. convolutional neural network
    3. deep learning
    4. optical flow

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICIGP 2020

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)9
    • Downloads (Last 6 weeks)3
    Reflects downloads up to 18 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2021)A Generalizable Sample Resolution Augmentation Method for Mechanical Fault Diagnosis Based on ESPCNJournal of Sensors10.1155/2021/74960072021:1Online publication date: 24-Nov-2021
    • (2021)Video Frame Rate Doubling Using Generative Adversarial NetworksComputer Communication, Networking and IoT10.1007/978-981-16-0980-0_43(463-474)Online publication date: 19-Jun-2021

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media