Abstract
Reference-based image super-resolution aims to improve the clarity of input low-resolution (LR) images by leveraging additional high-resolution reference (Ref) images. Although existing methods effectively address the problems associated with super-resolution reconstruction based on reference images, there are still significant challenges in effectively bridging the gap between LR and Ref and enhancing the finesse of texture details. This paper presents MARP-SR, which utilizes attention and residual pooling techniques to bridge their resolution gaps while enhancing texture details. The key innovations include the pooled residual feature enhancement (PRFE) module and the attention extraction feature (AEF) module. PRFE employs residual learning through convolution and pooling to boost texture and salient high-frequency features. AEF utilizes attention mechanisms to extract multi-scale features from LR and Ref images, aligning their resolutions. Experiments on benchmark datasets (CUFED5, Urban100, Manga109, and WR-SR) demonstrate that our proposed MARP-SR achieves state-of-the-art performance, generating more realistic and natural-looking super-resolution results than previous methods.
Similar content being viewed by others
Data availability
The CUFED5 dataset was developed by Zhang et al. [54] for training and testing reference-based image super-resolution and is publicly available for download. The WR-SR dataset was developed by Jiang et al. [22] as a benchmark for reference-based image super-resolution and can be downloaded publicly. The Urban100 dataset was developed by Huang et al. [20] for super-resolution benchmarking and is available for public download. The Manga109 dataset was developed by Matsui et al. [36] for super-resolution benchmarking and is available for open download.
References
Ahmad W et al (2022) A new generative adversarial network for medical images super resolution. Sci Rep 12(1):9533
Yang C et al (2022) Review of image super-resolution reconstruction algorithms based on deep learning. J Front Comput Sci Technol 16(9):1990
Cao J et al (2022) Reference-based image super-resolution with deformable attention transformer. European Conference on Computer Vision. Springer, Berlin, pp 325–342
Chiche BN et al (2022) Stable long-term recurrent video superresolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 837–846
Conde MV et al (2022) Swin2SR: Swinv2 transformer for compressed image super-resolution and restoration. European Conference on Computer Vision. Springer, Berlin, pp 669–687
Cornebise J, Orsolić I, Kalaitzis F (2022) Open high-resolution satellite imagery: the worldstrat dataset-with application to super-resolution. Adv Neural Inf Process Syst 35:25979–25991
Cristani M et al (2004) Distilling information with super-resolution for video surveillance. In: Proceedings of the ACM 2nd International Workshop on Video Surveillance and Sensor Networks. pp 2–11
Dai J et al (2017) Deformable convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp 764–773
Demiray BZ, Sit M, Demir I (2021) D-SRGAN: DEM superresolution with generative adversarial networks. SN Comput Sci 2:1–11
Deng J et al (2009) Imagenet: a large-scale hierarchical image database. In: IEEE conference on computer vision and pattern recognition. IEEE, pp 248–255
Dong C et al (2015) Image super-resolution using deep convolutional networks. IEEE Trans Pattern Anal Mach Intell 38(2):295–307
Dou H et al (2020) PCA-SRGAN: incremental orthogonal projection discrimination for face super-resolution. In: Proceedings of the 28th ACM International Conference on Multimedia, pp 1891–1899
Feng C-M et al (2021) Task transformer network for joint MRI reconstruction and super-resolution. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part VI 24. Springer, pp 307– 317
Georgescu M-I et al (2023) Multimodal multi-head convolutional attention with various kernel sizes for medical image super-resolution. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp 2195–2205
Greenspan H (2009) Super-resolution in medical imaging. Comput J 52(1):43–63
Gulrajani I et al (2017) Improved training of wasserstein gans. Adv Neural Inf Process Syst 30
Haris M, Shakhnarovich G, Ukita N (2018) Deep backprojection networks for super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1664–1673
He K et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 770–778
Huang H et al (2021) Lightweight image super-resolution with hierarchical and differentiable neural architecture search. arXiv preprint arXiv:2105.03939
Huang J-B, Singh A, Ahuja N (2015) Single image superresolution from transformed self-exemplars. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 5197–5206
Ignatov A et al (2017) Dslr-quality photos on mobile devices with deep convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp 3277–3285
Jiang Y et al (2021) Robust reference-based super-resolution via c2-matching. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 2103–2112
Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for realtime style transfer and super-resolution. In: Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part II 14. Springer, pp 694–711
Kim J, Lee JK, Lee KM (2016) Accurate image superresolution using very deep convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1646–1654
Kim J, Lee JK, Lee KM (2016) Deeply-recursive convolutional network for image super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1637–1645
Lai W-S et al (2017) Deep laplacian pyramid networks for fast and accurate super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 624–632
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444
Ledig C et al (2017) Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4681–4690
Lian S, Zhou H, Sun Y (2019) FG-SRGAN: a feature-guided superresolution generative adversarial network for unpaired image super-resolution. In: Advances in Neural Networks–ISNN 2019: 16th International Symposium on Neural Networks, ISNN 2019, Moscow, Russia, July 10–12, 2019, Proceedings, Part I 16. Springer, pp 151–161
Liang J et al (2021) Swinir: image restoration using swin transformer. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 1833–1844
Lim B et al (2017) Enhanced deep residual networks for single image superresolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp 136–144
Liu C et al (2022) Learning trajectory-aware transformer for video superresolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 5687–5696
Liu J, Tang J, Wu G (2020) Residual feature distillation network for lightweight image super-resolution. In: Computer Vision–ECCV 2020 Workshops: Glasgow, UK, August 23–28, 2020, Proceedings, Part III 16. Springer, pp 41–55
Lu L et al (2021) Masa-sr: matching acceleration and spatial adaptation for reference-based image super-resolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 6368–6377
Lu T et al (2019) Satellite image super-resolution via multi-scale residual deep neural network. Remote Sens 11(13):1588
Matsui Y et al (2017) Sketch-based manga retrieval using manga109 dataset. Multimed Tools Appl 76:21811–21838
Rasti P et al (2016) Convolutional neural network super resolution for face recognition in surveillance monitoring. In: Articulated Motion and Deformable Objects: 9th International Conference, AMDO 2016, Palma de Mallorca, Spain, July 13-15, 2016, Proceedings 9. Springer, pp 175–184
Sajjadi MSM, Scholkopf B, Hirsch M (2017) Enhancenet: single image super-resolution through automated texture synthesis. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4491–4500
Shermeyer J, Van Etten A (2019) The effects of super-resolution on object detection performance in satellite imagery. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops
Shim G, Park J, Kweon IS (2020) Robust reference-based superresolution with similarity-aware deformable convolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 8425–8434
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
Sun J, Zongben X, Shum H-Y (2008) Image super-resolution using gradient profile prior. In: IEEE Conference on Computer Vision and Pattern Recognition. IEEE, pp 1–8
Vaswani A et al (2017) Attention is all you need. Adv Neural Inf Process Syst 30
Wang L et al (2019) Learning for video super-resolution through HR optical flow estimation. In: Computer Vision–ACCV 2018: 14th Asian Conference on Computer Vision, Perth, Australia, December 2–6, 2018, Revised Selected Papers, Part I 14. Springer, pp 514–529
Wang X et al (2018) Esrgan: enhanced super-resolution generative adversarial networks. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops
Xia B et al (2022) Coarse-to-fine embedded patchmatch and multi-scale dynamic aggregation for reference-based super-resolution. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol 36, issue 3, pp 2768–2776
Xiong Y et al (2020) Improved SRGAN for remote sensing image superresolution across locations and sensors. Remote Sens 12(8):1263
Yan X et al (2020) Towards content-independent multi-reference super-resolution: Adaptive pattern matching and feature aggregation. In: Computer Vision– ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXV 16. Springer, pp 52–68
Yang F et al (2020) Learning texture transformer network for image superresolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 5791–5800
Yang J, Lin Z, Cohen S (2013) Fast image super-resolution based on in-place example regression. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1059–1066
Image JY et al (2008) Super-resolution as sparse representation of raw image patches. In: IEEE Conference on Computer Vision and Pattern Recognition. IEEE, pp 1–8
Zhang L et al (2010) A super-resolution reconstruction algorithm for surveillance images. Signal Process 90(3):848–859
Zhang W et al (2019) Ranksrgan: Generative adversarial networks with ranker for image super-resolution. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 3096–3105
Zhang Z et al (2019) Image super-resolution by neural texture transfer. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 7982–7991
Zhao M et al (2018) Cross-scale reference-based light field super-resolution. IEEE Trans Comput Imaging 4(3):406–418
Zheng H et al (2018) Crossnet: an end-to-end reference-based super resolution network using cross-scale warping. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 88–104
Zheng H et al (2017) Learning cross-scale correspondence and patch-based synthesis for reference-based super-resolution. In: BMVC, vol 1, p 2
Acknowledgements
This work is supported by the Sichuan Science and Tech nology program (Grant Nos. 2024YFG0001, 2023ZHCG0018, 2023NSFSC0470, 2020JDTD0020, 2022YFG0026, 2021YFG0018, 2023NSFSC0469) and partially supported by Opening Foundation of Agile and Intelligent Computing Key Laboratory of Sichuan Province and CUIT Science and Technology Innovation Capacity Enhancement Program Project under Grant KYTD202330.
Author information
Authors and Affiliations
Contributions
J.D. (Junkun Du) designed the research methods, conducted experiments, and performed data analysis, providing the primary conceptual framework and experimental support for the entire study. M.W. (Mingqing Wang) assisted in experimental design, collected and processed experimental data, and made contributions to the creation and organization of figures and tables. Xi.W. (Xi Wu) provided critical research background and related work and revised and supplemented the discussion and conclusions sections of the paper. Z.Y. (Zhipeng Yang) assisted in experimental design and provided research suggestions, contributing to the literature review section. X.L. (Xiaojie Li), as one of the corresponding authors, was responsible for writing the main sections of the paper, including the introduction, methods, and results and revised and proofread the entire manuscript. X.W. (Xin Wang), as one of the corresponding authors, provided statistical analysis and interpretation of experimental data and made contributions to the discussion section. All authors reviewed and approved the final version of the manuscript.
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare that they have no competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Ethical approval
This declaration is not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Du, J., Wang, M., Wang, X. et al. Reference-based image super-resolution with attention extraction and pooling of residuals. J Supercomput 81, 240 (2025). https://doi.org/10.1007/s11227-024-06587-8
Accepted:
Published:
DOI: https://doi.org/10.1007/s11227-024-06587-8