[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN114202459A - Blind image super-resolution method based on depth prior - Google Patents

Blind image super-resolution method based on depth prior Download PDF

Info

Publication number
CN114202459A
CN114202459A CN202111244295.3A CN202111244295A CN114202459A CN 114202459 A CN114202459 A CN 114202459A CN 202111244295 A CN202111244295 A CN 202111244295A CN 114202459 A CN114202459 A CN 114202459A
Authority
CN
China
Prior art keywords
image
resolution
network
blur kernel
resolution image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111244295.3A
Other languages
Chinese (zh)
Other versions
CN114202459B (en
Inventor
李一秾
禹晶
肖创柏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN202111244295.3A priority Critical patent/CN114202459B/en
Publication of CN114202459A publication Critical patent/CN114202459A/en
Application granted granted Critical
Publication of CN114202459B publication Critical patent/CN114202459B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/043Architecture, e.g. interconnection topology based on fuzzy logic, fuzzy membership or fuzzy inference, e.g. adaptive neuro-fuzzy inference systems [ANFIS]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Automation & Control Theory (AREA)
  • Computational Mathematics (AREA)
  • Fuzzy Systems (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Image Processing (AREA)

Abstract

本发明公开了一种基于深度先验的盲图像超分辨率方法,使用深度卷积神经网络DIP‑Net并结合非局部注意力估计高分辨率图像;通过求解关于模糊核最优化问题的精确解估计模糊核;交替迭代更新模糊核与高分辨率图像,利用估计模糊核生成低分辨率图像的降采样图像,结合低分辨率图像利用损失更新网络参数。对模糊核和高分辨率图像联合建模,采用网络模型与数学模型结合的方式同时估计模糊核与高分辨率图像;结合非局部注意力模块显式地利用图像的自相似性,为重建图像提供潜在的附加信息;仅利用单幅低分辨率图像作为监督信号实现了端到端的盲图像超分辨率重建,无需数据集的训练过程。本发明能够准确地估计模糊核,有效重建高分辨率图像的边缘与细节。

Figure 202111244295

The invention discloses a blind image super-resolution method based on depth prior, which uses the deep convolutional neural network DIP-Net and combines non-local attention to estimate high-resolution images; Estimate the blur kernel; iteratively update the blur kernel and the high-resolution image alternately, use the estimated blur kernel to generate a down-sampled image of the low-resolution image, and use the loss to update the network parameters in combination with the low-resolution image. The fuzzy kernel and the high-resolution image are jointly modeled, and the fuzzy kernel and the high-resolution image are estimated at the same time by the combination of the network model and the mathematical model. Provides potential additional information; end-to-end blind image super-resolution reconstruction is achieved using only a single low-resolution image as a supervision signal, without the training process of the dataset. The invention can accurately estimate the blur kernel and effectively reconstruct the edge and details of the high-resolution image.

Figure 202111244295

Description

Blind image super-resolution method based on depth prior
Technical Field
The invention relates to the field of image super-resolution, in particular to a blind image super-resolution method based on depth prior.
Background
In the image acquisition processes of video monitoring, mobile phone camera shooting, medical imaging, remote sensing imaging, video capturing and the like, a high-resolution image cannot be acquired generally due to the limitation of an imaging mechanism, an imaging environment or imaging equipment. The high-resolution image can provide more detailed information, and is beneficial to analysis and understanding tasks such as image interpretation. The method for improving the resolution of the acquired image in terms of hardware quality has the problems of cost, process and the like, and under the premise that hardware conditions cannot be changed, the image super-resolution method breaks through the limitation of the inherent sampling frequency of a sensor through an image processing means, so that the expansion of a low-resolution image on the spatial resolution or cut-off frequency is realized, and the purpose of improving the spatial resolution of the image is achieved. The single image super-resolution only utilizes the single low-resolution image to reconstruct a high-resolution image, recovers high-frequency components lost in the down-sampling process, and improves the effective resolution of the image.
The process of acquiring images by the imaging device can be regarded as a degradation process from a high-resolution image to a low-resolution image, and the image super-resolution reconstruction is the inverse process of the degradation process, namely the process of reconstructing the high-resolution image by a low-resolution observation image, which is a pathological inverse problem, the solution is not unique, and therefore, a space of feasible solution is required to be constrained by image prior. Traditional priors generally mathematically model the statistical properties of natural images, whereas depth priors refer to prior information that learns or represents images through a depth convolutional neural network. The mathematical expression models complex image prior difficultly in display, a deep network can be used for implicitly learning the mapping relation between data through a deep convolutional neural network, and the deep convolutional neural network can be flexibly used as a submodule of a regularization method. Depth priors can be divided into explicit and implicit modeled image priors. One type of method requires training the network using a data set to learn some potential prior information of the image; the other is to realize the prior constraint of the image by using a network structure without supervision training.
The image super-resolution method can be divided into a non-blind super-resolution method and a blind super-resolution method according to whether a fuzzy kernel is known or not. When the fuzzy kernel is known, the method is a non-blind image super-resolution method, and when the fuzzy kernel is unknown, the method is a blind image super-resolution method. Non-blind image super-resolution methods typically assume a blur kernel as a gaussian function or a bicubic interpolation function. Since the blur kernel in the real scene is complex, when there is a large difference between the real blur kernel and the assumed blur kernel, the reconstruction effect cannot be guaranteed. The blind image super-resolution method estimates a real fuzzy kernel in a degradation process by utilizing potential information in a low-resolution image and reconstructs a high-resolution image.
Although the research of deep learning in image super-resolution has made remarkable progress, most of the existing image super-resolution methods based on deep learning are non-blind, and the blind image super-resolution research work is not much. The image super-resolution method based on deep learning is divided into a method of supervised learning and a method of unsupervised learning according to whether a training data set is needed or not. The supervised learning method constructs a high-resolution image data set and a low-resolution image data set through a predefined degradation process, and trains a network by using the data sets. The self-supervised learning is a common non-supervised learning method in the image super-resolution problem, a low-resolution image is used as a supervision signal, additional information is mined by the image to reconstruct the image, and a true value training network of the high-resolution image or a fuzzy core is not needed. The DRN needs a training network of a paired simulation image data set and an unpaired real image data set, and comprises two network structures of a DRNS with a smaller parameter quantity and a DRNL with a larger parameter quantity. The SRGAN introduces a content loss function training network into the generating countermeasure network, wherein the content loss is defined as the Euclidean distance between the characteristics of a high-resolution reconstructed image and the characteristics of a true image, and the content loss function is used for reconstructing the high-frequency content of the image and improving the visual effect of the reconstructed image. DRN and SRGAN are non-blind image super-resolution methods for supervised learning. The ZSSR uses the assumed or estimated fuzzy core to construct an image pyramid of the low-resolution images according to the self-similarity of the images, and uses the high-low resolution images formed by the multi-layer low-resolution images and the downsampled images to train network parameters. In the ZSSR, although fuzzy kernel information is introduced into the image reconstruction process in the reconstruction process, the fuzzy kernel is not estimated, and the ZSSR does not belong to a blind image super-resolution method in a strict sense.
The blind image super-resolution method based on deep learning can be divided into independent solving and combined modeling according to a fuzzy kernel solving process. In the independently solved blind image super-resolution method, fuzzy kernel estimation and high-resolution image estimation are two independent stages, the error of the fuzzy kernel estimation stage can influence the result of the subsequent high-resolution image estimation, and the result of the high-resolution image estimation cannot correct the error of the fuzzy kernel estimation. And the fuzzy kernel and the high-resolution image can be estimated simultaneously by jointly modeling, the estimation problem of the fuzzy kernel and the high-resolution image is expressed as an optimization problem comprising two decision variables, and the fuzzy kernel and the high-resolution image are corrected mutually by alternately solving. The KernelGAN model proposed by Bell-Kligler et al uses a generative confrontation network, uses a deep linear network as a generator to model an image degradation process and generate a down-sampled version of a low-resolution image, uses a discriminator to judge the cross-scale self-similarity between the low-resolution image and the down-sampled image, and estimates a blur kernel by maximizing the cross-scale self-similarity therebetween. The method belongs to a blind image super-resolution method for self-supervised learning for independently solving a fuzzy kernel, and the existing blind image super-resolution method for combined modeling is few.
The invention discloses a blind image super-resolution method based on depth prior self-supervision learning, which combines a network model and a mathematical model and adopts a combined modeling mode to simultaneously estimate a fuzzy kernel and a high-resolution image. The method estimates the high-resolution image by using a depth convolution neural network DIP-Net, introduces a non-local attention module, provides additional information required by image reconstruction by using complementary constraint among similar image blocks, estimates a blur kernel by solving an accurate solution about a blur kernel optimization problem, and alternately and iteratively estimates the blur kernel and the high-resolution image. The depth prior has the smoothness constraint effect on the image, and the convolutional neural network has the invariance, implicitly utilizes the self-similarity of the image, and combines a non-local attention module to explicitly utilize the self-similarity of the image. The invention takes the low-resolution image as the self-supervision signal, does not need the high-resolution image or the fuzzy check value and has no training process. The method disclosed by the invention can accurately estimate the fuzzy core and effectively reconstruct the edge and detail information of the image.
Disclosure of Invention
In view of this, the embodiment of the present invention provides a depth prior-based blind image super-resolution method, so as to implement blind super-resolution reconstruction of a low-resolution image.
In order to achieve the above object, an embodiment of the present invention provides the following solutions:
a blind image super-resolution method based on depth prior is characterized by comprising the following 5 steps:
step 1, constructing an image generation network model
The method uses DIP-Net as an image generation network to realize the mapping x (z; theta) from a random vector z to a high-resolution image x, utilizes the network to inhibit noise, and implicitly models a smoothness constraint prior term. The DIP-Net network input z is a random vector uniformly distributed on the interval (0; 1), namely z-U (0; 1). Randomly initializing a group of network parameters, using the degraded image as a self-monitoring signal, and updating the parameters by using a gradient descent method
Figure BDA0003320391880000031
Causing the loss function to converge.
According to the invention, non-local constraint is introduced into DIP-Net, and additional information contained in similar image blocks is explicitly obtained through mutual constraint among the similar image blocks. Non-local operations are defined as follows:
Figure BDA0003320391880000032
in the formula, xiIn order to be able to process the features,
Figure BDA0003320391880000033
as output characteristics, xjIs xiThe neighborhood feature of (c), ρ (x)i;xj) As a function of similarity, for measuring xiAnd xjThe correlation between the two, g (-) is a feature extraction function,c (x) is a normalization parameter. The similarity function generally has four forms of a gaussian function, an embedded gaussian function, an inner product and a cascade.
The non-local attention module is defined as the following form on the basis of non-local operation:
Figure BDA0003320391880000034
non-local attention module uses residual connection' xi+ "so that it can be embedded in any pre-trained network without affecting the network's task, i.e., when W iszWhen 0, the network keeps the original structure.
The image generation network DIP-Net used by the invention has a U-shaped coding and decoding structure and comprises five groups of downsampling and upsampling convolution structures, wherein the third, fourth and fifth groups of downsampling layer characteristics of the network are added with a non-local attention module. Each group of convolution operations fuses the features of the down-sampling layer and the features of the up-sampling layer with the same corresponding dimension through cross-layer connection, and the number of channels connected in the cross-layer mode is fixed to be 16. The network input z is a random vector subject to uniform distribution, the size of the random vector is consistent with that of a high-resolution image, the number of channels is generally set to be 8 or 16, and the invention is set to be 8.
Step 2, initializing network parameters
Randomly initializing a parameter θ of an image generation network0I.e. to obtain an initial estimate x of the high resolution image0=f(z;θ0). Setting learning rate eta, down-sampling factor a, fuzzy kernel size s, network input random vector z and fuzzy kernel regularization parameter lambdahAnd a maximum iteration number K. Since the blind super-resolution method does not know the true size of the blur kernel, the size of the blur kernel needs to be estimated or preset when reconstructing a high-resolution image.
Step 3, estimating fuzzy core
Fixed network parameter thetak-1Estimate the blur kernel h byk
Figure BDA0003320391880000035
In the formula (I), the compound is shown in the specification,
Figure BDA0003320391880000041
which represents the fourier transform of the signal,
Figure BDA0003320391880000042
which represents the inverse of the fourier transform,
Figure BDA0003320391880000045
complex conjugate representing Fourier transform, etcaRepresenting the elemental dot product operation on a x a image blocks,
Figure BDA0003320391880000046
carrying out average processing operation on a multiplied by a image blocks, ° caIndicating a-fold upsampling with zero padding.
Step 4, estimating high-resolution image
Step 4.1 calculate loss function:
fixing the estimate h of the current blur kernelkGiven of thetak-1Update thetak. With updated fuzzy kernel hkWith estimated high-resolution image xk-1Generating a down-sampled image through variable step convolution operation, and calculating a loss function according to the low-resolution image and the down-sampled image:
Figure BDA0003320391880000043
the above equation is a mean square error loss function, and other continuously derivable functions may be used as the loss function of the network.
Step 4.2, updating the image generation network parameters:
calculating the gradient of the loss function with respect to the network parameter, and updating theta by using a gradient descent method shown in the following formulak
Figure BDA0003320391880000044
In the formula, η represents a learning rate. The invention updates the parameters using the Adam gradient descent algorithm.
Step 4.3 generating a high resolution image: generation of high resolution image x using updated parameter image generation networkk=f(z;θk)。
Step 5, judging convergence, outputting fuzzy core and estimating high-resolution image
Through the steps 3 and 4, the objective function is solved in an iterative mode once to obtain the estimation h of the fuzzy kernelkAnd estimate x of the high resolution imagek-1Is updated to xk. If the maximum iteration times or iteration convergence is reached, stopping iteration and outputting a final fuzzy kernel and high-resolution image estimation; otherwise, let k be k +1, and then repeat steps 3 and 4.
Preferably, the similarity function of the non-local attention module of the algorithm takes the form of an inner product represented by the following formula,
ρ(xi,xj)=θ(xi)Tφ(xj)
in the formula, θ (-) and Φ (-) are feature extraction functions, and c (x) ═ N is used as a normalization parameter, where N represents the total number of features in x, so as to simplify the calculation of the gradient. Since the size of the input features varies, it is more appropriate to use the number of features N of x as the normalization parameter.
Preferably, the number of iterations of the algorithm is 3000.
Preferably, the learning rate η has an initial value of 0.001, and the attenuation is 0.5 times that of the previous time per 500 iterations.
Preferably, the blur kernel regularization parameter λhIs 2 x 10 as an initial value-5The 1000 increases per iteration are 1.2 times the previous.
The invention discloses a blind image super-resolution method based on depth prior self-supervision learning, which is used for reconstructing a high-resolution image from end to end. The method combines a network model and a mathematical model, and simultaneously estimates the fuzzy core and the high-resolution image by adopting a combined modeling mode. Estimating a high-resolution image by using a depth convolution neural network DIP-Net, introducing a non-local attention module, providing additional information required for reconstructing the image by using complementary constraint among similar image blocks, and estimating a blur kernel by solving an accurate solution about a blur kernel optimization problem, and alternately and iteratively estimating the blur kernel and the high-resolution image. The invention takes the low-resolution image as the self-supervision signal, and does not need the training process of a data set. The method disclosed by the invention can accurately estimate the fuzzy core and effectively reconstruct the edge and detail information of the image.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic diagram of a blind image super-resolution reconstruction provided by an embodiment of the present invention;
FIG. 2 is a schematic diagram of a structure of a blind image super-resolution method based on depth prior provided by an embodiment of the present invention;
FIG. 3 is a flowchart of a method for super-resolution of a blind image based on depth prior according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a non-local attention module according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of an image generation network provided by an embodiment of the present invention;
FIG. 6 is a comparison of the average PSNR and SSIM for various methods on the DIVRK data set provided by embodiments of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
In the image super-resolution problem, the degradation model between the low-resolution image and the original high-resolution image is usually expressed in the form of a convolution as follows:
y=(h*x)↓a+n (1)
in the formula, y is a low-resolution image, h is a blur kernel, x is a high-resolution image, a is a down-sampling factor, n is additive noise, and x is a two-dimensional convolution operation. Under the convolution model, the image blind super-resolution method is to study how to estimate the blur kernel h and the high-resolution image x from the low-resolution image y at the same time, as shown in fig. 1.
The invention uses the image generation network to reconstruct the high-resolution image, and introduces the regularization constraint term of the fuzzy kernel h as fuzzy kernel prior, and the objective function can be expressed as:
Figure BDA0003320391880000061
wherein h is a fuzzy kernel, f (z; theta) is an image generation network with the network input being z and the network parameter being theta, and lambdahIs a regularization parameter. The former item in the target function is a data fidelity item, and the reconstructed image is ensured to accord with a degradation model; the latter term is a regularization constraint term of the fuzzy kernel h, and the smoothness of the fuzzy kernel is constrained.
Fig. 2 is an overall structure of the blind image super-resolution method based on depth prior disclosed by the invention, and the optimization problem shown in formula (2) is solved by an alternative iteration method in a mode of combining a network model and a mathematical model. The image is estimated by using DIP-Net, the process of calculating the loss function and updating the network parameter theta is the image updating process, the fuzzy kernel is estimated by using the optimality condition of the fuzzy kernel minimization problem, the high-resolution image and the fuzzy kernel are alternately updated in an iterative mode, and finally the network output x ═ f (z; theta) is the reconstructed high-resolution image.
The embodiment of the invention discloses a depth prior-based blind image super-resolution method, which is used for realizing blind super-resolution reconstruction of a low-resolution image. Referring to fig. 3, the above method includes at least the following 5 steps.
Step 1, constructing an image generation network model
The image generation network DIP-Net implements the random vector z to sharp image x mapping x ═ f (z; θ). The DIP-Net aims at solving the problem of image denoising, a network takes a noisy image as a supervision signal, a loss function is calculated to update network parameters, and a noise-free clear image can be preferentially generated in the process of fitting the noisy image by a random vector. DIP-Net can be effectively extended to solve a variety of image inverse problems, and the image super-resolution problem can be generally modeled as an optimization problem as follows:
Figure BDA0003320391880000062
wherein y is a low resolution image, h is a blur kernel, x is a high resolution image, ↓, andafor down-sampling operations, R (x) is a smoothness function, λxIs a regular term coefficient. The former item in the target function is a data fidelity item, and the reconstructed image is ensured to accord with a degradation model; the latter term is a smoothness prior constraint term, and suppresses noise in the reconstruction process.
The DIP-Net generates a network f (z; theta) through an image, noise is suppressed by using the network, the method is equivalent to implicitly establishing a smoothness constraint prior term R (x) in a mode (3), and a network loss function is as follows:
Figure BDA0003320391880000063
where theta is a network parameter, z is a network input,
Figure BDA0003320391880000064
i.e. the reconstructed image. The DIP-Net network input z is a random vector uniformly distributed on the interval (0; 1), namely z-U (0; 1). Randomly initializing a set of network parameters with a degraded image asUpdating parameters by gradient descent method based on self-supervision signal
Figure BDA0003320391880000065
Causing the loss function to converge. DIP-Net is essentially a regularization method, network parameters need to be estimated for each image, and the parameter update process is actually the solution process to the optimization problem. DIP-Net uses the degraded image itself as a supervision signal, without the need for a training process and training data set, and can be considered as self-supervised learning.
According to the invention, non-local constraint is introduced into DIP-Net, and additional information contained in similar image blocks is explicitly obtained through mutual constraint among the similar image blocks. There is a redundant similar image structure in a natural image, and the similar structure is embodied as similar image blocks in the image, and 80% of the image blocks in the same image have 9 or more similar image blocks. A large number of similar image blocks embedded in the natural image can provide additional information for super-resolution of the image. For any image block in an image, there are multiple similar image blocks in the entire image or a range of search windows. The non-local regularized super-resolution method utilizes complementary information provided by the similar image blocks in the form of regularization constraints by constructing non-local constraint terms. For the target image block xiSearching for L image blocks similar thereto
Figure BDA0003320391880000071
And estimating the target by using the similar image blocks, wherein the difference between the target value and the estimated value is a non-local constraint term.
Non-local operations are defined as follows:
Figure BDA0003320391880000072
in the formula, xiIn order to be able to process the features,
Figure BDA0003320391880000073
as output characteristics, xjIs xiThe neighborhood feature of (c), ρ (x)i;xj) As a function of similarity, for measuring xiAnd xjThe correlation between g (-) and c (x) is a normalization parameter.
The non-local attention module is defined as the following form on the basis of non-local operation:
Figure BDA0003320391880000074
non-local attention module uses residual connection' xi+ "so that it can be embedded in any pre-trained network without affecting the network's task, i.e., when W iszWhen 0, the network keeps the original structure.
The similarity function generally has four forms, namely a gaussian function, an embedded gaussian function, an inner product and a cascade, and the four forms of similarity functions are described below.
1. Gaussian function
The similarity function in the form of a Gaussian function is defined as a function of the non-local mean and bilateral filtering
Figure BDA0003320391880000075
Inner product operation can be replaced by Euclidean distance generally, but the inner product operation is easier to realize in a deep convolutional neural network, and the normalization parameter is defined as
Figure BDA0003320391880000076
2. Embedded gaussian function
A simple variant of the Gaussian function is to compute the similarity in the embedded space, i.e.
ρ(xi,xj)=exp(θ(xi)Tφ(xj)) (8)
In the formula, θ (-) and φ (-) are feature extraction functions, and the normalization parameters are defined as in the Gaussian function.
3. Inner product form
The similarity function can also be defined directly in the form of an inner product, i.e.
ρ(xi;xj)=θ(xi)Tφ(xj) (9)
In the formula, c (x) ═ N is used as the normalization parameter, where N represents the total number of features in x, which can simplify the gradient calculation. Since the size of the input features varies, it is more appropriate to use the number of features N of x as the normalization parameter.
4. Form of cascade
Using cascades as similarity functions, i.e.
Figure BDA0003320391880000081
Wherein [:]represents a cascade, wpFor the weight vector, the concatenated vector may be converted into a scalar, where c (x) is also set to N.
The similarity function of this embodiment adopts an inner product form, and similarity functions of other forms all belong to the protection scope of the present invention. FIG. 4 is a network structure of a non-local attention module used in the present embodiment, wherein Wθ、WφAnd WgThe convolution matrices, theta (-), phi (-), and g (-), respectively, are implemented by a 1 × 1 convolution. In the drawings
Figure BDA0003320391880000082
It is shown that the matrix multiplication operation,
Figure BDA0003320391880000083
indicating an element-by-element addition operation, and 1/N indicating an operation of scaling the features after the convolution operation by 1/N. The non-local attention module has no requirement on the size of the input feature, can be flexibly embedded into any position of the deep convolutional neural network, constructs constraint through the similarity of the feature to be processed and the neighborhood features of different positions, and utilizes the deep convolutional neural network to acquire additional information contained in the self-similarity of the image to reconstruct the height of the high-resolution imageFrequency details.
The image generation network DIP-Net used by the invention has a U-shaped coding and decoding structure, the specific structure of which is shown in figure 5, and comprises five groups of downsampling and upsampling convolution structures, wherein the third, fourth and fifth groups of downsampling layer features of the network are added with a non-local attention module. Each group of convolution operations fuses the features of the down-sampling layer and the features of the up-sampling layer with the same corresponding dimension through cross-layer connection, and the number of channels connected in the cross-layer mode is fixed to be 16. The network input z is a random vector subject to uniform distribution, the size of the random vector is consistent with that of a high-resolution image, the number of channels is generally set to be 8 or 16, and the invention is set to be 8. The size of the input vector in the graph is 8 × 271 × 271, the size of the input feature map is reduced from 271 × 271 to 9 × 9 through the first to sixth groups of convolution layers, and the input feature map is restored to the same size as the input random vector through the first to fifth groups of upsampling layers.
Step 2, initializing network parameters
Randomly initializing a parameter θ of an image generation network0I.e. to obtain an initial estimate x of the high resolution image0=f(z;θ0). Setting learning rate eta, down-sampling factor a, fuzzy kernel size s, network input random vector z and fuzzy kernel regularization parameter lambdahAnd a maximum iteration number K. Since the blind super-resolution method does not know the true size of the blur kernel, the size of the blur kernel needs to be estimated or preset when reconstructing a high-resolution image. The invention solves the network parameter theta and the fuzzy kernel h in the formula (2) by adopting an alternative solving mode, namely, the network parameter theta of the high-resolution image is fixedly estimated firstlyk-1Solving the fuzzy kernel hkThen fix the estimate h of the blur kernelkSolving for the network parameter θ of the estimated high resolution imagekUntil convergence or a maximum number of iterations is reached.
Step 3, estimating fuzzy core
Fixed network parameter thetak-1Estimating a blur kernel hkIn this case, the optimization problem can be expressed as:
Figure BDA0003320391880000084
wherein f (z; theta)k-1) Is the network output for the (k-1) th iteration. To simplify the mathematical expression, let xk-1=f(z;θk-1) Equation (11) can be written as:
Figure BDA0003320391880000085
converting equation (12) into the form of a matrix-vector product:
Figure BDA0003320391880000091
where D is a down-sampling matrix, Xk-1For high resolution image xk-1The corresponding block circulant matrix. Equation (13) is a quadratic function with respect to h, there is a closed solution:
Figure BDA0003320391880000092
equation (14) requires computation of large-scale matrix inversions, and the present invention employs solving a closed solution in the frequency domain. Is provided with
Figure BDA0003320391880000093
Having a diagonal element of Xk-1Fourier coefficients of the first column, denoted Λ as a block diagonal matrix, i.e.
Figure BDA0003320391880000094
Each sub-block
Figure BDA0003320391880000095
Also a diagonal matrix, denoted Γ as
Figure BDA0003320391880000096
Derived from equation (14):
Figure BDA0003320391880000097
wherein F is a Fourier transform matrix, FHIs an inverse fourier transform matrix. The frequency domain solution of the blur kernel can be written according to equation (15):
Figure BDA0003320391880000098
in the formula (I), the compound is shown in the specification,
Figure BDA0003320391880000099
which represents the fourier transform of the signal,
Figure BDA00033203918800000910
which represents the inverse of the fourier transform,
Figure BDA00033203918800000911
complex conjugate representing Fourier transform, etcaRepresenting the elemental dot product operation on a x a image blocks,
Figure BDA00033203918800000914
carrying out average processing operation on a multiplied by a image blocks, ° caIndicating a-fold upsampling with zero padding.
Step 4, estimating high-resolution image
Step 4.1 calculate loss function:
fixing the estimate h of the current blur kernelkGiven of thetak-1Update thetakAt this time, the objective function is simplified as:
Figure BDA00033203918800000912
with updated fuzzy kernel hkWith estimated high-resolution image xk-1The down-sampled image is generated by a variable step convolution operation. Calculating a loss function according to the low-resolution image and the down-sampled image, wherein an objective function in the formula (17) is a loss function of the network:
Figure BDA00033203918800000913
equation (18) is the mean square error loss function, but other continuously derivable functions may be used as the loss function of the network.
Step 4.2, updating the image generation network parameters:
the gradient of the loss function of the calculation formula (18) with respect to the network parameter is updated by a gradient descent methodk
Figure BDA0003320391880000101
In the formula, η represents a learning rate. In order to simplify the mathematical expressions, let the gradient
Figure BDA0003320391880000102
The invention updates network parameters by using an Adam gradient descent method, and the Adam algorithm uses momentum vkAnd second order momentum s in the RMSProp algorithmk. Initialization v0s 00, given a hyperparameter of 0 ≦ β1< 1, momentum v of the k-th iterationkExpressed as the gradient gk-1Exponentially weighted moving average of (d):
vk=β1vk-1+(1-β1)gk-1 (20)
given a hyperparameter of 0 ≦ β2<1,skExpressed as the gradient squared term gk-1⊙gk-1Exponentially weighted moving average of (d):
sk=β2sk-1+(1-β2)gk-1⊙gk-1 (21)
in the formula, "-" indicates element-by-element multiplication. Due to v0And s0Are all initialized to zero, and at the k-th iteration, the momentum vkExpressed as:
Figure BDA0003320391880000103
the gradient weights of each previous iteration are added to obtain,
Figure BDA0003320391880000104
when k is small, the sum of the gradient weights for each iteration is small. To eliminate such effects, for the k-th iteration, v iskIs divided by
Figure BDA0003320391880000105
And making the sum of the gradient weights of each past iteration be 1, and called deviation correction. In the Adam algorithm, the variable v is addedkAnd skAnd (4) correcting deviation:
Figure BDA0003320391880000106
Figure BDA0003320391880000107
adam algorithm uses a bias-corrected variable v'kAnd s'kAnd learning rate eta update gradient g'k-1
Figure BDA0003320391880000108
In the formula, eta is a learning rate, and each element of the independent variable in Adam has a different learning rate; e is a constant to avoid the case where the denominator is 0 in equation (26). Use g 'in the kth iteration'k-1The parameters of the network are updated and,
θk=θk-1-g′k-1 (27)
step 4.3 generating a high resolution image: generating network generator by using image with updated parametersInto a high resolution image xk=f(z;θk)。
Step 5, judging convergence, outputting fuzzy core and estimating high-resolution image
Through the steps 3 and 4, the objective function is solved in an iterative mode once to obtain the estimation h of the fuzzy kernelkAnd estimate x of the high resolution imagek-1Is updated to xk. If the maximum iteration times or iteration convergence is reached, stopping iteration and outputting a final fuzzy kernel and high-resolution image estimation; otherwise, let k be k +1, and then repeat steps 3 and 4.
Preferably, the iteration number of the algorithm is set to 3000, and the learning rate is set to be 3000ηIs 0.001, the attenuation of 500 iterations is 0.5 times that of the last iteration, and the fuzzy kernel regularization parameter lambda ishIs 2 x 10 as an initial value-5The 1000-time improvement per iteration is 1.2 times that of the previous time, and the blur kernel sizes of 2-time and 4-time super-resolution are set to 11 × 11 and 23 × 23, respectively.
For video memory reasons, the present invention constructs a DIVRK simulated low resolution dataset screening 20 different classes of images, including animals, sculptures, aerial photographs, buildings, plants, text, and people, of size 816 × 816 to 904 × 904 from the DIV2K dataset for evaluation of blind image super-resolution algorithms. According to the construction mode of a DIV2KRK data set, different and randomly generated anisotropic Gaussian blur kernels are used for each image to carry out convolution and down-sampling operation, and low-resolution images with down-sampling factors of 2 and 4 are generated respectively. The generation process of the anisotropic Gaussian blur kernel comprises the following steps: randomly setting the variance λ in the horizontal and vertical directions1;λ2U (0:35, 10), randomly rotating by an angle theta to U (-pi, pi), and performing normalization, wherein U (a, b) represents an interval [ a, b ]]Is uniformly distributed. When the down-sampling factor is 2, the blur kernel size is 11 × 11, and when the down-sampling factor is 4, the blur kernel size is 23 × 23.
The peak signal-to-noise ratio PSNR and the structural similarity SSIM are used as quantitative evaluation indexes. PSNR calculates the average pixel error between a reconstructed image and a true value image, SSIM calculates the structural similarity between the reconstructed image and the true value image, the result is between [0 and 1], and the higher the two indexes are, the better the reconstruction quality is. SRGAN and DRN are currently generally accepted non-blind image super-resolution methods, and Kernelgan + ZSSR represents a method for realizing blind image super-resolution by using a fuzzy kernel estimated by Kernelgan and combining ZSSR. Fig. 6 lists the average PSNR and SSIM of various image super-resolution algorithms on the DIVRK data set. As can be seen from the table, for the super-resolution reconstruction of 2 times and 4 times, the average PSNR and SSIM of the invention are both the highest, and the quantitative experiment result proves that the invention achieves better image reconstruction quality.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (5)

1.一种基于深度先验的盲图像超分辨率方法,其特征在于,包含以下5个步骤:1. a blind image super-resolution method based on depth prior, is characterized in that, comprises following 5 steps: 步骤1.构建图像生成网络模型Step 1. Build an image generation network model 使用DIP-Net作为图像生成网络实现随机向量z到高分辨率图像x的映射x=f(z;θ),利用网络本身抑制噪声,隐式地建模平滑性约束先验项;DIP-Net网络输入z为区间(0,1)上均匀分布的随机向量,即z~U(0,1);随机初始化一组网络参数,以降质图像作为自监督信号,利用梯度下降法更新参数
Figure FDA0003320391870000011
使损失函数收敛;
Use DIP-Net as an image generation network to realize the mapping of random vector z to high-resolution image x x = f(z; θ), use the network itself to suppress noise, and implicitly model the smoothness constraint prior term; DIP-Net The network input z is a random vector uniformly distributed in the interval (0, 1), that is, z~U(0, 1); a set of network parameters are randomly initialized, and the degraded image is used as a self-supervision signal, and the parameters are updated by the gradient descent method
Figure FDA0003320391870000011
make the loss function converge;
将非局部约束引入DIP-Net中,通过相似图像块间的相互约束显式地获取相似图像块蕴含的附加信息;非局部运算定义为如下所示形式:Non-local constraints are introduced into DIP-Net, and the additional information contained in similar image blocks is explicitly obtained through mutual constraints between similar image blocks; non-local operations are defined as follows:
Figure FDA0003320391870000012
Figure FDA0003320391870000012
式中,xi为待处理的特征,
Figure FDA0003320391870000013
为输出特征,xj为xi的邻域特征,ρ(xi,xj)为相似度函数,用于度量xi和xj之间的相关性,g(·)为特征提取函数,C(x)为归一化参数;其中,相似度函数通常具有高斯函数、嵌入式高斯函数、内积与级联四种形式;
where x i is the feature to be processed,
Figure FDA0003320391870000013
is the output feature, x j is the neighborhood feature of x i , ρ(x i , x j ) is the similarity function, used to measure the correlation between x i and x j , g( ) is the feature extraction function, C(x) is the normalization parameter; among them, the similarity function usually has four forms: Gaussian function, embedded Gaussian function, inner product and cascade;
在非局部运算的基础上非局部注意力模块定义为如下形式:On the basis of non-local operations, the non-local attention module is defined as follows:
Figure FDA0003320391870000014
Figure FDA0003320391870000014
非局部注意力模块使用残差连接“xi+”,使其可以嵌入到任何预训练的网络中,且不会影响网络的任务,即当Wz=0时,网络保持原有结构;The non-local attention module uses the residual connection " xi +", so that it can be embedded in any pre-trained network without affecting the task of the network, that is, when W z = 0, the network maintains the original structure; 使用的图像生成网络DIP-Net具有U型编解码结构,包含五组下采样和上采样卷积结构,其中网络的第三、四、五组下采样层特征加入非局部注意力模块;每一组卷积操作均通过跨层连接将下采样层的特征与对应相同维数的上采样层特征进行融合,跨层连接的通道数固定为16;网络输入z为服从均匀分布的随机向量,其尺寸与高分辨率图像一致,通道数一般设置为8或16,设置为8;The used image generation network DIP-Net has a U-shaped codec structure, including five groups of downsampling and upsampling convolution structures, in which the third, fourth, and fifth groups of downsampling layer features of the network are added to the non-local attention module; each The group convolution operation fuses the features of the down-sampling layer with the features of the up-sampling layer corresponding to the same dimension through cross-layer connections, and the number of channels for cross-layer connections is fixed at 16; the network input z is a random vector that obeys a uniform distribution, and its The size is consistent with the high-resolution image, and the number of channels is generally set to 8 or 16, and set to 8; 步骤2.初始化网络参数Step 2. Initialize network parameters 随机初始化图像生成网络的参数θ0,即可得高分辨率图像的初始估计x0=f(z;θ0);设置学习率η、降采样因子a、模糊核尺寸s、网络输入随机向量z、模糊核正则化参数λh、最大迭代次数K;由于盲超分辨率方法不知道模糊核的真实尺寸,在重建高分辨率图像时需要估计模糊核尺寸,或者预先设定模糊核尺寸;Randomly initialize the parameters θ 0 of the image generation network to obtain the initial estimate of the high-resolution image x 0 =f(z; θ 0 ); set the learning rate η, the downsampling factor a, the size of the blur kernel s, and the network input random vector z, the blur kernel regularization parameter λ h , the maximum number of iterations K; since the blind super-resolution method does not know the true size of the blur kernel, it is necessary to estimate the size of the blur kernel when reconstructing a high-resolution image, or pre-set the size of the blur kernel; 步骤3.估计模糊核Step 3. Estimating the blur kernel 固定网络参数θk-1,用下式估计模糊核hkWith the network parameters θ k-1 fixed, the blur kernel h k is estimated by the following equation:
Figure FDA0003320391870000021
Figure FDA0003320391870000021
式中,
Figure FDA0003320391870000022
表示傅里叶变换,
Figure FDA0003320391870000023
表示傅里叶逆变换,
Figure FDA0003320391870000024
表示傅里叶变换的复共轭,⊙a表示对a×a个图像块进行元素点乘运算,
Figure FDA0003320391870000025
是对a×a个图像块进行平均处理操作,↑a表示利用零填充的方式进行a倍上采样;
In the formula,
Figure FDA0003320391870000022
represents the Fourier transform,
Figure FDA0003320391870000023
represents the inverse Fourier transform,
Figure FDA0003320391870000024
represents the complex conjugate of the Fourier transform, ⊙ a represents the element-wise point multiplication of a×a image blocks,
Figure FDA0003320391870000025
It is an average processing operation on a×a image blocks, ↑ a means a times upsampling by zero padding;
步骤4.估计高分辨率图像Step 4. Estimating high-resolution images 步骤4.1计算损失函数:Step 4.1 Calculate the loss function: 固定当前模糊核的估计hk,给定θk-1,更新θk;利用更新的模糊核hk与估计的高分辨率图像xk-1通过可变步长的卷积操作生成降采样图像,根据低分辨率图像与降采样图像计算损失函数:Fix the current blur kernel estimate h k , given θ k-1 , update θ k ; use the updated blur kernel h k and the estimated high-resolution image x k-1 to generate downsampling through a variable-step convolution operation Image, calculate the loss function based on the low-resolution image and the downsampled image:
Figure FDA0003320391870000026
Figure FDA0003320391870000026
上式为均方误差损失函数,此外也能够使用其他连续可导的函数作为网络的损失函数;The above formula is the mean square error loss function, and other continuously derivable functions can also be used as the loss function of the network; 步骤4.2更新图像生成网络参数:Step 4.2 Update image generation network parameters: 计算损失函数关于网络参数的梯度,利用下式所示的梯度下降法更新θkCalculate the gradient of the loss function with respect to the network parameters, and update θ k using the gradient descent method shown in the following equation,
Figure FDA0003320391870000031
Figure FDA0003320391870000031
式中η表示学习率;使用Adam梯度下降算法更新参数;where η represents the learning rate; the parameters are updated using the Adam gradient descent algorithm; 步骤4.3生成高分辨率图像:利用更新参数后的图像生成网络生成高分辨率图像xk=f(z;θk);Step 4.3 Generate a high-resolution image: generate a high-resolution image x k =f(z; θ k ) using the image generation network after updating the parameters; 步骤5.判断收敛,输出模糊核与高分辨率图像的估计Step 5. Judging convergence, output the estimation of blur kernel and high-resolution image 通过步骤3和步骤4,进行对目标函数的一次迭代求解,获得模糊核的估计hk,并将高分辨率图像的估计xk-1更新为xk;如果此时达到最大迭代次数或者迭代收敛,则停止迭代,输出最终的模糊核与高分辨率图像估计;否则,令k=k+1,然后重复步骤3和步骤4。Through steps 3 and 4, an iterative solution of the objective function is performed to obtain the estimated blur kernel h k , and the estimated high-resolution image x k-1 is updated to x k ; if the maximum number of iterations or iterations is reached at this time If it converges, stop the iteration and output the final blur kernel and high-resolution image estimation; otherwise, let k=k+1, and then repeat steps 3 and 4.
2.如权利要求1所述的一种基于深度先验的盲图像超分辨率方法,其特征在于,所述算法非局部注意力模块的相似度函数采用下式所示内积形式,2. a kind of blind image super-resolution method based on depth prior as claimed in claim 1, is characterized in that, the similarity function of described algorithm non-local attention module adopts the inner product form shown in the following formula, ρ(xi,xj)=θ(xi)Tφ(xj)ρ(x i , x j )=θ(x i ) T φ(x j ) 式中,θ(·)与φ(·)为特征提取函数,归一化参数选用C(x)=N,其中N表示x中特征的总数目,简化梯度的计算;由于输入特征的尺寸是变化的,将x的特征数目N作为归一化参数。In the formula, θ( ) and φ( ) are the feature extraction functions, and the normalization parameter selects C(x)=N, where N represents the total number of features in x, which simplifies the calculation of the gradient; since the size of the input feature is Variation, take the number of features N of x as the normalization parameter. 3.如权利要求1所述的一种基于深度先验的盲图像超分辨率方法,其特征在于,所述算法迭代次数为3000。3 . The blind image super-resolution method based on depth prior according to claim 1 , wherein the number of iterations of the algorithm is 3000. 4 . 4.如权利要求1所述的一种基于深度先验的盲图像超分辨率方法,其特征在于,所述学习率η的初始值为0.001,每迭代500次衰减为上一次的0.5倍。4 . The blind image super-resolution method based on depth prior according to claim 1 , wherein the initial value of the learning rate η is 0.001, and the decay is 0.5 times of the previous time every 500 iterations. 5 . 5.如权利要求1所述的一种基于深度先验的盲图像超分辨率方法,其特征在于,所述模糊核正则化参数λh的初始值为2×10-5,每迭代1000次提高为上一次的1.2倍。5 . The blind image super-resolution method based on depth prior according to claim 1 , wherein the initial value of the blur kernel regularization parameter λ h is 2×10 −5 , and each iteration is 1000 times. 6 . The increase is 1.2 times the previous time.
CN202111244295.3A 2021-10-26 2021-10-26 Blind image super-resolution method based on depth priori Active CN114202459B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111244295.3A CN114202459B (en) 2021-10-26 2021-10-26 Blind image super-resolution method based on depth priori

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111244295.3A CN114202459B (en) 2021-10-26 2021-10-26 Blind image super-resolution method based on depth priori

Publications (2)

Publication Number Publication Date
CN114202459A true CN114202459A (en) 2022-03-18
CN114202459B CN114202459B (en) 2024-03-22

Family

ID=80646349

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111244295.3A Active CN114202459B (en) 2021-10-26 2021-10-26 Blind image super-resolution method based on depth priori

Country Status (1)

Country Link
CN (1) CN114202459B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114626989A (en) * 2022-03-30 2022-06-14 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Blind image super-resolution method and system based on domain-confrontation transformation
CN114972036A (en) * 2022-06-10 2022-08-30 上海交通大学 Blind image super-resolution reconstruction method and system based on fusion degradation prior
CN115131210A (en) * 2022-06-28 2022-09-30 闽江学院 Alternating optimization image blind super-resolution reconstruction method based on precise kernel estimation
CN115761529A (en) * 2023-01-09 2023-03-07 阿里巴巴(中国)有限公司 Image processing method and electronic device
CN118537229A (en) * 2024-06-03 2024-08-23 哈尔滨工业大学 Ultrasonic high-resolution imaging method based on memory-enhanced prior dynamic model

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107730468A (en) * 2017-10-23 2018-02-23 鹰艾思科技(深圳)有限公司 The restoration methods of picture rich in detail under a kind of UAV Fuzzy noise image
CN109767389A (en) * 2019-01-15 2019-05-17 四川大学 Adaptive weighted dual-norm remote sensing image blind super-resolution reconstruction method based on local and non-local joint priors
US20200302249A1 (en) * 2019-03-19 2020-09-24 Mitsubishi Electric Research Laboratories, Inc. Systems and Methods for Multi-Spectral Image Fusion Using Unrolled Projected Gradient Descent and Convolutinoal Neural Network
CN113096013A (en) * 2021-03-31 2021-07-09 南京理工大学 Blind image super-resolution reconstruction method and system based on imaging modeling and knowledge distillation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107730468A (en) * 2017-10-23 2018-02-23 鹰艾思科技(深圳)有限公司 The restoration methods of picture rich in detail under a kind of UAV Fuzzy noise image
CN109767389A (en) * 2019-01-15 2019-05-17 四川大学 Adaptive weighted dual-norm remote sensing image blind super-resolution reconstruction method based on local and non-local joint priors
US20200302249A1 (en) * 2019-03-19 2020-09-24 Mitsubishi Electric Research Laboratories, Inc. Systems and Methods for Multi-Spectral Image Fusion Using Unrolled Projected Gradient Descent and Convolutinoal Neural Network
CN113096013A (en) * 2021-03-31 2021-07-09 南京理工大学 Blind image super-resolution reconstruction method and system based on imaging modeling and knowledge distillation

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
李滔;何小海;滕奇志;吴小强;: "基于自适应双l_p-l_2范数的单幅模糊图像超分辨率盲重建", 计算机应用, no. 08, 10 August 2017 (2017-08-10) *
李键红;吕巨建;吴亚榕;: "基于最大化自相似性先验的盲单帧图像超分辨率算法", 计算机科学, no. 02, 15 February 2018 (2018-02-15) *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114626989A (en) * 2022-03-30 2022-06-14 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Blind image super-resolution method and system based on domain-confrontation transformation
CN114972036A (en) * 2022-06-10 2022-08-30 上海交通大学 Blind image super-resolution reconstruction method and system based on fusion degradation prior
CN114972036B (en) * 2022-06-10 2024-09-24 上海交通大学 A blind image super-resolution reconstruction method and system based on fusion degradation prior
CN115131210A (en) * 2022-06-28 2022-09-30 闽江学院 Alternating optimization image blind super-resolution reconstruction method based on precise kernel estimation
CN115761529A (en) * 2023-01-09 2023-03-07 阿里巴巴(中国)有限公司 Image processing method and electronic device
CN118537229A (en) * 2024-06-03 2024-08-23 哈尔滨工业大学 Ultrasonic high-resolution imaging method based on memory-enhanced prior dynamic model
CN118537229B (en) * 2024-06-03 2024-12-31 哈尔滨工业大学 A method for high-resolution ultrasound imaging based on memory-enhanced prior dynamic model

Also Published As

Publication number Publication date
CN114202459B (en) 2024-03-22

Similar Documents

Publication Publication Date Title
CN114202459B (en) Blind image super-resolution method based on depth priori
CN109146784B (en) An Image Super-Resolution Reconstruction Method Based on Multi-scale Generative Adversarial Networks
CN106952228B (en) A single image super-resolution reconstruction method based on image non-local self-similarity
CN114418883B (en) Blind image deblurring method based on depth priori
CN108765511B (en) Ultrasonic image super-resolution reconstruction method based on deep learning
CN110675321B (en) Super-resolution image reconstruction method based on progressive depth residual error network
CN101794440B (en) Weighted adaptive super-resolution reconstruction method for image sequences
CN108090871A (en) A kind of more contrast MR image reconstruction methods based on convolutional neural networks
CN111754403A (en) An Image Super-Resolution Reconstruction Method Based on Residual Learning
CN111123183B (en) A fast magnetic resonance imaging method based on complex R2U_Net network
CN111047515A (en) Cavity convolution neural network image super-resolution reconstruction method based on attention mechanism
CN107784628B (en) A Super-Resolution Implementation Method Based on Reconstruction Optimization and Deep Neural Networks
CN107993194B (en) Super-resolution reconstruction method based on stationary wavelet transform
CN111667424A (en) Unsupervised real image denoising method
CN102231204A (en) Sequence image self-adaptive regular super resolution reconstruction method
CN104008539A (en) Image super-resolution rebuilding method based on multiscale geometric analysis
CN107133923A (en) A kind of blurred picture non-blind deblurring method based on self-adaption gradient sparse model
Gao et al. Bayesian image super-resolution with deep modeling of image statistics
CN105488759B (en) An Image Super-Resolution Reconstruction Method Based on Local Regression Model
CN115496663A (en) Video super-resolution reconstruction method based on fusion network within D3D convolutional group
CN115829834A (en) Image super-resolution reconstruction method based on half-coupling depth convolution dictionary learning
CN114299185A (en) Magnetic resonance image generation method, magnetic resonance image generation device, computer equipment and storage medium
CN111008930B (en) Fabric image super-resolution reconstruction method
CN111860449B (en) Multispectral and hyperspectral image fusion method based on deep learning and model driving
CN110648291B (en) Unmanned aerial vehicle motion blurred image restoration method based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant