CN113920018A - Improved MSCNN underwater image defogging method - Google Patents
Improved MSCNN underwater image defogging method Download PDFInfo
- Publication number
- CN113920018A CN113920018A CN202111066603.8A CN202111066603A CN113920018A CN 113920018 A CN113920018 A CN 113920018A CN 202111066603 A CN202111066603 A CN 202111066603A CN 113920018 A CN113920018 A CN 113920018A
- Authority
- CN
- China
- Prior art keywords
- image
- mscnn
- algorithm according
- defogging
- underwater
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 42
- 238000005286 illumination Methods 0.000 claims abstract description 26
- 238000012545 processing Methods 0.000 claims abstract description 21
- 238000004422 calculation algorithm Methods 0.000 claims description 40
- 238000002834 transmittance Methods 0.000 claims description 22
- 230000004927 fusion Effects 0.000 claims description 17
- 238000012549 training Methods 0.000 claims description 13
- 238000007781 pre-processing Methods 0.000 claims description 9
- 230000002194 synthesizing effect Effects 0.000 claims description 6
- 238000010521 absorption reaction Methods 0.000 claims description 5
- 238000010606 normalization Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 12
- 238000011156 evaluation Methods 0.000 abstract description 5
- 230000008569 process Effects 0.000 abstract description 4
- 125000001475 halogen functional group Chemical group 0.000 abstract description 3
- 238000012937 correction Methods 0.000 abstract description 2
- 238000013527 convolutional neural network Methods 0.000 description 9
- 230000006870 function Effects 0.000 description 7
- 238000011176 pooling Methods 0.000 description 7
- 230000005540 biological transmission Effects 0.000 description 6
- 238000003384 imaging method Methods 0.000 description 6
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 6
- 230000008859 change Effects 0.000 description 5
- 238000013135 deep learning Methods 0.000 description 4
- 238000010801 machine learning Methods 0.000 description 4
- 238000011161 development Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000002474 experimental method Methods 0.000 description 3
- 239000002245 particle Substances 0.000 description 3
- 238000011084 recovery Methods 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 241000282414 Homo sapiens Species 0.000 description 2
- 230000002238 attenuated effect Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000013441 quality evaluation Methods 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 238000007637 random forest analysis Methods 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 238000001308 synthesis method Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/73—Deblurring; Sharpening
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/40—Image enhancement or restoration using histogram techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30181—Earth observation
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Processing (AREA)
Abstract
The invention has proposed a modified MSCNN underwater image defogging method, carry on the illumination based on processing uniform picture of Retinex and CLAHE to the fogging picture first, carry on pretreatment such as white balance correction, contrast enhancement, etc. to the image after processing, then input the image into MSCNN model trained to estimate out the transmissivity of the image, use the method of dark channel prior to get the background light intensity at the same time, get the image after defogging finally; the method combines Retinex and CLAHE to balance the brightness and improve the contrast of the underwater image, so that the method has more advantages aiming at the complex conditions of low illumination, uneven illumination, obvious Rayleigh scattering phenomenon and the like of the underwater environment, and the problems that in the prior art, halo and overexposure are easy to generate in the process of restoring the underwater image with uneven illumination are solved; compared with the existing method, the method has the advantages that the defogging effect of the underwater image is greatly improved, and various objective evaluation indexes are improved.
Description
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to an improved MSCNN underwater image defogging method.
Background
The underwater image is an important carrier and a presentation mode of underwater information, and has important effects on exploration, development and utilization of ocean resources. However, due to the limitations of objective imaging environment and equipment, the quality of underwater images is always poor, and the underwater images have degradation phenomena such as low contrast, blurred details, color deviation and the like, thereby seriously restricting the development of related fields. Therefore, the interest of students is increasing how to enhance and restore the degraded underwater image through the later algorithm. In recent years, with the rapid development of deep learning technology, the underwater image enhancement and restoration technology based on deep learning has made great progress.
In recent years, in the hot tide of machine learning, many algorithms for image defogging using a machine learning technique emerge. Tang et al systematically studied different fog-related features in a learning framework to determine the best combination of image defogging features, and recovered clear images by using random forests to regress atmospheric transmittance of foggy images from a variety of prior features. The use of synthesis methods to generate training data proposed by Levin et al is also possible with many of the latter machine learning-based methods. Zhu et al propose to use color attenuation priors to defogg from an input foggy image, to model the scene depth of the blurred image under the new priors by establishing a linear model, and to learn the parameters of the model by using a supervised learning method, which can recover the depth information well. The image algorithm based on machine learning improves a great effect on the basis of the traditional algorithm, but compared with image reconstruction tasks such as image super-resolution and the like, single image defogging has great progress space on subjective evaluation and objective evaluation, and many problems still exist and are not optimized and solved.
It can be seen from the atmospheric scattering model that the key of image defogging under the active light illumination condition is to correctly solve the atmospheric light and transmittance values, and image defogging by using a deep learning method can achieve a relatively ideal effect, but because the training data set with fog of the method is artificially synthesized, the model fitted by the network has a certain difference with a real scene, and the loss function adopts a mean square error form, which means that the network can only reduce the difference of the pixel values of the whole image, and the similarity of the defogged image and the whole area of the fogged image cannot be ensured, so that the deep learning-based method has the defects that training is difficult, and the defogging effect depends heavily on the quality of the training data set, and meanwhile, the traditional method such as the Retinex method cannot obtain an effective processing effect on the image with a large-area sky area, and the estimated value of the transmittance is low.
Disclosure of Invention
In view of this, the present invention provides an improved method for defogging an underwater image of MSCNN, which can effectively improve the defogging effect.
An image defogging algorithm comprising the steps of:
step 2, extracting three channels of R, G and B of the image, processing the three channels of image by using an MSRCR algorithm in a Retinex algorithm respectively, and then synthesizing the processed three channels of image into an image;
step 3, performing weighted fusion on the images processed in the step 1 and the step 2 to obtain a fused image, and setting the fused image as an image D;
step 4, preprocessing the image D;
step 5, training the MSCNN network, and then importing the processed image D into a trained MSCNN model to estimate the transmittance t (x) of the MSCNN model;
step 6, estimating the maximum illumination value of the image D by using a dark channel prior method to obtain a background light value C;
and 7, substituting the formula I (x) ═ J (x) t (x) + C [1-t (x) ], and recovering the fog-free image I (x).
Preferably, in the MSCNN network training, a CNN network capable of estimating the depth of field d (x) of an image is used, the transmittance of an input clear image e is estimated by using the formula t (x) ═ e- β d (x) (where β is an atmospheric absorption coefficient), then the model randomly generates a background light value a, a fogging image is synthesized by using the formula i (x) ═ j (x) t (x) + a [1-t (x) ], and the synthesized fogging image is sent to the MSCNN network for training.
Preferably, in step 1, an 8 × 8 rectangular block is used to divide the image into a plurality of different sub-regions,
preferably, in the step 1, the Clip-Limit is 0.01.
Preferably, in step 2, the scales adopted for the R, G, and B three-channel images during the fusion are gaussian functions of σ ═ 20, σ ═ 100, and σ ═ 200, respectively.
Preferably, in step 3, by adjusting the weighted fusion coefficient, the formula is as follows:
D=tA+(1-t)B
where D represents the last fused image, a represents the image after CLAHE processing, B represents the image after MSRCR processing, and t is a fusion weighting coefficient of 0.95.
Preferably, in the step 4, the preprocessing includes white balance.
Preferably, in the step 4, the preprocessing includes contrast enhancement.
Preferably, in the step 4, the preprocessing includes gamma enhancement.
Preferably, in the step 4, the gamma value used in the gamma enhancement is 0.9.
The invention has the following beneficial effects:
the invention has proposed a modified MSCNN underwater image defogging method, carry on the illumination based on processing uniform picture of Retinex and CLAHE to the fogging picture first, carry on pretreatment such as white balance correction, contrast enhancement, etc. to the image after processing, then input the image into MSCNN model trained to estimate out the transmissivity of the image, use the method of dark channel prior to get the background light intensity at the same time, get the image after defogging finally; the method combines Retinex and CLAHE to carry out brightness balance and contrast improvement on the underwater image, so that the method has more advantages aiming at the complex conditions of low illumination, uneven illumination, obvious Rayleigh scattering phenomenon and the like of the underwater environment, and solves the problems that halos and overexposure are easy to generate in the restoration process of the underwater image with uneven illumination in the prior art; compared with the existing method, the method has the advantages that the defogging effect of the underwater image is greatly improved, and various objective evaluation indexes are improved.
Drawings
FIG. 1 is a general operational flow diagram of the present invention.
Fig. 2 is a flow chart of the MSCNN defogging algorithm used in the present invention.
FIG. 3(a) is an original underwater foggy image; fig. 3(b) is a defogged image obtained by using the defogging method according to the present invention.
Detailed Description
The invention is described in detail below by way of example with reference to the accompanying drawings.
1. Principle of underwater imaging
In an underwater environment, an image obtained by an imaging system mainly comprises two parts, namely target reflected light attenuated by water body particle absorption and scattering, and background light formed by water body particle scattering, wherein the mathematical expression of the background light is as follows:
Itotal(x,y)=D(x,y)+B(x,y) (1)
B(x,y)=∫ΘB(Θ)approxdΘ=B∞[1-t(x,y)] (3)
wherein ItotalObtaining an original image which is an underwater blurred image for an imaging system; d (x, y) is the light intensity of the light reflected by the surface of the scene and reaching an imaging system after being attenuated by a water body and subjected to forward scattering; b (x, y) is the underwater background light intensity resulting from the underwater particle backscattering.
The expression for D (x, y) is:
whereinFor effective radiation information of underwater target, t (x, y) is e-β(x,y)ρ(x,y)For the transmission rate of light under water, ρ (x, y) is the relative distance between the light source and the imaging system, and β (x, y) is the total attenuation coefficient of the water body for the absorption and scattering effects of the transmitted light.
The expression for B (x, y) is:
B(x,y)=∫ΘB(Θ)approxdΘ=B∞[1-t(x,y)] (5)
wherein B (theta)approxFor the backscatter volume element function, Θ represents the set of all possible angles of scatter, B, that can occur for a given water volume element∞Background light intensity at infinity underwater.
2. Nonuniform illumination picture enhancement technology based on Retinex and CLAHE
1) Multi-scale Retinex algorithm with color recovery
The Retinex theory was first proposed by the american physicist Edwin Land and is based on the color constancy theory. Color constancy means that human judgment of the color of an object does not change due to a change in illumination conditions, and shows constancy. Retinex theory considers that the color of an object perceived by human beings is not influenced by the external illumination condition, and is only related to the reflection property of the object. Therefore, the Retinex theory can be used for removing the color offset of the underwater blurred image and realizing the color fidelity effect. Retinex theory considers an image S (x, y) as the product of an illumination component L (x, y) determined by environmental factors and a reflection component R (x, y) determined by the intrinsic properties of the object, as follows:
S(x,y)=L(x,y)·R(x,y) (6)
the illumination component L (x, y) determines the dynamic range of the image S (x, y), and the reflection component L (x, y) reflects the reflective properties of the object itself. The essence of the Retinex algorithm is to take the image as the sum of the illumination component and the reflection component, remove the illumination component, eliminate the influence of the external conditions on the image, and obtain the reflection component, i.e. the original attribute of the object.
The principle of the multi-scale Retinex algorithm for color recovery is as follows:
wherein, ak(x, y) represents a nonlinear adjustment coefficient, N represents the number of bands, Ik(x, y) denotes an image of the k-th band, and β denotes a gain adjustment coefficient. It can be seen that compared with the general multi-scale Retinex algorithm, the multi-scale Retinex algorithm for color recovery adds the adjustment factor a on the basis of the multi-scale Retinex algorithmk(x, y). Adjustment factor akThe function of (x, y) is to adjust the R, G, B components of the image enhanced by the multi-scale Retinex algorithm correspondingly according to the ratio of the R, G, B components of the original image, so as to avoid color distortion.
2) Equalization of CLAHE histograms
The histogram distribution of the low-illumination underwater image is narrow and generally low, and the CLAHE can change the distribution of the histogram again to enable the pixel level distribution to be within 0 to 255, so that the pixel level distribution is uniform, the color cast is removed to a certain extent, the contrast is improved, and more details are shown. The CLAHE method comprises the following concrete implementation steps:
the lost person image is divided into M × N non-repeating subregions.
b) The gray level histogram of each sub-region is equally distributed to pixels with the same calculation result, namely the average pixel is
Wherein N isgrayTotal pixel representing the gray level in the sub-area, NxRepresenting the number of pixels on the X-axis of each small area, NyThe number of pixels per small area on the Y-axis is indicated.
c) Make each small region histogram larger thanK represents a clipping coefficient, and the total pixel of the clipped part is calculated as ∑ NvThen, the number of pixels to which the truncated total number of images is assigned per gray level is found as:
d) and (4) after the steps (a), (b) and (c) are repeated for each subregion, histogram equalization is carried out on the obtained new histogram after each subregion is cut, and a transformation function is used for obtaining a new gray value.
3) Novel idea based on CLARE and multi-scale Retinex fusion
In order to solve the problem of underwater images affected by uneven illumination, the invention considers the defects of Retinex algorithm, namely, the color of the color picture is whitened and distorted due to the fact that the image with large illumination change has halo phenomenon. And the CLAHE algorithm can enhance the local contrast of the image by histogram amplitude limiting to reduce the illumination change.
Generally, the fused image needs to increase the amount of detail information as much as possible while retaining the advantages of the original image. The data level fusion is the basis of high-level image fusion, and directly processes data acquired by a sensor under a plurality of scenes, and reserves the original scene as much as possible, provides texture information which cannot be provided by other layer systems, and is also a common method for researching the current fusion image, such as a Weighted average method, a wavelet transform method and the like, wherein the Weighted average method (WA) can directly fuse a plurality of image information with different qualities, has the advantages of simplicity, intuition and high operation speed, meanwhile, the signal-to-noise ratio of the image can be improved, so that the invention selects and adopts a pixel weighted average method to fuse the images respectively processed by CLARE and the multi-scale Retinex algorithm.
3. Image defogging algorithm based on multi-scale convolutional network MSCNN
The performance of existing image defogging methods is limited by manually extracted features such as dark channels, color differences and maximum contrast, and complex fusion schemes. Ren et al trained the MSCNN network by learning the mapping synthesis data between the foggy image and its corresponding transmittance image, and the trained network could directly estimate its corresponding transmittance from the foggy image, and in the defogging stage, directly use the trained transmittance estimation model to obtain the foggy image transmittance, and then use the formula i (x) ═ j (x) t (x) + a [1-t (x) ]) to solve the image to be restored. For networks that estimate transmission, the network designed by Ren et al is divided into two parts, a coarse transmission estimation network and a refinement transmission network, where the input of the second convolutional layer of the refinement network combines the output of the coarse estimation network and both sub-networks are trained separately. Experiments have shown that MSCNN performs better than the most advanced methods for both synthetic and real-world images in terms of quality and speed.
1) Coarse scale convolutional neural network
The task of the coarse-scale network is to estimate the overall transmittance map of the scene. The coarse scale network comprises four operations, convolution, max pooling, upsampling, and linear combining.
Convolutional layer the network takes the RGB image as input. The response of each convolutional layer is given by the following equation.
WhereinAndcharacteristic diagrams of the l layer and the l +1 layer are respectively. In addition, k is a convolution kernel, and the index (m, n) represents the mapping from the current m-th layer to the n-th layer, and x represents the convolution operator. The function σ (-) represents the modified linear unit (ReLU) of the filter response, and b is the offset.
Maximum pooling layer each convolutional layer is followed by a maximum boosting layer with a sampling factor of 2.
And (4) upsampling, wherein the size of the ground real transmission rate graph is the same as the input in the whole frame. While the size of the feature map is reduced by half after the maximum pooling. Therefore, an upsampling layer is added to ensure that the output transmission rate map and the input hazy image are equal in size. Although the deletion of the max pooling layer and the upsampling layer may be chosen to achieve the same goal, this approach will reduce the non-linearity of the network, which is less efficient. The upsampling layer follows the pooling layer to recover the sub-sampled feature map size while preserving network non-linearity. The response of each upsampling layer may be defined as
Where l denotes the layer number and (x, y) denotes the two-dimensional spatial coordinates of the image, the essence of this equation is to copy the pixel value at (x, y) into its corresponding 2 x 2 region. Whereas the linear combination layer is essentially the result of a convolution kernel calculation using a spatial dimension of 1 x 1 for all feature maps of the penultimate layer.
Structurally, the network is an end-to-end regression network for estimating fog map transmittance, and the output of the rough estimation network is used as an important feature map to be combined into a fine estimation network, so that the multi-scale (multi scale) characteristic of the network is fully embodied. However, due to the connection mode of the upper sampling layer, the local receptive field of each layer in the network is small, so that the method has certain limitation on learning the global characteristics of the fog image, and in addition, the estimation effect of the network on the fog image transmissivity is also influenced to a certain extent by using the synthetic data based on the indoor image as training data.
2) Fine scale convolutional neural network:
the task of the fine-scale convolutional neural network is to refine the coarse transmittance map output by the coarse-scale convolutional neural network, and a more accurate transmittance map can be obtained by inputting the foggy image and the coarse transmittance map into the fine-scale convolutional neural network. The fine-scale convolutional neural network model comprises convolution layers with 3 different scales, the sizes of convolution kernels are 7 multiplied by 7,5 multiplied by 5 and 3 multiplied by 3 respectively, and the sizes of the pooling kernels in the pooling layers are consistent with those in the coarse-scale convolutional network and are 2 multiplied by 2 respectively. And finally, refining the fine-scale convolutional neural network, wherein the output of the model is a refined transmittance graph.
The estimation of the atmospheric illumination in the MSCNN algorithm is based on the dark channel prior theory. After the thinned transmittance graph is obtained, selecting darker pixel points in the transmittance graph, and solving the intensity of the pixel points in the original fog image corresponding to the darker pixel points, wherein the value with the maximum intensity is the atmospheric illumination value. And substituting the output transmittance graph of the MSCNN model and the solved atmospheric illumination value into an atmospheric scattering model formula to obtain a defogged image.
In order to integrate the advantages of the algorithm based on multi-scale Retinex and CLAHE and the MSCNN algorithm, the invention provides an improved MSCNN image defogging algorithm.
As shown in fig. 1, the specific process is as follows:
And 2, extracting three channels of R, G and B of the image, processing the three-channel image by using an MSRCR algorithm in a Retinex algorithm respectively, and then synthesizing the processed three-channel image into an image. In the fusion, the corresponding adopted scales of the R, G and B three-channel images are 20, 100 and 200 gaussian functions respectively.
And 3) carrying out weighted fusion on the images processed in the steps 1) and 2). And a better visual effect is obtained by adjusting the weighting fusion coefficient. The formula is as follows:
D=tA+(1-t)B (13)
wherein D represents the final fused image, A represents the image after CLAHE processing, B represents the image after MSRCR processing, and t is a fusion weighting coefficient, and the optimal fusion coefficient t is 0.95 through a large number of experiments.
And 4, preprocessing the processed image D such as white balance, contrast enhancement, gamma enhancement and the like, wherein the gamma value is determined to be 0.9 through repeated experiments.
And 5, as shown in fig. 2, training the MSCNN network, and then importing the processed image D into the trained MSCNN model to estimate the transmittance t (x) of the image D.
When the MSCNN network is trained, firstly, a CNN network capable of estimating the depth of field d (x) of an image is used, the transmittance of an input clear image e is estimated by using a formula t (x) ═ e- β d (x) (where β is an atmospheric absorption coefficient), then a model randomly generates a background light value a, a fogging image is synthesized by using a formula i (x) ═ j (x) t (x) + a [1-t (x) ], and the synthesized fogging image is sent to the MSCNN network for training.
And 6, estimating the maximum illumination value of the image D by using a dark channel prior method to obtain a background light value C.
And 7, substituting the formula I (x) ═ J (x) t (x) + C [1-t (x) ], and recovering the fog-free image I (x).
The effect chart of the defogging method of the present invention is shown in fig. 3, and the evaluation results are shown in table 1:
TABLE 1 Underwater image quality evaluation results
Entropy of information | Mean gradient | UCIQE | Visible edge gradient ratio | |
Original drawing | 7.0339 | 1.1158 | 0.2312 | 1 |
Method for producing a composite material | 7.0870 | 5.5931 | 0.3264 | 5.0402 |
In summary, the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (10)
1. An image defogging algorithm is characterized by comprising the following steps:
step 1, extracting three channels of R, G and B of an image, carrying out normalization processing, respectively processing by using a CLAHE algorithm, and then synthesizing the processed three channels of the image into an image;
step 2, extracting three channels of R, G and B of the image, processing the three-channel image by using an MSRCR algorithm in a Retinex algorithm respectively, and then synthesizing the processed three-channel image into an image;
step 3, performing weighted fusion on the images processed in the step 1 and the step 2 to obtain a fused image, and setting the fused image as an image D;
step 4, preprocessing the image D;
step 5, training the MSCNN network, and then importing the processed image D into a trained MSCNN model to estimate the transmittance t (x) of the MSCNN model;
step 6, estimating the maximum illumination value of the image D by using a dark channel prior method to obtain a background light value C;
and 7, substituting the formula I (x) ═ J (x) t (x) + C [1-t (x) ], and recovering the fog-free image I (x).
2. The image defogging algorithm according to claim 1, wherein in the training of the MSCNN network, firstly using the CNN network capable of estimating the depth of field d (x), the transmittance of the input clear image e is estimated using the formula t (x) -e- β d (x), then the model randomly generates the background light value a, synthesizes the fog image using the formula i (x) -j (x) t (x) + a [1-t (x) ], and then sends the synthesized fog image into the MSCNN network for training, wherein β is the atmospheric absorption coefficient.
3. An image defogging algorithm according to claim 1 or 2 wherein in step 1, an 8 x 8 rectangular block is used to divide the image into a plurality of different subregions.
4. An image defogging algorithm according to claim 1 or 2, wherein in the step 1, the Clip-Limit is 0.01.
5. An image defogging algorithm according to claim 1 or 2, wherein in the step 2, the scales adopted by the R, G and B three-channel images during fusion are 20, 100 and 200 gaussians respectively.
6. The image defogging algorithm according to claim 1 or 2, wherein in the step 3, the fused image is:
D=tA+(1-t)B
where D represents the last fused image, a represents the image after CLAHE processing, B represents the image after MSRCR processing, and t is a fusion weighting coefficient of 0.95.
7. An image defogging algorithm according to claim 1 or 2 wherein in step 4, the preprocessing includes white balancing.
8. The image defogging algorithm according to claim 7, wherein in step 4, the preprocessing comprises contrast enhancement.
9. The image defogging algorithm according to claim 8, wherein in step 4, the preprocessing comprises gamma enhancement.
10. The image defogging algorithm according to claim 8, wherein in step 4, the gamma value adopted in the gamma enhancement is 0.9.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111066603.8A CN113920018A (en) | 2021-09-13 | 2021-09-13 | Improved MSCNN underwater image defogging method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111066603.8A CN113920018A (en) | 2021-09-13 | 2021-09-13 | Improved MSCNN underwater image defogging method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113920018A true CN113920018A (en) | 2022-01-11 |
Family
ID=79234858
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111066603.8A Pending CN113920018A (en) | 2021-09-13 | 2021-09-13 | Improved MSCNN underwater image defogging method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113920018A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114511480A (en) * | 2022-01-25 | 2022-05-17 | 江苏科技大学 | Underwater image enhancement method based on fractional order convolution neural network |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105761227A (en) * | 2016-03-04 | 2016-07-13 | 天津大学 | Underwater image enhancement method based on dark channel prior algorithm and white balance |
CN106530246A (en) * | 2016-10-28 | 2017-03-22 | 大连理工大学 | Image dehazing method and system based on dark channel and non-local prior |
CN106548463A (en) * | 2016-10-28 | 2017-03-29 | 大连理工大学 | Based on dark and the sea fog image automatic defogging method and system of Retinex |
CN112508814A (en) * | 2020-12-07 | 2021-03-16 | 重庆邮电大学 | Image tone restoration type defogging enhancement method based on unmanned aerial vehicle at low altitude view angle |
CN112950589A (en) * | 2021-03-03 | 2021-06-11 | 桂林电子科技大学 | Dark channel prior defogging algorithm of multi-scale convolution neural network |
-
2021
- 2021-09-13 CN CN202111066603.8A patent/CN113920018A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105761227A (en) * | 2016-03-04 | 2016-07-13 | 天津大学 | Underwater image enhancement method based on dark channel prior algorithm and white balance |
CN106530246A (en) * | 2016-10-28 | 2017-03-22 | 大连理工大学 | Image dehazing method and system based on dark channel and non-local prior |
CN106548463A (en) * | 2016-10-28 | 2017-03-29 | 大连理工大学 | Based on dark and the sea fog image automatic defogging method and system of Retinex |
CN112508814A (en) * | 2020-12-07 | 2021-03-16 | 重庆邮电大学 | Image tone restoration type defogging enhancement method based on unmanned aerial vehicle at low altitude view angle |
CN112950589A (en) * | 2021-03-03 | 2021-06-11 | 桂林电子科技大学 | Dark channel prior defogging algorithm of multi-scale convolution neural network |
Non-Patent Citations (1)
Title |
---|
HAOYUE WANG; XIANGNING CHEN; BIJIE XU; SHUHAN DU; YINAN LI: "An improved MSCNN method for underwater image defogging", 2021 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND INDUSTRIAL DESIGN (AIID), 3 September 2021 (2021-09-03) * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114511480A (en) * | 2022-01-25 | 2022-05-17 | 江苏科技大学 | Underwater image enhancement method based on fractional order convolution neural network |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Bai et al. | Underwater image enhancement based on global and local equalization of histogram and dual-image multi-scale fusion | |
Hu et al. | Single image defogging based on illumination decomposition for visual maritime surveillance | |
Wang et al. | Dehazing for images with large sky region | |
CN111292258B (en) | Image defogging method based on dark channel prior and bright channel prior | |
Pan et al. | Underwater image de-scattering and enhancing using dehazenet and HWD | |
CN107798661B (en) | Self-adaptive image enhancement method | |
CN111861896A (en) | UUV-oriented underwater image color compensation and recovery method | |
CN110428389B (en) | Low-light-level image enhancement method based on MSR theory and exposure fusion | |
Pan et al. | De-scattering and edge-enhancement algorithms for underwater image restoration | |
Yao et al. | The Retinex-based image dehazing using a particle swarm optimization method | |
Tang et al. | A local flatness based variational approach to retinex | |
CN107093173A (en) | A kind of method of estimation of image haze concentration | |
Fu et al. | An anisotropic Gaussian filtering model for image de-hazing | |
CN112419163A (en) | Single image weak supervision defogging method based on priori knowledge and deep learning | |
CN117611501A (en) | Low-illumination image enhancement method, device, equipment and readable storage medium | |
Wen et al. | A survey of image dehazing algorithm based on retinex theory | |
Wang et al. | Nighttime image dehazing using color cast removal and dual path multi-scale fusion strategy | |
Hao et al. | Texture enhanced underwater image restoration via laplacian regularization | |
CN110349113B (en) | Adaptive image defogging method based on dark primary color priori improvement | |
Prasath et al. | Application of different techniques for underwater image processing-a systematic review | |
Kansal et al. | Fusion-based image de-fogging using dual tree complex wavelet transform | |
CN113920018A (en) | Improved MSCNN underwater image defogging method | |
CN109544470A (en) | A kind of convolutional neural networks single image to the fog method of boundary constraint | |
Liu et al. | Single image defogging method based on image patch decomposition and multi-exposure image fusion | |
Mi et al. | Multi-scale gradient domain underwater image enhancement |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |