[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
A Blockchain-Based Trust Model for the Internet of Things Supply Chain Management
Previous Article in Journal
D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Saliency-Guided Nonsubsampled Shearlet Transform for Multisource Remote Sensing Image Fusion

Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(5), 1756; https://doi.org/10.3390/s21051756
Submission received: 20 January 2021 / Revised: 26 February 2021 / Accepted: 26 February 2021 / Published: 4 March 2021
(This article belongs to the Section Remote Sensors)
Figure 1
<p>The nonsubsampled shearlet transform (NSST) decomposition of a zoneplate image. (<b>a</b>) original zoneplate image, (<b>b</b>) the low-frequency component, (<b>c</b>) the high-frequency sub-bands of NSST decomposition at level 1, (<b>d</b>) the high-frequency sub-bands of NSST decomposition at level 2, (<b>e</b>) the high-frequency sub-bands of NSST decomposition at level 3.</p> ">
Figure 2
<p>The flow chart of the proposed remote sensing image fusion method.</p> ">
Figure 3
<p>Multisource remote sensing image data sets.</p> ">
Figure 4
<p>Fusion results of the first group of images. (<b>a</b>) Source A, (<b>b</b>) Source B, (<b>c</b>) guided image filter (GFF), (<b>d</b>) image matting for fusion (IFM), (<b>e</b>) dual-tree complex wavelet transform (DTCWT), (<b>f</b>) curvelet transform-based image fusion (CVT), (<b>g</b>) phase congruency (PC), (<b>h</b>) structure-aware image fusion (SAIF), (<b>i</b>) different resolutions via total variation (DRTV), (<b>j</b>) multimodal image seamless fusion (MISF), (<b>k</b>) nonsubsampled shearlet transform (NSST), (<b>l</b>) proposed method.</p> ">
Figure 5
<p>Fusion results of the second group of images. (<b>a</b>) Source A, (<b>b</b>) Source B, (<b>c</b>) GFF, (<b>d</b>) IFM, (<b>e</b>) DTCWT, (<b>f</b>) CVT, (<b>g</b>) PC, (<b>h</b>) SAIF, (<b>i</b>) DRTV, (<b>j</b>) MISF, (<b>k</b>) NSST, (<b>l</b>) proposed method.</p> ">
Figure 5 Cont.
<p>Fusion results of the second group of images. (<b>a</b>) Source A, (<b>b</b>) Source B, (<b>c</b>) GFF, (<b>d</b>) IFM, (<b>e</b>) DTCWT, (<b>f</b>) CVT, (<b>g</b>) PC, (<b>h</b>) SAIF, (<b>i</b>) DRTV, (<b>j</b>) MISF, (<b>k</b>) NSST, (<b>l</b>) proposed method.</p> ">
Figure 6
<p>Fusion results of the third group of images. (<b>a</b>) Source A, (<b>b</b>) Source B, (<b>c</b>) GFF, (<b>d</b>) IFM, (<b>e</b>) DTCWT, (<b>f</b>) CVT, (<b>g</b>) PC, (<b>h</b>) SAIF, (<b>i</b>) DRTV, (<b>j</b>) MISF, (<b>k</b>) NSST, (<b>l</b>) proposed method.</p> ">
Figure 6 Cont.
<p>Fusion results of the third group of images. (<b>a</b>) Source A, (<b>b</b>) Source B, (<b>c</b>) GFF, (<b>d</b>) IFM, (<b>e</b>) DTCWT, (<b>f</b>) CVT, (<b>g</b>) PC, (<b>h</b>) SAIF, (<b>i</b>) DRTV, (<b>j</b>) MISF, (<b>k</b>) NSST, (<b>l</b>) proposed method.</p> ">
Figure 7
<p>Fusion results of the fourth group of images. (<b>a</b>) Source A, (<b>b</b>) Source B, (<b>c</b>) GFF, (<b>d</b>) IFM, (<b>e</b>) DTCWT, (<b>f</b>) CVT, (<b>g</b>) PC, (<b>h</b>) SAIF, (<b>i</b>) DRTV, (<b>j</b>) MISF, (<b>k</b>) NSST, (<b>l</b>) proposed method.</p> ">
Figure 8
<p>The line chart of objective metric data in <a href="#sensors-21-01756-t005" class="html-table">Table 5</a>. (<b>a</b>) VIFF; (<b>b</b>) Q<sub>S</sub>; (<b>c</b>) AG; (<b>d</b>) CC; (<b>e</b>) SF; (<b>f</b>) Q<sub>W</sub>.</p> ">
Figure 8 Cont.
<p>The line chart of objective metric data in <a href="#sensors-21-01756-t005" class="html-table">Table 5</a>. (<b>a</b>) VIFF; (<b>b</b>) Q<sub>S</sub>; (<b>c</b>) AG; (<b>d</b>) CC; (<b>e</b>) SF; (<b>f</b>) Q<sub>W</sub>.</p> ">
Versions Notes

Abstract

:
The rapid development of remote sensing and space technology provides multisource remote sensing image data for earth observation in the same area. Information provided by these images, however, is often complementary and cooperative, and multisource image fusion is still challenging. This paper proposes a novel multisource remote sensing image fusion algorithm. It integrates the contrast saliency map (CSM) and the sum-modified-Laplacian (SML) in the nonsubsampled shearlet transform (NSST) domain. The NSST is utilized to decompose the source images into low-frequency sub-bands and high-frequency sub-bands. Low-frequency sub-bands reflect the contrast and brightness of the source images, while high-frequency sub-bands reflect the texture and details of the source images. Using this information, the contrast saliency map and SML fusion rules are introduced into the corresponding sub-bands. Finally, the inverse NSST reconstructs the fusion image. Experimental results demonstrate that the proposed multisource remote image fusion technique performs well in terms of contrast enhancement and detail preservation.

1. Introduction

Remote sensing images play an important role in urban planning, environmental monitoring, and military defense [1]. As a basic step of target classification, detection, and recognition in remote sensing images, remote sensing image fusion has attracted more and more research interest across the world. Due to the incident wavelengths of the remote sensing images in the same region being different, multiband remote sensing images have significant differences. The high-band remote sensing image can provide an overall view of the scene, which is similar to optical imaging, while the low-band remote sensing image is relatively bleak and has deeper penetration. Remote sensing image fusion can integrate multiband remote sensing images into a comprehensive image, which is conducive to the recognition and observation of ground objects [1,2,3].
Multisource remote sensing image fusion is an information processing technology for the fusion of multisensor, multiplatform remote sensing and multispectral band remote sensing data. The fusion image contains different spatial, temporal, and spectral information of multisensor, which allows for preparation for further analysis and processing. Many image fusion methods have been proposed in recent decades; however, image fusion algorithms based on transform domain and edge-preserving filters are widely used [4]. In terms of transform domain-based image fusion frameworks, the wavelet transform, discrete wavelet transform (DWT) [5], dual-tree complex wavelet transform (DTCWT) [5], dual-tree complex wavelet package transform (DTCWPT) [6], framelet transform [7], curvelet transform [5], contourlet transform [8], nonsubsampled contourlet transform (NSCT) [9], shearlet transform [10], and nonsubsampled shearlet transform (NSST) [11], etc., are adapted to the field of image fusion. Iqbal et al. [12] introduced a multifocus image fusion approach using a DWT and a guided image filter to improve the definition of the fused images. Aishwarya et al. [13] used a DTCWT and an adaptive combined clustered dictionary for visible and infrared image fusion to enhance the target information. Wang et al. [14] proposed a multispectral (MS) and panchromatic (PAN) image fusion technique based on the hidden Markov tree model in a complex tight framelet transform domain to improve the spatial resolution of the MS image while keeping the spectral information. Due to the fact that the wavelet transform cannot capture the abundant directional information of remote sensing images and can introduce spatial distortion, a contourlet transform and an NSCT are introduced to resolve this shortcoming. Yang et al. [15] proposed a remote sensing image fusion algorithm via a contourlet hidden Markov tree and a clarity–saliency-driven pulse couple neural network (PCNN) model to enhance the edges and contours of fused remote sensing images. Li et al. [16] introduced an image fusion method using dynamic threshold neural P systems and NSCT for multimodality medical imaging to improve the visual quality and fusion performance. Because the contourlet transform- and NSCT-based image fusion approaches are computationally complex, the shearlet transform and the NSST are proposed to increase computational efficiency. Because the shearlet transform lacks translation invariance, the NSST has become more widely used as the improved version of the shearlet transform in the field of image processing. Yin et al. [17] proposed an image fusion technique via NSST and parameter-adaptive pulse coupled neural network (PAPCNN) to improve the contrast and brightness of the fused medical images. Wang et al. [18] introduced the nonsubsampled shearlet transform hidden Markov forest (NSST-HMF) model for pansharpening to improve the spatial resolution of hyperspectral images while preserving spectral features.
In terms of edge preserving filter-based image fusion approaches, the guided image filter, cross bilateral filter, and rolling guidance filter, etc., are widely used. Li et al. [19] first introduced the guided image filter for image fusion, for which the computational complexity is relatively low. Then, the combination of guided image filtering and other transform domain algorithms such as DTCWT, NSCT, and NSST is introduced into the field of image fusion, and good results are achieved. Shreyamsha et al. [20] introduced the cross bilateral filter for image fusion based on pixel significance to enhance the visual quality of the fused images. Jian et al. [21] proposed a multiscale image fusion method using a rolling guidance filter to preserve the details and suppress the artifacts of the fused images.
In this work, a novel remote sensing image fusion algorithm using a contrast saliency map (CSM) and SML in the NSST domain is proposed. The contrast saliency map-based fusion rule and SML-based fusion rule are used to merge the low- and high-frequency sub-bands, respectively. Experimental results demonstrate the effectiveness of the proposed remote sensing image fusion method over the traditional and state-of-the-art fusion algorithms in terms of qualitative and quantitative analysis.
The rest of this work is organized as follows: Section 2 shows the related works, Section 3 depicts the proposed remote sensing image fusion method, the experiments and results are summarized in Section 4, and the conclusions are provided in Section 5.

2. Related Works

Nonsubsampled Shearlet Transform

Nonsubsampled shearlet transform (NSST) is a kind of nonsubsampled multiscale transform, which was introduced based on the theory of shearlet transform [11,18]. The image is decomposed by NSST into multiple scales with multiple directions by multiscale and multidirectional decompositions. Firstly, the nonsubsampled pyramid (NSP) is adopted as the multiscale decomposition filter to decompose the image into one low-frequency sub-band and one high-frequency sub-band. Then, the high-frequency sub-band is decomposed by the shearing filter (SF) to achieve the multidirectional sub-bands. Due to the NSST decomposition process having no subsampling for the NSP and the SF, the NSST is shift-invariant. Figure 1 denotes the example of three levels of NSST decomposition of a zoneplate image, where all the images are displayed in the “jet” colormap and the direction numbers from coarser to finer are 4, 8, and 8. Figure 1a depicts the original zoneplate image, Figure 1b shows the low-frequency component, and Figure 1c–e show the high-frequency sub-band images with the direction numbers 4, 8, and 8, respectively.

3. Proposed Fusion Method

In this section, a novel remote sensing image fusion method based on the NSST is proposed, and the whole process can be divided into four parts: NSST decomposition, low-frequency sub-band fusion, high-frequency sub-band fusion, and inverse NSST image reconstruction. Suppose the input remote sensing images are A and B, then the two images are decomposed up to N levels utilizing the NSST to generate the decomposed sub-bands { L A , H A l , d } and { L B , H B l , d } , respectively. The H X l , d | X { A , B } represents the high-frequency sub-bands of X achieved at the lth decomposition with the direction d, the L X | X { A , B } represents the low-frequency sub-band of X, where l [ 1 , N ] , d [ 1 , D ( l ) ] , N equals the number of NSST decomposition levels, and D denotes the vector which concludes the number of directions at each l. The fused image F is generated by inverse NSST performed on the fused sub-bands { L F , H F l , d } . The flow chart of the proposed remote sensing image fusion approach is shown in Figure 2. The fusion rules for low-frequency and high-frequency components are summarized as follows.

3.1. Fusion of Low-Frequency Components

The low-frequency sub-bands present the brightness and contrast information of the source remote sensing images [22]. In this section, in order to preserve the contrast, the contrast saliency maps (CSM) of the low-frequency components are constructed based on the brightness distribution. The contrast of the image denotes the difference between the lowest and highest brightness levels in the remote sensing images, and where the difference in brightness is more significant, a higher contrast is implied. Therefore, we can infer that the brighter or darker the pixel value is relative to the average value of the image, the greater its contribution to the image contrast and the stronger the contrast significance. The L2 norm is used to judge the deviation between pixel value and average value, and the significance of each pixel is expressed. When the L2 norm is performed on the low-frequency sub-bands LA and LB, the contrast saliency maps S L A and S L B of the low-frequency sub-bands are generated by the following:
S L A = n o r m ( L A m e a n ( L A ) 2 )
S L B = n o r m ( L B m e a n ( L B ) 2 )
where the m e a n ( ) denotes the average value of the image. L2 norm is used to eliminate the effect of symbols, and the n o r m ( ) function is defined as follows:
n o r m ( x ) = x m i n ( x ) m a x ( x ) m i n ( x )
The weight matrices W L A and W L B of the low-frequency components are calculated by the following formulas performed on the saliency maps of the low-frequency sub-bands:
W L A = 0.5 + 0.5 ( S L A S L B )
W L B = 0.5 + 0.5 ( S L B S L A )
The fused low-frequency sub-bands are computed by the Hadamard product performed on the low-frequency components and the corresponding weight matrices, and the corresponding equation is defined as follows:
L F = W L A L A + W L B L B
where L F represents the fused low-frequency component, and * shows the Hadamard product.

3.2. Fusion of High-Frequency Components

The high-frequency components contain the texture information and details. In this section, the sum-modified-Laplacian (SML) is used to process the high-frequency sub-bands. The SML is defined for the local window with the size ( 2 P + 1 ) ( 2 Q + 1 ) , and the corresponding formula is calculated by [23]:
S M L l , d ( i , j ) = p = P P q = Q Q [ M L l , d ( i + p , j + q ) ] 2
M L l , d ( i , j ) = | 2 H l , d ( i , j ) H l , d ( i s t e p , j ) H l , d ( i + s t e p , j ) | + | 2 H l , d ( i , j ) H l , d ( i , j s t e p ) H l , d ( i , j + s t e p ) |
where step denotes the changeable interval among the high-frequency coefficients. It is usually defined as 1.
The fused high-frequency sub-bands can be computed by:
H F l , d ( i , j ) = { H A l , d ( i , j )   if   S M L A l , d ( i , j ) S M L B l , d ( i , j ) H B l , d ( i , j )   if   S M L A l , d ( i , j ) < S M L B l , d ( i , j )
where HF denotes the fused high-frequency components.
The whole procedure of the proposed remote sensing image fusion method can be summarized in Algorithm 1.
Algorithm 1 Remote sensing image fusion via NSST
Input: the source remote sensing images A and B
Output: fused image F
Parameters: the number of NSST decomposition levels—N; the number of directions at each decomposition level— D ( l ) , l [ 1 , N ]
Step 1: NSST decomposition
The input images A and B are decomposed into low- and high-frequency sub-bands { L A , H A l , d } and { L B , H B l , d } , respectively.
Step 2: low-frequency band fusion rule
(1) The saliency maps ( S L A , S L B ) and the corresponding weight matrices ( W L A , W L B ) of the low-frequency bands are calculated by Equations (1)–(5).
(2) The fused low-frequency band LF is obtained by Equation (6).
Step 3: high-frequency band fusion rule
(1) The SML of the high-frequency bands is constructed via Equations (7)–(8).
(2) The fused high-frequency band HF is computed by Equation (9).
Step 4: inverse NSST and image reconstruction
The fused image F is reconstructed by inverse NSST performed on the fused low- and high-frequency bands { L F , H F l , d } .

4. Experimental Results and Discussion

In this section, in order to demonstrate the effectiveness of the proposed multisource remote sensing image fusion method via NSST, public data sets (https://sites.google.com/view/durgaprasadbavirisetti/datasets (accessed on 15 December 2020)) are used for simulation, and several state-of-the-art image fusion algorithms are adapted for comparison, namely image fusion based on a guided image filter (GFF) [19], image matting for the fusion of multifocus image (IFM) [24], image fusion using a dual-tree complex wavelet transform (DTCWT) [5], curvelet transform-based image fusion (CVT) [5], image fusion utilizing phase congruency (PC) [25], structure-aware image fusion (SAIF) [26], fusing infrared and visible images of different resolutions via total variation model (DRTV) [27], multimodal image seamless fusion (MISF) [28], and parameter-adaptive pulse-coupled neural network-based image fusion via a nonsubsampled shearlet transform (NSST) [17]. In order to reflect the fairness of the algorithm, the parameters of the comparison algorithms are consistent with the original published papers. In the proposed fusion technique, the number of NSST decomposition levels is four, and the direction numbers from coarser to finer are 8, 8, 16, and 16. The selected remote sensing image data sets are shown in Figure 3.
In order to objectively assess the fusion performances of all the different fusion techniques, a lot of image fusion evaluation indexes have been introduced in these years. It is known to us that just one evaluation index could not well demonstrate the quality of fused images in quantitative assessment. Thus, for the sake of making a comprehensive evaluation for the fusion images, six popular fusion evaluation metrics are introduced in this section, namely visual information fidelity for fusion (VIFF) [29,30,31,32,33], QS [34], average gradient (AG) [20,35,36], correlation coefficient (CC) [20,37,38], spatial frequency (SF) [20,39,40,41], and QW [34,42]. In terms of all the six metrics, the higher the value data of the evaluation index, the better the fusion performance will be. The experimental results are depicted in Figure 4, Figure 5, Figure 6 and Figure 7 and Table 1, Table 2, Table 3, Table 4 and Table 5.

4.1. Qualitative Analysis

In this section, the fusion results obtained by the proposed method and the compared results calculated by nine other fusion algorithms are given in Figure 4, Figure 5, Figure 6 and Figure 7. The Figure 4, Figure 5, Figure 6 and Figure 7a,b show the source images A and B, respectively. As seen from Figure 4, the GFF, DTCWT, CVT and DRTV algorithms decrease the contrast of the fusion images, making some details invisible (see Figure 4c,e,f,i). The IFM, SAIF, and MISF methods appear to generate a block effect and artifacts, affecting the observation of the fused images (see Figure 4d,h,j). The PC algorithm makes the image distorted (see Figure 4g). The NSST technique provides overly high brightness (see Figure 4k). The proposed fusion technique can provide a high-definition image and preserve spatial detail information in the fused image (see Figure 4l).
From Figure 5, we can see that the GFF, IFM, and DRTV methods make the fused image darker in some regions (see Figure 5c,d,i). The DTCWT and CVT methods make the fused images better compared to the previous methods (see Figure 5e,f). The PC approach provides a poor fusion performance (see Figure 5g). The SAIF and MISF algorithms introduce artifacts (see Figure 5h,j). The NSST method makes the fused image brighter, and it is not conducive to the acquisition of target information from the fused image (see Figure 5k). The proposed fusion method provides a better fusion effect (see Figure 5l).
From Figure 6, it can be seen that the GFF, IFM, DTCWT, and CVT algorithms decrease the contrast and make the images darker (see Figure 6c–f). The PC technique appears to generate a block effect (see Figure 6g). The SAIF, MISF, and NSST methods produce artifacts, and the brightness is over-enhanced in some regions (see Figure 6h,j,k). The DRTV method produces over-enhanced brightness in some regions and an overly smooth fusion image (see Figure 6i). The proposed algorithm can enhance the contrast and definition, which is helpful in obtaining the target information from the fused image (see Figure 6l).
From Figure 7, we can see that the GFF, IFM, SAIF, and MISF algorithms make the fusion image darker (see Figure 7c,d,h,j). The DTCWT and CVT methods produce a good fusion visual effect (see Figure 7e,f). The PC, DRTV, and NSST techniques produce distortion and artifacts (see Figure 7g,i,k). The proposed fusion technique can produce relatively higher contrast and preserve the texture information (see Figure 7l).
In summary, the analysis of the subjective assessment of the fusion results demonstrates the super-performance of the proposed remote sensing image fusion technique when compared with the state-of-the-art fusion algorithms.

4.2. Quantitative Analysis

In this section, the six indexes (VIFF, QS, AG, CC, SF, QW) are used to evaluate the fusion results quantitatively. The data for the evaluation metrics of the different fusion algorithms for Figure 4, Figure 5, Figure 6 and Figure 7 are shown in Table 1, Table 2, Table 3 and Table 4. From Table 1, we can see that the value of VIFF as computed by the proposed method is slightly worse than the NSST algorithm, while the data for the other five metrics as calculated by the proposed fusion technique are the best. From Table 2, we can see that the metric values given by the proposed method are the largest except for the metric of CC. From Table 3, the values of CC and QW as computed by the proposed technique are a little smaller than the corresponding values obtained by the CVT and NSST methods, respectively. From Table 4, we can see that all six values of the metrics achieved by the proposed method are the best compared to the other fusion methods.
In order to demonstrate the effectiveness of the proposed method, the sixteen image groups given in Figure 3 are simulated, and the average values of their objective evaluation are given in Table 5. The line charts of the objective metrics data in Table 5 are given in Figure 8, and the proposed method has the best values in the data for all metrics. Therefore, it is demonstrated that better fusion performance can be generated by the proposed remote sensing image fusion work.

5. Conclusions

In this work, a novel saliency-guided nonsubsampled shearlet transform for multisource remote sensing image fusion is introduced. First, the input images are transformed from the spatial domain to the shearlet domain according to a nonsubsampled shearlet transform. Second, the contrast saliency maps and corresponding weighted matrices are introduced for fusing the low-frequency coefficients, and the SML-based fusion rule is performed on the high-frequency coefficients, which can improve the contrast and definition of the fused images. To prove the universality of the proposed fusion algorithm, sixteen sets of remote sensing images are simulated, and six image fusion evaluation indexes are utilized for the quantitative analysis. From the experimental results, we can conclude that the proposed fusion approach has superior performance compared to the state-of-the-art fusion methods. In future work, we will extend the algorithm to panchromatic and multispectral [43,44,45,46,47,48], hyperspectral and multispectral image fusion [49,50].

Author Contributions

The experimental measurements and data collection were carried out by L.L. and H.M. The manuscript was written by L.L. with the assistance of H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Shanghai Aerospace Science and Technology Innovation Fund under Grant No. SAST2019-048.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, S.; Wang, M. Fusion of multiparametric SAR images based on SW-nonsubsampled contourlet and PCNN. Signal Process. 2009, 89, 2596–2608. [Google Scholar] [CrossRef]
  2. Azarang, A.; Kehtarnavaz, N. Image fusion in remote sensing by multi-objective deep learning. Int. J. Remote Sens. 2020, 41, 9507–9524. [Google Scholar] [CrossRef]
  3. Liao, B.; Liu, W. Multispectral image fusion based on joint sparse subspace recovery. J. Appl. Remote Sens. 2015, 9, 095068. [Google Scholar] [CrossRef]
  4. Liu, Y.; Wang, L. Multi-focus image fusion: A survey of the state of the art. Inf. Fusion 2020, 64, 71–91. [Google Scholar] [CrossRef]
  5. Liu, Y.; Liu, S. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  6. Bayram, I.; Selesnick, I. On the dual-tree complex wavelet packet and m-band transforms. IEEE Trans. Signal Process. 2008, 56, 2298–2310. [Google Scholar] [CrossRef]
  7. Yang, X.; Wang, J. Random walks for synthetic aperture radar image fusion in framelet domain. IEEE Trans. Image Process. 2018, 27, 851–865. [Google Scholar] [CrossRef]
  8. Do, M.; Vetterli, M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Da, C.; Zhou, J. The nonsubsampled contourlet transform: Theory, design, and applications. IEEE Trans. Image Process. 2006, 15, 3089–3101. [Google Scholar]
  10. Guo, K.; Labate, D. Optimally sparse multidimensional representation using shearlets. SIAM J. Math. Anal. 2007, 39, 298–318. [Google Scholar] [CrossRef] [Green Version]
  11. Easley, G.; Labate, D.; Lim, W. Sparse directional image representations using the discrete shearlet transform. Appl. Comput. Harmon. Anal. 2008, 25, 25–46. [Google Scholar] [CrossRef] [Green Version]
  12. Iqbal, M.; Riaz, M.; Iltaf, N. A multifocus image fusion using highlevel DWT components and guided filter. Multimed. Tools Appl. 2020, 79, 12817–12828. [Google Scholar]
  13. Aishwarya, N.; Thangammal, C.B. Visible and infrared image fusion using DTCWT and adaptive combined clustered dictionary. Infrared Phys. Technol. 2018, 93, 300–309. [Google Scholar] [CrossRef]
  14. Wang, J.; Yang, X.; Zhu, R. Random walks for pansharpening in complex tight framelet domain. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5121–5134. [Google Scholar] [CrossRef]
  15. Yang, S.; Wang, M. Contourlet hidden markov tree and clarity-saliency driven PCNN based remote sensing images fusion. Appl. Soft Comput. 2012, 12, 228–237. [Google Scholar] [CrossRef]
  16. Li, B.; Peng, H. A novel fusion method based on dynamic threshold neural P systems and nonsubsampled contourlet transform for multi-modality medical images. Signal Process. 2021, 178, 107793. [Google Scholar] [CrossRef]
  17. Yin, M.; Liu, X.; Liu, Y. Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans. Instrum. Meas. 2019, 68, 49–64. [Google Scholar] [CrossRef]
  18. Wang, X.; Mu, Z.; Song, R. A hyperspectral image NSST-HMF model and its application in HS-pansharpening. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4803–4817. [Google Scholar] [CrossRef]
  19. Li, S.; Kang, X. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  20. Shreyamsha Kumar, B.K. Image fusion based on pixel significance using cross bilateral filter. Signal Image Video Process. 2015, 9, 1193–1204. [Google Scholar] [CrossRef]
  21. Jian, L.; Yang, X.; Zhou, Z. Multi-scale image fusion through rolling guidance filter. Future Gener. Comput. Syst. Int. J. Esci. 2018, 83, 310–325. [Google Scholar] [CrossRef]
  22. Ma, J.; Zhou, Y. Infrared and visible image fusion via gradientlet filter. Comput. Vis. Image Underst. 2020, 197, 103016. [Google Scholar] [CrossRef]
  23. Li, L.; Si, Y.; Wang, L. A novel approach for multi-focus image fusion based on SF-PAPCNN and ISML in NSST domain. Multimed. Tools Appl. 2020, 79, 24303–24328. [Google Scholar] [CrossRef]
  24. Li, S.; Kang, X.; Hu, J. Image matting for fusion of multi-focus images in dynamic scenes. Inf. Fusion 2013, 14, 147–162. [Google Scholar] [CrossRef]
  25. Zhan, K.; Li, Q.; Teng, J. Multifocus image fusion using phase congruency. J. Electron. Imaging 2015, 24, 033014. [Google Scholar] [CrossRef]
  26. Li, W.; Xie, Y. Structure-aware image fusion. Optik 2018, 172, 1–11. [Google Scholar] [CrossRef]
  27. Du, Q.; Xu, H.; Ma, Y. Fusing infrared and visible images of different resolutions via total variation model. Sensors 2018, 18, 3827. [Google Scholar] [CrossRef] [Green Version]
  28. Zhan, K.; Kong, L.; Liu, B. Multimodal image seamless fusion. J. Electron. Imaging 2019, 28, 023027. [Google Scholar] [CrossRef]
  29. Han, Y.; Cai, Y.; Cao, Y. A new image fusion performance metric based on visual information fidelity. Inf. Fusion 2013, 14, 127–135. [Google Scholar] [CrossRef]
  30. Li, L.; Ma, H.; Jia, Z. A novel multiscale transform decomposition based multi-focus image fusion framework. Multimed. Tools Appl. 2021, 80, 1–21. [Google Scholar]
  31. Li, H.; Wu, X.; Durrani, T. NestFuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Trans. Instrum. Meas. 2020, 69, 9645–9656. [Google Scholar] [CrossRef]
  32. Wang, K.; Zheng, M.; Wei, H. Multi-modality medical image fusion using convolutional neural network and contrast pyramid. Sensors 2020, 20, 2169. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Li, L.; Ma, H.; Lv, M. Multimodal medical image fusion via PCNN and WSEML in nonsubsampled contourlet transform domain. J. Med. Imaging Health Inform. 2021, 11, 1–18. [Google Scholar]
  34. Liu, Z.; Blasch, E.; Xue, Z. Objective assessment of multiresolution image fusion algorithms for context enhancement in night vision: A comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 94–109. [Google Scholar] [CrossRef] [PubMed]
  35. Yang, Y.; Zhang, Y.; Huang, S. Infrared and visible image fusion using visual saliency sparse representation and detail injection model. IEEE Trans. Instrum. Meas. 2021, 70, 5001715. [Google Scholar]
  36. Li, L.; Si, Y.; Wang, L. Brain image enhancement approach based on singular value decomposition in nonsubsampled shearlet transform domain. J. Med. Imaging Health Inform. 2020, 10, 1785–1794. [Google Scholar] [CrossRef]
  37. Li, H.; Zhang, L.; Jiang, M. Multi-focus image fusion algorithm based on supervised learning for fully convolutional neural network. Pattern Recognit. Lett. 2021, 141, 45–53. [Google Scholar] [CrossRef]
  38. Li, L.; Wang, L.; Wang, Z. A novel medical image fusion approach based on nonsubsampled shearlet transform. J. Med. Imaging Health Inform. 2019, 9, 1815–1826. [Google Scholar] [CrossRef]
  39. Raudonis, V.; Paulauskaite-Taraseviciene, A.; Sutiene, K. Fast multi-focus fusion based on deep learning for early-stage embryo image enhancement. Sensors 2021, 21, 863. [Google Scholar] [CrossRef]
  40. Subbiah Parvathy, V.; Pothiraj, S.; Sampson, J. A novel approach in multimodality medical image fusion using optimal shearlet and deep learning. Int. J. Imaging Syst. Technol. 2020, 30, 847–859. [Google Scholar] [CrossRef]
  41. Du, J.; Li, W. Two-scale image decomposition based image fusion using structure tensor. Int. J. Imaging Syst. Technol. 2020, 30, 271–284. [Google Scholar] [CrossRef]
  42. Ganasala, P.; Prasad, A.D. Medical image fusion based on laws of texture energy measures in stationary wavelet transform domain. Int. J. Imaging Syst. Technol. 2020, 30, 544–557. [Google Scholar] [CrossRef]
  43. Li, X.; Yan, H.; Xie, W. An improved pulse-coupled neural network model for Pansharpening. Sensors 2020, 20, 2764. [Google Scholar] [CrossRef]
  44. Zhang, H.; Ma, J. GTP-PNet: A residual learning network based on gradient transformation prior for pansharpening. ISPRS J. Photogramm. Remote Sens. 2021, 172, 223–239. [Google Scholar] [CrossRef]
  45. Lee, C.; Oh, J. Rigorous co-registration of KOMPSAT-3 multispectral and panchromatic images for pan-sharpening image fusion. Sensors 2020, 20, 2100. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Saxena, N.; Balasubramanian, R. A pansharpening scheme using spectral graph wavelet transforms and convolutional neural networks. Int. J. Remote Sens. 2021, 42, 2898–2919. [Google Scholar] [CrossRef]
  47. Wu, S.; Chen, H. Smart city oriented remote sensing image fusion methods based on convolution sampling and spatial transformation. Comput. Commun. 2020, 157, 444–450. [Google Scholar] [CrossRef]
  48. Ma, J.; Yu, W.; Chen, C. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  49. Peng, Y.; Li, W.; Luo, X. Integrated fusion framework based on semicoupled sparse tensor factorization for spatio-temporal-spectra fusion of remote sensing images. Inf. Fusion 2021, 65, 21–36. [Google Scholar] [CrossRef]
  50. Li, X.; Yuan, Y.; Wang, Q. Hyperspectral and multispectral image fusion via nonlocal low-rank tensor approximation and sparse representation. IEEE Trans. Geosci. Remote Sens. 2021, 59, 550–562. [Google Scholar] [CrossRef]
Figure 1. The nonsubsampled shearlet transform (NSST) decomposition of a zoneplate image. (a) original zoneplate image, (b) the low-frequency component, (c) the high-frequency sub-bands of NSST decomposition at level 1, (d) the high-frequency sub-bands of NSST decomposition at level 2, (e) the high-frequency sub-bands of NSST decomposition at level 3.
Figure 1. The nonsubsampled shearlet transform (NSST) decomposition of a zoneplate image. (a) original zoneplate image, (b) the low-frequency component, (c) the high-frequency sub-bands of NSST decomposition at level 1, (d) the high-frequency sub-bands of NSST decomposition at level 2, (e) the high-frequency sub-bands of NSST decomposition at level 3.
Sensors 21 01756 g001
Figure 2. The flow chart of the proposed remote sensing image fusion method.
Figure 2. The flow chart of the proposed remote sensing image fusion method.
Sensors 21 01756 g002
Figure 3. Multisource remote sensing image data sets.
Figure 3. Multisource remote sensing image data sets.
Sensors 21 01756 g003
Figure 4. Fusion results of the first group of images. (a) Source A, (b) Source B, (c) guided image filter (GFF), (d) image matting for fusion (IFM), (e) dual-tree complex wavelet transform (DTCWT), (f) curvelet transform-based image fusion (CVT), (g) phase congruency (PC), (h) structure-aware image fusion (SAIF), (i) different resolutions via total variation (DRTV), (j) multimodal image seamless fusion (MISF), (k) nonsubsampled shearlet transform (NSST), (l) proposed method.
Figure 4. Fusion results of the first group of images. (a) Source A, (b) Source B, (c) guided image filter (GFF), (d) image matting for fusion (IFM), (e) dual-tree complex wavelet transform (DTCWT), (f) curvelet transform-based image fusion (CVT), (g) phase congruency (PC), (h) structure-aware image fusion (SAIF), (i) different resolutions via total variation (DRTV), (j) multimodal image seamless fusion (MISF), (k) nonsubsampled shearlet transform (NSST), (l) proposed method.
Sensors 21 01756 g004
Figure 5. Fusion results of the second group of images. (a) Source A, (b) Source B, (c) GFF, (d) IFM, (e) DTCWT, (f) CVT, (g) PC, (h) SAIF, (i) DRTV, (j) MISF, (k) NSST, (l) proposed method.
Figure 5. Fusion results of the second group of images. (a) Source A, (b) Source B, (c) GFF, (d) IFM, (e) DTCWT, (f) CVT, (g) PC, (h) SAIF, (i) DRTV, (j) MISF, (k) NSST, (l) proposed method.
Sensors 21 01756 g005aSensors 21 01756 g005b
Figure 6. Fusion results of the third group of images. (a) Source A, (b) Source B, (c) GFF, (d) IFM, (e) DTCWT, (f) CVT, (g) PC, (h) SAIF, (i) DRTV, (j) MISF, (k) NSST, (l) proposed method.
Figure 6. Fusion results of the third group of images. (a) Source A, (b) Source B, (c) GFF, (d) IFM, (e) DTCWT, (f) CVT, (g) PC, (h) SAIF, (i) DRTV, (j) MISF, (k) NSST, (l) proposed method.
Sensors 21 01756 g006aSensors 21 01756 g006b
Figure 7. Fusion results of the fourth group of images. (a) Source A, (b) Source B, (c) GFF, (d) IFM, (e) DTCWT, (f) CVT, (g) PC, (h) SAIF, (i) DRTV, (j) MISF, (k) NSST, (l) proposed method.
Figure 7. Fusion results of the fourth group of images. (a) Source A, (b) Source B, (c) GFF, (d) IFM, (e) DTCWT, (f) CVT, (g) PC, (h) SAIF, (i) DRTV, (j) MISF, (k) NSST, (l) proposed method.
Sensors 21 01756 g007
Figure 8. The line chart of objective metric data in Table 5. (a) VIFF; (b) QS; (c) AG; (d) CC; (e) SF; (f) QW.
Figure 8. The line chart of objective metric data in Table 5. (a) VIFF; (b) QS; (c) AG; (d) CC; (e) SF; (f) QW.
Sensors 21 01756 g008aSensors 21 01756 g008b
Table 1. The objective evaluation of the methods in Figure 4.
Table 1. The objective evaluation of the methods in Figure 4.
VIFFQSAGCCSFQW
GFF0.40570.80648.79030.749314.45900.8079
IFM0.28710.71749.50610.683415.67300.7091
DTCWT0.53800.814010.03840.781615.77870.8214
CVT0.55340.798410.23970.777115.58990.8165
PC0.42460.74779.14940.666814.67790.6555
SAIF0.56620.80389.38840.679815.20250.8261
DRTV0.28950.73167.80060.717611.66890.6561
MISF0.52260.80519.13650.657514.91360.8142
NSST0.61580.821810.07660.727215.45830.8304
Proposed0.61300.843810.45920.789316.21490.8434
Table 2. The objective evaluation of the methods in Figure 5.
Table 2. The objective evaluation of the methods in Figure 5.
VIFFQSAGCCSFQW
GFF0.39820.719726.74010.892635.13800.7640
IFM0.36790.692527.47350.884036.65620.7345
DTCWT0.52550.738428.85000.889937.56510.7866
CVT0.53960.731029.27260.889637.62900.7828
PC0.37120.637924.66700.874834.98340.6894
SAIF0.46890.723927.96490.887537.69710.7872
DRTV0.36330.608222.45630.869431.28560.6744
MISF0.46300.725227.27440.885936.60620.7721
NSST0.51190.752128.89610.882037.04270.7872
Proposed0.59400.762530.11320.892138.98780.8034
Table 3. The objective evaluation of the methods in Figure 6.
Table 3. The objective evaluation of the methods in Figure 6.
VIFFQSAGCCSFQW
GFF0.40480.796522.77790.630033.98690.7602
IFM0.25640.677823.41840.631534.62520.5919
DTCWT0.41200.777224.52380.658335.95600.7537
CVT0.42580.761424.85280.661035.61060.7490
PC0.33810.718622.98230.622635.09670.6680
SAIF0.34930.768924.15200.621736.01280.7543
DRTV0.29700.643018.52590.597225.20820.5422
MISF0.38380.772223.65380.611236.17460.7535
NSST0.42990.791124.22490.632435.54510.7750
Proposed0.54300.796525.31220.651236.53620.7706
Table 4. The objective evaluation of the methods in Figure 7.
Table 4. The objective evaluation of the methods in Figure 7.
VIFFQSAGCCSFQW
GFF0.73390.952013.44160.932517.03490.9294
IFM0.68860.946513.53120.930217.14100.9100
DTCWT0.79970.949713.76630.941317.60680.9306
CVT0.80470.948513.82260.940917.59720.9304
PC0.69680.81249.45840.872614.60770.8451
SAIF0.74750.951013.26810.932017.00350.9297
DRTV0.52620.69005.43410.917910.89940.7934
MISF0.74290.949813.35930.930117.16030.9235
NSST0.71330.940613.08940.925015.99540.9068
Proposed0.82600.952913.91890.941417.79910.9366
Table 5. The average objective evaluation of the methods on the sixteen group images.
Table 5. The average objective evaluation of the methods on the sixteen group images.
VIFFQSAGCCSFQW
GFF0.50400.816517.26880.802524.77220.8166
IFM0.41670.759617.93190.770625.74610.7344
DTCWT0.56890.822918.33040.827125.86260.8266
CVT0.57590.814518.59070.827125.80540.8230
PC0.41880.724814.64690.775822.25730.6786
SAIF0.57300.819117.81520.786325.66500.8366
DRTV0.38850.707714.59270.787320.15730.6742
MISF0.55630.817017.65020.781125.51960.8265
NSST0.59020.820816.78400.801823.45100.8168
Proposed0.63720.839418.88700.827326.39300.8401
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, L.; Ma, H. Saliency-Guided Nonsubsampled Shearlet Transform for Multisource Remote Sensing Image Fusion. Sensors 2021, 21, 1756. https://doi.org/10.3390/s21051756

AMA Style

Li L, Ma H. Saliency-Guided Nonsubsampled Shearlet Transform for Multisource Remote Sensing Image Fusion. Sensors. 2021; 21(5):1756. https://doi.org/10.3390/s21051756

Chicago/Turabian Style

Li, Liangliang, and Hongbing Ma. 2021. "Saliency-Guided Nonsubsampled Shearlet Transform for Multisource Remote Sensing Image Fusion" Sensors 21, no. 5: 1756. https://doi.org/10.3390/s21051756

APA Style

Li, L., & Ma, H. (2021). Saliency-Guided Nonsubsampled Shearlet Transform for Multisource Remote Sensing Image Fusion. Sensors, 21(5), 1756. https://doi.org/10.3390/s21051756

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop