[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Soil Moisture Monitoring at Kilometer Scale: Assimilation of Sentinel-1 Products in ISBA
Previous Article in Journal
Low-Illumination Image Enhancement Using Local Gradient Relative Deviation for Retinex Models
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CCC-SSA-UNet: U-Shaped Pansharpening Network with Channel Cross-Concatenation and Spatial–Spectral Attention Mechanism for Hyperspectral Image Super-Resolution

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610041, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(17), 4328; https://doi.org/10.3390/rs15174328
Submission received: 18 July 2023 / Revised: 14 August 2023 / Accepted: 16 August 2023 / Published: 2 September 2023
(This article belongs to the Section AI Remote Sensing)
Graphical abstract
">
Figure 1
<p>The schematic diagram of the training phase and testing phase of the deep learning-based HSI-PAN fusion network. (<b>a</b>) Training phase; (<b>b</b>) testing phase.</p> ">
Figure 2
<p>The architecture of the proposed CCC-SSA-UNet. CCC-SSA-UNet takes an LR-HSI and a PAN as input and takes an HR-HSI as output. It combines UNet and SSA-Net, exploits the “Conv Block” as the main building block of the UNet backbone, and adopts the Res-SSA block as the main building block of the SSA-Net. In each “Conv Block” of the UNet, the input is first passed through a 3 × 3 convolution and subsequently is passed through a batch normalization layer (BN) and a LeakyReLU layer. In this way, the feature map is extracted and passed to the next layer. Finally, the output of UNet is passed through a 1 × 1 convolution to produce the residual map of the Up-HSI and then the residual map and Up-HSI are element-wise summed to produce the final output HR-HSI. “DownSample 2×” denotes 2 × 2 maxpooling, and “UpSample 2×” and “UpSample 4×” denote bilinear interpolation with scale 2 and scale 4, respectively. “Input CCC” and “Fea CCC” denote channel cross-concatenation of input images and feature maps, respectively.</p> ">
Figure 3
<p>Schematic illustration of the proposed CCC method. (<b>a</b>) Schematic illustration of the proposed Input CCC operation. The tensors corresponding to Up-HSI and PAN are taken as inputs, and the result of channel cross-concatenation is used as the output. The spatial resolution of Up-HSI is <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>×</mo> <mi>W</mi> </mrow> </semantics></math> pixels, and it has a spectral bandwidth of <math display="inline"><semantics> <mi>C</mi> </semantics></math>. The spatial resolution of PAN is <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>×</mo> <mi>W</mi> </mrow> </semantics></math> pixels, and it has a spectral bandwidth of <math display="inline"><semantics> <mrow> <msub> <mi>C</mi> <mi>p</mi> </msub> </mrow> </semantics></math>. The spatial resolution of output is <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>×</mo> <mi>W</mi> </mrow> </semantics></math> pixels, and it has a spectral bandwidth of <math display="inline"><semantics> <mrow> <msub> <mi>C</mi> <mi mathvariant="normal">o</mi> </msub> </mrow> </semantics></math>. (<b>b</b>) Schematic illustration of the proposed Feature CCC operation. Feature 1 denotes the first input feature map of Feature CCC, and Feature 2 denotes the second input feature map of Feature CCC. The tensors corresponding to Feature 1 and Feature 2 are taken as inputs, and the result of channel cross-concatenation is used as the output. The spatial resolution of Feature 1 is <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>×</mo> <mi>W</mi> </mrow> </semantics></math> pixels, and it has a spectral bandwidth of <math display="inline"><semantics> <mrow> <msub> <mi>C</mi> <mi>q</mi> </msub> </mrow> </semantics></math>. The spatial resolution of Feature 2 is <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>×</mo> <mi>W</mi> </mrow> </semantics></math> pixels, and it has a spectral bandwidth of <math display="inline"><semantics> <mrow> <msub> <mi>C</mi> <mi>r</mi> </msub> </mrow> </semantics></math>. The spatial resolution of output is <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>×</mo> <mi>W</mi> </mrow> </semantics></math> pixels, and it has a spectral bandwidth of <math display="inline"><semantics> <mrow> <msub> <mi>C</mi> <msup> <mi mathvariant="normal">o</mi> <mo>′</mo> </msup> </msub> </mrow> </semantics></math>.</p> ">
Figure 4
<p>Schematic illustration of the proposed SSA-Net. The SSA-Net consists of <span class="html-italic">N</span> Res-SSA blocks connected sequentially. And “Res-SSA” denotes residual spatial–spectral attention. In the proposed Res-SSA block, spectral attention is parallel to spatial attention and is embedded in the basic residual module.</p> ">
Figure 5
<p>Visual results generated by different pansharpening algorithms for the 10th patch of the Pavia University dataset. (<b>a</b>) GS [<a href="#B5-remotesensing-15-04328" class="html-bibr">5</a>], (<b>b</b>) GSA [<a href="#B6-remotesensing-15-04328" class="html-bibr">6</a>], (<b>c</b>) PCA [<a href="#B8-remotesensing-15-04328" class="html-bibr">8</a>], (<b>d</b>) GFPCA [<a href="#B7-remotesensing-15-04328" class="html-bibr">7</a>], (<b>e</b>) BayesNaive [<a href="#B14-remotesensing-15-04328" class="html-bibr">14</a>], (<b>f</b>) BayesSparse [<a href="#B44-remotesensing-15-04328" class="html-bibr">44</a>], (<b>g</b>) MTF-GLP [<a href="#B9-remotesensing-15-04328" class="html-bibr">9</a>], (<b>h</b>) MTF-GLP-HPM [<a href="#B11-remotesensing-15-04328" class="html-bibr">11</a>], (<b>i</b>) CNMF [<a href="#B16-remotesensing-15-04328" class="html-bibr">16</a>], (<b>j</b>) HySure [<a href="#B12-remotesensing-15-04328" class="html-bibr">12</a>], (<b>k</b>) HyperPNN1 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>l</b>) HyperPNN2 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>m</b>) DHP-DARN [<a href="#B30-remotesensing-15-04328" class="html-bibr">30</a>], (<b>n</b>) DIP-HyperKite [<a href="#B31-remotesensing-15-04328" class="html-bibr">31</a>], (<b>o</b>) Hyper-DSNet [<a href="#B29-remotesensing-15-04328" class="html-bibr">29</a>], (<b>p</b>) CCC-SSA-UNet-S (Ours), (<b>q</b>) CCC-SSA-UNet-L (Ours), and (<b>r</b>) reference ground truth. The RGB images are generated using the HSI’s 60th, 30th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).</p> ">
Figure 6
<p>Mean absolute error maps of different pansharpening algorithms for the 10th patch of the Pavia University dataset. (<b>a</b>) GS [<a href="#B5-remotesensing-15-04328" class="html-bibr">5</a>], (<b>b</b>) GSA [<a href="#B6-remotesensing-15-04328" class="html-bibr">6</a>], (<b>c</b>) PCA [<a href="#B8-remotesensing-15-04328" class="html-bibr">8</a>], (<b>d</b>) GFPCA [<a href="#B7-remotesensing-15-04328" class="html-bibr">7</a>], (<b>e</b>) BayesNaive [<a href="#B14-remotesensing-15-04328" class="html-bibr">14</a>], (<b>f</b>) BayesSparse [<a href="#B44-remotesensing-15-04328" class="html-bibr">44</a>], (<b>g</b>) MTF-GLP [<a href="#B9-remotesensing-15-04328" class="html-bibr">9</a>], (<b>h</b>) MTF-GLP-HPM [<a href="#B11-remotesensing-15-04328" class="html-bibr">11</a>], (<b>i</b>) CNMF [<a href="#B16-remotesensing-15-04328" class="html-bibr">16</a>], (<b>j</b>) HySure [<a href="#B12-remotesensing-15-04328" class="html-bibr">12</a>], (<b>k</b>) HyperPNN1 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>l</b>) HyperPNN2 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>m</b>) DHP-DARN [<a href="#B30-remotesensing-15-04328" class="html-bibr">30</a>], (<b>n</b>) DIP-HyperKite [<a href="#B31-remotesensing-15-04328" class="html-bibr">31</a>], (<b>o</b>) Hyper-DSNet [<a href="#B29-remotesensing-15-04328" class="html-bibr">29</a>], (<b>p</b>) CCC-SSA-UNet-S (Ours), (<b>q</b>) CCC-SSA-UNet-L (Ours), and (<b>r</b>) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands; the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).</p> ">
Figure 7
<p>Visual results generated by different pansharpening algorithms for the 15th patch of the Pavia Center dataset. (<b>a</b>) GS [<a href="#B5-remotesensing-15-04328" class="html-bibr">5</a>], (<b>b</b>) GSA [<a href="#B6-remotesensing-15-04328" class="html-bibr">6</a>], (<b>c</b>) PCA [<a href="#B8-remotesensing-15-04328" class="html-bibr">8</a>], (<b>d</b>) GFPCA [<a href="#B7-remotesensing-15-04328" class="html-bibr">7</a>], (<b>e</b>) BayesNaive [<a href="#B14-remotesensing-15-04328" class="html-bibr">14</a>], (<b>f</b>) BayesSparse [<a href="#B44-remotesensing-15-04328" class="html-bibr">44</a>], (<b>g</b>) MTF-GLP [<a href="#B9-remotesensing-15-04328" class="html-bibr">9</a>], (<b>h</b>) MTF-GLP-HPM [<a href="#B11-remotesensing-15-04328" class="html-bibr">11</a>], (<b>i</b>) CNMF [<a href="#B16-remotesensing-15-04328" class="html-bibr">16</a>], (<b>j</b>) HySure [<a href="#B12-remotesensing-15-04328" class="html-bibr">12</a>], (<b>k</b>) HyperPNN1 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>l</b>) HyperPNN2 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>m</b>) DHP-DARN [<a href="#B30-remotesensing-15-04328" class="html-bibr">30</a>], (<b>n</b>) DIP-HyperKite [<a href="#B31-remotesensing-15-04328" class="html-bibr">31</a>], (<b>o</b>) Hyper-DSNet [<a href="#B29-remotesensing-15-04328" class="html-bibr">29</a>], (<b>p</b>) CCC-SSA-UNet-S (Ours), (<b>q</b>) CCC-SSA-UNet-L (Ours), and (<b>r</b>) reference ground truth. The RGB images are generated using the HSI’s 60th, 30th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).</p> ">
Figure 8
<p>Mean absolute error maps of different pansharpening algorithms for the 15th patch of the Pavia Center dataset. (<b>a</b>) GS [<a href="#B5-remotesensing-15-04328" class="html-bibr">5</a>], (<b>b</b>) GSA [<a href="#B6-remotesensing-15-04328" class="html-bibr">6</a>], (<b>c</b>) PCA [<a href="#B8-remotesensing-15-04328" class="html-bibr">8</a>], (<b>d</b>) GFPCA [<a href="#B7-remotesensing-15-04328" class="html-bibr">7</a>], (<b>e</b>) BayesNaive [<a href="#B14-remotesensing-15-04328" class="html-bibr">14</a>], (<b>f</b>) BayesSparse [<a href="#B44-remotesensing-15-04328" class="html-bibr">44</a>], (<b>g</b>) MTF-GLP [<a href="#B9-remotesensing-15-04328" class="html-bibr">9</a>], (<b>h</b>) MTF-GLP-HPM [<a href="#B11-remotesensing-15-04328" class="html-bibr">11</a>], (<b>i</b>) CNMF [<a href="#B16-remotesensing-15-04328" class="html-bibr">16</a>], (<b>j</b>) HySure [<a href="#B12-remotesensing-15-04328" class="html-bibr">12</a>], (<b>k</b>) HyperPNN1 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>l</b>) HyperPNN2 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>m</b>) DHP-DARN [<a href="#B30-remotesensing-15-04328" class="html-bibr">30</a>], (<b>n</b>) DIP-HyperKite [<a href="#B31-remotesensing-15-04328" class="html-bibr">31</a>], (<b>o</b>) Hyper-DSNet [<a href="#B29-remotesensing-15-04328" class="html-bibr">29</a>], (<b>p</b>) CCC-SSA-UNet-S (Ours), (<b>q</b>) CCC-SSA-UNet-L (Ours), and (<b>r</b>) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands, the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).</p> ">
Figure 9
<p>Visual results generated by different pansharpening algorithms for the 31st patch of the Chikusei dataset. (<b>a</b>) GS [<a href="#B5-remotesensing-15-04328" class="html-bibr">5</a>], (<b>b</b>) GSA [<a href="#B6-remotesensing-15-04328" class="html-bibr">6</a>], (<b>c</b>) PCA [<a href="#B8-remotesensing-15-04328" class="html-bibr">8</a>], (<b>d</b>) GFPCA [<a href="#B7-remotesensing-15-04328" class="html-bibr">7</a>], (<b>e</b>) BayesNaive [<a href="#B14-remotesensing-15-04328" class="html-bibr">14</a>], (<b>f</b>) BayesSparse [<a href="#B44-remotesensing-15-04328" class="html-bibr">44</a>], (<b>g</b>) MTF-GLP [<a href="#B9-remotesensing-15-04328" class="html-bibr">9</a>], (<b>h</b>) MTF-GLP-HPM [<a href="#B11-remotesensing-15-04328" class="html-bibr">11</a>], (<b>i</b>) CNMF [<a href="#B16-remotesensing-15-04328" class="html-bibr">16</a>], (<b>j</b>) HySure [<a href="#B12-remotesensing-15-04328" class="html-bibr">12</a>], (<b>k</b>) HyperPNN1 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>l</b>) HyperPNN2 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>m</b>) DHP-DARN [<a href="#B30-remotesensing-15-04328" class="html-bibr">30</a>], (<b>n</b>) DIP-HyperKite [<a href="#B31-remotesensing-15-04328" class="html-bibr">31</a>], (<b>o</b>) Hyper-DSNet [<a href="#B29-remotesensing-15-04328" class="html-bibr">29</a>], (<b>p</b>) CCC-SSA-UNet-S (Ours), (<b>q</b>) CCC-SSA-UNet-L (Ours), and (<b>r</b>) reference ground truth. The RGB images are generated using the HSI’s 61st, 35th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).</p> ">
Figure 10
<p>Mean absolute error maps of different pansharpening algorithms for the 31st patch of the Chikusei dataset. (<b>a</b>) GS [<a href="#B5-remotesensing-15-04328" class="html-bibr">5</a>], (<b>b</b>) GSA [<a href="#B6-remotesensing-15-04328" class="html-bibr">6</a>], (<b>c</b>) PCA [<a href="#B8-remotesensing-15-04328" class="html-bibr">8</a>], (<b>d</b>) GFPCA [<a href="#B7-remotesensing-15-04328" class="html-bibr">7</a>], (<b>e</b>) BayesNaive [<a href="#B14-remotesensing-15-04328" class="html-bibr">14</a>], (<b>f</b>) BayesSparse [<a href="#B44-remotesensing-15-04328" class="html-bibr">44</a>], (<b>g</b>) MTF-GLP [<a href="#B9-remotesensing-15-04328" class="html-bibr">9</a>], (<b>h</b>) MTF-GLP-HPM [<a href="#B11-remotesensing-15-04328" class="html-bibr">11</a>], (<b>i</b>) CNMF [<a href="#B16-remotesensing-15-04328" class="html-bibr">16</a>], (<b>j</b>) HySure [<a href="#B12-remotesensing-15-04328" class="html-bibr">12</a>], (<b>k</b>) HyperPNN1 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>l</b>) HyperPNN2 [<a href="#B28-remotesensing-15-04328" class="html-bibr">28</a>], (<b>m</b>) DHP-DARN [<a href="#B30-remotesensing-15-04328" class="html-bibr">30</a>], (<b>n</b>) DIP-HyperKite [<a href="#B31-remotesensing-15-04328" class="html-bibr">31</a>], (<b>o</b>) Hyper-DSNet [<a href="#B29-remotesensing-15-04328" class="html-bibr">29</a>], (<b>p</b>) CCC-SSA-UNet-S (Ours), (<b>q</b>) CCC-SSA-UNet-L (Ours), and (<b>r</b>) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands, the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).</p> ">
Figure 11
<p>Visual comparison of the computational complexities among different pansharpening methods on the Pavia University dataset. (<b>a</b>) Comparison of PSNR, FLOPs, and GPU memory usage. (<b>b</b>) Comparison of SAM, FLOPs, and GPU memory usage. (<b>c</b>) Comparison of PSNR and Runtime. (<b>d</b>) Comparison of SAM and Runtime.</p> ">
Figure 12
<p>Schematic illustration of several attention block variants. (<b>a</b>) Residual block baseline, (<b>b</b>) CA, (<b>c</b>) SA, (<b>d</b>) CSA, and (<b>e</b>) DAU.</p> ">
Versions Notes

Abstract

:
A hyperspectral image (HSI) has a very high spectral resolution, which can reflect the target’s material properties well. However, the limited spatial resolution poses a constraint on its applicability. In recent years, some hyperspectral pansharpening studies have attempted to integrate HSI with PAN to improve the spatial resolution of HSI. Although some achievements have been made, there are still shortcomings, such as insufficient utilization of multi-scale spatial and spectral information, high computational complexity, and long network model inference time. To address the above issues, we propose a novel U-shaped hyperspectral pansharpening network with channel cross-concatenation and spatial–spectral attention mechanism (CCC-SSA-UNet). A novel channel cross-concatenation (CCC) method was designed to effectively enhance the fusion ability of different input source images and the fusion ability between feature maps at different levels. Regarding network design, integrating a UNet based on an encoder–decoder architecture with a spatial–spectral attention network (SSA-Net) based on residual spatial–spectral attention (Res-SSA) blocks further enhances the ability to extract spatial and spectral features. The experiment shows that our proposed CCC-SSA-UNet exhibits state-of-the-art performance and has a shorter inference runtime and lower GPU memory consumption than most of the existing hyperspectral pansharpening methods.

Graphical Abstract">

Graphical Abstract

1. Introduction

As a data cube, HSI often contains hundreds of spectral bands. HSI can reconstruct any point in space through its continuous and fine spectral curves, thus simultaneously obtaining the target’s spatial and material properties. However, due to the limitations of sensors, HSI often has low spatial resolution. In many applications, such as fine classification of land cover [1], medical diagnosis [2], and anomaly detection [3], HSI is required to have the characteristics of high spatial resolution and hyperspectral resolution at the same time. For these reasons, HSI SR has become a research hotspot. At present, RGB image super-resolution technology is very mature. There are many methods for single RGB image super-resolution (SISR), including the interpolation method, reconstruction method, traditional machine-learning method, and deep learning method. However, because hyperspectral images have higher spectral dimensions than RGB images, HSI SR technology is more challenging than SISR. At present, HSI SR technology has achieved some research results but is still under development. How to improve the spatial resolution and maintain as much spectral information as possible is a pressing problem to be solved [4].
According to whether additional information is used (such as multispectral image (MSI), RGB image, or panchromatic image (PAN)), the existing HSI SR methods can be roughly classified into the following three categories: (1) single HSI SR method without any other auxiliary image; (2) RGB or MSI spectral super-resolution reconstruction method; and (3) the method based on fusion is to fuse HSI and its corresponding auxiliary image, and the method of fusing HSI and PAN is also called hyperspectral pansharpening.
Hyperspectral pansharpening has evolved from the research on pansharpening, which involves the fusion of MSI and PAN. This method achieves a balance between improving spatial resolution and maintaining spectral information. The study of MSI and PAN image fusion has been ongoing for several decades and has reached a performance bottleneck, with minimal differences in the performance of different fusion methods, making them visually indistinguishable. On the other hand, research on the fusion of HSI and PAN started relatively late and still holds significant potential for development.
Early pansharpening methods primarily relied on traditional techniques such as component substitution (CS) [5,6,7,8], multiresolution analysis (MRA) [9,10,11], and optimization-based methods [12,13,14,15,16]. In recent years, deep learning methods based on artificial neural networks have exhibited remarkable efficacy in diverse computer vision tasks. These tasks include high-level applications such as object detection [17,18], object tracking [19], and image classification [20], as well as low-level applications such as image denoising [21], image deblurring [22], and image super-resolution [23,24,25,26,27], achieving excellent results. Inspired by these studies, researchers have gradually started to introduce various deep learning methods into the field of hyperspectral pansharpening. Examples include HyperPNN [28] and Hyper-DSNet [29] based on convolutional neural networks (CNNs), DHP-DARN [30], and DIP-HyperKite [31] based on deep image prior network (DIP-Net) and CNNs, PS-GDANet [32], and HPGAN [33] based on generative adversarial networks (GANs), as well as HyperTransformer [34] based on Transformer.
After an extensive literature review, we identified the following issues with existing deep learning-based hyperspectral pansharpening methods. HyperPNN focuses solely on the fusion of single-scale information and fails to achieve satisfactory fusion quality. DIP-HyperKite employs an encoder–decoder network with layer-wise upsampling and downsampling for multiscale information fusion, which significantly increases computational complexity and GPU memory consumption. Both DHP-DAR and DIP-HyperKite utilize a two-step pansharpening approach, involving upsampling the LR-HSI using a deep prior network followed by image fusion using a convolutional neural network. This approach substantially increases the inference runtime of the network model, and the adopted deep prior network exhibits significant instability. While recent Transformer-based pansharpening networks such as HyperTransformer have achieved good fusion results, their large parameter and computational requirements pose high demands on computer performance, thereby reducing their practicality. To address the drawbacks of the aforementioned methods and effectively extract spatial and spectral features from the input HSI and PAN, we propose a U-shaped network with channel cross-connection and spatial–spectral attention mechanism.
The contributions of this paper are summarized as follows:
  • We propose a novel framework for hyperspectral pansharpening named the CCC-SSA-UNet, which integrates the UNet architecture with the SSA-Net.
  • We propose a novel channel cross-concatenation method called Input CCC at the network’s entrance. This method effectively enhances the fusion capability of different input source images while introducing only a minimal number of additional parameters. Furthermore, we propose a Feature CCC approach within the decoder. This approach effectively strengthens the fusion capacity between different hierarchical feature maps without introducing any extra parameters or computational complexity.
  • We propose an improved Res-SSA block to enhance the representation capacity of spatial and spectral features. Experimental results demonstrate the effectiveness of our proposed hybrid attention module and its superiority over other attention module variants.
The remaining sections of this paper are organized as follows. Section 2 reviews related works about pansharpening methods, including classical pansharpening methods and deep learning-based methods. Section 3 provides a detailed exposition of the proposed methodology. Section 4 presents the experimental results and provides a comprehensive discussion. The conclusion of the paper is given in Section 5.

2. Related Work

HSI holds versatile applications in pansharpening [35], change detection [36], object detection [37], and classification [38] tasks, making it invaluable in various domains. This research specifically delves into the realm of pansharpening, where our focus lies. Within this section, we meticulously investigate both classical pansharpening methods and deep learning-based pansharpening methods, aiming to unravel their potential and advancements in this domain.

2.1. Classical Pansharpening Methods

Classical pansharpening methods can be roughly classified into the following three categories: CS, MRA, and optimization-based methods.
The CS approach first transforms the HSI into a new projection space, decomposes it into spectral components and spatial components, and then replaces its spatial components with a panchromatic image. Subsequently, the inverse transformation is applied to generate the reconstructed image. Typical CS approaches include IHS color space transformation [39], Gram–Schmidt (GS) transformation [5], Gram–Schmidt transformation with adaptive weights (GSA) [6], principal component analysis (PCA) [8], and guided filter principal component analysis (GFPCA) [7]. This category of methods is easy to implement and exhibits fast processing speeds while effectively preserving spatial information. However, these methods might introduce certain degrees of distortion to spectral information.
The MRA approach involves initially downsampling the HR-PAN image at multiple scales and decomposing it into high-frequency and low-frequency components. Subsequently, these components are then fused with the upsampled HSI according to various fusion rules, and finally, an inverse transformation is applied to obtain the reconstructed image. Typical methods within the MRA category can be classified as follows: the Laplacian pyramid method [40], the approach [41] based on undecimated discrete wavelet transform (UDWT) and the generalized Laplacian pyramid (GLP), the method using modulation transfer function and GLP (MTF-GLP) [9], and MTF-GLP with High Pass Modulation (MTF-GLP-HPM) [11], as well as the integration of MRA with CNNs, as seen in LPPNet [42]. The advantages of these methods lie in their ability to incorporate high-frequency spatial details into the HSI while preserving spectral information. However, they may result in the loss of some spatial information and introduce ringing artifacts.
CS-based and MRA-based methods are primarily employed in the field of MSI pansharpening. Due to the relatively low spatial resolution of HSI, pixel-level ambiguity often arises, rendering CS and MRA less suitable for addressing the fusion of HSI and PAN. Instead, optimization-based approaches are suitable to tackle this problem.
The core idea of the optimization-based approach lies in treating the fusion problem as an inverse reconstruction problem. By establishing a relationship model between the original image and the reference ground truth image, the model is mathematically optimized to obtain a solution. Bayesian estimation methods [12,13,14,43,44] and matrix factorization methods [15,16,45] are commonly used optimization-based methods. In contrast to the CS-based and MRA-based methods, these methods perform well in preserving both spatial and spectral information. However, due to their high computational demands, these methods also require massive computational resources.

2.2. Deep Learning-Based Pansharpening Methods

In the earlier research on deep learning-based pansharpening, the Pansharping Neural Network (PNN) [46] treated HR-PAN as an additional spectral band of LR-MSI, and employed three convolutional layers to learn the mapping relationship between the composited image of HR-PAN and LR-MSI and the reference ground truth MSI. However, this fusion method only combined the two input images at a basic level. Moreover, the utilized convolutional network was overly simplistic, which hindered the extraction of intricate spectral and spatial information. Consequently, the fusion performance was compromised. Yuan et al. [47] introduced a multi-scale and multi-depth CNN (MSDCNN) that improved upon the PNN by employing parallel multi-scale convolutional blocks to enhance the representational capacity.
Certain approaches employ a strategy of separately extracting features from the two input images before fusion. Liu et al. [48] proposed a TFNet, which employs two sub-networks to extract spectral and spatial features from the upsampled LR-MSI and the corresponding HR-PAN images, respectively. These features are then fused using a fusion network comprising multiple convolutional layers. Finally, an image reconstruction process is carried out using a reconstruction network.
In addition to fusing the features extracted from the two inputs, some methods propose the fusion of an original HR-PAN with deep features extracted from LR-HSI. He et al. [28] introduced two spectral prediction CNNs, called HyperPNN1 and HyperPNN2, which initially extract spectral features from the upsampled LR-HSI using two convolutional layers. These spectral features are then concatenated with the HR-PAN image and subjected to fusion reconstruction through multiple convolutional layers. While both HyperPNN1 and HyperPNN2 share a fundamental network structure, HyperPNN2 incorporates an additional residual structure with skip connections. These methods effectively extract features from the LR-HSI input. However, the process of feature fusion does not adequately account for the correlation between LR-HSI and HR-PAN. Additionally, some methods adopt the fusion of the upsampled LR-HSI with high-frequency detail features extracted from HR-PAN. Zhuo et al. [29] devised an HSI pansharpening network named Hyper-DSNet. This network employs five spatial domain high-pass filter templates to extract high-frequency detail characteristics from the HR-PAN. Subsequently, these extracted details are concatenated with the upsampled LR-HSI in the spectral dimension. The network architecture incorporates multi-scale convolutional modules, shallow-to-deep fusion structures, and a spectral attention mechanism. This method retains inherent spatial details and spectral fidelity.
In recent years, attention mechanisms have been widely applied in various computer vision tasks such as image super-resolution, object detection, and object recognition. The principle underlying attention mechanism is to automatically highlight the most informative components while suppressing less relevant ones, thereby enhancing computational efficiency. Hu et al. [49] initially introduced the channel attention mechanism, where a Squeeze-and-Excitation (SE) module, constructed using global average pooling along the spatial dimensions and two 1 × 1 convolutions, was employed to improve the object recognition performance of networks. Building upon the SE module, Roy et al. [50] proposed a concurrent spatial and channel Squeeze-and-Excitation (scSE) module, which utilized convolutional layers with 1 × 1 × C kernels and sigmoid activation functions to generate spatial attention maps. The scSE module then combined with the channel attention mechanism, yielding promising results in medical image segmentation. Motivated by this, Zheng et al. [30] proposed a hyperspectral pansharpening approach based on Deep Hyperspectral Prior (DHP) and Dual Attention Residual Network (DARN) that combines spatial–spectral attention mechanisms. In this approach, the DHP process solely employs spectral constraints, overlooking spatial constraints. Moreover, the fusion network only employs single-scale feature maps for fusion, neglecting multi-scale feature information. To overcome these limitations, Bandara et al. [31] introduced a novel spatial constraint in the Deep Image Prior (DIP) upsampling process and proposed the HyperKite network for residual reconstruction. HyperKite employs an encoder–decoder network that sequentially performs upsampling and downsampling layers for multi-scale feature fusion. However, the simplistic encoder–decoder architecture in HyperKite hinders the extraction of fine spectral and spatial information. Additionally, the layered upsampling and downsampling design imposes a significant computational burden.
In addition to the aforementioned CNN-based pansharpening methods, in recent years, pansharpening approaches based on Generative Adversarial Networks (GANs) have also emerged. Dong et al. [32] developed a specific pansharpening framework using a Paired-Shared Generative Dual Adversarial Network (PS-GDANet), featuring two discriminators. The spatial discriminator enforces the similarity between the intensity component of the pansharpened image and the panchromatic (PAN) image, while the spectral discriminator aids in preserving the spectral characteristics of the original HSI image. This configuration enables the network to generate high-resolution pansharpened images. Xie et al. [33] introduced a high-dimensional pansharpening framework called HPGAN based on a 3D Generative Adversarial Network (3D-GAN) and devised a loss function that comprehensively considers global, spectral, and spatial constraints. Despite the favorable perceptual quality of images generated by GANs, the instability in generating images was not well received in the remote sensing field.
The Transformer architecture initially emerged in the field of natural language processing and was later introduced to computer vision. In recent years, with the continuous development and expansion of the Transformer, it has also made its mark in the domain of hyperspectral pansharpening. Bandara et al. [34] introduced a novel Transformer-based pansharpening network known as HyperTransformer. This network comprises three core modules: two separate PAN and HSI feature extractors, a multi-head feature attention module, and a spatial–spectral feature fusion module. Despite its enhancement of the spatial and spectral quality of the pansharpened HSI, the network’s large parameter count and high computational load pose challenges to its practical applicability.

3. Proposed Method

This section will provide a detailed introduction to the proposed method from three aspects: problem statement and formulation, network architecture design, and loss function design.

3.1. Problem Statement and Formulation

Original hyperspectral image (LR-HSI) possesses high spectral resolution but suffers from low spatial resolution, whereas panchromatic image (PAN) exhibits high spatial resolution but lacks spectral information. Therefore, employing a fusion approach to combine these two types of image data is an effective means to obtain high-spatial-resolution hyperspectral image (HR-HSI). The main objective of this paper is to design a deep neural network model that can fuse LR-HSI and PAN to generate high-quality HR-HSI.
Let Χ h × w × C represent LR-HSI, with a spatial resolution of h × w pixels and C spectral bands. Let P H × W × 1 represent PAN, with a spatial resolution of H × W pixels and a single spectral band. Let Y ^ H × W × C denote the reconstructed hyperspectral image (HR-HSI), with a spatial resolution of H × W pixels and C spectral bands. Let Y H × W × C represent the reference ground truth HR-HSI (Ref-HR-HSI), with a spatial resolution of H × W pixels and C spectral bands. Additionally, it is assumed that conditions H > h , W > w and C > > 1 hold. Then, the training process of the HSI-PAN fusion network can be described as follows: The training dataset { [ Χ 1 , P 1 , Y 1 ] , , [ Χ D , P D , Y D ] } consists of D pairs of images. These images are processed by the neural network model Φ ( · , · ; Θ ) , resulting in the output image set [ Y ^ 1 , , Y ^ D ] . The parameters Θ of the neural network are continuously optimized and adjusted using an optimization algorithm, aiming to minimize the difference between Y ^ d ( 1 d D ) and Y d until it converges to a certain value. The training process of the network can be represented by the following equation:
Θ ^ = argmin Θ 1 D d = 1 D Loss ( Y ^ d , Y d ) s . t . Y ^ d = Φ ( X d , P d ; Θ )
where Θ ^ represents the optimized network parameters and Loss ( · , · ) refers to the loss function employed by the network. The loss function quantifies the dissimilarity between the predicted output Y ^ d and the desired target Y d , facilitating the training and optimization of the network.
During the testing phase, test image pairs [ Χ t , P t ] are processed using a neural network model Φ ( · , · ; Θ ^ ) with pre-trained parameters Θ ^ , resulting in the final output fused image Y ^ t . The testing process of the network can be represented by the following equation:
Y ^ t = Φ ( X t , P t ; Θ ^ )
where, X t and P t respectively represent the input LR-HSI and PAN images used for testing, while Y ^ t denotes the fused image, which is the final output of the network.
The schematic diagram of the training and testing phases of the deep learning-based HSI-PAN fusion network is illustrated in Figure 1.

3.2. Network Design

Figure 2 illustrates the overall network architecture of the proposed CCC-SSA-UNet. CCC-SSA-UNet takes an LR-HSI (represented by X h × w × C ) and a PAN (represented by P H × W × 1 ) as initial inputs and outputs an HR-HSI (represented by Y ^ H × W × C ). Following the design of DHP-DARN [30] and DIP-HyperKite [31], our pansharpening network CCC-SSA-UNet adopts a residual learning-based framework. Firstly, LR-HSI X is upsampled using bilinear interpolation to obtain the image U H × W × C , which has the same spatial resolution as P . Then, the proposed Input CCC method is applied to cross-concatenate U and P in the channel dimension, resulting in the image O H × W × ( C + m ) . Image O is fed into a U-shaped network to learn the residual image X r e s H × W × C for HR-HSI. Finally, the residual image X r e s is pixel-wise added to the image U to obtain the final fusion result Y ^ . The aforementioned process can be described using the following equation:
U = ( X )
X r e s = f C C C - S S A - U N e t ( U , P )
Y ^ = U + X r e s
where ( · ) represents bilinear interpolation for upsampling, while f C C C S S A U N e t ( · , · ) represents the proposed CCC-SSA-UNet network introduced in this paper.
The main idea behind CCC-SSA-UNet is to integrate the U-Net [51], based on an encoder–decoder architecture, with the SSA-Net, which incorporates spatial–spectral residual attention modules. The encoder–decoder architecture is applicable in many areas such as medical science [52], HSI classification [53], and agriculture science [54]. Firstly, we construct a U-shaped encoder–decoder network similar to U-Net, named UNet. Within UNet, we introduce the SSA-Net, which utilizes spatial–spectral attention mechanisms, between the layers of the encoder and their corresponding decoder counterparts. This design aims to enhance the expression capability of both spatial and spectral features. Additionally, we propose novel channel cross-concatenation methods, namely Input CCC and Feature CCC, at the network’s entrance and within the decoder, respectively. These methods effectively enhance the fusion capability of different input source images and the fusion capability between different hierarchical feature maps while minimizing additional computational complexity.

3.2.1. UNet Backbone

The network design of CCC-SSA-UNet draws inspiration from the state-of-the-art RGB image denoising method DRUNet [55] and the hyperspectral pansharpening method DIP-HyperKite [31]. Similar to DRUNet, our UNet backbone network consists of four scales with skip connections between the encoder and decoder at each scale. The number of channels at each layer varies across scales, denoted as C f 0 , C f 1 , C f 2 , and C f 2 from the first to the fourth scale, respectively. These parameters are determined through experimentation, and the specific parameter settings will be described in Section 4.3.
The UNet backbone network we propose is composed of several key components. The encoder consists of three Conv Block modules and three downsampling modules, which are arranged in an alternating manner. Similarly, the decoder follows the same structure, with three upsampling modules and three Conv Block modules. The skip connections between each layer of the encoder and its corresponding decoder are equipped with the SSA-Net, enhancing the fusion and representation capabilities. Additionally, a Bottleneck layer, comprising one Conv Block module, resides between the last downsampling module and the first upsampling module, facilitating the information flow between the encoder and decoder pathways.
The Conv Block module consists of consecutive layers, including a 3 × 3 convolutional layer with a stride of 1, a batch normalization layer, and a LeakyReLU activation function layer. This module plays a crucial role in feature extraction in the encoder and feature reconstruction in the decoder. Mathematically, the Conv Block module can be represented as follows:
C B o u t = f C B ( C B i n ) = δ ( BN ( Con v 3 × 3 ( C B i n ) ) )
where C B i n represents the input and C B o u t represents the output of the Conv Block module, f C B ( · ) represents the function representation of the Conv Block module, δ ( · ) represents the LeakyReLU activation function, BN ( · ) represents the batch normalization layer, and C o n v 3 × 3 ( · ) represents the 3 × 3 convolutional layer. It should be noted that the Conv Block module sequentially connects these layers to effectively capture and process the input features throughout the network.
In the encoder section, the output image O H × W × ( C + m ) from Input CCC undergoes a series of operations. First, it passes through the first Conv Block module, resulting in the first-level feature map E 1 H × W × C f 0 . Subsequently, E 1 is processed by a downsampling layer, yielding a feature map E d 1 H / 2 × W / 2 × C f 0 with half the spatial dimensions. E d 1 then undergoes the second Conv Block module, generating the second-level feature map E 2 H / 2 × W / 2 × C f 1 , which is further downsampled to obtain the feature map E d 2 H / 4 × W / 4 × C f 1 . Similarly, E d 2 goes through the third Conv Block module, producing the third-level feature map E 3 H / 4 × W / 4 × C f 2 , followed by downsampling to obtain the feature map E d 3 H / 8 × W / 8 × C f 2 . Finally, E d 3 is processed by the fourth Conv Block module, generating the feature map B H / 8 × W / 8 × C f 2 in the Bottleneck layer. This process can be mathematically represented as follows:
E 1 = f C B 1 ( O )
E 2 = f C B 2 ( E d 1 ) = f C B 2 ( ( E 1 ) )
E 3 = f C B 3 ( E d 2 ) = f C B 3 ( ( E 2 ) )
B = f C B 4 ( E d 3 ) = f C B 4 ( ( E 3 ) )
where f C B i ( · ) , i = 1 , 2 , 3 , 4 represents the functional notation for the i-th Conv Block module, which greatly contributes to feature extraction and reconstruction. On the other hand, ( · ) signifies the 2 × 2 maxpooling operation, employed for downsampling the feature maps to reduce their spatial dimensions.
Subsequently, the feature maps from the first three levels denoted as E 1 , E 2 , and E 3 undergo spatial and spectral feature enhancement using the respective SSA-Net modules at their corresponding levels. This enhancement process yields refined feature maps Q 1 H × W × C f 0 , Q 2 H / 2 × W / 2 × C f 1 , and Q 3 H / 4 × W / 4 × C f 2 , which embody strengthened spatial and spectral characteristics. These transformations can be mathematically represented by the equations as follows:
Q 1 = f S S A N e t 1 ( E 1 )
Q 2 = f S S A N e t 2 ( E 2 )
Q 3 = f S S A N e t 3 ( E 3 )
where f S S A N e t 1 ( · ) , f S S A N e t 2 ( · ) , and f S S A N e t 3 ( · ) denote the functional representations of the SSA-Net modules at the three corresponding levels. The comprehensive design details of these modules will be presented in Section 3.2.3.
Within the decoder section, the feature map B is initially subjected to an upsampling operation, yielding the feature map R 3 H / 4 × W / 4 × C f 2 . Subsequently, the Feature CCC method is utilized to concatenate the feature map R 3 with the output feature map Q 3 from the third-level SSA-Net, leading to the formation of the feature map O 3 H / 4 × W / 4 × 2 C f 2 . Following this, O 3 is processed through the fifth Conv Block module, resulting in the refined feature map D 1 H / 4 × W / 4 × C f 2 , which is further upsampled to generate the feature map R 2 H / 2 × W / 2 × C f 1 . Simultaneously, the feature map R 2 is merged with the output feature map Q 2 from the second-level SSA-Net via channel concatenation, resulting in the composite feature map O 2 H / 2 × W / 2 × 2 C f 1 . O 2 is then passed through the sixth Conv Block module, generating the enhanced feature map D 2 H / 2 × W / 2 × C f 0 . Further upsampling is performed on D 2 , producing the feature map R 1 H × W × C f 0 . Lastly, the feature map R 1 is merged with the output feature map Q 1 from the first-level SSA-Net through channel concatenation, yielding the combined feature map O 1 H × W × 2 C f 0 . O 1 subsequently undergoes processing through the seventh Conv Block module, culminating in the final feature map D 3 H × W × C . These transformations can be mathematically represented by the equations as follows:
O 3 = Fea CCC ( Q 3 , R 3 ) = Fea CCC ( Q 3 , ( B ) )
D 1 = f C B 5 ( O 3 )
O 2 = Fea CCC ( Q 2 , R 2 ) = Fea CCC ( Q 2 , ( D 1 ) )
D 2 = f C B 6 ( O 2 )
O 1 = Fea CCC ( Q 1 , R 1 ) = Fea CCC ( Q 1 , ( D 2 ) )
D 3 = f C B 7 ( O 1 )
where f C B i ( · ) , i = 5 , 6 , 7 represents the functional notation for the i-th Conv Block module, while ( · ) denotes the bilinear interpolation operation for upsampling, which increases the spatial resolution of the feature maps by a factor of 2. Meanwhile, Fea CCC ( · ) denotes the specific procedure of the proposed Feature CCC, which will be elaborated on in Section 3.2.2.
Finally, following the encoder section, we introduce a 1 × 1 convolutional layer for the reconstruction of residual maps. This process can be mathematically represented as follows:
X r e s = Con v 1 × 1 ( D 3 )
where the symbol Con v 1 × 1 ( · ) represents a 1 × 1 convolutional layer, and X r e s H × W × C represents the residual image obtained after reconstruction.
The aforementioned details elucidate the design specifics of our proposed CCC-SSA-UNet backbone network, which draws inspiration from the design principles of DRUNet [55] and DIP-HyperKite [31] while featuring notable differences. CCC-SSA-UNet distinguishes itself from DRUNe in three key aspects. Firstly, our backbone network employs 2 × 2 maxpooling downsampling and bilinear interpolation upsampling, in contrast to DRUNet’s use of 2 × 2 stride convolution (SConv) and 2 × 2 transpose convolution (TConv). This design choice reduces the parameter count in the sampling layers. Secondly, CCC-SSA-UNet utilizes a single Conv Block in each encoder or decoder block, as opposed to DRUNet’s employment of four residual convolution blocks. This design decision reduces the complexity of the encoder–decoder network. Lastly, CCC-SSA-UNet places the skip connections before each downsampling layer of the encoder and after the corresponding upsampling layer of the decoder, in contrast to DRUNet’s placement after each downsampling layer and before each upsampling layer. This arrangement maximizes the preservation of extracted features by the encoder, mitigating spatial information loss resulting from immediate upsampling following downsampling.
The CCC-SSA-UNet and DIP-HyperKite [31] exhibit three notable differences. First, DIP-HyperKite employs an architecture that performs layer-wise upsampling followed by downsampling, while our CCC-SSA-UNet adheres to the U-Net [51] architecture, which performs downsampling followed by upsampling. This design choice results in reduced intermediate feature map size, decreased computational complexity, and minimized GPU memory consumption. Second, DIP-HyperKite incorporates a DIP-Net for upsampling preprocessing of LR-HSI, while our CCC-SSA-UNet directly employs bilinear interpolation for upsampling, significantly reducing the inference time of the network model. Last, DIP-HyperKite employs direct skip connections between each encoder and its corresponding decoder, whereas our CCC-SSA-UNet integrates the SSA-Net, leveraging spatial–spectral attention mechanisms, between each encoder and its corresponding decoder. This design choice further enhances the representation capability of spatial–spectral features.

3.2.2. CCC

To enhance the fusion capability of different information sources, we propose a novel channel cross-concatenation method, referred to as CCC, which is shown in Figure 3. Based on the nature of the input sources, CCC can be further divided into two categories: Input CCC and Feature CCC. Input CCC refers to the channel cross-concatenation method between the HSI and PAN inputs, aiming to enhance the fusion capability of different input source images. On the other hand, Feature CCC refers to the channel cross-concatenation method between two feature maps, aiming to enhance the fusion capability between feature maps at different levels.
  • Input CCC
Figure 3a illustrates the working principle of Input CCC. The input consists of two tensors corresponding to different source images, namely Up-HSI and PAN. The output is the tensor obtained by performing channel cross-concatenation.
First, the Up-HSI is divided into m parts along the channel dimension, ensuring that the first m 1 parts have the same number of channels, and the number of channels in the m-th part should not exceed that of each previous part. Specifically, we set:
C = i = 1 m C i = C 1 + + C m 1 + C m
where C 1 = = C m 1 C m when m 2 . Particularly, C 1 = C when m = 1 , which signifies that no splitting is performed on Up-HSI.
The aforementioned splitting process can be represented by the following formula:
U 1 , U 2 , , U m 1 , U m = Split ( U )
where U H × W × C denotes the tensor corresponding to Up-HSI, while Split ( · ) signifies the operation of tensor splitting. Notably, U i H × W × C i ( 1 i m ) denotes the individual sub-tensors obtained through the splitting of U .
In the subsequent steps, the sub-tensor U i obtained from the splitting, with a channel number of C i , is sequentially inserted after the PAN tensor P , which has a channel number of C p = 1 . By concatenating them along the channel dimension, the resulting tensor O is obtained, with the number of channels in the output tensor O being determined as follows:
C O = i = 1 m ( C i + C p )   = C 1 + C p + + C m + C p   = C + m · C p   = C + m
The process of channel cross-concatenation mentioned above can be expressed mathematically as follows:
O = Concat ( U 1 , P , U 2 , P , , U m , P )
where P H × W × C p represents the tensor corresponding to PAN, while Concat ( · ) denotes the operation of channel concatenation. O H × W × C O signifies the resulting tensor.
In summary, the entire process of channel-wise cross-concatenation between Up-HSI and PAN can be represented by the following equation:
O = InputCCC ( U , P )
where InputCCC ( · ) represents the operation process of Input CCC.
  • Feature CCC
Figure 3b illustrates the working principle of Feature CCC. The input consists of two feature maps, Feature 1 and Feature 2, extracted from different levels of the UNet architecture. The output is a tensor obtained by cross-concatenating the two feature maps along the channel dimension.
First, Feature 1 and Feature 2 are split into n = 2 k ( 0 k 5 ) equal parts along the channel dimension, ensuring each sub-tensor has the same number of channels. Mathematically, we can express this as:
C q i = C q n , 1 i n C r j = C r n , 1 j n
Specifically, when n = 1 , indicating k = 0 , it represents no split operation is performed on Feature 1 and Feature 2. The aforementioned split process can be represented by the following formula:
Q 1 , Q 2 , , Q n = Split ( Q ) R 1 , R 2 , , R n = Split ( R )
where Q H × W × C q represents the tensor corresponding to Feature 1, R H × W × C r represents the tensor corresponding to Feature 2, Split ( ) represents the tensor split operation, Q i H × W × C q i ( 1 i n ) represents the sub-tensors obtained by splitting Q , and R j H × W × C r j ( 1 j n ) represents the sub-tensors obtained by splitting R .
The channel number of the output tensor O can be produced by inserting a sub-tensor R j with C r j channels after a sub-tensor Q i with C q i channels obtained from splitting, and then sequentially concatenating them along the channel dimension. The channel number of the output tensor O can be determined using the following formula:
C o = i = 1 j = 1 n ( C q i + C r j )   = C q 1 + C r 1 + + C q n + C r n   = C q + C r
The process of channel cross-concatenation mentioned above can be expressed mathematically as follows:
O = Concat ( Q 1 , R 1 , Q 2 , R 2 , , Q n , R n )
where Concat ( · ) denotes the operation of channel concatenation, while O H × W × C o signifies the resulting tensor.
In summary, the entire process of channel-wise cross-concatenation between Feature 1 and Feature 2 can be represented by the following equation:
O = Fea CCC ( Q , R )
where Fea CCC ( · ) represents the operation process of Feature CCC.
There are two main differences between Feature CCC and Input CCC. First, in Feature CCC, the channel numbers of input tensors Feature 1 and Feature 2 can be evenly divided by n , while in Input CCC, the channel number of the input tensor Up-HSI may not be evenly divisible by m . Second, in Feature CCC, the second input tensor Feature 2 needs to be split and inserted into each sub-tensor of Feature 1, while in Input CCC, the second input tensor PAN, having only one channel, cannot be split further. Instead, it is duplicated m times and then inserted into each sub-tensor of Up-HSI.

3.2.3. SSA-Net

In order to further enhance the expression capability of spatial–spectral features in the high-dimensional feature maps of CCC-SSA-UNet, we introduce SSA-Net, which is based on the spatial–spectral attention mechanism. Inspired by the design principles of the DARN network in DHP-DARN [30], SSA-Net incorporates N sequentially stacked Res-SSA blocks to adaptively highlight important spectral and spatial feature information. It is important to note that SSA-Net is positioned between the encoder and decoder in our network architecture, with its input being the feature maps extracted by the encoder and its output being the feature maps to be reconstructed by the decoder. Therefore, SSA-Net omits the front-end feature extraction module and the back-end feature reconstruction module of the DARN network. The schematic diagram of SSA-Net is shown in Figure 4, and it can be formulated as follows:
F N = f Re s S S A B N ( F N 1 ) = f Re s S S A B N ( f Re s S S A B N 1 ( f Re s S S A B 1 ( F 0 ) ) ) = f S S A N e t ( F 0 )
where F 0 represents the input feature map of SSA-Net, corresponding to the encoder output feature maps E 1 , E 2 , and E 3 of CCC-SSA-UNet in Figure 2. F N represents the output feature map of SSA-Net, corresponding to the decoder input feature maps Q 1 , Q 2 , and Q 3 of CCC-SSA-UNet in Figure 2. F k ( 1 k N 1 ) represents the intermediate feature map of SSA-Net. f Re s S S A B k ( · ) , 1 k N represents the function representation of the i-th Res-SSA block.
Similar to the CSA ResBlock in DHP-DARN [30] and the DAU in MIRNet [56], the design principle of our Res-SSA block also incorporates channel attention and spatial attention into the basic residual module. This integration aims to improve both spatial–spectral feature representation and the stability of network training, while accelerating convergence speed. In the field of hyperspectral image processing, channel attention is often referred to as spectral attention. To differentiate it from CSA, we name the entire residual spatial–spectral attention module as the Res-SSA block. The bottom half of Figure 4 illustrates the network structure of the Res-SSA block. For the Nth Res-SSA block, its input is the feature map F N 1 . F N 1 goes through a 3 × 3 convolutional layer to reduce the number of channels to 64. It then undergoes a ReLU activation layer and another 3 × 3 convolutional layer to extract the feature map F U , which serves as the input to the attention modules. F U is divided into two paths: one path obtains the spectral mask M C A through the spectral attention module, which is multiplied element-wise with the feature map F U to obtain F C A ; the other path obtains the spatial mask M S A through the spatial attention module, which is multiplied element-wise with the feature map F U to obtain F S A . Finally, F C A and F S A are added element-wise and combined with the input F N 1 to obtain the output F N of the Res-SSA block. This can be expressed mathematically as:
F U = Con v 3 × 3 ( δ ( Con v 3 × 3 ( F N 1 ) ) )
F C A = F U M C A
F S A = F U M S A
F N = F C A + F S A + F N 1
where δ ( · ) represents the ReLU activation layer, Con v 3 × 3 ( · ) represents the 3 × 3 convolutional layer, M C A and M S A represent the spectral mask and spatial mask, respectively, and represents the element-wise product operation.
Specifically, the backbone of the spectral attention module consists of a global average pooling layer along the spatial dimension, a 1 × 1 convolutional layer that is employed to reduce the number of channels from 64 to 64/r, a ReLU activation layer, a 1 × 1 convolutional layer that is employed to expand the number of channels from 64/r to 64, and a sigmoid activation layer in sequence. Here, r is referred to as the channel reduction ratio, which can be used to decrease the computational complexity of the network model. The backbone of the spatial attention module consists of a parallel arrangement of global average pooling and global maxpooling layers, a 1 × 1 convolutional layer, and a sigmoid activation layer in sequence. This can be expressed mathematically as:
M C A = σ ( Con v 1 × 1 ( δ ( Con v 1 × 1 ( GAP ( F U ) ) ) ) )
M S A = σ ( Con v 1 × 1 ( Contact ( GAP ( F U ) , GMP ( F U ) ) ) )
where σ ( · ) represents the sigmoid activation layer, Con v 1 × 1 ( · ) represents the 1 × 1 convolutional layer, GAP ( · ) stands for Global Average Pooling operation, GMP ( · ) stands for Global Maximum Pooling operation, and Concat ( · ) represents the operation of channel concatenation.
The spectral attention module (CA) filters out spectral information in the feature tensor that is less important for the fusion results, allowing the network to adaptively select crucial spectral information. The spatial attention module (SA) enables the network to focus more on the features in regions closely related to enhancing spatial details in the hyperspectral image. By combining channel attention with spatial attention and embedding them into the basic residual module, both the spatial–spectral feature representation capability of the network and the stability of network training are enhanced.

3.3. Loss Function

In order to improve the fidelity of the fused reconstruction results to the ground truth HSIs, common loss functions used in the current literature include 1 loss [30,31,48,57], 2 loss [58,59], perceptual loss [34], and adversarial loss [60]. Perceptual and adversarial losses are capable of restoring details that do not exist in the original image, which may not be desirable in the field of remote sensing. Conversely, 1 and 2 losses are considered more reliable [57]. 2 loss tends to penalize larger errors while disregarding smaller errors, which can result in networks utilizing 2 loss producing slightly blurred reconstructions [57,60]. On the other hand, 1 loss effectively penalizes smaller errors and promotes better convergence during training. Therefore, we employ 1 loss to evaluate the fusion performance of our network. Specifically, the mean absolute error (MAE) between all reconstructed images in a training batch and the reference HSIs is used to define the 1 loss, which can be mathematically expressed using the equation provided below:
1 ( Θ ) = 1 D d = 1 D Y ^ d Y d 1 = 1 D d = 1 D Φ ( X d , P d ; Θ ) Y d 1
where Y ^ d and Y d correspond to the d-th reconstructed HR-HSI and the reference HSI (GT), respectively. D represents the batch size, indicating the number of images included in a training batch. Θ encompasses all the parameters within the network. Φ ( · , · ; Θ ) signifies the comprehensive hyperspectral pansharpening neural network model proposed in this paper. Finally, 1 denotes the utilization of the 1 norm as a mathematical measure.

4. Experiments and Discussion

4.1. Datasets

In order to assess the effectiveness of the proposed hyperspectral pansharpening algorithm in this study, several hyperspectral image datasets were utilized for the experiments. These datasets include:
  • Pavia University Dataset [61]: The Pavia University dataset comprises aerial images acquired over Pavia University in Italy, utilizing the Reflective Optics System Imaging Spectrometer (ROSIS). The original image has a spatial resolution of 1.3 m and dimensions of 610 × 610 pixels. The ROSIS sensor captures 115 spectral bands, covering the spectral range of 430–860 nm. After excluding noisy bands, the image dataset contains 103 spectral bands. To remove uninformative regions, the right-side portion of the image was cropped, leaving a 610 × 340 pixel area for further analysis. Subsequently, a non-overlapping region of 576 × 288 pixels, situated at the top-left corner, was extracted and divided into 18 sub-images measuring 96 × 96 pixels each. These sub-images constitute the reference HR-HSI dataset, serving as the ground truth. To generate corresponding PAN and LR-HSI, the Wald protocol [62] was employed. Specifically, a Gaussian filter with an 8 × 8 kernel size was applied to blur the HR-HSI, followed by a downsampling process, reducing its spatial dimensions by a factor of four to obtain the LR-HSI. The PAN was created by computing the average of the first 100 spectral bands of the HR-HSI. Fourteen image pairs were randomly selected for the training set, while the remaining four pairs were reserved for the test set.
  • Pavia Centre Dataset [61]: The Pavia Centre dataset consists of aerial images captured over the city center of Pavia, located in northern Italy, using the Reflective Optics System Imaging Spectrometer (ROSIS). The original image has dimensions of 1096 × 1096 pixels and a spatial resolution of 1.3 m, similar to the Pavia University dataset. After excluding 13 noisy bands, the dataset contains 102 spectral bands, covering the spectral range of 430–860 nm. Due to the lack of informative content in the central region of the image, this portion was cropped, and only the remaining 1096 × 715 pixel area containing the relevant information was used for analysis. Subsequently, a non-overlapping region of 960 × 640 pixels, situated at the top-left corner, was extracted and divided into 24 sub-images measuring 160 × 160 pixels each. These sub-images constitute the reference HR-HSI dataset, serving as the ground truth. Similar to the Pavia University Dataset, the PAN and LR-HSI corresponding to the HR-HSI were generated using the same methodology. Eighteen image pairs were randomly selected as the training set, while the remaining seven pairs were designated as the test set.
  • Chikusei Dataset [63]: The Chikusei dataset comprises aerial images captured over the agricultural and urban areas of Chikusei, Japan, in 2014, using the Headwall Hyperspec-VNIR-C sensor. The original image has pixel dimensions of 2517 × 2355 and a spatial resolution of 2.5 m. It encompasses a total of 128 spectral bands, covering the spectral range of 363–1018 nm. For the experiments, a non-overlapping region of 2304 × 2304 pixels was selected from the top-left corner and divided into 81 sub-images of 256 × 256 pixels. These sub-images constitute the reference HR-HSI dataset, serving as the ground truth. Similar to the Pavia University dataset, LR-HSI corresponding to the HR-HSI were generated using the same method. The PAN image was obtained by averaging the spectral bands from 60 to 100 of the HR-HSI. Sixty-one image pairs were randomly selected as the training set, while the remaining 20 pairs were allocated to the test set.
Consistent with previous studies [30,31], the standard deviation ( σ ) of the Gaussian filter employed for LR-HSI generation is determined through the following formula:
σ = β 2 2 × 2.7725887
Here, β represents the downsampling scale factor used during the dataset generation process, indicating the linear spatial resolution ratio between the reference HR-HSI and the generated LR-HSI. In this study, the value of β is set to four.

4.2. Evaluation Metrics

To quantitatively assess the performance of our proposed method, we employed five widely used evaluation metrics: correlation coefficient (CC), spectral angle mapping (SAM) [64], root-mean-square error (RMSE), Erreur Relative Globale Adimensionnelle De Synthsès (ERGAS) [65], and peak signal-to-noise ratio (PSNR).
The CC metric provides insight into the geometric distortion present in the images, with values ranging from zero to one. RMSE measures the intensity differences between the super-resolved reconstruction and the ground truth. SAM evaluates the spectral fidelity of the reconstructed images. ERGAS assesses the overall quality of the generated images. PSNR serves as an important indicator of image reconstruction quality, and it is directly related to RMSE. For RMSE, SAM, and ERGAS, lower values indicate superior reconstruction quality. Conversely, higher values of CC and PSNR signify improved image quality, with an ideal value of one for CC and positive infinity for PSNR.

4.3. Implementation Details

The CCC-SSA-UNet proposed in this study comprises a four-level UNet architecture. Each level is characterized by the number of channels, represented as C f 0 , C f 1 , C f 2 , and C f 2 , ranging from the first to the fourth level, respectively. For the larger model, CCC-SSA-UNet-L, the channel values of C f 0 , C f 1 , and C f 2 are assigned as 32, 64, and 128, respectively. Conversely, for the smaller model, CCC-SSA-UNet-S, the channel values of C f 0 , C f 1 , and C f 2 are set as 32, 32, and 32, respectively. The SSA-Net module within CCC-SSA-UNet consists of ten sequential Res-SSA blocks. Each Res-SSA block employs a channel reduction factor of 16 within its spectral attention module. In the Input CCC module, the input tensor, Up-HSI, is partitioned into eight segments, while in the Feature CCC module, the input tensors, Feature 1 and Feature 2, are equally divided into eight partitions.
We set the batch size to four and employed the Adam [66] optimizer for training, with β1 = 0.9 and β2 = 0.999 as hyperparameters. The initial learning rate was set to 0.001. Specifically, for the Chikusei dataset, the learning rate was halved every 1000 epochs, whereas for the Pavia University dataset and Pavia Centre dataset, the learning rate was reduced by half every 2000 epochs. To optimize the model, the 1 loss function was utilized, and a total of 10,500 epochs were conducted. Our model was implemented using the PyTorch framework, and the training process was executed on a single GeForce RTX 3090 GPU. The training duration was approximately 40 h for the Chikusei dataset, 1.5 h for the Pavia University dataset, and 4 h for the Pavia Centre dataset.

4.4. Comparison with State-of-the-Art Methods

In order to demonstrate the effectiveness, efficiency, and state-of-the-art performance of the proposed CCC-SSA-UNet, comparative experiments were conducted on three datasets: Pavia University dataset, Pavia Centre dataset, and Chikusei dataset. Our method was compared against ten traditional pansharpening methods and five state-of-the-art deep learning-based methods. The traditional methods included in the comparison were GS [5], GSA [6], PCA [8], GFPCA [7], BayesNaive [14], BayesSparse [44], MTF-GLP [9], MTF-GLP-HPM [11], CNMF [16], and HySure [12]. The deep learning-based methods consisted of HyperPNN1 [28], HyperPNN2 [28], DHP-DARN [30], DIP-HyperKite [31], and Hyper-DSNet [29]. The traditional methods were implemented using the open-source MATLAB toolbox provided by Loncan et al. [67]. For the deep learning methods, we reproduced the experiments on our computer following the original papers’ descriptions and parameter settings, presenting the best results obtained. Notably, all evaluation metrics for the test datasets were recalculated using MATLAB to ensure a fair comparison between traditional and deep learning methods. During the calculations, the reconstructed images were normalized concerning the reference images. The following sections present the detailed comparative experimental results of the different pansharpening methods on the three datasets.

4.4.1. Experiments on Pavia University Dataset

We compared our proposed CCC-SSA-UNet with 15 other methods on the test set of the Pavia University dataset. The quantitative evaluation results are presented in Table 1, with the best values highlighted in red, the second-best values in blue, and the third-best values in green. It is evident that deep learning-based methods outperform traditional methods across various objective metrics. Among the traditional methods, HySure achieves the best performance, with a low SAM of 5.673 and a high PSNR of 32.663. However, there still exists a considerable gap compared to deep learning methods. Among the other five deep learning methods used for comparison, DHP-DARN delivers the best results, possibly due to its utilization of the spatial–spectral dual attention mechanism in the residual blocks. Our proposed CCC-SSA-UNet-S and CCC-SSA-UNet-L benefit from the fusion capability of the CCC operation and the feature extraction capability of the SSA-Net. They significantly outperform all other comparative methods across the objective metrics. Specifically, CCC-SSA-UNet-S improves upon the state-of-the-art comparative methods by 0.002 in CC, 1.675 in RSNR, and 0.786 in PSNR, while reducing SAM by 0.276, RMSE by 0.0014, and ERGAS by 0.176. CCC-SSA-UNet-L achieves improvements of 0.003 in CC, 2.002 in RSNR, and 0.928 in PSNR, along with reductions of 0.321 in SAM, 0.0016 in RMSE, and 0.204 in ERGAS, compared to the state-of-the-art comparative methods.
In addition to the aforementioned quantitative comparison results, we also present the visual results of various pansharpening methods on a randomly selected image patch (10th patch) from the test subset of the Pavia University dataset in Figure 5. To better showcase the reconstruction results, the regions of interest (ROI) in each image are magnified and highlighted with yellow rectangular boxes. Furthermore, in Figure 6, we display the mean absolute error (MAE) maps, which illustrate the differences between the reconstructed images and the reference images for the 10th patch, generated by each method. It can be observed that the images reconstructed by traditional methods are relatively blurry, with larger average absolute errors, resulting in poorer visual quality. In contrast, deep learning-based methods benefit from the powerful learning capabilities of deep neural networks, resulting in less blurriness in the reconstructed images and smaller average absolute errors, indicating better visual quality. Among them, our proposed CCC-SSA-UNet-S and CCC-SSA-UNet-L demonstrate the closest resemblance to the reference images, demonstrating their outstanding ability to maintain spectral fidelity and restore precise spatial details.

4.4.2. Experiments on Pavia Centre Dataset

We also compared our proposed CCC-SSA-UNet with 15 other methods on the test set of the Pavia Centre dataset, and the quantitative evaluation results are presented in Table 2. Among the traditional methods, HySure still achieved the best performance, with a SAM as low as 6.723 and a PSNR as high as 34.444, but there is still a significant gap compared to the deep learning methods. Consistent with the results on the Pavia University dataset, the deep learning-based methods outperformed the traditional methods across all objective metrics. Among the other five deep learning methods used for comparison, Hyper-DSNe achieved the best results. This may be attributed to the fact that the Pavia Centre dataset contains more high-frequency information than the Pavia University dataset, and Hyper-DSNet, with its five types of high-pass filter templates serving as multi-detail extractors, is more effective in recovering high-frequency details.
Our proposed CCC-SSA-UNet-S and CCC-SSA-UNet-L benefit from the fusion ability of the CCC and the feature extraction capability of the SSA-Net, demonstrating significant improvements over all other comparison methods in terms of various objective metrics. Specifically, CCC-SSA-UNet-S outperformed the state-of-the-art methods by 0.003 in CC, 1.905 in RSNR, and 0.868 in PSNR, while reducing SAM, RMSE, and ERGAS by 0.284, 0.0013, and 0.190, respectively. Similarly, CCC-SSA-UNet-L achieved improvements of 0.005 in CC, 1.895 in RSNR, and 0.873 in PSNR, along with reductions of 0.295, 0.0026, and 0.194 in SAM, RMSE, and ERGAS, respectively, compared to the state-of-the-art methods.
In addition to quantitative comparison results, Figure 7 presents the visual results of various pan-sharpening methods on randomly selected image patches (the 15th patch) from the test set of the Pavia Centre dataset. Figure 8 displays the mean absolute error (MAE) maps between the reconstructed images by different methods and the reference images for the 15th patch. Clearly, the images reconstructed by traditional methods appear blurrier, with larger mean absolute errors and lower visual quality. In contrast, the images reconstructed by deep learning-based methods exhibit less blurriness, smaller mean absolute errors, and higher visual quality. Notably, our proposed CCC-SSA-UNet-S and CCC-SSA-UNet-L demonstrate the closest resemblance to the reference images, providing evidence for the effectiveness and advancement of our proposed approach.

4.4.3. Experiments on Chikusei Dataset

We also compared our proposed CCC-SSA-UNet with 15 other methods on the Chikusei dataset. Table 3 presents the average quantitative evaluation results on the test set. Among the traditional methods, HySure achieved the best performance, with a SAM as low as 3.139 and a PSNR as high as 39.615, surpassing the other nine classical methods by a significant margin. Among the five other deep learning methods used for comparison, Hyper-DSNet achieved the best results, consistent with the findings on the Pavia Centre dataset. Our proposed CCC-SSA-UNet-S and CCC-SSA-UNet-L outperformed Hyper-DSNet in terms of SAM, RSNR, and PSNR metrics, while being comparable in terms of CC and RMSE metrics. Specifically, CCC-SSA-UNet-S improved RSNR and PSNR by 0.116 and 0.047, respectively, and reduced SAM by 0.012. CCC-SSA-UNet-L improved RSNR and PSNR by 0.176 and 0.076, respectively, and reduced SAM by 0.011.
In addition to the quantitative comparison results mentioned above, Figure 9 showcases the visual results of various pansharpening methods on the randomly selected 31st image patch from the Chikusei dataset’s test set. Figure 10 presents the average absolute error map (MAE map) between the reconstructed images by different methods and the reference images for the 31st patch. It is evident that the images reconstructed by the first nine traditional methods appear blurry with larger average absolute errors, indicating poorer visual quality. Particularly, the reconstructed image by MTF-GLP-HP exhibits numerous red spots, indicating significant spectral distortion. In contrast, HySure and the deep learning-based methods produce less blurry reconstructed images with smaller average absolute errors, demonstrating better visual quality. Among them, our proposed CCC-SSA-UNet-S and CCC-SSA-UNet-L show the closest resemblance to the reference images and exhibit the lightest colors on the MAE map, further confirming the effectiveness and superiority of our approach.

4.5. Analysis of the Computational Complexity

Table 4 presents a comparison of the computational complexities among different pansharpening methods on the Pavia University dataset. The metrics PSNR and SAM are representative indicators used to evaluate the quality of the network’s reconstructed images, while the number of parameters (#Params), multiply accumulate operations (MACs), floating-point operations (FLOPs), GPU memory usage (GPU Memory), and average inference runtime (Runtime) are employed to assess the computational complexity of the neural network. From Table 4, it can be observed that our CCC-SSA-UNet-S achieves leading image reconstruction performance while maintaining a smaller #Params, MACs, FLOPs, and GPU memory usage. In comparison, CCC-SSA-UNet-L demonstrates a five-fold increase in #Params compared to CCC-SSA-UNet-S, while MACs and FLOPs only double. Moreover, GPU memory usage and runtime remain relatively unchanged. Remarkably, CCC-SSA-UNet-L achieves optimal performance in terms of image quality. CCC-SSA-UNet-S and CCC-SSA-UNet-L demonstrate superior performance compared to other methods, as evidenced by extensive experiments conducted on publicly available datasets. Our models effectively leverage multiscale image feature information for fusion reconstruction while maintaining lower memory usage. Furthermore, they achieve shorter inference runtime while ensuring fusion quality.
Figure 11 visually illustrates the balance and superiority of our network in terms of performance and computational complexity. Figure 11a illustrates a comparison of our method with the current state-of-the-art methods in terms of PSNR, FLOPs, and GPU memory usage on the Pavia University dataset. Figure 11b presents a comparison of SAM, FLOPs, and GPU memory usage. It can be observed that our method achieves superior fusion quality with lower GPU memory consumption than the three state-of-the-art deep learning-based pansharpening methods. Figure 11c and Figure 11d, respectively demonstrate the comparisons of PSNR versus runtime and SAM versus runtime for different methods on the test set of the Pavia University dataset. It can be observed that our CCC-SSA-UNet achieves the highest image reconstruction performance while surpassing the majority of existing pansharpening methods in terms of inference runtime. This solidly demonstrates the effectiveness, advancement, and efficiency of our proposed approach.

4.6. Sensitivity Analysis of the Network Parameters

In order to select the optimal network parameters, we conducted extensive experiments and conducted detailed research on the number of Filter Channels, Input CCC groups, Feature CCC groups, SSA blocks, and the choice of initial learning rate. The following sections will provide a detailed description of each aspect.

4.6.1. Analysis of the Filter Channel Numbers

As described in Section 3.2.1, the proposed CCC-SSA-UNet adopts a four-layer UNet architecture with varying numbers of channels in each layer, denoted as C f 0 , C f 1 , C f 2 , and C f 2 , which can be referred to as Filter Channels. Clearly, the sizes of Filter Channels will have an impact on the model’s computational complexity and performance. Therefore, we conducted comparative experiments on the Pavia University dataset to investigate the influence of channel numbers in each layer on the model. As shown in Table 5, while keeping other parameters consistent, we created different models by setting the channel numbers in each layer. Specifically, Model 1 has all channels set to 32, Model 2 has all channels set to 64, Model 3 has all channels set to 128, Model 4 has C f 0 , C f 1 , and C f 2 set to 128, 64, and 32, respectively, and Model 5 has C f 0 , C f 1 , and C f 2 set to 32, 64, and 128, respectively. The experimental results indicate that Model 5 achieved the best performance, which corresponds to the CCC-SSA-UNet-L model mentioned in Section 4.3. Considering that Model 5 has slightly higher computational complexity, we also selected Model 1, which has comparable performance with the smallest #Params and MACs, as an alternative model. This corresponds to the CCC-SSA-UNet-S model mentioned in Section 4.3.

4.6.2. Analysis of the Input CCC Group Numbers

The number of groups, denoted as m, in the Input CCC operation is another hyperparameter of the network. To determine the optimal number of groups, we conducted comparative experiments on the Pavia University dataset to evaluate the impact of the group number on the performance of CCC-SSA-UNet-L. By changing the value of m while keeping other parameters constant, the experimental results are shown in Table 6. It can be observed that when m is set to 8, the network achieves optimal performance in all evaluation metrics. Compared to the base model with m = 1, setting m to 8 only increases the #Params by 0.002 M, MACs by 0.018 G, and Runtime by 0.5 ms. This indicates that the Input CCC operation can effectively enhance the fusion capability of different input source images with minimal increases in parameters and computational complexity.

4.6.3. Analysis of the Feature CCC Group Numbers

Similar to the Input CCC operation, the number of groups, denoted as n, in the Feature CCC operation is also one of the network’s hyperparameters. We conducted comparative experiments on the Pavia University dataset to evaluate the impact of the group number in Feature CCC on the performance of CCC-SSA-UNet-L. By changing the value of n while keeping other parameters constant, the experimental results are shown in Table 7. It can be observed that when n is set to 8, the network achieves optimal performance in all evaluation metrics. Compared to the base model with n = 1, setting n to 8 does not increase the #Params or MACs. Moreover, the Runtime decreases by 0.2 ms. This indicates that the Feature CCC operation can effectively enhance the fusion capability between different levels of feature maps without adding any parameters or computational complexity.

4.6.4. Analysis of the SSA Block Numbers

The SSA Block enhances the spatial–spectral feature representation of the network by combining channel attention and spatial attention and embedding them into the basic residual module. Multiple SSA blocks are concatenated to form the SSA-Net. Therefore, the number of SSA blocks, denoted as N, affects the performance of the SSA-Net. To investigate how the number of SSA blocks influences the network’s performance, we constructed several variants of SSA-Net, each containing a different number of SSA blocks while keeping other settings the same. Table 8 presents the comparative experiments conducted on the Pavia University dataset with these 12 SSA-Net variants.
The results show that, initially, as N increases, the network’s performance improves. When N reaches 10, the network achieves its maximum performance. However, beyond N = 10, as the network deepens, the computational complexity increases significantly without a substantial performance improvement. In fact, the performance may even degrade. Considering both performance and computational complexity, we set the number of SSA blocks to 10 in CCC-SSA-UNet.

4.6.5. Analysis of the Learning Rate

The learning rate is one of the most important hyperparameters in neural networks. Table 9 presents the performance of the CCC-SSA-UNet-L on the Pavia University dataset with different initial learning rates. It can be observed that the network achieves the best results when the initial learning rate is set to 0.001 and undergoes a halving decay every 2000 epochs.

4.7. Ablation Study

In this section, we conducted detailed ablation experiments to validate the effectiveness of the proposed Input CCC, Feature CCC, SSA-Net, and Res-SSA block. We constructed several variants of the CCC-SSA-UNet-L network, labeled as model 1 to model 8, each variant incorporating different combinations of the Input CCC, Feature CCC, and SSA-Net modules. Specifically, model 1 did not use any of the modules and employed regular channel connections instead of Input CCC, and used skip connections instead of SSA-Net. Models 2 to 4 utilized only one of the modules, while models 5 to 7 excluded one of the modules. Model 8 incorporated all three modules simultaneously. The quantitative results of the ablation experiments conducted on the Pavia University dataset are presented in Table 10. It can be observed that model 8, which incorporates Input CCC, Feature CCC, and SSA-Net, achieved the best performance. A detailed analysis of the effectiveness of each submodule will be discussed in the following subsections.

4.7.1. Effect of the Proposed Input CCC

From Table 10, it is evident that when comparing model 2 to model 1, the inclusion of the Input CCC module in the structure resulted in improvements in the CC, RSNR, and PSNR metrics by 0.001, 0.289, and 0.207, respectively. Additionally, the SAM, RMSE, and ERGAS metrics decreased by 0.037, 0.0004, and 0.044, respectively. Moreover, for model 8 compared to model 5, the addition of the Input CCC module to the structure led to an increase in RSNR and PSNR metrics by 0.061 and 0.047, respectively, while the SAM and ERGAS metrics decreased by 0.020 and 0.012, respectively. These findings strongly demonstrate the effectiveness of the proposed Input CCC method.

4.7.2. Effect of the Proposed Feature CCC

From Table 10, it can be observed that model 3, which incorporates the Feature CCC module into the structure of model 1, showed improvements in the CC, RSNR, and PSNR metrics by 0.001, 0.500, and 0.093, respectively. Moreover, the SAM, RMSE, and ERGAS metrics decreased by 0.047, 0.0003, and 0.024, respectively. On the other hand, model 8, which incorporates the Feature CCC module into the structure of model 6, demonstrated increases in the RSNR and PSNR metrics by 0.184 and 0.116, respectively, while the SAM, RMSE, and ERGAS metrics decreased by 0.036, 0.0002, and 0.037, respectively. These results provide strong evidence for the effectiveness of our proposed Feature CCC method.

4.7.3. Effect of the Proposed SSA-Net

Table 10 also provides evidence for the effectiveness of our proposed SSA-Net. When SSA-Net was added to the structure of model 1, significant performance improvements were observed. Model 4, which incorporated SSA-Net, exhibited notable improvements in CC, RSNR, and PSNR metrics by 0.005, 3.386, and 1.426, respectively. Additionally, SAM, RMSE, and ERGAS metrics decreased by 0.540, 0.0028, and 0.335, respectively. Similarly, for model 8, which incorporated SSA-Net in the structure of model 7, improvements were observed in CC, RSNR, and PSNR metrics by 0.005, 3.079, and 1.309, respectively. Furthermore, SAM, RMSE, and ERGAS metrics decreased by 0.492, 0.0025, and 0.309, respectively.

4.7.4. Effect of the Proposed Res-SSA Block

SSA-Net is composed of multiple Res-SSA blocks connected in series. To demonstrate the effectiveness of the proposed Res-SSA block, we designed several variants of attention modules for comparison in Figure 12. (a) Residual block baseline, which utilizes the basic residual module only. (b) CA, which employs the spectral attention module from the Res-SSA block. (c) SA, which utilizes the spatial attention module from the Res-SSA block. (d) CSA, which adopts the channel-spatial attention module from DHP-DARN [30]. (e) DAU, which incorporates the dual attention module from MIRNet [56]. Table 11 presents the quantitative comparison results of CCC-SSA-UNet-L with different attention modules on the Pavia University dataset. It can be observed that using CA or SA alone had a negative impact on the network performance. However, CSA showed performance improvement compared to the baseline residual block, while DAU did not demonstrate significant improvement in network performance. The experimental results indicate that our Res-SSA block achieved the best performance while having fewer parameters and MACs compared to CSA and DAU, suggesting a good balance between image reconstruction performance and computational complexity.

5. Conclusions

This paper has proposed a novel U-shaped hyperspectral pansharpening network CCC-SSA-UNet for hyperspectral image super-resolution. The channel cross-concatenation mechanism and the spatial–spectral attention mechanism have been incorporated in UNet, which effectively enhanced the network’s ability to extract spatial and spectral features. In detail, a novel Input CCC method at the network entrance and a novel Feature CCC method within the decoder have been proposed, which effectively enhanced the fusion capability of different input source images and facilitated the fusion of features at different levels without introducing additional parameters, respectively. Furthermore, the effectiveness of SSA-Net, which is composed of Res-SSA blocks, has been demonstrated by comparing it with other attention module variants. Furthermore, an ablation study has been performed to verify the effectiveness of each module proposed in our framework. By conducting comparative experiments with ten traditional pansharpening methods and five state-of-the-art deep learning-based methods on three datasets, the effectiveness, efficiency, and advancement of our proposed CCC-SSA-UNet have been proven. The results show a satisfactory performance of CCC-SSA-UNet and its superiority over the reference methods.
Although our method has achieved a state-of-the-art performance, there are some limitations and room for improvement too. First, the number of spectral bands varies across different datasets, which affects the equal partitioning of the input tensor Up-HSI in the Input CCC module and limits the flexibility in choosing specific values. In the future, we will consider adding a convolutional layer before the Input CCC module to adjust the channel number of the input tensor. Second, both the encoder and decoder in our network are composed of simple Conv Block modules, which have limited feature extraction capability. In the future, we plan to incorporate lightweight Transformer modules with global self-attention into the encoder and decoder to enhance the network’s ability to extract global features. The last problem is the scarcity of training data, which causes a limited generalization ability and poor performance on off-training test images. In the future, we will conduct in-depth research on unsupervised pansharpening methods to promote their application in practical scenarios.

Author Contributions

Conceptualization, Z.L.; validation, Z.L. and G.H.; formal analysis, Z.L.; investigation, Z.L., D.L. and G.H.; data curation, Z.L., A.D. and D.L.; original draft preparation, Z.L.; review and editing, Z.L., H.Y., P.L., D.C., D.L., A.D. and G.H.; Supervision, G.H., H.Y., P.L. and D.C.; funding acquisition, G.H. and D.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Development Project of Jilin Province, Key R&D Programs No. 20210201132GX and No. 20210201078GX.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the editors and reviewers for their insightful suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Audebert, N.; Saux, B.L.; Lefevre, S. Deep Learning for Classification of Hyperspectral Data: A Comparative Review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef]
  2. Fabelo, H.; Ortega, S.; Ravi, D.; Kiran, B.R.; Sosa, C.; Bulters, D.; Callicó, G.M.; Bulstrode, H.; Szolna, A.; Piñeiro, J.F.; et al. Spatio-spectral classification of hyperspectral images for brain cancer detection during surgical operations. PLoS ONE 2018, 13, e0193721. [Google Scholar] [CrossRef] [PubMed]
  3. Xie, W.; Lei, J.; Fang, S.; Li, Y.; Jia, X.; Li, M. Dual feature extraction network for hyperspectral image analysis. Pattern Recognit. 2021, 118, 107992. [Google Scholar] [CrossRef]
  4. Zhang, M.; Sun, X.; Zhu, Q.; Zheng, G. A Survey of Hyperspectral Image Super-Resolution Technology. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 4476–4479. [Google Scholar]
  5. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  6. Aiazzi, B.; Baronti, S.; Selva, M. Improving Component Substitution Pansharpening through Multivariate Regression of MS + Pan Data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  7. Liao, W.; Huang, X.; Van Coillie, F.; Gautama, S.; Pižurica, A.; Philips, W.; Liu, H.; Zhu, T.; Shimoni, M.; Moser, G. Processing of multiresolution thermal hyperspectral and digital color data: Outcome of the 2014 IEEE GRSS data fusion contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2984–2996. [Google Scholar] [CrossRef]
  8. Shah, V.P.; Younan, N.H.; King, R.L. An Efficient Pan-Sharpening Method via a Combined Adaptive PCA Approach and Contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  9. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored multiscale fusion of high-resolution MS and Pan imagery. Photogramm. Eng. Remote Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  10. Liu, J. Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  11. Vivone, G.; Restaino, R.; Mura, M.D.; Licciardi, G.; Chanussot, J. Contrast and Error-Based Fusion Schemes for Multispectral Image Pansharpening. IEEE Geosci. Remote Sens. Lett. 2014, 11, 930–934. [Google Scholar] [CrossRef]
  12. Simões, M.; Bioucas-Dias, J.; Almeida, L.B.; Chanussot, J. A Convex Formulation for Hyperspectral Image Superresolution via Subspace-Based Regularization. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3373–3388. [Google Scholar] [CrossRef]
  13. Wei, Q.; Bioucas-Dias, J.; Dobigeon, N.; Tourneret, J.Y. Hyperspectral and Multispectral Image Fusion Based on a Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3658–3668. [Google Scholar] [CrossRef]
  14. Wei, Q.; Dobigeon, N.; Tourneret, J.Y. Bayesian Fusion of Multi-Band Images. IEEE J. Sel. Top. Signal Process. 2015, 9, 1117–1127. [Google Scholar] [CrossRef]
  15. Kawakami, R.; Matsushita, Y.; Wright, J.; Ben-Ezra, M.; Tai, Y.; Ikeuchi, K. High-resolution hyperspectral imaging via matrix factorization. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2329–2336. [Google Scholar]
  16. Yokoya, N.; Yairi, T.; Iwasaki, A. Coupled Nonnegative Matrix Factorization Unmixing for Hyperspectral and Multispectral Data Fusion. IEEE Trans. Geosci. Remote Sens. 2012, 50, 528–537. [Google Scholar] [CrossRef]
  17. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  18. Zhang, Z.; Lu, X.; Cao, G.; Yang, Y.; Jiao, L.; Liu, F. ViT-YOLO: Transformer-Based YOLO for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, BC, Canada, 11–17 October 2021; pp. 2799–2808. [Google Scholar]
  19. Cao, Z.; Fu, C.; Ye, J.; Li, B.; Li, Y. SiamAPN++: Siamese Attentional Aggregation Network for Real-Time UAV Tracking. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 3086–3092. [Google Scholar]
  20. Zhang, G.; Li, Z.; Li, J.; Hu, X. CFNet: Cascade Fusion Network for Dense Prediction. arXiv 2023, arXiv:2302.06052. [Google Scholar]
  21. Zhang, K.; Li, Y.; Liang, J.; Cao, J.; Zhang, Y.; Tang, H.; Timofte, R.; Van Gool, L. Practical Blind Denoising via Swin-Conv-UNet and Data Synthesis. arXiv 2022, arXiv:2203.13278. [Google Scholar]
  22. Cho, S.-J.; Ji, S.-W.; Hong, J.-P.; Jung, S.-W.; Ko, S.-J. Rethinking coarse-to-fine approach in single image deblurring. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 4641–4650. [Google Scholar]
  23. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Gool, L.V.; Timofte, R. SwinIR: Image Restoration Using Swin Transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  24. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  25. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  26. Yoo, J.; Kim, T.; Lee, S.; Kim, S.H.; Lee, H.; Kim, T.H. Enriched CNN-Transformer Feature Aggregation Networks for Super-Resolution. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 4956–4965. [Google Scholar]
  27. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  28. He, L.; Zhu, J.; Li, J.; Plaza, A.; Chanussot, J.; Li, B. HyperPNN: Hyperspectral Pansharpening via Spectrally Predictive Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3092–3100. [Google Scholar] [CrossRef]
  29. Zhuo, Y.W.; Zhang, T.J.; Hu, J.F.; Dou, H.X.; Huang, T.Z.; Deng, L.J. A Deep-Shallow Fusion Network With Multidetail Extractor and Spectral Attention for Hyperspectral Pansharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 7539–7555. [Google Scholar] [CrossRef]
  30. Zheng, Y.; Li, J.; Li, Y.; Guo, J.; Wu, X.; Chanussot, J. Hyperspectral Pansharpening Using Deep Prior and Dual Attention Residual Network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8059–8076. [Google Scholar] [CrossRef]
  31. Bandara, W.G.C.; Valanarasu, J.M.J.; Patel, V.M. Hyperspectral Pansharpening Based on Improved Deep Image Prior and Residual Reconstruction. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  32. Dong, W.; Hou, S.; Xiao, S.; Qu, J.; Du, Q.; Li, Y. Generative Dual-Adversarial Network With Spectral Fidelity and Spatial Enhancement for Hyperspectral Pansharpening. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 7303–7317. [Google Scholar] [CrossRef]
  33. Xie, W.; Cui, Y.; Li, Y.; Lei, J.; Du, Q.; Li, J. HPGAN: Hyperspectral Pansharpening Using 3-D Generative Adversarial Networks. IEEE Trans. Geosci. Remote Sens. 2021, 59, 463–477. [Google Scholar] [CrossRef]
  34. Bandara, W.G.C.; Patel, V.M. HyperTransformer: A Textural and Spectral Feature Fusion Transformer for Pansharpening. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1767–1777. [Google Scholar]
  35. He, L.; Xi, D.; Li, J.; Lai, H.; Plaza, A.; Chanussot, J. Dynamic Hyperspectral Pansharpening CNNs. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–19. [Google Scholar] [CrossRef]
  36. Luo, F.; Zhou, T.; Liu, J.; Guo, T.; Gong, X.; Ren, J. Multiscale Diff-Changed Feature Fusion Network for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5502713. [Google Scholar] [CrossRef]
  37. He, X.; Tang, C.; Liu, X.; Zhang, W.; Sun, K.; Xu, J. Object Detection in Hyperspectral Image via Unified Spectral-Spatial Feature Aggregation. arXiv 2023, arXiv:2306.08370. [Google Scholar]
  38. Kordi Ghasrodashti, E. Hyperspectral image classification using a spectral–spatial random walker method. Int. J. Remote Sens. 2019, 40, 3948–3967. [Google Scholar] [CrossRef]
  39. Chavez, P.; Sides, S.C.; Anderson, J.A. Comparison of three different methods to merge multiresolution and multispectral data-Landsat TM and SPOT panchromatic. Photogramm. Eng. Remote Sens. 1991, 57, 295–303. [Google Scholar]
  40. Burt, P.J.; Adelson, E.H. The Laplacian pyramid as a compact image code. In Readings in Computer Vision; Fischler, M.A., Firschein, O., Eds.; Morgan Kaufmann: San Francisco, CA, USA, 1987; pp. 671–679. [Google Scholar]
  41. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-driven fusion of high spatial and spectral resolution images based on oversampled multiresolution analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  42. Dong, W.; Zhang, T.; Qu, J.; Xiao, S.; Liang, J.; Li, Y. Laplacian Pyramid Dense Network for Hyperspectral Pansharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5507113. [Google Scholar] [CrossRef]
  43. Fasbender, D.; Radoux, J.; Bogaert, P. Bayesian Data Fusion for Adaptable Image Pansharpening. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1847–1857. [Google Scholar] [CrossRef]
  44. Wei, Q.; Dobigeon, N.; Tourneret, J. Fast Fusion of Multi-Band Images Based on Solving a Sylvester Equation. IEEE Trans. Image Process. 2015, 24, 4109–4121. [Google Scholar] [CrossRef]
  45. Xue, J.; Zhao, Y.Q.; Bu, Y.; Liao, W.; Chan, J.C.W.; Philips, W. Spatial-Spectral Structured Sparse Low-Rank Representation for Hyperspectral Image Super-Resolution. IEEE Trans. Image Process. 2021, 30, 3084–3097. [Google Scholar] [CrossRef] [PubMed]
  46. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef]
  47. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef]
  48. Liu, X.; Liu, Q.; Wang, Y. Remote sensing image fusion based on two-stream fusion network. Inf. Fusion 2020, 55, 1–15. [Google Scholar] [CrossRef]
  49. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  50. Roy, A.G.; Navab, N.; Wachinger, C. Concurrent spatial and channel ‘squeeze & excitation’ in fully convolutional networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; pp. 421–429. [Google Scholar]
  51. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  52. Bass, C.; Silva, M.d.; Sudre, C.; Williams, L.Z.J.; Sousa, H.S.; Tudosiu, P.D.; Alfaro-Almagro, F.; Fitzgibbon, S.P.; Glasser, M.F.; Smith, S.M.; et al. ICAM-Reg: Interpretable Classification and Regression With Feature Attribution for Mapping Neurological Phenotypes in Individual Scans. IEEE Trans. Med. Imaging 2023, 42, 959–970. [Google Scholar] [CrossRef]
  53. Kordi Ghasrodashti, E.; Sharma, N. Hyperspectral image classification using an extended Auto-Encoder method. Signal Process. Image Commun. 2021, 92, 116111. [Google Scholar] [CrossRef]
  54. Adkisson, M.; Kimmell, J.C.; Gupta, M.; Abdelsalam, M. Autoencoder-based Anomaly Detection in Smart Farming Ecosystem. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 3390–3399. [Google Scholar]
  55. Zhang, K.; Li, Y.; Zuo, W.; Zhang, L.; Gool, L.V.; Timofte, R. Plug-and-Play Image Restoration With Deep Denoiser Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 6360–6376. [Google Scholar] [CrossRef]
  56. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H.; Shao, L. Learning enriched features for real image restoration and enhancement. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 492–511. [Google Scholar]
  57. Jiang, J.; Sun, H.; Liu, X.; Ma, J. Learning Spatial-Spectral Prior for Super-Resolution of Hyperspectral Imagery. IEEE Trans. Comput. Imaging 2020, 6, 1082–1096. [Google Scholar] [CrossRef]
  58. Hu, J.F.; Huang, T.Z.; Deng, L.J.; Jiang, T.X.; Vivone, G.; Chanussot, J. Hyperspectral Image Super-Resolution via Deep Spatiospectral Attention Convolutional Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 7251–7265. [Google Scholar] [CrossRef]
  59. He, L.; Zhu, J.; Li, J.; Meng, D.; Chanussot, J.; Plaza, A. Spectral-Fidelity Convolutional Neural Networks for Hyperspectral Pansharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5898–5914. [Google Scholar] [CrossRef]
  60. Li, J.; Cui, R.; Li, B.; Song, R.; Li, Y.; Dai, Y.; Du, Q. Hyperspectral Image Super-Resolution by Band Attention through Adversarial Learning. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4304–4318. [Google Scholar] [CrossRef]
  61. Available online: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes#Pavia_University_scene (accessed on 1 August 2023).
  62. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  63. Yokoya, N.; Iwasaki, A. Airborne Hyperspectral Data over Chikusei; Technical Report SAL-2016-05-27; The University of Tokyo: Tokyo, Japan, 2016. [Google Scholar]
  64. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Proceedings of the JPL, Summaries of the Third Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; Volume 1. [Google Scholar]
  65. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion? In Proceedings of the Third Conference Fusion of Earth Data: Merging Point Measurements, Raster Maps and Remotely Sensed Images, Sophia Antipolis, France, 26–28 January 2000; pp. 99–103. [Google Scholar]
  66. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  67. Loncan, L.; de Almeida, L.B.; Bioucas-Dias, J.M.; Briottet, X.; Chanussot, J.; Dobigeon, N.; Fabre, S.; Liao, W.; Licciardi, G.A.; Simões, M.; et al. Hyperspectral Pansharpening: A Review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 27–46. [Google Scholar] [CrossRef]
Figure 1. The schematic diagram of the training phase and testing phase of the deep learning-based HSI-PAN fusion network. (a) Training phase; (b) testing phase.
Figure 1. The schematic diagram of the training phase and testing phase of the deep learning-based HSI-PAN fusion network. (a) Training phase; (b) testing phase.
Remotesensing 15 04328 g001
Figure 2. The architecture of the proposed CCC-SSA-UNet. CCC-SSA-UNet takes an LR-HSI and a PAN as input and takes an HR-HSI as output. It combines UNet and SSA-Net, exploits the “Conv Block” as the main building block of the UNet backbone, and adopts the Res-SSA block as the main building block of the SSA-Net. In each “Conv Block” of the UNet, the input is first passed through a 3 × 3 convolution and subsequently is passed through a batch normalization layer (BN) and a LeakyReLU layer. In this way, the feature map is extracted and passed to the next layer. Finally, the output of UNet is passed through a 1 × 1 convolution to produce the residual map of the Up-HSI and then the residual map and Up-HSI are element-wise summed to produce the final output HR-HSI. “DownSample 2×” denotes 2 × 2 maxpooling, and “UpSample 2×” and “UpSample 4×” denote bilinear interpolation with scale 2 and scale 4, respectively. “Input CCC” and “Fea CCC” denote channel cross-concatenation of input images and feature maps, respectively.
Figure 2. The architecture of the proposed CCC-SSA-UNet. CCC-SSA-UNet takes an LR-HSI and a PAN as input and takes an HR-HSI as output. It combines UNet and SSA-Net, exploits the “Conv Block” as the main building block of the UNet backbone, and adopts the Res-SSA block as the main building block of the SSA-Net. In each “Conv Block” of the UNet, the input is first passed through a 3 × 3 convolution and subsequently is passed through a batch normalization layer (BN) and a LeakyReLU layer. In this way, the feature map is extracted and passed to the next layer. Finally, the output of UNet is passed through a 1 × 1 convolution to produce the residual map of the Up-HSI and then the residual map and Up-HSI are element-wise summed to produce the final output HR-HSI. “DownSample 2×” denotes 2 × 2 maxpooling, and “UpSample 2×” and “UpSample 4×” denote bilinear interpolation with scale 2 and scale 4, respectively. “Input CCC” and “Fea CCC” denote channel cross-concatenation of input images and feature maps, respectively.
Remotesensing 15 04328 g002
Figure 3. Schematic illustration of the proposed CCC method. (a) Schematic illustration of the proposed Input CCC operation. The tensors corresponding to Up-HSI and PAN are taken as inputs, and the result of channel cross-concatenation is used as the output. The spatial resolution of Up-HSI is H × W pixels, and it has a spectral bandwidth of C . The spatial resolution of PAN is H × W pixels, and it has a spectral bandwidth of C p . The spatial resolution of output is H × W pixels, and it has a spectral bandwidth of C o . (b) Schematic illustration of the proposed Feature CCC operation. Feature 1 denotes the first input feature map of Feature CCC, and Feature 2 denotes the second input feature map of Feature CCC. The tensors corresponding to Feature 1 and Feature 2 are taken as inputs, and the result of channel cross-concatenation is used as the output. The spatial resolution of Feature 1 is H × W pixels, and it has a spectral bandwidth of C q . The spatial resolution of Feature 2 is H × W pixels, and it has a spectral bandwidth of C r . The spatial resolution of output is H × W pixels, and it has a spectral bandwidth of C o .
Figure 3. Schematic illustration of the proposed CCC method. (a) Schematic illustration of the proposed Input CCC operation. The tensors corresponding to Up-HSI and PAN are taken as inputs, and the result of channel cross-concatenation is used as the output. The spatial resolution of Up-HSI is H × W pixels, and it has a spectral bandwidth of C . The spatial resolution of PAN is H × W pixels, and it has a spectral bandwidth of C p . The spatial resolution of output is H × W pixels, and it has a spectral bandwidth of C o . (b) Schematic illustration of the proposed Feature CCC operation. Feature 1 denotes the first input feature map of Feature CCC, and Feature 2 denotes the second input feature map of Feature CCC. The tensors corresponding to Feature 1 and Feature 2 are taken as inputs, and the result of channel cross-concatenation is used as the output. The spatial resolution of Feature 1 is H × W pixels, and it has a spectral bandwidth of C q . The spatial resolution of Feature 2 is H × W pixels, and it has a spectral bandwidth of C r . The spatial resolution of output is H × W pixels, and it has a spectral bandwidth of C o .
Remotesensing 15 04328 g003
Figure 4. Schematic illustration of the proposed SSA-Net. The SSA-Net consists of N Res-SSA blocks connected sequentially. And “Res-SSA” denotes residual spatial–spectral attention. In the proposed Res-SSA block, spectral attention is parallel to spatial attention and is embedded in the basic residual module.
Figure 4. Schematic illustration of the proposed SSA-Net. The SSA-Net consists of N Res-SSA blocks connected sequentially. And “Res-SSA” denotes residual spatial–spectral attention. In the proposed Res-SSA block, spectral attention is parallel to spatial attention and is embedded in the basic residual module.
Remotesensing 15 04328 g004
Figure 5. Visual results generated by different pansharpening algorithms for the 10th patch of the Pavia University dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. The RGB images are generated using the HSI’s 60th, 30th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).
Figure 5. Visual results generated by different pansharpening algorithms for the 10th patch of the Pavia University dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. The RGB images are generated using the HSI’s 60th, 30th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).
Remotesensing 15 04328 g005
Figure 6. Mean absolute error maps of different pansharpening algorithms for the 10th patch of the Pavia University dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands; the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).
Figure 6. Mean absolute error maps of different pansharpening algorithms for the 10th patch of the Pavia University dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands; the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).
Remotesensing 15 04328 g006
Figure 7. Visual results generated by different pansharpening algorithms for the 15th patch of the Pavia Center dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. The RGB images are generated using the HSI’s 60th, 30th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).
Figure 7. Visual results generated by different pansharpening algorithms for the 15th patch of the Pavia Center dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. The RGB images are generated using the HSI’s 60th, 30th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).
Remotesensing 15 04328 g007
Figure 8. Mean absolute error maps of different pansharpening algorithms for the 15th patch of the Pavia Center dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands, the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).
Figure 8. Mean absolute error maps of different pansharpening algorithms for the 15th patch of the Pavia Center dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands, the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).
Remotesensing 15 04328 g008
Figure 9. Visual results generated by different pansharpening algorithms for the 31st patch of the Chikusei dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. The RGB images are generated using the HSI’s 61st, 35th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).
Figure 9. Visual results generated by different pansharpening algorithms for the 31st patch of the Chikusei dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. The RGB images are generated using the HSI’s 61st, 35th, and 10th bands as red, green, and blue bands, respectively. The yellow box indicates the magnified region of interest (ROI).
Remotesensing 15 04328 g009
Figure 10. Mean absolute error maps of different pansharpening algorithms for the 31st patch of the Chikusei dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands, the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).
Figure 10. Mean absolute error maps of different pansharpening algorithms for the 31st patch of the Chikusei dataset. (a) GS [5], (b) GSA [6], (c) PCA [8], (d) GFPCA [7], (e) BayesNaive [14], (f) BayesSparse [44], (g) MTF-GLP [9], (h) MTF-GLP-HPM [11], (i) CNMF [16], (j) HySure [12], (k) HyperPNN1 [28], (l) HyperPNN2 [28], (m) DHP-DARN [30], (n) DIP-HyperKite [31], (o) Hyper-DSNet [29], (p) CCC-SSA-UNet-S (Ours), (q) CCC-SSA-UNet-L (Ours), and (r) reference ground truth. MAE colormap denotes the colormap of normalized mean absolute error across all spectral bands, the minimum value is set to 0 and the maximum value is set to 0.3 for better visual comparison. The yellow box indicates the magnified region of interest (ROI).
Remotesensing 15 04328 g010
Figure 11. Visual comparison of the computational complexities among different pansharpening methods on the Pavia University dataset. (a) Comparison of PSNR, FLOPs, and GPU memory usage. (b) Comparison of SAM, FLOPs, and GPU memory usage. (c) Comparison of PSNR and Runtime. (d) Comparison of SAM and Runtime.
Figure 11. Visual comparison of the computational complexities among different pansharpening methods on the Pavia University dataset. (a) Comparison of PSNR, FLOPs, and GPU memory usage. (b) Comparison of SAM, FLOPs, and GPU memory usage. (c) Comparison of PSNR and Runtime. (d) Comparison of SAM and Runtime.
Remotesensing 15 04328 g011
Figure 12. Schematic illustration of several attention block variants. (a) Residual block baseline, (b) CA, (c) SA, (d) CSA, and (e) DAU.
Figure 12. Schematic illustration of several attention block variants. (a) Residual block baseline, (b) CA, (c) SA, (d) CSA, and (e) DAU.
Remotesensing 15 04328 g012
Table 1. Quantitative results of different methods on the Pavia University dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better.
Table 1. Quantitative results of different methods on the Pavia University dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better.
TypeMethodCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑
TraditionalGS [5]0.9416.2730.032937.9334.75530.572
GSA [6]0.9326.9750.032638.6874.74530.709
PCA [8]0.8079.4170.049829.1566.97727.059
GFPCA [7]0.8559.1000.051628.7387.24726.754
BayesNaive [14]0.8695.9400.044331.8336.59827.662
BayesSparse [44]0.8928.5410.042832.2206.21128.210
MTF-GLP [9]0.9416.1700.030339.4984.27331.570
MTF-GLP-HPM [11]0.9176.4480.034836.4595.56930.401
CNMF [16]0.9196.2520.036935.9055.35629.617
HySure [12]0.9535.6730.026142.6333.80932.663
Deep
learning
HyperPNN1 [28]0.9764.1170.017949.9032.700 35.771
HyperPNN2 [28]0.9764.0450.017650.270 2.66335.900
DHP-DARN [30]0.9803.7930.016152.0152.44436.667
DIP-HyperKite [31]0.9804.1270.016851.1262.54536.270
Hyper-DSNet [29]0.9774.0380.017350.6182.59136.097
CCC-SSA-UNet-S (Ours)0.9823.5170.014753.6902.26837.453
CCC-SSA-UNet-L (Ours)0.9833.4720.014554.0172.24037.595
Table 2. Quantitative results of different methods on the Pavia Centre dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better.
Table 2. Quantitative results of different methods on the Pavia Centre dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better.
TypeMethodCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑
Traditional GS [5]0.9647.5270.028137.0034.95631.694
GSA [6]0.9557.9150.026338.8914.73232.236
PCA [8]0.9467.9780.032434.5605.55530.917
GFPCA [7]0.9039.4630.045326.9407.77727.526
BayesNaive [14]0.8856.9640.043128.2927.59327.760
BayesSparse [44]0.9298.9080.035231.9996.47129.507
MTF-GLP [9]0.9607.1340.024839.9624.42932.852
MTF-GLP-HPM [11]0.9527.5850.026539.0335.17432.468
CNMF [16]0.9487.4020.029336.3855.20031.287
HySure [12]0.9716.7230.020843.6243.79234.444
Deep
learning
HyperPNN1 [28]0.9815.3650.015949.1482.99036.910
HyperPNN2 [28]0.9815.4150.016148.9113.01636.814
DHP-DARN [30]0.9816.1750.015849.1853.03836.678
DIP-HyperKite [31]0.9816.1620.015449.6712.97536.869
Hyper-DSNet [29]0.9844.9400.014151.5472.68037.971
CCC-SSA-UNet-S (Ours)0.9864.6560.012853.4522.49038.839
CCC-SSA-UNet-L (Ours)0.9864.6450.012853.4422.48638.844
Table 3. Quantitative results of different methods on the Chikusei dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better.
Table 3. Quantitative results of different methods on the Chikusei dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better.
TypeMethodCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑
Traditional GS [5]0.9423.8650.017645.0535.95036.334
GSA [6]0.9473.7520.015248.3735.72837.467
PCA [8]0.7936.2140.034331.7669.52231.524
GFPCA [7]0.8805.2370.026336.8438.50232.937
BayesNaive [14]0.9103.3670.023739.2356.52234.449
BayesSparse [44]0.8994.8400.021940.3967.96334.145
MTF-GLP [9]0.9384.0510.015747.5596.21136.994
MTF-GLP-HPM [11]0.7656.3220.043228.78224.00131.610
CNMF [16]0.9014.7590.020842.2517.22935.224
HySure [12]0.9623.1390.011753.5714.82539.615
Deep
learning
HyperPNN1 [28]0.9662.8740.010555.7094.45840.404
HyperPNN2 [28]0.9672.8600.010555.8204.41040.464
DHP-DARN [30]0.9563.6310.011753.5725.02939.268
DIP-HyperKite [31]0.9523.8840.012152.8175.32438.894
Hyper-DSNet [29]0.9802.2740.008460.2323.46042.535
CCC-SSA-UNet-S (Ours)0.9802.2620.008460.3483.49242.582
CCC-SSA-UNet-L (Ours)0.9802.2630.008460.4083.47842.611
Table 4. Computational complexity comparison of different pansharpening methods on the Pavia University dataset. The best value is marked in red, and the second-best value is marked in blue. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Table 4. Computational complexity comparison of different pansharpening methods on the Pavia University dataset. The best value is marked in red, and the second-best value is marked in blue. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
TypeMethodPSNR
SAM
#Params
(M)
MACs
(G)
FLOPs
(G)
Memory
(G)
Runtime
(ms)
Traditional GS [5]30.5726.273----79.0
GSA [6]30.7096.975----205.8
PCA [8]27.0599.417----102.8
GFPCA [7]26.7549.100----139.5
BayesNaive [14]27.6625.940----136.5
BayesSparse [44]28.2108.541----79.5
MTF-GLP [9]31.5706.170----30.0
MTF-GLP-HPM [11]30.4016.448----118.5
CNMF [16]29.6176.252----8976.3
HySure [12]32.6635.673----2409.0
Deep
learning
HyperPNN1 [28]35.7714.1170.1331.2222.4440.89824.3
HyperPNN2 [28]35.9004.0450.1371.2592.5181.12422.5
DHP-DARN [30]36.6673.7930.4173.8217.6422.367176,695.5
DIP-HyperKite [31]36.2704.1270.526122.981245.9627.08223,375.5
Hyper-DSNet [29]36.0974.0380.2722.490 4.980 1.57127.5
CCC-SSA-UNet-S (Ours)37.4533.5170.7273.2596.5191.44647.8
CCC-SSA-UNet-L (Ours)37.5953.4724.4326.32312.6461.462 47.8
Table 5. Performance comparison of CCC-SSA-UNet with different numbers of filter channels on the Pavia University dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Table 5. Performance comparison of CCC-SSA-UNet with different numbers of filter channels on the Pavia University dataset. The best value is marked in red, the second-best value is marked in blue, and the third-best value is marked in green. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Model C f 0   C f 1 C f 2 CC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑#Params (M)MACs (G)Runtime (ms)
13232320.9823.5170.014753.690 2.26837.4530.727 3.25947.8
26464640.9823.5280.014653.850 2.25737.5122.686 11.03848.8
31281281280.9823.5270.014753.7092.28237.43210.331 40.45854.3
412864320.9823.4950.014853.6972.27437.4594.568 33.01454.5
532641280.9833.4720.014554.0172.240 37.5954.4326.32347.8
Table 6. Performance comparison of CCC-SSA-UNet-L with different numbers of Input CCC group on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Table 6. Performance comparison of CCC-SSA-UNet-L with different numbers of Input CCC group on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
mCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑#Params (M)MACs (G)Runtime (ms)
10.9833.4920.014553.9562.25237.5484.4306.30547.3
20.9833.4860.014653.8772.270 37.490 4.4306.30747.3
40.9823.5380.014753.7952.26137.4984.431 6.31247.5
80.9833.4720.014554.0172.24037.5954.432 6.32347.8
120.9833.4960.014653.9412.25137.5554.433 6.33448.0
150.9823.5060.014553.9672.25837.5334.434 6.34248.5
260.9823.5770.014853.6212.27937.4224.437 6.37149.3
350.9823.4910.014653.8852.25837.5024.440 6.39550.0
Table 7. Performance comparison of CCC-SSA-UNet-L with different numbers of Feature CCC group on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Table 7. Performance comparison of CCC-SSA-UNet-L with different numbers of Feature CCC group on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
nCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑#Params (M)MACs (G)Runtime (ms)
10.9833.5080.014753.8332.27737.4794.4326.32348.0
20.9823.4830.014653.8582.26837.4754.4326.32348.8
40.9823.4970.014653.8742.25937.5274.4326.32349.8
80.9833.4720.014554.0172.24037.5954.4326.32347.8
160.9823.4880.014553.9892.25137.5594.4326.32350.0
320.9823.5150.014653.8682.25837.5164.4326.32350.0
Table 8. Performance comparison of CCC-SSA-UNet-L with different numbers of SSA blocks on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Table 8. Performance comparison of CCC-SSA-UNet-L with different numbers of SSA blocks on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
NCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑#Params (M)MACs (G)Runtime (ms)
00.9783.9640.017050.9382.54936.2860.5281.22224.5
10.9774.0590.017650.3492.60836.0390.9181.73226.5
20.9803.7790.016151.9882.43636.7421.3092.24227.5
40.9803.7240.015852.3862.40036.8922.0893.26232.5
60.9813.6790.015752.5202.37536.9812.8704.28238.0
80.9823.5110.014753.8072.26437.4913.6515.30345.3
100.9833.4720.014554.0172.24037.5954.4326.32347.8
120.9823.4990.014753.7512.26437.4705.2137.34361.5
140.9823.5020.014853.6052.29037.3845.9948.36469.3
160.9833.4780.014554.0512.24637.5726.7759.38471.5
180.9833.4630.014553.9922.24537.5717.55610.40472.0
200.9823.5260.014653.8572.27037.4828.33711.42576.5
Table 9. Performance comparison of CCC-SSA-UNet-L with different initial learning rates on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better.
Table 9. Performance comparison of CCC-SSA-UNet-L with different initial learning rates on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better.
Learning RateDecay RateCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑
0.0040.50.9823.5400.014853.6462.30237.374
0.0020.50.9823.5260.014753.7932.26337.496
0.0010.50.9833.4720.014554.0172.24037.595
0.00050.50.9823.5820.014853.6012.29437.388
0.00010.50.9754.4410.018549.4112.77035.497
Table 10. Quantitative results of ablation study of the CCC-SSA-UNet-L on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better.
Table 10. Quantitative results of ablation study of the CCC-SSA-UNet-L on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better.
ModelInput CCCFeature CCCSSA-NetCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑
10.9774.0460.017550.4232.60436.064
20.9784.0090.017150.9232.56036.271
30.9783.9990.017250.7122.58036.157
40.9823.5060.014753.8092.26937.490
50.9833.4920.014553.9562.25237.548
60.9833.5080.014753.8332.27737.479
70.9783.9640.017050.9382.54936.286
80.9833.4720.014554.0172.24037.595
Table 11. Performance comparison of CCC-SSA-UNet-L with different attention blocks on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
Table 11. Performance comparison of CCC-SSA-UNet-L with different attention blocks on the Pavia University dataset. The best value is marked in bold. means that the larger the value, the better, while means that the smaller the value, the better. # means the number of something.
AttentionCC ↑SAM ↓RMSE ↓RSNR ↑ERGAS ↓PSNR ↑#Params (M)MACs (G)Runtime (ms)
Baseline0.9823.5810.0150 53.3852.29937.3264.4036.31832.8
CA0.9813.6520.015552.7222.35537.0664.432 6.32341.8
SA0.980 3.7140.015952.3062.39536.8914.405 6.32338.0
CSA0.9823.5610.014753.710 2.28837.4144.434 6.32844.0
DAU0.9823.5680.0150 53.4252.30337.3274.892 6.89554.0
Res-SSA0.9833.4720.014554.0172.24037.5954.432 6.32347.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Z.; Han, G.; Yang, H.; Liu, P.; Chen, D.; Liu, D.; Deng, A. CCC-SSA-UNet: U-Shaped Pansharpening Network with Channel Cross-Concatenation and Spatial–Spectral Attention Mechanism for Hyperspectral Image Super-Resolution. Remote Sens. 2023, 15, 4328. https://doi.org/10.3390/rs15174328

AMA Style

Liu Z, Han G, Yang H, Liu P, Chen D, Liu D, Deng A. CCC-SSA-UNet: U-Shaped Pansharpening Network with Channel Cross-Concatenation and Spatial–Spectral Attention Mechanism for Hyperspectral Image Super-Resolution. Remote Sensing. 2023; 15(17):4328. https://doi.org/10.3390/rs15174328

Chicago/Turabian Style

Liu, Zhichao, Guangliang Han, Hang Yang, Peixun Liu, Dianbing Chen, Dongxu Liu, and Anping Deng. 2023. "CCC-SSA-UNet: U-Shaped Pansharpening Network with Channel Cross-Concatenation and Spatial–Spectral Attention Mechanism for Hyperspectral Image Super-Resolution" Remote Sensing 15, no. 17: 4328. https://doi.org/10.3390/rs15174328

APA Style

Liu, Z., Han, G., Yang, H., Liu, P., Chen, D., Liu, D., & Deng, A. (2023). CCC-SSA-UNet: U-Shaped Pansharpening Network with Channel Cross-Concatenation and Spatial–Spectral Attention Mechanism for Hyperspectral Image Super-Resolution. Remote Sensing, 15(17), 4328. https://doi.org/10.3390/rs15174328

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop