[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (149)

Search Parameters:
Keywords = superpixel image classification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 33763 KiB  
Article
MSIMRS: Multi-Scale Superpixel Segmentation Integrating Multi-Source Remote Sensing Data for Lithology Identification in Semi-Arid Area
by Jiaxin Lu, Liangzhi Li, Junfeng Wang, Ling Han, Zhaode Xia, Hongjie He and Zongfan Bai
Remote Sens. 2025, 17(3), 387; https://doi.org/10.3390/rs17030387 - 23 Jan 2025
Viewed by 359
Abstract
Lithology classification stands as a pivotal research domain within geological Remote Sensing (RS). In recent years, extracting lithology information from multi-source RS data has become an inevitable trend. Various classification image primitives yield distinct outcomes in lithology classification. The current research on lithology [...] Read more.
Lithology classification stands as a pivotal research domain within geological Remote Sensing (RS). In recent years, extracting lithology information from multi-source RS data has become an inevitable trend. Various classification image primitives yield distinct outcomes in lithology classification. The current research on lithology classification utilizing RS data has predominantly concentrated on pixel-level classification, which suffers from a long classification time and high sensitivity to noise. In order to explore the application potential of superpixel segmentation in lithology classification, this study proposed the Multi-scale superpixel Segmentation Integrating Multi-source RS data (MSIMRS), and conducted a lithology classification study in Duolun County, Inner Mongolia Autonomous Region, China combining MSIMRS and the Support Vector Machine (MSIMRS-SVM). In addition, pixel-level K-Nearest Neighbor (KNN), Random Forest (RF) and SVM classification algorithms, as well as deep-learning models including Resnet50 (Res50), Efficientnet_B8 (Effi_B8), and Vision Transformer (ViT) were chosen for a comparative analysis. Among these methods, our proposed MSIMRS-SVM achieved the highest accuracy in lithology classification in a typical semi-arid area, Duolun County, with an overall accuracy and Kappa coefficient of 92.9% and 0.92. Moreover, the findings indicate that incorporating superpixel segmentation into lithology classification resulted in notably fewer fragmented patches and significantly improved the visualization effect. The results showcase the application potential of superpixel primitives in lithology information extraction within semi-arid areas. Full article
Show Figures

Figure 1

Figure 1
<p>The location map of the study area.</p>
Full article ">Figure 2
<p>Geologic schematic map of the study area.</p>
Full article ">Figure 3
<p>The superpixel clustering criterion integrating multi-source RS data.</p>
Full article ">Figure 4
<p>Flowchart of single-scale superpixel segmentation.</p>
Full article ">Figure 5
<p>Flowchart of the MSIMRS algorithm.</p>
Full article ">Figure 6
<p>Segmentation results in the region of 1000 × 1000 pixels in our study area: (<b>a</b>) single-scale segmentation result; (<b>b</b>) MSIMRS segmentation result; (<b>c</b>) classification result corresponding to single-scale segmentation; (<b>d</b>) classification result corresponding to MSIMRS segmentation; and (<b>e</b>) legend.</p>
Full article ">Figure 6 Cont.
<p>Segmentation results in the region of 1000 × 1000 pixels in our study area: (<b>a</b>) single-scale segmentation result; (<b>b</b>) MSIMRS segmentation result; (<b>c</b>) classification result corresponding to single-scale segmentation; (<b>d</b>) classification result corresponding to MSIMRS segmentation; and (<b>e</b>) legend.</p>
Full article ">Figure 7
<p>Classification maps of different algorithms in our study area.</p>
Full article ">Figure 8
<p>Lithology classification results of the whole study area through different algorithms: (<b>a</b>) SVM; (<b>b</b>) MSIMRS-SVM; and (<b>c</b>) legend.</p>
Full article ">Figure 9
<p>Accuracy statistics of different lithology categories in our study area: (<b>a</b>) precision (%); and (<b>b</b>) recall (%). The category numbers are the same as those in the legend in <a href="#remotesensing-17-00387-f007" class="html-fig">Figure 7</a>.</p>
Full article ">
10 pages, 2010 KiB  
Proceeding Paper
Learnable Weight Graph Neural Network for River Ice Classification
by Yifan Qu, Armina Soleymani, Denise Sudom and Katharine Andrea Scott
Proceedings 2024, 110(1), 30; https://doi.org/10.3390/proceedings2024110030 - 13 Jan 2025
Viewed by 301
Abstract
Monitoring river ice is crucial for planning safe navigation routes, with ice–water classification being one of the most important tasks in ice mapping. While high-resolutions satellite imagery, such as synthetic aperture radar (SAR), is well-suited to this task, manual interpretation of these data [...] Read more.
Monitoring river ice is crucial for planning safe navigation routes, with ice–water classification being one of the most important tasks in ice mapping. While high-resolutions satellite imagery, such as synthetic aperture radar (SAR), is well-suited to this task, manual interpretation of these data is challenging due to the large data volume. Machine learning approaches are suitable methods to overcome this; however, training the models might not be time-effective when the desired result is a narrow structure, such as a river, within a large image. To address this issue, we proposed a model incorporating a graph neural network (GNN), called learnable weights graph convolution network (LWGCN). Focusing on the winters of 2017–2021 with emphasis on the Beauharnois Canal and Lake St Lawrence regions of the Saint Lawrence River. The model first converts the SAR image into graph-structured data using simple linear iterative clustering (SLIC) to segment the SAR image, then connecting the centers of each superpixel to form graph-structured data. For the training model, the LWGCN learns the weights on each edge to determine the relationship between ice and water. By using the graph-structured data as input, the proposed model training time is eight times faster, compared to a convolution neural network (CNN) model. Our findings also indicate that the LWGCN model can significantly enhance the accuracy of ice and water classification in SAR imagery. Full article
(This article belongs to the Proceedings of The 31st International Conference on Geoinformatics)
Show Figures

Figure 1

Figure 1
<p>The study region consists of the Beauharnois Canal and Lake Saint Lawrence. The central coordinates for the Beauharnois Canal are approximately 45.26° N and 73.94° W. The central coordinates for Lake Saint Lawrence are approximately 44.99° N and 74.88° W.</p>
Full article ">Figure 2
<p>The process of generating graphs from SAR imagery for an arbitrary chosen date (2017-01-12) in the Beauharnois Canal. (<b>a</b>) Sentinel-1 SAR image. (<b>b</b>) Use simple linear iterative clustering (SLIC) to segment the image into superpixels. (<b>c</b>) Connect the centers of each superpixel. (<b>d</b>) Remove the land area (this is the graph structure used in the LWGCN model).</p>
Full article ">Figure 3
<p>(<b>a</b>) Sentinel-1 VV SAR image of Lake Saint Lawrence (see <a href="#proceedings-110-00030-f001" class="html-fig">Figure 1</a> for location within larger study region), (<b>b</b>) ground truth from manually labeled shapefile, where blue indicates water and red indicates ice, and (<b>c</b>) LWGCN model output, where the colors are represented in <a href="#proceedings-110-00030-t005" class="html-table">Table 5</a>. The arbitrary chosen date is 2018-01-07.</p>
Full article ">
22 pages, 16745 KiB  
Article
Unsupervised PolSAR Image Classification Based on Superpixel Pseudo-Labels and a Similarity-Matching Network
by Lei Wang, Lingmu Peng, Rong Gui, Hanyu Hong and Shenghui Zhu
Remote Sens. 2024, 16(21), 4119; https://doi.org/10.3390/rs16214119 - 4 Nov 2024
Viewed by 1093
Abstract
Supervised polarimetric synthetic aperture radar (PolSAR) image classification demands a large amount of precisely labeled data. However, such data are difficult to obtain. Therefore, many unsupervised methods have been proposed for unsupervised PolSAR image classification. The classification maps of unsupervised methods contain many [...] Read more.
Supervised polarimetric synthetic aperture radar (PolSAR) image classification demands a large amount of precisely labeled data. However, such data are difficult to obtain. Therefore, many unsupervised methods have been proposed for unsupervised PolSAR image classification. The classification maps of unsupervised methods contain many high-confidence samples. These samples, which are often ignored, can be used as supervisory information to improve classification performance on PolSAR images. This study proposes a new unsupervised PolSAR image classification framework. The framework combines high-confidence superpixel pseudo-labeled samples and semi-supervised classification methods. The experiments indicated that this framework could achieve higher-level effectiveness in unsupervised PolSAR image classification. First, superpixel segmentation was performed on PolSAR images, and the geometric centers of the superpixels were generated. Second, the classification maps of rotation-domain deep mutual information (RDDMI), an unsupervised PolSAR image classification method, were used as the pseudo-labels of the central points of the superpixels. Finally, the unlabeled samples and the high-confidence pseudo-labeled samples were used to train an excellent semi-supervised method, similarity matching (SimMatch). Experiments on three real PolSAR datasets illustrated that, compared with the excellent RDDMI, the accuracy of the proposed method was increased by 1.70%, 0.99%, and 0.8%. The proposed framework provides significant performance improvements and is an efficient method for improving unsupervised PolSAR image classification. Full article
(This article belongs to the Special Issue SAR in Big Data Era III)
Show Figures

Figure 1

Figure 1
<p>The five parts of the framework. The Wide ResNet model adopts the classic wide residual networks (WRNs) [<a href="#B37-remotesensing-16-04119" class="html-bibr">37</a>]. The useful features from the input data are extracted by the backbone to obtain an embedding vector. <math display="inline"><semantics> <msub> <mi>L</mi> <mi>s</mi> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>L</mi> <mi>u</mi> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>L</mi> <mrow> <mi>i</mi> <mi>n</mi> </mrow> </msub> </semantics></math> represent the supervised loss, unsupervised loss, and similarity distribution, respectively.</p>
Full article ">Figure 2
<p>The pseudo-label generation structure of SimMatch. SimMatch generates semantic and instance pseudo-labels using weakly augmented views and calculates semantic and instance similarities through class centers. These two similarities are then propagated to each other using expansion and aggregation to obtain better pseudo-labels.</p>
Full article ">Figure 3
<p>The propagation of pseudo-labels’ information. As the example in the red box shows. If the similarity between semantics and instances are different, the histogram will be flatter, and if the semantics similarities and the instances similarities are similar, the resulting histogram will be sharper.</p>
Full article ">Figure 4
<p>RS-2 Flevoland dataset. (<b>a</b>) Pauli pseudo-color image. (<b>b</b>) Ground-truth map.</p>
Full article ">Figure 5
<p>RS-2 Wuhan dataset. (<b>a</b>) Pauli pseudo-color image. (<b>b</b>) Ground-truth map. (<b>c</b>) An optical image of ROI_1. (<b>d</b>) An optical image of ROI_2.</p>
Full article ">Figure 6
<p>AIRSAR Flevoland dataset. (<b>a</b>) Pauli pseudo-color image. (<b>b</b>) Ground-truth map.</p>
Full article ">Figure 7
<p>Classification results on the RS-2 Flevoland dataset. The black boxes show that SP-SIM has more fine classification results than RDDMI. (<b>a</b>) Ground-truth map. (<b>b</b>) Wishart. (<b>c</b>) RDDMI. (<b>d</b>) SP-SIM.</p>
Full article ">Figure 8
<p>Classification results on the RS-2 Wuhan dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) Wishart. (<b>c</b>) RDDMI. (<b>d</b>) SP-SIM.</p>
Full article ">Figure 9
<p>Similar backscattering properties on the AIRSAR Flevoland dataset. (<b>a</b>) Four similar backscattering properties. (<b>b</b>) Water. (<b>c</b>) Bare soil. (<b>d</b>) Lucerne. (<b>e</b>) Rape seed.</p>
Full article ">Figure 10
<p>Classification results on the AIRSAR Flevoland dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) Wishart. (<b>c</b>) RDDMI. (<b>d</b>) SP-SIM.</p>
Full article ">
17 pages, 2991 KiB  
Article
Feature Extraction and Identification of Rheumatoid Nodules Using Advanced Image Processing Techniques
by Azmath Mubeen and Uma N. Dulhare
Rheumato 2024, 4(4), 176-192; https://doi.org/10.3390/rheumato4040014 - 24 Oct 2024
Viewed by 656
Abstract
Background/Objectives: Accurate detection and classification of nodules in medical images, particularly rheumatoid nodules, are critical due to the varying nature of these nodules, where their specific type is often unknown before analysis. This study addresses the challenges of multi-class prediction in nodule detection, [...] Read more.
Background/Objectives: Accurate detection and classification of nodules in medical images, particularly rheumatoid nodules, are critical due to the varying nature of these nodules, where their specific type is often unknown before analysis. This study addresses the challenges of multi-class prediction in nodule detection, with a specific focus on rheumatoid nodules, by employing a comprehensive approach to feature extraction and classification. We utilized a diverse dataset of nodules, including rheumatoid nodules sourced from the DermNet dataset and local rheumatologists. Method: This study integrates 62 features, combining traditional image characteristics with advanced graph-based features derived from a superpixel graph constructed through Delaunay triangulation. The key steps include image preprocessing with anisotropic diffusion and Retinex enhancement, superpixel segmentation using SLIC, and graph-based feature extraction. Texture analysis was performed using Gray-Level Co-occurrence Matrix (GLCM) metrics, while shape analysis was conducted with Fourier descriptors. Vascular pattern recognition, crucial for identifying rheumatoid nodules, was enhanced using the Frangi filter. A Hybrid CNN–Transformer model was employed for feature fusion, and feature selection and hyperparameter tuning were optimized using Gray Wolf Optimization (GWO) and Particle Swarm Optimization (PSO). Feature importance was assessed using SHAP values. Results: The proposed methodology achieved an accuracy of 85%, with a precision of 0.85, a recall of 0.89, and an F1 measure of 0.87, demonstrating the effectiveness of the approach in detecting and classifying rheumatoid nodules in both binary and multi-class classification scenarios. Conclusions: This study presents a robust tool for the detection and classification of nodules, particularly rheumatoid nodules, in medical imaging, offering significant potential for improving diagnostic accuracy and aiding in the early identification of rheumatoid conditions. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Gray-scale lesion image; (<b>b</b>) resized image.</p>
Full article ">Figure 2
<p>(<b>a</b>) Resized image; (<b>b</b>) segmented image.</p>
Full article ">Figure 3
<p>Graph density output.</p>
Full article ">Figure 4
<p>Rheumatoid nodule identified.</p>
Full article ">Figure 5
<p>Confusion matrix.</p>
Full article ">
22 pages, 4894 KiB  
Article
SMALE: Hyperspectral Image Classification via Superpixels and Manifold Learning
by Nannan Liao, Jianglei Gong, Wenxing Li, Cheng Li, Chaoyan Zhang and Baolong Guo
Remote Sens. 2024, 16(18), 3442; https://doi.org/10.3390/rs16183442 - 17 Sep 2024
Viewed by 977
Abstract
As an extremely efficient preprocessing tool, superpixels have become more and more popular in various computer vision tasks. Nevertheless, there are still several drawbacks in the application of hyperspectral image (HSl) processing. Firstly, it is difficult to directly apply superpixels because of the [...] Read more.
As an extremely efficient preprocessing tool, superpixels have become more and more popular in various computer vision tasks. Nevertheless, there are still several drawbacks in the application of hyperspectral image (HSl) processing. Firstly, it is difficult to directly apply superpixels because of the high dimension of HSl information. Secondly, existing superpixel algorithms cannot accurately classify the HSl objects due to multi-scale feature categorization. For the processing of high-dimensional problems, we use the principle of PCA to extract three principal components from numerous bands to form three-channel images. In this paper, a novel superpixel algorithm called Seed Extend by Entropy Density (SEED) is proposed to alleviate the seed point redundancy caused by the diversified content of HSl. It also focuses on breaking the dilemma of manually setting the number of superpixels to overcome the difficulty of classification imprecision caused by multi-scale targets. Next, a space–spectrum constraint model, termed Hyperspectral Image Classification via superpixels and manifold learning (SMALE), is designed, which integrates the proposed SEED to generate a dimensionality reduction framework. By making full use of spatial context information in the process of unsupervised dimension reduction, it could effectively improve the performance of HSl classification. Experimental results show that the proposed SEED could effectively promote the classification accuracy of HSI. Meanwhile, the integrated SMALE model outperforms existing algorithms on public datasets in terms of several quantitative metrics. Full article
Show Figures

Figure 1

Figure 1
<p>Two typical superpixel algorithms. (<b>a</b>) The segmentation process of SNIC [<a href="#B39-remotesensing-16-03442" class="html-bibr">39</a>] and its variants algorithm; (<b>b</b>) The segmentation process of SEEDS [<a href="#B48-remotesensing-16-03442" class="html-bibr">48</a>].</p>
Full article ">Figure 2
<p>Two frequently used seed point methods. (<b>a</b>) Input image; (<b>b</b>) grid-based seed point method; (<b>c</b>) gradient-based seed point method.</p>
Full article ">Figure 3
<p>Two frequently used seed point methods. (<b>a</b>) Input image; (<b>b</b>) seeds are distributed adaptively according to their entropy density; (<b>c</b>) seeds are evenly distributed throughout the image.</p>
Full article ">Figure 4
<p>Dynamic segmentation procedure of SEED. (<b>a</b>) Input image; (<b>b</b>) grid-sampled seeds; (<b>c</b>) gray level image; (<b>d</b>) entropy density distribution diagram Entropy density values range from small to large, indicating colors from blue to yellow; (<b>e</b>–<b>g</b>) seed distribution is dynamically adjusted according to different values of the entropy density. Among them, yellow boxes are examples of seed extend paths in regions with larger entropy density, and blue boxes are examples of seed extend paths in regions with smaller entropy density.</p>
Full article ">Figure 5
<p>The procedure of the spatial–spectral dimensionality reduction model. (<b>a</b>) Neighbor selection; (<b>b</b>) data normalization; (<b>c</b>) superpixel segmentation; (<b>d</b>) compute weights; (<b>e</b>) calculation of embedding.</p>
Full article ">Figure 6
<p>Visual comparison of segmentation results with 100 expected superpixels. (<b>a</b>) ERS [<a href="#B38-remotesensing-16-03442" class="html-bibr">38</a>]; (<b>b</b>) SNIC [<a href="#B39-remotesensing-16-03442" class="html-bibr">39</a>]; (<b>c</b>) MBS [<a href="#B40-remotesensing-16-03442" class="html-bibr">40</a>]; (<b>d</b>) WSGL [<a href="#B41-remotesensing-16-03442" class="html-bibr">41</a>]; (<b>e</b>) IBIS [<a href="#B42-remotesensing-16-03442" class="html-bibr">42</a>]; (<b>f</b>) SCALE [<a href="#B43-remotesensing-16-03442" class="html-bibr">43</a>]; (<b>g</b>) SEED. Alternating columns show each segmented image followed by the zoom-in performance.</p>
Full article ">Figure 7
<p>Quantitative evaluation of different algorithms on four evaluation indicators. (<b>a</b>) Boundary recall; (<b>b</b>) under-segmentation error; (<b>c</b>) precision recall; (<b>d</b>) achievable segmentation accuracy; (<b>e</b>) compactness; (<b>f</b>) execution time. The expected number of superpixels ranges from 50 to 500.</p>
Full article ">Figure 8
<p>Classification maps for the Indian Pines dataset using all DR methods under comparison with the optimal parameters. (<b>a</b>) Input image; (<b>b</b>) ground truth; (<b>c</b>) PCA [<a href="#B13-remotesensing-16-03442" class="html-bibr">13</a>]; (<b>d</b>) KPCA [<a href="#B16-remotesensing-16-03442" class="html-bibr">16</a>]; (<b>e</b>) LLE [<a href="#B20-remotesensing-16-03442" class="html-bibr">20</a>]; (<b>f</b>) LTSA [<a href="#B22-remotesensing-16-03442" class="html-bibr">22</a>]; (<b>g</b>) SuperPCA [<a href="#B32-remotesensing-16-03442" class="html-bibr">32</a>]; (<b>h</b>) RLMR [<a href="#B44-remotesensing-16-03442" class="html-bibr">44</a>]; (<b>i</b>) S<sup>2</sup>DL [<a href="#B52-remotesensing-16-03442" class="html-bibr">52</a>]; (<b>j</b>) D-VIC [<a href="#B53-remotesensing-16-03442" class="html-bibr">53</a>]; (<b>k</b>) ERS-LLE; (<b>l</b>) MBS-LLE; (<b>m</b>) SMALE.</p>
Full article ">Figure 9
<p>OA and AA indicators of the proposed method and other methods using the superpixels in the Indian Pines dataset.</p>
Full article ">Figure 10
<p>Classification maps for the Salinas A dataset using all DR methods under comparison with the optimal parameters. (<b>a</b>) Input image; (<b>b</b>) ground truth; (<b>c</b>) PCA [<a href="#B13-remotesensing-16-03442" class="html-bibr">13</a>]; (<b>d</b>) KPCA [<a href="#B16-remotesensing-16-03442" class="html-bibr">16</a>]; (<b>e</b>) LLE [<a href="#B20-remotesensing-16-03442" class="html-bibr">20</a>]; (<b>f</b>) LTSA [<a href="#B22-remotesensing-16-03442" class="html-bibr">22</a>]; (<b>g</b>) SuperPCA [<a href="#B32-remotesensing-16-03442" class="html-bibr">32</a>]; (<b>h</b>) RLMR [<a href="#B44-remotesensing-16-03442" class="html-bibr">44</a>]; (<b>i</b>) S<sup>2</sup>DL [<a href="#B52-remotesensing-16-03442" class="html-bibr">52</a>]; (<b>j</b>) D-VIC [<a href="#B53-remotesensing-16-03442" class="html-bibr">53</a>]; (<b>k</b>) ERS-DR; (<b>l</b>) MBS-DR; (<b>m</b>) SMALE.</p>
Full article ">Figure 10 Cont.
<p>Classification maps for the Salinas A dataset using all DR methods under comparison with the optimal parameters. (<b>a</b>) Input image; (<b>b</b>) ground truth; (<b>c</b>) PCA [<a href="#B13-remotesensing-16-03442" class="html-bibr">13</a>]; (<b>d</b>) KPCA [<a href="#B16-remotesensing-16-03442" class="html-bibr">16</a>]; (<b>e</b>) LLE [<a href="#B20-remotesensing-16-03442" class="html-bibr">20</a>]; (<b>f</b>) LTSA [<a href="#B22-remotesensing-16-03442" class="html-bibr">22</a>]; (<b>g</b>) SuperPCA [<a href="#B32-remotesensing-16-03442" class="html-bibr">32</a>]; (<b>h</b>) RLMR [<a href="#B44-remotesensing-16-03442" class="html-bibr">44</a>]; (<b>i</b>) S<sup>2</sup>DL [<a href="#B52-remotesensing-16-03442" class="html-bibr">52</a>]; (<b>j</b>) D-VIC [<a href="#B53-remotesensing-16-03442" class="html-bibr">53</a>]; (<b>k</b>) ERS-DR; (<b>l</b>) MBS-DR; (<b>m</b>) SMALE.</p>
Full article ">
22 pages, 9228 KiB  
Article
Cross-Hopping Graph Networks for Hyperspectral–High Spatial Resolution (H2) Image Classification
by Tao Chen, Tingting Wang, Huayue Chen, Bochuan Zheng and Wu Deng
Remote Sens. 2024, 16(17), 3155; https://doi.org/10.3390/rs16173155 - 27 Aug 2024
Cited by 1 | Viewed by 1045
Abstract
As we take stock of the contemporary issue, remote sensing images are gradually advancing towards hyperspectral–high spatial resolution (H2) double-high images. However, high resolution produces serious spatial heterogeneity and spectral variability while improving image resolution, which increases the difficulty of feature [...] Read more.
As we take stock of the contemporary issue, remote sensing images are gradually advancing towards hyperspectral–high spatial resolution (H2) double-high images. However, high resolution produces serious spatial heterogeneity and spectral variability while improving image resolution, which increases the difficulty of feature recognition. So as to make the best of spectral and spatial features under an insufficient number of marking samples, we would like to achieve effective recognition and accurate classification of features in H2 images. In this paper, a cross-hop graph network for H2 image classification(H2-CHGN) is proposed. It is a two-branch network for deep feature extraction geared towards H2 images, consisting of a cross-hop graph attention network (CGAT) and a multiscale convolutional neural network (MCNN): the CGAT branch utilizes the superpixel information of H2 images to filter samples with high spatial relevance and designate them as the samples to be classified, then utilizes the cross-hop graph and attention mechanism to broaden the range of graph convolution to obtain more representative global features. As another branch, the MCNN uses dual convolutional kernels to extract features and fuse them at various scales while attaining pixel-level multi-scale local features by parallel cross connecting. Finally, the dual-channel attention mechanism is utilized for fusion to make image elements more prominent. This experiment on the classical dataset (Pavia University) and double-high (H2) datasets (WHU-Hi-LongKou and WHU-Hi-HongHu) shows that the H2-CHGN can be efficiently and competently used in H2 image classification. In detail, experimental results showcase superior performance, outpacing state-of-the-art methods by 0.75–2.16% in overall accuracy. Full article
(This article belongs to the Special Issue Deep Learning for Spectral-Spatial Hyperspectral Image Classification)
Show Figures

Figure 1

Figure 1
<p>The framework of H<sup>2</sup>-CHGN model for H<sup>2</sup> image classification.</p>
Full article ">Figure 2
<p>Superpixel and pixel feature conversion process.</p>
Full article ">Figure 3
<p>The procedure for k-hop matrices.</p>
Full article ">Figure 4
<p>Cross-hop graph attention module: (<b>a</b>) pyramid structure by cross-connect feature and (<b>b</b>) graph attention mechanism.</p>
Full article ">Figure 5
<p>The structure of ConvBlock.</p>
Full article ">Figure 6
<p>Convolutional block attention network module: (<b>a</b>) channel attention module and (<b>b</b>) spatial attention module.</p>
Full article ">Figure 7
<p>Dual-channel attention fusion module.</p>
Full article ">Figure 8
<p>OAs under a different number of heads and epochs: (<b>a</b>) Pavia University; (<b>b</b>) WHU-Hi-LongKou; and (<b>c</b>) WHU-Hi-HongHu.</p>
Full article ">Figure 9
<p>Mean color visualization of superpixel segmented regions on different datasets: (<b>a</b>–<b>d</b>): Pavia University; (<b>e</b>–<b>h</b>): WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>): WHU-Hi-HongHu.</p>
Full article ">Figure 9 Cont.
<p>Mean color visualization of superpixel segmented regions on different datasets: (<b>a</b>–<b>d</b>): Pavia University; (<b>e</b>–<b>h</b>): WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>): WHU-Hi-HongHu.</p>
Full article ">Figure 10
<p>Effect of different superpixel segmentation scales on classification accuracy.</p>
Full article ">Figure 11
<p>Effect of different cross-hopping methods on classification accuracy.</p>
Full article ">Figure 12
<p>Classification maps for the Pavia University dataset: (<b>a</b>) False-color image; (<b>b</b>) ground truth; (<b>c</b>) SVM (OA = 79.54%); (<b>d</b>) CEGCN (OA = 97.81%); (<b>e</b>) SGML (OA = 94.30%); (<b>f</b>) WFCG (OA = 97.53%); (<b>g</b>) MSSG-UNet (OA = 98.52%); (<b>h</b>) MS-RPNet (OA = 96.96%); (<b>i</b>) AMGCFN (OA = 98.24%); and (<b>j</b>) H<sup>2</sup>-CHGN (OA = 99.24%).</p>
Full article ">Figure 13
<p>Classification maps for the WHU-Hi-LongKou dataset: (<b>a</b>) False-color image; (<b>b</b>) ground truth; (<b>c</b>) SVM (OA = 92.88%); (<b>d</b>) CEGCN (OA = 98.72%); (<b>e</b>) SGML (OA = 96.03%); (<b>f</b>) WFCG (OA = 98.29%); (<b>g</b>) MSSG-UNet (OA = 98.56%); (<b>h</b>) MS-RPNet (OA = 97.17%); (<b>i</b>) AMGCFN (OA = 98.44%); and (<b>j</b>) H<sup>2</sup>-CHGN (OA = 99.19%).</p>
Full article ">Figure 14
<p>Classification maps for the WHU-Hi-HongHu dataset: (<b>a</b>) False-color image; (<b>b</b>) Ground truth; (<b>c</b>) SVM (OA = 66.34%); (<b>d</b>) CEGCN (OA = 94.01%); (<b>e</b>) SGML (OA = 92.51%); (<b>f</b>) WFCG (OA = 93.98%); (<b>g</b>) MSSG-UNet (OA = 93.73%); (<b>h</b>) MS-RPNet (OA = 93.56%); (<b>i</b>) AMGCFN (OA = 94.44%); (<b>j</b>) H2-CHGN (OA = 96.60%).</p>
Full article ">Figure 15
<p>Effect of different numbers of training samples for the methods: (<b>a</b>) Pavia University; (<b>b</b>) WHU-Hi-LongKou; and (<b>c</b>) WHU-Hi-HongHu.</p>
Full article ">Figure 16
<p>t-SNE results of different methods on three datasets: (<b>a</b>–<b>d</b>) Pavia University; (<b>e</b>–<b>h</b>) WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>) WHU-Hi-HongHu.</p>
Full article ">Figure 16 Cont.
<p>t-SNE results of different methods on three datasets: (<b>a</b>–<b>d</b>) Pavia University; (<b>e</b>–<b>h</b>) WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>) WHU-Hi-HongHu.</p>
Full article ">
21 pages, 1668 KiB  
Article
DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration
by Wenkai Zhu, Xueying Sun and Qiang Zhang
Electronics 2024, 13(16), 3271; https://doi.org/10.3390/electronics13163271 - 18 Aug 2024
Viewed by 1087
Abstract
In recent years, graph convolutional neural networks (GCNs) and convolutional neural networks (CNNs) have made significant strides in hyperspectral image (HSI) classification. However, existing models often encounter information redundancy and feature mismatch during feature fusion, and they struggle with small-scale refined features. To [...] Read more.
In recent years, graph convolutional neural networks (GCNs) and convolutional neural networks (CNNs) have made significant strides in hyperspectral image (HSI) classification. However, existing models often encounter information redundancy and feature mismatch during feature fusion, and they struggle with small-scale refined features. To address these issues, we propose DCG-Net, an innovative classification network integrating CNN and GCN architectures. Our approach includes the development of a double-branch expanding network (E-Net) to enhance spectral features and efficiently extract high-level features. Additionally, we incorporate a GCN with an attention mechanism to facilitate the integration of multi-space scale superpixel-level and pixel-level features. To further improve feature fusion, we introduce a feature aggregation module (FAM) that adaptively learns channel features, enhancing classification robustness and accuracy. Comprehensive experiments on three widely used datasets show that DCG-Net achieves superior classification results compared to other state-of-the-art methods. Full article
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)
Show Figures

Figure 1

Figure 1
<p>DCG-Net architecture diagram. Our model is structured into four primary stages: superpixel segmentation, feature extraction, feature aggregation, and feature classification. We employ a dual-branch architecture, where each branch processes both pixel-level and superpixel-level features. The 1st branch retains the original image resolution for feature processing, whereas the 2nd branch processes upscaled feature images to facilitate multi-scale feature analysis.</p>
Full article ">Figure 2
<p>E-Net structure diagram. Both the encoder and decoder use a customized convolution module. E-Net can effectively combine GCN in the dual-branch encoding and decoding processes to achieve the effective combination of different spatial features.</p>
Full article ">Figure 3
<p>Flowchart of superpixel segmentation and graph structure construction. (<b>a</b>) Original hyperspectral image. (<b>b</b>) Hyperspectral image after PCA downscaling, with the 4-connected graph as a feature graph constructed using the pixels of (<b>b</b>) as nodes. (<b>c</b>) Hyperspectral image after superpixel segmentation, with the superpixel graph structure as a feature graph constructed using the superpixels of (<b>c</b>) as nodes. The orange dots represent the graph’s nodes, and the orange dotted lines represent the graph’s edges.</p>
Full article ">Figure 4
<p>Feature aggregation module structure diagram. The module begins with channel feature learning through the channel attention module, then it continues with feature extraction using the customized convolution module.</p>
Full article ">Figure 5
<p>Description of Indian Pines dataset.</p>
Full article ">Figure 6
<p>Description of Kennedy Space Center dataset.</p>
Full article ">Figure 7
<p>Description of the Salinas dataset.</p>
Full article ">Figure 8
<p>Classification maps of different methods for the Indian Pines dataset.</p>
Full article ">Figure 9
<p>Classification maps of different methods for the Kennedy Space Center dataset.</p>
Full article ">Figure 10
<p>Classification maps of different methods for the Salinas dataset.</p>
Full article ">Figure 11
<p>Comparison of classification performance of seven methods with different training set ratios for three datasets.</p>
Full article ">Figure 12
<p>Comparison results of different superpixel numbers on three datasets.</p>
Full article ">Figure 13
<p>Comparison results of different K values on three datasets.</p>
Full article ">
22 pages, 12904 KiB  
Article
Intelligent Classification and Segmentation of Sandstone Thin Section Image Using a Semi-Supervised Framework and GL-SLIC
by Yubo Han and Ye Liu
Minerals 2024, 14(8), 799; https://doi.org/10.3390/min14080799 - 5 Aug 2024
Cited by 1 | Viewed by 1121
Abstract
This study presents the development and validation of a robust semi-supervised learning framework specifically designed for the automated segmentation and classification of sandstone thin section images from the Yanchang Formation in the Ordos Basin. Traditional geological image analysis methods encounter significant challenges due [...] Read more.
This study presents the development and validation of a robust semi-supervised learning framework specifically designed for the automated segmentation and classification of sandstone thin section images from the Yanchang Formation in the Ordos Basin. Traditional geological image analysis methods encounter significant challenges due to the labor-intensive and error-prone nature of manual labeling, compounded by the diversity and complexity of rock thin sections. Our approach addresses these challenges by integrating the GL-SLIC algorithm, which combines Gabor filters and Local Binary Patterns for effective superpixel segmentation, laying the groundwork for advanced component identification. The primary innovation of this research is the semi-supervised learning model that utilizes a limited set of manually labeled samples to generate high-confidence pseudo labels, thereby significantly expanding the training dataset. This methodology effectively tackles the critical challenge of insufficient labeled data in geological image analysis, enhancing the model’s generalization capability from minimal initial input. Our framework improves segmentation accuracy by closely aligning superpixels with the intricate boundaries of mineral grains and pores. Additionally, it achieves substantial improvements in classification accuracy across various rock types, reaching up to 96.3% in testing scenarios. This semi-supervised approach represents a significant advancement in computational geology, providing a scalable and efficient solution for detailed petrographic analysis. It not only enhances the accuracy and efficiency of geological interpretations but also supports broader hydrocarbon exploration efforts. Full article
Show Figures

Figure 1

Figure 1
<p>Thin section image of sandstone under plane-polarized light: the main components are quartz, kaolinite, matrix, pores and lithic fragments.</p>
Full article ">Figure 2
<p>Workflow for recognizing minerals using GL-SLIC segmentation and semi-supervised training.</p>
Full article ">Figure 3
<p>GL-BP feature extraction workflow integrating LBP operator and Gabor filters for sandstone thin section images.</p>
Full article ">Figure 4
<p>Feature extraction visualization using Gabor filters at various scales and orientations for sandstone thin section images.</p>
Full article ">Figure 5
<p>Mean feature comparison chart: (<b>a</b>) mean feature chart at scale 1; (<b>b</b>) mean feature chart at scale 2; (<b>c</b>) mean feature chart at scale 3; (<b>d</b>) mean feature chart at scale 4; (<b>e</b>) mean feature chart at scale 5; (<b>f</b>) mean feature chart at scale 6.</p>
Full article ">Figure 6
<p>LBP Feature Extraction: (<b>a</b>) mean feature chart at scale 1; (<b>b</b>) mean feature chart at scale 2; (<b>c</b>) mean feature chart at scale 3; (<b>d</b>) mean feature chart at scale 4; (<b>e</b>) mean feature chart at scale 5; (<b>f</b>) mean feature chart at scale 6.</p>
Full article ">Figure 7
<p>Semi-supervised self-training process.</p>
Full article ">Figure 8
<p>Modified VGG16 Classifier Architecture.</p>
Full article ">Figure 9
<p>Discriminator model architecture.</p>
Full article ">Figure 10
<p>Comparison of superpixel segmentation algorithms on sandstone images: (<b>a</b>) original sandstone image; (<b>b</b>) FH; (<b>c</b>) QS; (<b>d</b>) SEEDS; (<b>e</b>) Watershed; (<b>f</b>) LSC; (<b>g</b>) SLIC; (<b>h</b>) GL-SLIC.</p>
Full article ">Figure 11
<p>Comparison of segmentation results between SLIC and GL-SLIC algorithms: (<b>a</b>) pre-segmentation result by the SLIC algorithm; (<b>b</b>) pre-segmentation result using the GL-SLIC algorithm.</p>
Full article ">Figure 12
<p>Detailed comparison between SLIC and GL-SLIC algorithms: (<b>a1</b>) detail area a1 from SLIC; (<b>a2</b>) detail area a2 from SLIC; (<b>a3</b>) detail area a3 from SLIC; (<b>b1</b>) detail area b1 from GL-SLIC; (<b>b2</b>) detail area b2 from GL-SLIC; (<b>b3</b>) detail area b3 from GL-SLIC.</p>
Full article ">Figure 13
<p>Comparison of superpixel merging in medium-coarse-grained quartz sandstone: (<b>a</b>) medium-coarse-grained quartz sandstone image; (<b>b</b>) pre-segmentation result; (<b>c</b>) result after superpixel merging.</p>
Full article ">Figure 14
<p>Iterative model training and data augmentation process using labeled and unlabeled rock data to mitigate overfitting and enhance classification accuracy: (<b>a</b>) primary model; (<b>b</b>) discriminator model.</p>
Full article ">Figure 15
<p>Curves of training and testing accuracy variation with epochs for the primary model.</p>
Full article ">Figure 16
<p>Classification accuracy analysis: (<b>a</b>) training set confusion matrix, (<b>b</b>) test set confusion matrix.</p>
Full article ">Figure 17
<p>Improved model accuracy post dataset cleansing and enhancement.</p>
Full article ">Figure 18
<p>Final confusion matrices for model evaluation: (<b>a</b>) training data confusion matrix, (<b>b</b>) testing data confusion matrix.</p>
Full article ">Figure 19
<p>Component identification results: (<b>a</b>) original petrographic thin section images; (<b>b</b>) proposed method results; (<b>c</b>) UNet-based semantic segmentation results.</p>
Full article ">
17 pages, 2393 KiB  
Article
A Modified Bio-Inspired Optimizer with Capsule Network for Diagnosis of Alzheimer Disease
by Praveena Ganesan, G. P. Ramesh, C. Puttamdappa and Yarlagadda Anuradha
Appl. Sci. 2024, 14(15), 6798; https://doi.org/10.3390/app14156798 - 4 Aug 2024
Cited by 6 | Viewed by 1082
Abstract
Recently, Alzheimer’s disease (AD) is one of the common neurodegenerative disorders, which primarily occurs in old age. Structural magnetic resonance imaging (sMRI) is an effective imaging technique used in clinical practice for determining the period of AD patients. An efficient deep learning framework [...] Read more.
Recently, Alzheimer’s disease (AD) is one of the common neurodegenerative disorders, which primarily occurs in old age. Structural magnetic resonance imaging (sMRI) is an effective imaging technique used in clinical practice for determining the period of AD patients. An efficient deep learning framework is proposed in this paper for AD detection, which is inspired from clinical practice. The proposed deep learning framework significantly enhances the performance of AD classification by requiring less processing time. Initially, in the proposed framework, the sMRI images are acquired from a real-time dataset and two online datasets including Australian Imaging, Biomarker and Lifestyle flagship work of ageing (AIBL), and Alzheimer’s Disease Neuroimaging Initiative (ADNI). Next, a fuzzy-based superpixel-clustering algorithm is introduced to segment the region of interest (RoI) in sMRI images. Then, the informative deep features are extracted in segmented RoI images by integrating the probabilistic local ternary pattern (PLTP), ResNet-50, and Visual Geometry Group (VGG)-16. Furthermore, the dimensionality reduction is accomplished by through the modified gorilla troops optimizer (MGTO). This process not only enhances the classification performance but also diminishes the processing time of the capsule network (CapsNet), which is employed to classify the classes of AD. In the MGTO algorithm, a quasi-reflection-based learning (QRBL) process is introduced for generating silverback’s quasi-refraction position for further improving the optimal position’s quality. The proposed fuzzy based superpixel-clustering algorithm and MGTO-CapsNet model obtained a pixel accuracy of 0.96, 0.94, and 0.98 and a classification accuracy of 99.88%, 96.38%, and 99.94% on the ADNI, real-time, and AIBL datasets, respectively. Full article
Show Figures

Figure 1

Figure 1
<p>Processes involved in the proposed deep learning framework.</p>
Full article ">Figure 2
<p>Acquired sample sMRI images: (<b>a</b>) ADNI dataset, (<b>b</b>) real-time dataset, and (<b>c</b>) AIBL dataset.</p>
Full article ">Figure 3
<p>Segmented hippocampus portion from the sMRIs: (<b>a</b>) ADNI dataset, (<b>b</b>) real-time dataset, and (<b>c</b>) AIBL dataset.</p>
Full article ">Figure 4
<p>Pictorial comparison of the state-of-the-art algorithms.</p>
Full article ">Figure 5
<p>Pictorial comparison of the MGTO-CapsNet model and the conventional models on the ADNI dataset.</p>
Full article ">Figure 6
<p>Pictorial comparison of the MGTO-CapsNet model and the conventional models on the real-time dataset.</p>
Full article ">Figure 7
<p>Pictorial comparison of the MGTO-CapsNet model and the conventional models on the AIBL dataset.</p>
Full article ">
23 pages, 7913 KiB  
Article
A Dual-Branch Fusion of a Graph Convolutional Network and a Convolutional Neural Network for Hyperspectral Image Classification
by Pan Yang and Xinxin Zhang
Sensors 2024, 24(14), 4760; https://doi.org/10.3390/s24144760 - 22 Jul 2024
Viewed by 994
Abstract
Semi-supervised graph convolutional networks (SSGCNs) have been proven to be effective in hyperspectral image classification (HSIC). However, limited training data and spectral uncertainty restrict the classification performance, and the computational demands of a graph convolution network (GCN) present challenges for real-time applications. To [...] Read more.
Semi-supervised graph convolutional networks (SSGCNs) have been proven to be effective in hyperspectral image classification (HSIC). However, limited training data and spectral uncertainty restrict the classification performance, and the computational demands of a graph convolution network (GCN) present challenges for real-time applications. To overcome these issues, a dual-branch fusion of a GCN and convolutional neural network (DFGCN) is proposed for HSIC tasks. The GCN branch uses an adaptive multi-scale superpixel segmentation method to build fusion adjacency matrices at various scales, which improves the graph convolution efficiency and node representations. Additionally, a spectral feature enhancement module (SFEM) enhances the transmission of crucial channel information between the two graph convolutions. Meanwhile, the CNN branch uses a convolutional network with an attention mechanism to focus on detailed features of local areas. By combining the multi-scale superpixel features from the GCN branch and the local pixel features from the CNN branch, this method leverages complementary features to fully learn rich spatial–spectral information. Our experimental results demonstrate that the proposed method outperforms existing advanced approaches in terms of classification efficiency and accuracy across three benchmark data sets. Full article
Show Figures

Figure 1

Figure 1
<p>An outline of the proposed DFGCN for HSIC. It consists of two branches: a GCN branch, based on multi-scale superpixel segmentation, and a CNN branch with an attention mechanism.</p>
Full article ">Figure 2
<p>Segmentation maps acquired from the Indian Pines data set using the first principal component (PC) and adaptive multi-scale superpixel segmentation. Superpixel numbers at varying scales make up the figures: (<b>a</b>) first PC, (<b>b</b>) 262, (<b>c</b>) 525, and (<b>d</b>) 1051.</p>
Full article ">Figure 3
<p>Implementation of the proposed spectral feature enhancement module.</p>
Full article ">Figure 4
<p>Structural diagram of SE attention mechanism.</p>
Full article ">Figure 5
<p>Indian Pines: (<b>a</b>) false-color synthetic image; (<b>b</b>) ground truth.</p>
Full article ">Figure 6
<p>Pavia University: (<b>a</b>) false-color synthetic image; (<b>b</b>) ground truth.</p>
Full article ">Figure 7
<p>Kennedy Space Center: (<b>a</b>) false-color synthetic image; (<b>b</b>) ground truth.</p>
Full article ">Figure 8
<p>Impact of the parameters of α and the spectral feature enhancement module.</p>
Full article ">Figure 9
<p>Performance of classification with varying values of γ in spectral feature enhancement.</p>
Full article ">Figure 10
<p>Maps of classification produced using the IndianP data set using various methods in an optimal close-up view: (<b>a</b>) original image; (<b>b</b>) ground truth; (<b>c</b>) RBF-SVM; (<b>d</b>) 2-D-CNN; (<b>e</b>) 3-D-CNN; (<b>f</b>) GCN; (<b>g</b>) S<sup>2</sup>GCN; (<b>h</b>) MDGCN; (<b>i</b>) SGML; and (<b>j</b>) DFGCN.</p>
Full article ">Figure 11
<p>Maps of classification produced using the PaviaU data set using various methods in an optimal close-up view: (<b>a</b>) original image; (<b>b</b>) ground truth; (<b>c</b>) RBF-SVM; (<b>d</b>) 2-D-CNN; (<b>e</b>) 3-D-CNN; (<b>f</b>) GCN; (<b>g</b>) S<sup>2</sup>GCN; (<b>h</b>) MDGCN; (<b>i</b>) SGML; and (<b>j</b>) DFGCN.</p>
Full article ">Figure 12
<p>Maps of classification produced using the KSC data set using various methods in an optimal close-up view: (<b>a</b>) original image; (<b>b</b>) ground truth; (<b>c</b>) RBF-SVM; (<b>d</b>) 2-D-CNN; (<b>e</b>) 3-D-CNN; (<b>f</b>) GCN; (<b>g</b>) S<sup>2</sup>GCN; (<b>h</b>) MDGCN; (<b>i</b>) SGML; and (<b>j</b>) DFGCN.</p>
Full article ">Figure 13
<p>ROC curves and AUC values of each category in the PU data set.</p>
Full article ">Figure 14
<p>Multi-scale channel importance visualization using the PaviaU data set. (<b>a</b>–<b>c</b>) are weight visualizations of different scales. The red boxes indicate that the importance of superpixels within these ranges remains highly consistent across all channels at different scales, while the blue, orange, and green boxes represent that the importance of all superpixels on channels within these ranges is almost the same.</p>
Full article ">Figure 15
<p>The visualization of features from the IndianP, PaviaU, and KSC data sets using 2-D t-SNE. (<b>a</b>–<b>c</b>) are the original feature spaces of labeled samples and (<b>d</b>–<b>f</b>) are the data distributions of labeled samples in the graph convolution feature space. Classes are represented by different colors.</p>
Full article ">
23 pages, 40689 KiB  
Article
Multiscale Feature Search-Based Graph Convolutional Network for Hyperspectral Image Classification
by Ke Wu, Yanting Zhan, Ying An and Suyi Li
Remote Sens. 2024, 16(13), 2328; https://doi.org/10.3390/rs16132328 - 26 Jun 2024
Cited by 3 | Viewed by 1360
Abstract
With the development of hyperspectral sensors, the availability of hyperspectral images (HSIs) has increased significantly, prompting advancements in deep learning-based hyperspectral image classification (HSIC) methods. Recently, graph convolutional networks (GCNs) have been proposed to process graph-structured data in non-Euclidean domains, and have been [...] Read more.
With the development of hyperspectral sensors, the availability of hyperspectral images (HSIs) has increased significantly, prompting advancements in deep learning-based hyperspectral image classification (HSIC) methods. Recently, graph convolutional networks (GCNs) have been proposed to process graph-structured data in non-Euclidean domains, and have been used for HSIC. The superpixel segmentation should be implemented first in the GCN-based methods, however, it is difficult to manually select the optimal superpixel segmentation sizes to obtain the useful information for classification. To solve this problem, we constructed a HSIC model based on a multiscale feature search-based graph convolutional network (MFSGCN) in this study. Firstly, pixel-level features of HSIs are extracted sequentially using 3D asymmetric decomposition convolution and 2D convolution. Then, superpixel-level features at different scales are extracted using multilayer GCNs. Finally, the neural architecture search (NAS) method is used to automatically assign different weights to different scales of superpixel features. Thus, a more discriminative feature map is obtained for classification. Compared with other GCN-based networks, the MFSGCN network can automatically capture features and obtain higher classification accuracy. The proposed MFSGCN model was implemented on three commonly used HSI datasets and compared to some state-of-the-art methods. The results confirm that MFSGCN effectively improves accuracy. Full article
(This article belongs to the Special Issue Deep Learning for the Analysis of Multi-/Hyperspectral Images II)
Show Figures

Figure 1

Figure 1
<p>Framework of the multiscale feature search-based graph convolutional network (MFSGCN) for HSIC.</p>
Full article ">Figure 2
<p>Graphical illustration of Indian Pines (IP), Pavia University (PU), and Salinas (SA). (<b>a</b>) False-color map. (<b>b</b>) Ground-truth map. (<b>c</b>) Color code.</p>
Full article ">Figure 3
<p>Multiscale feature weights of IP dataset search process. The experiments were repeated five times for the weights of encoder and decoder, respectively.</p>
Full article ">Figure 4
<p>Multiscale feature weights of PU dataset search process. The experiments were repeated five times for the weights of encoder and decoder, respectively.</p>
Full article ">Figure 5
<p>Multiscale feature weights of SA dataset search process. The experiments were repeated five times for the weights of encoder and decoder, respectively.</p>
Full article ">Figure 6
<p>Classification maps for the IP dataset. (<b>a</b>) False-color image. (<b>b</b>) Ground truth. (<b>c</b>) 3D CNN. (<b>d</b>) SSRN. (<b>e</b>) HybridSN. (<b>f</b>) A2S2K-ResNet. (<b>g</b>) MDGCN. (<b>h</b>) CEGCN. (<b>i</b>) MSSGU. (<b>j</b>) MFSGCN.</p>
Full article ">Figure 7
<p>Classification maps for the PU dataset. (<b>a</b>) False-color image. (<b>b</b>) Ground truth. (<b>c</b>) 3D CNN. (<b>d</b>) SSRN. (<b>e</b>) HybridSN. (<b>f</b>) A2S2K-ResNet. (<b>g</b>) MDGCN. (<b>h</b>) CEGCN. (<b>i</b>) MSSGU. (<b>j</b>) MFSGCN.</p>
Full article ">Figure 8
<p>Classification maps for the SA dataset. (<b>a</b>) False-color image. (<b>b</b>) Ground truth. (<b>c</b>) 3D CNN. (<b>d</b>) SSRN. (<b>e</b>) HybridSN. (<b>f</b>) A2S2K-ResNet. (<b>g</b>) MDGCN. (<b>h</b>) CEGCN. (<b>i</b>) MSSGU. (<b>j</b>) MFSGCN.</p>
Full article ">
21 pages, 12118 KiB  
Article
Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction
by Qiansen Dai, Chencong Ma and Qizhong Zhang
Sensors 2024, 24(13), 4072; https://doi.org/10.3390/s24134072 - 22 Jun 2024
Cited by 1 | Viewed by 1334
Abstract
Hyperspectral images (HSIs) possess an inherent three-order structure, prompting increased interest in extracting 3D features. Tensor analysis and low-rank representations, notably truncated higher-order SVD (T-HOSVD), have gained prominence for this purpose. However, determining the optimal order and addressing sensitivity to changes in data [...] Read more.
Hyperspectral images (HSIs) possess an inherent three-order structure, prompting increased interest in extracting 3D features. Tensor analysis and low-rank representations, notably truncated higher-order SVD (T-HOSVD), have gained prominence for this purpose. However, determining the optimal order and addressing sensitivity to changes in data distribution remain challenging. To tackle these issues, this paper introduces an unsupervised Superpixelwise Multiscale Adaptive T-HOSVD (SmaT-HOSVD) method. Leveraging superpixel segmentation, the algorithm identifies homogeneous regions, facilitating the extraction of local features to enhance spatial contextual information within the image. Subsequently, T-HOSVD is adaptively applied to the obtained superpixel blocks for feature extraction and fusion across different scales. SmaT-HOSVD harnesses superpixel blocks and low-rank representations to extract 3D features, effectively capturing both spectral and spatial information of HSIs. By integrating optimal-rank estimation and multiscale fusion strategies, it acquires more comprehensive low-rank information and mitigates sensitivity to data variations. Notably, when trained on subsets comprising 2%, 1%, and 1% of the Indian Pines, University of Pavia, and Salinas datasets, respectively, SmaT-HOSVD achieves impressive overall accuracies of 93.31%, 97.21%, and 99.25%, while maintaining excellent efficiency. Future research will explore SmaT-HOSVD’s applicability in deep-sea HSI classification and pursue additional avenues for advancing the field. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Schematic of the proposed SmaT-HOSVD-based HSI classification.</p>
Full article ">Figure 2
<p>Schematic diagram of Multiscale Adaptive T-HOSVD usage on HSI Global.</p>
Full article ">Figure 3
<p>Schematic diagram of SmaT-HOSVD on HSI.</p>
Full article ">Figure 4
<p>OA obtained by different methods with different training percentages over (<b>a</b>) IP, (<b>b</b>) PU, and (<b>c</b>) SD datasets.</p>
Full article ">Figure 5
<p>Classification results for the Indian Pines dataset (10% training samples): (<b>a</b>) Original; (<b>b</b>) Ground Truth; (<b>c</b>) SVM (66.09); (<b>d</b>) T-HOSVD (65.53); (<b>e</b>) SpectralFormer (60.23); (<b>f</b>) SSTN (89.79); (<b>g</b>) SuperPCA (90.9); (<b>h</b>) SpaSSA (75.14); (<b>i</b>) TensorSSA (85.22); (<b>j</b>) Ours (93.31).</p>
Full article ">Figure 6
<p>Classification results for the Pavia University dataset (1% training samples): (<b>a</b>) Origin; (<b>b</b>) Ground Truth; (<b>c</b>) SVM (88.81); (<b>d</b>) T-HOSVD (89.33); (<b>e</b>) SpectralFormer (77.94); (<b>f</b>) SSTN (95.12); (<b>g</b>) SuperPCA (85.34); (<b>h</b>) SpaSSA (93.61); (<b>i</b>) TensorSSA (94.78); (<b>j</b>) Ours (97.34).</p>
Full article ">Figure 7
<p>Classification results for the Salinas dataset (1% training samples): (<b>a</b>) Original; (<b>b</b>) Ground Truth; (<b>c</b>) SVM (86.47); (<b>d</b>) T-HOSVD (85.54); (<b>e</b>) SpectralFormer (86.55); (<b>f</b>) SSTN (97.66); (<b>g</b>) SuperPCA (98.89); (<b>h</b>) SpaSSA (97.06); (<b>i</b>) TensorSSA (90.00); (<b>j</b>) Ours (99.24).</p>
Full article ">Figure 7 Cont.
<p>Classification results for the Salinas dataset (1% training samples): (<b>a</b>) Original; (<b>b</b>) Ground Truth; (<b>c</b>) SVM (86.47); (<b>d</b>) T-HOSVD (85.54); (<b>e</b>) SpectralFormer (86.55); (<b>f</b>) SSTN (97.66); (<b>g</b>) SuperPCA (98.89); (<b>h</b>) SpaSSA (97.06); (<b>i</b>) TensorSSA (90.00); (<b>j</b>) Ours (99.24).</p>
Full article ">Figure 8
<p>The influence of parameter S on overall accuracy (OA) concerning (<b>a</b>) IP with 2% training, (<b>b</b>) PU with 1% training, and (<b>c</b>) SD with 1% training.</p>
Full article ">Figure 9
<p>The influence of reduction dimensionality n on overall accuracy (OA) concerning (<b>a</b>) IP with 2% training, (<b>b</b>) PU with 1% training, and (<b>c</b>) SD with 1% training.</p>
Full article ">Figure 10
<p>The influence of parameter T<sup>1</sup> and T<sup>2</sup> on overall accuracy (OA) concerning (<b>a</b>) IP with 2% training, (<b>b</b>) PU with 1% training, and (<b>c</b>) SD with 1% training.</p>
Full article ">
16 pages, 2675 KiB  
Article
Superpixel-Based Graph Convolutional Network for UAV Forest Fire Image Segmentation
by Yunjie Mu, Liyuan Ou, Wenjing Chen, Tao Liu and Demin Gao
Drones 2024, 8(4), 142; https://doi.org/10.3390/drones8040142 - 3 Apr 2024
Cited by 1 | Viewed by 1885
Abstract
Given the escalating frequency and severity of global forest fires, it is imperative to develop advanced detection and segmentation technologies to mitigate their impact. To address the challenges of these technologies, the development of deep learning-based forest fire surveillance has significantly accelerated. Nevertheless, [...] Read more.
Given the escalating frequency and severity of global forest fires, it is imperative to develop advanced detection and segmentation technologies to mitigate their impact. To address the challenges of these technologies, the development of deep learning-based forest fire surveillance has significantly accelerated. Nevertheless, the integration of graph convolutional networks (GCNs) in forest fire detection remains relatively underexplored. In this context, we introduce a novel superpixel-based graph convolutional network (SCGCN) for forest fire image segmentation. Our proposed method utilizes superpixels to transform images into a graph structure, thereby reinterpreting the image segmentation challenge as a node classification task. Additionally, we transition the spatial graph convolution operation to a GraphSAGE graph convolution mechanism, mitigating the class imbalance issue and enhancing the network’s versatility. We incorporate an innovative loss function to contend with the inconsistencies in pixel dimensions within superpixel clusters. The efficacy of our technique is validated on two different forest fire datasets, demonstrating superior performance compared to four alternative segmentation methodologies. Full article
(This article belongs to the Special Issue Drones for Wildfire and Prescribed Fire Science)
Show Figures

Figure 1

Figure 1
<p>Frame samples of the normal spectrum palette.</p>
Full article ">Figure 2
<p>Frame samples of thermal images of Fusion, WhiteHot, and GreenHot palettes from top row to the bottom row.</p>
Full article ">Figure 3
<p>Overview of the proposed SCGCN framework.</p>
Full article ">Figure 4
<p>Similarity graph construction of image superpixel area.</p>
Full article ">Figure 5
<p>The structure of GCN.</p>
Full article ">Figure 6
<p>Results from testing images. (<b>a</b>) Original images, (<b>b</b>) ground truth, (<b>c</b>) DeepLabv3+, (<b>d</b>) Unet++, (<b>e</b>) HRnet, (<b>f</b>) PSPnet, (<b>g</b>) SCGCN.</p>
Full article ">Figure 7
<p>Results from testing images. (<b>a</b>) Original images, (<b>b</b>) ground truth, (<b>c</b>) PSPnet, (<b>d</b>) HRnet, (<b>e</b>) Deeplabv3+, (<b>f</b>) Unet++, (<b>g</b>) SCGCN.</p>
Full article ">Figure 8
<p>The segmentation by SLIC of FLAME dataset. (<b>a</b>) Examples of the original images. (<b>b</b>,<b>c</b>) The superpixel representation for FLAME dataset; K is the number of superpixels (nodes in our graphs).</p>
Full article ">Figure 9
<p>The segmentation by SLIC of Chongli dataset. (<b>a</b>) Examples of the original images. (<b>b</b>,<b>c</b>) The superpixel representation for Chongli dataset; K is the number of superpixels (nodes in our graphs).</p>
Full article ">Figure 10
<p>Performance comparisons of different superpixel numbers when evaluating with FLAME and Chongli datasets. (<b>a</b>) FLAME dataset; (<b>b</b>) Chongli dataset.</p>
Full article ">
19 pages, 3610 KiB  
Article
Spatial-Pooling-Based Graph Attention U-Net for Hyperspectral Image Classification
by Qi Diao, Yaping Dai, Jiacheng Wang, Xiaoxue Feng, Feng Pan and Ce Zhang
Remote Sens. 2024, 16(6), 937; https://doi.org/10.3390/rs16060937 - 7 Mar 2024
Cited by 2 | Viewed by 1728
Abstract
In recent years, graph convolutional networks (GCNs) have attracted increasing attention in hyperspectral image (HSI) classification owing to their exceptional representation capabilities. However, the high computational requirements of GCNs have led most existing GCN-based HSI classification methods to utilize superpixels as graph nodes, [...] Read more.
In recent years, graph convolutional networks (GCNs) have attracted increasing attention in hyperspectral image (HSI) classification owing to their exceptional representation capabilities. However, the high computational requirements of GCNs have led most existing GCN-based HSI classification methods to utilize superpixels as graph nodes, thereby limiting the spatial topology scale and neglecting pixel-level spectral–spatial features. To address these limitations, we propose a novel HSI classification network based on graph convolution called the spatial-pooling-based graph attention U-net (SPGAU). Specifically, unlike existing GCN models that rely on fixed graphs, our model involves a spatial pooling method that emulates the region-growing process of superpixels and constructs multi-level graphs by progressively merging adjacent graph nodes. Inspired by the CNN classification framework U-net, SPGAU’s model has a U-shaped structure, realizing multi-scale feature extraction from coarse to fine and gradually fusing features from different graph levels. Additionally, the proposed graph attention convolution method adaptively aggregates adjacency information, thereby further enhancing feature extraction efficiency. Moreover, a 1D-CNN is established to extract pixel-level features, striking an optimal balance between enhancing the feature quality and reducing the computational burden. Experimental results on three representative benchmark datasets demonstrate that the proposed SPGAU outperforms other mainstream models both qualitatively and quantitatively. Full article
Show Figures

Figure 1

Figure 1
<p>The conceptual architecture of SPGAU. Firstly, the superpixels obtained by hyperspectral image segmentation (SLIC) are used as graph nodes, the hidden states of the nodes are the feature vectors extracted by the convolutional neural network, and the graph is constructed according to the spatial relationship between the superpixels. Then, multi-scale feature information is extracted by graph evolution through a set of stacked graph attention convolution and spatial pooling modules. Finally, the segmentation result of HSI is obtained by the softmax function. The blue dashed line indicates the skip connection, and the black arrow is the direction of the information flow.</p>
Full article ">Figure 2
<p>Progressive merging of superpixels and evolution of graph structure, where the top half is the superpixel merging process. Different colors represent different classes. We take the red region as an example to encode the superpixels and intuitively show the process of superpixel merging. Correspondingly, the bottom half is the graph structure evolution process inspired by the idea of superpixel merging.</p>
Full article ">Figure 3
<p>Illustration of the spatial pooling process. Given the calculated probabilities for merging nodes, the space pooling process involves multiple attempts to merge nodes and generate a new graph until an accepted graph is obtained. The generation of a new graph entails randomly merging some nodes with high merging probability, resulting in the creation of new edges.</p>
Full article ">Figure 4
<p>Classification maps obtained by eight different methods on the Indian Pines dataset. (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. (<b>c</b>) 3-DCNN. (<b>d</b>) GCN. (<b>e</b>) MDGCN. (<b>f</b>) SSRN. (<b>g</b>) SSGCN. (<b>h</b>) HybridSN. (<b>i</b>) SVM. (<b>j</b>) SPGAU.</p>
Full article ">Figure 5
<p>Classification maps obtained by eight different methods on the Pavia University dataset. (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. (<b>c</b>) 3-DCNN. (<b>d</b>) GCN. (<b>e</b>) MDGCN. (<b>f</b>) SSRN. (<b>g</b>) SSGCN. (<b>h</b>) HybridSN. (<b>i</b>) SVM. (<b>j</b>) SPGAU.</p>
Full article ">Figure 6
<p>Classification maps obtained by seven different methods on the WHU-Hi-LongKou dataset. (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. (<b>c</b>) 3-DCNN. (<b>d</b>) GCN. (<b>e</b>) SSRN. (<b>f</b>) SSGCN. (<b>g</b>) HybridSN. (<b>h</b>) SVM. (<b>i</b>) SPGAU.</p>
Full article ">Figure 7
<p>Classification accuracies of SPAGUN with different network depths. (<b>a</b>) Indian Pines. (<b>b</b>) Pavia University. (<b>c</b>) WHU-Hi-LongKou.</p>
Full article ">Figure 8
<p>The classification performance of various methods with different training set ratios on three datasets. (<b>a</b>) Indian Pines. (<b>b</b>) Pavia University. (<b>c</b>) WHU-Hi-LongKou.</p>
Full article ">
23 pages, 22134 KiB  
Article
Multiobjective Evolutionary Superpixel Segmentation for PolSAR Image Classification
by Boce Chu, Mengxuan Zhang, Kun Ma, Long Liu, Junwei Wan, Jinyong Chen, Jie Chen and Hongcheng Zeng
Remote Sens. 2024, 16(5), 854; https://doi.org/10.3390/rs16050854 - 29 Feb 2024
Cited by 1 | Viewed by 1081
Abstract
Superpixel segmentation has been widely used in the field of computer vision. The generations of PolSAR superpixels have also been widely studied for their feasibility and high efficiency. The initial numbers of PolSAR superpixels are usually designed manually by experience, which has a [...] Read more.
Superpixel segmentation has been widely used in the field of computer vision. The generations of PolSAR superpixels have also been widely studied for their feasibility and high efficiency. The initial numbers of PolSAR superpixels are usually designed manually by experience, which has a significant impact on the final performance of superpixel segmentation and the subsequent interpretation tasks. Additionally, the effective information of PolSAR superpixels is not fully analyzed and utilized in the generation process. Regarding these issues, a multiobjective evolutionary superpixel segmentation for PolSAR image classification is proposed in this study. It contains two layers, an automatic optimization layer and a fine segmentation layer. Fully considering the similarity information within the superpixels and the difference information among the superpixels simultaneously, the automatic optimization layer can determine the suitable number of superpixels automatically by the multiobjective optimization for PolSAR superpixel segmentation. Considering the difficulty of the search for accurate boundaries of complex ground objects in PolSAR images, the fine segmentation layer can further improve the qualities of superpixels by fully using the boundary information of good-quality superpixels in the evolution process for generating PolSAR superpixels. The experiments on different PolSAR image datasets validate that the proposed approach can automatically generate high-quality superpixels without any prior information. Full article
Show Figures

Figure 1

Figure 1
<p>Overall framework of MOES.</p>
Full article ">Figure 2
<p>Individual encoding.</p>
Full article ">Figure 3
<p>Differential evolution strategy.</p>
Full article ">Figure 4
<p>Individual encoding and production of new superpixel centers.</p>
Full article ">Figure 5
<p>Evolutionary operator with boundary information.</p>
Full article ">Figure 6
<p>Flevoland dataset. (<b>a</b>) Flevoland image (PauliRGB); (<b>b</b>) ground truth.</p>
Full article ">Figure 7
<p>Wei River in Xi’an dataset. (<b>a</b>) Wei River in Xi’an image (PauliRGB); (<b>b</b>) ground truth.</p>
Full article ">Figure 8
<p>San Francisco dataset. (<b>a</b>) San Francisco image (PauliRGB); (<b>b</b>) ground truth.</p>
Full article ">Figure 9
<p>Sensitivity of parameter <math display="inline"><semantics> <mrow> <mi>m</mi> <mi>p</mi> <mi>o</mi> <mi>l</mi> </mrow> </semantics></math> on different PolSAR datasets. (<b>a</b>) Flevoland dataset; (<b>b</b>) Wei River in Xi’an dataset; (<b>c</b>) San Francisco dataset.</p>
Full article ">Figure 10
<p>Sensitivities of parameters <math display="inline"><semantics> <mrow> <mi>p</mi> <mi>o</mi> <msub> <mi>p</mi> <mn>1</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>1</mn> </msub> </mrow> </semantics></math> in automatic optimization layer in Flevoland dataset. (<b>a</b>) UE; (<b>b</b>) BR.</p>
Full article ">Figure 11
<p>Sensitivities of parameters <math display="inline"><semantics> <mrow> <mi>p</mi> <mi>o</mi> <msub> <mi>p</mi> <mn>2</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>2</mn> </msub> </mrow> </semantics></math> in fine segmentation layer in Flevoland dataset. (<b>a</b>) UE; (<b>b</b>) BR.</p>
Full article ">Figure 12
<p>PFs of two layers of MOES in Flevoland dataset.</p>
Full article ">Figure 13
<p>PFs of two layers of MOES in Wei River in Xi’an dataset.</p>
Full article ">Figure 14
<p>PFs of two layers of MOES in San Francisco dataset.</p>
Full article ">Figure 15
<p>Visual superpixel segmentation results in Flevoland dataset. (<b>a</b>) SLIC; (<b>b</b>) SEEDS; (<b>c</b>) TP; (<b>d</b>) QS; (<b>e</b>) POL-HLT; (<b>f</b>) HCI; (<b>g</b>) MOES.</p>
Full article ">Figure 16
<p>The enlarged images of the selected regions in visual results in Flevoland dataset. (<b>a</b>) SLIC; (<b>b</b>) SEEDS; (<b>c</b>) TP; (<b>d</b>) QS; (<b>e</b>) POL-HLT; (<b>f</b>) HCI; (<b>g</b>) MOES.</p>
Full article ">Figure 17
<p>Visual superpixel segmentation results in Wei River in Xi’an dataset. (<b>a</b>) SLIC; (<b>b</b>) SEEDS; (<b>c</b>) TP; (<b>d</b>) QS; (<b>e</b>) POL-HLT; (<b>f</b>) HCI; (<b>g</b>) MOES.</p>
Full article ">Figure 18
<p>The enlarged images of the selected regions in visual results in Wei River in Xi’an dataset. (<b>a</b>) SLIC; (<b>b</b>) SEEDS; (<b>c</b>) TP; (<b>d</b>) QS; (<b>e</b>) POL-HLT; (<b>f</b>) HCI; (<b>g</b>) MOES.</p>
Full article ">Figure 19
<p>Visual superpixel segmentation results in San Francisco dataset. (<b>a</b>) SLIC; (<b>b</b>) SEEDS; (<b>c</b>) TP; (<b>d</b>) QS; (<b>e</b>) POL-HLT; (<b>f</b>) HCI; (<b>g</b>) MOES.</p>
Full article ">Figure 20
<p>The enlarged images of the selected regions in visual results in San Francisco dataset. (<b>a</b>) SLIC; (<b>b</b>) SEEDS; (<b>c</b>) TP; (<b>d</b>) QS; (<b>e</b>) POL-HLT; (<b>f</b>) HCI; (<b>g</b>) MOES.</p>
Full article ">
Back to TopTop