[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (84)

Search Parameters:
Keywords = differential channel attention

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 4035 KiB  
Article
Atmospheric Turbulence Intensity Image Acquisition Method Based on Convolutional Neural Network
by Yuan Mu, Liangping Zhou, Shiyong Shao, Zhiqiang Wang, Pei Tang, Zhiyuan Hu and Liwen Ye
Remote Sens. 2025, 17(1), 103; https://doi.org/10.3390/rs17010103 - 30 Dec 2024
Viewed by 331
Abstract
An algorithmic model of a neural network with channel attention and spatial attention (CASANet) is proposed to estimate the value of atmospheric coherence length, which in turn provides a quantitative description of atmospheric turbulence intensity. By processing the acquired spot image data, the [...] Read more.
An algorithmic model of a neural network with channel attention and spatial attention (CASANet) is proposed to estimate the value of atmospheric coherence length, which in turn provides a quantitative description of atmospheric turbulence intensity. By processing the acquired spot image data, the channel attention and spatial attention mechanisms are utilized, and the convolutional neural network learns the interdependence between the channel and space of the feature image and adaptively recalibrates the feature response in terms of the channel to increase the contribution of the foreground spot and suppress the background features. Based on the experimental data, an analysis of the CASANet model subject to turbulence intensity perturbations, fluctuations in outgoing power, and fluctuations in beam quality at the outlet is carried out. Comparison of the results of the convolutional neural network with those of the inverse method and the differential image motion method shows that the convolutional neural network is optimal in three evaluation indexes, namely, the mean deviation, the root-mean-square error, and the correlation coefficient, which are 2.74, 3.35, and 0.94, respectively. The convolutional neural network exhibits high accuracy under moderate and weak turbulence, and the estimation values under strong turbulence conditions are still mostly within the 95% confidence interval. The above results fully demonstrate that the proposed convolutional neural network method can effectively estimate the atmospheric coherence length, which provides technical support for the inversion of atmospheric turbulence intensity based on images. Full article
(This article belongs to the Section Atmospheric Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Structure of the CASANet model network for atmospheric coherence length calculation.</p>
Full article ">Figure 2
<p>Schematic diagram of CASA module structure.</p>
Full article ">Figure 3
<p>Flowchart of CASANet atmospheric coherence length calculation model training module.</p>
Full article ">Figure 4
<p>(<b>a</b>–<b>d</b>) Comparison curves of computational results of the inversion model, DIMM model, and CASANet model.</p>
Full article ">Figure 5
<p>Comparison of the CASANet model and inverse model affected by turbulence intensity. (<b>a</b>) r<sub>0</sub> ratio of the CASANet model to the DIMM model and its exponential fit, and (<b>b</b>) r<sub>0</sub> ratio of the inversion model to the DIMM model and its exponential fit.</p>
Full article ">Figure 6
<p>Trend of atmospheric coherence length and luminous power over time. (<b>a</b>) Under strong turbulence conditions, trends of atmospheric coherence length and outgoing optical power with time in two sets of experiments with serial numbers 65 and 78. (<b>b</b>) Under medium turbulence conditions, trends of atmospheric coherence length as well as outgoing optical power with time in two sets of experiments with serial numbers 91 and 125. (<b>c</b>) Under weak turbulence conditions, trends of atmospheric coherence length as well as outgoing optical power with time in two sets of experiments with serial numbers 96 and 155.</p>
Full article ">Figure 7
<p>(<b>a</b>–<b>d</b>) Comparison of beam quality <math display="inline"><semantics> <mrow> <msub> <mi>β</mi> <mn>0</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>β</mi> </mrow> </semantics></math> of the laser at 5 s.</p>
Full article ">Figure 8
<p>The beam quality as a function of <math display="inline"><semantics> <mrow> <mi>D</mi> <mo>/</mo> <msub> <mi>r</mi> <mn>0</mn> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>Effect of beam quality on the computational accuracy of CASANet models.</p>
Full article ">
20 pages, 3176 KiB  
Article
Spectral Weaver: A Study of Forest Image Classification Based on SpectralFormer
by Haotian Yu, Xuyang Li, Xinggui Xu, Hong Li and Xiangsuo Fan
Forests 2025, 16(1), 21; https://doi.org/10.3390/f16010021 - 26 Dec 2024
Viewed by 229
Abstract
In forest ecosystems, the application of hyperspectral (HS) imagery offers unprecedented opportunities for refined identification and classification. The diversity and complexity of forest cover make it challenging for traditional remote-sensing techniques to capture subtle spectral differences. Hyperspectral imagery, however, can reveal the nuanced [...] Read more.
In forest ecosystems, the application of hyperspectral (HS) imagery offers unprecedented opportunities for refined identification and classification. The diversity and complexity of forest cover make it challenging for traditional remote-sensing techniques to capture subtle spectral differences. Hyperspectral imagery, however, can reveal the nuanced changes in different tree species, vegetation health status, and soil composition through its nearly continuous spectral information. This detailed spectral information is crucial for the monitoring, management, and conservation of forest resources. While Convolutional Neural Networks (CNNs) have demonstrated excellent local context modeling capabilities in HS image classification, their inherent network architecture limits the exploration and representation of spectral feature sequence properties. To address this issue, we have rethought HS image classification from a sequential perspective and proposed a hybrid model, the Spectral Weaver, which combines CNNs and Transformers. The Spectral Weaver replaces the traditional Multi-Head Attention mechanism with a Channel Attention mechanism (MCA) and introduces Centre-Differential Convolutional Layers (Conv2d-cd) to enhance spatial feature extraction capabilities. Additionally, we designed a cross-layer skip connection that adaptively learns to fuse “soft” residuals, transferring memory-like components from shallow to deep layers. Notably, the proposed model is a highly flexible backbone network, adaptable to both hyperspectral and multispectral image inputs. In comparison to traditional Visual Transformers (ViT), the Spectral Weaver innovates in several ways: (1) It introduces the MCA mechanism to enhance the mining of spectral feature sequence properties; (2) It employs Centre-Differential Convolutional Layers to strengthen spatial feature extraction; (3) It designs cross-layer skip connections to reduce information loss; (4) It supports both multispectral and hyperspectral inputs, increasing the model’s flexibility and applicability. By integrating global and local features, our model significantly improves the performance of HS image classification. We have conducted extensive experiments on the Gaofen dataset, multispectral data, and multiple hyperspectral datasets, validating the superiority of the Spectral Weaver model in forest hyperspectral image classification. The experimental results show that our model achieves 98.59% accuracy on multispectral data, surpassing ViT’s 96.30%. On the Jilin-1 dataset, our proposed algorithm achieved an accuracy of 98.95%, which is 2.17% higher than ViT. The model significantly outperforms classic ViT and other state-of-the-art backbone networks in classification performance. Not only does it effectively capture the spectral features of forest vegetation, but it also significantly improves the accuracy and robustness of classification, providing strong technical support for the refined management and conservation of forest resources. Full article
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Situation of the study area.</p>
Full article ">Figure 2
<p>Schematic of the Spectrum Weaver architecture.</p>
Full article ">Figure 3
<p>Classification results obtained by different models on the Jilin 1 high-resolution L3D dataset, (<b>a</b>) Image. (<b>b</b>) ViT. (<b>c</b>) speformer. (<b>d</b>) m3ddcnn. (<b>e</b>) rssan. (<b>f</b>) ablstm. (<b>g</b>) dffn. (<b>h</b>) Spectral Weaver.</p>
Full article ">Figure 4
<p>Classification results obtained by different models on Sentinel 2 dataset, (<b>a</b>) Image. (<b>b</b>) ViT. (<b>c</b>) speformer. (<b>d</b>) m3ddcnn. (<b>e</b>) rssan. (<b>f</b>) ablstm. (<b>g</b>) dffn. (<b>h</b>) Spectral Weaver.</p>
Full article ">Figure 5
<p>Classification results obtained by different models on the HyRANK-Loukia dataset, (<b>a</b>) Image. (<b>b</b>) gt. (<b>c</b>) speformer. (<b>d</b>) m3ddcnn. (<b>e</b>) Spectral Weaver.</p>
Full article ">Figure 6
<p>Classification results obtained by different models on the WHU-Hi-HongHu dataset, (<b>a</b>) Image.(<b>b</b>) gt. (<b>c</b>) ViT. (<b>d</b>) speformer. (<b>e</b>) m3ddcnn. (<b>f</b>) rssan. (<b>g</b>) ablstm. (<b>h</b>) dffn. (<b>i</b>) ssftt. (<b>j</b>) Spectral Weaver.</p>
Full article ">Figure 7
<p>Classification results obtained by different models on the WHU-Hi-LongKou dataset, (<b>a</b>) Image.(<b>b</b>) gt. (<b>c</b>) ViT. (<b>d</b>) speformer. (<b>e</b>) m3ddcnn. (<b>f</b>) rssan. (<b>g</b>) ablstm. (<b>h</b>) dffn. (<b>i</b>) ssftt. (<b>j</b>) Spectral Weaver.</p>
Full article ">Figure 8
<p>Houston 2018 Dataset Data Classification Chart.</p>
Full article ">Figure 9
<p>Houston 2018 Dataset LOSS and correctness plot.</p>
Full article ">
24 pages, 7396 KiB  
Article
Smoke Detection Transformer: An Improved Real-Time Detection Transformer Smoke Detection Model for Early Fire Warning
by Baoshan Sun and Xin Cheng
Fire 2024, 7(12), 488; https://doi.org/10.3390/fire7120488 - 23 Dec 2024
Viewed by 487
Abstract
As one of the important features in the early stage of fires, the detection of smoke can provide a faster early warning of a fire, thus suppressing the spread of the fire in time. However, the features of smoke are not apparent; the [...] Read more.
As one of the important features in the early stage of fires, the detection of smoke can provide a faster early warning of a fire, thus suppressing the spread of the fire in time. However, the features of smoke are not apparent; the shape of smoke is not fixed, and it is easy to be confused with the background outdoors, which leads to difficulties in detecting smoke. Therefore, this study proposes a model called Smoke Detection Transformer (Smoke-DETR) for smoke detection, which is based on a Real-Time Detection Transformer (RT-DETR). Considering the limited computational resources of smoke detection devices, Enhanced Channel-wise Partial Convolution (ECPConv) is introduced to reduce the number of parameters and the amount of computation. This approach improves Partial Convolution (PConv) by using a selection strategy that selects channels containing more information for each convolution, thereby increasing the network’s ability to learn smoke features. To cope with smoke images with inconspicuous features and irregular shapes, the Efficient Multi-Scale Attention (EMA) module is used to strengthen the feature extraction capability of the backbone network. Additionally, in order to overcome the problem of smoke being easily confused with the background, the Multi-Scale Foreground-Focus Fusion Pyramid Network (MFFPN) is designed to strengthen the model’s attention to the foreground of images, which improves the accuracy of detection in situations where smoke is not well differentiated from the background. Experimental results demonstrate that Smoke-DETR has achieved significant improvements in smoke detection. In the self-building dataset, compared to RT-DETR, Smoke-DETR achieves a Precision that has reached 86.2%, marking an increase of 3.6 percentage points. Similarly, Recall has achieved 80%, showing an improvement of 3.6 percentage points. In terms of mAP50, it has reached 86.2%, with a 3.8 percentage point increase. Furthermore, mAP50 has reached 53.9%, representing a 3.6 percentage point increase. Full article
Show Figures

Figure 1

Figure 1
<p>Smoke-DETR network architecture.</p>
Full article ">Figure 2
<p>Illustration of Enhanced Channel-wise Partial Convolution.</p>
Full article ">Figure 3
<p>Processing flow of feature map by EMA.</p>
Full article ">Figure 4
<p>Improved backbone network structure of Smoke-DETR.</p>
Full article ">Figure 5
<p>Overall framework of the Multi-Scale Foreground-Focus Fusion Pyramid Network.</p>
Full article ">Figure 6
<p>Components of the Rectangular Self-Calibration Module.</p>
Full article ">Figure 7
<p>Details of pyramid context extraction.</p>
Full article ">Figure 8
<p>Presentation of smoke images within the dataset.</p>
Full article ">Figure 9
<p>Scatterplot of location and size distribution of actual labels.</p>
Full article ">Figure 10
<p>The results of valid sets for RT-DETR and Smoke-DETR during the training process.</p>
Full article ">Figure 11
<p>Comparison of parameters and FLOPs between RT-DETR and Smoke-DETR.</p>
Full article ">Figure 12
<p>Comparison of RT-DETR and Smoke-DETR in terms of four metrics: Precision, Recall, mAP50, and mAP95.</p>
Full article ">Figure 13
<p>The detection results using different methods. (<b>a</b>) The input image to be detected. (<b>b</b>) The manually labeled ground truth. (<b>c</b>) The detection result of RT-DETR. (<b>d</b>) The detection result of Smoke-DETR.</p>
Full article ">Figure 14
<p>Heat map results under different methods. A darker shade of red indicates a higher probability ascribed by the model to the presence of smoke in that particular section. (<b>a</b>) Input image. (<b>b</b>) The heat map of RT-DETR model). (<b>c</b>) The heat map of backbone with ECPConvBlock. (<b>d</b>) The heat map of further introducing MFFPN.</p>
Full article ">
18 pages, 14931 KiB  
Article
Wavelet-Driven Multi-Band Feature Fusion for RGB-T Salient Object Detection
by Jianxun Zhao, Xin Wen, Yu He, Xiaowei Yang and Kechen Song
Sensors 2024, 24(24), 8159; https://doi.org/10.3390/s24248159 - 20 Dec 2024
Viewed by 473
Abstract
RGB-T salient object detection (SOD) has received considerable attention in the field of computer vision. Although existing methods have achieved notable detection performance in certain scenarios, challenges remain. Many methods fail to fully utilize high-frequency and low-frequency features during information interaction among different [...] Read more.
RGB-T salient object detection (SOD) has received considerable attention in the field of computer vision. Although existing methods have achieved notable detection performance in certain scenarios, challenges remain. Many methods fail to fully utilize high-frequency and low-frequency features during information interaction among different scale features, limiting detection performance. To address this issue, we propose a method for RGB-T salient object detection that enhances performance through wavelet transform and channel-wise attention fusion. Through feature differentiation, we effectively extract spatial characteristics of the target, enhancing the detection capability for global context and fine-grained details. First, input features are passed through the channel-wise criss-cross module (CCM) for cross-modal information fusion, adaptively adjusting the importance of features to generate rich fusion information. Subsequently, the multi-scale fusion information is input into the feature selection wavelet transforme module (FSW), which selects beneficial low-frequency and high-frequency features to improve feature aggregation performance and achieves higher segmentation accuracy through long-distance connections. Extensive experiments demonstrate that our method outperforms 22 state-of-the-art methods. Full article
(This article belongs to the Special Issue Multi-Modal Image Processing Methods, Systems, and Applications)
Show Figures

Figure 1

Figure 1
<p>The design concept of our method.</p>
Full article ">Figure 2
<p>The overall architecture of our proposed method. Where <math display="inline"><semantics> <mrow> <mi>U</mi> <mi>P</mi> </mrow> </semantics></math> represents upsampling.</p>
Full article ">Figure 3
<p>The detailed architecture of CCM.</p>
Full article ">Figure 4
<p>The structure of an FSW.</p>
Full article ">Figure 5
<p>The dataset used in this experiment: (<b>a</b>) VT821, (<b>b</b>) VT1000, (<b>c</b>) VT5000.</p>
Full article ">Figure 6
<p>Quantitative comparison of our method with other state-of-the-art methods: (<b>a</b>) PR curve. (<b>b</b>) Fm curve.</p>
Full article ">Figure 7
<p>Qualitative comparison of our model with eleven recent state-of-the-art models.</p>
Full article ">Figure 8
<p>The visualization of ablation, where ’w/o’ stands for the absence of the corresponding module.</p>
Full article ">Figure 9
<p>Visualization of some typical failure cases in our method.</p>
Full article ">
18 pages, 4228 KiB  
Article
A Lightweight Method for Peanut Kernel Quality Detection Based on SEA-YOLOv5
by Zhixia Liu, Chunyu Wang, Xilin Zhong, Genhua Shi, He Zhang, Dexu Yang and Jing Wang
Agriculture 2024, 14(12), 2273; https://doi.org/10.3390/agriculture14122273 - 11 Dec 2024
Viewed by 566
Abstract
Peanuts are susceptible to defects such as insect damage, breakage, germinant, and mildew, leading to varying qualities of peanuts. The disparity in peanut kernel quality results in significant differences in their prices and economic value. Conducting real-time, accurate, and non-destructive quality inspections of [...] Read more.
Peanuts are susceptible to defects such as insect damage, breakage, germinant, and mildew, leading to varying qualities of peanuts. The disparity in peanut kernel quality results in significant differences in their prices and economic value. Conducting real-time, accurate, and non-destructive quality inspections of peanut kernels can effectively increase the utilization rate and commercial value of peanuts. Manual inspections are inefficient and subjective, while photoelectric sorting is costly and less precise. Therefore, this study proposes a peanut kernel quality detection algorithm based on an enhanced YOLO v5 model. Compared to other models, this model is practical, highly accurate, lightweight, and easy to integrate. Initially, YOLO v5s was chosen as the foundational training model through comparison. Subsequently, the original backbone network was replaced with a lightweight ShuffleNet v2 network to improve the model’s ability to differentiate features among various types of peanut kernels and reduce the parameters. The ECA (Efficient Channel Attention) mechanism was introduced into the C3 module to enhance feature extraction capabilities, thereby improving average accuracy. The CIoU loss function was replaced with the alpha-IoU loss function to boost detection accuracy. The experimental results indicated that the improved model, SEA-YOLOv5, achieved an accuracy of 98.8% with a parameter count of 0.47 M and an average detection time of 11.2 ms per image. When compared to other detection models, there was an improvement in accuracy, demonstrating the effectiveness of the proposed peanut kernel quality detection model. Furthermore, this model is suitable for deployment on resource-limited embedded devices such as mobile terminals, enabling real-time and precise detection of peanut kernel quality. Full article
(This article belongs to the Special Issue Agricultural Products Processing and Quality Detection)
Show Figures

Figure 1

Figure 1
<p>Comparison of samples in different categories: (<b>a</b>) complete; (<b>b</b>) mildew (<b>c</b>) breakage (<b>d</b>) germinant (<b>e</b>) broken (<b>f</b>) insect erosion; and (<b>g</b>) soil/stone.</p>
Full article ">Figure 2
<p>Peanut kernel data-acquisition test bed.</p>
Full article ">Figure 3
<p>Examples of peanut kernel images: (<b>a</b>) original image; (<b>b</b>) increased exposure; (<b>c</b>) changed brightness; (<b>d</b>) flip; (<b>e</b>) Gaussian noise; (<b>f</b>) salt-and-pepper noise.</p>
Full article ">Figure 4
<p>YOLOv5 model structure.</p>
Full article ">Figure 5
<p>ShuffleNetV2 building block.</p>
Full article ">Figure 6
<p>ECA mechanism structure diagram.</p>
Full article ">Figure 7
<p>IoU loss-function prediction box and real box. Note: the green part of the figure is the prediction box, the purple part is the real box; W<sub>g</sub> and H<sub>g</sub> represent the width and height of the minimum bounding box, respectively; W<sub>i</sub> and H<sub>i</sub> represent the width and height of the overlap between the real box and the prediction box, respectively; B and B<sup>gt</sup> represent the center points of the prediction and real boxes, respectively.</p>
Full article ">Figure 8
<p>Network structure of SEA-YOLOv5.</p>
Full article ">Figure 9
<p>Raspberry Pi 4B device.</p>
Full article ">Figure 10
<p>Raspberry Pi 4B test sample.</p>
Full article ">Figure 11
<p>Comparison of the detection results of the model.</p>
Full article ">
14 pages, 1464 KiB  
Article
An Improved Neural Network Model Based on DenseNet for Fabric Texture Recognition
by Li Tan, Qiang Fu and Jing Li
Sensors 2024, 24(23), 7758; https://doi.org/10.3390/s24237758 - 4 Dec 2024
Viewed by 576
Abstract
In modern knitted garment production, accurate identification of fabric texture is crucial for enabling automation and ensuring consistent quality control. Traditional manual recognition methods not only demand considerable human effort but also suffer from inefficiencies and are prone to subjective errors. Although machine [...] Read more.
In modern knitted garment production, accurate identification of fabric texture is crucial for enabling automation and ensuring consistent quality control. Traditional manual recognition methods not only demand considerable human effort but also suffer from inefficiencies and are prone to subjective errors. Although machine learning-based approaches have made notable advancements, they typically rely on manual feature extraction. This dependency is time-consuming and often limits recognition accuracy. To address these limitations, this paper introduces a novel model, called the Differentiated Leaning Weighted DenseNet (DLW-DenseNet), which builds upon the DenseNet architecture. Specifically, DLW-DenseNet introduces a learnable weight mechanism that utilizes channel attention to enhance the selection of relevant channels. The proposed mechanism reduces information redundancy and expands the feature search space of the model. To maintain the effectiveness of channel selection in the later stages of training, DLW-DenseNet incorportes a differentiated learning strategy. By assigning distinct learning rates to the learnable weights, the model ensures continuous and efficient channel selection throughout the training process, thus facilitating effective model pruning. Furthermore, in response to the absence of publicly available datasets for fabric texture recognition, we construct a new dataset named KF9 (knitted fabric). Compared to the fabric recognition network based on the improved ResNet, the recognition accuracy has increased by five percentage points, achieving a higher recognition rate. Experimental results demonstrate that DLW-DenseNet significantly outperforms other representative methods in terms of recognition accuracy on the KF9 dataset. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The original 4-layer DenseNet block.</p>
Full article ">Figure 2
<p>A 4-layer weighted DenseNet block. In this block, each feature channel is assigned a weight w, denoted by two subscripts. The first subscript indicates the layer to which the weight is applied, while the second subscript corresponds to the specific feature channel. For example, w<sub>4,1</sub> represents the weight applied by the fourth layer to the first feature channel. For clarity, only the weights associated with cross-layer feature channels are illustrated, while the weights within non-cross-layer channels are omitted.</p>
Full article ">Figure 3
<p>The process of input feature weighting involves multiplying each feature channel by its corresponding weight parameter, represented by the symbol <math display="inline"><semantics> <mrow> <mo>◯</mo> <mspace width="-8.5359pt"/> <mo>×</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>The 4-layer Weighted DenseNet block with the introduction of a differential learning strategy. The dashed lines indicate channels where the corresponding weights have been reduced to near zero during the learning process, effectively performing pruning.</p>
Full article ">Figure 5
<p>Photography setup, where “light” represents the strong, uniform lighting conditions. The “camera” refers to the high-resolution digital camera used for capturing detailed images, while the “table” serves as the platform for positioning the fabric samples. The “fabric” refers to the knitted fabric textures being photographed.</p>
Full article ">Figure 6
<p>Class cardinality of different categories in the KF9 knitted fabric dataset.</p>
Full article ">Figure 7
<p>The legend corresponding to the nine categories of knitted fabric textures.</p>
Full article ">Figure 8
<p>The comparison of performance among the improved ResNet, VGGNet-16, and our proposed network. The horizontal axis represents the number of iterations, while the vertical axis denotes classification accuracy. The solid line represents the smoothed data, whereas the dashed line corresponds to the original data.</p>
Full article ">Figure 9
<p>Comparison between the original DenseNet and the Weighted DenseNet.</p>
Full article ">Figure 10
<p>Comparison between the Weighted DenseNet and the DLW-DenseNet.</p>
Full article ">
23 pages, 7013 KiB  
Article
Attention Swin Transformer UNet for Landslide Segmentation in Remotely Sensed Images
by Bingxue Liu, Wei Wang, Yuming Wu and Xing Gao
Remote Sens. 2024, 16(23), 4464; https://doi.org/10.3390/rs16234464 - 28 Nov 2024
Viewed by 575
Abstract
The development of artificial intelligence makes it possible to rapidly segment landslides. However, there are still some challenges in landslide segmentation based on remote sensing images, such as low segmentation accuracy, caused by similar features, inhomogeneous features, and blurred boundaries. To address these [...] Read more.
The development of artificial intelligence makes it possible to rapidly segment landslides. However, there are still some challenges in landslide segmentation based on remote sensing images, such as low segmentation accuracy, caused by similar features, inhomogeneous features, and blurred boundaries. To address these issues, we propose a novel deep learning model called AST-UNet in this paper. This model is based on structure of SwinUNet, attaching a channel Attention and spatial intersection (CASI) module as a parallel branch of the encoder, and a spatial detail enhancement (SDE) module in the skip connection. Specifically, (1) the spatial intersection module expands the spatial attention range, alleviating noise in the image and enhances the continuity of landslides in segmentation results; (2) the channel attention module refines the spatial attention weights by feature modeling in the channel dimension, improving the model’s ability to differentiate targets that closely resemble landslides; and (3) the spatial detail enhancement module increases the accuracy for landslide boundaries by strengthening the attention of the decoder to detailed features. We use the landslide data from the area of Luding, Sichuan to conduct experiments. The comparative analyses with state-of-the-art (SOTA) models, including FCN, UNet, DeepLab V3+, TransFuse, TranUNet, and SwinUNet, prove the superiority of our AST-UNet for landslide segmentation. The generalization of our model is also verified in the experiments. The proposed AST-UNet obtains an F1-score of 90.14%, mIoU of 83.45%, foreground IoU of 70.81%, and Hausdorff distance of 3.73, respectively, on the experimental datasets. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Examples of the landslide images and the corresponding labels.</p>
Full article ">Figure 2
<p>The study region for landslide segmentation experiments.</p>
Full article ">Figure 3
<p>The architecture of the proposed AST-UNet: CASI represents the channel attention and spatial intersection module and SDE represents the spatial detail enhancement module.</p>
Full article ">Figure 4
<p>The interrelated characteristics between landslides and surroundings. (<b>a</b>) the correlation between landslides and surrounding water bodies; (<b>b</b>,<b>c</b>) the correlation between landslides and surrounding vegetation.</p>
Full article ">Figure 5
<p>The challenges in landslide segmentation. (<b>a</b>,<b>b</b>) the “hole” phenomenon within the landslide; (<b>c</b>) the noise within the landslide; (<b>d</b>–<b>f</b>) disruptive land features around landslides.</p>
Full article ">Figure 6
<p>The structure of the channel attention and spatial intersection (CASI) module.</p>
Full article ">Figure 7
<p>The comparison of feature maps from the SDE module and the swin transformer blocks.</p>
Full article ">Figure 8
<p>The structure of spatial detail enhancement (SDE) module.</p>
Full article ">Figure 9
<p>Landslide segmentation results comparison between the proposed AST-UNet and other comparative models on the validation datasets.</p>
Full article ">Figure 10
<p>Landslide segmentation results comparison between the proposed AST-UNet and other comparative models on the test datasets.</p>
Full article ">Figure 11
<p>The results of landslide segmentation with AST-UNet and the object-oriented algorithm.</p>
Full article ">Figure 12
<p>Bar charts of the comparison of model properties: (<b>a</b>) parameters of the models; (<b>b</b>) FLOPs of the models; (<b>c</b>) FPS of the models; (<b>d</b>) APC of the models; (<b>e</b>) APL of the models; and (<b>f</b>) UTB of the models.</p>
Full article ">Figure 13
<p>The improvement of evaluation indexes after adding our proposed modules: (<b>a</b>) the improvement from the CA module; (<b>b</b>) the improvement from the SI module; (<b>c</b>) the improvement from the SDE module.</p>
Full article ">Figure 14
<p>Visualizations of results extracted from ablation experiments.</p>
Full article ">
19 pages, 2872 KiB  
Article
Channel and Spatial Attention in Chest X-Ray Radiographs: Advancing Person Identification and Verification with Self-Residual Attention Network
by Hazem Farah, Akram Bennour, Neesrin Ali Kurdi, Samir Hammami and Mohammed Al-Sarem
Diagnostics 2024, 14(23), 2655; https://doi.org/10.3390/diagnostics14232655 - 25 Nov 2024
Viewed by 539
Abstract
Background/Objectives: In contrast to traditional biometric modalities, such as facial recognition, fingerprints, and iris scans or even DNA, the research orientation towards chest X-ray recognition has been spurred by its remarkable recognition rates. Capturing the intricate anatomical nuances of an individual’s skeletal structure, [...] Read more.
Background/Objectives: In contrast to traditional biometric modalities, such as facial recognition, fingerprints, and iris scans or even DNA, the research orientation towards chest X-ray recognition has been spurred by its remarkable recognition rates. Capturing the intricate anatomical nuances of an individual’s skeletal structure, the ribcage of the chest, lungs, and heart, chest X-rays have emerged as a focal point for identification and verification, especially in the forensic field, even in scenarios where the human body damaged or disfigured. Discriminative feature embedding is essential for large-scale image verification, especially in applying chest X-ray radiographs for identity identification and verification. This study introduced a self-residual attention-based convolutional neural network (SRAN) aimed at effective feature embedding, capturing long-range dependencies and emphasizing critical spatial features in chest X-rays. This method offers a novel approach to person identification and verification through chest X-ray categorization, relevant for biometric applications and patient care, particularly when traditional biometric modalities are ineffective. Method: The SRAN architecture integrated a self-channel and self-spatial attention module to minimize channel redundancy and enhance significant spatial elements. The attention modules worked by dynamically aggregating feature maps across channel and spatial dimensions to enhance feature differentiation. For the network backbone, a self-residual attention block (SRAB) was implemented within a ResNet50 framework, forming a Siamese network trained with triplet loss to improve feature embedding for identity identification and verification. Results: By leveraging the NIH ChestX-ray14 and CheXpert datasets, our method demonstrated notable improvements in accuracy for identity verification and identification based on chest X-ray images. This approach effectively captured the detailed anatomical characteristics of individuals, including skeletal structure, ribcage, lungs, and heart, highlighting chest X-rays as a viable biometric tool even in cases of body damage or disfigurement. Conclusions: The proposed SRAN with self-residual attention provided a promising solution for biometric identification through chest X-ray imaging, showcasing its potential for accurate and reliable identity verification where traditional biometric approaches may fall short, especially in postmortem cases or forensic investigations. This methodology could play a transformative role in both biometric security and healthcare applications, offering a robust alternative modality for identity verification. Full article
Show Figures

Figure 1

Figure 1
<p>Proposed architecture.</p>
Full article ">Figure 2
<p>Attention block architecture.</p>
Full article ">Figure 3
<p>Reverse attention block architecture.</p>
Full article ">Figure 4
<p>Identification system.</p>
Full article ">Figure 5
<p>Example of testing a sample for identification.</p>
Full article ">Figure 6
<p>Verification System.</p>
Full article ">Figure 7
<p>Training loss against epochs 40 and 80.</p>
Full article ">
25 pages, 314 KiB  
Article
Digital Literacy and the Livelihood Resilience of Livestock Farmers: Empirical Evidence from the Old Revolutionary Base Areas in Northwest China
by Xuefeng Ma, Liang Cheng, Yahui Li and Minjuan Zhao
Agriculture 2024, 14(11), 1941; https://doi.org/10.3390/agriculture14111941 - 31 Oct 2024
Viewed by 997
Abstract
Enhancing the livelihood resilience of livestock farmers in the old revolutionary base areas helps them to cope with the increasingly complex external risk shocks of recent years and promotes the sustainable development of regional agriculture. This study is based on survey data from [...] Read more.
Enhancing the livelihood resilience of livestock farmers in the old revolutionary base areas helps them to cope with the increasingly complex external risk shocks of recent years and promotes the sustainable development of regional agriculture. This study is based on survey data from 1047 livestock farmers in the Ningxia and Gansu provinces of the northwest old revolutionary base area. It incorporates the characteristics of livestock farmers and the elements of psychological capital into the sustainable livelihood analysis framework to construct a livelihood resilience index system. After measuring livelihood resilience, this paper uses a general linear regression model and a probit model to explore the impact and mechanism of digital literacy on the livelihood resilience of livestock farmers. The results show the following: (1) digital literacy has a significant positive effect on the livelihood resilience of livestock farmers, and the impact of different dimensions of digital literacy on different dimensions of livelihood resilience also varies. Additionally, this effect also shows the heterogeneity in different village clustering forms and different income groups. In areas inhabited by ethnic minorities and among moderate-income groups, the role of digital literacy on the livelihood resilience of livestock farmers is more significant. (2) The improvement of digital literacy has a significant positive impact on livelihood resilience through three different pathways: the “differential mode of association”, learning channels, and types of income. (3) Digital literacy has led to the psychological aspects of rural hollowing-out problems among livestock farmers, which is particularly evident in families with only one type of caregiving burden (either only left-behind elderly people or only left-behind children). This problem is more evident. Therefore, this paper poses that the advancement of agricultural and rural economic development in China should not only focus on the cultivation of farmers’ digital literacy but also accelerate the construction of digital infrastructure to ensure the long-term effective mechanism of improving digital literacy. At the same time, in the process of promoting digital rural areas, attention should be paid to the psychological isolation issues that the network era brings to farmers. Full article
(This article belongs to the Section Agricultural Economics, Policies and Rural Management)
22 pages, 2472 KiB  
Article
DASR-Net: Land Cover Classification Methods for Hybrid Multiattention Multispectral High Spectral Resolution Remote Sensing Imagery
by Xuyang Li, Xiangsuo Fan, Jinlong Fan, Qi Li, Yuan Gao and Xueqiang Zhao
Forests 2024, 15(10), 1826; https://doi.org/10.3390/f15101826 - 19 Oct 2024
Viewed by 1178
Abstract
The prompt acquisition of precise land cover categorization data is indispensable for the strategic development of contemporary farming practices, especially within the realm of forestry oversight and preservation. Forests are complex ecosystems that require precise monitoring to assess their health, biodiversity, and response [...] Read more.
The prompt acquisition of precise land cover categorization data is indispensable for the strategic development of contemporary farming practices, especially within the realm of forestry oversight and preservation. Forests are complex ecosystems that require precise monitoring to assess their health, biodiversity, and response to environmental changes. The existing methods for classifying remotely sensed imagery often encounter challenges due to the intricate spacing of feature classes, intraclass diversity, and interclass similarity, which can lead to weak perceptual ability, insufficient feature expression, and a lack of distinction when classifying forested areas at various scales. In this study, we introduce the DASR-Net algorithm, which integrates a dual attention network (DAN) in parallel with the Residual Network (ResNet) to enhance land cover classification, specifically focusing on improving the classification of forested regions. The dual attention mechanism within DASR-Net is designed to address the complexities inherent in forested landscapes by effectively capturing multiscale semantic information. This is achieved through multiscale null attention, which allows for the detailed examination of forest structures across different scales, and channel attention, which assigns weights to each channel to enhance feature expression using an improved BSE-ResNet bilinear approach. The two-channel parallel architecture of DASR-Net is particularly adept at resolving structural differences within forested areas, thereby avoiding information loss and the excessive fusion of features that can occur with traditional methods. This results in a more discriminative classification of remote sensing imagery, which is essential for accurate forest monitoring and management. To assess the efficacy of DASR-Net, we carried out tests with 10m Sentinel-2 multispectral remote sensing images over the Heshan District, which is renowned for its varied forestry. The findings reveal that the DASR-Net algorithm attains an accuracy rate of 96.36%, outperforming classical neural network models and the transformer (ViT) model. This demonstrates the scientific robustness and promise of the DASR-Net model in assisting with automatic object recognition for precise forest classification. Furthermore, we emphasize the relevance of our proposed model to hyperspectral datasets, which are frequently utilized in agricultural and forest classification tasks. DASR-Net’s enhanced feature extraction and classification capabilities are particularly advantageous for hyperspectral data, where the rich spectral information can be effectively harnessed to differentiate between various forest types and conditions. By doing so, DASR-Net contributes to advancing remote sensing applications in forest monitoring, supporting sustainable forestry practices and environmental conservation efforts. The findings of this study have significant practical implications for urban forestry management. The DASR-Net algorithm can enhance the accuracy of forest cover classification, aiding urban planners in better understanding and monitoring the status of urban forests. This, in turn, facilitates the development of effective forest conservation and restoration strategies, promoting the sustainable development of the urban ecological environment. Full article
Show Figures

Figure 1

Figure 1
<p>Sentinel-2 remote sensing image of the study area. (<b>a</b>) Heshan District cartography: displays the geographical layout of Heshan District in Hunan Province. (<b>b</b>) Sentinel-2 remote sensing imagery of Heshan District: Shows the different land cover types and their spatial distribution.</p>
Full article ">Figure 2
<p>Dataset training and testing sample ratio.</p>
Full article ">Figure 3
<p>Schematic of the DASR-Net architecture for the hyperspectral multispectral image classification task.</p>
Full article ">Figure 4
<p>Visual representation of the evolution in feature embedding via collective spectral embedding.</p>
Full article ">Figure 5
<p>DilateFormer attention mechanism.</p>
Full article ">Figure 6
<p>SE module.</p>
Full article ">Figure 7
<p>Schematic diagram of the improved BSE-ResNet structure based on the residual structure.</p>
Full article ">Figure 8
<p>Results of different algorithms. (<b>a</b>) Image. (<b>b</b>) SVM. (<b>c</b>) KNN. (<b>d</b>) RF. (<b>e</b>) CNN. (<b>f</b>) RNN. (<b>g</b>) Transformer (ViT). (<b>h</b>) SF. (<b>i</b>) DASR-Net.</p>
Full article ">Figure 9
<p>Classification maps of HyRANK-Loukia dataset ((<b>a</b>) Image. (<b>b</b>) Ground Truth. (<b>c</b>) SVM. (<b>d</b>) KNN. (<b>e</b>) RF. (<b>f</b>) CNN. (<b>g</b>) RNN. (<b>h</b>) Transformer (ViT). (<b>i</b>) SF. (<b>j</b>) DASR-Net).</p>
Full article ">Figure 10
<p>Classification maps of Pavia University dataset ((<b>a</b>) Image. (<b>b</b>) Ground Truth. (<b>c</b>) SVM. (<b>d</b>) KNN. (<b>e</b>) RF. (<b>f</b>) CNN. (<b>g</b>) RNN. (<b>h</b>) Transformer (ViT). (<b>i</b>) SF. (<b>j</b>) DASR-Net).</p>
Full article ">Figure 11
<p>Classification maps of WHU-Hi-HanChuan dataset ((<b>a</b>) Image. (<b>b</b>) Ground Truth. (<b>c</b>) SVM. (<b>d</b>) KNN. (<b>e</b>) RF. (<b>f</b>) CNN. (<b>g</b>) RNN. (<b>h</b>) Transformer (ViT). (<b>i</b>) SF. (<b>j</b>) DASR-Net).</p>
Full article ">
24 pages, 918 KiB  
Article
Quality of Service-Aware Multi-Objective Enhanced Differential Evolution Optimization for Time Slotted Channel Hopping Scheduling in Heterogeneous Internet of Things Sensor Networks
by Aida Vatankhah and Ramiro Liscano
Sensors 2024, 24(18), 5987; https://doi.org/10.3390/s24185987 - 15 Sep 2024
Viewed by 690
Abstract
The emergence of the Internet of Things (IoT) has attracted significant attention in industrial environments. These applications necessitate meeting stringent latency and reliability standards. To address this, the IEEE 802.15.4e standard introduces a novel Medium Access Control (MAC) protocol called Time Slotted Channel [...] Read more.
The emergence of the Internet of Things (IoT) has attracted significant attention in industrial environments. These applications necessitate meeting stringent latency and reliability standards. To address this, the IEEE 802.15.4e standard introduces a novel Medium Access Control (MAC) protocol called Time Slotted Channel Hopping (TSCH). Designing a centralized scheduling system that simultaneously achieves the required Quality of Service (QoS) is challenging due to the multi-objective optimization nature of the problem. This paper introduces a novel optimization algorithm, QoS-aware Multi-objective enhanced Differential Evolution optimization (QMDE), designed to handle the QoS metrics, such as delay and packet loss, across multiple services in heterogeneous networks while also achieving the anticipated service throughput. Through co-simulation between TSCH-SIM and Matlab, R2023a we conducted multiple simulations across diverse sensor network topologies and industrial QoS scenarios. The evaluation results illustrate that an optimal schedule generated by QMDE can effectively fulfill the QoS requirements of closed-loop supervisory control and condition monitoring industrial services in sensor networks from 16 to 100 nodes. Through extensive simulations and comparative evaluations against the Traffic-Aware Scheduling Algorithm (TASA), this study reveals the superior performance of QMDE, achieving significant enhancements in both Packet Delivery Ratio (PDR) and delay metrics. Full article
(This article belongs to the Special Issue Advanced Applications of WSNs and the IoT)
Show Figures

Figure 1

Figure 1
<p>Sample tree topology showing sink, transmitting nodes, and flows.</p>
Full article ">Figure 2
<p>Simple wireless network topology with an example TSCH schedule.</p>
Full article ">Figure 3
<p>QoS-oriented Multi-objective Differential Evolution Optimization flowchart.</p>
Full article ">Figure 4
<p>Sample of six pool statuses corresponding to six time slots.</p>
Full article ">Figure 5
<p>Process of mapping the generated matrix values to sensors for TSCH schedule creation: (<b>a</b>) random matrix generation, (<b>b</b>) normalization, (<b>c</b>) mapping the sensor’s position in the pool, and (<b>d</b>) assign nodes and matching pairs.</p>
Full article ">Figure 6
<p>Co-simulation: sequence diagram of QMDE using Matlab and TSCH-SIM.</p>
Full article ">Figure 7
<p>Optimization progress in scenario 5 with 64 nodes.</p>
Full article ">Figure 8
<p>Slotframe size of QMDE algorithm in various scenarios.</p>
Full article ">Figure 9
<p>Evaluation of delay between applications in (<b>a</b>) Scn 1, (<b>b</b>) Scn 2, (<b>c</b>) Scn 3, (<b>d</b>) Scn 4, and (<b>e</b>) Scn 5.</p>
Full article ">Figure 10
<p>Evaluation of PDR for two applications in (<b>a</b>) Scn 1, (<b>b</b>) Scn 2, (<b>c</b>) Scn 3, (<b>d</b>) Scn 4, and (<b>e</b>) Scn 5.</p>
Full article ">Figure 11
<p>Time complexity of QMDE algorithm in various scenarios.</p>
Full article ">Figure 12
<p>Delay comparison between QMDE and TASA in (<b>a</b>) Scn 1, (<b>b</b>) Scn 2, (<b>c</b>) Scn 3, (<b>d</b>) Scn 4, and (<b>e</b>) Scn 5.</p>
Full article ">Figure 13
<p>PDR comparison between QMDE and TASA in (<b>a</b>) Scn 1, (<b>b</b>) Scn 2, (<b>c</b>) Scn 3, (<b>d</b>) Scn 4, and (<b>e</b>) Scn 5.</p>
Full article ">
19 pages, 21202 KiB  
Article
Distribution Characteristics and Genesis Mechanism of Ground Fissures in Three Northern Counties of the North China Plain
by Chao Xue, Mingdong Zang, Zhongjian Zhang, Guoxiang Yang, Nengxiong Xu, Feiyong Wang, Cheng Hong, Guoqing Li and Fujiang Wang
Sustainability 2024, 16(18), 8027; https://doi.org/10.3390/su16188027 - 13 Sep 2024
Viewed by 879
Abstract
The North China Plain is among the regions most afflicted by ground fissure disasters in China. Recent urbanization has accelerated ground fissure activity in the three counties of the northern North China Plain, posing significant threats to both the natural environment and socioeconomic [...] Read more.
The North China Plain is among the regions most afflicted by ground fissure disasters in China. Recent urbanization has accelerated ground fissure activity in the three counties of the northern North China Plain, posing significant threats to both the natural environment and socioeconomic sustainability. Despite the increased attention, a lack of comprehensive understanding persists due to delayed recognition and limited research. This study conducted field visits and geological surveys across 43 villages and 80 sites to elucidate the spatial distribution patterns of ground fissures in the aforementioned counties. By integrating these findings with regional geological data, we formulated a causative model to explain ground fissure formation. Our analysis reveals a concentration of ground fissures near the Niuxi and Rongxi faults, with the former exhibiting the most extensive distribution. The primary manifestations of ground fissures include linear cracks and patch-shaped collapse pits, predominantly oriented in east-west and north-south directions, indicating tensile failure with minimal vertical displacement. Various factors contribute to ground fissure development, including fault activity, ancient river channel distribution, bedrock undulations, rainfall, and ground settlement. Fault activity establishes a concealed fracture system in shallow geotechnical layers, laying the groundwork for ground fissure formation. Additionally, the distribution of ancient river channels and bedrock undulations modifies regional stress fields, further facilitating ground fissure emergence. Rainfall and differential ground settlement serve as triggering mechanisms, exposing ground fissures at the surface. This research offers new insights into the causes of ground fissures in the northern North China Plain, providing crucial scientific evidence for sustaining both the natural environment and the socio-economic stability of the region. Full article
(This article belongs to the Topic Natural Hazards and Disaster Risks Reduction, 2nd Volume)
Show Figures

Figure 1

Figure 1
<p>Distribution of the ground fissures and the faults in the study area. F1—Rongxi fault, F2—Rongdong fault, F3—Rongcheng fault, F4—Niuxi fault, F5—Niudong fault; A1—Xushui depression, A2—Rongcheng uplift, A3—Langgu depression, A4—Niutuo Town uplift, A5—Baxian depression; A-A′—section line; and D<sub>1</sub> and D<sub>2</sub>—drilling wells.</p>
Full article ">Figure 2
<p>Tectonic profile A-A′ in the study area [<a href="#B37-sustainability-16-08027" class="html-bibr">37</a>]. Location of the profile line is shown in <a href="#sustainability-16-08027-f001" class="html-fig">Figure 1</a>. F1—Rongxi fault, F2—Rongdong fault, F3—Rongcheng fault, F4—Niuxi fault, F5—Niudong fault.</p>
Full article ">Figure 3
<p>Groundwater level change in the study area [<a href="#B38-sustainability-16-08027" class="html-bibr">38</a>].</p>
Full article ">Figure 4
<p>Stratigraphic column profile in the study area based on drilling data [<a href="#B34-sustainability-16-08027" class="html-bibr">34</a>,<a href="#B35-sustainability-16-08027" class="html-bibr">35</a>]. The sites of drilling wells D<sub>1</sub> and D<sub>2</sub> are shown in <a href="#sustainability-16-08027-f001" class="html-fig">Figure 1</a>.</p>
Full article ">Figure 5
<p>Distribution of the ground fissures in Zhangweizhuangtou village and surrounding areas: (<b>a</b>–<b>i</b>) typical photos of the ground fissures; f1—Zhangweizhuangtou ground fissure.</p>
Full article ">Figure 6
<p>Distribution of the ground fissures in Beihoutai village, Nanhoutai village, Jiaguang village and surrounding areas: (<b>a</b>–<b>c</b>,<b>e</b>) typical photos of the wall fissures; (<b>d</b>) typical photos of the house ground subsidence; and (<b>f</b>) typical photos of the house floor fissures; F1—Rongxi fault; f2—Beihoutai ground fissure.</p>
Full article ">Figure 7
<p>Distribution of the ground fissures in Longwanxi village and surrounding areas: (<b>a</b>–<b>h</b>) typical photos of the ground fissures; f3–f5: Longwanxi ground fissures.</p>
Full article ">Figure 8
<p>Distribution of the ground fissures in Beizhang village and surrounding areas: (<b>a</b>–<b>g</b>) typical photos of the wall fissures; f6—Beizhang ground fissure.</p>
Full article ">Figure 9
<p>Distribution of the ground fissures in Dongangezhuang village and surrounding areas: (<b>a</b>–<b>e</b>): typical photos of the ground fissures; f7–f9: Dongangezhuang ground fissures.</p>
Full article ">Figure 10
<p>Formation process of rainfall-induced ground fissures under the combined influence of fault activity and rainfall erosion: (<b>a</b>) fault activity initiates the formation of concealed fissures near the surface; (<b>b</b>) infiltration of surface water leads to erosion of the soil layer, migration of soil particles, widening of cracks, and the creation of cavities; (<b>c</b>) fissures propagate upward, causing surface soil to collapse into linear fissures or collapse pits.</p>
Full article ">Figure 11
<p>Contour map of land subsidence rate in the study area (2016) [<a href="#B39-sustainability-16-08027" class="html-bibr">39</a>]. f1–f9: typical ground fissures in the study area and the details are shown in <a href="#sustainability-16-08027-t002" class="html-table">Table 2</a>.</p>
Full article ">Figure 12
<p>Pre-Cenozoic bedrock buried depth contour map and paleochannel distribution map in the study area [<a href="#B40-sustainability-16-08027" class="html-bibr">40</a>,<a href="#B41-sustainability-16-08027" class="html-bibr">41</a>]. f1–f9: typical ground fissures in the study area and the details are shown in <a href="#sustainability-16-08027-t002" class="html-table">Table 2</a>.</p>
Full article ">Figure 13
<p>Formation process of palaeochannel-type ground fissures: (<b>a</b>) original formation state; (<b>b</b>) the initial pumping resulted in uneven settlement of the strata, resulting in a tensile stress concentration area at the shoulder of the palaeochannel and forming hidden cracks; (<b>c</b>) further pumping causes uneven ground settlement to intensify, and hidden cracks develop and then appear on the surface; and (<b>d</b>) stereogram of genetic mechanism of palaeochannel type ground fissures.</p>
Full article ">Figure 14
<p>Formation process of bedrock ridge-type ground fissures: (<b>a</b>) original formation state; (<b>b</b>) the initial pumping results in uneven settlement of the strata, resulting in a tensile stress concentration area at the bedrock ridge and forming hidden cracks; (<b>c</b>) further pumping causes uneven ground settlement to intensify, and hidden cracks develop and then appear on the surface; and (<b>d</b>) stereogram of genetic mechanism of bedrock ridge-type ground fissures.</p>
Full article ">Figure 15
<p>Formation process of bedrock step-type ground fissures: (<b>a</b>) original formation state; (<b>b</b>) the initial pumping results in uneven formation settlement, and the tension stress concentration area is generated in the sudden change of terrain, forming hidden cracks; (<b>c</b>) further pumping causes uneven ground settlement to intensify, and hidden cracks develop and then appear on the surface; and (<b>d</b>) stereogram of genetic mechanism of bedrock step-type ground fissures.</p>
Full article ">
23 pages, 11538 KiB  
Article
A Novel Deep Learning Model for Breast Tumor Ultrasound Image Classification with Lesion Region Perception
by Jinzhu Wei, Haoyang Zhang and Jiang Xie
Curr. Oncol. 2024, 31(9), 5057-5079; https://doi.org/10.3390/curroncol31090374 - 28 Aug 2024
Viewed by 1215
Abstract
Multi-task learning (MTL) methods are widely applied in breast imaging for lesion area perception and classification to assist in breast cancer diagnosis and personalized treatment. A typical paradigm of MTL is the shared-backbone network architecture, which can lead to information sharing conflicts and [...] Read more.
Multi-task learning (MTL) methods are widely applied in breast imaging for lesion area perception and classification to assist in breast cancer diagnosis and personalized treatment. A typical paradigm of MTL is the shared-backbone network architecture, which can lead to information sharing conflicts and result in the decline or even failure of the main task’s performance. Therefore, extracting richer lesion features and alleviating information-sharing conflicts has become a significant challenge for breast cancer classification. This study proposes a novel Multi-Feature Fusion Multi-Task (MFFMT) model to effectively address this issue. Firstly, in order to better capture the local and global feature relationships of lesion areas, a Contextual Lesion Enhancement Perception (CLEP) module is designed, which integrates channel attention mechanisms with detailed spatial positional information to extract more comprehensive lesion feature information. Secondly, a novel Multi-Feature Fusion (MFF) module is presented. The MFF module effectively extracts differential features that distinguish between lesion-specific characteristics and the semantic features used for tumor classification, and enhances the common feature information of them as well. Experimental results on two public breast ultrasound imaging datasets validate the effectiveness of our proposed method. Additionally, a comprehensive study on the impact of various factors on the model’s performance is conducted to gain a deeper understanding of the working mechanism of the proposed framework. Full article
(This article belongs to the Topic Artificial Intelligence in Cancer Pathology and Prognosis)
Show Figures

Figure 1

Figure 1
<p>Example of BUSI dataset preprocessing.</p>
Full article ">Figure 2
<p>The Multi-Feature Fusion Multi-Task model (MFFMT) architecture diagram.</p>
Full article ">Figure 3
<p>Schematic diagram of the CLEP module.</p>
Full article ">Figure 4
<p>Schematic diagram of the Coordinate Attention module (CA).</p>
Full article ">Figure 5
<p>Schematic diagram of Convolution Block Attention Module (CBAM).</p>
Full article ">Figure 6
<p>Schematic diagram of the Multi-Feature Fusion.</p>
Full article ">Figure 7
<p>Schematic diagram of the Differential Feature Fusion Module (DFFM).</p>
Full article ">Figure 8
<p>Schematic diagram of the Common Feature Enhancement Fusion Module (CFEFM).</p>
Full article ">Figure 9
<p><math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>18</mn> </msup> </semantics></math>: The results of ablation experiments on the BUSI dataset.</p>
Full article ">Figure 10
<p><math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>50</mn> </msup> </semantics></math>: The results of ablation experiments on the BUSI dataset.</p>
Full article ">Figure 11
<p><math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>18</mn> </msup> </semantics></math>: The results of ablation experiments on the MIBU dataset.</p>
Full article ">Figure 12
<p><math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>50</mn> </msup> </semantics></math>: The results of ablation experiments on the MIBU dataset.</p>
Full article ">Figure 13
<p>Accuracy rate of <math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>18</mn> </msup> </semantics></math> with different <math display="inline"><semantics> <mi>λ</mi> </semantics></math> values on the BUSI dataset.</p>
Full article ">Figure 14
<p>Accuracy rate of <math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>50</mn> </msup> </semantics></math> with different <math display="inline"><semantics> <mi>λ</mi> </semantics></math> values on the BUSI dataset.</p>
Full article ">Figure 15
<p>Accuracy rate of <math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>18</mn> </msup> </semantics></math> with different <math display="inline"><semantics> <mi>λ</mi> </semantics></math> values on the MIBU dataset.</p>
Full article ">Figure 16
<p>Accuracy rate of <math display="inline"><semantics> <msup> <mi>MFFMT</mi> <mn>50</mn> </msup> </semantics></math> with different <math display="inline"><semantics> <mi>λ</mi> </semantics></math> values on the MIBU dataset.</p>
Full article ">Figure 17
<p>Example of Grad-CAM heatmap on the BUSI dataset.</p>
Full article ">Figure 18
<p>Example of Grad-CAM heatmap on the MIBU dataset.</p>
Full article ">
25 pages, 4813 KiB  
Article
Land Cover Classification of Remote Sensing Imagery with Hybrid Two-Layer Attention Network Architecture
by Xiangsuo Fan, Xuyang Li and Jinlong Fan
Forests 2024, 15(9), 1504; https://doi.org/10.3390/f15091504 - 28 Aug 2024
Viewed by 921
Abstract
In remote sensing image processing, when categorizing images from multiple remote sensing data sources, the deepening of the network hierarchy is prone to the problems of feature dispersion, as well as the loss of semantic information. In order to solve this problem, this [...] Read more.
In remote sensing image processing, when categorizing images from multiple remote sensing data sources, the deepening of the network hierarchy is prone to the problems of feature dispersion, as well as the loss of semantic information. In order to solve this problem, this paper proposes to integrate a parallel network architecture HDAM-Net algorithm with a hybrid dual attention mechanism Hybrid dual attention mechanism for forest land cover change. Firstly, we propose a fusion MCA + SAM (MS) attention mechanism to improve VIT network, which can capture the correlation information between features; secondly, we propose a multilayer residual cascade convolution (MSCRC) network model using Double Cross-Attention Module (DCAM) attention mechanism, which is able to efficiently utilize the spatial dependency between multiscale encoder features: the spatial dependency between multiscale encoder features. Finally, the dual-channel parallel architecture is utilized to solve the structural differences and realize the enhancement of forestry image classification differentiation and effective monitoring of forest cover changes. In order to compare the performance of HDAM-Net, mountain urban forest types are classified based on multiple remote sensing data sources, and the performance of the model is evaluated. The experimental results show that the overall accuracy of the algorithm proposed in this paper is 99.42%, while the Transformer (ViT) is 96.92%, which indicates that the proposed classifier is able to accurately determine the cover type.The HDAM-Net model emphasizes the effectiveness in terms of accurately classifying the land, as well as the forest types by using multiple remote sensing data sources for predicting the future trend of the forest ecosystem. In addition, the land utilization rate and land cover change can clearly show the forest cover change and support the data to predict the future trend of the forest ecosystem so that the forest resource survey can effectively monitor deforestation and evaluate forest restoration projects. Full article
(This article belongs to the Special Issue Artificial Intelligence and Machine Learning Applications in Forestry)
Show Figures

Figure 1

Figure 1
<p>Research data graphs used for the research data in this paper.</p>
Full article ">Figure 2
<p>Example of class labels in the sample library of Heshan District.</p>
Full article ">Figure 3
<p>Schematic structure of the HDAM algorithm proposed in this paper.</p>
Full article ">Figure 4
<p>Half-sequence grouping spectral embedding schematic.</p>
Full article ">Figure 5
<p>DCA block.</p>
Full article ">Figure 6
<p>Testing accuracy of various algorithms in Xiangyin County study area: (<b>a</b>) image, (<b>b</b>) SVM, (<b>c</b>) KNN, (<b>d</b>) RF, (<b>e</b>) CNN, (<b>f</b>) RNN, (<b>g</b>) ViT, (<b>h</b>) HDAM-Net, (<b>i</b>) VITCNN.</p>
Full article ">Figure 7
<p>Results of different algorithms on the Houston dataset: (<b>a</b>) SVM. (<b>b</b>) KNN. (<b>c</b>) RF. (<b>d</b>) CNN. (<b>e</b>) RNN. (<b>f</b>) ViT. (<b>g</b>) SF. (<b>h</b>) HDAM-Net.</p>
Full article ">Figure 8
<p>Results of different algorithms on the Indian Pines dataset, as well as spatial distribution of Indian Pines training and test sets: (<b>a</b>) SVM. (<b>b</b>) KNN. (<b>c</b>) RF. (<b>d</b>) CNN. (<b>e</b>) RNN. (<b>f</b>) ViT. (<b>g</b>) SF. (<b>h</b>) HDAM-Net.</p>
Full article ">Figure 9
<p>Results of different algorithms on the Pavia University dataset: (<b>a</b>) SVM. (<b>b</b>) KNN. (<b>c</b>) RF. (<b>d</b>) CNN. (<b>e</b>) RNN. (<b>f</b>) ViT. (<b>g</b>) SF. (<b>h</b>) HDAM-Net.</p>
Full article ">Figure 10
<p>Loss curves of the proposed HDAM-Net algorithm during training.</p>
Full article ">Figure 11
<p>Spatial feature distribution of Sentinel-2 imagery in Heshan District—(<b>a</b>) 2019, (<b>b</b>) 2021, (<b>c</b>) 2023—for three years.</p>
Full article ">Figure 12
<p>Three -year tree migration of trees in Heshan District: (<b>a</b>) 2019–2021, (<b>b</b>) 2021–2023, (<b>c</b>) 2019–2023. Three-year dynamics of forests in Heshan district: (<b>d</b>) 2019–2021, (<b>e</b>) 2021–2023, (<b>f</b>) 2019–2023.</p>
Full article ">
23 pages, 7568 KiB  
Article
1D-CLANet: A Novel Network for NLoS Classification in UWB Indoor Positioning System
by Qiu Wang, Mingsong Chen, Jiajie Liu, Yongcheng Lin, Kai Li, Xin Yan and Chizhou Zhang
Appl. Sci. 2024, 14(17), 7609; https://doi.org/10.3390/app14177609 - 28 Aug 2024
Cited by 1 | Viewed by 1345
Abstract
Ultra-Wideband (UWB) technology is crucial for indoor localization systems due to its high accuracy and robustness in multipath environments. However, Non-Line-of-Sight (NLoS) conditions can cause UWB signal distortion, significantly reducing positioning accuracy. Thus, distinguishing between NLoS and LoS scenarios and mitigating positioning errors [...] Read more.
Ultra-Wideband (UWB) technology is crucial for indoor localization systems due to its high accuracy and robustness in multipath environments. However, Non-Line-of-Sight (NLoS) conditions can cause UWB signal distortion, significantly reducing positioning accuracy. Thus, distinguishing between NLoS and LoS scenarios and mitigating positioning errors is crucial for enhancing UWB system performance. This research proposes a novel 1D-ConvLSTM-Attention network (1D-CLANet) for extracting UWB temporal channel impulse response (CIR) features and identifying NLoS scenarios. The model combines the convolutional neural network (CNN) and Long Short-Term memory (LSTM) architectures to extract temporal CIR features and introduces the Squeeze-and-Excitation (SE) attention mechanism to enhance critical features. Integrating SE attention with LSTM outputs boosts the model’s ability to differentiate between various NLoS categories. Experimental results show that the proposed 1D-CLANet with SE attention achieves superior performance in differentiating multiple NLoS scenarios with limited computational resources, attaining an accuracy of 95.58%. It outperforms other attention mechanisms and the version of 1D-CLANet without attention. Compared to advanced methods, the SE-enhanced 1D-CLANet significantly improves the ability to distinguish between LoS and similar NLoS scenarios, such as human obstructions, enhancing overall recognition accuracy in complex environments. Full article
(This article belongs to the Section Electrical, Electronics and Communications Engineering)
Show Figures

Figure 1

Figure 1
<p>Example of NLoS and LoS propagation in a UWB IPS.</p>
Full article ">Figure 2
<p>Example of a trilateration-based 3-anchor positioning model: (<b>a</b>) positioning under LoS conditions, (<b>b</b>) positioning under NLoS conditions.</p>
Full article ">Figure 3
<p>CIR curve from typical (<b>a</b>) LoS, (<b>b</b>) other NLoS scenarios.</p>
Full article ">Figure 4
<p>The network structure diagram of 1D-CLANet.</p>
Full article ">Figure 5
<p>The structure diagram of 1D-CNN.</p>
Full article ">Figure 6
<p>The architecture of a LSTM cell.</p>
Full article ">Figure 7
<p>The architecture of SE attention block.</p>
Full article ">Figure 8
<p>Instruments and experimental environment. The anchor point (“<math display="inline"><semantics> <mo>Δ</mo> </semantics></math>”) and tag (“<math display="inline"><semantics> <mo>□</mo> </semantics></math>”) are positioned as shown. LoS ranging positions are shown in blue and NLoS ranging positions are shown in red. (<b>a</b>) Stairway passage, (<b>b</b>) office corridor.</p>
Full article ">Figure 9
<p>Explanation of network structures: (<b>a</b>) CNSM, (<b>b</b>) ResNet without ECA.</p>
Full article ">Figure 10
<p>ROC curve for different methods in NLoS binary classification.</p>
Full article ">Figure 11
<p>Performance comparison of different methods for NLoS multi-classification.</p>
Full article ">Figure 12
<p>Confusion matrix outcomes for multiclassification. The Sce.1, Sce.2, Sce.3, Sce.4, and Sce.5 correspond to LoS, human, glass, door, and wall, respectively. (<b>a</b>) LSTM. (<b>b</b>) SVM. (<b>c</b>) HQCNN. (<b>d</b>) MLP. (<b>e</b>) CNSM. (<b>f</b>) 1D-CLANet.</p>
Full article ">
Back to TopTop