[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (379)

Search Parameters:
Keywords = shadow extraction

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 10942 KiB  
Article
MambaShadowDet: A High-Speed and High-Accuracy Moving Target Shadow Detection Network for Video SAR
by Xiaowo Xu, Tianwen Zhang, Xiaoling Zhang, Wensi Zhang, Xiao Ke and Tianjiao Zeng
Remote Sens. 2025, 17(2), 214; https://doi.org/10.3390/rs17020214 - 9 Jan 2025
Viewed by 331
Abstract
Existing convolution neural network (CNN)-based video synthetic aperture radar (SAR) moving target shadow detectors are difficult to model long-range dependencies, while transformer-based ones often suffer from greater complexity. To handle these issues, this paper proposes MambaShadowDet, a novel lightweight deep learning (DL) detector [...] Read more.
Existing convolution neural network (CNN)-based video synthetic aperture radar (SAR) moving target shadow detectors are difficult to model long-range dependencies, while transformer-based ones often suffer from greater complexity. To handle these issues, this paper proposes MambaShadowDet, a novel lightweight deep learning (DL) detector based on a state space model (SSM), dedicated to high-speed and high-accuracy moving target shadow detection in video SAR images. By introducing SSM with the linear complexity into YOLOv8, MambaShadowDet effectively captures the global feature dependencies while relieving computational load. Specifically, it designs Mamba-Backbone, combining SSM and CNN to effectively extract both global contextual and local spatial information, as well as a slim path aggregation feature pyramid network (Slim-PAFPN) to enhance multi-level feature extraction and further reduce complexity. Abundant experiments on the Sandia National Laboratories (SNL) video SAR data show that MambaShadowDet achieves superior moving target shadow detection performance with a detection accuracy of 80.32% F1 score and an inference speed of 44.44 frames per second (FPS), outperforming existing models in both accuracy and speed. Full article
Show Figures

Figure 1

Figure 1
<p>Challenges in shadow detection for video SAR. (<b>a</b>) Background occlusion; (<b>b</b>) low contrast of shadows; (<b>c</b>) shape variation of shadows.</p>
Full article ">Figure 1 Cont.
<p>Challenges in shadow detection for video SAR. (<b>a</b>) Background occlusion; (<b>b</b>) low contrast of shadows; (<b>c</b>) shape variation of shadows.</p>
Full article ">Figure 2
<p>The network architecture of MambaShadowDet. Specifically, MambaShadowDet mainly consists of three parts, Mamba-Backbone (in the blue dashed box), Slim-PAFPN (in the orange dashed box), and YOLO-Head (in the gray dashed box).</p>
Full article ">Figure 3
<p>The network architecture of the SSM-Block. The SSM-Block mainly consists of three parts, i.e., SS2D, local-spatial block (LSB), and residual-gated block (RGB).</p>
Full article ">Figure 4
<p>The overall process of the SS2D algorithm. The SS2D algorithm consists of three primary steps: scan expansion, S6-Block, and scan merge.</p>
Full article ">Figure 5
<p>The network architecture of the local-spatial block (LSB).</p>
Full article ">Figure 6
<p>The network architecture of the residual-gated block (RGB).</p>
Full article ">Figure 7
<p>The network architecture of Slim-PAFPN. Specifically, Slim-PAFPN composed of PAFPN (Path Aggregation Feature Pyramid Network) with SCSP blocks.</p>
Full article ">Figure 8
<p>The network architecture of SCSP.</p>
Full article ">Figure 9
<p>The loss curves during training. (<b>a</b>) The Box_loss (i.e., regression loss) curve during training; (<b>b</b>) The Cls_loss (i.e., classification loss) curve during training; (<b>c</b>) The Dfl_loss (i.e., dynamic feature learning loss) curve during training. The blue curve denotes the loss result while the orange dashed line denotes the corresponding smooth curve.</p>
Full article ">Figure 10
<p>Qualitative results of different models in the SNL dataset. (<b>a</b>) YOLOX; (<b>b</b>) RetinaNet; (<b>c</b>) CenterNet; (<b>d</b>) Faster R-CNN; (<b>e</b>) Cascade R-CNN; (<b>f</b>) Deformable DETR; (<b>g</b>) ShadowDeNet; (<b>h</b>) MambaShadowDet (ours). The detection boxes are marked in green, where the number in the upper right corner represents the confidence score. The false alarm boxes are marked in red, and the missed detection boxes are marked in white. FP denotes false positives, i.e., the amount of false detection results, and FN denotes false negatives, i.e., the amount of missed detection results. The smaller the FP and FN, the better the detection results.</p>
Full article ">Figure 10 Cont.
<p>Qualitative results of different models in the SNL dataset. (<b>a</b>) YOLOX; (<b>b</b>) RetinaNet; (<b>c</b>) CenterNet; (<b>d</b>) Faster R-CNN; (<b>e</b>) Cascade R-CNN; (<b>f</b>) Deformable DETR; (<b>g</b>) ShadowDeNet; (<b>h</b>) MambaShadowDet (ours). The detection boxes are marked in green, where the number in the upper right corner represents the confidence score. The false alarm boxes are marked in red, and the missed detection boxes are marked in white. FP denotes false positives, i.e., the amount of false detection results, and FN denotes false negatives, i.e., the amount of missed detection results. The smaller the FP and FN, the better the detection results.</p>
Full article ">Figure 11
<p>The loss curves (including regression loss, i.e., box_cls, classification loss, i.e., cls_loss, and dynamic feature learning loss, i.e., dfl_loss) during training using different mini-batch sizes. The blue curve denotes the loss result while the orange dashed line denotes the corresponding smooth curve. (<b>a</b>) The loss curve with the mini-batch size of 1. (<b>b</b>) The loss curve with the mini-batch size of 2. (<b>c</b>) The loss curve with the mini-batch size of 4. (<b>d</b>) The loss curve with the mini-batch size of 8. (<b>e</b>) The loss curve with the mini-batch size of 16 (ours).</p>
Full article ">Figure 11 Cont.
<p>The loss curves (including regression loss, i.e., box_cls, classification loss, i.e., cls_loss, and dynamic feature learning loss, i.e., dfl_loss) during training using different mini-batch sizes. The blue curve denotes the loss result while the orange dashed line denotes the corresponding smooth curve. (<b>a</b>) The loss curve with the mini-batch size of 1. (<b>b</b>) The loss curve with the mini-batch size of 2. (<b>c</b>) The loss curve with the mini-batch size of 4. (<b>d</b>) The loss curve with the mini-batch size of 8. (<b>e</b>) The loss curve with the mini-batch size of 16 (ours).</p>
Full article ">
25 pages, 3292 KiB  
Article
Lane Detection Based on CycleGAN and Feature Fusion in Challenging Scenes
by Eric Hsueh-Chan Lu and Wei-Chih Chiu
Vehicles 2025, 7(1), 2; https://doi.org/10.3390/vehicles7010002 - 1 Jan 2025
Viewed by 355
Abstract
Lane detection is a pivotal technology of the intelligent driving system. By identifying the position and shape of the lane, the vehicle can stay in the correct lane and avoid accidents. Image-based deep learning is currently the most advanced method for lane detection. [...] Read more.
Lane detection is a pivotal technology of the intelligent driving system. By identifying the position and shape of the lane, the vehicle can stay in the correct lane and avoid accidents. Image-based deep learning is currently the most advanced method for lane detection. Models using this method already have a very good recognition ability in general daytime scenes, and can almost achieve real-time detection. However, these models often fail to accurately identify lanes in challenging scenarios such as night, dazzle, or shadows. Furthermore, the lack of diversity in the training data restricts the capacity of the models to handle different environments. This paper proposes a novel method to train CycleGAN with existing daytime and nighttime datasets. This method can extract features of different styles and multi-scales, thereby increasing the richness of model input. We use CycleGAN as a domain adaptation model combined with an image segmentation model to boost the model’s performance in different styles of scenes. The proposed consistent loss function is employed to mitigate performance disparities of the model in different scenarios. Experimental results indicate that our method enhances the detection performance of original lane detection models in challenging scenarios. This research helps improve the dependability and robustness of intelligent driving systems, ultimately making roads safer and enhancing the driving experience. Full article
Show Figures

Figure 1

Figure 1
<p>Illustration of the two-stage proposed method.</p>
Full article ">Figure 2
<p>The architecture of the proposed lane detection model training.</p>
Full article ">Figure 3
<p>The architecture of the proposed domain adaption model training.</p>
Full article ">Figure 4
<p>Qualitative analysis result of Grad-CAM.</p>
Full article ">Figure 5
<p>The example of challenge scenarios from the CULane test dataset. (<b>a</b>) Crowded; (<b>b</b>) Dazzle; (<b>c</b>) Shadow; (<b>d</b>) Arrow.</p>
Full article ">Figure 6
<p>The example of other challenge scenarios from the CULane test dataset. (<b>a</b>) Indoor; (<b>b</b>) wet ground.</p>
Full article ">Figure 7
<p>Visualization results of challenging scenes in the CULane test dataset. (<b>a</b>) Night; (<b>b</b>) crowded; (<b>c</b>) dazzle; (<b>d</b>) shadow; (<b>e</b>) arrow; (<b>f</b>) indoor; (<b>g</b>) wet ground.</p>
Full article ">Figure 7 Cont.
<p>Visualization results of challenging scenes in the CULane test dataset. (<b>a</b>) Night; (<b>b</b>) crowded; (<b>c</b>) dazzle; (<b>d</b>) shadow; (<b>e</b>) arrow; (<b>f</b>) indoor; (<b>g</b>) wet ground.</p>
Full article ">
28 pages, 1540 KiB  
Article
Integrating Macroeconomic and Technical Indicators into Forecasting the Stock Market: A Data-Driven Approach
by Saima Latif, Faheem Aslam, Paulo Ferreira and Sohail Iqbal
Economies 2025, 13(1), 6; https://doi.org/10.3390/economies13010006 - 31 Dec 2024
Viewed by 570
Abstract
Forecasting stock markets is challenging due to the influence of various internal and external factors compounded by the effects of globalization. This study introduces a data-driven approach to forecast S&P 500 returns by incorporating macroeconomic indicators including gold and oil prices, the volatility [...] Read more.
Forecasting stock markets is challenging due to the influence of various internal and external factors compounded by the effects of globalization. This study introduces a data-driven approach to forecast S&P 500 returns by incorporating macroeconomic indicators including gold and oil prices, the volatility index, economic policy uncertainty, the financial stress index, geopolitical risk, and shadow short rate, with ten technical indicators. We propose three hybrid deep learning models that sequentially combine convolutional and recurrent neural networks for improved feature extraction and predictive accuracy. These models include the deep belief network with gated recurrent units, the LeNet architecture with gated recurrent units, and the LeNet architecture combined with highway networks. The results demonstrate that the proposed hybrid models achieve higher forecasting accuracy than the single deep learning models. This outcome is attributed to the complementary strengths of convolutional networks in feature extraction and recurrent networks in pattern recognition. Additionally, an analysis using the Shapley method identifies the volatility index, financial stress index, and economic policy uncertainty as the most significant predictors, underscoring the effectiveness of our data-driven approach. These findings highlight the substantial impact of contemporary uncertainty factors on stock markets, emphasizing their importance in studies analyzing market behaviour. Full article
(This article belongs to the Special Issue Financial Market Volatility under Uncertainty)
Show Figures

Figure 1

Figure 1
<p>Data-Driven Approach for Forecasting S&amp;P 500 Returns.</p>
Full article ">Figure 2
<p>Merging Data.</p>
Full article ">Figure 3
<p>Architecture of a GRU Cell.</p>
Full article ">Figure 4
<p>Architecture of Recurrent Highway Network.</p>
Full article ">Figure 5
<p>Architecture of LeNet.</p>
Full article ">Figure 6
<p>Architecture of DBN.</p>
Full article ">Figure 7
<p>S&amp;P500 Return Prediction with Hybrid DBN-GRU Model.</p>
Full article ">Figure 8
<p>S&amp;P500 Return Prediction with Hybrid LeNet-GRU Model.</p>
Full article ">Figure 9
<p>S&amp;P500 Return Prediction with Hybrid LeNet-Highway Model.</p>
Full article ">Figure 10
<p>RMSE of the Applied Deep Models.</p>
Full article ">Figure 11
<p>MAE of the Applied Deep Models.</p>
Full article ">Figure 12
<p>MAPE of the Applied Deep Models.</p>
Full article ">Figure 13
<p>Comparison of Error Values of DBN, GRU and their Hybrid Model.</p>
Full article ">Figure 14
<p>Comparison of Error Values of LeNet, GRU and their Hybrid Model.</p>
Full article ">Figure 15
<p>Comparison of Error Values of LeNet, Highway Network and their Hybrid Model.</p>
Full article ">Figure 16
<p>SHAP Summary Plot for LeNet-Highway.</p>
Full article ">Figure 17
<p>Force Plot for LeNet-Highway.</p>
Full article ">
20 pages, 5692 KiB  
Article
Combining UAV Remote Sensing with Ensemble Learning to Monitor Leaf Nitrogen Content in Custard Apple (Annona squamosa L.)
by Xiangtai Jiang, Lutao Gao, Xingang Xu, Wenbiao Wu, Guijun Yang, Yang Meng, Haikuan Feng, Yafeng Li, Hanyu Xue and Tianen Chen
Agronomy 2025, 15(1), 38; https://doi.org/10.3390/agronomy15010038 - 27 Dec 2024
Viewed by 319
Abstract
One of the most important nutrients needed for fruit tree growth is nitrogen. For orchards to get targeted, well-informed nitrogen fertilizer, accurate, large-scale, real-time monitoring, and assessment of nitrogen nutrition is essential. This study examines the Leaf Nitrogen Content (LNC) of the custard [...] Read more.
One of the most important nutrients needed for fruit tree growth is nitrogen. For orchards to get targeted, well-informed nitrogen fertilizer, accurate, large-scale, real-time monitoring, and assessment of nitrogen nutrition is essential. This study examines the Leaf Nitrogen Content (LNC) of the custard apple tree, a noteworthy fruit tree that is extensively grown in China’s Yunnan Province. This study uses an ensemble learning technique based on multiple machine learning algorithms to effectively and precisely monitor the leaf nitrogen content in the tree canopy using multispectral canopy footage of custard apple trees taken via Unmanned Aerial Vehicle (UAV) across different growth phases. First, canopy shadows and background noise from the soil are removed from the UAV imagery by using spectral shadow indices across growth phases. The noise-filtered imagery is then used to extract a number of vegetation indices (VIs) and textural features (TFs). Correlation analysis is then used to determine which features are most pertinent for LNC estimation. A two-layer ensemble model is built to quantitatively estimate leaf nitrogen using the stacking ensemble learning (Stacking) principles. Random Forest (RF), Adaptive Boosting (ADA), Gradient Boosting Decision Trees (GBDT), Linear Regression (LR), and Extremely Randomized Trees (ERT) are among the basis estimators that are integrated in the first layer. By detecting and eliminating redundancy among base estimators, the Least Absolute Shrinkage and Selection Operator regression (Lasso)model used in the second layer improves nitrogen estimation. According to the analysis results, Lasso successfully finds redundant base estimators in the suggested ensemble learning approach, which yields the maximum estimation accuracy for the nitrogen content of custard apple trees’ leaves. With a root mean square error (RMSE) of 0.059 and a mean absolute error (MAE) of 0.193, the coefficient of determination (R2) came to 0. 661. The significant potential of UAV-based ensemble learning techniques for tracking nitrogen nutrition in custard apple leaves is highlighted by this work. Additionally, the approaches investigated might offer insightful information and a point of reference for UAV remote sensing applications in nitrogen nutrition monitoring for other crops. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Geographical location of the study area.</p>
Full article ">Figure 2
<p>DJI Mavic 3M.</p>
Full article ">Figure 3
<p>(<b>a</b>–<b>f</b>) show a comparison between the original images and the images with shadows and soil removed across three growth stages.</p>
Full article ">Figure 4
<p>Design of an Ensemble Learning Workflow for Estimating LNC in Custard Apple.</p>
Full article ">Figure 5
<p>Presents the heatmap of spectral features with moderate or stronger correlations with the LNC of custard apple.</p>
Full article ">Figure 6
<p>Heatmap of the correlation between custard apple LNC and the optimal input variables.</p>
Full article ">Figure 7
<p>This illustrates the remote sensing estimation of custard apple leaf nitrogen content using different learning methods. (<b>a</b>–<b>e</b>) represent the fitting curves of the base models RF, GBDT, ADA, ERT, and LR, respectively. (<b>f</b>) shows the fitting curve of the meta-model, and (<b>g</b>) presents the Lasso weight graph.</p>
Full article ">Figure 8
<p>Remote Sensing Monitoring of Custard Apple Leaf Nitrogen Content Based on UAV Multispectral Imagery. (<b>a</b>–<b>c</b>) represent the remote sensing monitoring images of leaf nitrogen content in May, August, and November, respectively.</p>
Full article ">
26 pages, 15301 KiB  
Article
Ethnic Architectural Heritage Identification Using Low-Altitude UAV Remote Sensing and Improved Deep Learning Algorithms
by Ting Luo, Xiaoqiong Sun, Weiquan Zhao, Wei Li, Linjiang Yin and Dongdong Xie
Buildings 2025, 15(1), 15; https://doi.org/10.3390/buildings15010015 - 24 Dec 2024
Viewed by 305
Abstract
Ethnic minority architecture is a vital carrier of the cultural heritage of ethnic minorities in China, and its quick and accurate extraction from remote sensing images is highly important for promoting the application of remote sensing information in urban management and architectural heritage [...] Read more.
Ethnic minority architecture is a vital carrier of the cultural heritage of ethnic minorities in China, and its quick and accurate extraction from remote sensing images is highly important for promoting the application of remote sensing information in urban management and architectural heritage protection. Taking Buyi architecture in China as an example, this paper proposes a minority architectural heritage identification method that combines low-altitude unmanned aerial vehicle (UAV) remote sensing technology and an improved deep learning algorithm. First, UAV images are used as the data source to provide high-resolution images for research on ethnic architecture recognition and to solve the problems associated with the high costs, time consumption, and destructiveness of traditional methods for ethnic architecture recognition. Second, to address the lack of edge pixel features in the sample images and reduce repeated labeling of the same sample, the ethnic architecture in entire remote sensing images is labeled on the Arcgis platform, and the sliding window method is used to cut the image data and the corresponding label file with a 10% overlap rate. Finally, an attention mechanism SE module is introduced to improve the DeepLabV3+ network model structure and achieve superior ethnic building recognition results. The experimental data fully show that the model’s accuracy reaches as high as 0.9831, with an excellent recall rate of 0.9743. Moreover, the F1 score is stable at a high level of 0.9787, which highlights the excellent performance of the model in terms of comprehensive evaluation indicators. Additionally, the intersection/union ratio (IoU) of the model is 0.9582, which further verifies its high precision in pixel-level recognition tasks. According to an in-depth comparative analysis, the innovative method proposed in this paper solves the problem of insufficient feature extraction of sample edge pixels and substantially reduces interference from complex environmental factors such as roads, building shadows, and vegetation with the recognition results for ethnic architecture. This breakthrough greatly improves the accuracy and robustness of the identification of architecture in low-altitude remote sensing images and provides strong technical support for the protection and intelligent analysis of architectural heritage. Full article
Show Figures

Figure 1

Figure 1
<p>Buyi stone–wood structure dry-column stone-slab house.</p>
Full article ">Figure 2
<p>Village distribution map.</p>
Full article ">Figure 3
<p>(<b>a</b>) Unmanned aerial vehicle and (<b>b</b>) camera.</p>
Full article ">Figure 4
<p>Architecture of the proposed automatic ethnic architecture recognition method.</p>
Full article ">Figure 5
<p>Image block strategy/sliding window method. The vector files (purple shapefile) come from the visual interpretation in ArcGIS.</p>
Full article ">Figure 6
<p>Example images and labels of ethnic buildings (areas represented by black and white pixels correspond to nonethnic and ethnic buildings, respectively).</p>
Full article ">Figure 7
<p>DeepLabv3+ network structure.</p>
Full article ">Figure 8
<p>Squeeze-and-excitation module structure.</p>
Full article ">Figure 9
<p>Loss value change curve of the model.</p>
Full article ">Figure 10
<p>Variation diagrams of (<b>a</b>) precision, (<b>b</b>) MIoU, and (<b>c</b>) recall with the number of iterations.</p>
Full article ">Figure 11
<p>(<b>a</b>) Partial validation samples, (<b>b</b>) the corresponding true values, and (<b>c</b>) the corresponding predicted values. Black represents the background and gray represents the Buyi architecture.</p>
Full article ">
18 pages, 7403 KiB  
Article
A Full-Scale Shadow Detection Network Based on Multiple Attention Mechanisms for Remote-Sensing Images
by Lei Zhang, Qing Zhang, Yu Wu, Yanfeng Zhang, Shan Xiang, Donghai Xie and Zeyu Wang
Remote Sens. 2024, 16(24), 4789; https://doi.org/10.3390/rs16244789 - 22 Dec 2024
Viewed by 419
Abstract
Shadows degrade image quality and complicate interpretation, underscoring the importance of accurate shadow detection for many image analysis tasks. However, due to the complex backgrounds and variable shadow characteristics of remote sensing images (RSIs), existing methods often struggle with accurately detecting shadows of [...] Read more.
Shadows degrade image quality and complicate interpretation, underscoring the importance of accurate shadow detection for many image analysis tasks. However, due to the complex backgrounds and variable shadow characteristics of remote sensing images (RSIs), existing methods often struggle with accurately detecting shadows of various scales and misclassifying dark, non-shaded areas as shadows. To address these issues, we proposed a comprehensive shadow detection network called MAMNet. Firstly, we proposed a multi-scale spatial channel attention fusion module, which extracted multi-scale features incorporating both spatial and channel information, allowing the model to flexibly adapt to shadows of different scales. Secondly, to address the issue of false detection in non-shadow areas, we introduced a criss-cross attention module, enabling non-shadow pixels to be compared with other shadow and non-shadow pixels in the same row and column, learning similar features of pixels in the same category, which improved the classification accuracy of non-shadow pixels. Finally, to address the issue of important information from the other two modules being lost due to continuous upsampling during the decoding phase, we proposed an auxiliary branch module to assist the main branch in decision-making, ensuring that the final output retained the key information from all stages. The experimental results demonstrated that the model outperformed the current state-of-the-art RSI shadow detection method on the aerial imagery dataset for shadow detection (AISD). The model achieved an overall accuracy (OA) of 97.50%, an F1 score of 94.07%, an intersection over union (IOU) of 88.87%, a precision of 95.06%, and a BER of 4.05%, respectively. Additionally, visualization results indicated that our model could effectively detect shadows of various scales while avoiding false detection in non-shadow areas. Therefore, this model offers an efficient solution for shadow detection in aerial imagery. Full article
Show Figures

Figure 1

Figure 1
<p>Existing challenges in remote-sensing image shadow detection.</p>
Full article ">Figure 2
<p>The entire flow of shadow detection.</p>
Full article ">Figure 3
<p>Illustration of the proposed MAMNet.</p>
Full article ">Figure 4
<p>Illustration of the proposed multi-scale spatial channel attention fusion module.</p>
Full article ">Figure 5
<p>Structure of the auxiliary branch.</p>
Full article ">Figure 6
<p>AISD dataset pictures and their labels.</p>
Full article ">Figure 7
<p>Comparison of the shadow detection results in the AISD dataset. (<b>a</b>) Input image. (<b>b</b>) Ground truth. (<b>c</b>) PSPNet. (<b>d</b>) ECANet. (<b>e</b>) MSASDNet. (<b>f</b>) MAMNet (ours).</p>
Full article ">
15 pages, 3149 KiB  
Article
Mapping Coastal Dynamics Induced Land Use Change in Sandwip Island, Bangladesh
by Philip Kibet Langat, Manoj Kumer Ghosh, Chandan Roy, Puspita Talukdar, Richard Koech and Arjun Neupane
Remote Sens. 2024, 16(24), 4686; https://doi.org/10.3390/rs16244686 - 16 Dec 2024
Viewed by 748
Abstract
Evaluating satellite water extraction indices, particularly for coastal environments, guarantees that satellite-derived water maps are as accurate and functional as possible, notwithstanding the unique complexities these areas present. Variability in salinity levels, intricate land-water boundaries, dynamic sediment loads, and tidal fluctuations often complicate [...] Read more.
Evaluating satellite water extraction indices, particularly for coastal environments, guarantees that satellite-derived water maps are as accurate and functional as possible, notwithstanding the unique complexities these areas present. Variability in salinity levels, intricate land-water boundaries, dynamic sediment loads, and tidal fluctuations often complicate coastal water mapping. Sandwip Island in Bangladesh is one of the most complex and dynamic coastal environments in the world and is our area of focus. Six water information extraction indices were evaluated: normalized-difference vegetation index (NDVI), modified normalized difference water index (MNDWI), automated water extraction index for built-up areas (AWEInsh) and shadows (AWEIsh), multi-band water index (MBWI), and normalized difference water index (NDWI), using Sandwip Island’s satellite Landsat imagery acquired in February 1990, 2000, 2010, and 2020. The results showed that NDWI performed the best based on the total area obtained and classification accuracy. NDWI was then used to assess the erosion and accretion dynamics of the island for the study period (1990–2020). In the period 1990–2000, the island saw significant erosion and accretion along its coastlines in all parts, while the 2000–2010 period indicated that the island eroded on all sides. However, the situation was totally opposite during 2010–2020. The results illustrated the best performance of the NDWI algorithm in mapping surface water in the complex and dynamic Sandwip coastal environment. Also, erosion and accretion change temporally and spatially on the island. While this study is confined to Sandwip Island in Bangladesh, the findings hold the potential for broader applicability in regions with comparable characteristics. Full article
Show Figures

Figure 1

Figure 1
<p>Location of the study area (Sandwip Island).</p>
Full article ">Figure 2
<p>Tidal water levels at Companiganj for the duration corresponding to the satellite image acquisition period. Red frames in the images represent image acquisition dates used in this research (Data source: Bangladesh Water Development Board).</p>
Full article ">Figure 3
<p>Locations of the ground control points (GCPs) used for image geo-rectification.</p>
Full article ">Figure 4
<p>Methodological structure used to determine the best water index in this study.</p>
Full article ">Figure 5
<p>Images produced using water indices of Sandwip Island, Bangladesh, to discriminate the land–water interface in 1990, 2000, 2010, and 2020. As different indices integrate different bands for discriminating land and water, the total area under land and water varied among the indices (see <a href="#remotesensing-16-04686-t002" class="html-table">Table 2</a>).</p>
Full article ">Figure 6
<p>Classified (land and water) images of Sandwip Island, Bangladesh, for 1990, 2000, 2010, and 2020.</p>
Full article ">Figure 7
<p>Coastal change of Sandwip Island, Bangladesh, based on image analysis from (<b>a</b>) 1990 to 2000, (<b>b</b>) 2000 to 2010, (<b>c</b>) 2010 to 2020, and (<b>d</b>) 1990–2020. Changes in the coastline were expressed for four sides to provide a better understanding.</p>
Full article ">
13 pages, 4284 KiB  
Article
Image Processing Technique for Enhanced Combustion Efficiency of Wood Pellets
by Thomas Gasperini, Andrea Pizzi, Lucia Olivi, Giuseppe Toscano, Alessio Ilari and Daniele Duca
Energies 2024, 17(23), 6144; https://doi.org/10.3390/en17236144 - 5 Dec 2024
Viewed by 556
Abstract
The combustion efficiency of wood pellets is partly affected by their average length. The ISO 17829 standard defines the methodology for assessing the average length of sample pellets, but the method does not always lead to representative data. Furthermore, a standard analysis is [...] Read more.
The combustion efficiency of wood pellets is partly affected by their average length. The ISO 17829 standard defines the methodology for assessing the average length of sample pellets, but the method does not always lead to representative data. Furthermore, a standard analysis is time-consuming as it requires manual measurement of the pellets using a caliper. This paper, whilst evaluating the effect of pellet length on combustion efficiency, proposes a pending-patented dimensional image processing method (DIP) for assessing pellet length. DIP allows the dimensional data of grouped and stacked pellets to be obtained by exploiting the shadows produced by pellets when exposed to a light source, assuming that different-sized pellets produce different shadows. Thus, the proposed method allows for the extraction of dimensional information from non-distinct objects, overcoming the reliance of classical image processing methods on object distance for effective segmentation. Combustion tests, carried out using pellets varying only in length, confirmed the influence of length on combustion efficiency. Shorter pellets, compared to longer ones, significantly reduced CO emissions by up to 94% (mg/MJ). However, they exhibited a higher fuel mass consumption rate (kg/h), with an increase of up to 22.8% compared to the longest sample. In addition, longer pellets produced fewer but larger shadows than shorter ones. Further studies are needed to correlate the number and size of shadows with samples’ average length so that DIP could be implemented in stoves and programmed to communicate with the control unit and automatically optimize the setting in order to improve combustion efficiency. Full article
(This article belongs to the Special Issue Decarbonization and Sustainability in Industrial and Tertiary Sectors)
Show Figures

Figure 1

Figure 1
<p>Diagram of the experimental combustion test system.</p>
Full article ">Figure 2
<p>Diagram of the DIP prototypal system.</p>
Full article ">Figure 3
<p>(<b>a</b>) Length class distribution of Sp, Mp, and Lp. (<b>b</b>) Boxplots of Sp2, Mp2, and Lp2.</p>
Full article ">Figure 4
<p>Average normalized CO data (mg/MJ) from 5 combustion tests. Each colored bin represents a different sample, whilst the red line represents the average temperature of flue gasses obtained during (<b>a</b>) program F3, (<b>b</b>) program F2, and (<b>c</b>) program F1.</p>
Full article ">Figure 5
<p>Average normalized NOx data (mg/MJ) from 5 combustion tests during the (<b>a</b>) F3, (<b>b</b>) F2, and (<b>c</b>) F1 programs. Each colored bin represents a different sample, the bars represent standard deviation, and the black dots represent the MFR of each sample.</p>
Full article ">Figure 6
<p>An extract of the images taken via DIP. The shorter pellet, Sp2, produced more but smaller shadows compared to the longer pellet Lp2.</p>
Full article ">Figure 7
<p>Distribution of objects detected via DIP in terms of Area classes.</p>
Full article ">Figure 8
<p>Biplots highlighting a difference between samples and confirming the inverse relationship between sample length and the number of objects detected via DIP analysis. The orange square corresponds to Sp2, the green square corresponds to Mp2, and the blue square corresponds to Lp2.</p>
Full article ">
15 pages, 3524 KiB  
Article
Effective Detection of Cloud Masks in Remote Sensing Images
by Yichen Cui, Hong Shen and Chan-Tong Lam
Sensors 2024, 24(23), 7730; https://doi.org/10.3390/s24237730 - 3 Dec 2024
Viewed by 511
Abstract
Effective detection of the contours of cloud masks and estimation of their distribution can be of practical help in studying weather changes and natural disasters. Existing deep learning methods are unable to extract the edges of clouds and backgrounds in a refined manner [...] Read more.
Effective detection of the contours of cloud masks and estimation of their distribution can be of practical help in studying weather changes and natural disasters. Existing deep learning methods are unable to extract the edges of clouds and backgrounds in a refined manner when detecting cloud masks (shadows) due to their unpredictable patterns, and they are also unable to accurately identify small targets such as thin and broken clouds. For these problems, we propose MDU-Net, a multiscale dual up-sampling segmentation network based on an encoder–decoder–decoder. The model uses an improved residual module to capture the multi-scale features of clouds more effectively. MDU-Net first extracts the feature maps using four residual modules at different scales, and then sends them to the context information full flow module for the first up-sampling. This operation refines the edges of clouds and shadows, enhancing the detection performance. Subsequently, the second up-sampling module concatenates feature map channels to fuse contextual spatial information, which effectively reduces the false detection rate of unpredictable targets hidden in cloud shadows. On a self-made cloud and cloud shadow dataset based on the Landsat8 satellite, MDU-Net achieves scores of 95.61% in PA and 84.97% in MIOU, outperforming other models in both metrics and result images. Additionally, we conduct experiments to test the model’s generalization capability on the landcover.ai dataset to show that it also achieves excellent performance in the visualization results. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The structure of MDU-Net.</p>
Full article ">Figure 2
<p>The structure of the residual module. (<b>a</b>) Downsampling Residual Module. (<b>b</b>) Standard Residual Module.</p>
Full article ">Figure 3
<p>The structure of dual up-sampling module.</p>
Full article ">Figure 4
<p>The backgrounds of self-made dataset. (<b>a</b>) Water areas. (<b>b</b>) Cities. (<b>c</b>) Vegetation. (<b>d</b>) Deserts.</p>
Full article ">Figure 5
<p>Visualization results of different models in rural and vegetated regions. (<b>a</b>) Original image, (<b>b</b>) Label image, (<b>c</b>) FCN, (<b>d</b>) UNet, (<b>e</b>) MultiResUNet, (<b>f</b>) PSPNet, (<b>g</b>) RSAGUNet, (<b>h</b>) AFMUNet, (<b>i</b>) MDU-Net.</p>
Full article ">Figure 6
<p>Prediction pictures of different algorithms in saline and snow-covered areas. (<b>a</b>) Original image, (<b>b</b>) Label image, (<b>c</b>) FCN, (<b>d</b>) UNet, (<b>e</b>) MultiResUNet, (<b>f</b>) PSPNet, (<b>g</b>) RSAGUNet, (<b>h</b>) AFMUNet, (<b>i</b>) MDU-Net.</p>
Full article ">Figure 7
<p>Visualization results of different models on landcover.ai dataset: (<b>a</b>) Original image, (<b>b</b>) FCN, (<b>c</b>) MultiResUNet, (<b>d</b>) ResUNet, (<b>e</b>) UNet, (<b>f</b>) PSPNet, (<b>g</b>) RSAGUNet, (<b>h</b>) AFMUNet, (<b>i</b>) MDU-Net.</p>
Full article ">
23 pages, 5897 KiB  
Article
A Large-Scale Building Unsupervised Extraction Method Leveraging Airborne LiDAR Point Clouds and Remote Sensing Images Based on a Dual P-Snake Model
by Zeyu Tian, Yong Fang, Xiaohui Fang, Yan Ma and Han Li
Sensors 2024, 24(23), 7503; https://doi.org/10.3390/s24237503 - 25 Nov 2024
Viewed by 445
Abstract
Automatic large-scale building extraction from the LiDAR point clouds and remote sensing images is a growing focus in the fields of the sensor applications and remote sensing. However, this building extraction task remains highly challenging due to the complexity of building sizes, shapes, [...] Read more.
Automatic large-scale building extraction from the LiDAR point clouds and remote sensing images is a growing focus in the fields of the sensor applications and remote sensing. However, this building extraction task remains highly challenging due to the complexity of building sizes, shapes, and surrounding environments. In addition, the discreteness, sparsity, and irregular distribution of point clouds, lighting, and shadows, as well as occlusions of the images, also seriously affect the accuracy of building extraction. To address the above issues, we propose a new unsupervised building extraction algorithm PBEA (Point and Pixel Building Extraction Algorithm) based on a new dual P-snake model (Dual Point and Pixel Snake Model). The proposed dual P-snake model is an enhanced active boundary model, which uses both point clouds and images simultaneously to obtain the inner and outer boundaries. The proposed dual P-snake model enables interaction and convergence between the inner and outer boundaries to improve the performance of building boundary detection, especially in complex scenes. Using the dual P-snake model and polygonization, this proposed PBEA can accurately extract large-scale buildings. We evaluated our PBEA and dual P-snake model on the ISPRS Vaihingen dataset and the Toronto dataset. The experimental results show that our PBEA achieves an area-based quality evaluation metric of 90.0% on the Vaihingen dataset and achieves the area-based quality evaluation metric of 92.4% on the Toronto dataset. Compared with other methods, our method demonstrates satisfactory performance. Full article
(This article belongs to the Special Issue Object Detection via Point Cloud Data)
Show Figures

Figure 1

Figure 1
<p>Flow chart of PBEA.</p>
Full article ">Figure 2
<p>Building boundary extraction from the point clouds: (<b>a</b>) Cluster results of FEC algorithm. (<b>b</b>) Boundary extracted by the boundary algorithm. (<b>c</b>) Projection onto the image.</p>
Full article ">Figure 3
<p>Building boundary extracted from the image: (<b>a</b>) Building boundary extracted by the edge closure algorithm, where white lines represent the canny edges and red lines indicate the extended closed edges. (<b>b</b>) Building boundary constrained by the point cloud projection boundary.</p>
Full article ">Figure 4
<p>Building boundary extracted by the dual P-snake model: (<b>a</b>) Comparison of the initial inner and outer boundaries. (<b>b</b>) Comparison among the initial inner boundary, initial outer boundary, and boundary extracted by the dual P-snake model. (<b>c</b>) Boundary extracted by the dual P-snake model.</p>
Full article ">Figure 5
<p>Influence of the sparse region of the point clouds on the building extraction: (<b>a</b>) Remote sensing image of the building. (<b>b</b>) Building boundary extracted from the sparse region of the point clouds. (<b>c</b>) Building boundary extracted from the image. (<b>d</b>) Building boundary extracted by the dual P-snake model.</p>
Full article ">Figure 5 Cont.
<p>Influence of the sparse region of the point clouds on the building extraction: (<b>a</b>) Remote sensing image of the building. (<b>b</b>) Building boundary extracted from the sparse region of the point clouds. (<b>c</b>) Building boundary extracted from the image. (<b>d</b>) Building boundary extracted by the dual P-snake model.</p>
Full article ">Figure 6
<p>Influence of the shadow region of the image on the building extraction: (<b>a</b>) Remote sensing image of the building. (<b>b</b>) Building boundary extracted from the shadow region of the image. (<b>c</b>) Building boundary extracted from the point clouds. (<b>d</b>) Building boundary extracted by the dual P-snake model.</p>
Full article ">Figure 7
<p>Influence of the occlusion regions of the image on the building extraction: (<b>a</b>) Remote sensing image of the building, and the lower right corner of the building is obscured by trees and shadows. (<b>b</b>) Building boundary extracted from the occlusion region of the image. (<b>c</b>) Building boundary extracted from the point clouds. (<b>d</b>) Building boundary extracted by the dual P-snake model.</p>
Full article ">Figure 7 Cont.
<p>Influence of the occlusion regions of the image on the building extraction: (<b>a</b>) Remote sensing image of the building, and the lower right corner of the building is obscured by trees and shadows. (<b>b</b>) Building boundary extracted from the occlusion region of the image. (<b>c</b>) Building boundary extracted from the point clouds. (<b>d</b>) Building boundary extracted by the dual P-snake model.</p>
Full article ">Figure 8
<p>Results of the building boundary polygonization.</p>
Full article ">Figure 9
<p>Performance of the snake models on buildings with varying complexity: (<b>a</b>) Simple rectangular building. (<b>b</b>) L-shaped building. (<b>c</b>) Building with the complex boundary.</p>
Full article ">Figure 10
<p>Buildings extracted from the Vaihingen dataset by our PBEA: (<b>a</b>) Area 1, (<b>b</b>) Area 2, (<b>c</b>) Area 3, (<b>d</b>) extraction results of Area 1, (<b>e</b>) extraction results of Area 2 (<b>f</b>) extraction results of Area 3.</p>
Full article ">Figure 10 Cont.
<p>Buildings extracted from the Vaihingen dataset by our PBEA: (<b>a</b>) Area 1, (<b>b</b>) Area 2, (<b>c</b>) Area 3, (<b>d</b>) extraction results of Area 1, (<b>e</b>) extraction results of Area 2 (<b>f</b>) extraction results of Area 3.</p>
Full article ">Figure 11
<p>Buildings extracted from the Toronto dataset by our PBEA: (<b>a</b>) Area 4, (<b>b</b>) Area 5, (<b>c</b>) extraction results of Area 4, (<b>d</b>) extraction results of Area 5.</p>
Full article ">
22 pages, 9723 KiB  
Article
AFENet: An Attention-Focused Feature Enhancement Network for the Efficient Semantic Segmentation of Remote Sensing Images
by Jiarui Li and Shuli Cheng
Remote Sens. 2024, 16(23), 4392; https://doi.org/10.3390/rs16234392 - 24 Nov 2024
Cited by 1 | Viewed by 530
Abstract
The semantic segmentation of high-resolution remote sensing images (HRRSIs) faces persistent challenges in handling complex architectural structures and shadow occlusions, limiting the effectiveness of existing deep learning approaches. To address these limitations, we propose an attention-focused feature enhancement network (AFENet) with a novel [...] Read more.
The semantic segmentation of high-resolution remote sensing images (HRRSIs) faces persistent challenges in handling complex architectural structures and shadow occlusions, limiting the effectiveness of existing deep learning approaches. To address these limitations, we propose an attention-focused feature enhancement network (AFENet) with a novel encoder–decoder architecture. The encoder architecture combines ResNet50 with a parallel multistage feature enhancement group (PMFEG), enabling robust feature extraction through optimized channel reduction, scale expansion, and channel reassignment operations. Building upon this foundation, we develop a global multi-scale attention mechanism (GMAM) in the decoder that effectively synthesizes spatial information across multiple scales by learning comprehensive global–local relationships. The architecture is further enhanced by an efficient feature-weighted fusion module (FWFM) that systematically integrates remote spatial features with local semantic information to improve segmentation accuracy. Experimental results across diverse scenarios demonstrate that AFENet achieves superior performance in building structure detection, exhibiting enhanced segmentation connectivity and completeness compared to state-of-the-art methods. Full article
Show Figures

Figure 1

Figure 1
<p>Challenges in the semantic segmentation of HRRSIs. (<b>a</b>) displays windows inside a building and a house extending outward, reflecting the rich texture variations within the building. (<b>b</b>) illustrates cars and buildings affected by shadows, which have indistinguishable segmentation edges. Furthermore, (<b>a</b>,<b>b</b>) show windows inside cars and buildings, respectively, which have very similar shapes, leading the model to easily confuse the two during segmentation.</p>
Full article ">Figure 2
<p>Overall architecture for attention-focused feature enhancement network.</p>
Full article ">Figure 3
<p>Illustration of parallel multistage feature enhancement group.</p>
Full article ">Figure 4
<p>The details of each component in GMAM. (<b>a</b>) Scale-variation relationship calculation: this module contains several matrices, and to clarify their product relationships, three priority levels are used in the diagram. The first priority represents the highest level. (<b>b</b>) Illustration of global multi-scale attention mechanism.</p>
Full article ">Figure 5
<p>Distribution of example images in datasets and the percentage of each category. (<b>a</b>,<b>b</b>) show example images and labels from the two datasets. (<b>c</b>) shows the proportion of pixels in each category relative to the total pixels in the dataset in response to the category imbalance in the dataset.</p>
Full article ">Figure 6
<p>Performance of each category in the Vaihingen dataset on the IoU metric.</p>
Full article ">Figure 7
<p>Qualitative comparison with state-of-the-art methods on the ISPRS Vaihingen dataset.</p>
Full article ">Figure 8
<p>Qualitative comparison with state-of-the-art methods on the ISPRS Potsdam dataset.</p>
Full article ">Figure 9
<p>The ablation studies of AFENet: (<b>a</b>) details its assembly from backbone to module integration and shows how each component affects mIoU performance, culminating in the complete AFENet. (<b>b</b>) demonstrates the portability of the modules within this network on different ResNet family backbones, showing that the performance remains stable across other backbones.</p>
Full article ">Figure 10
<p>Visualization of the effects of the progressive addition of PMFEG.</p>
Full article ">
31 pages, 4631 KiB  
Article
Environmental Impact of Wind Farms
by Mladen Bošnjaković, Filip Hrkać, Marija Stoić and Ivan Hradovi
Environments 2024, 11(11), 257; https://doi.org/10.3390/environments11110257 - 16 Nov 2024
Cited by 1 | Viewed by 2186
Abstract
The aim of this article is to analyse the global environmental impact of wind farms, i.e., the effects on human health and the local ecosystem. Compared to conventional energy sources, wind turbines emit significantly fewer greenhouse gases, which helps to mitigate global warming. [...] Read more.
The aim of this article is to analyse the global environmental impact of wind farms, i.e., the effects on human health and the local ecosystem. Compared to conventional energy sources, wind turbines emit significantly fewer greenhouse gases, which helps to mitigate global warming. During the life cycle of a wind farm, 86% of CO2 emissions are generated by the extraction of raw materials and the manufacture of wind turbine components. The water consumption of wind farms is extremely low. In the operational phase, it is 4 L/MWh, and in the life cycle, one water footprint is only 670 L/MWh. However, wind farms occupy a relatively large total area of 0.345 ± 0.224 km2/MW of installed capacity on average. For this reason, wind farms will occupy more than 10% of the land area in some EU countries by 2030. The impact of wind farms on human health is mainly reflected in noise and shadow flicker, which can cause insomnia, headaches and various other problems. Ice flying off the rotor blades is not mentioned as a problem. On a positive note, the use of wind turbines instead of conventionally operated power plants helps to reduce the emission of particulate matter 2.5 microns or less in diameter (PM 2.5), which are a major problem for human health. In addition, the non-carcinogenic toxicity potential of wind turbines for humans over the entire life cycle is one of the lowest for energy plants. Wind farms can have a relatively large impact on the ecological system and biodiversity. The destruction of animal migration routes and habitats, the death of birds and bats in collisions with wind farms and the negative effects of wind farm noise on wildlife are examples of these impacts. The installation of a wind turbine at sea generates a lot of noise, which can have a significant impact on some marine animals. For this reason, planners should include noise mitigation measures when selecting the site for the future wind farm. The end of a wind turbine’s service life is not a major environmental issue. Most components of a wind turbine can be easily recycled and the biggest challenge is the rotor blades due to the composite materials used. Full article
(This article belongs to the Collection Trends and Innovations in Environmental Impact Assessment)
Show Figures

Figure 1

Figure 1
<p>Average emissions of CO<sub>2</sub> eq.kg/MWh.</p>
Full article ">Figure 2
<p>Water footprint for different electricity generation technologies. The red line represents the range and the circle represents the median.</p>
Full article ">Figure 3
<p>Lifecycle human toxicity potential, non-carcinogenic. The red line represents the range and the circle represents the median.</p>
Full article ">Figure 4
<p>Lifecycle human toxicity potential, carcinogenic. The red line represents the range and the circle represents the median.</p>
Full article ">Figure 5
<p>Illustration of the noise level of wind turbines as a function of distance.</p>
Full article ">Figure 6
<p>Illustration of the flickering shadow effect, with permission of WKC Group.</p>
Full article ">Figure 7
<p>Share of land used by wind power.</p>
Full article ">Figure 8
<p>Development of the offshore wind farm project over time [<a href="#B124-environments-11-00257" class="html-bibr">124</a>].</p>
Full article ">Figure 9
<p>Sound transmission path of an offshore windturbine.</p>
Full article ">
17 pages, 5156 KiB  
Article
Identifying Alpine Lakes with Shoreline Features
by Zhimin Hu, Min Feng, Yijie Sui, Dezhao Yan, Kuo Zhang, Jinhao Xu, Rui Liu and Earina Sthapit
Water 2024, 16(22), 3287; https://doi.org/10.3390/w16223287 - 15 Nov 2024
Viewed by 861
Abstract
Alpine lakes located in high-altitude mountainous regions act as vital sentinels of environmental change. Remote-sensing-based identification of these lakes is crucial for understanding their response to climate variations and for assessing associated disaster risks. However, the complex terrain and weather conditions in these [...] Read more.
Alpine lakes located in high-altitude mountainous regions act as vital sentinels of environmental change. Remote-sensing-based identification of these lakes is crucial for understanding their response to climate variations and for assessing associated disaster risks. However, the complex terrain and weather conditions in these areas pose significant challenges to accurate detection. This paper proposes a method that leverages the high precision of deep learning for small lake and lake boundary extraction combined with deep learning to eliminate noise and errors in the identification results. Using Sentinel-2 data, we accurately identified and delineated alpine lakes in the eastern Himalayas. A total of 2123 lakes were detected, with an average lake area of 0.035 km². Notably, 76% of these lakes had areas smaller than 0.01 km². The slope data is crucial for the lake classification model in eliminating shadow noise. The accuracy of the proposed lake classification model reached 97.7%. In the identification of small alpine lakes, the recognition rate of this method was 96.4%, significantly surpassing that of traditional deep learning approaches. Additionally, this method effectively eliminated most shadow noise present in water body detection results obtained through machine learning techniques. Full article
(This article belongs to the Topic Advances in Hydrological Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Geographic extent and elevation distribution of the study area.</p>
Full article ">Figure 2
<p>Sentinel-2 RGB true-color imagery and corresponding labels for classification.</p>
Full article ">Figure 3
<p>Flow diagram of the methods in the study.</p>
Full article ">Figure 4
<p>Training loss for different input data configurations. Input1 consists of Sentinel-2 RGB images; Input2 combines Sentinel-2 RGB images with slope data; Input3 includes Sentinel-2 RGB images with MNDWI; and Input4 integrates Sentinel-2 RGB images with both slope and MNDWI data.</p>
Full article ">Figure 5
<p>Distribution density map of alpine lakes in the study area.</p>
Full article ">Figure 6
<p>(<b>a</b>) Distribution of alpine lakes by size, (<b>b</b>) distribution of alpine lakes by elevation.</p>
Full article ">Figure 7
<p>Sentinel-2 RGB composite displaying alpine lakes; (<b>a</b>) shows RF method results;(<b>b</b>) shows results from the proposed method. Blue indicates alpine lakes and red represents noise.</p>
Full article ">Figure 8
<p>Sentinel-2 RGB composite displaying alpine lakes, with columns labeled as follows: (<b>a</b>) for the manual interpretation results, (<b>b</b>) for the ConvNeXt model segmentation results, and (<b>c</b>) for the results obtained from the proposed method.</p>
Full article ">Figure 9
<p>The area comparison between the dataset from this method and the manually interpreted water body dataset. (<b>a</b>–<b>c</b>) Comparison of classification results with manual interpretation for small, medium, and large alpine lakes. (<b>d</b>–<b>f</b>) Comparison of segmentation results with manual interpretation for small, medium, and large alpine lakes.</p>
Full article ">
15 pages, 924 KiB  
Article
Novel Approach in Vegetation Detection Using Multi-Scale Convolutional Neural Network
by Fatema A. Albalooshi
Appl. Sci. 2024, 14(22), 10287; https://doi.org/10.3390/app142210287 - 8 Nov 2024
Viewed by 731
Abstract
Vegetation segmentation plays a crucial role in accurately monitoring and analyzing vegetation cover, growth patterns, and changes over time, which in turn contributes to environmental studies, land management, and assessing the impact of climate change. This study explores the potential of a multi-scale [...] Read more.
Vegetation segmentation plays a crucial role in accurately monitoring and analyzing vegetation cover, growth patterns, and changes over time, which in turn contributes to environmental studies, land management, and assessing the impact of climate change. This study explores the potential of a multi-scale convolutional neural network (MSCNN) design for object classification, specifically focusing on vegetation detection. The MSCNN is designed to integrate multi-scale feature extraction and attention mechanisms, enabling the model to capture both fine and coarse vegetation patterns effectively. Moreover, the MSCNN architecture integrates multiple convolutional layers with varying kernel sizes (3 × 3, 5 × 5, and 7 × 7), enabling the model to extract features at different scales, which is vital for identifying diverse vegetation patterns across various landscapes. Vegetation detection is demonstrated using three diverse datasets: the CamVid dataset, the FloodNet dataset, and the multispectral RIT-18 dataset. These datasets present a range of challenges, including variations in illumination, the presence of shadows, occlusion, scale differences, and cluttered backgrounds, which are common in real-world scenarios. The MSCNN architecture allows for the integration of information from multiple scales, facilitating the detection of diverse vegetation types under varying conditions. The performance of the proposed MSCNN method is rigorously evaluated and compared against state-of-the-art techniques in the field. Comprehensive experiments showcase the effectiveness of the approach, highlighting its robustness in accurately segmenting and classifying vegetation even in complex environments. The results indicate that the MSCNN design significantly outperforms traditional methods, achieving a remarkable global accuracy and boundary F1 score (BF score) of up to 98%. This superior performance underscores the MSCNN’s capability to enhance vegetation detection in imagery, making it a promising tool for applications in environmental monitoring and land use management. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Vegetation cover segmentation. (<b>a</b>) Input image, (<b>b</b>) percentage of vegetation cover marked in yellow (input image obtained from [<a href="#B4-applsci-14-10287" class="html-bibr">4</a>]).</p>
Full article ">Figure 2
<p>Proposed MSCNN network architecture.</p>
Full article ">Figure 3
<p>Visual results using Camvid dataset. (<b>a</b>) Input image, (<b>b</b>) ground truth, (<b>c</b>) proposed MSCNN result, (<b>d</b>) DeepLab V3+ result, (<b>e</b>) U-Net result, and (<b>f</b>) SegNet result.</p>
Full article ">Figure 4
<p>Visual results using FloodNet dataset. (<b>a</b>) Input image, (<b>b</b>) ground truth, (<b>c</b>) proposed MSCNN result, (<b>d</b>) DeepLab V3+ result, (<b>e</b>) U-Net result, and (<b>f</b>) SegNet result.</p>
Full article ">Figure 5
<p>Visual results using RIT-18 dataset. (<b>a</b>) Input RGB image, (<b>b</b>) ground truth, (<b>c</b>) proposed MSCNN result, (<b>d</b>) DeepLab V3+ result, (<b>e</b>) U-Net result, and (<b>f</b>) SegNet result.</p>
Full article ">Figure 6
<p>Proposed MSCNN network evaluation metrics compared to the DeepLab V3+, UNet, and SegNet networks’ performance.</p>
Full article ">
23 pages, 20937 KiB  
Article
Lunarminer Framework for Nature-Inspired Swarm Robotics in Lunar Water Ice Extraction
by Joven Tan, Noune Melkoumian, David Harvey and Rini Akmeliawati
Biomimetics 2024, 9(11), 680; https://doi.org/10.3390/biomimetics9110680 - 7 Nov 2024
Viewed by 990
Abstract
The Lunarminer framework explores the use of biomimetic swarm robotics, inspired by the division of labor in leafcutter ants and the synchronized flashing of fireflies, to enhance lunar water ice extraction. Simulations of water ice extraction within Shackleton Crater showed that the framework [...] Read more.
The Lunarminer framework explores the use of biomimetic swarm robotics, inspired by the division of labor in leafcutter ants and the synchronized flashing of fireflies, to enhance lunar water ice extraction. Simulations of water ice extraction within Shackleton Crater showed that the framework may improve task allocation, by reducing the extraction time by up to 40% and energy consumption by 31% in scenarios with high ore block quantities. This system, capable of producing up to 181 L of water per day from excavated regolith with a conversion efficiency of 0.8, may allow for supporting up to eighteen crew members. It has demonstrated robust fault tolerance and sustained operational efficiency, even for a 20% robot failure rate. The framework may help to address key challenges in lunar resource extraction, particularly in the permanently shadowed regions. To refine the proposed strategies, it is recommended that further studies be conducted on their large-scale applications in space mining operations at the Extraterrestrial Environmental Simulation (EXTERRES) laboratory at the University of Adelaide. Full article
(This article belongs to the Special Issue Recent Advances in Robotics and Biomimetics)
Show Figures

Figure 1

Figure 1
<p>Landing sites near Shackleton Crater on the lunar south pole, marked with blue squares, and highlighting the geological formations near Shackleton Crater [<a href="#B41-biomimetics-09-00680" class="html-bibr">41</a>].</p>
Full article ">Figure 2
<p>Bio-inspired design concepts for the Lunarminer framework. The orange boxes represent specific tasks, i.e. task allocation and material handling (inspired by leafcutter ants), and recruitment and fault tolerance (inspired by fireflies),which contribute to broader goals stated in the green boxes, such as efficient navigation, swarm automation, and resource handling.</p>
Full article ">Figure 3
<p>RASSOR 2.0 computer-aided design [<a href="#B14-biomimetics-09-00680" class="html-bibr">14</a>].</p>
Full article ">Figure 4
<p>Simulated virtual lunar environment. ROS simulation of Shackleton Crater’s floor (gray areas) with a central hub for collection (black circles), maintenance (yellow squares), base stations (green squares), processing (blue squares), mining (green areas), and transportation (blue areas). The robot fleet includes 4 orange explorers, 2 green excavators, 4 yellow haulers, and 2 blue transporters.</p>
Full article ">Figure 5
<p>Lunarminer finite state machine.</p>
Full article ">Figure 6
<p>(<b>a</b>) Strip search and piecewise tracking function for resource prospecting. The green arrows indicate the strip search path, while the red arrow highlights a prioritized direction or specific target location within the search area. Blue-shaded areas represent zones covered by individual units as they scan for resources; (<b>b</b>) fireflies’ bioluminescent function inspired recruitment protocol, where light beacons are placed at ore locations to signal and attract other units to ore locations.</p>
Full article ">Figure 7
<p>(<b>a</b>) Selection of the mining site based on light proximity and intensity, with red arrows indicating sensed skylight directions guiding site selection; (<b>b</b>) mining excavation process showing ore block detection and hauler positioning system, with light beacon areas representing operational zones.; and (<b>c</b>) division of labor in transporting ore blocks: yellow arrows indicate transport paths from the mine site to the central hub, and green arrows show paths from the central hub to the processing plant.</p>
Full article ">Figure 8
<p>(<b>a</b>) A red-light signal emitted by a malfunctioning robot, inspired by the flashing behavior of fireflies, with blue shaded areas representing the communication range of each robot; (<b>b</b>) activation of the fault-tolerance protocol to replace the malfunctioning robot, indicated by red arrows guiding the replacement robot toward its target within the blue communication zones. The base station and maintenance site are shown in green and yellow, respectively, facilitating the coordination of the replacement process.</p>
Full article ">Figure 9
<p>Highlights of various stages of the Lunarminer mining process from the exploration stage to the recovery stage.</p>
Full article ">Figure 10
<p>(<b>a</b>) Resource extraction time and (<b>b</b>) energy distribution across different scenarios.</p>
Full article ">Figure 11
<p>Fault tolerance and system robustness across three scenarios, i.e., normal, with failure, and with recovery settings.</p>
Full article ">Figure 12
<p>Lunarminer framework classification.</p>
Full article ">
Back to TopTop