[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (6,839)

Search Parameters:
Keywords = fusion feature

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 3820 KiB  
Article
A Deep Learning-Based Watershed Feature Fusion Approach for Tunnel Crack Segmentation in Complex Backgrounds
by Haozheng Wang, Qiang Wang, Weikang Zhang, Junli Zhai, Dongyang Yuan, Junhao Tong, Xiongyao Xie, Biao Zhou and Hao Tian
Materials 2025, 18(1), 142; https://doi.org/10.3390/ma18010142 (registering DOI) - 1 Jan 2025
Abstract
As highway tunnel operations continue over time, structural defects, particularly cracks, have been observed to increase annually. Coupled with the rapid expansion of tunnel networks, traditional manual inspection methods have proven inadequate to meet current demands. In recent years, machine vision and deep [...] Read more.
As highway tunnel operations continue over time, structural defects, particularly cracks, have been observed to increase annually. Coupled with the rapid expansion of tunnel networks, traditional manual inspection methods have proven inadequate to meet current demands. In recent years, machine vision and deep learning technologies have gained significant attention in civil engineering for the detection and analysis of structural defects. However, rapid and accurate defect identification in highway tunnels presents challenges due to complex background conditions, numerous interfering factors, and the relatively low proportion of cracks within the structure. Additionally, the intensive labor requirements and limited efficiency in labeling training datasets for deep learning pose significant constraints on the deployment of intelligent crack segmentation algorithms. To address these limitations, this study proposes an automatic labeling and optimization algorithm for crack sample sets, utilizing crack features and the watershed algorithm to enable efficient automated segmentation with minimal human input. Furthermore, the deep learning-based crack segmentation network was optimized through comparative analysis of various network depths and residual structure configurations to achieve the best possible model performance. Enhanced accuracy was attained by incorporating axis extraction and watershed filling algorithms to refine segmentation outcomes. Under diverse lining surface conditions and multiple interference factors, the proposed approach achieved a crack segmentation accuracy of 98.78%, with an Intersection over Union (IoU) of 72.41%, providing a robust solution for crack segmentation in tunnels with complex backgrounds. Full article
28 pages, 3263 KiB  
Review
High Energy Density Welding of Ni-Based Superalloys: An Overview
by Riccardo Donnini, Alessandra Varone, Alessandra Palombi, Saveria Spiller, Paolo Ferro and Giuliano Angella
Metals 2025, 15(1), 30; https://doi.org/10.3390/met15010030 (registering DOI) - 1 Jan 2025
Abstract
High energy density technologies for welding processes provide opportune solutions to joint metal materials and repair components in several industrial applications. Their high-performance levels are related to the high penetration depth and welding speed achievable. Moreover, the localized thermal input helps in reducing [...] Read more.
High energy density technologies for welding processes provide opportune solutions to joint metal materials and repair components in several industrial applications. Their high-performance levels are related to the high penetration depth and welding speed achievable. Moreover, the localized thermal input helps in reducing distortion and residual stresses in the welds, minimizing the extension of the fusion zone and heat-affected zone. The use of these welding technologies can be decisive in the employment of sophisticated alloys such as Ni-based superalloys, which are notoriously excellent candidates for industrial components subjected to high temperatures and corrosive work conditions. Nonetheless, the peculiar crystallographic and chemical complexity of Ni-based superalloys (whether characterized by polycrystalline, directionally solidified, or single-crystal microstructure) leads to high susceptibility to welding processes and, in general, challenging issues related to the microstructural features of the welded joints. The present review highlights the advantages and drawbacks of high energy density (Laser Beam and Electron Beam) welding techniques applied to Ni-based superalloy. The effects of process parameters on cracking susceptibility have been analyzed to better understand the correlation between them and the microstructure-mechanical properties of the welds. The weldability of three different polycrystalline Ni superalloys, one solid solution-strengthened alloy, Inconel 625, and two precipitation-strengthen alloys, Nimonic 263 and Inconel 718, is reviewed in detail. In addition, a variant of the latter, the AF955 alloy, is also presented for its great potential in terms of weldability. Full article
(This article belongs to the Special Issue Advanced Welding Technology in Metals III)
25 pages, 3292 KiB  
Article
Lane Detection Based on CycleGAN and Feature Fusion in Challenging Scenes
by Eric Hsueh-Chan Lu and Wei-Chih Chiu
Vehicles 2025, 7(1), 2; https://doi.org/10.3390/vehicles7010002 (registering DOI) - 1 Jan 2025
Viewed by 80
Abstract
Lane detection is a pivotal technology of the intelligent driving system. By identifying the position and shape of the lane, the vehicle can stay in the correct lane and avoid accidents. Image-based deep learning is currently the most advanced method for lane detection. [...] Read more.
Lane detection is a pivotal technology of the intelligent driving system. By identifying the position and shape of the lane, the vehicle can stay in the correct lane and avoid accidents. Image-based deep learning is currently the most advanced method for lane detection. Models using this method already have a very good recognition ability in general daytime scenes, and can almost achieve real-time detection. However, these models often fail to accurately identify lanes in challenging scenarios such as night, dazzle, or shadows. Furthermore, the lack of diversity in the training data restricts the capacity of the models to handle different environments. This paper proposes a novel method to train CycleGAN with existing daytime and nighttime datasets. This method can extract features of different styles and multi-scales, thereby increasing the richness of model input. We use CycleGAN as a domain adaptation model combined with an image segmentation model to boost the model’s performance in different styles of scenes. The proposed consistent loss function is employed to mitigate performance disparities of the model in different scenarios. Experimental results indicate that our method enhances the detection performance of original lane detection models in challenging scenarios. This research helps improve the dependability and robustness of intelligent driving systems, ultimately making roads safer and enhancing the driving experience. Full article
Show Figures

Figure 1

Figure 1
<p>Illustration of the two-stage proposed method.</p>
Full article ">Figure 2
<p>The architecture of the proposed lane detection model training.</p>
Full article ">Figure 3
<p>The architecture of the proposed domain adaption model training.</p>
Full article ">Figure 4
<p>Qualitative analysis result of Grad-CAM.</p>
Full article ">Figure 5
<p>The example of challenge scenarios from the CULane test dataset. (<b>a</b>) Crowded; (<b>b</b>) Dazzle; (<b>c</b>) Shadow; (<b>d</b>) Arrow.</p>
Full article ">Figure 6
<p>The example of other challenge scenarios from the CULane test dataset. (<b>a</b>) Indoor; (<b>b</b>) wet ground.</p>
Full article ">Figure 7
<p>Visualization results of challenging scenes in the CULane test dataset. (<b>a</b>) Night; (<b>b</b>) crowded; (<b>c</b>) dazzle; (<b>d</b>) shadow; (<b>e</b>) arrow; (<b>f</b>) indoor; (<b>g</b>) wet ground.</p>
Full article ">Figure 7 Cont.
<p>Visualization results of challenging scenes in the CULane test dataset. (<b>a</b>) Night; (<b>b</b>) crowded; (<b>c</b>) dazzle; (<b>d</b>) shadow; (<b>e</b>) arrow; (<b>f</b>) indoor; (<b>g</b>) wet ground.</p>
Full article ">
31 pages, 9112 KiB  
Article
Intelligent Target Detection in Synthetic Aperture Radar Images Based on Multi-Level Fusion
by Qiaoyu Liu, Ziqi Ye, Chenxiang Zhu, Dongxu Ouyang, Dandan Gu and Haipeng Wang
Remote Sens. 2025, 17(1), 112; https://doi.org/10.3390/rs17010112 (registering DOI) - 1 Jan 2025
Viewed by 148
Abstract
Due to the unique imaging mechanism of SAR, targets in SAR images present complex scattering characteristics. As a result, intelligent target detection in SAR images has been facing many challenges, which mainly lie in the insufficient exploitation of target characteristics, inefficient characterization of [...] Read more.
Due to the unique imaging mechanism of SAR, targets in SAR images present complex scattering characteristics. As a result, intelligent target detection in SAR images has been facing many challenges, which mainly lie in the insufficient exploitation of target characteristics, inefficient characterization of scattering features, and inadequate reliability of decision models. In this respect, we propose an intelligent target detection method based on multi-level fusion, where pixel-level, feature-level, and decision-level fusions are designed for enhancing scattering feature mining and improving the reliability of decision making. The pixel-level fusion method through the channel fusion of original images and their features after scattering feature enhancement represents an initial exploration of image fusion. Two feature-level fusion methods are conducted using respective migratable fusion blocks, namely DBAM and FDRM, presenting higher-level fusion. Decision-level fusion based on DST can not only consolidate complementary strengths in different models but also incorporate human or expert involvement in proposition for guiding effective decision making. This represents the highest-level fusion integrating results by proposition setting and statistical analysis. Experiments of different fusion methods integrating different features were conducted on typical target detection datasets. As shown in the results, the proposed method increases the mAP by 16.52%, 7.1%, and 3.19% in ship, aircraft, and vehicle target detection, demonstrating high effectiveness and robustness. Full article
(This article belongs to the Special Issue SAR-Based Signal Processing and Target Recognition (Second Edition))
Show Figures

Figure 1

Figure 1
<p>Relationship of methods driven by model, data, and target characteristics.</p>
Full article ">Figure 2
<p>Structure of intelligent method based on multi-level image fusion for target detection in SAR images.</p>
Full article ">Figure 3
<p>Processing levels of image fusion.</p>
Full article ">Figure 4
<p>Structure of backbone network.</p>
Full article ">Figure 5
<p>Structure of PA-FPN.</p>
Full article ">Figure 6
<p>Structure of ST-PA_RCNN based on pixel-level fusion.</p>
Full article ">Figure 7
<p>Channel fusion.</p>
Full article ">Figure 8
<p>Structure of DBAM ST-PA_RCNN.</p>
Full article ">Figure 9
<p>Structure of DB-PA-FPN.</p>
Full article ">Figure 10
<p>Structure of attention mechanism fusion block.</p>
Full article ">Figure 11
<p>Structure of FDRM ST-PA_RCNN.</p>
Full article ">Figure 12
<p>Structure of FDRM fusion block.</p>
Full article ">Figure 13
<p>Structure of decision-level fusion based on DST.</p>
Full article ">Figure 14
<p>Examples of different scene slices in SRSDD-v1.0.</p>
Full article ">Figure 15
<p>Examples of different scene slices in GF3-ADD.</p>
Full article ">Figure 16
<p>Details in fine-to-broad categorization.</p>
Full article ">Figure 17
<p>Examples of different scene slices in MSTAR-VDD.</p>
Full article ">
23 pages, 4009 KiB  
Article
Remaining Life Prediction Modeling Method for Rotating Components of Complex Intelligent Equipment
by Yaohua Deng, Zilin Zhang, Hao Huang and Xiali Liu
Electronics 2025, 14(1), 136; https://doi.org/10.3390/electronics14010136 - 31 Dec 2024
Viewed by 243
Abstract
This paper aims to address the challenges of significant data distribution differences and extreme data imbalances in the remaining useful life prediction modeling of rotating components of complex intelligent equipment under various working conditions. Grounded in deep learning modeling, it considers the multi-dimensional [...] Read more.
This paper aims to address the challenges of significant data distribution differences and extreme data imbalances in the remaining useful life prediction modeling of rotating components of complex intelligent equipment under various working conditions. Grounded in deep learning modeling, it considers the multi-dimensional extraction method for degraded data features in the data feature extraction stage, proposes a network structure with multiple attention data extraction channels, and explores the extraction method for valuable data segments in the channel and time series dimensions. This paper also proposes a domain feature fusion network based on feature migration and examines methods that leverage abundant labeled data from the source domain to assist in target domain learning. Finally, in combination with a long short-term memory neural network (LSTM), this paper constructs an intelligent model to estimate the remaining lifespan of rotating components. Experiments demonstrate that, when integrating the foundational deep convolution network with the domain feature fusion network, the comprehensive loss error for life prediction on the target domain test set can be reduced by up to 6.63%. Furthermore, when adding the dual attention feature extraction network, the maximum reduction in the comprehensive loss error is 3.22%. This model can effectively enhance the precision of life prediction in various operating conditions; thus, it provides a certain theoretical basis and technical support for the operation and maintenance management of complex intelligent equipment. It has certain practical value and application prospects in the remaining life prediction of rotating components under multiple working conditions. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Life prediction modeling framework for rotating components of complex intelligent equipment.</p>
Full article ">Figure 2
<p>Channel attention module structure.</p>
Full article ">Figure 3
<p>Temporal attention module structure.</p>
Full article ">Figure 4
<p>Domain feature fusion network framework.</p>
Full article ">Figure 5
<p>Domain feature fusion network framework.</p>
Full article ">Figure 6
<p>Kurtosis plots of bearing dataset 1: (<b>a</b>) kurtosis of bearing data 2-3; (<b>b</b>) kurtosis of bearing data 2-4.</p>
Full article ">Figure 7
<p>Kurtosis plots of bearing dataset 1: (<b>a</b>) CBAM (0–100); (<b>b</b>) DAM (0–100); (<b>c</b>) CBAM (100–200); (<b>d</b>) DAM (100–200); (<b>e</b>) CBAM (200–300); (<b>f</b>) DAM (200–300); (<b>g</b>) CBAM (300–400); (<b>h</b>) DAM (300–400); (<b>i</b>) AM (400–500); (<b>j</b>) DAM (400–500).</p>
Full article ">Figure 7 Cont.
<p>Kurtosis plots of bearing dataset 1: (<b>a</b>) CBAM (0–100); (<b>b</b>) DAM (0–100); (<b>c</b>) CBAM (100–200); (<b>d</b>) DAM (100–200); (<b>e</b>) CBAM (200–300); (<b>f</b>) DAM (200–300); (<b>g</b>) CBAM (300–400); (<b>h</b>) DAM (300–400); (<b>i</b>) AM (400–500); (<b>j</b>) DAM (400–500).</p>
Full article ">Figure 8
<p>Variance changing curves of features: (<b>a</b>) bearing dataset 1; (<b>b</b>) bearing dataset 2.</p>
Full article ">Figure 9
<p>MAE loss plots for different models: (<b>a</b>) CNN-LSTM; (<b>b</b>) DANN-LSTM; (<b>c</b>) DFFN-LSTM; (<b>d</b>) DAM-DFFN-LSTM.</p>
Full article ">Figure 10
<p>Life prediction trend curves: (<b>a</b>) bearing dataset 1; (<b>b</b>) bearing dataset 2.</p>
Full article ">
20 pages, 4757 KiB  
Article
Infrared Image Detection and Recognition of Substation Electrical Equipment Based on Improved YOLOv8
by Haotian Tao, Agyemang Paul and Zhefu Wu
Appl. Sci. 2025, 15(1), 328; https://doi.org/10.3390/app15010328 - 31 Dec 2024
Viewed by 248
Abstract
To address the challenges associated with lightweight design and small object detection in infrared imaging for substation electrical equipment, this paper introduces an enhanced YOLOv8_Adv network model. This model builds on YOLOv8 through several strategic improvements. The backbone network incorporates PConv and FasterNet [...] Read more.
To address the challenges associated with lightweight design and small object detection in infrared imaging for substation electrical equipment, this paper introduces an enhanced YOLOv8_Adv network model. This model builds on YOLOv8 through several strategic improvements. The backbone network incorporates PConv and FasterNet modules to substantially reduce the computational load and memory usage, thereby achieving model lightweighting. In the neck layer, GSConv and VoVGSCSP modules are utilized for multi-stage, multi-feature map fusion, complemented by the integration of the EMA attention mechanism to improve feature extraction. Additionally, a specialized detection layer for small objects is added to the head of the network, enhancing the model’s performance in detecting small infrared targets. Experimental results demonstrate that YOLOv8_Adv achieves a 4.1% increase in [email protected] compared to the baseline YOLOv8n. It also outperforms five existing baseline models, with the highest accuracy of 98.7%, and it reduces the computational complexity by 18.5%, thereby validating the effectiveness of the YOLOv8_Adv model. Furthermore, the effectiveness of the model in detecting small targets in infrared images makes it suitable for use in areas such as infrared surveillance, military target detection, and wildlife monitoring. Full article
(This article belongs to the Special Issue Signal and Image Processing: From Theory to Applications)
Show Figures

Figure 1

Figure 1
<p>YOLOv8 network architecture diagram.</p>
Full article ">Figure 2
<p>Structural diagram of the C2f-fast.</p>
Full article ">Figure 3
<p>Structural diagram of the FasterNet block.</p>
Full article ">Figure 4
<p>Structural diagram of the GSConv.</p>
Full article ">Figure 5
<p>Diagram illustrating the structure of VoVGSCSP.</p>
Full article ">Figure 6
<p>Structural diagram of the EMA.</p>
Full article ">Figure 7
<p>Diagram of the feature fusion structure with the addition of a small target detection layer.</p>
Full article ">Figure 8
<p>YOLOv8_Adv network structure diagram.</p>
Full article ">Figure 9
<p>Training dataset.</p>
Full article ">Figure 10
<p>Confusion matrix.</p>
Full article ">Figure 11
<p>Comparative analysis of precision–recall (P–R) curves for different models.</p>
Full article ">Figure 12
<p>mAP curve comparison of various models.</p>
Full article ">Figure 13
<p>Visualized detection results (<b>a</b>) YOLOv8n; (<b>b</b>) YOLOv8_Adv.</p>
Full article ">
26 pages, 34163 KiB  
Article
Navigating ALICE: Advancements in Deployable Docking and Precision Detection for AUV Operations
by Yevgeni Gutnik, Nir Zagdanski, Sharon Farber, Tali Treibitz and Morel Groper
Robotics 2025, 14(1), 5; https://doi.org/10.3390/robotics14010005 - 31 Dec 2024
Viewed by 226
Abstract
Autonomous Underwater Vehicles (AUVs) operate independently using onboard batteries and data storage, necessitating periodic recovery for battery recharging and data transfer. Traditional surface-based launch and recovery (L&R) operations pose significant risks to personnel and equipment, particularly in adverse weather conditions. Subsurface docking stations [...] Read more.
Autonomous Underwater Vehicles (AUVs) operate independently using onboard batteries and data storage, necessitating periodic recovery for battery recharging and data transfer. Traditional surface-based launch and recovery (L&R) operations pose significant risks to personnel and equipment, particularly in adverse weather conditions. Subsurface docking stations provide a safer alternative but often involve complex fixed installations and costly acoustic positioning systems.This work introduces a comprehensive docking solution featuring the following two key innovations: (1) a novel deployable docking station (DDS) designed for rapid deployment from vessels of opportunity, operating without active acoustic transmitters; and (2) an innovative sensor fusion approach that combines the AUV’s onboard forward-looking sonar and camera data. The DDS comprises a semi-submersible protective frame and a subsurface, heave-compensated docking component equipped with backlit visual markers, an electromagnetic (EM) beacon, and an EM lifting device. This adaptable design is suitable for temporary installations and in acoustically sensitive or covert operations.The positioning and guidance system employs a multi-sensor approach, integrating range and azimuth data from the sonar with elevation data from the vision camera to achieve precise 3D positioning and robust navigation in varying underwater conditions. This paper details the design considerations and integration of the AUV system and the docking station, highlighting their innovative features.The proposed method was validated through software-in-the-loop simulations, controlled seawater pool experiments, and preliminary open-sea trials, including several docking attempts. While further sea trials are planned, current results demonstrate the potential of this solution to enhance AUV operational capabilities in challenging underwater environments while reducing deployment complexity and operational costs. Full article
(This article belongs to the Special Issue Navigation Systems of Autonomous Underwater and Surface Vehicles)
23 pages, 1411 KiB  
Article
Applications of the FusionScratchNet Algorithm Based on Convolutional Neural Networks and Transformer Models in the Detection of Cell Phone Screen Scratches
by Zhihong Cao, Kun Liang, Sheng Tang and Cheng Zhang
Electronics 2025, 14(1), 134; https://doi.org/10.3390/electronics14010134 - 31 Dec 2024
Viewed by 181
Abstract
Screen defect detection has become a crucial research domain, propelled by the growing necessity of precise and effective quality control in mobile device production. This study presents the FusionScratchNet (FS-Net), a novel algorithm developed to overcome the challenges of noise interference and to [...] Read more.
Screen defect detection has become a crucial research domain, propelled by the growing necessity of precise and effective quality control in mobile device production. This study presents the FusionScratchNet (FS-Net), a novel algorithm developed to overcome the challenges of noise interference and to characterize indistinct defects and subtle scratches on mobile phone screens. By integrating the transformer and convolutional neural network (CNN) architectures, FS-Net effectively captures both global and local features, thereby enhancing feature representation. The global–local feature integrator (GLFI) module effectively fuses global and local information through unique channel splitting, feature dependency characterization, and attention mechanisms, thereby enhancing target features and suppressing noise. The bridge attention (BA) module calculates an attention feature map based on the multi-layer fused features, precisely focusing on scratch characteristics and recovering details lost during downsampling. Evaluations using the PKU-Market-Phone dataset demonstrated an overall accuracy of 98.04%, an extended intersection over union (EIoU) of 88.03%, and an F1-score of 65.13%. In comparison to established methods like you only look once (YOLO) and retina network (RetinaNet), FS-Net demonstrated enhanced detection accuracy, computational efficiency, and resilience against noise. The experimental results demonstrated that the proposed method effectively enhances the accuracy of scratch segmentation. Full article
28 pages, 4471 KiB  
Article
Remaining Life Prediction of Automatic Fare Collection Systems from the Perspective of Sustainable Development: A Sparse and Weak Feature Fault Data-Based Approach
by Jing Xiong, Youchao Sun, Zhihao Xu, Yongbing Wan and Gang Yu
Sustainability 2025, 17(1), 230; https://doi.org/10.3390/su17010230 - 31 Dec 2024
Viewed by 252
Abstract
The most effective way to solve urban traffic congestion in mega cities is to develop rail transit, which is also an important strategy for sustainable urban development. Improving the service performance of rail transit equipment is the key to ensuring the sustainable operation [...] Read more.
The most effective way to solve urban traffic congestion in mega cities is to develop rail transit, which is also an important strategy for sustainable urban development. Improving the service performance of rail transit equipment is the key to ensuring the sustainable operation of urban rail transit. Automatic fare collection (AFC) is an indispensable system in urban rail transit. AFC directly serves passengers, and its condition directly affects the sustainability and safety of urban rail transit. This study proposes remaining useful life (RUL) prediction framework for AFC systems. Firstly, it proposes the quantification of AFC health state based on health degree, and proposes a health state assessment method based on digital analog fusion, which compensates for the shortcomings of single data-driven or model driven health methods. Secondly, it constructs a multi feature extraction method based on multi-layer LSTM, which can capture long-term temporal dependencies and multi-dimensional feature, overcoming the limitation of low model accuracy because of the weak data features. Then, the SSA-XGBoost model for AFC RUL prediction is proposed, which effectively performs global and local searches, reduces the possibility of overfitting, and improves the accuracy of the prediction model. Finally, we put it into practice of the AFC system of Shanghai Metro Line 10. The experiment shows that the proposed model has an MSE of 0.00111 and MAE of 0.02869 on the test set, while on the validation set, MSE is 0.00004 and MAE is 0.00659. These indicators are significantly better than other comparative models such as XGBoost, random forest regression, and linear regression. In addition, the SSA-XGBoost model also performs well on R-squared, further verifying its effectiveness in prediction accuracy and model fitting. Full article
Show Figures

Figure 1

Figure 1
<p>A framework for prediction of the remaining life of AFC.</p>
Full article ">Figure 2
<p>Multi-layer LSTM based feature extraction.</p>
Full article ">Figure 3
<p>Structure of XGboost model.</p>
Full article ">Figure 4
<p>SSA-XGBoost based prediction process.</p>
Full article ">Figure 5
<p>AFC system health state assessment and RUL prediction flowchart.</p>
Full article ">Figure 6
<p>Loss curve of GAN network.</p>
Full article ">Figure 7
<p>Health degree of AFC system based on weight calculation.</p>
Full article ">Figure 8
<p>Health degree of AFC system based on digital analog fusion.</p>
Full article ">Figure 9
<p>Comparison of predicted health values with actual health values (validation set).</p>
Full article ">Figure 10
<p>Comparison of prediction health grade and actual health grade (validation set).</p>
Full article ">Figure 11
<p>Mean square error for different models.</p>
Full article ">Figure 12
<p>Mean absolute errors for different models.</p>
Full article ">Figure 13
<p>R-squared for different models.</p>
Full article ">Figure 14
<p>Feature importance based on PFI.</p>
Full article ">
19 pages, 11243 KiB  
Article
A Simple Polarization-Based Fringe Projection Profilometry Method for Three-Dimensional Reconstruction of High-Dynamic-Range Surfaces
by Xiang Sun, Zhenjun Luo, Shizhao Wang, Jianhua Wang, Yunpeng Zhang and Dandan Zou
Photonics 2025, 12(1), 27; https://doi.org/10.3390/photonics12010027 - 31 Dec 2024
Viewed by 195
Abstract
Three-dimensional (3D) reconstruction of high-dynamic-range (HDR) surfaces plays an important role in the fields of computer vision and image processing. Traditional 3D measurement methods often face the risk of information loss when dealing with surfaces that have HDR characteristics. To address this issue, [...] Read more.
Three-dimensional (3D) reconstruction of high-dynamic-range (HDR) surfaces plays an important role in the fields of computer vision and image processing. Traditional 3D measurement methods often face the risk of information loss when dealing with surfaces that have HDR characteristics. To address this issue, this paper proposes a simple 3D reconstruction method, which combines the features of non-overexposed regions in polarized and unpolarized images to improve the reconstruction quality of HDR surface objects. The optimum fringe regions are extracted from images with different polarization angles, and the non-overexposed regions in normally captured unpolarized images typically contain complete fringe information and are less affected by specular highlights. The optimal fringe information from different polarized image groups is gradually used to replace the incorrect fringe information in the unpolarized image, resulting in a complete set of fringe data. Experimental results show that the proposed method requires only 24~36 images and simple phase fusion to achieve successful 3D reconstruction. It can effectively mitigate the negative impact of overexposed regions on absolute phase calculation and 3D reconstruction when reconstructing objects with strongly reflective surfaces. Full article
(This article belongs to the Special Issue New Perspectives in Optical Design)
Show Figures

Figure 1

Figure 1
<p>The overexposed region <span class="html-italic">R<sub>G</sub></span> and non-overexposed region <span class="html-italic">R<sub>F</sub></span> in the unpolarized image (from a set of images).</p>
Full article ">Figure 2
<p>(<b>a</b>) is the overexposed region <span class="html-italic">R<sub>G</sub></span> and non-overexposed region <span class="html-italic">R<sub>F</sub></span> in the brighter polarization set images, and (<b>b</b>) is the non-overexposed region <span class="html-italic">R<sub>F</sub></span> in the darker polarization set images.</p>
Full article ">Figure 3
<p>Experimental design diagram.</p>
Full article ">Figure 4
<p>The flowchart of the method presented in this paper.</p>
Full article ">Figure 5
<p>The physical diagram of the system built in this paper.</p>
Full article ">Figure 6
<p>The sum of the non-overexposed regions in the polarized image set can cover the overexposed region in the unpolarized image.</p>
Full article ">Figure 7
<p>The unpolarized image of the metal plate is shown in (<b>a</b>), the calculated absolute phase is presented in (<b>b</b>), and the generated 3D point cloud is displayed in (<b>c</b>).</p>
Full article ">Figure 8
<p>The fusion process of fringe information.</p>
Full article ">Figure 9
<p>After fusing the phase information, the calculated absolute phase is shown in (<b>a</b>), and the generated 3D point cloud is displayed in (<b>b</b>).</p>
Full article ">Figure 10
<p>Fringe image diagram.</p>
Full article ">Figure 11
<p>The point cloud maps obtained from the additional experiments.</p>
Full article ">Figure 12
<p>(<b>a</b>) is the point cloud map obtained from the experimental group images processed using traditional methods, (<b>b</b>) is the point cloud map obtained using the proposed method, (<b>c</b>) shows the region without fringe information, which corresponds to the fused region of the proposed method, and (<b>d</b>) provides a comparison of the same region from both point clouds.</p>
Full article ">Figure 13
<p>Physical drawings of water bottles and standard part.</p>
Full article ">Figure 14
<p>The unpolarized image of the kettle is shown in (<b>a</b>), the calculated absolute phase is presented in (<b>b</b>), and the generated 3D point cloud is displayed in (<b>c</b>); after fusing the phase information, the resulting absolute phase and 3D point cloud images are shown in (<b>d</b>) and (<b>e</b>), respectively.</p>
Full article ">Figure 15
<p>The unpolarized image of the standard part is shown in (<b>a</b>), the calculated absolute phase is presented in (<b>b</b>), and the generated 3D point cloud is displayed in (<b>c</b>); after fusing the phase information, the resulting absolute phase and 3D point cloud images are shown in (<b>d</b>) and (<b>e</b>), respectively.</p>
Full article ">Figure 16
<p>(<b>a</b>) shows the fringe pattern, point cloud, and local magnified image obtained using the method from Reference [<a href="#B14-photonics-12-00027" class="html-bibr">14</a>], (<b>b</b>) shows the fringe pattern, point cloud, and local magnified image obtained using the method from Reference [<a href="#B15-photonics-12-00027" class="html-bibr">15</a>], and (<b>c</b>) shows the fringe pattern, point cloud, and local magnified image obtained using the proposed method.</p>
Full article ">
16 pages, 2102 KiB  
Article
Semantic Segmentation Method for High-Resolution Tomato Seedling Point Clouds Based on Sparse Convolution
by Shizhao Li, Zhichao Yan, Boxiang Ma, Shaoru Guo and Hongxia Song
Agriculture 2025, 15(1), 74; https://doi.org/10.3390/agriculture15010074 (registering DOI) - 31 Dec 2024
Viewed by 166
Abstract
Semantic segmentation of three-dimensional (3D) plant point clouds at the stem-leaf level is foundational and indispensable for high-throughput tomato phenotyping systems. However, existing semantic segmentation methods often suffer from issues such as low precision and slow inference speed. To address these challenges, we [...] Read more.
Semantic segmentation of three-dimensional (3D) plant point clouds at the stem-leaf level is foundational and indispensable for high-throughput tomato phenotyping systems. However, existing semantic segmentation methods often suffer from issues such as low precision and slow inference speed. To address these challenges, we propose an innovative encoding-decoding structure, incorporating voxel sparse convolution (SpConv) and attention-based feature fusion (VSCAFF) to enhance semantic segmentation of the point clouds of high-resolution tomato seedling images. Tomato seedling point clouds from the Pheno4D dataset labeled into semantic classes of ‘leaf’, ‘stem’, and ‘soil’ are applied for the semantic segmentation. In order to reduce the number of parameters so as to further improve the inference speed, the SpConv module is designed to function through the residual concatenation of the skeleton convolution kernel and the regular convolution kernel. The feature fusion module based on the attention mechanism is designed by giving the corresponding attention weights to the voxel diffusion features and the point features in order to avoid the ambiguity of points with different semantics having the same characteristics caused by the diffusion module, in addition to suppressing noise. Finally, to solve model training class bias caused by the uneven distribution of point cloud classes, the composite loss function of Lovász-Softmax and weighted cross-entropy is introduced to supervise the model training and improve its performance. The results show that mIoU of VSCAFF is 86.96%, which outperformed the performance of PointNet, PointNet++, and DGCNN, respectively. IoU of VSCAFF achieves 99.63% in the soil class, 64.47% in the stem class, and 96.72% in the leaf class. The time delay of 35ms in inference speed is better than PointNet++ and DGCNN. The results demonstrate that VSCAFF has high performance and inference speed for semantic segmentation of high-resolution tomato point clouds, and can provide technical support for the high-throughput automatic phenotypic analysis of tomato plants. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>The raw tomato seedling point cloud and point cloud labeled into semantic classes of ‘leaf’, and ‘stem’, and ‘soil’.</p>
Full article ">Figure 2
<p>The structure of network.</p>
Full article ">Figure 3
<p>Encoding-decodin g architecture based on SpConv.</p>
Full article ">Figure 4
<p>Three kinds of convolution kernel structures.</p>
Full article ">Figure 5
<p>Attention-based feature fusion method.</p>
Full article ">Figure 6
<p>Semantic segmentation of the tomato plant point clouds. Note1: GT represents ground truth. Note2: Four seedling point clouds scanned in four discrete days represented.</p>
Full article ">
20 pages, 2870 KiB  
Article
Research on Mine-Personnel Helmet Detection Based on Multi-Strategy-Improved YOLOv11
by Lei Zhang, Zhipeng Sun, Hongjing Tao, Meng Wang and Weixun Yi
Sensors 2025, 25(1), 170; https://doi.org/10.3390/s25010170 - 31 Dec 2024
Viewed by 215
Abstract
In the complex environment of fully mechanized mining faces, the current object detection algorithms face significant challenges in achieving optimal accuracy and real-time detection of mine personnel and safety helmets. This difficulty arises from factors such as uneven lighting conditions and equipment obstructions, [...] Read more.
In the complex environment of fully mechanized mining faces, the current object detection algorithms face significant challenges in achieving optimal accuracy and real-time detection of mine personnel and safety helmets. This difficulty arises from factors such as uneven lighting conditions and equipment obstructions, which often lead to missed detections. Consequently, these limitations pose a considerable challenge to effective mine safety management. This article presents an enhanced algorithm based on YOLOv11n, referred to as GCB-YOLOv11. The proposed improvements are realized through three key aspects: Firstly, the traditional convolution is replaced with GSConv, which significantly enhances feature extraction capabilities while simultaneously reducing computational costs. Secondly, a novel C3K2_FE module was designed that integrates Faster_block and ECA attention mechanisms. This design aims to improve detection accuracy while also accelerating detection speed. Finally, the introduction of the Bi FPN mechanism in the Neck section optimizes the efficiency of multi-scale feature fusion and addresses issues related to feature loss and redundancy. The experimental results demonstrate that GCB-YOLOv11 exhibits strong performance on the dataset concerning mine personnel and safety helmets, achieving a mean average precision of 93.6%. Additionally, the frames per second reached 90.3 f·s−1, representing increases of 3.3% and 9.4%, respectively, compared to the baseline model. In addition, when compared to models such as YOLOv5s, YOLOv8s, YOLOv3 Tiny, Fast R-CNN, and RT-DETR, GCB-YOLOv11 demonstrates superior performance in both detection accuracy and model complexity. This highlights its advantages in mining environments and offers a viable technical solution for enhancing the safety of mine personnel. Full article
(This article belongs to the Special Issue Recent Advances in Optical Sensor for Mining)
Show Figures

Figure 1

Figure 1
<p>YOLOv11 model structure.</p>
Full article ">Figure 2
<p>Two types of residual structures: (<b>a</b>) C3 structure and (<b>b</b>) C3K2 structure.</p>
Full article ">Figure 3
<p>C2PSA module structure.</p>
Full article ">Figure 4
<p>Comparison of YOLOv11 and YOLOv8 model detection head structures.</p>
Full article ">Figure 5
<p>GCB-YOLOv11 network structure.</p>
Full article ">Figure 6
<p>GSConv structure.</p>
Full article ">Figure 7
<p>Faster_block structure.</p>
Full article ">Figure 8
<p>ECA attention mechanism structure.</p>
Full article ">Figure 9
<p>C3K2_faster structure.</p>
Full article ">Figure 10
<p>Feature pyramid structure: (<b>a</b>) FPN + PAN structure and (<b>b</b>) Bi FPN structure.</p>
Full article ">Figure 11
<p>Sample example of dataset.</p>
Full article ">Figure 12
<p>Comparison of training process curves between two types of models: (<b>a</b>) mAP@0.5 curve and (<b>b</b>) loss curve.</p>
Full article ">Figure 13
<p>Comparison of P-R curves of two types of models on the validation set: (<b>a</b>) P-R curve of YOLOv11n and (<b>b</b>) P-R curve of GCB-YOLOv11.</p>
Full article ">Figure 14
<p>Different model detection results.</p>
Full article ">Figure 15
<p>GCB-YOLOv11 heatmap.</p>
Full article ">
19 pages, 21061 KiB  
Article
Enhancing Rooftop Photovoltaic Segmentation Using Spatial Feature Reconstruction and Multi-Scale Feature Aggregation
by Yu Xiao, Long Lin, Jun Ma and Maoqiang Bi
Energies 2025, 18(1), 119; https://doi.org/10.3390/en18010119 - 31 Dec 2024
Viewed by 244
Abstract
Amidst the dual challenges of energy shortages and global warming, photovoltaic (PV) power generation has emerged as a critical technology due to its efficient utilization of solar energy. Rooftops, as underutilized spaces, are ideal locations for installing solar panels, avoiding the need for [...] Read more.
Amidst the dual challenges of energy shortages and global warming, photovoltaic (PV) power generation has emerged as a critical technology due to its efficient utilization of solar energy. Rooftops, as underutilized spaces, are ideal locations for installing solar panels, avoiding the need for additional land. However, the accurate and generalized segmentation of large-scale PV panel images remains a technical challenge, primarily due to varying image resolutions, large image scales, and the significant imbalance between foreground and background categories. To address these challenges, this paper proposes a novel model based on the Res2Net architecture, an enhanced version of the classic ResNet optimized for multi-scale feature extraction. The model integrates Spatial Feature Reconstruction and multi-scale feature aggregation modules, enabling effective extraction of multi-scale data features and precise reconstruction of spatial features. These improvements are particularly designed to handle the small proportion of PV panels in images, effectively distinguishing target features from redundant ones and improving recognition accuracy. Comparative experiments conducted on a publicly available rooftop PV dataset demonstrate that the proposed method achieves superior performance compared to mainstream techniques, showcasing its effectiveness in precise PV panel segmentation. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Figure 1

Figure 1
<p>An illustrative diagram depicting rooftop photovoltaic panels with different spatial resolutions. The top row shows the original images, while the bottom row displays the corresponding segmentation labels.</p>
Full article ">Figure 2
<p>The proposed network consists of three key components: the encoding phase, skip connections, and the decoding phase. The skip connection segment incorporates the Spatial Feature Reconstruction (SFR) module to strengthen feature extraction. To enlarge the receptive field, Multi-scale Feature Aggregation (MFA) is applied in the lower layers using parallel dilated convolutions. During the decoding phase, high-level semantic features from the MFA module are merged with low-level features to improve feature fusion.</p>
Full article ">Figure 3
<p>The structural diagram of the Res2Net module.</p>
Full article ">Figure 4
<p>The structural diagram of the proposed Spatial Feature Reconstruction module.</p>
Full article ">Figure 5
<p>The training loss values of different methods are presented.</p>
Full article ">Figure 6
<p>The mIoU, mAcc, and mFscore performance metrics across different methods.</p>
Full article ">Figure 7
<p>Visualization results obtained by various methods selected from the Rooftop PV dataset. From left to right column: (<b>a</b>) raw image; (<b>b</b>) ground truth; (<b>c</b>) Ours (<b>d</b>); U-Net [<a href="#B31-energies-18-00119" class="html-bibr">31</a>]; (<b>e</b>) DeepLabv3+ [<a href="#B32-energies-18-00119" class="html-bibr">32</a>]; (<b>f</b>) HRnet [<a href="#B33-energies-18-00119" class="html-bibr">33</a>]; (<b>g</b>) Mask2Former [<a href="#B38-energies-18-00119" class="html-bibr">38</a>]; (<b>h</b>) PSPNet [<a href="#B34-energies-18-00119" class="html-bibr">34</a>]; (<b>i</b>) SegFormer [<a href="#B35-energies-18-00119" class="html-bibr">35</a>]. (<b>j</b>) Beit [<a href="#B36-energies-18-00119" class="html-bibr">36</a>]; (<b>k</b>) MaskFormer [<a href="#B37-energies-18-00119" class="html-bibr">37</a>].</p>
Full article ">Figure 8
<p>The visualized features map of the proposed different modules are shown, where (<b>a</b>,<b>b</b>) represent the original image and the label map, respectively. (<b>c</b>) represents the visualized feature map of the Baseline. (<b>d</b>–<b>f</b>) represent the visualized feature maps after applying Res2Net, MFA, and SFR, respectively. (<b>g</b>) represents the visualized feature map after combining MFA and SFR, and (<b>h</b>) represents the visualized feature map with all modules combined.</p>
Full article ">Figure 9
<p>Failure cases of our method in high-resolution scenarios.</p>
Full article ">
24 pages, 4784 KiB  
Article
Facies-Constrained Kriging Interpolation Method for Parameter Modeling
by Zhenbo Nie, Bo Feng, Huazhong Wang, Chengliang Wu, Rongwei Xu and Chao Ning
Remote Sens. 2025, 17(1), 102; https://doi.org/10.3390/rs17010102 - 30 Dec 2024
Viewed by 218
Abstract
In seismic exploration, establishing a reliable parameter model (such as velocity, density, impedance) is crucial for seismic migration imaging and reservoir characterization. The interpolation of well data to obtain a complete spatial model is an important aspect of parameter modeling. However, in practical [...] Read more.
In seismic exploration, establishing a reliable parameter model (such as velocity, density, impedance) is crucial for seismic migration imaging and reservoir characterization. The interpolation of well data to obtain a complete spatial model is an important aspect of parameter modeling. However, in practical applications, well data are often sparse and irregularly distributed, which complicates the accurate construction of subsurface parameter models. The Kriging method is an effective interpolation method based on discrete well data, but its theoretical assumptions do not meet the practical requirements in seismic exploration, resulting in low modeling accuracy. This article introduces seismic facies information into the Kriging method and proposes a novel parameter modeling method named the facies-constrained Kriging (FC-Kriging) method. The FC-Kriging method modifies the Euclidean distance metric used in Kriging so that the distance between two points depends not only on their spatial coordinates but also on their associated facies categories. The proposed method is a multi-information fusion method that integrates facies information based on well data, enabling good interpolation results even with a limited number of wells. The parameter modeling results based on the FC-Kriging method are more consistent with geological logic, exhibiting clearer boundary features and higher resolution. Furthermore, the FC-Kriging method does not introduce additional computational complexity, making it convenient to implement in a 3D situation. The FC-Kriging method is applied to the 2D Sigsbee model, the 3D Standford V reservoir model and F3 block field data. The results demonstrate its accuracy and effectiveness. Full article
(This article belongs to the Special Issue Multi-data Applied to Near-Surface Geophysics (Second Edition))
18 pages, 30213 KiB  
Article
Prediction of Full-Frequency Deep-Sea Noise Based on Sea Surface Wind Speed and Real-Time Noise Data
by Bo Yuan, Licheng Lu, Zhenzhu Wang, Guoli Song, Li Ma and Wenbo Wang
Remote Sens. 2025, 17(1), 101; https://doi.org/10.3390/rs17010101 - 30 Dec 2024
Viewed by 202
Abstract
The prediction of ocean ambient noise is crucial for protecting the marine ecosystem and ensuring communication and navigation safety, especially under extreme weather conditions such as typhoons and strong winds. Ocean ambient noise is primarily caused by ship activities, wind waves, and other [...] Read more.
The prediction of ocean ambient noise is crucial for protecting the marine ecosystem and ensuring communication and navigation safety, especially under extreme weather conditions such as typhoons and strong winds. Ocean ambient noise is primarily caused by ship activities, wind waves, and other factors, and its complexity makes it a significant challenge to effectively utilize limited data to observe future changes in noise energy. To address this issue, we have designed a multi-modal linear model based on a “decomposition-prediction-modal trend fusion-total fusion” framework. This model simultaneously decomposes wind speed data and ocean ambient noise data into trend and residual components, enabling the wind speed information to effectively extract key trend features of ocean ambient noise. Compared to polynomial fitting methods, single-modal models, and LSTM multi-modal models, the average error of the relative sound pressure level was reduced by 1.3 dB, 0.5 dB, and 0.3 dB, respectively. Our approach demonstrates significant improvements in predicting future trends and detailed fittings of the data. Full article
Show Figures

Figure 1

Figure 1
<p>The schematic diagram of environmental noise measurement.</p>
Full article ">Figure 2
<p>Numerical calculation of Transmission loss, source depth 10 m, receiver depth 4000 m. (<b>a</b>) Point source Transmission loss; (<b>b</b>) distributed source Transmission loss.</p>
Full article ">Figure 3
<p>Single-modal Prediction Workflow.</p>
Full article ">Figure 4
<p>Multi-modal prediction workflow.</p>
Full article ">Figure 5
<p>Decomposition idea of LTSF-Linear.</p>
Full article ">Figure 6
<p>Pseudo-color map of measured sound speed profile and surface source (depth 10 m, frequency 100 Hz) Transmission loss in deep sea.</p>
Full article ">Figure 7
<p>Environmental noise monitoring stations and the path of Typhoon “Kompasu”.</p>
Full article ">Figure 8
<p>Wind speed data for the experimental area from 1 October to 20 October 2021.</p>
Full article ">Figure 9
<p>Pseudo-color plot of relative sound pressure level of different frequency components of ocean ambient noise varying with time.</p>
Full article ">Figure 10
<p>Changes in relative sound pressure level of ocean ambient noise October 1 to 20.</p>
Full article ">Figure 11
<p>Correlation coefficient curve between ocean ambient noise level and wind speed.</p>
Full article ">Figure 12
<p>The normalized comparison of environmental noise and wind speed at frequencies of 80, 203, 1024, and 4096 Hz.</p>
Full article ">Figure 13
<p>Polynomial interpolation noise prediction results.</p>
Full article ">Figure 13 Cont.
<p>Polynomial interpolation noise prediction results.</p>
Full article ">Figure 14
<p>LSTM-sm model prediction performance.</p>
Full article ">Figure 15
<p>LSTM-mm model prediction performance.</p>
Full article ">Figure 15 Cont.
<p>LSTM-mm model prediction performance.</p>
Full article ">Figure 16
<p>LTSF-Linear-sm model prediction effect.</p>
Full article ">Figure 17
<p>LTSF-Linear-mm model prediction effect.</p>
Full article ">Figure 18
<p>Correlation results between actual values and predicted values.</p>
Full article ">Figure 19
<p>MSE results for actual values and predicted values of the top 25% of data.</p>
Full article ">Figure 20
<p>Comprehensive comparison graph for MSE, MAE.</p>
Full article ">Figure 21
<p>LSTM-sm model prediction performance (second test).</p>
Full article ">Figure 22
<p>LSTM-mm model prediction performance (second test).</p>
Full article ">Figure 23
<p>LTSF-Linear-sm model prediction effect (second test).</p>
Full article ">Figure 24
<p>LTSF-Linear-mm model prediction effect (second test).</p>
Full article ">
Back to TopTop