[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,681)

Search Parameters:
Keywords = visibility network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 9095 KiB  
Article
BMFusion: Bridging the Gap Between Dark and Bright in Infrared-Visible Imaging Fusion
by Chengwen Liu, Bin Liao and Zhuoyue Chang
Electronics 2024, 13(24), 5005; https://doi.org/10.3390/electronics13245005 (registering DOI) - 19 Dec 2024
Abstract
The fusion of infrared and visible light images is a crucial technology for enhancing visual perception in complex environments. It plays a pivotal role in improving visual perception and subsequent performance in advanced visual tasks. However, due to the significant degradation of visible [...] Read more.
The fusion of infrared and visible light images is a crucial technology for enhancing visual perception in complex environments. It plays a pivotal role in improving visual perception and subsequent performance in advanced visual tasks. However, due to the significant degradation of visible light image quality in low-light or nighttime scenes, most existing fusion methods often struggle to obtain sufficient texture details and salient features when processing such scenes. This can lead to a decrease in fusion quality. To address this issue, this article proposes a new image fusion method called BMFusion. Its aim is to significantly improve the quality of fused images in low-light or nighttime scenes and generate high-quality fused images around the clock. This article first designs a brightness attention module composed of brightness attention units. It extracts multimodal features by combining the SimAm attention mechanism with a Transformer architecture. Effective enhancement of brightness and features has been achieved, with gradual brightness attention performed during feature extraction. Secondly, a complementary fusion module was designed. This module deeply fuses infrared and visible light features to ensure the complementarity and enhancement of each modal feature during the fusion process, minimizing information loss to the greatest extent possible. In addition, a feature reconstruction network combining CLIP-guided semantic vectors and neighborhood attention enhancement was proposed in the feature reconstruction stage. It uses the KAN module to perform channel adaptive optimization on the reconstruction process, ensuring semantic consistency and detail integrity of the fused image during the reconstruction phase. The experimental results on a large number of public datasets demonstrate that the BMFusion method can generate fusion images with higher visual quality and richer details in night and low-light environments compared with various existing state-of-the-art (SOTA) algorithms. At the same time, the fusion image can significantly improve the performance of advanced visual tasks. This shows the great potential and application prospect of this method in the field of multimodal image fusion. Full article
(This article belongs to the Topic Computer Vision and Image Processing, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>An example of an illumination imbalance. From left to right: IR image, visible image, fusion results of various algorithms and our proposed BMFusion. Existing methods ignore the problem of nighttime illumination degradation, leading to detail loss and thermal target degradation. Our algorithm can enhance the brightness while integrating meaningful information, mining a large amount of information lost in the dark.</p>
Full article ">Figure 2
<p>Overview of the BMFusion network architecture, showcasing modules for brightness adjustment, mutual feature enhancement, and progressive semantic-guided reconstruction for efficient multimodal image fusion.</p>
Full article ">Figure 3
<p>Illustration of detailed architectures, including Transformer, simAM mechanism, Brightness Attention Unit (BAU), and Mutually Reinforcing Fusion (MRF) modules, showcasing their roles in feature extraction, attention, and multimodal fusion.</p>
Full article ">Figure 4
<p>Qualitative comparison of BMFusion with 9 state-of-the-art methods in different scenes on LLVIP datasets.</p>
Full article ">Figure 5
<p>Quantitative results of six metrics, i.e., SD, MI, VlF, AG, EN, and SF, on any 20 image pairs from the LLVIP dataset. Nine SOTA methods are used for comparison.</p>
Full article ">Figure 6
<p>Qualitative comparison of BMFusion with 9 state-of-the-art methods in different scenes on MSRS datasets.</p>
Full article ">Figure 7
<p>Quantitative results of six metrics, i.e., SD, MI, VlF, AG, EN, and SE, on any 20 image pairs from the MSRS dataset. Nine SOTA methods are used for comparison.</p>
Full article ">Figure 8
<p>Visualized results of images and feature maps. The first column presents the infrared image, visible image, and fused image, respectively. The next five columns show the feature maps corresponding to the infrared, visible, and fused images in various channel dimensions.</p>
Full article ">Figure 9
<p>Visualized results of ablation on three typical infrared and visible image pairs. From top to bottom: infrared images, visible images, fused results of BMFusion, BMFusion without BAU, and BMFusion without MEF.</p>
Full article ">Figure 10
<p>Visual ablation results of five typical infrared and visible light images. From top to bottom, they are: infrared image, visible light image, fusion result of BMFusion, BMFusion without VGG loss, and BMFusion without gradient loss.</p>
Full article ">Figure 11
<p>Detection performance of our results with nine SOTA fusion results on different images from the LLVIP dataset.</p>
Full article ">
26 pages, 1206 KiB  
Article
Network Coding-Enhanced Polar Codes for Relay-Assisted Visible Light Communication Systems
by Congduan Li, Mingyang Zhong, Yiqian Zhang, Dan Song, Nanfeng Zhang and Jingfeng Yang
Entropy 2024, 26(12), 1112; https://doi.org/10.3390/e26121112 (registering DOI) - 19 Dec 2024
Abstract
This paper proposes a novel polar coding scheme tailored for indoor visible light communication (VLC) systems. Simulation results demonstrate a significant reduction in bit error rate (BER) compared to uncoded transmission, with a coding gain of at least 5 dB. Furthermore, the reliable [...] Read more.
This paper proposes a novel polar coding scheme tailored for indoor visible light communication (VLC) systems. Simulation results demonstrate a significant reduction in bit error rate (BER) compared to uncoded transmission, with a coding gain of at least 5 dB. Furthermore, the reliable communication area of the VLC system is substantially extended. Building on this foundation, this study explores the joint design of polar codes and physical-layer network coding (PNC) for VLC systems. Simulation results illustrate that the BER of our scheme closely approaches that of the conventional VLC relay scheme. Moreover, our approach doubles the throughput, cuts equipment expenses in half, and boosts effective bit rates per unit time-slot twofold. This proposed design noticeably advances the performance of VLC systems and is particularly well-suited for scenarios with low-latency demands. Full article
(This article belongs to the Special Issue Advances in Modern Channel Coding)
30 pages, 13159 KiB  
Article
GLMAFuse: A Dual-Stream Infrared and Visible Image Fusion Framework Integrating Local and Global Features with Multi-Scale Attention
by Fu Li, Yanghai Gu, Ming Zhao, Deji Chen and Quan Wang
Electronics 2024, 13(24), 5002; https://doi.org/10.3390/electronics13245002 - 19 Dec 2024
Abstract
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to [...] Read more.
Integrating infrared and visible-light images facilitates a more comprehensive understanding of scenes by amalgamating dual-sensor data derived from identical environments. Traditional CNN-based fusion techniques are predominantly confined to local feature emphasis due to their inherently limited receptive fields. Conversely, Transformer-based models tend to prioritize global information, which can lead to a deficiency in feature diversity and detail retention. Furthermore, methods reliant on single-scale feature extraction are inadequate for capturing extensive scene information. To address these limitations, this study presents GLMAFuse, an innovative dual-stream encoder–decoder network, which utilizes a multi-scale attention mechanism to harmoniously integrate global and local features. This framework is designed to maximize the extraction of multi-scale features from source images while effectively synthesizing local and global information across all layers. We introduce the global-aware and local embedding (GALE) module to adeptly capture and merge global structural attributes and localized details from infrared and visible imagery via a parallel dual-branch architecture. Additionally, the multi-scale attention fusion (MSAF) module is engineered to optimize attention weights at the channel level, facilitating an enhanced synergy between high-frequency edge details and global backgrounds. This promotes effective interaction and fusion of dual-modal features. Extensive evaluations using standard datasets demonstrate that GLMAFuse surpasses the existing leading methods in both qualitative and quantitative assessments, highlighting its superior capability in infrared and visible image fusion. On the TNO and MSRS datasets, our method achieves outstanding performance across multiple metrics, including EN (7.15, 6.75), SD (46.72, 47.55), SF (12.79, 12.56), MI (2.21, 3.22), SCD (1.75, 1.80), VIF (0.79, 1.08), Qbaf (0.58, 0.71), and SSIM (0.99, 1.00). These results underscore its exceptional proficiency in infrared and visible image fusion. Full article
(This article belongs to the Special Issue Artificial Intelligence Innovations in Image Processing)
Show Figures

Figure 1

Figure 1
<p>The qualitative fusion results based on CNN, AE, GAN, Transformer frameworks, and GLMAFuse.</p>
Full article ">Figure 2
<p>Deep model based on feature-level fusion.</p>
Full article ">Figure 3
<p>The framework of the proposed GLMAFuse for IVIF (where <math display="inline"><semantics> <mrow> <msub> <mrow> <mi mathvariant="bold">I</mi> </mrow> <mrow> <mi mathvariant="bold">F</mi> </mrow> </msub> </mrow> </semantics></math> means fused image, <math display="inline"><semantics> <mrow> <msub> <mrow> <mi mathvariant="bold">I</mi> </mrow> <mrow> <mi mathvariant="bold">i</mi> <mi mathvariant="bold">r</mi> </mrow> </msub> </mrow> </semantics></math> means infrared image, and <math display="inline"><semantics> <mrow> <msub> <mrow> <mi mathvariant="bold">I</mi> </mrow> <mrow> <mi mathvariant="bold">v</mi> <mi mathvariant="bold">i</mi> <mi mathvariant="bold">s</mi> </mrow> </msub> </mrow> </semantics></math> means visible image).</p>
Full article ">Figure 4
<p>Overview of the global-aware and local embedding module.</p>
Full article ">Figure 5
<p>Overview of the dual-modal interactive residual fusion block.</p>
Full article ">Figure 6
<p>Comparative analysis of visual fusion methods on the TNO dataset. Subfigure (<b>a</b>) displays images of rural scene, while subfigure (<b>b</b>) shows images of urban scene.</p>
Full article ">Figure 7
<p>Cumulative distribution of 8 metrics from the TNO dataset. The point on the curve <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="normal">x</mi> <mo>,</mo> <mi mathvariant="normal">y</mi> <mo>)</mo> </mrow> </semantics></math> indicates that there are <math display="inline"><semantics> <mrow> <mo>(</mo> <mn>100</mn> <mo>×</mo> <mi mathvariant="normal">x</mi> <mo>)</mo> <mi mathvariant="normal">%</mi> </mrow> </semantics></math> image pairs with metric values not exceeding y.</p>
Full article ">Figure 8
<p>Qualitative comparison results on the MSRS dataset. Subfigure (<b>a</b>) displays images of nighttime road scene, while subfigure (<b>b</b>) shows images of daytime road scene.</p>
Full article ">Figure 9
<p>Cumulative distribution of 8 metrics from the MSRS dataset. The point on the curve <math display="inline"><semantics> <mrow> <mo>(</mo> <mi mathvariant="normal">x</mi> <mo>,</mo> <mi mathvariant="normal">y</mi> <mo>)</mo> </mrow> </semantics></math> indicates that there are <math display="inline"><semantics> <mrow> <mo>(</mo> <mn>100</mn> <mo>×</mo> <mi mathvariant="normal">x</mi> <mo>)</mo> <mi mathvariant="normal">%</mi> </mrow> </semantics></math> of image pairs with metric values not exceeding y.</p>
Full article ">Figure 10
<p>Qualitative results of the ablation experiment of GALE on the Roadscene dataset.</p>
Full article ">Figure 11
<p>Qualitative results of the ablation experiment of MSAF on the Roadscene dataset.</p>
Full article ">
19 pages, 4353 KiB  
Article
Fusarium Wilt of Banana Latency and Onset Detection Based on Visible/Near Infrared Spectral Technology
by Cuiling Li, Dandan Xiang, Shuo Yang, Xiu Wang and Chunyu Li
Agronomy 2024, 14(12), 2994; https://doi.org/10.3390/agronomy14122994 - 16 Dec 2024
Viewed by 317
Abstract
Fusarium wilt of banana is a soil-borne vascular disease caused by Fusarium oxysporum f. sp. cubense. The rapid and accurate detection of this disease is of great significance to controlling its spread. The research objective was to explore rapid banana Fusarium wilt [...] Read more.
Fusarium wilt of banana is a soil-borne vascular disease caused by Fusarium oxysporum f. sp. cubense. The rapid and accurate detection of this disease is of great significance to controlling its spread. The research objective was to explore rapid banana Fusarium wilt latency and onset detection methods and establish a disease severity grading model. Visible/near-infrared spectroscopy analysis combined with machine learning methods were used for the rapid in vivo detection of banana Fusarium wilt. A portable visible/near-infrared spectrum acquisition system was constructed to collect the spectra data of banana Fusarium wilt leaves representing five different disease grades, totaling 106 leaf samples which were randomly divided into a training set with 80 samples and a test set with 26 samples. Different data preprocessing methods were utilized, and Fisher discriminant analysis (FDA), an extreme learning machine (ELM), and a one-dimensional convolutional neural network (1D-CNN) were used to establish the classification models of the disease grades. The classification accuracies of the FDA, ELM, and 1D-CNN models reached 0.891, 0.989, and 0.904, respectively. The results showed that the proposed visible/near infrared spectroscopy detection method could realize the detection of the incubation period of banana Fusarium wilt and the classification of the disease severity and could be a favorable tool for the field diagnosis of banana Fusarium wilt. Full article
(This article belongs to the Section Pest and Disease Management)
Show Figures

Figure 1

Figure 1
<p>Structure schematic diagram of the visible/near-infrared spectrum acquisition system.</p>
Full article ">Figure 2
<p>Internal structure diagram of the portable mainframe.</p>
Full article ">Figure 3
<p>Spectral data acquisition and the red dot on the left image represents the site where the symptom was observed and the spectral data were taken.</p>
Full article ">Figure 4
<p>Original reflectance spectrum of banana leaf samples. G0, G1, G2, G3, and G4 represent disease “Grade 0”, “Grade 1”, “Grade 2”, “Grade 3”, and “Grade 4”, respectively.</p>
Full article ">Figure 5
<p>Average original reflectance spectral curves of banana leaves for the five disease grades.</p>
Full article ">Figure 6
<p>The 1D-CNN network structure of the banana Fusarium wilt disease grading model.</p>
Full article ">Figure 7
<p>Reflectance spectra curves after pretreatment with SG smoothing method.</p>
Full article ">Figure 8
<p>Score coefficient curves of the first principal component PC1 based on the original spectral data.</p>
Full article ">Figure 9
<p>The 3D distribution of PC1, PC2, and PC3 based on the original spectral data.</p>
Full article ">Figure 10
<p>Disease grade classification results on test set of FD-MSC-SG-FDA model. G0, G1, G2, G3, and G4 represent disease “Grade 0”, “Grade 1”, “Grade 2”, “Grade 3”, and “Grade 4”, respectively.</p>
Full article ">Figure 11
<p>Influence of the number of hidden layer neurons on the MSC-SG-ELM model.</p>
Full article ">Figure 12
<p>Disease grade classification results of MSC-SG-ELM model on test set. G0, G1, G2, G3, and G4 represent disease Grade 0, Grade 1, Grade 2, Grade 3, and Grade 4, respectively.</p>
Full article ">Figure 13
<p>Loss values of the test set at different learning rates, each color represents a learning rate.</p>
Full article ">Figure 14
<p>Accuracies of the training set and test set at different learning rates, each color represents a learning rate.</p>
Full article ">Figure 15
<p>Loss values and accuracies when the learning rate was 0.01.</p>
Full article ">
20 pages, 10226 KiB  
Article
Coffee Leaf Rust Disease Detection and Implementation of an Edge Device for Pruning Infected Leaves via Deep Learning Algorithms
by Raka Thoriq Araaf, Arkar Minn and Tofael Ahamed
Sensors 2024, 24(24), 8018; https://doi.org/10.3390/s24248018 - 16 Dec 2024
Viewed by 312
Abstract
Global warming and extreme climate conditions caused by unsuitable temperature and humidity lead to coffee leaf rust (Hemileia vastatrix) diseases in coffee plantations. Coffee leaf rust is a severe problem that reduces productivity. Currently, pesticide spraying is considered the most effective [...] Read more.
Global warming and extreme climate conditions caused by unsuitable temperature and humidity lead to coffee leaf rust (Hemileia vastatrix) diseases in coffee plantations. Coffee leaf rust is a severe problem that reduces productivity. Currently, pesticide spraying is considered the most effective solution for mitigating coffee leaf rust. However, the application of pesticide spray is still not efficient for most farmers worldwide. In these cases, pruning the most infected leaves with leaf rust at coffee plantations is important to help pesticide spraying to be more efficient by creating a more targeted, accessible treatment. Therefore, detecting coffee leaf rust is important to support the decision on pruning infected leaves. The dataset was acquired from a coffee farm in Majalengka Regency, Indonesia. Only images with clearly visible spots of coffee leaf rust were selected. Data collection was performed via two devices, a digital mirrorless camera and a phone camera, to diversify the dataset and test it with different datasets. The dataset, comprising a total of 2024 images, was divided into three sets with a ratio of 70% for training (1417 images), 20% for validation (405 images), and 10% for testing (202 images). Images with leaves infected by coffee leaf rust were labeled via LabelImg® with the label “CLR”. All labeled images were used to train the YOLOv5 and YOLOv8 algorithms through the convolutional neural network (CNN). The trained model was tested with a test dataset, a digital mirrorless camera image dataset (100 images), a phone camera dataset (100 images), and real-time detection with a coffee leaf rust image dataset. After the model was trained, coffee leaf rust was detected in each frame. The mean average precision (mAP) and recall for the trained YOLOv5 model were 69% and 63.4%, respectively. For YOLOv8, the mAP and recall were approximately 70.2% and 65.9%, respectively. To evaluate the performance of the two trained models in detecting coffee leaf rust on trees, 202 original images were used for testing with the best-trained weight from each model. Compared to YOLOv5, YOLOv8 demonstrated superior accuracy in detecting coffee leaf rust. With a mAP of 73.2%, YOLOv8 outperformed YOLOv5, which achieved a mAP of 70.5%. An edge device was utilized to deploy real-time detection of CLR with the best-trained model. The detection was successfully executed with high confidence in detecting CLR. The system was further integrated into pruning solutions for Arabica coffee farms. A pruning device was designed using Autodesk Fusion 360® and fabricated for testing on a coffee plantation in Indonesia. Full article
(This article belongs to the Special Issue Deep Learning for Intelligent Systems: Challenges and Opportunities)
Show Figures

Figure 1

Figure 1
<p>Coffee leaf rust disease under severe conditions.</p>
Full article ">Figure 2
<p>Conceptual research framework.</p>
Full article ">Figure 3
<p>Arabica coffee farm in Majalengka District.</p>
Full article ">Figure 4
<p>Image augmentation: (<b>a</b>) original image, (<b>b</b>) flipped horizontally, (<b>c</b>) flipped vertically, (<b>d</b>) rotated 90°, (<b>e</b>) rotated 180°, and (<b>f</b>) rotated 270°.</p>
Full article ">Figure 5
<p>Stages of coffee leaf rust disease in the dataset: (<b>a</b>) healthy, (<b>b</b>) early stage, (<b>c</b>) severe stage with chlorosis, (<b>d</b>) severe stage with chlorosis and defoliation.</p>
Full article ">Figure 6
<p>CLR detection process on the YOLO framework.</p>
Full article ">Figure 7
<p>YOLOv5 training configuration.</p>
Full article ">Figure 8
<p>YOLOv8 training configuration.</p>
Full article ">Figure 9
<p>Detection results on the testing dataset comprising images from the digital mirrorless camera: (<b>a</b>) original image, (<b>b</b>) YOLOv5 detection, and (<b>c</b>) YOLOv8 detection.</p>
Full article ">Figure 10
<p>Detection results on the testing dataset comprising images from the phone camera: (<b>a</b>) original image, (<b>b</b>) YOLOv5 detection, and (<b>c</b>) YOLOv8 detection.</p>
Full article ">Figure 11
<p>Real-time detection of CLR on Jetson Nano using YOLOv8: (<b>a</b>) infected and (<b>b</b>) healthy images.</p>
Full article ">Figure 12
<p>Challenge in dataset quality: (<b>a</b>) low-quality image (blurry) and (<b>b</b>) occluded object.</p>
Full article ">Figure 13
<p>Design of the pruning device components. (<b>a</b>) Pruning device, i.e., cutting part, (<b>b</b>) slicing part, and (<b>c</b>) vacuum tank.</p>
Full article ">Figure 14
<p>Design of the test for real-time detection before deployment in the field at an Arabica coffee farm: (<b>a</b>) Device components and (<b>b</b>) real-time detection.</p>
Full article ">
22 pages, 6302 KiB  
Article
Field Grading of Longan SSC via Vis-NIR and Improved BP Neural Network
by Jun Li, Meiqi Zhang, Kaixuan Wu, Hengxu Chen, Zhe Ma, Juan Xia and Guangwen Huang
Agriculture 2024, 14(12), 2297; https://doi.org/10.3390/agriculture14122297 - 14 Dec 2024
Viewed by 490
Abstract
Soluble solids content (SSC) measurements are crucial for managing longan production and post-harvest handling. However, most traditional SSC detection methods are destructive, cumbersome, and unsuitable for field applications. This study proposes a novel field detection model (Brix-back propagation neural network, Brix-BPNN), designed for [...] Read more.
Soluble solids content (SSC) measurements are crucial for managing longan production and post-harvest handling. However, most traditional SSC detection methods are destructive, cumbersome, and unsuitable for field applications. This study proposes a novel field detection model (Brix-back propagation neural network, Brix-BPNN), designed for longan SSC grading based on an improved BP neural network. Initially, nine preprocessing methods were combined with six classification algorithms to develop the longan SSC grading prediction model. Among these, the model preprocessed with Savitzky–Golay smoothing and the first derivative (SG-D1) demonstrated a 7.02% improvement in accuracy compared to the original spectral model. Subsequently, the BP network structure was refined, and the competitive adaptive reweighted sampling (CARS) algorithm was employed for feature wavelength extraction. The results show that the improved Brix-BPNN model, integrated with the CARS, achieves the highest prediction performance, with a 2.84% increase in classification accuracy relative to the original BPNN model. Additionally, the number of wavelengths is reduced by 92% compared to the full spectrum, making this model both lightweight and efficient for rapid field detection. Furthermore, a portable detection device based on visible-near-infrared (Vis-NIR) spectroscopy was developed for longan SSC grading, achieving a prediction accuracy of 83.33% and enabling fast, nondestructive testing in field conditions. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Block diagram of the hardware system structure.</p>
Full article ">Figure 2
<p>Perspective view of the device.</p>
Full article ">Figure 3
<p>Physical drawing of the device: (<b>a</b>) internal structure of the device; (<b>b</b>) overall view of the device.</p>
Full article ">Figure 4
<p>Spectral collection points.</p>
Full article ">Figure 5
<p>Overall structures of the BPNN and Brix-BPNN.</p>
Full article ">Figure 6
<p>Structural diagram of the ECA-Brix attention mechanism.</p>
Full article ">Figure 7
<p>H-swish and ReLU activation functions.</p>
Full article ">Figure 8
<p>Max pooling process for longan spectral data.</p>
Full article ">Figure 9
<p>Original spectral curves.</p>
Full article ">Figure 10
<p>Average spectral curves of the three SSC grades.</p>
Full article ">Figure 11
<p>Feature wavelength selection of longan SSC based on the SPA algorithm: (<b>a</b>) RMSE variation of the model; (<b>b</b>) optimal feature wavelength selected by SPA.</p>
Full article ">Figure 11 Cont.
<p>Feature wavelength selection of longan SSC based on the SPA algorithm: (<b>a</b>) RMSE variation of the model; (<b>b</b>) optimal feature wavelength selected by SPA.</p>
Full article ">Figure 12
<p>Feature wavelength selection of longan SSC based on the CARS algorithm: (<b>a</b>) number of sample variables; (<b>b</b>) RMSECV.</p>
Full article ">Figure 12 Cont.
<p>Feature wavelength selection of longan SSC based on the CARS algorithm: (<b>a</b>) number of sample variables; (<b>b</b>) RMSECV.</p>
Full article ">Figure 13
<p>Statistical chart of the true and predicted labels for the SSC grade of the test samples.</p>
Full article ">
15 pages, 5399 KiB  
Article
Studies on Morphological Evolution of Gravure-Printed ZnO Thin Films Induced by Low-Temperature Vapor Post-Treatment
by Giuliano Sico, Vincenzo Guarino, Carmela Borriello and Maria Montanino
Nanomaterials 2024, 14(24), 2006; https://doi.org/10.3390/nano14242006 - 13 Dec 2024
Viewed by 498
Abstract
In recent years, the morphology control of semiconductor nanomaterials has been attracting increasing attention toward maximizing their functional properties and reaching their end use in real-world devices. However, the development of easy and cost-effective methods for preparing large-scale patterned semiconductor structures on flexible [...] Read more.
In recent years, the morphology control of semiconductor nanomaterials has been attracting increasing attention toward maximizing their functional properties and reaching their end use in real-world devices. However, the development of easy and cost-effective methods for preparing large-scale patterned semiconductor structures on flexible temperature-sensitive substrates remains ever in demand. In this study, vapor post-treatment (VPT) is investigated as a potential, simple and low-cost post-preparative method to morphologically modify gravure-printed zinc oxide (ZnO) nanoparticulate thin films at low temperatures. Exposing nanoparticles (NPs) to acidic vapor solution, spontaneous restructuring pathways are observed as a consequence of NPs tending to reduce their high interfacial energy. Depending on the imposed environmental conditions during the treatment (e.g., temperature, vapor composition), various ZnO thin-film morphologies are produced, from dense to porous ones, as a result of the activation and interplay of different spontaneous interface elimination mechanisms, including dissolution–precipitation, grain boundary migration and grain rotation–coalescence. The influence of VPT on structural/optical properties has been examined via XRD, UV–visible and photoluminescence measurements. Controlling NP junctions and network nanoporosity, VPT appears as promising cost-effective, low-temperature and pressureless post-preparative platform for preparing supported ZnO NP-based films with improved connectivity and mechanical stability, favoring their practical use and integration in flexible devices. Full article
(This article belongs to the Section Physical Chemistry at Nanoscale)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>As-printed (untreated) ZnO NP film exposed to the vapor of a 1 M acetic acid aqueous solution in a closed oven at 50 °C: t = 0 (<b>a</b>); t &gt; 150 min (<b>b</b>).</p>
Full article ">Figure 2
<p>Microstructure at t = 150 min of as-printed ZnO NP films changing VPT environmental conditions at t &gt; 75 min: by replacing the vaporizing acid solution with distilled water maintaining the temperature constant at 50 °C (<b>a</b>); by isothermal heating at 70 °C maintaining the acidic atmospheric composition (<b>b</b>).</p>
Full article ">Figure 3
<p>Schematic of the studied two-step vapor post-treatment profiles reported in the Materials and Methods section.</p>
Full article ">Figure 4
<p>Microstructural evolution of as-printed ZnO NP film as a result of the treatment profile A: early stage of NP sliding and reorientation (<b>a</b>); intermediate stage of NP coalescence (<b>b</b>); porous leafage-like nanostructured network (<b>c</b>).</p>
Full article ">Figure 5
<p>Microstructural evolution of as-printed ZnO NP films as a result of the treatment profile B (<b>a</b>) and C (<b>b</b>).</p>
Full article ">Figure 6
<p>Structural and optical characterization of gravure-printed ZnO films on aluminum foil subjected to VPT profiles as reported in the Materials and Methods section: (<b>a</b>) XRD patterns; (<b>b</b>) optical transmittance; (<b>c</b>) photoluminescence spectra. Pristine refers to the as-printed (untreated) ZnO sample.</p>
Full article ">Figure 7
<p>Diagram of the studied VPT processes.</p>
Full article ">
18 pages, 5411 KiB  
Article
Leveraging Neural Radiance Fields for Large-Scale 3D Reconstruction from Aerial Imagery
by Max Hermann, Hyovin Kwak, Boitumelo Ruf and Martin Weinmann
Remote Sens. 2024, 16(24), 4655; https://doi.org/10.3390/rs16244655 - 12 Dec 2024
Viewed by 432
Abstract
Since conventional photogrammetric approaches struggle with with low-texture, reflective, and transparent regions, this study explores the application of Neural Radiance Fields (NeRFs) for large-scale 3D reconstruction of outdoor scenes, since NeRF-based methods have recently shown very impressive results in these areas. We evaluate [...] Read more.
Since conventional photogrammetric approaches struggle with with low-texture, reflective, and transparent regions, this study explores the application of Neural Radiance Fields (NeRFs) for large-scale 3D reconstruction of outdoor scenes, since NeRF-based methods have recently shown very impressive results in these areas. We evaluate three approaches: Mega-NeRF, Block-NeRF, and Direct Voxel Grid Optimization, focusing on their accuracy and completeness compared to ground truth point clouds. In addition, we analyze the effects of using multiple sub-modules, estimating the visibility by an additional neural network and varying the density threshold for the extraction of the point cloud. For performance evaluation, we use benchmark datasets that correspond to the setting off standard flight campaigns and therefore typically have nadir camera perspective and relatively little image overlap, which can be challenging for NeRF-based approaches that are typically trained with significantly more images and varying camera angles. We show that despite lower quality compared to classic photogrammetric approaches, NeRF-based reconstructions provide visually convincing results in challenging areas. Furthermore, our study shows that in particular increasing the number of sub-modules and predicting the visibility using an additional neural network improves the quality of the resulting reconstructions significantly. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of our methodology. The input data consists of images and their camera calibration. Based on this, the images are clustered using the camera centers and assigned to individual sub-modules, which are then reconstructed as NeRF volumes. The point clouds are then extracted from the camera viewpoint using random sampling.</p>
Full article ">Figure 2
<p>Visualization of the cluster pre-processing based on the position of the camera centers. The colors symbolize the four sub-modules.</p>
Full article ">Figure 3
<p>Point clouds derived for the TMB dataset. Each row shows examples of one approach, with COLMAP as a reference in the first row.</p>
Full article ">Figure 4
<p>Point clouds derived for the UseGeo3 dataset. Each row shows examples of one approach, with COLMAP as a reference in the first row.</p>
Full article ">Figure 5
<p>Point clouds derived for the Hessigheim 3D dataset. Each row shows examples of one approach, with COLMAP as a reference in the first row.</p>
Full article ">Figure 6
<p>Qualitative evaluation of different density thresholds using Mega-NeRF. As the density threshold increases, roads and canals in particular show significantly more holes.</p>
Full article ">Figure 7
<p>Relationship between accuracy, completeness and the density threshold. The letters represent the three methods analyzed: Mega-NeRF (M), Block-NeRF (B) and DVGO (D). The numbers in front of them refer to their position in <a href="#remotesensing-16-04655-t002" class="html-table">Table 2</a>, whereas the number after the letter indicates the number of sub-modules. The black arrows indicate the approximate direction of the values over the increasing threshold.</p>
Full article ">Figure 8
<p>Reconstructions with and without VisibilityNet derived for the UseGeo and Hessigheim 3D dataset. B in the first column denotes Block-NeRF. Point clouds extracted using the VisibilityNet show a better level of detail in general.</p>
Full article ">Figure 9
<p>Mip-NeRF based point cloud reconstruction without and with an identity function for the positional encoding, which means additionally concatenating the original input coordinates. Positional encoded features without an identity function lead to points that are scattered over several periodic layers, whereas with an identity function these artifacts disappear.</p>
Full article ">
23 pages, 18600 KiB  
Article
Cross-Modality Data Augmentation for Aerial Object Detection with Representation Learning
by Chiheng Wei, Lianfa Bai, Xiaoyu Chen and Jing Han
Remote Sens. 2024, 16(24), 4649; https://doi.org/10.3390/rs16244649 - 12 Dec 2024
Viewed by 294
Abstract
Data augmentation methods offer a cost-effective and efficient alternative to the acquisition of additional data, significantly enhancing data diversity and model generalization, making them particularly favored in object detection tasks. However, existing data augmentation techniques primarily focus on the visible spectrum and are [...] Read more.
Data augmentation methods offer a cost-effective and efficient alternative to the acquisition of additional data, significantly enhancing data diversity and model generalization, making them particularly favored in object detection tasks. However, existing data augmentation techniques primarily focus on the visible spectrum and are directly applied to RGB-T object detection tasks, overlooking the inherent differences in image data between the two tasks. Visible images capture rich color and texture information during the daytime, while infrared images are capable of imaging under low-light complex scenarios during the nighttime. By integrating image information from both modalities, their complementary characteristics can be exploited to improve the overall effectiveness of data augmentation methods. To address this, we propose a cross-modality data augmentation method tailored for RGB-T object detection, leveraging masked image modeling within representation learning. Specifically, we focus on the temporal consistency of infrared images and combine them with visible images under varying lighting conditions for joint data augmentation, thereby enhancing the realism of the augmented images. Utilizing the masked image modeling method, we reconstruct images by integrating multimodal features, achieving cross-modality data augmentation in feature space. Additionally, we investigate the differences and complementarities between data augmentation methods in data space and feature space. Building upon existing theoretical foundations, we propose an integrative framework that combines these methods for improved augmentation effectiveness. Furthermore, we address the slow convergence observed with the existing Mosaic method in aerial imagery by introducing a multi-scale training strategy and proposing a full-scale Mosaic method as a complement. This optimization significantly accelerates network convergence. The experimental results validate the effectiveness of our proposed method and highlight its potential for further advancements in cross-modality object detection tasks. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The label errors that may be introduced by data augmentation methods based on the data space. Figure (<b>a</b>) shows an example of a label error in an image classification task: after aggressive cropping, the image’s class changes from “bicycle” to “wheel”. Figure (<b>b</b>) provides an example of a label error in an object detection task, where after applying cropping, the retained portions of two different objects exhibit a high degree of class similarity, potentially introducing a labeling error.</p>
Full article ">Figure 2
<p>The cross-modality data augmentation framework in the feature space consists of an object filtering and editing module and an RGBTMAE reconstruction module. This figure uses the object replacement mode as an example of the object filtering and editing module. In the object filtering and editing module, a classifier is trained to focus on foreground objects, filtering and editing them based on confidence scores. In the RGBTMAE reconstruction module, the correlation between different modality images is utilized to reconstruct the edited images, thereby reducing unnatural transitions between the foreground and background after object editing. In the figure, (<b>a</b>) represents the input raw image, (<b>b</b>) illustrates the image processed through the object filtering and editing module, and (<b>c</b>) shows the reconstructed image produced by the data augmentation method.</p>
Full article ">Figure 3
<p>Comparison of original and processed visible images using different foreground operations in the object filtering and editing module. Images marked in red represent the original input, while those marked in green depict the processed results: (<b>a</b>) shows the outcome of object removal, (<b>b</b>) illustrates object replacement results, and (<b>c</b>) demonstrates the object copy–pasting results.</p>
Full article ">Figure 4
<p>The training and reconstruction phases of the RGBTMAE reconstruction module. During the training phase, the network framework takes two masked modality images as input, with the original images serving as the reconstruction target. In the reconstruction phase, the network input consists of visible and infrared images, where the foreground has been edited. Infrared image information, unaffected by lighting variations, is fully retained, while masked regions cover the replaced foreground areas in the visible images. After reconstruction by the network and subsequent image post-processing, a new set of reconstructed images is obtained.</p>
Full article ">Figure 5
<p>The structure of the RGBT oriented object detection network based on RoI Transformer. The middle fusion strategy is employed to balance fusion performance and model complexity.</p>
Full article ">Figure 6
<p>The effects of different data augmentation methods on the DroneVehicle dataset. (<b>a</b>) shows the original visible and infrared images; (<b>b</b>) illustrates the image augmentation effect of the CutOut method, where portions of the objects are obscured by masks; (<b>c</b>) presents the image augmentation effect of the MixUp method, showing darkened tones in the visible images and overlapping of some objects; (<b>d</b>) demonstrates the image augmentation effect of our proposed full-scale Mosaic method, which significantly enhances the training efficiency of the network. The second row depicts image augmentation methods based on feature space, where visible images are reconstructed from infrared images for data augmentation; (<b>e</b>) shows the image reconstruction results of the CycleGAN method; (<b>f</b>) presents the results of the Pixel2PixelHD method; (<b>g</b>) illustrates the reconstruction results of the DR-AVIT method; and (<b>h</b>) shows the augmentation effect of our proposed cross-modality data augmentation method based on representation learning. It can be observed that other feature-space-based augmentation methods result in reconstructed images with significant deviations from the distribution of real images. In contrast, our method edits the foreground objects while preserving the background of real images as much as possible. The green highlighted regions in the figure indicate the effects of foreground editing, while the red highlighted regions correspond to the respective regions in the original image. In the subsequent experiments, we will integrate methods (<b>d</b>,<b>h</b>) to evaluate the performance of combining both data augmentation approaches.</p>
Full article ">Figure 7
<p>Visual comparison of different feature-space data augmentation methods on dual-modality datasets. The first row presents the reconstruction results on the DroneVehicle dataset, while the second row shows the reconstruction results on the VEDAI dataset. “RGB” denotes the corresponding real visible images. Highlighted regions indicate slight color discrepancies between the reconstruction results of the RGBTMAE method and the real images.</p>
Full article ">Figure 8
<p>The effects of different object filtering and editing methods on the DroneVehicle dataset are presented. (<b>a</b>) displays the results of randomly replacing the foreground object, (<b>b</b>) displays the foreground object replacement results based on our representation learning approach, and (<b>c</b>) shows the effect of removing the foreground objects based on our representation learning approach. It is important to note that the optimization of image effects was conducted only on visible images, where brightness differences are more pronounced. The highlighted regions in the images represent areas with dense objects. As shown in (<b>b</b>), the replaced foreground objects reconstructed using our method better align with the overall image distribution and transition more naturally with the background regions.</p>
Full article ">Figure 9
<p>Ablation experiment results for the two Mosaic methods under different data conditions, where the dataset were expanded by replicating the original data. The experimental results show that the full-scale Mosaic method achieves the fastest convergence speed.</p>
Full article ">
18 pages, 10262 KiB  
Article
Fault Diagnosis of Mechanical Rolling Bearings Using a Convolutional Neural Network–Gated Recurrent Unit Method with Envelope Analysis and Adaptive Mean Filtering
by Huiyi Zhu, Zhen Sui, Jianliang Xu and Yeshen Lan
Processes 2024, 12(12), 2845; https://doi.org/10.3390/pr12122845 - 12 Dec 2024
Viewed by 340
Abstract
Rolling bearings are vital components in rotating machinery, and their reliable operation is crucial for maintaining the stability and efficiency of mechanical systems. However, fault detection in rolling bearings is often hindered by noise interference in complex industrial environments. To overcome this challenge, [...] Read more.
Rolling bearings are vital components in rotating machinery, and their reliable operation is crucial for maintaining the stability and efficiency of mechanical systems. However, fault detection in rolling bearings is often hindered by noise interference in complex industrial environments. To overcome this challenge, this paper presents a novel fault diagnosis method for rolling bearings, combining Convolutional Neural Networks (CNNs) and Gated Recurrent Units (GRUs), integrated with the envelope analysis and adaptive mean filtering techniques. Initially, envelope analysis and adaptive mean filtering are applied to suppress random noise in the bearing signals, thereby enhancing the visibility of fault features. Subsequently, a deep learning model that combines a CNN and a GRU is developed: the CNN extracts spatial features, while the GRU captures the temporal dependencies between these features. The integration of the CNN and GRU significantly improves the accuracy and robustness of fault diagnosis. The proposed method is validated using the CWRU dataset, with the experimental results achieving an average accuracy of 99.25%. Additionally, the method is compared to four classical fault diagnosis models, demonstrating superior performance in terms of both diagnostic accuracy and generalization ability. The results, supported by various visualization techniques, show that the proposed approach effectively addresses the challenges of fault detection in rolling bearings under complex industrial conditions. Full article
Show Figures

Figure 1

Figure 1
<p>CNN structure diagram.</p>
Full article ">Figure 2
<p>GRU network structure diagram.</p>
Full article ">Figure 3
<p>Experimental diagram of CWRU bearing equipment.</p>
Full article ">Figure 4
<p>Initial data curve of bearings.</p>
Full article ">Figure 5
<p>Comparison results of different filtering algorithms for bearings.</p>
Full article ">Figure 6
<p>Box plot of 10 categories of bearing data ((<b>a</b>) original bearing data; (<b>b</b>) filtered bearing data).</p>
Full article ">Figure 7
<p>CNN-GRU fault diagnosis model.</p>
Full article ">Figure 8
<p>Accuracy and loss value variation curve during training process.</p>
Full article ">Figure 9
<p>A visual display of the confusion matrix in the test set.</p>
Full article ">Figure 9 Cont.
<p>A visual display of the confusion matrix in the test set.</p>
Full article ">Figure 10
<p>Visualization of t-SNE clustering in different processes of the model.</p>
Full article ">
17 pages, 6640 KiB  
Article
Analysis of Tidal Cycle Wave Breaking Distribution Characteristics on a Low-Tide Terrace Beach Using Video Imagery Segmentation
by Hang Yin, Feng Cai, Hongshuai Qi, Yuwu Jiang, Gen Liu, Zhubin Cao, Yi Sun and Zheyu Xiao
Remote Sens. 2024, 16(24), 4616; https://doi.org/10.3390/rs16244616 - 10 Dec 2024
Viewed by 410
Abstract
Wave breaking is a fundamental process in ocean energy dissipation and plays a crucial role in the exchange between ocean and nearshore sediments. Foam, the primary visible feature of wave breaking areas, serves as a direct indicator of wave breaking processes. Monitoring the [...] Read more.
Wave breaking is a fundamental process in ocean energy dissipation and plays a crucial role in the exchange between ocean and nearshore sediments. Foam, the primary visible feature of wave breaking areas, serves as a direct indicator of wave breaking processes. Monitoring the distribution of foam via remote sensing can reveal the spatiotemporal patterns of nearshore wave breaking. Existing studies on wave breaking processes primarily focus on individual wave events or short timescales, limiting their effectiveness for nearshore regions where hydrodynamic processes are often represented at tidal cycles. In this study, video imagery from a typical low-tide terrace (LTT) beach was segmented into four categories, including the wave breaking foam, using the DeepLabv3+ architecture, a convolutional neural networks (CNNs)-based model suitable for semantic segmentation in complex visual scenes. After training and testing on a manually labelled dataset, which was divided into training, validation, and testing sets based on different time periods, the overall classification accuracy of the model was 96.4%, with an accuracy of 96.2% for detecting wave breaking foam. Subsequently, a heatmap of the wave breaking foam distribution over a tidal cycle on the LTT beach was generated. During the tidal cycle, the foam distribution density exhibited both alongshore variability, and a pronounced bimodal structure in the cross-shore direction. Analysis of morphodynamical data collected in the field indicated that the bimodal structure is primarily driven by tidal variations. The wave breaking process is a key factor in shaping the profile morphology of LTT beaches. High-frequency video monitoring further showed the wave breaking patterns vary significantly with tidal levels, leading to diverse geomorphological features at various cross-shore locations. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>,<b>b</b>) Geographical location of the study area, monitoring profile positions, and wave gauge locations (red dots); (<b>c</b>) shore-based video monitoring system at Xisha Bay; (<b>d</b>) deployment of the RBR solo3D|wave16 wave gauge.</p>
Full article ">Figure 2
<p>(<b>a</b>) Timex video imagery example; (<b>b</b>) the corresponding manually annotated labels used for model training; (<b>c</b>) the proportion of pixels with each label.</p>
Full article ">Figure 3
<p>The DeepLabv3+ architecture used in this study.</p>
Full article ">Figure 4
<p>Field dynamic observation results at P1 and P2 during the experimental period at Xisha Bay beach. (<b>a</b>) Tidal level; (<b>b</b>) significant wave height; (<b>c</b>) wave energy.</p>
Full article ">Figure 5
<p>The aggregated confusion matrix for all categories. The percentages represent the normalized number of pixel points.</p>
Full article ">Figure 6
<p>The segmentation results at different tidal levels.</p>
Full article ">Figure 7
<p>Heatmap of the wave breaking foam distribution at Xisha Bay beach over a tidal cycle.</p>
Full article ">Figure 8
<p>Cross-shore sampling results of the normalized foam density. (<b>a</b>) Profile elevations of P1; (<b>b</b>) the corresponding normalized foam density of P1 positions; (<b>c</b>) profile elevations of P2; (<b>d</b>) the corresponding normalized foam density of P2 positions.</p>
Full article ">Figure 9
<p>(<b>a</b>) Sampling positions at different tidal stages: on the low-tide terrace, near the inflection area, and near the beach cusp; (<b>b</b>) the width of the wave breaking foam distribution areas at the different sampling positions for profiles P1 and P2; (<b>c</b>) the corresponding tidal levels.</p>
Full article ">Figure 10
<p>Schematics of the wave breaker types on LTT beaches. (<b>a</b>,<b>d</b>) The wave breaker type at low tide with the corresponding snap video imagery; (<b>b</b>,<b>e</b>) the wave breaker type at mid-tide with the corresponding snap video imagery; (<b>c</b>,<b>f</b>) the wave breaker type at high tide with the corresponding snap video imagery.</p>
Full article ">Figure 11
<p>(<b>a</b>) Normalized foam density vs. tidal slope at P1 positions; (<b>b</b>) normalized foam density vs. wave energy at P1 positions; (<b>c</b>) normalized foam density vs. beach slope at P1 positions; (<b>d</b>) normalized foam density vs. tidal slope at P2 positions; (<b>e</b>) normalized foam density vs. wave energy at P2 positions; (<b>f</b>) normalized foam density vs. beach slope at P2 positions.</p>
Full article ">
16 pages, 6306 KiB  
Article
L-Cysteine/Silver Nitrate/Iodate Anions System: Peculiarities of Supramolecular Gel Formation with and Without Visible-Light Exposure
by Dmitry V. Vishnevetskii, Elizaveta E. Polyakova, Yana V. Andrianova, Arif R. Mekhtiev, Alexandra I. Ivanova, Dmitry V. Averkin, Vladimir G. Alekseev, Alexey V. Bykov and Mikhail G. Sulman
Gels 2024, 10(12), 809; https://doi.org/10.3390/gels10120809 - 9 Dec 2024
Viewed by 556
Abstract
In this study, novel anion photo-responsive supramolecular hydrogels based on cysteine–silver sol (CSS) and iodate anions (IO3) were prepared. The peculiarities of the self-assembly process of gel formation in the dark and under visible-light exposure were studied using a complex [...] Read more.
In this study, novel anion photo-responsive supramolecular hydrogels based on cysteine–silver sol (CSS) and iodate anions (IO3) were prepared. The peculiarities of the self-assembly process of gel formation in the dark and under visible-light exposure were studied using a complex of modern physico-chemical methods of analysis, including viscosimetry, UV spectroscopy, dynamic light scattering, electrophoretic light scattering, scanning electron microscopy, energy-dispersive X-ray spectroscopy, and X-ray photoelectron spectroscopy. In the dark phase, the formation of weak snot-like gels takes place in a quite narrow IO3 ion concentration range. The visible-light exposure of these gels leads to an increase in their viscosity and dramatic change in their color. The morphology of gels alters after light irradiation that is reflected in the formation of a huge number of spherical/elliptical particles and the thickening of the fibers of the gel network. The interaction of CSS with IO3 anions has features of a redox process, which leads to the formation of silver iodide/silver oxide nanoparticles inside and on the surface of CSS particles. CSS possesses selectivity only to IO3 anions compared to many other inorganic ions relevant for humans and the environment. Thus, the CSS/IO3 system is non-trivial and can be considered as a novel low-molecular-weight gelator with photosensitive properties, as another way to produce silver iodide nanoparticles, and as a new approach for IO3 ion detection. Full article
(This article belongs to the Special Issue Synthesis and Applications of Hydrogels (2nd Edition))
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>(<b>A</b>) Hydrogel formation in the CSS/IO<sub>3</sub><sup>−</sup> system. (<b>B</b>) The anion (XO<sub>3</sub>—ClO<sub>3</sub>, BrO<sub>3</sub> or IO<sub>3</sub>) content in the systems (Table); photo of CSGXO<sub>3</sub> systems 2, 4, 6, 8, 10, 12, 14, 16, 18, 20 (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B) after visible-light exposure for 1 h. CSGXO<sub>3</sub>—cysteine silver gels based on XO<sub>3</sub><sup>−</sup> anions.</p>
Full article ">Figure 2
<p>Concentration diagrams and viscosity dependence over time for CSS/IO<sub>3</sub><sup>−</sup>-based gels before (<b>a</b>,<b>c</b>) and after (<b>b</b>,<b>d</b>) visible-light exposure for 1 h. Numbers 1, 2, 3, 4, and 5 on graphs (<b>c</b>,<b>d</b>) correspond to samples 8, 10, 12, 14, and 16 (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B).</p>
Full article ">Figure 3
<p>SEM images and EDS of hydrogels before (dark) and after visible-light exposure (light) for 1 h. Images (<b>a</b>–<b>e</b>) correspond to gel systems 8, 10, 12, 14, and 16 (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B). EDS data are presented for irradiated samples. Photos are provided for the corresponding irradiated gels.</p>
Full article ">Figure 4
<p>UV spectra of the systems before (<b>a</b>) and after (<b>b</b>) visible-light exposure: 1, 2, and 3—aqueous solutions of KClO<sub>3</sub>, KBrO<sub>3</sub>, and KIO<sub>3</sub> respectively; 4—CSS; 5—CSS/ClO<sub>3</sub><sup>−</sup> and CSS/BrO<sub>3</sub><sup>−</sup>; 6—CSS/IO<sub>3</sub><sup>−</sup>. Data are presented for systems 2, 4, 6, 8, 10, 12, 14, 16, 18, and 20 (<a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B). (<b>c</b>) Comparison of UV spectra of the CSS/IO<sub>3</sub><sup>−</sup> systems before (bold lines) and after (dash lines) light irradiation. (<b>d</b>) Kinetics of UV spectra evolution for the CSS/IO<sub>3</sub><sup>−</sup> system 20 (<a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B) in the dark and under light exposure.</p>
Full article ">Figure 5
<p>Particle size and zeta potential distributions for CSS/IO<sub>3</sub><sup>−</sup> systems before (<b>a</b>,<b>c</b>) and after (<b>b</b>,<b>d</b>) visible-light exposure for 1 h. 1, 2, 3, 4, and 5 correspond to systems 8, 10, 12, 14, and 16 (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B), respectively.</p>
Full article ">Figure 6
<p>The XPS survey scan of CSS/IO<sub>3</sub><sup>−</sup> gel <b>16</b> (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B) after visible-light exposure for 1 h. The XPS spectrum for the non-irradiated hydrogel <b>16</b> remained unchanged.</p>
Full article ">Figure 7
<p>High-resolution XPS spectra of (<b>a</b>) I 3d, (<b>b</b>) Ag 3d, (<b>c</b>) Ag MNN, (<b>d</b>) S 2p, (<b>e</b>) O 1s, and (<b>f</b>) N 1s for the CSS/IO<sub>3</sub><sup>−</sup> gel <b>16</b> (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B) after visible-light exposure for 1 h.</p>
Full article ">Figure 8
<p>Photos of CSS with various anions: (<b>a</b>) immediately after adding electrolyte; (<b>b</b>) after 24 h in the dark; (<b>c</b>–<b>e</b>) after visible-light exposure of samples (photo (<b>b</b>)) for 20, 40, and 60 min, respectively. (<b>f</b>) the UV spectra of CSS with various anions (samples on the photo (<b>e</b>)) after visible-light exposure for 1 h. The final electrolyte concentration corresponds to the system 16 (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B).</p>
Full article ">Figure 8 Cont.
<p>Photos of CSS with various anions: (<b>a</b>) immediately after adding electrolyte; (<b>b</b>) after 24 h in the dark; (<b>c</b>–<b>e</b>) after visible-light exposure of samples (photo (<b>b</b>)) for 20, 40, and 60 min, respectively. (<b>f</b>) the UV spectra of CSS with various anions (samples on the photo (<b>e</b>)) after visible-light exposure for 1 h. The final electrolyte concentration corresponds to the system 16 (see Table, <a href="#gels-10-00809-f001" class="html-fig">Figure 1</a>B).</p>
Full article ">
17 pages, 1791 KiB  
Article
Apple Defect Detection in Complex Environments
by Wei Shan and Yurong Yue
Electronics 2024, 13(23), 4844; https://doi.org/10.3390/electronics13234844 - 9 Dec 2024
Viewed by 416
Abstract
Aiming at the problem of high false detection and missed detection rate of apple surface defects in complex environments, a new apple surface defect detection network: space-to-depth convolution-Multi-scale Empty Attention-Context Guided Feature Pyramid Network-You Only Look Once version 8 nano (SMC-YOLOv8n) is designed. [...] Read more.
Aiming at the problem of high false detection and missed detection rate of apple surface defects in complex environments, a new apple surface defect detection network: space-to-depth convolution-Multi-scale Empty Attention-Context Guided Feature Pyramid Network-You Only Look Once version 8 nano (SMC-YOLOv8n) is designed. Firstly, space-to-depth convolution (SPD-Conv) is introduced before each Faster Implementation of CSP Bottleneck with 2 convolutions (C2f) in the backbone network as a preprocessing step to improve the quality of input data. Secondly, the Bottleneck in C2f is removed in the neck, and Multi-scale Empty Attention (MSDA) is introduced to enhance the feature extraction ability. Finally, the Context Guided Feature Pyramid Network (CGFPN) is used to replace the Concat method of the neck for feature fusion, thereby improving the expression ability of the features. Compared with the YOLOv8n baseline network, mean Average Precision (mAP) 50 increased by 2.7% and 1.1%, respectively, and mAP50-95 increased by 4.1% and 2.7%, respectively, on the visible light apple surface defect data set and public data set in the self-made complex environments.The experimental results show that SMC-YOLOv8n shows higher efficiency in apple defect detection, which lays a solid foundation for intelligent picking and grading of apples. Full article
Show Figures

Figure 1

Figure 1
<p>ADDCE research work classification.</p>
Full article ">Figure 2
<p>The overall architecture of YOLOv8.</p>
Full article ">Figure 3
<p>SPD-Conv structure diagram.</p>
Full article ">Figure 4
<p>C2f structure diagram.</p>
Full article ">Figure 5
<p>MSDA structure diagram.</p>
Full article ">Figure 6
<p>C2f-MSDA structure diagram.</p>
Full article ">Figure 7
<p>SE Attention module.</p>
Full article ">Figure 8
<p>SE Attention structure diagram.</p>
Full article ">Figure 9
<p>Context guide feature pyramid network architecture diagram.</p>
Full article ">Figure 10
<p>SMC-YOLOV8n.</p>
Full article ">Figure 11
<p>Examples of some data sets.</p>
Full article ">Figure 12
<p>Training curve and test curve.</p>
Full article ">Figure 13
<p>Test set confusion matrix.</p>
Full article ">Figure 14
<p>Part of the apple detection map.</p>
Full article ">
14 pages, 2037 KiB  
Article
Design of a Deep Learning-Based Metalens Color Router for RGB-NIR Sensing
by Hua Mu, Yu Zhang, Zhenyu Liang, Haoqi Gao, Haoli Xu, Bingwen Wang, Yangyang Wang and Xing Yang
Nanomaterials 2024, 14(23), 1973; https://doi.org/10.3390/nano14231973 - 8 Dec 2024
Viewed by 446
Abstract
Metalens can achieve arbitrary light modulation by controlling the amplitude, phase, and polarization of the incident waves and have been applied across various fields. This paper presents a color router designed based on metalens, capable of effectively separating spectra from visible light to [...] Read more.
Metalens can achieve arbitrary light modulation by controlling the amplitude, phase, and polarization of the incident waves and have been applied across various fields. This paper presents a color router designed based on metalens, capable of effectively separating spectra from visible light to near-infrared light. Traditional design methods for meta-lenses require extensive simulations, making them time-consuming. In this study, we propose a deep learning network capable of forward prediction across a broad wavelength range, combined with a particle swarm optimization algorithm to design metalens efficiently. The simulation results align closely with theoretical predictions. The designed color router can simultaneously meet the theoretical transmission phase of the target spectra, specifically for red, green, blue, and near-infrared light, and focus them into designated areas. Notably, the optical efficiency of this design reaches 40%, significantly surpassing the efficiency of traditional color filters. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) The schematic diagram of the RGB-NIR color router. (<b>b</b>) The schematic diagram of the meta-atom.</p>
Full article ">Figure 2
<p>Flowchart of the prediction process for neural network architecture.</p>
Full article ">Figure 3
<p>(<b>a</b>) Convergence curve of the prediction error for the deep learning network. (<b>b</b>) Statistical chart of the transmittance prediction error of the model on the test set. (<b>c</b>) Statistical chart of the phase prediction error of the model on the test set. (<b>d</b>) Comparison curve of transmittance and phase predictions against simulations for a randomly selected structure in the test set.</p>
Full article ">Figure 4
<p>Flowchart of the inverse design process combining deep learning network architecture with particle swarm optimization algorithm.</p>
Full article ">Figure 5
<p>Comparison of the theoretical phase distribution and the phase distribution of the meta-atoms obtained through inverse design, along with the corresponding transmittance curves.</p>
Full article ">Figure 6
<p>(<b>a</b>) The electric field distribution in the focusing cross-section and (<b>b</b>) the electric field distribution in the focal plane of the metalens.</p>
Full article ">Figure 7
<p>Comparison of the theoretical phase distribution and the phase distribution of the meta-atoms obtained through inverse design at center wavelengths of (<b>a</b>) 450 nm, (<b>b</b>) 540 nm, (<b>c</b>) 630 nm, and (<b>d</b>) 800 nm.</p>
Full article ">Figure 8
<p>The normalized electric field distributions at center wavelengths of (<b>a</b>) 450 nm, (<b>b</b>) 540 nm, (<b>c</b>) 630 nm, and (<b>d</b>) 800 nm.</p>
Full article ">Figure 9
<p>The optical efficiency of the metalens over a broadband spectral range.</p>
Full article ">
16 pages, 7435 KiB  
Article
Reconfigurable Intelligent Surface-Aided Security Enhancement for Vehicle-to-Vehicle Visible Light Communications
by Xiaoqiong Jing, Yating Wu, Fei Yu, Yuru Xu and Xiaoyong Wang
Photonics 2024, 11(12), 1151; https://doi.org/10.3390/photonics11121151 - 6 Dec 2024
Viewed by 441
Abstract
Vehicle-to-vehicle (V2V) visible light communication (VLC) systems are increasingly being deployed for real-time data exchange in intelligent transportation systems (ITS). However, these systems are highly vulnerable to eavesdropping, especially in scenarios such as road intersections where signals may be exposed to unauthorized receivers. [...] Read more.
Vehicle-to-vehicle (V2V) visible light communication (VLC) systems are increasingly being deployed for real-time data exchange in intelligent transportation systems (ITS). However, these systems are highly vulnerable to eavesdropping, especially in scenarios such as road intersections where signals may be exposed to unauthorized receivers. To address these security challenges, we propose a novel reconfigurable intelligent surface (RIS)-assisted security enhancement scheme for V2V VLC networks. The proposed scheme leverages RIS to improve the reception of legitimate signals at the destination vehicle while simultaneously introducing artificial noise (AN) to interfere with potential eavesdroppers. Optimization problems are formulated to maximize the SINR of the destination vehicle and simultaneously minimize the worst-case SINR of eavesdroppers. The simulation results demonstrate that the proposed scheme achieves a notable improvement in the system’s secrecy rate by 1.64 bit/s/Hz and enhances the overall security performance, offering a robust solution to the security challenges in V2V VLC systems. Full article
Show Figures

Figure 1

Figure 1
<p>Illustration of RIS-based V2V VLC system at road intersection.</p>
Full article ">Figure 2
<p>Abstraction system model of the intersection scenario.</p>
Full article ">Figure 3
<p>RIS-based artificial noise scheme.</p>
Full article ">Figure 4
<p>The SINR of PD of the destination vehicle versus reflection coefficient under different numbers of RIS units.</p>
Full article ">Figure 5
<p>The SINR at the destination vehicle versus source-to-RIS distance under different transmit power.</p>
Full article ">Figure 6
<p>The secrecy rate of the proposed scheme versus the position of eavesdroppers under different numbers of RIS units.</p>
Full article ">Figure 7
<p>The secrecy rate versus the position of eavesdroppers with and without RIS.</p>
Full article ">Figure 8
<p>The secrecy rate versus the position of eavesdroppers with and without AN.</p>
Full article ">
Back to TopTop