[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,996)

Search Parameters:
Keywords = visual recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 3652 KiB  
Article
Differential Signaling Pathways Identified in Aqueous Humor, Anterior Capsule, and Crystalline Lens of Age-Related, Diabetic, and Post-Vitrectomy Cataract
by Christina Karakosta, Martina Samiotaki, Anastasios Bisoukis, Konstantinos I. Bougioukas, George Panayotou, Dimitrios Papaconstantinou and Marilita M. Moschos
Proteomes 2025, 13(1), 7; https://doi.org/10.3390/proteomes13010007 (registering DOI) - 3 Feb 2025
Abstract
Background: The purpose of this study was to detect proteomic alterations and corresponding signaling pathways involved in the formation of age-related cataract (ARC), diabetic cataract (DC), and post-vitrectomy cataract (PVC). Methods: Three sample types, the aqueous humor (AH), the anterior capsule [...] Read more.
Background: The purpose of this study was to detect proteomic alterations and corresponding signaling pathways involved in the formation of age-related cataract (ARC), diabetic cataract (DC), and post-vitrectomy cataract (PVC). Methods: Three sample types, the aqueous humor (AH), the anterior capsule (AC), and the content of the phaco cassette, were collected during phacoemulsification surgery. The samples were obtained from 12 participants without diabetes mellitus (DM), 11 participants with DM, and 7 participants without DM, with a history of vitrectomy surgery in the past 12 months. The Sp3 protocol (Single-Pot, Solid-Phase, Sample-Preparation) was used for the sample preparation. The recognition and quantification of proteins were carried out with liquid chromatography online with tandem mass spectrometry. The DIA-NN software was applied for the identification and quantification of peptides/proteins. Statistical analysis and data visualization were conducted on Perseus software. Data are available via ProteomeXchange. Results: A very rich atlas of the lens and AH proteome has been generated. Glycosaminoglycan biosynthesis and the non-canonical Wnt receptor signaling pathway were differentially expressed in ARC compared to both the DC and PVC groups. In the PVC group, complement activation was differentially expressed in AH samples, while glutathione metabolism and oxidoreductase activity were differentially expressed in AC samples. Microfilament motor activity, microtubule cytoskeleton organization, and microtubule binding were differentially expressed in the DC and PVC groups in both AH and AC samples. Conclusions: The results of this study expand the existing knowledge on pathways involved in the pathophysiology of cataract, and suggest possible important druggable targets for slower progression or even prevention of cataract. Full article
(This article belongs to the Special Issue Clinical Proteomics: Fourth Edition)
Show Figures

Figure 1

Figure 1
<p>Sample collection: (<b>A</b>) The aqueous humor was collected from the anterior chamber using an insulin syringe. (<b>B</b>) The anterior capsule was collected using Utrata forceps after capsulerhexis was completed and was washed with Balanced Salt Solution (BSS) and stored immediately in a sterile box filled with BSS. (<b>C</b>) The content of the phaco cassette was collected at the end of the surgery, which contained the phacoemulsified particles of crystalline lens together with a small portion of the re-secreted aqueous humor in Balanced Salt Solution (BSS).</p>
Full article ">Figure 2
<p>(<b>a</b>) Venn diagram of the identified proteins in aqueous humor, anterior capsule, and phaco cassette sample types. (<b>b</b>) Venn diagram of the three sample types and the eye proteome of the Human Protein Atlas, demonstrating that 12 elements (0.3%) were common.</p>
Full article ">Figure 3
<p>(<b>a</b>) Aqueous humor samples. Dotplot of gene enrichment analysis for proteins significantly more abundant in ARC samples when compared with DC samples (top left). Dotplot of gene enrichment analysis for proteins significantly more abundant in DC samples when compared with ARC sample (top right). Dotplot of gene enrichment analysis for proteins significantly more abundant in ARC samples when compared with PVC samples (bottom left). Dotplot of gene enrichment analysis for proteins significantly more abundant in PVC samples when compared with ARC samples (bottom right). (<b>b</b>) Anterior capsule samples. Dotplot of gene enrichment analysis for proteins significantly more abundant in ARC samples when compared with DC samples (top left). Dotplot of gene enrichment analysis for proteins significantly more abundant in DC samples when compared with ARC sample (top right). Dotplot of gene enrichment analysis for proteins significantly more abundant in ARC samples when compared with PVC samples (bottom left). Dotplot of gene enrichment analysis for proteins significantly more abundant in PVC samples when compared with ARC samples (bottom right). (<b>c</b>) Phaco cassette content samples. Dotplot of gene enrichment analysis for proteins significantly more abundant in ARC samples when compared with DC samples (top left). Dotplot of gene enrichment analysis for proteins significantly more abundant in DC samples when compared with ARC sample (top right). Dotplot of gene enrichment analysis of proteins that presented with significant changes in degradation in ARC samples when compared with PVC samples (bottom left). Dotplot of gene enrichment analysis of proteins that presented with significant changes in degradation in PVC samples when compared with ARC samples (bottom right).</p>
Full article ">Figure 4
<p>(<b>a</b>) Volcano plot of AH sample results, showing DC-ARC (left) and ARC-PVC comparison (right). (<b>b</b>) Volcano plot of AC sample results, showing DC-ARC (left) and ARC-PVC comparison (right). (<b>c</b>) Volcano plot of phaco cassette content sample results, showing DC-ARC (left) and ARC-PVC comparison (right).</p>
Full article ">Figure 5
<p>(<b>a</b>) Aqueous humor sample heatmaps of proteins involved in complement activation (left) and in glycosaminoglycan biosynthetic process (right). (<b>b</b>) Anterior capsule samples heatmaps of myosins (left) and of proteins involved in oxidoreduction coenzyme metabolic process. (<b>c</b>) Phaco cassette content sample heatmaps of principal cytoskeletal proteins (left) and of proteins involved in proteasome complex.</p>
Full article ">Figure 6
<p>Raincloud plots of VTN protein involved in complement and coagulation cascades in AH samples (<b>a</b>) (ARC-DC <span class="html-italic">p</span> = 0.193353, ARC-PVC <span class="html-italic">p</span> &lt; 0.0001), ST3GAL5 protein involved in glycosphingolipid biosynthetic process in AC samples (<b>b</b>) (ARC-DC <span class="html-italic">p</span> &lt; 0.0421014, ARC-PVC <span class="html-italic">p</span> = 0.0218166), FRZB (<b>c</b>) (ARC-DC <span class="html-italic">p</span> = 0.00017, ARC-PVC <span class="html-italic">p</span> &lt; 0.0001), and SFRP5 proteins (<b>d</b>) involved in non-canonical Wnt signaling receptor pathway (ARC-DC <span class="html-italic">p</span> &lt; 0.0001, ARC-PVC <span class="html-italic">p</span> &lt; 0.0001) in phaco cassette content samples.</p>
Full article ">
15 pages, 2947 KiB  
Article
Neem and Gliricidia Plant Leaf Extracts Improve Yield and Quality of Leaf Mustard by Managing Insect Pests’ Abundance Without Harming Beneficial Insects and Some Sensory Attributes
by Rowland Maganizo Kamanga, Salifu Bhikha, Felix Dalitso Kamala, Vincent Mgoli Mwale, Yolice Tembo and Patrick Alois Ndakidemi
Insects 2025, 16(2), 156; https://doi.org/10.3390/insects16020156 - 3 Feb 2025
Viewed by 6
Abstract
Production and consumption of vegetable crops has seen a sharp increase in the recent past owing to an increasing recognition of their nutraceutical benefits. In tandem, there has been unwarranted application of agrochemicals such as insecticides to enhance productivity and vegetable quality, at [...] Read more.
Production and consumption of vegetable crops has seen a sharp increase in the recent past owing to an increasing recognition of their nutraceutical benefits. In tandem, there has been unwarranted application of agrochemicals such as insecticides to enhance productivity and vegetable quality, at the cost of human health, and fundamental environmental and ecosystem functions and services. This study was conducted to evaluate the efficacy of neem and gliricidia botanical extracts in managing harmful insect pest populations in leaf mustard. Our results report that neem and gliricidia plant extracts enhance the yield and quality of leaf mustard by reducing the prevalence and feeding activity of harmful insect pests in a manner similar to synthetic insecticides. Some of the key insect pests reduced were Lipaphis erysimi, Pieris oleracea, Phyllotreta Cruciferae, Melanoplus sanguinipes, and Murgantia histrionica. However, compared to synthetic insecticides, neem and gliricidia plant extracts were able to preserve beneficial insects such as the Coccinellidae spp., Trichogramma minutum, Araneae spp., Lepidoptera spp., and Blattodea spp. Furthermore, plant extracts did not significantly alter sensory attributes, especially taste and odor, whereas the visual appearance of leaf mustard was greater in plants sprayed with neem and synthetic insecticides. Physiologically, plant extracts were also able to significantly lower leaf membrane damage as shown through the electrolyte leakage assay. Therefore, these plant extracts represent promising pesticidal plant materials and botanically active substances that can be leveraged to develop environmentally friendly commercial pest management products. Full article
(This article belongs to the Section Insect Pest and Vector Management)
Show Figures

Figure 1

Figure 1
<p>Leaf mustard plots with 4 blocks in a randomized complete block design.</p>
Full article ">Figure 2
<p>Average number of insect pests in treated and untreated plots per scouting week. The values represent weekly means from 4 plots in the 4 blocks.</p>
Full article ">Figure 3
<p>Effect of botanical extracts and synthetic insecticide on biological (<b>A</b>) and economic yield (<b>B</b>) of mustard leaves. The values represent means from 10 biological replicates. Different letters indicate significant differences using the Tukey test at a 0.05 level of significance, whereas similar letters indicate no significant differences at a 0.05 level of significance.</p>
Full article ">Figure 4
<p>Effect of botanical extracts and synthetic insecticide on economic leaf area of mustard leaves. The values represent means from 10 biological replicates. Different letters indicate significant differences using the Tukey test at a 0.05 level of significance, whereas similar letters indicate no significant differences at a 0.05 level of significance.</p>
Full article ">Figure 5
<p>Effect of plant botanical extracts and synthetic insecticides on beneficial insect abundance. The values represent means from 4 plots in the 4 blocks.</p>
Full article ">Figure 6
<p>Effect of plant botanical extracts and synthetic insecticides on membrane integrity. The values represent means from 10 biological replicates. Different letters indicate significant differences using the Tukey test at a 0.05 level of significance, whereas similar letters indicate no significant differences at a 0.05 level of significance.</p>
Full article ">
26 pages, 3470 KiB  
Article
A Novel Face Frontalization Method by Seamlessly Integrating Landmark Detection and Decision Forest into Generative Adversarial Network (GAN)
by Mahmood H. B. Alhlffee and Yea-Shuan Huang
Mathematics 2025, 13(3), 499; https://doi.org/10.3390/math13030499 (registering DOI) - 2 Feb 2025
Viewed by 240
Abstract
In real-world scenarios, posture variation and low-quality image resolution are two well-known factors that compromise the accuracy and reliability of face recognition system. These challenges can be overcome using various methods, including Generative Adversarial Networks (GANs). Despite this, concerns over the accuracy and [...] Read more.
In real-world scenarios, posture variation and low-quality image resolution are two well-known factors that compromise the accuracy and reliability of face recognition system. These challenges can be overcome using various methods, including Generative Adversarial Networks (GANs). Despite this, concerns over the accuracy and reliability of GAN methods are increasing as the facial recognition market expands rapidly. The existing framework such as Two-Pathway GAN (TP-GAN) method has demonstrated that it is superior to numerous GAN methods that provide better face-texture details due to its unique deep neural network structure that allows it to perceive local details and global structure in a supervised manner. TP-GAN overcomes some of the obstacle associated with face frontalization tasks through the use of landmark detection and synthesis functions, but it remains challenging to achieve the desired performance across a wide range of datasets. To address the inherent limitations of TP-GAN, we propose a novel face frontalization method (NFF) combining landmark detection, decision forests, and data augmentation. NFF provides 2D landmark detection to integrate global structure with local details of the generator model so that more accurate facial feature representations and robust feature extractions can be achieved. NFF enhances the stability of the discriminator model over time by integrating decision forest capabilities into the TP-GAN discriminator core architecture that allows us to perform a wide range of facial pose tasks. Moreover, NFF uses data augmentation techniques to maximize training data by generating completely new synthetic data from existing data. Our evaluations are based on the Multi-PIE, FEI, and CAS-PEAL datasets. NFF results indicate that TP-GAN performance can be significantly enhanced by resolving the challenges described above, leading to high quality visualizations and rank-1 face identification. Full article
(This article belongs to the Special Issue Advanced Machine Vision with Mathematics)
21 pages, 5270 KiB  
Article
Three-Dimensional Object Recognition Using Orthogonal Polynomials: An Embedded Kernel Approach
by Aqeel Abdulazeez Mohammed, Ahlam Hanoon Al-sudani, Alaa M. Abdul-Hadi, Almuntadher Alwhelat, Basheera M. Mahmmod, Sadiq H. Abdulhussain, Muntadher Alsabah and Abir Hussain
Algorithms 2025, 18(2), 78; https://doi.org/10.3390/a18020078 (registering DOI) - 1 Feb 2025
Viewed by 410
Abstract
Computer vision seeks to mimic the human visual system and plays an essential role in artificial intelligence. It is based on different signal reprocessing techniques; therefore, developing efficient techniques becomes essential to achieving fast and reliable processing. Various signal preprocessing operations have been [...] Read more.
Computer vision seeks to mimic the human visual system and plays an essential role in artificial intelligence. It is based on different signal reprocessing techniques; therefore, developing efficient techniques becomes essential to achieving fast and reliable processing. Various signal preprocessing operations have been used for computer vision, including smoothing techniques, signal analyzing, resizing, sharpening, and enhancement, to reduce reluctant falsifications, segmentation, and image feature improvement. For example, to reduce the noise in a disturbed signal, smoothing kernels can be effectively used. This is achievedby convolving the distributed signal with smoothing kernels. In addition, orthogonal moments (OMs) are a crucial technique in signal preprocessing, serving as key descriptors for signal analysis and recognition. OMs are obtained by the projection of orthogonal polynomials (OPs) onto the signal domain. However, when dealing with 3D signals, the traditional approach of convolving kernels with the signal and computing OMs beforehand significantly increases the computational cost of computer vision algorithms. To address this issue, this paper develops a novel mathematical model to embed the kernel directly into the OPs functions, seamlessly integrating these two processes into a more efficient and accurate approach. The proposed model allows the computation of OMs for smoothed versions of 3D signals directly, thereby reducing computational overhead. Extensive experiments conducted on 3D objects demonstrate that the proposed method outperforms traditional approaches across various metrics. The average recognition accuracy improves to 83.85% when the polynomial order is increased to 10. Experimental results show that the proposed method exhibits higher accuracy and lower computational costs compared to the benchmark methods in various conditions for a wide range of parameter values. Full article
(This article belongs to the Section Algorithms for Multidisciplinary Applications)
Show Figures

Figure 1

Figure 1
<p>Samples showing the effect of kernels on ant 3D objects using conventional convolution and OP-based convolution.</p>
Full article ">Figure 2
<p>Samples showing the effect of kernels on plane 3D objects using conventional convolution and OP-based convolution.</p>
Full article ">Figure 3
<p>Flow diagram of the proposed embedded kernel.</p>
Full article ">Figure 4
<p>Flow diagram of object recognition using the proposed technique.</p>
Full article ">Figure 5
<p>Samples of 3D objects extracted from McGill dataset.</p>
Full article ">
21 pages, 16401 KiB  
Article
High-Resolution Mapping of Maize in Mountainous Terrain Using Machine Learning and Multi-Source Remote Sensing Data
by Luying Liu, Jingyi Yang, Fang Yin and Linsen He
Land 2025, 14(2), 299; https://doi.org/10.3390/land14020299 - 31 Jan 2025
Viewed by 309
Abstract
In recent years, machine learning methods have garnered significant attention in the field of crop recognition, playing a crucial role in obtaining spatial distribution information and understanding dynamic changes in planting areas. However, research in smaller plots within mountainous regions remains relatively limited. [...] Read more.
In recent years, machine learning methods have garnered significant attention in the field of crop recognition, playing a crucial role in obtaining spatial distribution information and understanding dynamic changes in planting areas. However, research in smaller plots within mountainous regions remains relatively limited. This study focuses on Shangzhou District in Shangluo City, Shaanxi Province, utilizing a dataset of high-resolution remote sensing images (GF-1, ZY1-02D, ZY-3) collected over seven months in 2021 to calculate the normalized difference vegetation index (NDVI) and construct a time series. By integrating field survey results with time series images and Google Earth for visual interpretation, the NDVI time series curve for maize was analyzed. The Random Forest (RF) classification algorithm was employed for maize recognition, and comparative analyses of classification accuracy were conducted using Support Vector Machine (SVM), Gaussian Naive Bayes (GNB), and Artificial Neural Network (ANN). The results demonstrate that the random forest algorithm achieved the highest accuracy, with an overall accuracy of 94.88% and a Kappa coefficient of 0.94, both surpassing those of the other classification methods and yielding satisfactory overall results. This study confirms the feasibility of using time series high-resolution remote sensing images for precise crop extraction in the southern mountainous regions of China, providing valuable scientific support for optimizing land resource use and enhancing agricultural productivity. Full article
(This article belongs to the Section Land – Observation and Monitoring)
Show Figures

Figure 1

Figure 1
<p>Location of the study area and land-use features: (<b>a</b>) administrative divisions of China; (<b>b</b>) administrative divisions of Shaanxi Province; (<b>c</b>) main natural rivers in Shangzhou District; (<b>d</b>) elevation map and distribution of maize planting points; (<b>e</b>) land-use status map of the study area in 2023.</p>
Full article ">Figure 2
<p>Maize phenological period and image acquisition dates.</p>
Full article ">Figure 3
<p>NDVI time series curves and the arithmetic mean spectral reflectance for different land cover types: (<b>a</b>) NDVI time series for major land cover types; (<b>b</b>) average spectral reflectance in the red band for major land cover types; (<b>c</b>) average spectral reflectance in the near-infrared band for major land cover types; the solid line is used to represent periods with continuous data, while the dashed line is used to connect periods with missing data.</p>
Full article ">Figure 4
<p>Annual variation trends of temperature, precipitation, and evapotranspiration in the study area from 2019 to 2023: (<b>a</b>) annual variation trend of temperature; (<b>b</b>) annual variation trend of precipitation; (<b>c</b>) annual variation trend of evapotranspiration.</p>
Full article ">Figure 5
<p>Distribution of accuracy for each machine learning method.</p>
Full article ">Figure 6
<p>Classification results of various machine learning methods: e, f, g, and h represent the four typical regions of the study area; (<b>a</b>) Gaussian Naive Bayes classification results; (<b>b</b>) Artificial Neural Network classification results; (<b>c</b>) Support Vector Machine classification results; (<b>d</b>) Random Forest classification results.</p>
Full article ">Figure 7
<p>Typical area classification results of various machine learning methods: (<b>a</b>–<b>d</b>) represent the classification results using Gaussian Naive Bayes for four typical regions of the study area; (<b>a1</b>,<b>b1</b>,<b>c1</b>,<b>d1</b>) represent the classification results using Artificial Neural Network; (<b>a2</b>,<b>b2</b>,<b>c2</b>,<b>d2</b>) represent the classification results using Support Vector Machine; (<b>a3</b>,<b>b3</b>,<b>c3</b>,<b>d3</b>) represent the classification results using Random Forest.</p>
Full article ">Figure 8
<p>Distribution of main crop maize in typical areas: (<b>a</b>–<b>d</b>) are the classification results of typical regions; (<b>a1</b>,<b>b1</b>,<b>c1</b>,<b>d1</b>) are the NDVI curves for maize in these regions; (<b>a2</b>,<b>b2</b>,<b>c2</b>,<b>d2</b>) are the images of these regions; (<b>a3</b>,<b>b3</b>,<b>c3</b>,<b>d3</b>) are field photographs of these regions.</p>
Full article ">
14 pages, 2761 KiB  
Article
Validation of Novel Image Processing Method for Objective Quantification of Intra-Articular Bleeding During Arthroscopic Procedures
by Olgar Birsel, Umut Zengin, Ilker Eren, Ali Ersen, Beren Semiz and Mehmet Demirhan
J. Imaging 2025, 11(2), 40; https://doi.org/10.3390/jimaging11020040 - 31 Jan 2025
Viewed by 357
Abstract
Visual clarity is crucial for shoulder arthroscopy, directly influencing surgical precision and outcomes. Despite advances in imaging technology, intraoperative bleeding remains a significant obstacle to optimal visibility, with subjective evaluation methods lacking consistency and standardization. This study proposes a novel image processing system [...] Read more.
Visual clarity is crucial for shoulder arthroscopy, directly influencing surgical precision and outcomes. Despite advances in imaging technology, intraoperative bleeding remains a significant obstacle to optimal visibility, with subjective evaluation methods lacking consistency and standardization. This study proposes a novel image processing system to objectively quantify bleeding and assess surgical effectiveness. The system uses color recognition algorithms to calculate a bleeding score based on pixel ratios by incorporating multiple color spaces to enhance accuracy and minimize errors. Moreover, 200 three-second video clips from prior arthroscopic rotator cuff repairs were evaluated by three senior surgeons trained on the system’s color metrics and scoring process. Assessments were repeated two weeks later to test intraobserver reliability. The system’s scores were compared to the average score given by the surgeons. The average surgeon-assigned score was 5.10 (range: 1–9.66), while the system scored videos from 1 to 9.46, with an average of 5.08. The mean absolute error between system and surgeon scores was 0.56, with a standard deviation of 0.50, achieving agreement ranging from [0.96,0.98] with 96.7% confidence (ICC = 0.967). This system provides a standardized method to evaluate intraoperative bleeding, enabling the precise detection of blood variations and supporting advanced technologies like autonomous arthropumps to enhance arthroscopy and surgical outcomes. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Figure 1
<p>An image processing system was designed to count pixels meeting threshold criteria and calculate their ratio to the total, assigning a score to each image. It evaluates three random frames per second (from 24 fps) and compares each second’s average score to the previous three seconds’ scores to suppress the effect of spontaneous red tissue artifacts.</p>
Full article ">Figure 2
<p>Grading scale from 1 to 9 according to the scores given by the fourth surgeon. This tutorial is created for training the raters, indicating the gradually increasing amount of bleeding into the surgical area. Note that the redness, saturation, and blur are increasing from 1 to 9.</p>
Full article ">Figure 3
<p>Comparison of scores 1 (<b>A</b>) and 10 (<b>B</b>) based on the scores of the fourth surgeon. These scores were verified by the system as the lowest (1) and the highest (9.46) scores of the whole series.</p>
Full article ">Figure 4
<p>Arthroscopic images extracted from video clips scored as 3.17 (<b>A</b>) and 4.08 (<b>B</b>) by the system. In image (<b>A</b>), the background inflammation was not classified as bleeding, resulting in a low score of 3.</p>
Full article ">Figure 5
<p>Arthroscopic images extracted from video clips scored as 4.21 (<b>A</b>) and 5.3 (<b>B</b>) by the system. Both images were uniformly scored as 4 by all three surgeons, demonstrating the system’s ability to distinguish between similar images.</p>
Full article ">Figure 6
<p>(<b>a</b>) Individual surgeons’ scores against the system-generated scores. (<b>b</b>) Individual scores of each surgeon for 200 video clips. (<b>c</b>) Absolute errors for 200 video clips.</p>
Full article ">Figure 7
<p>The system’s analysis report of a 62 min procedure. The total bleeding score is a novel parameter that indicates the average bleeding score of this arthroscopic rotator cuff repair procedure and calculated as 1.90/second. Ersin et al. previously reported that the occurrence of bleeding varies along different stages of an arthroscopic shoulder procedure [<a href="#B11-jimaging-11-00040" class="html-bibr">11</a>]. Indeed, the total bleeding score is highest in subacromial decompression (2.28) and biceps tenodesis (2.26) phases of the surgery, while it is lowest in the final phase where the rotator cuff tendons are repaired (1.59).</p>
Full article ">
13 pages, 1070 KiB  
Review
Primary Congenital and Childhood Glaucoma—A Complex Clinical Picture and Surgical Management
by Valeria Coviltir, Maria Cristina Marinescu, Bianca Maria Urse and Miruna Gabriela Burcel
Diagnostics 2025, 15(3), 308; https://doi.org/10.3390/diagnostics15030308 - 28 Jan 2025
Viewed by 341
Abstract
Childhood glaucoma encompasses a group of rare but severe ocular disorders characterized by increased intraocular pressure (IOP), posing significant risks to vision and quality of life. Primary congenital glaucoma has a prevalence of one in 10,000–68,000 people in Western countries. More worryingly, it [...] Read more.
Childhood glaucoma encompasses a group of rare but severe ocular disorders characterized by increased intraocular pressure (IOP), posing significant risks to vision and quality of life. Primary congenital glaucoma has a prevalence of one in 10,000–68,000 people in Western countries. More worryingly, it is responsible for 5–18% of all childhood blindness cases. According to the Childhood Glaucoma Research Network (CGRN), this spectrum of disease is classified into primary glaucoma (primary congenital glaucoma and juvenile open-angle glaucoma) and secondary glaucomas (associated with non-acquired ocular anomalies, non-acquired systemic disease, acquired conditions, and glaucoma after cataract surgery). They present very specific ocular characteristics, such as buphthalmos or progressive myopic shift, corneal modifications such as Haab striae, corneal edema or increased corneal diameter, and also glaucoma findings including high intraocular pressure, specific visual fields abnormalities, and optic nerve damage such as increased cup-disc ratio, cup-disc ratio asymmetry of at least 0.2 and focal rim thinning. Surgical intervention remains the cornerstone of treatment, and initial surgical options include angle surgeries such as goniotomy and trabeculotomy, aimed at improving aqueous outflow. For refractory cases, trabeculectomy and glaucoma drainage devices (GDDs) serve as second-line therapies. Advanced cases may require cyclodestructive procedures, including transscleral cyclophotocoagulation, reserved for eyes with limited visual potential. All in all, with appropriate management, the prognosis of PCG may be quite favorable: stationary disease has been reported in 90.3% of cases after one year, with a median visual acuity in the better eye of 20/30. Immediate recognition of the specific signs and symptoms by caregivers, primary care providers, and ophthalmologists, followed by prompt diagnosis, comprehensive surgical planning, and involving the caregivers in the follow-up schedule remain critical for optimizing outcomes in childhood glaucoma management. Full article
(This article belongs to the Special Issue Diagnosis, Treatment and Management of Eye Diseases, Second Edition)
Show Figures

Figure 1

Figure 1
<p>Buphthalmic aspect of an eye in a patient with secondary childhood glaucoma associated with systemic diseases or syndromes, in this case, Sturge–Weber Syndrome (personal archive of author V.C.).</p>
Full article ">Figure 2
<p>Slit-lamp photograph of Haab striae (Personal archive of author V.C.).</p>
Full article ">Figure 3
<p>Trabeculotome insertion into the anterior chamber and cutting the walls of the Schlemm’s canal (personal archive of author V.C.).</p>
Full article ">Figure 4
<p>Final aspect of GDD implantation, with a tube placed in the anterior chamber (personal archive of author V.C.).</p>
Full article ">
42 pages, 1207 KiB  
Review
Advancements in Ocular Neuro-Prosthetics: Bridging Neuroscience and Information and Communication Technology for Vision Restoration
by Daniele Giansanti
Biology 2025, 14(2), 134; https://doi.org/10.3390/biology14020134 - 28 Jan 2025
Viewed by 342
Abstract
Background: Neuroprosthetics for vision restoration have advanced significantly, incorporating technologies like retinal implants, cortical implants, and non-invasive stimulation methods. These advancements hold the potential to tackle major challenges in visual prosthetics, such as enhancing functionality, improving biocompatibility, and enabling real-time object recognition. Aim: [...] Read more.
Background: Neuroprosthetics for vision restoration have advanced significantly, incorporating technologies like retinal implants, cortical implants, and non-invasive stimulation methods. These advancements hold the potential to tackle major challenges in visual prosthetics, such as enhancing functionality, improving biocompatibility, and enabling real-time object recognition. Aim: The aim of this review overview is to provide a comprehensive analysis of the latest advancements in ocular neuroprostheses. Methods: A narrative review was conducted, focusing on the latest developments in visual neuroprosthetics. Comprehensive searches were carried out on Google Scholar, PubMed, and Scopus using specific keywords. A specific narrative checklist was applied, alongside a tailored quality assessment methodology, to evaluate the quality of the studies included. A total of sixteen relevant studies from the past three years were included in the review. Results and discussion: The integration of artificial retinas, cortical implants, high technology-enabled prosthetics, gene therapies, nanotechnology, and bioprinting has shown significant promise in enhancing the quality and functionality of vision restoration systems, offering potential to address complex visual impairments and improve independence and mobility for individuals with blindness. These innovations appear to have the potential to transform healthcare systems in the future by enabling more efficient and personalized therapies and prosthetic devices. However, challenges such as energy efficiency, scalability, and the neural dynamics of vision restoration persist, requiring continued interdisciplinary collaboration to refine these technologies, overcome ethical and regulatory hurdles, and ensure their effectiveness in real-world applications. Conclusions: While visual neuroprosthetics have made remarkable progress, addressing challenges related to energy consumption and regulatory and ethical concerns will be crucial for ensuring that neuroprosthetic devices can effectively meet the needs of individuals with visual impairments. Full article
(This article belongs to the Special Issue The Convergence of Neuroscience and ICT: From Data to Insights)
27 pages, 12866 KiB  
Article
Multimodal Augmented Reality System for Real-Time Roof Type Recognition and Visualization on Mobile Devices
by Bartosz Kubicki, Artur Janowski and Adam Inglot
Appl. Sci. 2025, 15(3), 1330; https://doi.org/10.3390/app15031330 - 27 Jan 2025
Viewed by 495
Abstract
The utilization of augmented reality (AR) is becoming increasingly prevalent in the integration of virtual reality (VR) elements into the tangible reality of the physical world. It facilitates a more straightforward comprehension of the interconnections, interdependencies, and spatial context of data. Furthermore, the [...] Read more.
The utilization of augmented reality (AR) is becoming increasingly prevalent in the integration of virtual reality (VR) elements into the tangible reality of the physical world. It facilitates a more straightforward comprehension of the interconnections, interdependencies, and spatial context of data. Furthermore, the presentation of analyses and the combination of spatial data with annotated data are facilitated. This is particularly evident in the context of mobile applications, where the combination of real-world and virtual imagery facilitates enhances visualization. This paper presents a proposal for the development of a multimodal system that is capable of identifying roof types in real time and visualizing them in AR on mobile devices. The current approach to roof identification is based on data made available by public administrations in an open-source format, including orthophotos and building contours. Existing computer processing technologies have been employed to generate objects representing the shapes of building masses, and in particular, the shape of roofs, in three-dimensional (3D) space. The system integrates real-time data obtained from multiple sources and is based on a mobile application that enables the precise positioning and detection of the recipient’s viewing direction (pose estimation) in real time. The data were integrated and processed in a Docker container system, which ensured the scalability and security of the solution. The multimodality of the system is designed to enhance the user’s perception of the space and facilitate a more nuanced interpretation of its intricacies. In its present iteration, the system facilitates the extraction and classification/generalization of two categories of roof types (gable and other) from aerial imagery through the utilization of deep learning methodologies. The outcomes achieved suggest considerable promise for the advancement and deployment of the system in domains pertaining to architecture, urban planning, and civil engineering. Full article
(This article belongs to the Special Issue Applications of Data Science and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Study area in Głusk Community, Lublin Poviat, Lubelskie Voivodeship, Poland (source: author’s own elaboration).</p>
Full article ">Figure 2
<p>Repository of certain roof type classes: axonometric view (<b>a</b>) and top view (<b>b</b>) with geometrical parameters involving examples from orthophoto (<b>c</b>) (source: author’s own elaboration).</p>
Full article ">Figure 3
<p>Scheme of assumptions for creating data for the learning set of YOLO (source: author’s own elaboration).</p>
Full article ">Figure 4
<p>Diagram with the percentage of the roof type class in each individual area (source: author’s own elaboration).</p>
Full article ">Figure 5
<p>Normal (Gaussian) distribution graph of the percentage of roof type classes. (source: author’s own elaboration).</p>
Full article ">Figure 6
<p>Statistics of training data from YOLO: (<b>a</b>) Centroid distribution of training objects. (<b>b</b>) Bounding box dimensions of training objects (source: author’s own elaboration).</p>
Full article ">Figure 7
<p>Training and validation loss and precision trends across 350 epochs. Early stopping at epoch 327 ensures optimal precision and prevents overfitting (source: author’s own elaboration).</p>
Full article ">Figure 8
<p>Examples of misclassification of roof types: (<b>a</b>) incorrect class, (<b>b</b>) roof is not located on the building, (<b>c</b>) incorrect object other than the roof, (<b>d</b>) unclassified (source: author’s own elaboration).</p>
Full article ">Figure 9
<p>Visualization of representative classes of building blocks: buildings with gable roofs (<b>a</b>) and others (<b>b</b>) (source: author’s own elaboration).</p>
Full article ">Figure 10
<p>Multimodal system component data flow architecture (source: author’s own elaboration).</p>
Full article ">Figure 11
<p>The options for presenting classes of building roof shapes in 2D and 3D (source: author’s own elaboration).</p>
Full article ">Figure A1
<p>Determining the optimal F1-score value for confidence threshold optimization, considering overfitting (source: author’s own elaboration).</p>
Full article ">
19 pages, 1177 KiB  
Article
Characterization of the Municipal Plastic and Multilayer Packaging Waste in Three Cities of the Baltic States
by Pavlo Lyshtva, Artūras Torkelis, Yaroslav Kobets, Estefania Carpio-Vallejo, Andrea Dobri, Jelena Barbir, Viktoria Voronova, Gintaras Denafas and Linas Kliucininkas
Sustainability 2025, 17(3), 986; https://doi.org/10.3390/su17030986 - 25 Jan 2025
Viewed by 521
Abstract
The composition of plastic and multilayer packaging waste was assessed in the mixed municipal solid waste (MSW) streams of the Kaunas (Lithuania), Daugavpils (Latvia) and Tallinn (Estonia) municipalities. For the analysis of samples in the mixed MSW streams, the authors used manual sorting [...] Read more.
The composition of plastic and multilayer packaging waste was assessed in the mixed municipal solid waste (MSW) streams of the Kaunas (Lithuania), Daugavpils (Latvia) and Tallinn (Estonia) municipalities. For the analysis of samples in the mixed MSW streams, the authors used manual sorting and a visual recognition method. Composition analysis of plastic and multilayer packaging waste from separately collected waste of multi-family and single-family households was performed in the Kaunas and Tallinn municipalities. For the analysis of samples in the separately collected waste streams, the research group combined manual sorting and near-infrared (NIR) spectroscopy methods. The findings reveal that the percentage distribution of plastic and multilayer packaging waste within the municipal solid waste (MSW) stream is relatively consistent across the municipalities of Kaunas, Daugavpils and Tallinn, comprising 40.16%, 36.83% and 35.09%, respectively. However, a notable variation emerges when examining separately collected plastic and multilayer packaging waste streams. In this category, the proportion of plastic and multilayer packaging within the total separately collected packaging waste stream ranges from 62.05% to 74.7% for multi-family residential buildings and from 44.66% to 56.89% for single-family residential buildings. The authors provided further insights for the enhanced recycling potential of different plastic materials through improved sorting. Full article
Show Figures

Figure 1

Figure 1
<p>Population and size of administrative territories in Kaunas, Daugavpils and Tallinn municipalities.</p>
Full article ">Figure 2
<p>Distribution of plastic and multilayer packaging waste by the type of polymer in the samples collected from multi-family households in Kaunas (mass percentages).</p>
Full article ">Figure 3
<p>Distribution of plastic and multilayer packaging waste by the type of polymer (mass percentages) in the samples collected from single-family households in Kaunas.</p>
Full article ">Figure 4
<p>Distribution of plastic and multilayer packaging waste by the type of polymer (mass percentages) in the samples collected from multi-family households in Tallinn.</p>
Full article ">Figure 5
<p>Distribution of plastic and multilayer packaging waste by type of polymer (mass percentages) in the samples collected from single-family households in Tallinn.</p>
Full article ">
21 pages, 1368 KiB  
Article
Radar Signal Processing and Its Impact on Deep Learning-Driven Human Activity Recognition
by Fahad Ayaz, Basim Alhumaily, Sajjad Hussain, Muhamamd Ali Imran, Kamran Arshad, Khaled Assaleh and Ahmed Zoha
Sensors 2025, 25(3), 724; https://doi.org/10.3390/s25030724 - 25 Jan 2025
Viewed by 478
Abstract
Human activity recognition (HAR) using radar technology is becoming increasingly valuable for applications in areas such as smart security systems, healthcare monitoring, and interactive computing. This study investigates the integration of convolutional neural networks (CNNs) with conventional radar signal processing methods to improve [...] Read more.
Human activity recognition (HAR) using radar technology is becoming increasingly valuable for applications in areas such as smart security systems, healthcare monitoring, and interactive computing. This study investigates the integration of convolutional neural networks (CNNs) with conventional radar signal processing methods to improve the accuracy and efficiency of HAR. Three distinct, two-dimensional radar processing techniques, specifically range-fast Fourier transform (FFT)-based time-range maps, time-Doppler-based short-time Fourier transform (STFT) maps, and smoothed pseudo-Wigner–Ville distribution (SPWVD) maps, are evaluated in combination with four state-of-the-art CNN architectures: VGG-16, VGG-19, ResNet-50, and MobileNetV2. This study positions radar-generated maps as a form of visual data, bridging radar signal processing and image representation domains while ensuring privacy in sensitive applications. In total, twelve CNN and preprocessing configurations are analyzed, focusing on the trade-offs between preprocessing complexity and recognition accuracy, all of which are essential for real-time applications. Among these results, MobileNetV2, combined with STFT preprocessing, showed an ideal balance, achieving high computational efficiency and an accuracy rate of 96.30%, with a spectrogram generation time of 220 ms and an inference time of 2.57 ms per sample. The comprehensive evaluation underscores the importance of interpretable visual features for resource-constrained environments, expanding the applicability of radar-based HAR systems to domains such as augmented reality, autonomous systems, and edge computing. Full article
(This article belongs to the Special Issue Non-Intrusive Sensors for Human Activity Detection and Recognition)
Show Figures

Figure 1

Figure 1
<p>Detailed flow diagram illustrating the structure of the paper sections and content.</p>
Full article ">Figure 2
<p>Radar-based HAR system depicting the workflow from data acquisition to radar maps’ generation, along with state-of-the-art neural networks.</p>
Full article ">Figure 3
<p>Two-dimensional images of six activities resulting from TR, STFT, and SPWVD techniques.</p>
Full article ">Figure 4
<p>Confusion matrices of best-performing pairs. (<b>a</b>) shows pair M1, (<b>b</b>) shows pair M7, and (<b>c</b>) shows pair M10.</p>
Full article ">Figure 5
<p>Generalization capability of the proposed HAR system.</p>
Full article ">Figure 6
<p>Performance and computational analysis comparison across radar domains as input to MobileNetV2.</p>
Full article ">
17 pages, 4918 KiB  
Article
CDKD-w+: A Keyframe Recognition Method for Coronary Digital Subtraction Angiography Video Sequence Based on w+ Space Encoding
by Yong Zhu, Haoyu Li, Shuai Xiao, Wei Yu, Hongyu Shang, Lin Wang, Yang Liu, Yin Wang and Jiachen Yang
Sensors 2025, 25(3), 710; https://doi.org/10.3390/s25030710 - 24 Jan 2025
Viewed by 361
Abstract
Currently, various deep learning methods can assist in medical diagnosis. Coronary Digital Subtraction Angiography (DSA) is a medical imaging technology used in cardiac interventional procedures. By employing X-ray sensors to visualize the coronary arteries, it generates two-dimensional images from any angle. However, due [...] Read more.
Currently, various deep learning methods can assist in medical diagnosis. Coronary Digital Subtraction Angiography (DSA) is a medical imaging technology used in cardiac interventional procedures. By employing X-ray sensors to visualize the coronary arteries, it generates two-dimensional images from any angle. However, due to the complexity of the coronary structures, the 2D images may sometimes lack sufficient information, necessitating the construction of a 3D model. Camera-level 3D modeling can be realized based on deep learning. Nevertheless, the beating of the heart results in varying degrees of arterial vasoconstriction and vasodilation, leading to substantial discrepancies between DSA sequences, which introduce errors in 3D modeling of the coronary arteries, resulting in the inability of the 3D model to reflect the coronary arteries. We propose a coronary DSA video sequence keyframe recognition method, CDKD-w+, based on w+ space encoding. The method utilizes a pSp encoder to encode the coronary DSA images, converting them into latent codes in the w+ space. Differential analysis of inter-frame latent codes is employed for heartbeat keyframe localization, aiding in coronary 3D modeling. Experimental results on a self-constructed coronary DSA heartbeat keyframe recognition dataset demonstrate an accuracy of 97%, outperforming traditional metrics such as L1, SSIM, and PSNR. Full article
(This article belongs to the Special Issue Image Processing in Sensors and Communication Systems)
Show Figures

Figure 1

Figure 1
<p>A DSA sequence containing a complete cardiac cycle.</p>
Full article ">Figure 2
<p>An intuitive display of the vasodilation and vasoconstriction frames in <a href="#sensors-25-00710-f001" class="html-fig">Figure 1</a>, with marked positions indicating significant differences.</p>
Full article ">Figure 3
<p>Overall framework of CDKD-w+. Different images are extracted features by the pSp encoder, then the cosine similarity between feature vectors is calculated, forming a confusion matrix based on cosine similarity. Keyframes are located by searching for local minimum in the confusion matrix.</p>
Full article ">Figure 4
<p>pSp schematic diagram.</p>
Full article ">Figure 5
<p>Illustration of obtaining a local minimum value.</p>
Full article ">Figure 6
<p>Illustration of non-minimum suppression. Assuming the 0th frame is a keyframe, the 1st frame within a distance of 2 frames will be suppressed, while keyframes with a distance greater than 2 frames will be retained.</p>
Full article ">Figure 7
<p>Dataset diagram. The dataset consists of two parts: images and labels. The images include sequences from multiple positions such as RAO CRA, AP CRA, etc. Each sequence has labels categorized as 1 or 0. In this context, 1 indicates that a particular image is a keyframe related to the heartbeat of that sequence, while 0 indicates it is not.</p>
Full article ">Figure 8
<p>A complete DSA sequence designed to visually demonstrate the actual performance of different methods for recognizing heartbeat keyframes within a unified sequence.</p>
Full article ">Figure 9
<p>CDKD-w+ confusion matrix.</p>
Full article ">Figure 10
<p>L1 LOSS confusion matrix.</p>
Full article ">Figure 11
<p>PSNR confusion matrix.</p>
Full article ">Figure 12
<p>SSIM confusion matrix.</p>
Full article ">Figure 13
<p>ResNet confusion matrix.</p>
Full article ">Figure 14
<p>Illustration of vasodilation frame, intermediate frame and vasoconstriction frame.</p>
Full article ">
30 pages, 5271 KiB  
Article
Cognitive Mechanisms Underlying Memory Advantages in Bridge Experts: Based on Suit Categorization and Honor Card Rules
by Yanzhe Liu, Yan Liu, Bingjie Zhao and Qihan Zhang
Behav. Sci. 2025, 15(2), 125; https://doi.org/10.3390/bs15020125 - 24 Jan 2025
Viewed by 401
Abstract
To explore the memory advantage and the underlying mechanisms of bridge experts, this study conducted two experiments. Experiment 1 investigated the effects of the suit categorization rule and the rank ordering rule on the memory performance of bridge experts when memorizing hands. The [...] Read more.
To explore the memory advantage and the underlying mechanisms of bridge experts, this study conducted two experiments. Experiment 1 investigated the effects of the suit categorization rule and the rank ordering rule on the memory performance of bridge experts when memorizing hands. The findings revealed that the suit categorization rule significantly influenced the memory advantage of bridge experts, regardless of whether the task involved recognition or free recall. Conversely, the rank ordering rule had no discernible effect on their memory performance, though the honor card information within this rule notably impacted their memory. Building on the first experiment, Experiment 2 further examined the roles of visual familiarity induced by the suit categorization and honor card rules, alongside the abstract knowledge embedded in these rules, on the memory performance of bridge experts. The results demonstrated that visual familiarity influenced recognition among bridge experts, while both visual familiarity and abstract knowledge jointly contributed to recall performance. These research findings concurrently support both chunking/template theory and SEEK theory. Full article
Show Figures

Figure 1

Figure 1
<p>Materials for Experiment 1.</p>
Full article ">Figure 2
<p>Recognition task procedure.</p>
Full article ">Figure 3
<p>Recall task procedure.</p>
Full article ">Figure 4
<p>Recognition task-related data figures. (<b>a</b>) IES of the expert and control groups under different experimental conditions. (<b>b</b>) Correct rejection rates of similar and different hand materials by the expert group under different experimental conditions. Note. EG: expert group; CG: control group; CR: regular suit categorization; CI: irregular suit categorization; OR: regular rank ordering; OI: irregular rank ordering.</p>
Full article ">Figure 5
<p>The recall performance of the expert and control groups under different experimental conditions. (<b>a</b>) Recall accuracy. (<b>b</b>) The proportion of correctly recalled honor cards. Note. *** <span class="html-italic">p</span> &lt; 0.001. EG: expert group; CG: control group; CR: regular suit categorization; CI: irregular suit categorization; OR: regular rank ordering; OI: irregular rank ordering.</p>
Full article ">Figure 6
<p>The correct recall rate of honor cards for the expert and the control groups under different experimental conditions. Note. EG: expert group; CG: control group; CR: regular suit categorization; CI: irregular suit categorization; OR: regular rank ordering; OI: irregular rank ordering.</p>
Full article ">Figure 7
<p>Memory strategies of the expert and control groups under recognition and recall tasks. (<b>a</b>) Memory strategies for recognition task. (<b>b</b>) Memory strategies for recall task.</p>
Full article ">Figure 8
<p>Experiment 2 materials.</p>
Full article ">Figure 9
<p>IES of experts under different experimental conditions. Note. * <span class="html-italic">p</span> &lt; 0.05; *** <span class="html-italic">p</span> &lt; 0.001, “ns” signifies “not significant”. CR: regular suit categorization; CI: irregular suit categorization; HU: unchanged honor card objects; HC: changed honor card objects; SU: unchanged suit objects; SC: changed suit objects.</p>
Full article ">Figure 10
<p>The rejection performance of bridge experts to unlearned materials. (<b>a</b>) HC-SC condition. (<b>b</b>) CI-SC condition. (<b>c</b>) CI-HC condition. Note. * <span class="html-italic">p</span> &lt; 0.05; ** <span class="html-italic">p</span> &lt; 0.01; *** <span class="html-italic">p</span> &lt; 0.001. CR: regular suit categorization; CI: irregular suit categorization; HU: unchanged honor card objects; HC: changed honor card objects; SU: unchanged suit objects; SC: changed suit objects.</p>
Full article ">Figure 11
<p>Experts’ recall performance under different experimental conditions. (<b>a</b>) Recall accuracy. (<b>b</b>) The correct recall rate of honor cards. Note. *** <span class="html-italic">p</span> &lt; 0.001, “ns” signifies “not significant”. CR: regular suit categorization; CI: irregular suit categorization; HU: unchanged honor card objects; HC: changed honor card objects; SU: unchanged suit objects; SC: changed suit objects.</p>
Full article ">Figure 12
<p>Memory strategies of experts in Experiment 2.</p>
Full article ">
18 pages, 2656 KiB  
Article
Multimodal Natural Disaster Scene Recognition with Integrated Large Model and Mamba
by Yuxuan Shao and Liwen Xu
Appl. Sci. 2025, 15(3), 1149; https://doi.org/10.3390/app15031149 - 23 Jan 2025
Viewed by 450
Abstract
The accurate identification of natural disasters is crucial in ensuring effective post-disaster relief efforts. However, the existing models for disaster classification often incur high costs. To address this, we propose leveraging the most advanced pre-trained large language models, which offer superior generative and [...] Read more.
The accurate identification of natural disasters is crucial in ensuring effective post-disaster relief efforts. However, the existing models for disaster classification often incur high costs. To address this, we propose leveraging the most advanced pre-trained large language models, which offer superior generative and multimodal understanding capabilities. Using a question-answering approach, we extract textual descriptions and category prediction probabilities for disaster scenarios, which are then used as input to our proposed Mamba Multimodal Disaster Recognition Network (Mamba-MDRNet). This model integrates a large pre-trained model with the Mamba mechanism, enabling the selection of the most reliable modality information as a robust basis for scene classification. Extensive experiments demonstrate consistent performance improvements across various visual models with heterogeneous architectures. Notably, integrating EfficientNet within Mamba-MDRNet yielded 97.82% accuracy for natural scene classification, surpassing the performance of the CNN (91.75%), ViT (94.50%), and ResNet18 (97.25%). These results highlight the potential of multimodal models combining large models and the Mamba mechanism for disaster type prediction. Full article
(This article belongs to the Special Issue Deep Learning for Image Processing and Computer Vision)
Show Figures

Figure 1

Figure 1
<p>The overall framework diagram of the Mamba-MDRNet learning algorithm. Only the model within the dashed box will be deployed and used for inference after training.</p>
Full article ">Figure 2
<p>Generated by Gemini1.5 using prompts.</p>
Full article ">Figure 3
<p>Grad-CAM heatmaps for 11 natural disaster categories.</p>
Full article ">Figure 4
<p>Comparison of confusion matrices for natural disaster classification.</p>
Full article ">
19 pages, 865 KiB  
Article
Reversible Adversarial Examples with Minimalist Evolution for Recognition Control in Computer Vision
by Shilong Yang, Lu Leng, Ching-Chun Chang and Chin-Chen Chang
Appl. Sci. 2025, 15(3), 1142; https://doi.org/10.3390/app15031142 - 23 Jan 2025
Viewed by 473
Abstract
As artificial intelligence increasingly automates the recognition and analysis of visual content, it poses significant risks to privacy, security, and autonomy. Computer vision systems can surveil and exploit data without consent. With these concerns in mind, we introduce a novel method to control [...] Read more.
As artificial intelligence increasingly automates the recognition and analysis of visual content, it poses significant risks to privacy, security, and autonomy. Computer vision systems can surveil and exploit data without consent. With these concerns in mind, we introduce a novel method to control whether images can be recognized by computer vision systems using reversible adversarial examples. These examples are generated to evade unauthorized recognition, allowing only systems with permission to restore the original image by removing the adversarial perturbation with zero-bit error. A key challenge with prior methods is their reliance on merely restoring the examples to a state in which they can be correctly recognized by the model; however, the restored images are not fully consistent with the original images, and they require excessive auxiliary information to achieve reversibility. To achieve zero-bit error restoration, we utilize the differential evolution algorithm to optimize adversarial perturbations while minimizing distortion. Additionally, we introduce a dual-color space detection mechanism to localize perturbations, eliminating the need for extra auxiliary information. Ultimately, when combined with reversible data hiding, adversarial attacks can achieve reversibility. Experimental results demonstrate that the PSNR and SSIM between the restored images by the method and the original images are ∞ and 1, respectively. The PSNR and SSIM between the reversible adversarial examples and the original images are 48.32 dB and 0.9986, respectively. Compared to state-of-the-art methods, the method maintains high visual fidelity at a comparable attack success rate. Full article
Show Figures

Figure 1

Figure 1
<p>Framework for the generation and recovery of reversible adversarial examples.</p>
Full article ">Figure 2
<p>The process involves the obtaining and recording of auxiliary information (the original RGB values).</p>
Full article ">Figure 3
<p>Differential computation process.</p>
Full article ">Figure 4
<p>Histogram at each stage. (<b>a</b>–<b>c</b>) represent the histograms of the original image, the histogram after shifting, and the histogram after embedding, respectively.</p>
Full article ">Figure 5
<p>Generation of reversible adversarial example (a 4 × 4 black as an instance).</p>
Full article ">Figure 6
<p>Visualization of perturbed pixel detection.</p>
Full article ">Figure 7
<p>Data extraction and image restoration (the pixels filled with green color remain unchanged).</p>
Full article ">Figure 8
<p>Visual results of the difference matrices generated by different approaches. (<b>a</b>–<b>d</b>) correspond to RGB, HSV, RGB + HSV, and Proposed, respectively.</p>
Full article ">Figure 9
<p>Visual results of proposed method. (<b>a</b>–<b>d</b>) based on CIFAR-10. (<b>a</b>) Original image. (<b>b</b>) AE. (<b>c</b>) RAE. (<b>d</b>) Restored image; (<b>e</b>,<b>f</b>) based on ImageNet. (<b>e</b>) Original image. (<b>f</b>) RAE.</p>
Full article ">
Back to TopTop