[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (9,707)

Search Parameters:
Keywords = image segmentation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 3123 KiB  
Article
DeSPPNet: A Multiscale Deep Learning Model for Cardiac Segmentation
by Elizar Elizar, Rusdha Muharar and Mohd Asyraf Zulkifley
Diagnostics 2024, 14(24), 2820; https://doi.org/10.3390/diagnostics14242820 (registering DOI) - 14 Dec 2024
Abstract
Background: Cardiac magnetic resonance imaging (MRI) plays a crucial role in monitoring disease progression and evaluating the effectiveness of treatment interventions. Cardiac MRI allows medical practitioners to assess cardiac function accurately by providing comprehensive and quantitative information about the structure and function, hence [...] Read more.
Background: Cardiac magnetic resonance imaging (MRI) plays a crucial role in monitoring disease progression and evaluating the effectiveness of treatment interventions. Cardiac MRI allows medical practitioners to assess cardiac function accurately by providing comprehensive and quantitative information about the structure and function, hence making it an indispensable tool for monitoring the disease and treatment response. Deep learning-based segmentation enables the precise delineation of cardiac structures including the myocardium, right ventricle, and left ventricle. The accurate segmentation of these structures helps in the diagnosis of heart failure, cardiac functional response to therapies, and understanding the state of the heart functions after treatment. Objectives: The objective of this study is to develop a multiscale deep learning model to segment cardiac organs based on MRI imaging data. Good segmentation performance is difficult to achieve due to the complex nature of the cardiac structure, which includes a variety of chambers, arteries, and tissues. Furthermore, the human heart is also constantly beating, leading to motion artifacts that reduce image clarity and consistency. As a result, a multiscale method is explored to overcome various challenges in segmenting cardiac MRI images. Methods: This paper proposes DeSPPNet, a multiscale-based deep learning network. Its foundation follows encoder–decoder pair architecture that utilizes the Spatial Pyramid Pooling (SPP) layer to improve the performance of cardiac semantic segmentation. The SPP layer is designed to pool features from densely convolutional layers at different scales or sizes, which will be combined to maintain a set of spatial information. By processing features at different spatial resolutions, the multiscale densely connected layer in the form of the Pyramid Pooling Dense Module (PPDM) helps the network to capture both local and global context, preserving finer details of the cardiac structure while also capturing the broader context required to accurately segment larger cardiac structures. The PPDM is incorporated into the deeper layer of the encoder section of the deep learning network to allow it to recognize complex semantic features. Results: An analysis of multiple PPDM placement scenarios and structural variations revealed that the 3-path PPDM, positioned at the encoder layer 5, yielded optimal segmentation performance, achieving dice, intersection over union (IoU), and accuracy scores of 0.859, 0.800, and 0.993, respectively. Conclusions: Different PPDM configurations produce a different effect on the network; as such, a shallower layer placement, like encoder layer 4, retains more spatial data that need more parallel paths to gather the optimal set of multiscale features. In contrast, deeper layers contain more informative features but at a lower spatial resolution, which reduces the number of parallel paths required to provide optimal multiscale context. Full article
20 pages, 5548 KiB  
Article
Spatial Sense of Safety for Seniors in Living Streets Based on Street View Image Data
by Xuyang Sun, Xinlei Nie, Lu Wang, Zichun Huang and Ruiming Tian
Buildings 2024, 14(12), 3973; https://doi.org/10.3390/buildings14123973 (registering DOI) - 14 Dec 2024
Abstract
As the global population ages, the friendliness of urban spaces towards seniors becomes increasingly crucial. This research primarily investigates the environmental factors that influence the safety perception of elderly people in living street spaces. Taking Dingzigu Street in Tianjin, China, as an example, [...] Read more.
As the global population ages, the friendliness of urban spaces towards seniors becomes increasingly crucial. This research primarily investigates the environmental factors that influence the safety perception of elderly people in living street spaces. Taking Dingzigu Street in Tianjin, China, as an example, by employing deep learning fully convolutional network (FCN-8s) technology and the semantic segmentation method based on computer vision, the objective measurement data of street environmental elements are acquired. Meanwhile, the subjective safety perception evaluation data of elderly people are obtained through SD semantic analysis combined with the Likert scale. Utilizing Pearson correlation analysis and multiple linear regression analysis, the study comprehensively examines the impact of the physical environment characteristics of living street spaces on the spatial safety perception of seniors. The results indicate that, among the objective environmental indicators, ① the street greening rate is positively correlated with the spatial sense of security of seniors; ② there is a negative correlation between sky openness and interface enclosure; and ③ the overall safety perception of seniors regarding street space is significantly influenced by the spatial sense of security, the sense of security during walking behavior, and the security perception in visual recognition. This research not only uncovers the impact mechanism of the street environment on the safety perception of seniors, but also offers valuable references for the age-friendly design of urban spaces. Full article
(This article belongs to the Section Architectural Design, Urban Science, and Real Estate)
Show Figures

Figure 1

Figure 1
<p>Current situation of living street space (author’s own photographs).</p>
Full article ">Figure 2
<p>Research scope (author’s own representation). (<b>a</b>) Research scope: Proportion of elderly population (aged 60 and above); (<b>b</b>) Research scope and road network situation in surrounding areas.</p>
Full article ">Figure 3
<p>Technical road map of street view image semantic segmentation (author’s own representation).</p>
Full article ">Figure 4
<p>Schematic diagram of measurable indicators for street space environment (author’s own representation).</p>
Full article ">Figure 5
<p>Semantic segmentation of street view images and spatial feature data collection process (author’s own representation).</p>
Full article ">Figure 6
<p>Seniors’ perception of street space safety based on Maslow’s hierarchy of needs theory (top image obtained from Abraham Maslow’s 1943 work <span class="html-italic">The Theory of Human Motivation</span> [<a href="#B30-buildings-14-03973" class="html-bibr">30</a>]; bottom picture created by the author).</p>
Full article ">Figure 7
<p>Score curve graph of street space safety perception evaluation based on SD semantic analysis method (author’s own representation).</p>
Full article ">
10 pages, 3682 KiB  
Article
NAILS: Normalized Artificial Intelligence Labeling Sensor for Self-Care Health
by Livio Tenze and Enrique Canessa
Sensors 2024, 24(24), 7997; https://doi.org/10.3390/s24247997 (registering DOI) - 14 Dec 2024
Abstract
Visual examination of nails can reflect human health status. Diseases such as nutritive imbalances and skin diseases can be identified by looking at the colors around the plate part of the nails. We present the AI-based NAILS method to detect fingernails through segmentation [...] Read more.
Visual examination of nails can reflect human health status. Diseases such as nutritive imbalances and skin diseases can be identified by looking at the colors around the plate part of the nails. We present the AI-based NAILS method to detect fingernails through segmentation and labeling. The NAILS leverages a pre-trained Convolutional Neural Network model to segment and label fingernail regions from fingernail images, normalizing RGB values to monitor tiny color changes via a GUI and the use of an HD webcam in real time. The use of normalized RGB values combined with AI-based segmentation for real-time health monitoring is novel and innovative. The NAILS algorithm could be used to self-extract and archive primary signs of diseases in humans, especially in rural areas or when other testing may be not available. Full article
(This article belongs to the Special Issue Feature Papers in Intelligent Sensors 2024)
Show Figures

Figure 1

Figure 1
<p>Diagram of the NAILS sequence algorithm. The hand images are examples from the dataset in [<a href="#B7-sensors-24-07997" class="html-bibr">7</a>].</p>
Full article ">Figure 2
<p>GUI for the Normalized Artificial Intelligence Labeling Sensor (NAILS) aiming to identify hidden signals from fingernails. The numbers 1-5 label each R-G-B curve with a corresponding fingernail.</p>
Full article ">Figure 3
<p>Webcam selection menu.</p>
Full article ">Figure 4
<p>Modalities to acquire (single or multiple) images for the fingernail regions.</p>
Full article ">Figure 5
<p>Fingernail areas regulated in size with the cursor on the right of the GUI.</p>
Full article ">Figure 6
<p>The visible RGB colors of the five fingernail regions being identified at each measurement. The empty white values correspond to missing measures with the CNN algorithm, usually due to the occlusion and position of the thumb with time. The observed nail color range is similar to the skin color and depends on luminosity.</p>
Full article ">Figure 7
<p>Red, green and blue plots of the mean color values for each fingernail.</p>
Full article ">Figure 8
<p>Critical assessment of NAILS behavior by forcing red and blue fingernail colors.</p>
Full article ">
30 pages, 10681 KiB  
Article
Exploring Transfer Learning for Anthropogenic Geomorphic Feature Extraction from Land Surface Parameters Using UNet
by Aaron E. Maxwell, Sarah Farhadpour and Muhammad Ali
Remote Sens. 2024, 16(24), 4670; https://doi.org/10.3390/rs16244670 (registering DOI) - 14 Dec 2024
Viewed by 103
Abstract
Semantic segmentation algorithms, such as UNet, that rely on convolutional neural network (CNN)-based architectures, due to their ability to capture local textures and spatial context, have shown promise for anthropogenic geomorphic feature extraction when using land surface parameters (LSPs) derived from digital terrain [...] Read more.
Semantic segmentation algorithms, such as UNet, that rely on convolutional neural network (CNN)-based architectures, due to their ability to capture local textures and spatial context, have shown promise for anthropogenic geomorphic feature extraction when using land surface parameters (LSPs) derived from digital terrain models (DTMs) as input predictor variables. However, the operationalization of these supervised classification methods is limited by a lack of large volumes of quality training data. This study explores the use of transfer learning, where information learned from another, and often much larger, dataset is used to potentially reduce the need for a large, problem-specific training dataset. Two anthropogenic geomorphic feature extraction problems are explored: the extraction of agricultural terraces and the mapping of surface coal mine reclamation-related valley fill faces. Light detection and ranging (LiDAR)-derived DTMs were used to generate LSPs. We developed custom transfer parameters by attempting to predict geomorphon-based landforms using a large dataset of digital terrain data provided by the United States Geological Survey’s 3D Elevation Program (3DEP). We also explored the use of pre-trained ImageNet parameters and initializing models using parameters learned from the other mapping task investigated. The geomorphon-based transfer learning resulted in the poorest performance while the ImageNet-based parameters generally improved performance in comparison to a random parameter initialization, even when the encoder was frozen or not trained. Transfer learning between the different geomorphic datasets offered minimal benefits. We suggest that pre-trained models developed using large, image-based datasets may be of value for anthropogenic geomorphic feature extraction from LSPs even given the data and task disparities. More specifically, ImageNet-based parameters should be considered as an initialization state for the encoder component of semantic segmentation architectures applied to anthropogenic geomorphic feature extraction even when using non-RGB image-based predictor variables, such as LSPs. The value of transfer learning between the different geomorphic mapping tasks may have been limited due to smaller sample sizes, which highlights the need for continued research in using unsupervised and semi-supervised learning methods, especially given the large volume of digital terrain data available, despite the lack of associated labels. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Location of 10 × 10 km USGS 3DEP tiles used in this study for which geomorphons were calculated; (<b>b</b>) training, validation, and testing extents for agricultural terrace (terraceDL) dataset [<a href="#B21-remotesensing-16-04670" class="html-bibr">21</a>] in Iowa, USA; (<b>c</b>) training, validation, and testing extents for surface coal mining valley fill face (vfillDL) dataset [<a href="#B22-remotesensing-16-04670" class="html-bibr">22</a>] in West Virginia, Kentucky, and Virginia, USA.</p>
Full article ">Figure 2
<p>Example agricultural terraces in Iowa, USA (<b>a</b>) and valley fill faces (<b>c</b>) in the Appalachian southern coalfields of the eastern United States. Red areas in (<b>a</b>,<b>c</b>) show the extents of terraces and valley fill faces, respectively, over a multidirectional hillshade. The LSPs used in this study are visualized in (<b>b</b>,<b>d</b>) (red = TPI calculated with a 50 m circular window; green = square root of slope; blue = TPI calculated with a 2 m inner and 5 m outer annulus window). Coordinates are relative to the NAD83 UTM Zone 15N projection for (<b>a</b>,<b>b</b>) and the NAD83 UTM Zone 17N projection for (<b>c</b>,<b>d</b>).</p>
Full article ">Figure 3
<p>Example land surface parameter (LSP) composite image chips (<b>a</b>) and associated geomorphon classifications (<b>b</b>). Chips were selected from random locations within the extent of the downloaded 3DEP DTM data. Each chip consists of 512 × 512 cells with a spatial resolution of 2 m.</p>
Full article ">Figure 4
<p>Conceptualization of UNet architecture [<a href="#B19-remotesensing-16-04670" class="html-bibr">19</a>] with the ResNet-34 [<a href="#B20-remotesensing-16-04670" class="html-bibr">20</a>] encoder backbone used in this study. E = encoder; D = decoder, CH = classification head, LSPs = land surface parameters; Conv = convolutional layer, BN = batch normalization, and ReLU = rectified linear unit.</p>
Full article ">Figure 5
<p>Example classification results using a random parameter initiation and 1000 training chips. (<b>a</b>) Multidirectional hillshade for example agricultural terrace classification; (<b>b</b>) reference agricultural terrace data; (<b>c</b>) agricultural terrace classification result; (<b>d</b>) multidirectional hillshade for example valley fill face classification; (<b>e</b>) reference valley fill face data; (<b>f</b>) valley fill face classification result; (<b>g</b>) multidirectional hillshade for example geomorphon classification; (<b>h</b>) reference geomorphon data; (<b>i</b>) geomorphon classification result.</p>
Full article ">Figure 6
<p>Training loss for terraceDL (<b>a</b>) and vfillDL (<b>b</b>) datasets using 1000 training samples, different weight initiations, and with the encoder frozen or unfrozen across all 50 training epochs. Magnified area shows results for epochs 40 through 50.</p>
Full article ">Figure 7
<p>Validation F1-score for terraceDL (<b>a</b>) and vfillDL (<b>b</b>) datasets using 1000 training samples, different weight initiations, and with the encoder frozen or unfrozen across all 50 training epochs. Magnified area shows results for epochs 40 through 50.</p>
Full article ">Figure 8
<p>Training loss for terraceDL (<b>a</b>) and vfillDL (<b>b</b>) datasets using varying training sample sizes, different weight initiations, and with the encoder frozen or unfrozen. Magnified area shows results for epochs 40 through 50.</p>
Full article ">Figure 9
<p>Validation F1-score for terraceDL (<b>a</b>) and vfillDL (<b>b</b>) datasets using varying training sample sizes, different weight initiations, and with the encoder frozen or unfrozen. Magnified area shows results for epochs 40 through 50.</p>
Full article ">Figure 10
<p>Assessment metrics calculated from the withheld test data for terraceDL (<b>top</b>) and vfillDL (<b>bottom</b>) datasets using different weight initiations and with the encoder frozen and unfrozen. Results reflect the experiment using 1000 training chips and the model parameters associated with the training epoch that provided the highest F1-score for the validation data.</p>
Full article ">Figure 11
<p>Assessment metrics for withheld test data for terraceDL dataset using different training sample sizes, weight initiations, and with the encoder frozen and unfrozen.</p>
Full article ">Figure 12
<p>Assessment metrics for withheld test data for vfillDL dataset using different training sample sizes, weight initiations, and with the encoder frozen and unfrozen.</p>
Full article ">Figure 13
<p>CKA analysis results for each convolutional layer in the architecture. Each graph represents a comparison of a pair of models. Each compared model was trained from a random initialization and using the largest training set available for the specific task. Since the ImageNet weights are not available for the decoder, the decoder blocks were not compared when ImageNet was included in the pair.</p>
Full article ">
19 pages, 7712 KiB  
Article
Research on Intelligent Diagnosis of Corrosion in the Operation and Maintenance Stage of Steel Structure Engineering Based on U-Net Attention
by Zhen Duan, Xinghong Huang, Jia Hou, Wei Chen and Lixiong Cai
Buildings 2024, 14(12), 3972; https://doi.org/10.3390/buildings14123972 (registering DOI) - 14 Dec 2024
Viewed by 106
Abstract
Intelligent corrosion diagnosis plays a crucial role in enhancing the efficiency of operation and maintenance for steel structures. Presently, corrosion detection primarily depends on manual visual inspections and non-destructive testing methods, which are inefficient, costly, and subject to human bias. While machine vision [...] Read more.
Intelligent corrosion diagnosis plays a crucial role in enhancing the efficiency of operation and maintenance for steel structures. Presently, corrosion detection primarily depends on manual visual inspections and non-destructive testing methods, which are inefficient, costly, and subject to human bias. While machine vision has demonstrated significant potential in controlled laboratory settings, most studies have focused on environments with limited background interference, restricting their practical applicability. To tackle the challenges posed by complex backgrounds and multiple interference factors in field-collected images of steel components, this study introduces an intelligent corrosion grading method designed specifically for images containing background elements. By integrating an attention mechanism into the traditional U-Net network, we achieve precise segmentation of component pixels from background pixels in engineering images, attaining an accuracy of up to 94.1%. The proposed framework is validated using images collected from actual engineering sites. A sliding window sampling technique divides on-site images into several rectangular windows, which are filtered based on U-Net Attention segmentation results. Leveraging a dataset of steel plate corrosion images with known grades, we train an Inception v3 corrosion classification model. Transfer learning techniques are then applied to determine the corrosion grade of each filtered window, culminating in a weighted average to estimate the overall corrosion grade of the target component. This study provides a quantitative index for assessing large-scale steel structure corrosion, significantly impacting the improvement of construction and maintenance quality while laying a solid foundation for further research and development in related fields. Full article
Show Figures

Figure 1

Figure 1
<p>Process for determining corrosion grades of components.</p>
Full article ">Figure 2
<p>U-Net architecture diagram.</p>
Full article ">Figure 3
<p>Structure of the CBAM module.</p>
Full article ">Figure 4
<p>Structure of the improved U-Net Attention model.</p>
Full article ">Figure 5
<p>Images of corrosion on various parts of the steel structure.</p>
Full article ">Figure 6
<p>Labelme data annotation.</p>
Full article ">Figure 7
<p>Example of the semantic segmentation dataset.</p>
Full article ">Figure 8
<p>Semantic segmentation dataset. (<b>a</b>) Original images from the dataset. (<b>b</b>) Annotated label images from the dataset.</p>
Full article ">Figure 9
<p>Comparison of segmentation results across models.</p>
Full article ">Figure 10
<p>Image size transformation.</p>
Full article ">Figure 11
<p>Simple linear interpolation.</p>
Full article ">Figure 12
<p>Bilinear interpolation.</p>
Full article ">Figure 13
<p>Calculation of bilinear interpolation.</p>
Full article ">Figure 14
<p>Sliding window sampling.</p>
Full article ">Figure 15
<p>Three different types of windows.</p>
Full article ">Figure 16
<p>The corrosion grade corresponds to the dataset.</p>
Full article ">Figure 17
<p>Feature map visualization.</p>
Full article ">Figure 18
<p>Model output of corrosion levels.</p>
Full article ">Figure 19
<p>Number of windows for each corrosion level of steel component samples.</p>
Full article ">
18 pages, 2056 KiB  
Article
Alternative Non-Destructive Approach for Estimating Morphometric Measurements of Chicken Eggs from Tomographic Images with Computer Vision
by Jean Pierre Brik López Vargas, Katariny Lima de Abreu, Davi Duarte de Paula, Denis Henrique Pinheiro Salvadeo, Lilian Francisco Arantes de Souza and Carlos Bôa-Viagem Rabello
Foods 2024, 13(24), 4039; https://doi.org/10.3390/foods13244039 (registering DOI) - 14 Dec 2024
Viewed by 116
Abstract
The egg has natural barriers that prevent microbiological contamination and promote food safety. The use of non-destructive methods to obtain morphometric measurements of chicken eggs has the potential to replace traditional invasive techniques, offering greater efficiency and accuracy. This paper aims to demonstrate [...] Read more.
The egg has natural barriers that prevent microbiological contamination and promote food safety. The use of non-destructive methods to obtain morphometric measurements of chicken eggs has the potential to replace traditional invasive techniques, offering greater efficiency and accuracy. This paper aims to demonstrate that estimates derived from non-invasive approaches, such as 3D computed tomography (CT) image analysis, can be comparable to conventional destructive methods. To achieve this goal, two widely recognized deep learning architectures, U-Net 3D and Fully Convolutional Networks (FCN) 3D, were modeled to segment and analyze 3D CT images of chicken eggs. A dataset of real CT images was created and labeled, allowing the extraction of important morphometric measurements, including height, width, shell thickness, and volume. The models achieved an accuracy of up to 98.69%, demonstrating their effectiveness compared to results from manual measurements. These findings highlight the potential of CT image analysis, combined with deep learning, as a non-invasive alternative in industrial and research settings. This approach not only minimizes the need for invasive procedures but also offers a scalable and reliable method for egg quality assessment. Full article
Show Figures

Figure 1

Figure 1
<p>Measuring Equipment: (<b>a</b>) Egg measurements, length and width measured with a digital caliper (<b>b</b>) Eggshell thickness, measured with digital micrometer (<b>c</b>) Acquisition of images of chicken eggs by computed tomography.</p>
Full article ">Figure 2
<p>Labeling of egg structures made in CVAT.</p>
Full article ">Figure 3
<p>Pipeline of Estimated Morphometrics Measurements since the Inputs (Images) to Outputs (Estimated Masks and Morphometrics Measures) of the Models.</p>
Full article ">Figure 4
<p>Approach to counting the number of voxels for estimating morphometric measurements.</p>
Full article ">Figure 5
<p>Metrics of the segmentation stage: (<b>a</b>) Accuracy metrics of the training and validation sets of the U-Net and FCN models. (<b>b</b>) Loss metrics of the training and validation sets of the U-Net and FCN models.</p>
Full article ">Figure 6
<p>Comparison between a slice (<b>right</b>) with low accuracy and a slice (<b>left</b>) with high accuracy.</p>
Full article ">Figure 7
<p>Comparison of real tomographic image slices (column 1), corresponding labeled mask slices (column 2), probability outputs of the networks for each class (column 3–12), showing in yellow the most probable class and the final predictions for the U-Net and FCN networks (column 13–14).</p>
Full article ">
26 pages, 18107 KiB  
Article
Tree Species Classification for Shelterbelt Forest Based on Multi-Source Remote Sensing Data Fusion from Unmanned Aerial Vehicles
by Kai Jiang, Qingzhan Zhao, Xuewen Wang, Yuhao Sheng and Wenzhong Tian
Forests 2024, 15(12), 2200; https://doi.org/10.3390/f15122200 - 13 Dec 2024
Viewed by 239
Abstract
Accurately understanding the stand composition of shelter forests is essential for the construction and benefit evaluation of shelter forest projects. This study explores classification methods for dominant tree species in shelter forests using UAV-derived RGB, hyperspectral, and LiDAR data. It also investigates the [...] Read more.
Accurately understanding the stand composition of shelter forests is essential for the construction and benefit evaluation of shelter forest projects. This study explores classification methods for dominant tree species in shelter forests using UAV-derived RGB, hyperspectral, and LiDAR data. It also investigates the impact of individual tree crown (ITC) delineation accuracy, crown morphological parameters, and various data sources and classifiers. First, as a result of the overlap and complex structure of tree crowns in shelterbelt forests, existing ITC delineation methods often lead to over-segmentation or segmentation errors. To address this challenge, we propose a watershed and multi-feature-controlled spectral clustering (WMF-SCS) algorithm for ITC delineation based on UAV RGB and LiDAR data, which offers clearer and more reliable classification objects, features, and training data for tree species classification. Second, spectral, texture, structural, and crown morphological parameters were extracted using UAV hyperspectral and LiDAR data combined with ITC delineation results. Twenty-one classification images were constructed using RF, SVM, MLP, and SAMME for tree species classification. The results show that (1) the proposed WMF-SCS algorithm demonstrates significant performance in ITC delineation in complex mixed forest scenarios (Precision = 0.88, Recall = 0.87, F1-Score = 0.87), resulting in a 1.85% increase in overall classification accuracy; (2) the inclusion of crown morphological parameters derived from LiDAR data improves the overall accuracy of the random forest classifier by 5.82%; (3) compared to using LiDAR or hyperspectral data alone, the classification accuracy using multi-source data improves by an average of 7.94% and 7.52%, respectively; (4) the random forest classifier combined with multi-source data achieves the highest classification accuracy and consistency (OA = 90.70%, Kappa = 0.8747). Full article
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of the study area.</p>
Full article ">Figure 2
<p>Flowchart of the technical roadmap.</p>
Full article ">Figure 3
<p>Flowchart of WMF-SCS algorithm for individual tree crown delineation.</p>
Full article ">Figure 4
<p>Delineation results of site 1 using (<b>a</b>) CHM-based watershed segmentation algorithm; (<b>b</b>) WMF-SCS algorithm; (<b>c</b>) overlap display.</p>
Full article ">Figure 5
<p>Delineation results of site 2 using (<b>a</b>) CHM-based watershed segmentation algorithm; (<b>b</b>) WMF-SCS algorithm; (<b>c</b>) overlap display.</p>
Full article ">Figure 6
<p>Delineation results of site 3 using (<b>a</b>) CHM-based watershed segmentation algorithm; (<b>b</b>) WMF-SCS algorithm; (<b>c</b>) overlap display.</p>
Full article ">Figure 7
<p>Spectral response curve of tree species using (<b>a</b>) full hyperspectral bands; (<b>b</b>) CV-SVM-RFE algorithm.</p>
Full article ">Figure 8
<p>Screening results and importance of variables.</p>
Full article ">Figure 9
<p>Box plots of CPA, CSI, I_PH75, CD, H_Skewness, and PD of LiDAR data.</p>
Full article ">Figure 10
<p>Box plots of CARI, PCA2, MTCI, MNF_Homogeneity, MNF_ASM, and B12_Mean of hyperspectral data.</p>
Full article ">Figure 11
<p>Box plots of CPA, CD, CARI, MTCI, PD, and REP of LiDAR and hyperspectral data.</p>
Full article ">Figure 12
<p>The relationship between the number of LiDAR data feature variables and classification accuracy.</p>
Full article ">Figure 13
<p>The relationship between the number of hyperspectral data feature variables and classification accuracy.</p>
Full article ">Figure 14
<p>The relationship between the number of LiDAR and hyperspectral data feature variables and classification accuracy.</p>
Full article ">Figure 15
<p>LiDAR point clouds of different tree species.</p>
Full article ">Figure 16
<p>Tree species classification results: (<b>a</b>) Site 1; (<b>b</b>) Site 2; (<b>c</b>) Site 3.</p>
Full article ">
17 pages, 8597 KiB  
Article
Automatic Segmentation of Metastatic Livers by Means of U-Net-Based Procedures
by Camilla Tiraboschi, Federica Parenti, Fabio Sangalli, Andrea Resovi, Dorina Belotti and Ettore Lanzarone
Cancers 2024, 16(24), 4159; https://doi.org/10.3390/cancers16244159 - 13 Dec 2024
Viewed by 198
Abstract
Background: The liver is one of the most common sites for the spread of pancreatic ductal adenocarcinoma (PDAC) cells, with metastases present in about 80% of patients. Clinical and preclinical studies of PDAC require quantification of the liver’s metastatic burden from several acquired [...] Read more.
Background: The liver is one of the most common sites for the spread of pancreatic ductal adenocarcinoma (PDAC) cells, with metastases present in about 80% of patients. Clinical and preclinical studies of PDAC require quantification of the liver’s metastatic burden from several acquired images, which can benefit from automatic image segmentation tools. Methods: We developed three neural networks based on U-net architecture to automatically segment the healthy liver area (HL), the metastatic liver area (MLA), and liver metastases (LM) in micro-CT images of a mouse model of PDAC with liver metastasis. Three alternative U-nets were trained for each structure to be segmented following appropriate image preprocessing and the one with the highest performance was then chosen and applied for each case. Results: Good performance was achieved, with accuracy of 92.6%, 88.6%, and 91.5%, specificity of 95.5%, 93.8%, and 99.9%, Dice of 71.6%, 74.4%, and 29.9%, and negative predicted value (NPV) of 97.9%, 91.5%, and 91.5% on the pilot validation set for the chosen HL, MLA, and LM networks, respectively. Conclusions: The networks provided good performance and advantages in terms of saving time and ensuring reproducibility. Full article
(This article belongs to the Special Issue Advanced Research in Pancreatic Ductal Adenocarcinoma)
Show Figures

Figure 1

Figure 1
<p>Proposed architecture consisting of three CNNs: the healthy liver (HL) network, the metastatic liver area (MLA) network, and the liver metastases (LM) network.</p>
Full article ">Figure 2
<p>Healthy liver slice (<b>a</b>) and metastatic liver slice (<b>b</b>) in the sagittal (∼2600 × 1500 pixels), frontal (∼2600 × 1600 pixels), and transverse planes (∼1500 × 1600 pixels), visualized with open source software 3DSlicer (version 5.7) [<a href="#B22-cancers-16-04159" class="html-bibr">22</a>].</p>
Full article ">Figure 3
<p>Sample images of healthy liver for HL: proximal slice (<b>a</b>), slice in the middle (<b>b</b>), and distal slice (<b>c</b>). The acquired image, the corresponding GT, the HL segmentations before binarization, and the predicted BPMs from all networks (U-net-1, U-net-2, and U-net-3) are reported for each slice.</p>
Full article ">Figure 4
<p>Sample images of liver with metastases for MLA: proximal slice (<b>a</b>), slice in the middle (<b>b</b>), and distal slice (<b>c</b>). The acquired image, the corresponding GT, the MLA segmentations before binarization and the predicted BPMs from all networks (U-net-1, U-net-2, and U-net-3) are reported for each slice.</p>
Full article ">Figure 5
<p>Sample images of liver with metastases for LM: proximal slice (<b>a</b>), slice in the middle (<b>b</b>), and distal slice (<b>c</b>). The acquired image, the corresponding GT, the LM segmentations before binarization, and the predicted BPMs from all networks (U-net-1, U-net-2, and U-net-3) are reported for each slice.</p>
Full article ">Figure 6
<p>Comparison between the original U-net-3 for LM (<b>left</b>) and the alternative one based on the manually cleaned images (<b>right</b>): image and GT common to both alternatives, LM segmentation before binarization, and predicted BPM.</p>
Full article ">Figure 7
<p>Examples of a combined mask: BPM for metastases in yellow and metastatic liver surface in blue (<b>a</b>); GT with metastases in fuchsia and metastatic liver surface in blue (<b>b</b>).</p>
Full article ">
17 pages, 3635 KiB  
Article
Automatic Segmentation in 3D CT Images: A Comparative Study of Deep Learning Architectures for the Automatic Segmentation of the Abdominal Aorta
by Christos Mavridis, Theodoros P. Vagenas, Theodore L. Economopoulos, Ioannis Vezakis, Ourania Petropoulou, Ioannis Kakkos and George K. Matsopoulos
Electronics 2024, 13(24), 4919; https://doi.org/10.3390/electronics13244919 - 13 Dec 2024
Viewed by 311
Abstract
Abdominal aortic aneurysm (AAA) is a complex vascular condition associated with high mortality rates. Accurate abdominal aorta segmentation is essential in medical imaging, facilitating diagnosis and treatment for a range of cardiovascular diseases. In this regard, deep learning-based automated segmentation has shown significant [...] Read more.
Abdominal aortic aneurysm (AAA) is a complex vascular condition associated with high mortality rates. Accurate abdominal aorta segmentation is essential in medical imaging, facilitating diagnosis and treatment for a range of cardiovascular diseases. In this regard, deep learning-based automated segmentation has shown significant promise in the precise delineation of the aorta. However, comparisons across different models remain limited, with most studies performing algorithmic training and testing on the same dataset. Furthermore, due to the variability in AAA presentation, using healthy controls for deep learning AAA segmentation poses a significant challenge. This study provides a detailed comparative analysis of four deep learning architectures—UNet, SegResNet, UNet Transformers (UNETR), and Shifted-Windows UNet Transformers (SwinUNETR)—for full abdominal aorta segmentation. The models were evaluated both qualitatively and quantitatively using private and public 3D (Computed Tomography) CT datasets. Moreover, they were successful in attaining high performance in delineating AAA aorta, while being trained on healthy aortic imaging data. Our findings indicate that the UNet architecture achieved the highest segmentation accuracy among the models tested. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of UNet.</p>
Full article ">Figure 2
<p>The architecture of UNETR.</p>
Full article ">Figure 3
<p>The architecture of SwinUNETR.</p>
Full article ">Figure 4
<p>The architecture of SegResNet.</p>
Full article ">Figure 5
<p>Detected regions of interest (aorta) superimposed on the original imaging data slices for two clinical cases (row 1 and 2) of the public dataset for (<b>a</b>) the initial image; (<b>b</b>) ground truth; (<b>c</b>) UNet model; (<b>d</b>) UNETR model; (<b>e</b>) SwinUNETR model; and (<b>f</b>) SegResNet.</p>
Full article ">Figure 6
<p>Detected regions of interest (aorta) superimposed on the original imaging data slices for two clinical cases (row 1 and 2) of the private dataset for (<b>a</b>) the initial image; (<b>b</b>) ground truth; (<b>c</b>) UNet model; (<b>d</b>) UNETR model; (<b>e</b>) SwinUNETR model; and (<b>f</b>) SegResNet.</p>
Full article ">Figure 7
<p>Three-dimensional fused models of the estimated aorta (blue) superimposed on the ground truth (coral) for three cases, using (<b>a</b>) UNet; (<b>b</b>) UNETR; (<b>c</b>) SwinUNETR; and (<b>d</b>) SegResNet.</p>
Full article ">Figure 7 Cont.
<p>Three-dimensional fused models of the estimated aorta (blue) superimposed on the ground truth (coral) for three cases, using (<b>a</b>) UNet; (<b>b</b>) UNETR; (<b>c</b>) SwinUNETR; and (<b>d</b>) SegResNet.</p>
Full article ">
15 pages, 285 KiB  
Perspective
Application of Artificial Intelligence in Otology: Past, Present, and Future
by Hajime Koyama, Akinori Kashio and Tatsuya Yamasoba
J. Clin. Med. 2024, 13(24), 7577; https://doi.org/10.3390/jcm13247577 - 13 Dec 2024
Viewed by 240
Abstract
Artificial Intelligence (AI) is a concept whose goal is to imitate human intellectual activity in computers. It emerged in the 1950s and has gone through three booms. We are in the third boom, and it will continue. Medical applications of AI include diagnosing [...] Read more.
Artificial Intelligence (AI) is a concept whose goal is to imitate human intellectual activity in computers. It emerged in the 1950s and has gone through three booms. We are in the third boom, and it will continue. Medical applications of AI include diagnosing otitis media from images of the eardrum, often outperforming human doctors. Temporal bone CT and MRI analyses also benefit from AI, with segmentation accuracy improved in anatomically significant structures or diagnostic accuracy improved in conditions such as otosclerosis and vestibular schwannoma. In treatment, AI predicts hearing outcomes for sudden sensorineural hearing loss and post-operative hearing outcomes for patients who have undergone tympanoplasty. AI helps patients with hearing aids hear in challenging situations, such as in noisy environments or when multiple people are speaking. It also provides fitting information to help improve hearing with hearing aids. AI also improves cochlear implant mapping and outcome prediction, even in cases of cochlear malformation. Future trends include generative AI, such as ChatGPT, which can provide medical advice and information, although its reliability and application in clinical settings requires further investigation. Full article
(This article belongs to the Section Otolaryngology)
23 pages, 3369 KiB  
Article
Transfer Learning-Driven Cattle Instance Segmentation Using Deep Learning Models
by Rotimi-Williams Bello, Pius A. Owolawi, Etienne A. van Wyk and Chunling Tu
Agriculture 2024, 14(12), 2282; https://doi.org/10.3390/agriculture14122282 - 12 Dec 2024
Viewed by 392
Abstract
Among the emerging applications of artificial intelligence is animal instance segmentation, which has provided a practical means for various researchers to accomplish some aim or execute some order. Though video and image processing are two of the several complex tasks in artificial intelligence, [...] Read more.
Among the emerging applications of artificial intelligence is animal instance segmentation, which has provided a practical means for various researchers to accomplish some aim or execute some order. Though video and image processing are two of the several complex tasks in artificial intelligence, these tasks have become more complex due to the large data and resources needed for training deep learning models. However, these challenges are beginning to be overcome by the transfer learning method of deep learning. In furtherance of the application of the transfer learning method, a system is proposed in this study that applies transfer learning to the detection and recognition of animal activity in a typical farm environment using deep learning models. Among the deep learning models compared, Enhanced Mask R-CNN obtained a significant computing time of 0.2 s and 97% mAP results, which are better than the results obtained by Mask R-CNN, Faster R-CNN, SSD, and YOLOv3, respectively. The findings from the results obtained in this study validate the innovative use of transfer learning to address challenges in cattle segmentation by optimizing the segmentation accuracy and processing time (0.2 s) of the proposed Enhanced Mask R-CNN. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

Figure 1
<p>Samples of cattle datasets in the open-range environment depicting Holstein Friesian cattle (<b>a</b>), Keteku-Muturu cattle (<b>b</b>).</p>
Full article ">Figure 2
<p>DJI Phantom 4 Rtk Drone.</p>
Full article ">Figure 3
<p>Flow diagram of the algorithm for cattle instance segmentation.</p>
Full article ">Figure 4
<p>Process flow of the transfer learning [<a href="#B44-agriculture-14-02282" class="html-bibr">44</a>].</p>
Full article ">Figure 5
<p>Framework of Mask R-CNN for instance segmentation [<a href="#B25-agriculture-14-02282" class="html-bibr">25</a>].</p>
Full article ">Figure 6
<p>Residual network architecture [<a href="#B48-agriculture-14-02282" class="html-bibr">48</a>].</p>
Full article ">Figure 7
<p>Residual learning as a building block [<a href="#B48-agriculture-14-02282" class="html-bibr">48</a>].</p>
Full article ">Figure 8
<p>Enhanced architecture of a residual network.</p>
Full article ">Figure 9
<p>The FCL of Mask R-CNN integrated with sub-network of Enhanced Mask R-CNN for enhanced output.</p>
Full article ">Figure 10
<p>Instance segmentation output of the input cattle images generated by different segmentation algorithms compared in this study, namely YOLOv3 (<b>a</b>,<b>e</b>), Faster R-CNN (<b>b</b>,<b>f</b>), SSD (<b>c</b>,<b>g</b>), Enhanced Mask R-CNN (<b>d</b>,<b>h</b>), Mask R-CNN (<b>i</b>,<b>k</b>), and Enhanced Mask R-CNN (<b>j</b>,<b>l</b>). Being the model whose bias and weights were utilized in training the Enhanced Mask R-CNN in the transfer learning process, we limit the confidence scores and class prediction evaluation to only Mask R-CNN and Enhanced Mask R-CNN models, as shown in subfigures (<b>i</b>,<b>k</b>) and subfigures (<b>j</b>,<b>l</b>), respectively.</p>
Full article ">Figure 10 Cont.
<p>Instance segmentation output of the input cattle images generated by different segmentation algorithms compared in this study, namely YOLOv3 (<b>a</b>,<b>e</b>), Faster R-CNN (<b>b</b>,<b>f</b>), SSD (<b>c</b>,<b>g</b>), Enhanced Mask R-CNN (<b>d</b>,<b>h</b>), Mask R-CNN (<b>i</b>,<b>k</b>), and Enhanced Mask R-CNN (<b>j</b>,<b>l</b>). Being the model whose bias and weights were utilized in training the Enhanced Mask R-CNN in the transfer learning process, we limit the confidence scores and class prediction evaluation to only Mask R-CNN and Enhanced Mask R-CNN models, as shown in subfigures (<b>i</b>,<b>k</b>) and subfigures (<b>j</b>,<b>l</b>), respectively.</p>
Full article ">Figure 10 Cont.
<p>Instance segmentation output of the input cattle images generated by different segmentation algorithms compared in this study, namely YOLOv3 (<b>a</b>,<b>e</b>), Faster R-CNN (<b>b</b>,<b>f</b>), SSD (<b>c</b>,<b>g</b>), Enhanced Mask R-CNN (<b>d</b>,<b>h</b>), Mask R-CNN (<b>i</b>,<b>k</b>), and Enhanced Mask R-CNN (<b>j</b>,<b>l</b>). Being the model whose bias and weights were utilized in training the Enhanced Mask R-CNN in the transfer learning process, we limit the confidence scores and class prediction evaluation to only Mask R-CNN and Enhanced Mask R-CNN models, as shown in subfigures (<b>i</b>,<b>k</b>) and subfigures (<b>j</b>,<b>l</b>), respectively.</p>
Full article ">Figure 10 Cont.
<p>Instance segmentation output of the input cattle images generated by different segmentation algorithms compared in this study, namely YOLOv3 (<b>a</b>,<b>e</b>), Faster R-CNN (<b>b</b>,<b>f</b>), SSD (<b>c</b>,<b>g</b>), Enhanced Mask R-CNN (<b>d</b>,<b>h</b>), Mask R-CNN (<b>i</b>,<b>k</b>), and Enhanced Mask R-CNN (<b>j</b>,<b>l</b>). Being the model whose bias and weights were utilized in training the Enhanced Mask R-CNN in the transfer learning process, we limit the confidence scores and class prediction evaluation to only Mask R-CNN and Enhanced Mask R-CNN models, as shown in subfigures (<b>i</b>,<b>k</b>) and subfigures (<b>j</b>,<b>l</b>), respectively.</p>
Full article ">Figure 10 Cont.
<p>Instance segmentation output of the input cattle images generated by different segmentation algorithms compared in this study, namely YOLOv3 (<b>a</b>,<b>e</b>), Faster R-CNN (<b>b</b>,<b>f</b>), SSD (<b>c</b>,<b>g</b>), Enhanced Mask R-CNN (<b>d</b>,<b>h</b>), Mask R-CNN (<b>i</b>,<b>k</b>), and Enhanced Mask R-CNN (<b>j</b>,<b>l</b>). Being the model whose bias and weights were utilized in training the Enhanced Mask R-CNN in the transfer learning process, we limit the confidence scores and class prediction evaluation to only Mask R-CNN and Enhanced Mask R-CNN models, as shown in subfigures (<b>i</b>,<b>k</b>) and subfigures (<b>j</b>,<b>l</b>), respectively.</p>
Full article ">Figure 10 Cont.
<p>Instance segmentation output of the input cattle images generated by different segmentation algorithms compared in this study, namely YOLOv3 (<b>a</b>,<b>e</b>), Faster R-CNN (<b>b</b>,<b>f</b>), SSD (<b>c</b>,<b>g</b>), Enhanced Mask R-CNN (<b>d</b>,<b>h</b>), Mask R-CNN (<b>i</b>,<b>k</b>), and Enhanced Mask R-CNN (<b>j</b>,<b>l</b>). Being the model whose bias and weights were utilized in training the Enhanced Mask R-CNN in the transfer learning process, we limit the confidence scores and class prediction evaluation to only Mask R-CNN and Enhanced Mask R-CNN models, as shown in subfigures (<b>i</b>,<b>k</b>) and subfigures (<b>j</b>,<b>l</b>), respectively.</p>
Full article ">
32 pages, 10548 KiB  
Article
An Unsupervised Remote Sensing Image Change Detection Method Based on RVMamba and Posterior Probability Space Change Vector
by Jiaxin Song, Shuwen Yang, Yikun Li and Xiaojun Li
Remote Sens. 2024, 16(24), 4656; https://doi.org/10.3390/rs16244656 - 12 Dec 2024
Viewed by 285
Abstract
Change vector analysis in posterior probability space (CVAPS) is an effective change detection (CD) framework that does not require sound radiometric correction and is robust against accumulated classification errors. Based on training samples within target images, CVAPS can generate a uniformly scaled change-magnitude [...] Read more.
Change vector analysis in posterior probability space (CVAPS) is an effective change detection (CD) framework that does not require sound radiometric correction and is robust against accumulated classification errors. Based on training samples within target images, CVAPS can generate a uniformly scaled change-magnitude map that is suitable for a global threshold. However, vigorous user intervention is required to achieve optimal performance. Therefore, to eliminate user intervention and retain the merit of CVAPS, an unsupervised CVAPS (UCVAPS) CD method, RFCC, which does not require rigorous user training, is proposed in this study. In the RFCC, we propose an unsupervised remote sensing image segmentation algorithm based on the Mamba model, i.e., RVMamba differentiable feature clustering, which introduces two loss functions as constraints to ensure that RVMamba achieves accurate segmentation results and to supply the CSBN module with high-quality training samples. In the CD module, the fuzzy C-means clustering (FCM) algorithm decomposes mixed pixels into multiple signal classes, thereby alleviating cumulative clustering errors. Then, a context-sensitive Bayesian network (CSBN) model is introduced to incorporate spatial information at the pixel level to estimate the corresponding posterior probability vector. Thus, it is suitable for high-resolution remote sensing (HRRS) imagery. Finally, the UCVAPS framework can generate a uniformly scaled change-magnitude map that is suitable for the global threshold and can produce accurate CD results. The experimental results on seven change detection datasets confirmed that the proposed method outperforms five state-of-the-art competitive CD methods. Full article
Show Figures

Figure 1

Figure 1
<p>Unsupervised change detection process based on RVMamba and Posterior Probability.</p>
Full article ">Figure 2
<p>Feature extraction network for visual state space modeling. (<b>a</b>) The overarching design of RVMamba. (<b>b</b>) VSS block; SS2D is the core operation in VSS block.</p>
Full article ">Figure 3
<p>Data flow of SS2D. It expands the inputs in four directions according to the serial number, scans them one by one through S6, and then merges them.</p>
Full article ">Figure 4
<p>Context-sensitive Bayesian network model.</p>
Full article ">Figure 5
<p>Experimental datasets and ground truth.</p>
Full article ">Figure 6
<p>Segmentation accuracies of RVMamba, UNet, and KMeans.</p>
Full article ">Figure 7
<p>Change maps obtained by the different most advanced methods on the dataset DS1. (<b>a</b>) ASEA, (<b>b</b>) PCANet, (<b>c</b>) KPCAMNet, (<b>d</b>) DeepCVA, (<b>e</b>) GMCD, (<b>f</b>) RFCC. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 8
<p>Change maps obtained by the different most advanced methods on the dataset DS2. (<b>a</b>) ASEA, (<b>b</b>) PCANet, (<b>c</b>) KPCAMNet, (<b>d</b>) DeepCVA, (<b>e</b>) GMCD, (<b>f</b>) RFCC. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 9
<p>Change maps obtained by the different most advanced methods on the dataset DS3. (<b>a</b>) ASEA, (<b>b</b>) PCANet, (<b>c</b>) KPCAMNet, (<b>d</b>) DeepCVA, (<b>e</b>) GMCD, (<b>f</b>) RFCC. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 10
<p>Change maps obtained with different algorithms tested on the dataset DS1. (<b>a</b>) RFCC, (<b>b</b>) UNet-FCM-CSBN-CVAPS, (<b>c</b>) RVMamba-FCM-SBN-CVAPS, (<b>d</b>) RVMamba-SVM-CVAPS. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 11
<p>Change maps obtained with different algorithms tested on the dataset DS2. (<b>a</b>) RFCC, (<b>b</b>) UNet-FCM-CSBN-CVAPS, (<b>c</b>) RVMamba-FCM-SBN-CVAPS, (<b>d</b>) RVMamba-SVM-CVAPS. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 12
<p>Change maps obtained with different algorithms tested on the dataset DS3. (<b>a</b>) RFCC, (<b>b</b>) UNet-FCM-CSBN-CVAPS, (<b>c</b>) RVMamba-FCM-SBN-CVAPS, (<b>d</b>) RVMamba-SVM-CVAPS. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 13
<p>Evaluation of change magnitude and entropy in bitemporal simulated posterior probability vectors. (<b>a</b>): Low uncertainty. (<b>b</b>): Appropriate reduction in certainty. (<b>c</b>): High uncertainty.</p>
Full article ">Figure 14
<p>Effect of fuzziness q on algorithm results.</p>
Full article ">Figure 15
<p>Effect of window size on algorithm results.</p>
Full article ">Figure 16
<p>Effect of the number of segmentation labels on Kappa and algorithm timeliness.</p>
Full article ">Figure 17
<p>Change maps generated by different techniques in the adaptive experiments. (Black is TN, white is TP, red is FA, and green is MD).</p>
Full article ">Figure 18
<p>Change detection with unsupervised segmentation.</p>
Full article ">Figure 19
<p>Change detection based on RVMamba, K-means, and Fuzzy C-means unsupervised segmentation.</p>
Full article ">
9 pages, 3409 KiB  
Case Report
Artificial Intelligence-Assisted Segmentation of a Falx Cerebri Calcification on Cone-Beam Computed Tomography: A Case Report
by Julien Issa, Alexandre Chidiac, Paul Mozdziak, Bartosz Kempisty, Barbara Dorocka-Bobkowska, Katarzyna Mehr and Marta Dyszkiewicz-Konwińska
Medicina 2024, 60(12), 2048; https://doi.org/10.3390/medicina60122048 - 12 Dec 2024
Viewed by 267
Abstract
Intracranial calcifications, particularly within the falx cerebri, serve as crucial diagnostic markers ranging from benign accumulations to signs of severe pathologies. The falx cerebri, a dural fold that separates the cerebral hemispheres, presents challenges in visualization due to its low contrast in standard [...] Read more.
Intracranial calcifications, particularly within the falx cerebri, serve as crucial diagnostic markers ranging from benign accumulations to signs of severe pathologies. The falx cerebri, a dural fold that separates the cerebral hemispheres, presents challenges in visualization due to its low contrast in standard imaging techniques. Recent advancements in artificial intelligence (AI), particularly in machine learning and deep learning, have significantly transformed radiological diagnostics. This study aims to explore the application of AI in the segmentation and detection of falx cerebri calcifications using Cone-Beam Computed Tomography (CBCT) images through a comprehensive literature review and a detailed case report. The case report presents a 59-year-old patient diagnosed with falx cerebri calcifications whose CBCT images were analyzed using a cloud-based AI platform, demonstrating effectiveness in segmenting these calcifications, although challenges persist in distinguishing these from other cranial structures. A specific search strategy was employed to search electronic databases, yielding four studies exploring AI-based segmentation of the falx cerebri. The review detailed various AI models and their accuracy across different imaging modalities in identifying and segmenting falx cerebri calcifications, also highlighting the gap in publications in this area. In conclusion, further research is needed to improve AI-driven methods for accurately identifying and measuring intracranial calcifications. Advancing AI applications in radiology, particularly for detecting falx cerebri calcifications, could significantly enhance diagnostic precision, support disease monitoring, and inform treatment planning. Full article
(This article belongs to the Section Neurology)
Show Figures

Figure 1

Figure 1
<p>CBCT image view of the falx cerebri calcification (with the red arrow indicating the calcification). (<b>a</b>) Axial view. (<b>b</b>) Sagittal view. (<b>c</b>) Coronal view.</p>
Full article ">Figure 2
<p>AI-based segmentation of the scanned area (with the red arrow indicating the calcification).</p>
Full article ">Figure 3
<p>Pineal gland calcification (with the red arrow indicating the calcification).</p>
Full article ">
22 pages, 3363 KiB  
Article
New Approaches to AI Methods for Screening Cardiomegaly on Chest Radiographs
by Patrycja S. Matusik, Zbisław Tabor, Iwona Kucybała, Jarosław D. Jarczewski and Tadeusz J. Popiela
Appl. Sci. 2024, 14(24), 11605; https://doi.org/10.3390/app142411605 - 12 Dec 2024
Viewed by 324
Abstract
Background: Cardiothoracic ratio (CTR) and transverse cardiac diameter (TCD) are parameters that are used to assess cardiac size on chest radiographs (CXRs). We aimed to investigate the performance and efficiency of artificial intelligence (AI) in screening for cardiomegaly on CXRs. Methods: The U-net [...] Read more.
Background: Cardiothoracic ratio (CTR) and transverse cardiac diameter (TCD) are parameters that are used to assess cardiac size on chest radiographs (CXRs). We aimed to investigate the performance and efficiency of artificial intelligence (AI) in screening for cardiomegaly on CXRs. Methods: The U-net architecture was designed for lung and heart segmentation. The CTR and TCD were then calculated using these labels and a mathematical algorithm. For the training set, we retrospectively included 65 randomly selected patients who underwent CXRs, while for the testing set, we chose 50 patients who underwent cardiac magnetic resonance (CMR) imaging and had available CXRs in the medical documentation. Results: Using U-net for the training set, the Dice coefficient for the lung was 0.984 ± 0.003 (min. 0.977), while for the heart it was 0.983 ± 0.004 (min. 0.972). For the testing set, the Dice coefficient for the lung was 0.970 ± 0.012 (min. 0.926), while for the heart it was 0.950 ± 0.021 (min. 0.871). The mean CTR and TCD measurements were slightly greater when calculated from either manual or automated segmentation than when manually read. Receiver operating characteristic analyses showed that both the CTR and TCD measurements calculated from either manual or automated segmentation, or when manually read, were good predictors of cardiomegaly diagnosed in CMR. However, McNemar tests have shown that diagnoses made with TCD, rather than CTR, were more consistent with CMR diagnoses. According to a different definition of cardiomegaly based on CMR imaging, accuracy for CTR measurements ranged from 62.0 to 74.0% for automatic segmentation (for TCD it ranged from 64.0 to 72.0%). Conclusion: The use of AI may optimize the screening process for cardiomegaly on CXRs. Future studies should focus on improving the accuracy of AI algorithms and on assessing the usefulness both of CTR and TCD measurements in screening for cardiomegaly. Full article
Show Figures

Figure 1

Figure 1
<p>The cardiothoracic ratio from manual reading measurement, manual segmentation, and automatic segmentation. Box plots show the medians and first and third quartiles in boxes, as well as minimum and maximum values of the cardiothoracic ratio from manual measurement, manual segmentation, and automatic segmentation.</p>
Full article ">Figure 2
<p>Bland–Altman plots of deep learning models based on manual (<b>A</b>) and automatic (<b>B</b>) segmentation compared to manual reading measurements.</p>
Full article ">Figure 3
<p>The transverse cardiac diameter from direct manual measurement, manual segmentation, and automatic segmentation. Box plots show the medians and first and third quartiles in boxes, as well as the minimum and maximum values of the cardiothoracic ratio from manual measurement, manual segmentation, and automatic segmentation. Abbreviations: TCD—transverse cardiac diameter.</p>
Full article ">Figure 4
<p>Bland–Altman plots of deep learning models based on manual (<b>A</b>) and automatic (<b>B</b>) segmentation compared to direct manual measurements.</p>
Full article ">Figure A1
<p>Dice coefficients for lung and heart using U-net for training (<b>A</b>) and testing (<b>B</b>) sets.</p>
Full article ">Figure A2
<p>Example of a chest radiograph showing a good correlation of the automatic measurements with the references. (<b>A</b>)—original chest radiograph; (<b>B</b>)—CTR and TCD estimations from the manual segmentation; (<b>C</b>)—CTR and TCD estimations from the automatic segmentation. Abbreviations: CTR—cardiothoracic ratio; TCD—transverse cardiac diameter.</p>
Full article ">Figure A3
<p>Example of chest radiograph showing a weak correlation of th eautomatic measurements with the references. (<b>A</b>)—original chest radiograph; (<b>B</b>)—CTR estimation from the manual segmentation; (<b>C</b>)—CTR estimation from the automatic segmentation. Abbreviations: CTR—cardiothoracic ratio.</p>
Full article ">Figure A4
<p>Correlations of the cardiothoracic ratio between the direct manual measurements made by radiologists, measurements from the manual segmentation, and from the automatic segmentation.</p>
Full article ">Figure A5
<p>Correlations of transverse cardiac diameter between the direct manual measurements made by radiologists, measurements from the manual segmentation, and from the automatic segmentation.</p>
Full article ">
19 pages, 4573 KiB  
Article
Comparing Different Data Partitioning Strategies for Segmenting Areas Affected by COVID-19 in CT Scans
by Anne de Souza Oliveira, Marly Guimarães Fernandes Costa, João Pedro Guimarães Fernandes Costa and Cícero Ferreira Fernandes Costa Filho
Diagnostics 2024, 14(24), 2791; https://doi.org/10.3390/diagnostics14242791 - 12 Dec 2024
Viewed by 332
Abstract
Background/Objectives: According to the World Health Organization, the gold standard for diagnosing COVID-19 is the Reverse Transcription Polymerase Chain Reaction (RT-PCR) test. However, to confirm the diagnosis in patients who have negative results but still show symptoms, imaging tests, especially computed tomography (CT), [...] Read more.
Background/Objectives: According to the World Health Organization, the gold standard for diagnosing COVID-19 is the Reverse Transcription Polymerase Chain Reaction (RT-PCR) test. However, to confirm the diagnosis in patients who have negative results but still show symptoms, imaging tests, especially computed tomography (CT), are used. In this study, using convolutional neural networks, we compared the following topics using manual and automatic lung segmentation methods: (1) the performance of an automatic segmentation of COVID-19 areas using two strategies for data partitioning, CT scans, and slice strategies; (2) the performance of an automatic segmentation method of COVID-19 when there was interobserver agreement between two groups of radiologists; and (3) the performance of the area affected by COVID-19. Methods: Two datasets and two deep neural network architectures are used to evaluate the automatic segmentation of lungs and COVID-19 areas. The performance of the U-Net architecture is compared with the performance of a new architecture proposed by the research group. Results: With automatic lung segmentation, the Dice metrics for the segmentation of the COVID-19 area were 73.01 ± 9.47% and 84.66 ± 5.41% for the CT-scan strategy and slice strategy, respectively. With manual lung segmentation, the Dice metrics for the automatic segmentation of COVID-19 were 74.47 ± 9.94% and 85.35 ± 5.41% for the CT-scan and the slice strategy, respectively. Conclusions: The main conclusions were as follows: COVID-19 segmentation was slightly better for the slice strategy than for the CT-scan strategy; a comparison of the performance of the automatic COVID-19 segmentation and the interobserver agreement, in a group of 7 CT scans, revealed that there was no statistically significant difference between any metric. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>Graphs showing, for each CT scan of the dataset, the total number of slices, the number of slices containing the lung, and the number of slices containing regions with COVID-19. (<b>a</b>) Dataset 1; (<b>b</b>) Dataset 2. Blue bars: number of slices; yellow bars: number of slices with lung; red bars: number of slices with COVID-19.</p>
Full article ">Figure 2
<p>The steps of the methodology adopted in this work.</p>
Full article ">Figure 3
<p>Steps used in image preprocessing.</p>
Full article ">Figure 4
<p>Examples of slices of CT scans after preprocessing. From top to bottom, a slice, a lung mask, and a COVID-19 mask.</p>
Full article ">Figure 5
<p>Training and test sets obtained with the CT-scan strategy. Different colors represent different CT-scans.</p>
Full article ">Figure 6
<p>Training and test sets obtained with the slice strategy. Different colors represent different CT-scans.</p>
Full article ">Figure 7
<p>Convolutional neural network architecture used for semantic segmentation of lungs and COVID-19—U-Net [<a href="#B29-diagnostics-14-02791" class="html-bibr">29</a>].</p>
Full article ">Figure 8
<p>Convolutional neural network architecture used for semantic segmentation of COVID-19—CNN2.</p>
Full article ">Figure 9
<p>A comparison of the segmented areas of the lungs and COVID-19 patients in test sets, expressed as percentages of the CT-scan area, with the same area segmented by a radiologist: (<b>a</b>) CT-scan strategy; (<b>b</b>) slice strategy.</p>
Full article ">Figure 10
<p>The first row shows the lung and COVID-19 images segmented with U-Net trained with the CT scan strategy. The second row shows the lung and COVID-19 images segmented with U-Net trained with the slice strategy. Radiologist segmentation (Dataset 1) is shown in blue, whereas automatic segmentation is shown in red.</p>
Full article ">Figure 11
<p>(<b>a</b>) Original image; (<b>b</b>) agreement between radiologists. Radiologists’ segmentation in Dataset 1 is shown in blue, whereas radiologists’ segmentation in Dataset 2 is shown in yellow.</p>
Full article ">Figure 12
<p>(<b>a</b>) Original image with low contrast; (<b>b</b>) radiologist segmentation (Dataset 1) is shown in blue, whereas automatic segmentation is shown in red.</p>
Full article ">
Back to TopTop