[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 

Advances in Biomedical Image Processing and Artificial Intelligence for Computer-Aided Diagnosis in Medicine

A special issue of Journal of Imaging (ISSN 2313-433X). This special issue belongs to the section "Medical Imaging".

Deadline for manuscript submissions: 31 March 2025 | Viewed by 25229

Special Issue Editors


E-Mail Website
Guest Editor
1. Ri.MED Foundation, via Bandiera 11, 90133 Palermo, Italy
2. Research Affiliate Long Term, Laboratory of Computational Computer Vision (LCCV), School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA, USA
Interests: biomedical image processing and analysis; radiomics; artificial intelligence; machine learning; deep learning
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Department of Mathematics and Computer Science, University of Cagliari, Via Ospedale 72, 09124 Cagliari, Italy
Interests: computer vision; image retrieval; biomedical image analysis; pattern recognition and machine learning
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Department of Electrical and Electronic Engineering, University of Cagliari, Piazza d’Armi, 09123 Cagliari, Italy
Interests: computer vision; medical image analysis; shape analysis and matching; image retrieval and classification
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Institute of Molecular Bioimaging and Physiology, National Research Council (IBFM-CNR), 90015 Cefalù, Italy
Interests: non-invasive imaging techniques: positron emission tomography (PET), computerized tomography (CT), and magnetic resonance (MR); radiomics and artificial intelligence in clinical health care applications; processing, quantification, and correction methods for ex vivo and in vivo medical images
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues, 

With the digitization of medical data, various artificial intelligence techniques are now being employed. Radiomics and texture analysis are possible through the use of positron emission tomography, computerized tomography, and magnetic resonance imaging. Machine and deep learning techniques can aid in improving therapeutic tools, diagnostic decisions, and rehabilitation. Nevertheless, diagnosing patients has become more difficult due to the abundance of data from different imaging techniques, patient diversity, and the necessity to consider data from various sources, leading to the domain shift problem. Radiologists and pathologists rely on computer-aided diagnosis (CAD) systems to analyze biomedical images and address these challenges. CAD systems help reduce inter- and intra-observer variability, which occurs when different physicians assess the area under the same assumptions or at different times. Additionally, data access can be prevented due to privacy, security, and intellectual property concerns. Synthetic data are increasingly being explored in this context to address these issues. This Special Issue is connected to the 2nd International Workshop on Artificial Intelligence and Radiomics in Computer-Aided Diagnosis (AIRCAD 2023) but is open to additional submissions addressing topics relevant to the Special Issue’s scope. This Special Issue will cover the latest developments in biomedical image processing using machine learning, deep learning, artificial intelligence, and radiomics features, focusing on practical applications and their integration into the medical image processing workflow.

Potential topics include but are not limited to the following: biomedical image processing; machine and deep learning techniques for image analysis (i.e., the segmentation of cells, tissues, organs, and lesions and the classification of cells, diseases, tumors, etc.); image registration techniques; image preprocessing techniques; image-based 3D reconstruction; computer-aided detection and diagnosis systems (CADs); biomedical image analysis; radiomics and artificial intelligence for personalized medicine; machine and deep learning as tools to support medical diagnoses and decisions; image retrieval (e.g., context-based retrieval and lesion similarity); CAD architectures; advanced architectures for biomedical image remote processing, elaboration, and transmission; 3D vision, virtual, augmented, and mixed reality applications for remote surgery; image processing techniques for privacy-preserving AI in medicine.

Dr. Andrea Loddo
Dr. Albert Comelli
Dr. Cecilia Di Ruberto
Dr. Lorenzo Putzu
Dr. Alessandro Stefano
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Journal of Imaging is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • CAD architectures
  • biomedical image processing
  • machine learning
  • deep learning
  • biomedical image analysis
  • radiomics
  • artificial intelligence
  • personalized medicine
  • privacy-preserving AI

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (16 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Review, Other

11 pages, 1525 KiB  
Article
Toward Closing the Loop in Image-to-Image Conversion in Radiotherapy: A Quality Control Tool to Predict Synthetic Computed Tomography Hounsfield Unit Accuracy
by Paolo Zaffino, Ciro Benito Raggio, Adrian Thummerer, Gabriel Guterres Marmitt, Johannes Albertus Langendijk, Anna Procopio, Carlo Cosentino, Joao Seco, Antje Christin Knopf, Stefan Both and Maria Francesca Spadea
J. Imaging 2024, 10(12), 316; https://doi.org/10.3390/jimaging10120316 - 10 Dec 2024
Viewed by 676
Abstract
In recent years, synthetic Computed Tomography (CT) images generated from Magnetic Resonance (MR) or Cone Beam Computed Tomography (CBCT) acquisitions have been shown to be comparable to real CT images in terms of dose computation for radiotherapy simulation. However, until now, there has [...] Read more.
In recent years, synthetic Computed Tomography (CT) images generated from Magnetic Resonance (MR) or Cone Beam Computed Tomography (CBCT) acquisitions have been shown to be comparable to real CT images in terms of dose computation for radiotherapy simulation. However, until now, there has been no independent strategy to assess the quality of each synthetic image in the absence of ground truth. In this work, we propose a Deep Learning (DL)-based framework to predict the accuracy of synthetic CT in terms of Mean Absolute Error (MAE) without the need for a ground truth (GT). The proposed algorithm generates a volumetric map as an output, informing clinicians of the predicted MAE slice-by-slice. A cascading multi-model architecture was used to deal with the complexity of the MAE prediction task. The workflow was trained and tested on two cohorts of head and neck cancer patients with different imaging modalities: 27 MR scans and 33 CBCT. The algorithm evaluation revealed an accurate HU prediction (a median absolute prediction deviation equal to 4 HU for CBCT-based synthetic CTs and 6 HU for MR-based synthetic CTs), with discrepancies that do not affect the clinical decisions made on the basis of the proposed estimation. The workflow exhibited no systematic error in MAE prediction. This work represents a proof of concept about the feasibility of synthetic CT evaluation in daily clinical practice, and it paves the way for future patient-specific quality assessment strategies. Full article
Show Figures

Figure 1

Figure 1
<p>Representation of the general MAE prediction pipeline. An axial sCT slice is given as input, and the associated MAE scalar for the image slice is predicted by using a DL pipeline.</p>
Full article ">Figure 2
<p>A more detailed graphical representation of the MAE prediction pipeline. The final MAE prediction is obtained as a result of two DL steps: First a raw MAE interval classification is performed, followed by a more precise MAE estimation based on a regression algorithm.</p>
Full article ">Figure 3
<p>Exemplary <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>C</mi> <msub> <mi>T</mi> <mrow> <mi>C</mi> <mi>B</mi> <mi>C</mi> <mi>T</mi> </mrow> </msub> </mrow> </semantics></math> overlaid with its <math display="inline"><semantics> <mrow> <mi>p</mi> <mi>M</mi> <mi>A</mi> <msub> <mi>E</mi> <mrow> <mi>v</mi> <mi>o</mi> <mi>l</mi> <mi>u</mi> <mi>m</mi> <mi>e</mi> </mrow> </msub> </mrow> </semantics></math>. In addition to the 2D views (axial, sagittal, and coronal planes), the 3D representation is also shown.</p>
Full article ">Figure 4
<p>Detailed workflow of MAE prediction. A single sCT axial slice is fed firstly into a DL model that classifies it as belonging to a specific MAE class. According to this prediction, the 2D image is then provided as input to a connected DL regression model, specifically trained to operate on a restricted range of MAE values. As a result, the MAE of a single sCT slice can be forecasted. In order to train the different models with a GT MAE, the ground truth CT is needed (dashed lines are needed only to train the models).</p>
Full article ">Figure 5
<p>PD distributions for modality-specific and mixed pipelines. Results for <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>C</mi> <msub> <mi>T</mi> <mrow> <mi>C</mi> <mi>B</mi> <mi>C</mi> <mi>T</mi> </mrow> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>C</mi> <msub> <mi>T</mi> <mrow> <mi>M</mi> <mi>R</mi> </mrow> </msub> </mrow> </semantics></math> are reported, respectively, in the left and in the right panel.</p>
Full article ">Figure 6
<p>APD distributions for modality-specific and mixed pipelines. Results for <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>C</mi> <msub> <mi>T</mi> <mrow> <mi>C</mi> <mi>B</mi> <mi>C</mi> <mi>T</mi> </mrow> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>C</mi> <msub> <mi>T</mi> <mrow> <mi>M</mi> <mi>R</mi> </mrow> </msub> </mrow> </semantics></math> are reported, respectively, in the left and in the right panel.</p>
Full article ">
21 pages, 2595 KiB  
Article
Joint Image Processing with Learning-Driven Data Representation and Model Behavior for Non-Intrusive Anemia Diagnosis in Pediatric Patients
by Tarek Berghout
J. Imaging 2024, 10(10), 245; https://doi.org/10.3390/jimaging10100245 - 2 Oct 2024
Cited by 1 | Viewed by 1376
Abstract
Anemia diagnosis is crucial for pediatric patients due to its impact on growth and development. Traditional methods, like blood tests, are effective but pose challenges, such as discomfort, infection risk, and frequent monitoring difficulties, underscoring the need for non-intrusive diagnostic methods. In light [...] Read more.
Anemia diagnosis is crucial for pediatric patients due to its impact on growth and development. Traditional methods, like blood tests, are effective but pose challenges, such as discomfort, infection risk, and frequent monitoring difficulties, underscoring the need for non-intrusive diagnostic methods. In light of this, this study proposes a novel method that combines image processing with learning-driven data representation and model behavior for non-intrusive anemia diagnosis in pediatric patients. The contributions of this study are threefold. First, it uses an image-processing pipeline to extract 181 features from 13 categories, with a feature-selection process identifying the most crucial data for learning. Second, a deep multilayered network based on long short-term memory (LSTM) is utilized to train a model for classifying images into anemic and non-anemic cases, where hyperparameters are optimized using Bayesian approaches. Third, the trained LSTM model is integrated as a layer into a learning model developed based on recurrent expansion rules, forming a part of a new deep network called a recurrent expansion network (RexNet). RexNet is designed to learn data representations akin to traditional deep-learning methods while also understanding the interaction between dependent and independent variables. The proposed approach is applied to three public datasets, namely conjunctival eye images, palmar images, and fingernail images of children aged up to 6 years. RexNet achieves an overall evaluation of 99.83 ± 0.02% across all classification metrics, demonstrating significant improvements in diagnostic results and generalization compared to LSTM networks and existing methods. This highlights RexNet’s potential as a promising alternative to traditional blood-based methods for non-intrusive anemia diagnosis. Full article
Show Figures

Figure 1

Figure 1
<p>Example photographs of collected data and regions of interest. (<b>a</b>–<b>c</b>) Eye conjunctiva, hand palm, and fingernail images of non-anemic individuals; (<b>d</b>–<b>f</b>) eye conjunctiva, hand palm, and fingernail images of anemic individuals, with green zones highlighting the regions of interest in this study. Adapted from [<a href="#B20-jimaging-10-00245" class="html-bibr">20</a>]: WILEY 2023, under open access license. The figure has been modified for better clarity, including realignment, recoloring of regions of interest, and denoising.</p>
Full article ">Figure 2
<p>Flowchart of the proposed methodology for image processing.</p>
Full article ">Figure 3
<p>Class proportions of image datasets: (<b>a</b>) palmar images; (<b>b</b>) eye conjunctiva images; and (<b>c</b>) fingernail images.</p>
Full article ">Figure 4
<p>Selected feature categories and their proportions based on the feature-extraction process for each dataset. (<b>a</b>,<b>b</b>) Proportions of selected feature categories for the palm dataset; (<b>c</b>,<b>d</b>) proportions of selected feature categories for the eye conjunctiva dataset; and (<b>e</b>,<b>f</b>) proportions of selected feature categories for the Fingernails dataset.</p>
Full article ">Figure 5
<p>Dataset class scatters after image processing. (<b>a</b>) Palmar images; (<b>b</b>) eye conjunctiva images; and (<b>c</b>) fingernail images.</p>
Full article ">Figure 6
<p>Flow diagram of proposed RexNet.</p>
Full article ">Figure 7
<p>Training behavior of the studied approaches. (<b>a</b>) Loss function of RexNet and LSTM for the fingernails dataset; (<b>b</b>) loss function of RexNet and LSTM for the eye conjunctiva dataset; and (<b>c</b>) loss function of RexNet and LSTM for the palmar images dataset.</p>
Full article ">Figure 8
<p>ROC curves behavior of the studied approaches. (<b>a</b>) ROC curves of RexNet and LSTM for the fingernails dataset; (<b>b</b>) ROC curves of RexNet and LSTM for the eye conjunctiva dataset; (<b>c</b>) ROC curves of RexNet and LSTM for the palmar images dataset; and (<b>d</b>) zoomed-in subplot for dataset 2—palm dataset.</p>
Full article ">Figure 9
<p>Confusion matrices for LSTM and RexNet models across different datasets. (<b>a</b>–<b>c</b>) LSTM results on fingernails, palm, and conjunctival eye datasets; (<b>d</b>–<b>f</b>) RexNet results for fingernails, palm, and conjunctival eye datasets.</p>
Full article ">
10 pages, 5992 KiB  
Article
Comparison of Visual and Quantra Software Mammographic Density Assessment According to BI-RADS® in 2D and 3D Images
by Francesca Morciano, Cristina Marcazzan, Rossella Rella, Oscar Tommasini, Marco Conti, Paolo Belli, Andrea Spagnolo, Andrea Quaglia, Stefano Tambalo, Andreea Georgiana Trisca, Claudia Rossati, Francesca Fornasa and Giovanna Romanucci
J. Imaging 2024, 10(9), 238; https://doi.org/10.3390/jimaging10090238 - 23 Sep 2024
Viewed by 815
Abstract
Mammographic density (MD) assessment is subject to inter- and intra-observer variability. An automated method, such as Quantra software, could be a useful tool for an objective and reproducible MD assessment. Our purpose was to evaluate the performance of Quantra software in assessing MD, [...] Read more.
Mammographic density (MD) assessment is subject to inter- and intra-observer variability. An automated method, such as Quantra software, could be a useful tool for an objective and reproducible MD assessment. Our purpose was to evaluate the performance of Quantra software in assessing MD, according to BI-RADS® Atlas Fifth Edition recommendations, verifying the degree of agreement with the gold standard, given by the consensus of two breast radiologists. A total of 5009 screening examinations were evaluated by two radiologists and analysed by Quantra software to assess MD. The agreement between the three assigned values was expressed as intraclass correlation coefficients (ICCs). The agreement between the software and the two readers (R1 and R2) was moderate with ICC values of 0.725 and 0.713, respectively. A better agreement was demonstrated between the software’s assessment and the average score of the values assigned by the two radiologists, with an index of 0.793, which reflects a good correlation. Quantra software appears a promising tool in supporting radiologists in the MD assessment and could be part of a personalised screening protocol soon. However, some fine-tuning is needed to improve its accuracy, reduce its tendency to overestimate, and ensure it excludes high-density structures from its assessment. Full article
Show Figures

Figure 1

Figure 1
<p>Raw images for MD assessment by Quantra Software.</p>
Full article ">Figure 2
<p>RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category A, “almost entirely fat”, by the Quantra 2.2.3 software according to BI-RADS lexicon.</p>
Full article ">Figure 3
<p>RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category B, “scattered fibroglandular”, by the Quantra 2.2.3 software according to BI-RADS lexicon.</p>
Full article ">Figure 4
<p>RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category C, “heterogeneously dense”, by the Quantra 2.2.3 software according to BI-RADS lexicon.</p>
Full article ">Figure 5
<p>RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms assigned to category D, “extremely dense”, by the Quantra 2.2.3 software according to BI-RADS lexicon.</p>
Full article ">Figure 6
<p>Density categories assigned by Quantra Software according to BI-RADS lexicon, fifth edition.</p>
Full article ">Figure 7
<p>RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms of the same patient. The visual assessment made by radiologists for this exam was “B”. Quantra software assigned the highest category, “D”, due to the presence of breast implants, resulting in an incorrect assessment.</p>
Full article ">Figure 8
<p>RCC, LCC, RMLO, and LMLO 2D-synthetic mammograms of the same patient. The visual assessment made by radiologists for this exam was “B”. Quantra software assigned the highest category, “D”, due to the presence of a loop recorder and some macrocalcifications in the left breast, resulting in an incorrect assessment.</p>
Full article ">Chart 1
<p>Graphic illustration of the data reported in <a href="#jimaging-10-00238-t001" class="html-table">Table 1</a> on the frequency of breast density categories assigned by each reader (R1 and R2), the mean of the two readers (R1-R2), and the Quantra software (R3).</p>
Full article ">
23 pages, 5832 KiB  
Article
Enhancing Deep Learning Model Explainability in Brain Tumor Datasets Using Post-Heuristic Approaches
by Konstantinos Pasvantis and Eftychios Protopapadakis
J. Imaging 2024, 10(9), 232; https://doi.org/10.3390/jimaging10090232 - 18 Sep 2024
Viewed by 1264
Abstract
The application of deep learning models in medical diagnosis has showcased considerable efficacy in recent years. Nevertheless, a notable limitation involves the inherent lack of explainability during decision-making processes. This study addresses such a constraint by enhancing the interpretability robustness. The primary focus [...] Read more.
The application of deep learning models in medical diagnosis has showcased considerable efficacy in recent years. Nevertheless, a notable limitation involves the inherent lack of explainability during decision-making processes. This study addresses such a constraint by enhancing the interpretability robustness. The primary focus is directed towards refining the explanations generated by the LIME Library and LIME image explainer. This is achieved through post-processing mechanisms based on scenario-specific rules. Multiple experiments have been conducted using publicly accessible datasets related to brain tumor detection. Our proposed post-heuristic approach demonstrates significant advancements, yielding more robust and concrete results in the context of medical diagnosis. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Proposed methodology.</p>
Full article ">Figure 2
<p>Demonstrating the overlap of the 3 most important segments (<b>right</b>) given an input image (<b>left</b>) and the LIE heatmap (<b>center</b>).</p>
Full article ">Figure 3
<p>Edge Detection example using Laplacian Filter.</p>
Full article ">Figure 4
<p>Edge detection example using sobel filters.</p>
Full article ">Figure 5
<p>Edge detection example using canny filter.</p>
Full article ">Figure 6
<p>Generated binary masks using Li’s and Otsu’s thresholding.</p>
Full article ">Figure 7
<p>An example of tumor coverage calculation.</p>
Full article ">Figure 8
<p>An example of brain coverage calculation.</p>
Full article ">Figure 9
<p>Classification performance scores for the utilized approaches.</p>
Full article ">Figure 10
<p>Tumor Segment Coverage average using Quickshift.</p>
Full article ">Figure 11
<p>Brain Segment Coverage average using Quickshift.</p>
Full article ">Figure 12
<p>Tumor Segment Coverage average using Felzenszwalb.</p>
Full article ">Figure 13
<p>Brain Segment Coverage average using Felzenszwalb.</p>
Full article ">Figure 14
<p>Tumor Segment Coverage Average using Slic.</p>
Full article ">Figure 15
<p>Brain Segment Coverage average using Slic.</p>
Full article ">Figure 16
<p>Improvements over the tumor segment coverage before and after the refinement process.</p>
Full article ">Figure 17
<p>Improvements over the Tumor Segment Coverage before and after the refinement process for images with absolute difference values greater that 0.01.</p>
Full article ">Figure 18
<p>Statistical measurements between techniques, with respective <span class="html-italic">p</span>-values and mean differences.</p>
Full article ">Figure 19
<p>Instances of wrong brain mask production.</p>
Full article ">
13 pages, 1308 KiB  
Article
Decoding Breast Cancer: Using Radiomics to Non-Invasively Unveil Molecular Subtypes Directly from Mammographic Images
by Manon A. G. Bakker, Maria de Lurdes Ovalho, Nuno Matela and Ana M. Mota
J. Imaging 2024, 10(9), 218; https://doi.org/10.3390/jimaging10090218 - 4 Sep 2024
Viewed by 1435
Abstract
Breast cancer is the most commonly diagnosed cancer worldwide. The therapy used and its success depend highly on the histology of the tumor. This study aimed to explore the potential of predicting the molecular subtype of breast cancer using radiomic features extracted from [...] Read more.
Breast cancer is the most commonly diagnosed cancer worldwide. The therapy used and its success depend highly on the histology of the tumor. This study aimed to explore the potential of predicting the molecular subtype of breast cancer using radiomic features extracted from screening digital mammography (DM) images. A retrospective study was performed using the OPTIMAM Mammography Image Database (OMI-DB). Four binary classification tasks were performed: luminal A vs. non-luminal A, luminal B vs. non-luminal B, TNBC vs. non-TNBC, and HER2 vs. non-HER2. Feature selection was carried out by Pearson correlation and LASSO. The support vector machine (SVM) and naive Bayes (NB) ML classifiers were used, and their performance was evaluated with the accuracy and the area under the receiver operating characteristic curve (AUC). A total of 186 patients were included in the study: 58 luminal A, 35 luminal B, 52 TNBC, and 41 HER2. The SVM classifier resulted in AUCs during testing of 0.855 for luminal A, 0.812 for luminal B, 0.789 for TNBC, and 0.755 for HER2, respectively. The NB classifier showed AUCs during testing of 0.714 for luminal A, 0.746 for luminal B, 0.593 for TNBC, and 0.714 for HER2. The SVM classifier outperformed NB with statistical significance for luminal A (p = 0.0268) and TNBC (p = 0.0073). Our study showed the potential of radiomics for non-invasive breast cancer subtype classification. Full article
Show Figures

Figure 1

Figure 1
<p>The in- and exclusion criteria flowchart used during this study.</p>
Full article ">Figure 2
<p>The tumor segmentation process. Starting with normalization of the original DM image, where breast lesions (red) classified as ‘calcification’ underwent image enhancement. Breast lesions classified as ‘mass’ underwent segmentation using a region-growing algorithm. The segmentations were finalized with the use of the image segmenter tool from MATLAB to obtain the final tumor segmentation.</p>
Full article ">Figure 3
<p>An example of image enhancement for a calcification region where (<b>a</b>) is the original DM image and (<b>b</b>) is the enhanced image, making the calcification more pronounced.</p>
Full article ">Figure 4
<p>Examples of breast tumor segmentations for (<b>a</b>) luminal A, (<b>b</b>) luminal B, (<b>c</b>) TNBC, and (<b>d</b>) HER2.</p>
Full article ">Figure 5
<p>The selected radiomic features for (<b>a</b>) luminal A vs. non-luminal A, (<b>b</b>) luminal B vs. non-luminal B, (<b>c</b>) TNBC vs. non-TNBC, and (<b>d</b>) HER2 vs. non-HER2 classification tasks.</p>
Full article ">Figure 6
<p>The ROC curves of the SVM (blue) and NB (yellow) classifiers for (<b>a</b>) luminal A vs. non-luminal A, (<b>b</b>) luminal B vs. non-luminal B, (<b>c</b>) TNBC vs. non-TNBC, and (<b>d</b>) HER2 vs. non-HER2.</p>
Full article ">
29 pages, 4861 KiB  
Article
A New Approach for Effective Retrieval of Medical Images: A Step towards Computer-Assisted Diagnosis
by Suchita Sharma and Ashutosh Aggarwal
J. Imaging 2024, 10(9), 210; https://doi.org/10.3390/jimaging10090210 - 26 Aug 2024
Viewed by 868
Abstract
The biomedical imaging field has grown enormously in the past decade. In the era of digitization, the demand for computer-assisted diagnosis is increasing day by day. The COVID-19 pandemic further emphasized how retrieving meaningful information from medical repositories can aid in improving the [...] Read more.
The biomedical imaging field has grown enormously in the past decade. In the era of digitization, the demand for computer-assisted diagnosis is increasing day by day. The COVID-19 pandemic further emphasized how retrieving meaningful information from medical repositories can aid in improving the quality of patient’s diagnosis. Therefore, content-based retrieval of medical images has a very prominent role in fulfilling our ultimate goal of developing automated computer-assisted diagnosis systems. Therefore, this paper presents a content-based medical image retrieval system that extracts multi-resolution, noise-resistant, rotation-invariant texture features in the form of a novel pattern descriptor, i.e., MsNrRiTxP, from medical images. In the proposed approach, the input medical image is initially decomposed into three neutrosophic images on its transformation into the neutrosophic domain. Afterwards, three distinct pattern descriptors, i.e., MsTrP, NrTxP, and RiTxP, are derived at multiple scales from the three neutrosophic images. The proposed MsNrRiTxP pattern descriptor is obtained by scale-wise concatenation of the joint histograms of MsTrP×RiTxP and NrTxP×RiTxP. To demonstrate the efficacy of the proposed system, medical images of different modalities, i.e., CT and MRI, from four test datasets are considered in our experimental setup. The retrieval performance of the proposed approach is exhaustively compared with several existing, recent, and state-of-the-art local binary pattern-based variants. The retrieval rates obtained by the proposed approach for the noise-free and noisy variants of the test datasets are observed to be substantially higher than the compared ones. Full article
Show Figures

Figure 1

Figure 1
<p>Neutrosophic images of input medical image when transformed into neutrosophic domain: (<b>a</b>) Sample noise-free and noisy medical images, (<b>b</b>) truth image (<math display="inline"><semantics> <msub> <mi>T</mi> <mrow> <mi>N</mi> <mi>S</mi> </mrow> </msub> </semantics></math>), (<b>c</b>) indeterminacy image (<math display="inline"><semantics> <msub> <mi>I</mi> <mrow> <mi>N</mi> <mi>S</mi> </mrow> </msub> </semantics></math>), and (<b>d</b>) falsity image (<math display="inline"><semantics> <msub> <mi>F</mi> <mrow> <mi>N</mi> <mi>S</mi> </mrow> </msub> </semantics></math>).</p>
Full article ">Figure 2
<p>A sample image patch (around a center pixel <math display="inline"><semantics> <msub> <mi>p</mi> <mi>c</mi> </msub> </semantics></math>, highlighted in red) from noise-free and noisy image for illustration of noise robustness of the <math display="inline"><semantics> <msub> <mrow> <mi>T</mi> <mi>r</mi> <mi>P</mi> </mrow> <mi>r</mi> </msub> </semantics></math> pattern. The figure also shows the multi-resolution view of the image patches at four scales <math display="inline"><semantics> <msub> <mi>S</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>2</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math>, corresponding to <math display="inline"><semantics> <mrow> <mi>r</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>2</mn> <mo>,</mo> <mn>3</mn> <mo>,</mo> <mn>4</mn> </mrow> </semantics></math>, respectively.</p>
Full article ">Figure 3
<p>Example illustrating the computation of proposed <math display="inline"><semantics> <msub> <mrow> <mi>T</mi> <mi>r</mi> <mi>P</mi> </mrow> <mi>r</mi> </msub> </semantics></math> pattern for center pixel <math display="inline"><semantics> <msub> <mi>p</mi> <mi>c</mi> </msub> </semantics></math> (highlighted in RED color) at multiples scales <math display="inline"><semantics> <msub> <mi>S</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>2</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math>, corresponding to <math display="inline"><semantics> <mrow> <mi>r</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>2</mn> <mo>,</mo> <mn>3</mn> <mo>,</mo> <mn>4</mn> </mrow> </semantics></math>, respectively, on noise-free and noisy image patch shown in <a href="#jimaging-10-00210-f002" class="html-fig">Figure 2</a>: (<b>a</b>) Neighbor vectors <math display="inline"><semantics> <msub> <mi mathvariant="bold-italic">p</mi> <mi>r</mi> </msub> </semantics></math> at scales <math display="inline"><semantics> <msub> <mi>S</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>2</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math> for noise-free image patch; (<b>b</b>) median quantized neighbor vectors <math display="inline"><semantics> <msub> <mi mathvariant="bold-italic">mqp</mi> <mi>r</mi> </msub> </semantics></math> at scales <math display="inline"><semantics> <msub> <mi>S</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>2</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math> for noise-free image patch; (<b>c</b>) proposed <math display="inline"><semantics> <msub> <mrow> <mi>T</mi> <mi>r</mi> <mi>P</mi> </mrow> <mi>r</mi> </msub> </semantics></math> binary pattern; (<b>d</b>) median quantized neighbor vectors <math display="inline"><semantics> <msub> <mi mathvariant="bold-italic">mqp</mi> <mi>r</mi> </msub> </semantics></math> at scales <math display="inline"><semantics> <msub> <mi>S</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>2</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math> for noisy image patch; (<b>e</b>) neighbor vectors <math display="inline"><semantics> <msub> <mi mathvariant="bold-italic">p</mi> <mi>r</mi> </msub> </semantics></math> at scales <math display="inline"><semantics> <msub> <mi>S</mi> <mn>1</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>2</mn> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>S</mi> <mn>3</mn> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>S</mi> <mn>4</mn> </msub> </semantics></math> for noisy image patch.</p>
Full article ">Figure 4
<p>Sample image from each class of (<b>a</b>) Emphysema CT database, (<b>b</b>) NEMA CT database, (<b>c</b>) OASIS MRI database, and (<b>d</b>) NEMA MRI database.</p>
Full article ">Figure 5
<p>Sample noisy image from each class of (<b>a</b>) Emphysema CT database, (<b>b</b>) NEMA CT database, (<b>c</b>) OASIS MRI database, and (<b>d</b>) NEMA MRI database.</p>
Full article ">Figure 6
<p>Query results of the proposed method for noise-free query images on (<b>a</b>) Emphysema CT database, (<b>b</b>) NEMA CT database, (<b>c</b>) OASIS MRI database, and (<b>d</b>) NEMA MRI database.</p>
Full article ">Figure 7
<p>Query results of the proposed method for noisy query image on (<b>a</b>) Emphysema CT database, (<b>b</b>) NEMA CT database, (<b>c</b>) OASIS MRI database, and (<b>d</b>) NEMA MRI database.</p>
Full article ">Figure 8
<p>Proposed approach’s retrieval performance in comparison to all other methods in terms of <math display="inline"><semantics> <mrow> <mi>a</mi> <mi>v</mi> <mi>g</mi> <mi>P</mi> </mrow> </semantics></math> on noisy and noise-free images of four test datasets.</p>
Full article ">Figure 9
<p>Proposed approach’s retrieval performance in comparison to all other methods in terms of <math display="inline"><semantics> <mrow> <mi>M</mi> <mi>a</mi> <mi>v</mi> <mi>g</mi> <mi>P</mi> </mrow> </semantics></math> on noisy and noise-free images of four test datasets.</p>
Full article ">Figure 10
<p>Proposed approach’s retrieval performance in comparison to all other methods in terms of <math display="inline"><semantics> <mrow> <mi>C</mi> <mi>V</mi> </mrow> </semantics></math> (coefficient of variation) on noisy and noise-free images of four test datasets.</p>
Full article ">
27 pages, 14394 KiB  
Article
Celiac Disease Deep Learning Image Classification Using Convolutional Neural Networks
by Joaquim Carreras
J. Imaging 2024, 10(8), 200; https://doi.org/10.3390/jimaging10080200 - 16 Aug 2024
Cited by 1 | Viewed by 1659
Abstract
Celiac disease (CD) is a gluten-sensitive immune-mediated enteropathy. This proof-of-concept study used a convolutional neural network (CNN) to classify hematoxylin and eosin (H&E) CD histological images, normal small intestine control, and non-specified duodenal inflammation (7294, 11,642, and 5966 images, respectively). The trained network [...] Read more.
Celiac disease (CD) is a gluten-sensitive immune-mediated enteropathy. This proof-of-concept study used a convolutional neural network (CNN) to classify hematoxylin and eosin (H&E) CD histological images, normal small intestine control, and non-specified duodenal inflammation (7294, 11,642, and 5966 images, respectively). The trained network classified CD with high performance (accuracy 99.7%, precision 99.6%, recall 99.3%, F1-score 99.5%, and specificity 99.8%). Interestingly, when the same network (already trained for the 3 class images), analyzed duodenal adenocarcinoma (3723 images), the new images were classified as duodenal inflammation in 63.65%, small intestine control in 34.73%, and CD in 1.61% of the cases; and when the network was retrained using the 4 histological subtypes, the performance was above 99% for CD and 97% for adenocarcinoma. Finally, the model added 13,043 images of Crohn’s disease to include other inflammatory bowel diseases; a comparison between different CNN architectures was performed, and the gradient-weighted class activation mapping (Grad-CAM) technique was used to understand why the deep learning network made its classification decisions. In conclusion, the CNN-based deep neural system classified 5 diagnoses with high performance. Narrow artificial intelligence (AI) is designed to perform tasks that typically require human intelligence, but it operates within limited constraints and is task-specific. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>General design of the convolutional neural network. A convolutional neural network (CNN) is a deep learning algorithm that takes an input image, assigns weights/biases to different components of the image, and classifies all the images. There are three major components of the network: the convolutional layer, the pooling layer, and the fully connected layer.</p>
Full article ">Figure 2
<p>Structure of the convolutional neural network of this study (based on ResNet-18).</p>
Full article ">Figure 3
<p>Characteristic histological images of the small intestine (duodenum). Duodenal control (<b>A</b>); celiac disease (<b>B</b>); inflammatory duodenum (<b>C</b>); duodenal adenocarcinoma (<b>D</b>); Cronh’s disease (<b>E</b>); examples of image cropping at 224-by-224 (224 × 224 × 3) size (F).</p>
Full article ">Figure 4
<p>Images of celiac disease. Celiac disease is a gluten-sensitive immune-mediated enteropathy that occurs in genetically predisposed individuals. Diagnosis of celiac disease is made by combining clinical data, serological tests, and histopathological features. Celiac disease is characterized by mucosal inflammation, intraepithelial lymphocytosis, chronic inflammation in the lamina propria, villous atrophy, and crypt hyperplasia. When the disease persists despite the gluten-free diet, it is called refractory celiac disease, type I (the intraepithelial lymphocytes display normal phenotype, associated with good prognosis when combined with immunosuppressive therapy) and type II (loss of phenotype of the intraepithelial lymphocytes, monoclonal rearrangement of T-cell receptor, and higher risk of developing enteropathy-associated T-cell lymphoma (EATL)) [<a href="#B41-jimaging-10-00200" class="html-bibr">41</a>,<a href="#B42-jimaging-10-00200" class="html-bibr">42</a>,<a href="#B43-jimaging-10-00200" class="html-bibr">43</a>,<a href="#B44-jimaging-10-00200" class="html-bibr">44</a>,<a href="#B45-jimaging-10-00200" class="html-bibr">45</a>,<a href="#B46-jimaging-10-00200" class="html-bibr">46</a>,<a href="#B47-jimaging-10-00200" class="html-bibr">47</a>,<a href="#B48-jimaging-10-00200" class="html-bibr">48</a>,<a href="#B49-jimaging-10-00200" class="html-bibr">49</a>]. The input size is 224-by-224 (224 × 224 × 3). Original magnification 200×.</p>
Full article ">Figure 5
<p>Images of small intestine control. The small intestine is located between the stomach and the large intestine. It is comprised of three parts: the duodenum, jejunum, and ileum. The histological structure has four main layers. The mucosa is the innermost layer that contains epithelium, lamina propria, and muscularis mucosae. The submucosa is a connective tissue layer with blood vessels, lymphatics, and the submucosal plexus. The muscularis externa contains two smooth muscle layers, the inner circular and the outer longitudinal layer. Between them, the myenteric plexus is found. The outermost layer is the adventitia comprised of fibroblasts, collagen, vessels, and nerves. The adventitia is covered by mesothelium (serosa). The small intestine is characterized by high absorption. The mucosa and submucosa form large folds (plicae) in circular manner. The plicae contain microvilli. The epithelium has several components, enterocytes, globet cells, crypts of Lieberkuhm, enteroendocrine cells, and Paneth cells [<a href="#B50-jimaging-10-00200" class="html-bibr">50</a>,<a href="#B51-jimaging-10-00200" class="html-bibr">51</a>]. This figure shows images of the ileum. Additionally, images obtained from the duodenum were included in the dataset. The input size is 224-by-224 (224 × 224 × 3). Original magnification 200×.</p>
Full article ">Figure 6
<p>Images of nonspecific inflammatory small intestine. The input size is 224-by-224 (224 × 224 × 3). Original magnification 200×.</p>
Full article ">Figure 7
<p>Images of duodenal adenocarcinoma. Small bowel adenocarcinomas are histologically very similar to colorectal adenocarcinomas. Adenocarcinomas are characterized by columnar epithelial cells with elongated and pseudostratified nuclei that form a complex glandular structure. There is nuclear polymorphism, loss of epithelial polarity, and necrosis. Inflammatory, autoimmune, genetic, and familiar diseases are common risk factors, including celiac disease, Crohn’s disease, familiar adenomatous polyposis, Peutz–Jeghers syndrome, and Lynch syndrome [<a href="#B53-jimaging-10-00200" class="html-bibr">53</a>,<a href="#B54-jimaging-10-00200" class="html-bibr">54</a>,<a href="#B55-jimaging-10-00200" class="html-bibr">55</a>,<a href="#B56-jimaging-10-00200" class="html-bibr">56</a>,<a href="#B57-jimaging-10-00200" class="html-bibr">57</a>]. The input size is 224-by-224 (224 × 224 × 3). Original magnification 200×.</p>
Full article ">Figure 8
<p>Images of Crohn’s disease. Crohn’s disease is an idiopathic chronic inflammatory condition that can affect both the upper and lower gastrointestinal tract, but usually involves the distal ileum and proximal large intestine. The diagnostic criteria are segmental disease, transmural inflammation, noncaseating granulomas, deep fissuring ulcers, and ileal involvement [<a href="#B58-jimaging-10-00200" class="html-bibr">58</a>,<a href="#B59-jimaging-10-00200" class="html-bibr">59</a>,<a href="#B60-jimaging-10-00200" class="html-bibr">60</a>]. The input size is 224-by-224 (224 × 224 × 3). Original magnification 200×.</p>
Full article ">Figure 9
<p>Training progress of the convolutional neural network for the classification of celiac disease and small intestine control.</p>
Full article ">Figure 10
<p>Confusion matrix of celiac disease and small intestine control. This image shows the confusion matrix of the test set, which includes the analysis of images not previously used in the training and validation steps (i.e., the holdout data). The accuracy of predicting celiac disease was 99.97%.</p>
Full article ">Figure 11
<p>Confusion matrix of celiac disease and small intestine control using different AI models. This image shows the confusion matrices of the test set, which includes the analysis of images not previously used in the training and validation steps (i.e., the holdout data).</p>
Full article ">Figure 12
<p>Confusion matrix of celiac disease, duodenal inflammation, and small intestine control. This image shows the confusion matrix of the test set, which includes the analysis of images not previously used in the training and validation steps (i.e., the holdout data).</p>
Full article ">Figure 13
<p>Confusion matrix of celiac disease, duodenal adenocarcinoma, duodenal inflammation, and small intestine control. This image shows the confusion matrix of the test set, which includes the analysis of images not previously used in the training and validation steps (i.e., the holdout data).</p>
Full article ">Figure 14
<p>Confusion matrix of celiac disease, Crohn’s disease, duodenal adenocarcinoma, duodenal inflammation, and small intestine control. This image shows the confusion matrix of the test set, which includes the analysis of images not previously used in the training and validation steps (i.e., the holdout data).</p>
Full article ">Figure 15
<p>Grad-CAM images of celiac disease. The gradient-weighted class activation mapping (Grad-CAM) visualizes which parts of an image are important to the classification decision of a network. From the histopathological point of view, the Grad-CAM showed that the CNN was focusing on the important parts of the tissue such as the epithelial layer and inflammation.</p>
Full article ">Figure 16
<p>Other Grad-CAM images. The gradient-weighted class activation mapping (Grad-CAM) visualizes which parts of an image are important for the classification decision by the CNN. In this example, the most important part of the image for classification is the epithelial layer.</p>
Full article ">Figure 17
<p>Grad-CAM analysis of discordant cases. The Grad-CAM visualizes which parts of an image are important to the classification decision of a network. Differences were due to images that did not have a clear diagnosis from a histological point of view and/or wrong focus area that was important to the classification of the network. Celiac D. (celiac Disease), Crohn’s (Crohn’s disease), D.Adk (duodenal adenocarcinoma), D.Infla. (duodenal inflammation), Small I.C. (small intestine control).</p>
Full article ">Figure A1
<p>Training progress of the convolutional neural network for the classification of celiac disease, small intestine control, and duodenal inflammation.</p>
Full article ">Figure A2
<p>Training progress of the convolutional neural network for the classification of celiac disease, small intestine control, duodenal inflammation, and duodenal adenocarcinoma.</p>
Full article ">
24 pages, 410 KiB  
Article
Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies
by Andrea Loddo, Marco Usai and Cecilia Di Ruberto
J. Imaging 2024, 10(8), 195; https://doi.org/10.3390/jimaging10080195 - 10 Aug 2024
Cited by 1 | Viewed by 1424
Abstract
Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists’ heavy workloads and the potential for diagnostic errors. Consequently, [...] Read more.
Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists’ heavy workloads and the potential for diagnostic errors. Consequently, there is a pressing need for automated and precise histopathological diagnostic tools. This study leverages Machine Learning and Deep Learning techniques to classify histopathological images into healthy and cancerous categories. By utilizing both handcrafted and deep features and shallow learning classifiers on the GasHisSDB dataset, we conduct a comparative analysis to identify the most effective combinations of features and classifiers for differentiating normal from abnormal histopathological images without employing fine-tuning strategies. Our methodology achieves an accuracy of 95% with the SVM classifier, underscoring the effectiveness of feature fusion strategies. Additionally, cross-magnification experiments produced promising results with accuracies close to 80% and 90% when testing the models on unseen testing images with different resolutions. Full article
Show Figures

Figure 1

Figure 1
<p>Sample images from the GasHisSDB dataset, acquired with the H&amp;E staining method. The hematoxylin is alkaline, and stains cell nuclei a purplish blue, and eosin is acidic and stains the extracellular matrix and cytoplasm pink, with other structures taking on different shades, hues, and combinations of these colors [<a href="#B2-jimaging-10-00195" class="html-bibr">2</a>].</p>
Full article ">
31 pages, 5788 KiB  
Article
Automated Lung Cancer Diagnosis Applying Butterworth Filtering, Bi-Level Feature Extraction, and Sparce Convolutional Neural Network to Luna 16 CT Images
by Nasr Y. Gharaibeh, Roberto De Fazio, Bassam Al-Naami, Abdel-Razzak Al-Hinnawi and Paolo Visconti
J. Imaging 2024, 10(7), 168; https://doi.org/10.3390/jimaging10070168 - 15 Jul 2024
Viewed by 1718
Abstract
Accurate prognosis and diagnosis are crucial for selecting and planning lung cancer treatments. As a result of the rapid development of medical imaging technology, the use of computed tomography (CT) scans in pathology is becoming standard practice. An intricate interplay of requirements and [...] Read more.
Accurate prognosis and diagnosis are crucial for selecting and planning lung cancer treatments. As a result of the rapid development of medical imaging technology, the use of computed tomography (CT) scans in pathology is becoming standard practice. An intricate interplay of requirements and obstacles characterizes computer-assisted diagnosis, which relies on the precise and effective analysis of pathology images. In recent years, pathology image analysis tasks such as tumor region identification, prognosis prediction, tumor microenvironment characterization, and metastasis detection have witnessed the considerable potential of artificial intelligence, especially deep learning techniques. In this context, an artificial intelligence (AI)-based methodology for lung cancer diagnosis is proposed in this research work. As a first processing step, filtering using the Butterworth smooth filter algorithm was applied to the input images from the LUNA 16 lung cancer dataset to remove noise without significantly degrading the image quality. Next, we performed the bi-level feature selection step using the Chaotic Crow Search Algorithm and Random Forest (CCSA-RF) approach to select features such as diameter, margin, spiculation, lobulation, subtlety, and malignancy. Next, the Feature Extraction step was performed using the Multi-space Image Reconstruction (MIR) method with Grey Level Co-occurrence Matrix (GLCM). Next, the Lung Tumor Severity Classification (LTSC) was implemented by using the Sparse Convolutional Neural Network (SCNN) approach with a Probabilistic Neural Network (PNN). The developed method can detect benign, normal, and malignant lung cancer images using the PNN algorithm, which reduces complexity and efficiently provides classification results. Performance parameters, namely accuracy, precision, F-score, sensitivity, and specificity, were determined to evaluate the effectiveness of the implemented hybrid method and compare it with other solutions already present in the literature. Full article
Show Figures

Figure 1

Figure 1
<p>The overall architecture of the proposed algorithm for lung tumor detection.</p>
Full article ">Figure 2
<p>Unprocessed (input) and filtered (output) images obtained by the Butterworth high-pass smooth filter: for a normal case (<b>a</b>), for a benign lung-cancer pathology (<b>b</b>), and in the case of a malignant lung cancer (<b>c</b>).</p>
Full article ">Figure 2 Cont.
<p>Unprocessed (input) and filtered (output) images obtained by the Butterworth high-pass smooth filter: for a normal case (<b>a</b>), for a benign lung-cancer pathology (<b>b</b>), and in the case of a malignant lung cancer (<b>c</b>).</p>
Full article ">Figure 3
<p>Architecture of the Sparse Convolutional Neural Network.</p>
Full article ">Figure 4
<p>Structure of the developed PNN.</p>
Full article ">Figure 5
<p>Obtained accuracy for DenseNet201, CNN + SVM, and the proposed model as a function of epochs.</p>
Full article ">Figure 6
<p>Obtained precision for DenseNet201, CNN + SVM, and the proposed model as a function of epochs.</p>
Full article ">Figure 7
<p>F1-score of DenseNet201, CNN + SVM, and proposed model as a function of epochs.</p>
Full article ">Figure 8
<p>ROC curve of the DensNet201, CNN + SVM, and proposed model as a function of epochs.</p>
Full article ">Figure 9
<p>Comparison between the algorithms proposed in Ref. [<a href="#B38-jimaging-10-00168" class="html-bibr">38</a>] on the left (<b>a</b>) and that proposed in this research work on the right (<b>b</b>); the dashed vertical line separates the two simplified schemes for clarity.</p>
Full article ">
21 pages, 1918 KiB  
Article
Residual-Based Multi-Stage Deep Learning Framework for Computer-Aided Alzheimer’s Disease Detection
by Najmul Hassan, Abu Saleh Musa Miah and Jungpil Shin
J. Imaging 2024, 10(6), 141; https://doi.org/10.3390/jimaging10060141 - 11 Jun 2024
Viewed by 1894
Abstract
Alzheimer’s Disease (AD) poses a significant health risk globally, particularly among the elderly population. Recent studies underscore its prevalence, with over 50% of elderly Japanese facing a lifetime risk of dementia, primarily attributed to AD. As the most prevalent form of dementia, AD [...] Read more.
Alzheimer’s Disease (AD) poses a significant health risk globally, particularly among the elderly population. Recent studies underscore its prevalence, with over 50% of elderly Japanese facing a lifetime risk of dementia, primarily attributed to AD. As the most prevalent form of dementia, AD gradually erodes brain cells, leading to severe neurological decline. In this scenario, it is important to develop an automatic AD-detection system, and many researchers have been working to develop an AD-detection system by taking advantage of the advancement of deep learning (DL) techniques, which have shown promising results in various domains, including medical image analysis. However, existing approaches for AD detection often suffer from limited performance due to the complexities associated with training hierarchical convolutional neural networks (CNNs). In this paper, we introduce a novel multi-stage deep neural network architecture based on residual functions to address the limitations of existing AD-detection approaches. Inspired by the success of residual networks (ResNets) in image-classification tasks, our proposed system comprises five stages, each explicitly formulated to enhance feature effectiveness while maintaining model depth. Following feature extraction, a deep learning-based feature-selection module is applied to mitigate overfitting, incorporating batch normalization, dropout and fully connected layers. Subsequently, machine learning (ML)-based classification algorithms, including Support Vector Machines (SVM), Random Forest (RF) and SoftMax, are employed for classification tasks. Comprehensive evaluations conducted on three benchmark datasets, namely ADNI1: Complete 1Yr 1.5T, MIRAID and OASIS Kaggle, demonstrate the efficacy of our proposed model. Impressively, our model achieves accuracy rates of 99.47%, 99.10% and 99.70% for ADNI1: Complete 1Yr 1.5T, MIRAID and OASIS datasets, respectively, outperforming existing systems in binary class problems. Our proposed model represents a significant advancement in the AD-analysis domain. Full article
Show Figures

Figure 1

Figure 1
<p>Visual information of brain MRIs: (<b>a</b>) Different view of MRI imaging plans, (<b>b</b>) Normal brains and affected by AD.</p>
Full article ">Figure 2
<p>Outline of the proposed model.</p>
Full article ">Figure 3
<p>Proposed Residual-Based Multi-Stage Deep Learning (RBMSDL) model architecture.</p>
Full article ">Figure 4
<p>(<b>a</b>) Enhanced module, (<b>b</b>) residual module.</p>
Full article ">Figure 5
<p>Accuracy curves of the proposed RBMSDL model with ADNI-1 and MIRIAD dataset.</p>
Full article ">Figure 6
<p>Loss curves of the proposed RBMSDL model with ADNI-1 and MIRAID dataset.</p>
Full article ">Figure 7
<p>Accuracy and loss curves for our proposed RBMSDL model with OASIS dataset.</p>
Full article ">Figure 8
<p>Confusion matrix of the proposed RBMSDL model with ADNI and MIRAID dataset.</p>
Full article ">Figure 9
<p>ROC curves of the proposed RBMSDL model. The diagonal (red) line indicates random chance, representing a classifier that has no discriminative power between the positive and negative classes.</p>
Full article ">Figure 10
<p>Confusion matrix and ROC curve of the proposed RBMSDL model with the OASIS dataset. The diagonal (red) line indicates random chance, representing a classifier that has no discriminative power between the positive and negative classes.</p>
Full article ">
21 pages, 12872 KiB  
Article
Optimizing Vision Transformers for Histopathology: Pretraining and Normalization in Breast Cancer Classification
by Giulia Lucrezia Baroni, Laura Rasotto, Kevin Roitero, Angelica Tulisso, Carla Di Loreto and Vincenzo Della Mea
J. Imaging 2024, 10(5), 108; https://doi.org/10.3390/jimaging10050108 - 30 Apr 2024
Cited by 3 | Viewed by 2400
Abstract
This paper introduces a self-attention Vision Transformer model specifically developed for classifying breast cancer in histology images. We examine various training strategies and configurations, including pretraining, dimension resizing, data augmentation and color normalization strategies, patch overlap, and patch size configurations, in order to [...] Read more.
This paper introduces a self-attention Vision Transformer model specifically developed for classifying breast cancer in histology images. We examine various training strategies and configurations, including pretraining, dimension resizing, data augmentation and color normalization strategies, patch overlap, and patch size configurations, in order to evaluate their impact on the effectiveness of the histology image classification. Additionally, we provide evidence for the increase in effectiveness gathered through geometric and color data augmentation techniques. We primarily utilize the BACH dataset to train and validate our methods and models, but we also test them on two additional datasets, BRACS and AIDPATH, to verify their generalization capabilities. Our model, developed from a transformer pretrained on ImageNet, achieves an accuracy rate of 0.91 on the BACH dataset, 0.74 on the BRACS dataset, and 0.92 on the AIDPATH dataset. Using a model based on the prostate small and prostate medium HistoEncoder models, we achieve accuracy rates of 0.89 and 0.86, respectively. Our results suggest that pretraining on large-scale general datasets like ImageNet is advantageous. We also show the potential benefits of using domain-specific pretraining datasets, such as extensive histopathological image collections as in HistoEncoder, though not yet with clear advantages. Full article
Show Figures

Figure 1

Figure 1
<p>On the left side, an in situ carcinoma image from the training set. On the right side, two sample patches in the two sizes used in our experiments.</p>
Full article ">Figure 2
<p>Example of a BRACS WSI with annotations and its detailed RoI adapted from Brancati et al. [<a href="#B42-jimaging-10-00108" class="html-bibr">42</a>].</p>
Full article ">Figure 3
<p>Overview of the proposed self-attention ViT model for classifying breast cancer in histopathology images.</p>
Full article ">Figure 4
<p>Confusion matrices for all 5 folds. (<b>a</b>) Confusion Matrix Fold 1. (<b>b</b>) Confusion Matrix Fold 2. (<b>c</b>) Confusion Matrix Fold 3. (<b>d</b>) Confusion Matrix Fold 4. (<b>e</b>) Confusion Matrix Fold 5.</p>
Full article ">Figure 5
<p>Probability distributions for Fold 2. (<b>a</b>) Probability distribution for class normal. (<b>b</b>) Probability distribution for class benign. (<b>c</b>) Probability distribution for class in situ. (<b>d</b>) Probability distribution for class invasive.</p>
Full article ">
14 pages, 5416 KiB  
Article
Attention-Enhanced Unpaired xAI-GANs for Transformation of Histological Stain Images
by Tibor Sloboda, Lukáš Hudec, Matej Halinkovič and Wanda Benesova
J. Imaging 2024, 10(2), 32; https://doi.org/10.3390/jimaging10020032 - 25 Jan 2024
Cited by 1 | Viewed by 2167
Abstract
Histological staining is the primary method for confirming cancer diagnoses, but certain types, such as p63 staining, can be expensive and potentially damaging to tissues. In our research, we innovate by generating p63-stained images from H&E-stained slides for metaplastic breast cancer. This is [...] Read more.
Histological staining is the primary method for confirming cancer diagnoses, but certain types, such as p63 staining, can be expensive and potentially damaging to tissues. In our research, we innovate by generating p63-stained images from H&E-stained slides for metaplastic breast cancer. This is a crucial development, considering the high costs and tissue risks associated with direct p63 staining. Our approach employs an advanced CycleGAN architecture, xAI-CycleGAN, enhanced with context-based loss to maintain structural integrity. The inclusion of convolutional attention in our model distinguishes between structural and color details more effectively, thus significantly enhancing the visual quality of the results. This approach shows a marked improvement over the base xAI-CycleGAN and standard CycleGAN models, offering the benefits of a more compact network and faster training even with the inclusion of attention. Full article
Show Figures

Figure 1

Figure 1
<p>Demonstration of significant differences in tissue in paired and aligned p63 stained tissue (<b>left</b>) compared with its H&amp;E counterpart (<b>right</b>) [<a href="#B9-jimaging-10-00032" class="html-bibr">9</a>].</p>
Full article ">Figure 2
<p>Demonstration of the context loss computation in one direction. <math display="inline"><semantics> <msup> <mi>G</mi> <mrow> <mi>A</mi> <mi>B</mi> </mrow> </msup> </semantics></math> represents the H&amp;E to p63 conversion and <math display="inline"><semantics> <msup> <mi>G</mi> <mrow> <mi>B</mi> <mi>A</mi> </mrow> </msup> </semantics></math> represents the p63 to H&amp;E conversion. <math display="inline"><semantics> <msup> <mi>B</mi> <mo>′</mo> </msup> </semantics></math> is the converted (fake) p63 image and <math display="inline"><semantics> <msup> <mi>A</mi> <mi>c</mi> </msup> </semantics></math> is the cycled H&amp;E image. <span class="html-italic">H</span> represents Huber Loss.</p>
Full article ">Figure 3
<p>Results show that the FID of the enhanced xAI-CycleGAN is significantly lower than that of the original, demonstrating significant improvement over the previous method.</p>
Full article ">Figure 4
<p>Comparing conversion from p63 to H&amp;E for both original and improved xAI-CycleGAN at the same point in training, demonstrating that the issue with corruption/artifacts in conversion is solved. (<b>a</b>) Demonstration of the corrupted conversion present in original xAI-CycleGAN due to explainability-driven training. (<b>b</b>) Demonstration of a clean proper conversion with our enhanced model, using the same test image, at the same point during training (the model has seen the same amount of training samples).</p>
Full article ">Figure 5
<p>A demonstration of successful editing capabilities. (<b>A</b>) contains image converted from H&amp;E to p63 without any modifications applied. (<b>B</b>) contains an image with modifications applied to best match the real image. (<b>C</b>) contains the unmodified original p63 image of the same region. The same base image as in our previous work [<a href="#B9-jimaging-10-00032" class="html-bibr">9</a>] has been used.</p>
Full article ">Figure 6
<p>Demonstration of conversion by our model from H&amp;E to p63 where no myoepithelial cells are present. (<b>a</b>) Raw unedited H&amp;E that has no myoepithelial cells present. (<b>b</b>) Edited p63 image after correct transformation to include brown coloring by p63 for myoepithelial cells, which is not meant to be present in the image. For this image, we used <math display="inline"><semantics> <mrow> <mi>r</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math> with <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mo>[</mo> <mo>−</mo> <mn>2</mn> <mo>,</mo> <mo>−</mo> <mn>1.6</mn> <mo>,</mo> <mn>0.2</mn> <mo>]</mo> </mrow> </semantics></math> for each vector, respectively.</p>
Full article ">Figure 7
<p>A grid of edited images with varying <span class="html-italic">r</span> from 1 to 5, and <math display="inline"><semantics> <mi>α</mi> </semantics></math> values for columns ranging from −5.0 to 5.0 with a step size of 2.5 (applied to all matrices <span class="html-italic">V</span> defined in Equation (<a href="#FD8-jimaging-10-00032" class="html-disp-formula">8</a>)).</p>
Full article ">

Review

Jump to: Research, Other

54 pages, 5089 KiB  
Review
The Neural Frontier of Future Medical Imaging: A Review of Deep Learning for Brain Tumor Detection
by Tarek Berghout
J. Imaging 2025, 11(1), 2; https://doi.org/10.3390/jimaging11010002 - 24 Dec 2024
Viewed by 433
Abstract
Brain tumor detection is crucial in medical research due to high mortality rates and treatment challenges. Early and accurate diagnosis is vital for improving patient outcomes, however, traditional methods, such as manual Magnetic Resonance Imaging (MRI) analysis, are often time-consuming and error-prone. The [...] Read more.
Brain tumor detection is crucial in medical research due to high mortality rates and treatment challenges. Early and accurate diagnosis is vital for improving patient outcomes, however, traditional methods, such as manual Magnetic Resonance Imaging (MRI) analysis, are often time-consuming and error-prone. The rise of deep learning has led to advanced models for automated brain tumor feature extraction, segmentation, and classification. Despite these advancements, comprehensive reviews synthesizing recent findings remain scarce. By analyzing over 100 research papers over past half-decade (2019–2024), this review fills that gap, exploring the latest methods and paradigms, summarizing key concepts, challenges, datasets, and offering insights into future directions for brain tumor detection using deep learning. This review also incorporates an analysis of previous reviews and targets three main aspects: feature extraction, segmentation, and classification. The results revealed that research primarily focuses on Convolutional Neural Networks (CNNs) and their variants, with a strong emphasis on transfer learning using pre-trained models. Other methods, such as Generative Adversarial Networks (GANs) and Autoencoders, are used for feature extraction, while Recurrent Neural Networks (RNNs) are employed for time-sequence modeling. Some models integrate with Internet of Things (IoT) frameworks or federated learning for real-time diagnostics and privacy, often paired with optimization algorithms. However, the adoption of eXplainable AI (XAI) remains limited, despite its importance in building trust in medical diagnostics. Finally, this review outlines future opportunities, focusing on image quality, underexplored deep learning techniques, expanding datasets, and exploring deeper learning representations and model behavior such as recurrent expansion to advance medical imaging diagnostics. Full article
Show Figures

Figure 1

Figure 1
<p>Illustration of changes or symptoms prior to brain tumor diagnosis. This figure, reproduced from [<a href="#B14-jimaging-11-00002" class="html-bibr">14</a>] under an open-access license permitting non-commercial use, has been edited using a curves adjustment layer. This adjustment was applied to manipulate the visual properties, enhancing clarity and detail.</p>
Full article ">Figure 2
<p>Key aspects of brain tumors: challenges, causes, treatments, and diagnostic methods.</p>
Full article ">Figure 3
<p>Distribution of research papers by type: (<b>a</b>) pie chart showing the percentage of publications by type over half a decade; (<b>b</b>) number of research papers per year by type.</p>
Full article ">Figure 4
<p>Simplified diagram of the papers framework for reviewing brain tumor detection related works: incorporating related review analyses, feature extraction, segmentation, and classification using medical images and deep learning, with future opportunities.</p>
Full article ">Figure 5
<p>Examples of healthy and cancerous brain images: (<b>a</b>) healthy brain; (<b>b</b>) glioma tumor; (<b>c</b>) meningioma tumor; (<b>d</b>) pituitary tumor.</p>
Full article ">Figure 6
<p>Examples of extracted features with an autoencoder from healthy and cancerous brain images: (<b>a</b>) healthy brain; (<b>b</b>) glioma tumor; (<b>c</b>) meningioma tumor; (<b>d</b>) pituitary tumor.</p>
Full article ">Figure 7
<p>Examples of data scatters of (<b>a</b>) original data and (<b>b</b>) extracted features with an autoencoder from healthy and cancerous brain images.</p>
Full article ">Figure 8
<p>(<b>a</b>–<b>c</b>) Summary of feature extraction methods, paradigms, and datasets.</p>
Full article ">Figure 9
<p>Example of MRI brain image segmentation: (<b>a</b>) original MRI brain image; (<b>b</b>–<b>d</b>) ResNet-18 activations at layers 5, 10, and 20 highlighting key segmentation areas.</p>
Full article ">Figure 10
<p>Overview and distribution of methods used for brain MRI image segmentation: (<b>a</b>) bar chart depicting the frequency of methods employed; (<b>b</b>) bar chart categorizing the learning paradigms; (<b>c</b>) bar chart illustrating the distribution of datasets used.</p>
Full article ">Figure 11
<p>Visualization of data distributions from validation images and ResNet mappings: (<b>a</b>) scatter plot of feature distributions extracted directly from the original validation images; (<b>b</b>) scatter plot of feature distributions obtained from ResNet layer activations, illustrating the network’s learned representations.</p>
Full article ">Figure 12
<p>Comprehensive analysis of brain tumor detection and classification techniques: (<b>a</b>) methods; (<b>b</b>) learning paradigms; and (<b>c</b>) datasets.</p>
Full article ">Figure 13
<p>Diagram of proposed future directions for brain tumors detection with deep learning image classification.</p>
Full article ">
35 pages, 7878 KiB  
Review
Advances in Real-Time 3D Reconstruction for Medical Endoscopy
by Alexander Richter, Till Steinmann, Jean-Claude Rosenthal and Stefan J. Rupitsch
J. Imaging 2024, 10(5), 120; https://doi.org/10.3390/jimaging10050120 - 14 May 2024
Viewed by 3343
Abstract
This contribution is intended to provide researchers with a comprehensive overview of the current state-of-the-art concerning real-time 3D reconstruction methods suitable for medical endoscopy. Over the past decade, there have been various technological advancements in computational power and an increased research effort in [...] Read more.
This contribution is intended to provide researchers with a comprehensive overview of the current state-of-the-art concerning real-time 3D reconstruction methods suitable for medical endoscopy. Over the past decade, there have been various technological advancements in computational power and an increased research effort in many computer vision fields such as autonomous driving, robotics, and unmanned aerial vehicles. Some of these advancements can also be adapted to the field of medical endoscopy while coping with challenges such as featureless surfaces, varying lighting conditions, and deformable structures. To provide a comprehensive overview, a logical division of monocular, binocular, trinocular, and multiocular methods is performed and also active and passive methods are distinguished. Within these categories, we consider both flexible and non-flexible endoscopes to cover the state-of-the-art as fully as possible. The relevant error metrics to compare the publications presented here are discussed, and the choice of when to choose a GPU rather than an FPGA for camera-based 3D reconstruction is debated. We elaborate on the good practice of using datasets and provide a direct comparison of the presented work. It is important to note that in addition to medical publications, publications evaluated on the KITTI and Middlebury datasets are also considered to include related methods that may be suited for medical 3D reconstruction. Full article
Show Figures

Figure 1

Figure 1
<p>Exemplary depiction of a laparoscopic MIS, also referred to as keyhole surgery, using an endoscope.</p>
Full article ">Figure 2
<p>Overview of the state-of-the-art real-time camera-based acquisition systems for 3D reconstruction that are discussed in this contribution.</p>
Full article ">Figure 3
<p>Illustration of accuracy vs. precision, where the bullseye represents the true value that is expected, while black dots represent measurements, hence throwing results.</p>
Full article ">Figure 4
<p>Visual representation of commonly used error metrics when comparing a method’s performance against ground truth data of a dataset. Where DEV is the deviation to the ground truth, and MAE and RMSE as defined in <a href="#sec2dot2-jimaging-10-00120" class="html-sec">Section 2.2</a>.</p>
Full article ">Figure 5
<p>Historic timeline of datasets for 3D reconstruction.</p>
Full article ">Figure 6
<p>Example image from the SCARED dataset, along with the corresponding depth map [<a href="#B15-jimaging-10-00120" class="html-bibr">15</a>].</p>
Full article ">Figure 7
<p>3D reconstruction of an in vivo video sequence from a monocular laparoscope, using the quasi-conformal method as presented by Malti et al. [<a href="#B64-jimaging-10-00120" class="html-bibr">64</a>].</p>
Full article ">Figure 8
<p>Intra-operative example of a measurement during the 3D reconstruction of an in vivo video sequence. The side-view of the intra-operative measurement example on the right shows that the measurement line closely follows the surface curvature. Reprinted with permission from Chen et al. [<a href="#B52-jimaging-10-00120" class="html-bibr">52</a>]. Copyright 2024 Elsevier.</p>
Full article ">Figure 9
<p>Porcine large bowel under Structured Light (SL). Reprinted with permission from Lin et al. [<a href="#B58-jimaging-10-00120" class="html-bibr">58</a>]. Copyright 2024 Springer Nature.</p>
Full article ">Figure 10
<p>Presentation of the calibration process of the SL method. The depicted lamb trachea was examined in an experiment. The resulting 3D reconstruction is shown on the right. The missing area at the top is caused by the camera connection cable. Reprinted with permission from Schmalz et al. [<a href="#B56-jimaging-10-00120" class="html-bibr">56</a>]. Copyright 2024 Elsevier.</p>
Full article ">Figure 11
<p>The ToF method uses the phase shift between the emitted and received light pulses to calculate the distance to an object. To determine the distance the phase shift is multiplied by the speed of light and then divided by four times pi, multiplied by the modulation frequency of the emitted light pulse.</p>
Full article ">Figure 12
<p>Schematic diagram of a laser distance sensing system based on the Michelson interferometer. The green path depicts the known reference path and the red path length is dependent on the distance to the sample object. The resulting interference on the detector is recorded, and a Fourier transform from the resulting interferogram leads to a spectrum that correlates to the measured depth.</p>
Full article ">Figure 13
<p>Stereo matching takes advantage of the information provided by two images of the same scene. To determine the 3D position of an object, the corresponding location is identified in both images by either using a correlation-based or a feature-based matching approach. With a known baseline, the resulting disparity can be used to triangulate the 3D position.</p>
Full article ">Figure 14
<p>Example of a 3D point cloud from different perspectives generated by a deterministic stereo matching algorithm using images from the SCARED dataset [<a href="#B15-jimaging-10-00120" class="html-bibr">15</a>] as input.</p>
Full article ">Figure 15
<p>Asimplified structure of a deep learning network such as a CNN. The blue input layers accept the pixels of the stereo images as input. Colored in green are the hidden layers that perform a combination of convolutional operations on the information passed from the blue input layers. To derive the disparity map, the yellow output layer weights the information received by the previous layer.</p>
Full article ">Figure 16
<p>The mosaicked 3D point cloud of a pig stomach obtained by SGBM on the <b>left</b>, and by StereoNet presented by Huo et al. [<a href="#B132-jimaging-10-00120" class="html-bibr">132</a>] on the <b>right</b>. Red rectangles indicate areas with outliers in the point cloud that affect the final stitching results due to a rough surface.</p>
Full article ">Figure 17
<p>Schematic of a trinocular endoscope observing organ tissue. The dashed lines represent the line of sight for each camera. Stereo matching is performed between all possible camera pairs in order to derive a 3D reconstruction.</p>
Full article ">

Other

Jump to: Research, Review

10 pages, 1746 KiB  
Technical Note
MOTH: Memory-Efficient On-the-Fly Tiling of Histological Image Annotations Using QuPath
by Thomas Kauer, Jannik Sehring, Kai Schmid, Marek Bartkuhn, Benedikt Wiebach, Slaven Crnkovic, Grazyna Kwapiszewska, Till Acker and Daniel Amsel
J. Imaging 2024, 10(11), 292; https://doi.org/10.3390/jimaging10110292 - 15 Nov 2024
Viewed by 775
Abstract
The emerging usage of digitalized histopathological images is leading to a novel possibility for data analysis. With the help of artificial intelligence algorithms, it is now possible to detect certain structures and morphological features on whole slide images automatically. This enables algorithms to [...] Read more.
The emerging usage of digitalized histopathological images is leading to a novel possibility for data analysis. With the help of artificial intelligence algorithms, it is now possible to detect certain structures and morphological features on whole slide images automatically. This enables algorithms to count, measure, or evaluate those areas when trained properly. To achieve suitable training, datasets must be annotated and curated by users in programs like QuPath. The extraction of this data for artificial intelligence algorithms is still rather tedious and needs to be saved on a local hard drive. We developed a toolkit for integration into existing pipelines and tools, like U-net, for the on-the-fly extraction of annotation tiles from existing QuPath projects. The tiles can be directly used as input for artificial intelligence algorithms, and the results are directly transferred back to QuPath for visual inspection. With the toolkit, we created a convenient way to incorporate QuPath into existing AI workflows. Full article
Show Figures

Figure 1

Figure 1
<p>MOTH overview. MOTH is a suite of tools that facilitates the import and export of annotations and images from and into QuPath. The system is capable of establishing a connection to local AI-based algorithms.</p>
Full article ">Figure 2
<p>(<b>A</b>,<b>B</b>) IoU and HD of exported shapes rendered with MOTH and Groovy in the artificial dataset. (<b>C</b>,<b>D</b>) IoU and HD of exported shapes rendered with MOTH and Groovy in the mitosis dataset. Groovy results are marked in orange and MOTH results are marked in green. Diamonds represent outliers.</p>
Full article ">Figure 3
<p>MOTH export of small shapes with pixel offsets. The figure shows the export of small ground truth shapes. The ground truth shapes are drawn as orange lines and the center of the shape is marked by an orange dot. Black areas are pixels set in the MOTH export. A high overlap with the ground truth shapes can be observed.</p>
Full article ">Figure 4
<p>Groovy export of small shapes with pixel offsets. The figure shows the export of small ground truth shapes. In comparison to the previous figure, a lower overlap between the ground truth and the export is visible.</p>
Full article ">Figure 5
<p>Real world example using MOTH. The proposals are generated via QuPath and extracted from the project via MOTH. The proposals are evaluated and improved via custom methods and loaded back into QuPath for visual inspection using MOTH.</p>
Full article ">
22 pages, 604 KiB  
Systematic Review
The Accuracy of Three-Dimensional Soft Tissue Simulation in Orthognathic Surgery—A Systematic Review
by Anna Olejnik, Laurence Verstraete, Tomas-Marijn Croonenborghs, Constantinus Politis and Gwen R. J. Swennen
J. Imaging 2024, 10(5), 119; https://doi.org/10.3390/jimaging10050119 - 14 May 2024
Cited by 1 | Viewed by 1486
Abstract
Three-dimensional soft tissue simulation has become a popular tool in the process of virtual orthognathic surgery planning and patient–surgeon communication. To apply 3D soft tissue simulation software in routine clinical practice, both qualitative and quantitative validation of its accuracy are required. The objective [...] Read more.
Three-dimensional soft tissue simulation has become a popular tool in the process of virtual orthognathic surgery planning and patient–surgeon communication. To apply 3D soft tissue simulation software in routine clinical practice, both qualitative and quantitative validation of its accuracy are required. The objective of this study was to systematically review the literature on the accuracy of 3D soft tissue simulation in orthognathic surgery. The Web of Science, PubMed, Cochrane, and Embase databases were consulted for the literature search. The systematic review (SR) was conducted according to the PRISMA statement, and 40 articles fulfilled the inclusion and exclusion criteria. The Quadas-2 tool was used for the risk of bias assessment for selected studies. A mean error varying from 0.27 mm to 2.9 mm for 3D soft tissue simulations for the whole face was reported. In the studies evaluating 3D soft tissue simulation accuracy after a Le Fort I osteotomy only, the upper lip and paranasal regions were reported to have the largest error, while after an isolated bilateral sagittal split osteotomy, the largest error was reported for the lower lip and chin regions. In the studies evaluating simulation after bimaxillary osteotomy with or without genioplasty, the highest inaccuracy was reported at the level of the lips, predominantly the lower lip, chin, and, sometimes, the paranasal regions. Due to the variability in the study designs and analysis methods, a direct comparison was not possible. Therefore, based on the results of this SR, guidelines to systematize the workflow for evaluating the accuracy of 3D soft tissue simulations in orthognathic surgery in future studies are proposed. Full article
Show Figures

Figure 1

Figure 1
<p>PRISMA 2020 flow diagram.</p>
Full article ">
Back to TopTop