Enhancing Early Breast Cancer Detection with Infrared Thermography: A Comparative Evaluation of Deep Learning and Machine Learning Models
<p>Proposed Framework for Feature Extraction and Classification from Thermal Breast Images.</p> "> Figure 2
<p>Sample of Thermal Images from Dataset.</p> "> Figure 3
<p>Example of a Full-Body Thermal Image Capturing the Breast Area.</p> "> Figure 4
<p>Effects of the preprocessing filters applied to infrared images.</p> "> Figure 5
<p>Distribution of pixel intensities on the real-world vs. augmented data.</p> "> Figure 6
<p>Workflow of 10-fold cross-validation implementation.</p> "> Figure 7
<p>PCA Visualization of Thermal Image Features for Breast Cancer Detection.</p> "> Figure 8
<p>Feature Correlation Heatmap for Thermal Image Dataset.</p> "> Figure 9
<p>Top 10 Most Important Features from the Thermal Images for Breast Cancer Detection.</p> "> Figure 10
<p>Confusion Matrix for SVM with ResNet-152 Features.</p> "> Figure 11
<p>Precision–Recall Curve of the Model.</p> "> Figure 12
<p>The ROC curve of the Model.</p> "> Figure 13
<p>Accuracy Comparison of Classifiers across Feature Models.</p> "> Figure 14
<p>AUC Comparison of Classifiers across Feature Models for Breast Cancer Classification.</p> "> Figure 15
<p>Grad-CAM Overlay for Normal Class (the original thermal image (<b>left</b>) alongside the Grad-CAM overlay (<b>right</b>) highlights the regions contributing to the model’s prediction of the “Normal” class with a confidence score of 0.80).</p> "> Figure 16
<p>Grad-CAM Overlay for Sick Class (the original thermal image (<b>left</b>) alongside the Grad-CAM overlay (<b>right</b>) demonstrates the model’s focus on specific regions, leading to the prediction of the “Sick” class with a confidence score of 0.85).</p> "> Figure 17
<p>Grad-CAM Overlay for Malignant Class (the original thermal image (<b>left</b>) and its corresponding Grad-CAM overlay (<b>right</b>) show the model’s focus on abnormal heat regions, supporting the “malignant” classification with a confidence score of 0.89).</p> "> Figure 18
<p>Grad-CAM Overlay for Benign Class (the original thermal image (<b>left</b>) and its Grad-CAM overlay (<b>right</b>) depict the regions contributing to the model’s prediction of the “benign” class with a confidence score of 0.88).</p> "> Figure 19
<p>(<b>Left</b>) Original thermal image highlighting the temperature distribution across the chest area, with warmer regions indicated by red/yellow hues and cooler regions by blue/green hues. (<b>Right</b>) Grad-CAM overlay demonstrating the areas of highest model attention during classification, with cooler colours indicating less attention and warmer colours indicating regions of interest.</p> ">
Abstract
:1. Introduction
- Utilizing thermography images in ML, DL, and transfer learning models to build an effective breast cancer detection system. The focus on thermal imaging offers a non-invasive diagnostic option with the potential to facilitate earlier detection and improve patient outcomes.
- Conducting thorough testing and benchmarking of the developed models against established methods to evaluate their performance and reliability. This comparative analysis highlights the strengths and limitations of various ML and DL techniques in breast cancer detection using thermal imaging, as measured by accuracy, precision, recall, and other performance indicators.
2. Related Work
3. Methods
3.1. Thermal Images
3.2. Data Preparation
3.2.1. Patient Guideline
- Avoid direct sun exposure prior to the imaging session.
- Refrain from any breast stimulation or treatments involving the breast area.
- Do not apply lotions, deodorants, antiperspirants, or makeup on the day of imaging.
- Avoid physical activities or exercises that may increase body temperature.
- Refrain from bathing or showering immediately before the imaging procedure.
- Remove clothing for approximately 12 min prior to imaging to allow the body to acclimate to the room temperature [40].
- Following these protocols helps ensure that the thermal images reflect true physiological conditions, enhancing the reliability of the diagnostic process.
3.2.2. Image Preprocessing
- Identifying and addressing incomplete or erroneous entries. Entries that could not be corrected were removed from the dataset.
- Excluding patients who lacked all five standard imaging angles: front, left 45°, right 45°, left 90°, and right 90°.
- Substituting dynamic protocol images for any missing or unclear static images in the front or side views.
3.3. Augmentation
3.4. Feature Extraction
- VGG16: Known for its depth and simplicity, it was used to extract high-level features, which is ideal for capturing intricate patterns within breast thermal images.
- InceptionV3: Utilizes inception modules for capturing multi-scale features, particularly valuable for distinguishing subtle temperature variations in thermal images.
- ResNet152: Leverages residual connections to handle deeper networks, enabling it to effectively capture complex visual patterns in breast cancer thermograms.
- MobileNetV2: Lightweight and efficient, MobileNetV2 is optimized for feature extraction in resource-constrained environments.
- DenseNet121: Uses dense connections to increase feature reuse and reduce redundancy, enhancing the model’s effectiveness in analyzing high-resolution thermal images.
- Xception: An extension of the Inception architecture employs depth-wise separable convolutions to efficiently capture essential features.
3.5. Dimensionality Reduction
3.6. Cross-Validation and Model Evaluation
3.7. Classifier
3.8. Evaluation Metrics
3.8.1. Confusion Matrix
3.8.2. Performance Metrics
- Accuracy: The overall correctness of the model is calculated as the ratio of correctly predicted instances to the total instances.
- Recall (Sensitivity or True Positive Rate): The ratio of correctly predicted positive observations to all observations in the actual class.
- Specificity (True Negative Rate): The ratio of correctly predicted negative observations to all observations in the actual negative class.
- F1 Score: The harmonic mean of precision and recall, providing a balance between the two.
3.8.3. Latency (Time and CPU)
4. Experiments and Results
4.1. Dataset
4.2. Experimental Setup
- Processor: Intel Core i7-11800H (8 cores, 16 threads)
- RAM: 32 GB DDR4
- Graphics Card: NVIDIA GeForce RTX 3060 (6 GB VRAM)
- Storage: 1 TB SSD
4.3. Effects of Feature Extraction
4.3.1. Data Cleaning and Feature Analysis
4.3.2. Dimensionality Reduction (PCA)
4.3.3. Feature Heatmap
- A heatmap was created to visualize the correlations between the top 10 features extracted from the thermal image dataset. These features include descriptors such as mean temperature, temperature variance, maximum intensity, edge gradient, and entropy. The color gradient in the heatmap indicates the strength of the correlations, with darker colors representing weaker correlations and lighter colors indicating stronger relationships.
- This visualization helps identify patterns and relationships among features, revealing which features are strongly correlated and which are independent.
- The absence of strong correlations among most features suggests minimal redundancy, meaning each feature provides unique information to the model.
- However, clusters of similar color patterns among certain features may suggest localized correlations, which could be further explored for dimensionality reduction or feature engineering strategies.
- The heatmap also offers insights into outliers or unique patterns across samples, supporting the development of more robust preprocessing methods.
- Additionally, similar color patterns among certain features may suggest correlations, which could be beneficial for dimensionality reduction techniques like PCA.
4.3.4. Feature Importance
- Feature importance was assessed using a RandomForestClassifier to rank the contributions of various features to the model’s predictive accuracy. The top-ranked features include mean temperature, maximum temperature, edge intensity, gradient magnitude, and texture complexity, which are crucial in distinguishing between benign and malignant cases.
- Features with lower importance scores, such as skewness or kurtosis, were found to contribute minimally and may be considered for removal in future iterations to streamline the model.
- The bar plot in Figure 9 highlights the top 10 most important features, with importance scores ranging from approximately 0.09 to 0.43. The highest-ranked feature, with a score of 0.4297, indicates a strong influence on the model’s predictive decisions.
- These results emphasize the importance of targeted feature engineering and data collection to further improve model accuracy. Additionally, identifying critical features allows for a focused approach in future studies, particularly in refining diagnostic tools.
4.4. Evaluation Metrics
4.4.1. Confusion Matrix of the Best Classifier
- True positives (TP) for healthy: 26—cases correctly identified as healthy.
- True positives (TP) for benign: 37—cases correctly identified as benign.
- True positives (TP) for malignant: 8—cases correctly identified as malignant.
- False positives (FP) for benign: 0—no healthy cases were misclassified as benign.
- False negatives (FN) for benign: 7—benign cases misclassified as malignant.
- False positives (FP) for malignant: 9—benign cases incorrectly classified as malignant.
4.4.2. Weighted Accuracy and Utility-Adjusted F1-Score
- Weighted Accuracy:
- TP: true positives
- TN: true negatives
- FP: false positives
- FN: false negatives
- Utility TP, Utility TN, Utility FP, Utility FN = Utility weights of each classification type
- Using the confusion matrix derived from the ResNet152 + SVM model (Figure 10),
- TP = 26 + 37 + 8 = 71; FP = 0 + 9 = 9; FN = 7; TN = 0
- Total Samples = TP + FP + FN + TN = 87; and
- Utility TP = Utility TN = 1 and Utility FP = Utility FN = −1
- ii.
- Utility-Adjusted F1—Score:
4.4.3. Precision–Recall Curve
4.4.4. The Receiver Operating Characteristic (ROC) Curve
4.5. Results and Discussion
4.5.1. Model Explainability Using GradCAM
4.5.2. Evaluation and Justification of ResNet152 as the Top-Performing Model in Breast Cancer Detection
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Sun, P.; Yu, C.; Yin, L.; Chen, Y.; Sun, Z.; Zhang, T.; Shuai, P.; Zeng, K.; Yao, X.; Chen, J.; et al. Global, regional, and national burden of female cancers in women of child-bearing age, 1990–2021: Analysis of data from the global burden of disease study 2021. eClinicalMedicine 2024, 74, 102713. [Google Scholar] [CrossRef] [PubMed]
- Arnold, M.; Morgan, E.; Rumgay, H.; Mafra, A.; Singh, D.; Laversanne, M.; Vignat, J.; Gralow, J.R.; Cardoso, F.; Siesling, S.; et al. Current and future burden of breast cancer: Global statistics for 2020 and 2040. Breast 2022, 66, 15–23. [Google Scholar] [CrossRef]
- Jalloul, R.; Chethan, H.K.; Alkhatib, R. A Review of Machine Learning Techniques for the Classification and Detection of Breast Cancer from Medical Images. Diagnostics 2023, 13, 2460. [Google Scholar] [CrossRef] [PubMed]
- Jiang, R.Y.; Fang, Z.R.; Zhang, H.P.; Xu, J.Y.; Zhu, J.Y.; Chen, K.Y.; Wang, W.; Jiang, X.; Wang, X.J. Ginsenosides: Changing the basic hallmarks of cancer cells to achieve the purpose of treating breast cancer. Chin. Med. 2023, 18, 125. [Google Scholar] [CrossRef]
- Upadhyaya, V. Predictive Analytics in Medical Diagnosis. In Intelligent Data Analytics for Bioinformatics and Biomedical Systems; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2024; pp. 27–66. [Google Scholar] [CrossRef]
- Guo, Y.; Zhang, H.; Yuan, L.; Chen, W.; Zhao, H.; Yu, Q.Q.; Shi, W. Machine learning and new insights for breast cancer diagnosis. J. Int. Med. Res. 2024, 52, 03000605241237867. [Google Scholar] [CrossRef]
- Tsietso, D.; Yahya, A.; Samikannu, R. A Review on Thermal Imaging-Based Breast Cancer Detection Using Deep Learning. Mob. Inf. Syst. 2022, 2022, 8952849. [Google Scholar] [CrossRef]
- Lakshman, K.; Dabhade, S.B.; Rode, Y.S.; Dabhade, K.; Deshmukh, S.; Maheshwari, R. Identification of Breast Cancer from Thermal Imaging using SVM and Random Forest Method. In Proceedings of the 2021 5th International Conference on Trends in Elec-tronics and Informatics (ICOEI), Tirunelveli, India, 3–5 June 2021. [Google Scholar] [CrossRef]
- Iacob, R.; Iacob, E.R.; Stoicescu, E.R.; Ghenciu, D.M.; Cocolea, D.M.; Constantinescu, A.; Ghenciu, L.A.; Manolescu, D.L. Evaluating the Role of Breast Ultrasound in Early Detection of Breast Cancer in Low- and Middle-Income Countries: A Comprehensive Narrative Review. Bioengineering 2024, 11, 262. [Google Scholar] [CrossRef] [PubMed]
- Mirasbekov, Y.; Aidossov, N.; Mashekova, A.; Zarikas, V.; Zhao, Y.; Ng, E.Y.K.; Midlenko, A. Fully Interpretable Deep Learning Model Using IR Thermal Images for Possible Breast Cancer Cases. Biomimetics 2024, 9, 609. [Google Scholar] [CrossRef]
- Sritharan, N.; Gutierrez, C.; Perez-Raya, I.; Gonzalez-Hernandez, J.L.; Owens, A.; Dabydeen, D.; Medeiros, L.; Kandlikar, S.; Phatak, P. Breast Cancer Screening Using Inverse Modeling of Surface Temperatures and Steady-State Thermal Imaging. Cancers 2024, 16, 2264. [Google Scholar] [CrossRef] [PubMed]
- Islam, T.; Sheakh, M.A.; Tahosin, M.S.; Hena, M.H.; Akash, S.; Jardan, Y.a.B.; FentahunWondmie, G.; Nafidi, H.A.; Bourhia, M. Predictive modeling for breast cancer classification in the context of Bangladeshi patients by use of machine learning approach with explainable AI. Sci. Rep. 2024, 14, 8487. [Google Scholar] [CrossRef] [PubMed]
- Ikechukwu, A.V.; Bhimshetty, S.; Deepu, R.; Mala, M.V. Advances in Thermal Imaging: A Convolutional Neural Network Ap-proach for Improved Breast Cancer Diagnosis. In Proceedings of the 2024 International Conference on Distributed Computing and Optimization Techniques (ICDCOT), Bengaluru, India, 15–16 March 2024; pp. 1–7. [Google Scholar] [CrossRef]
- Tiwari, D.; Dixit, M.; Gupta, K. Deep Multi-View Breast Cancer Detection: A Multi-View Concatenated Infrared Thermal Images Based Breast Cancer Detection System Using Deep Transfer Learning. Trait. Signal 2021, 38, 1699–1711. [Google Scholar] [CrossRef]
- Lakkis, N.A.; Abdallah, R.M.; Musharrafieh, U.M.; Issa, H.G.; Osman, M.H. Epidemiology of Breast, Corpus Uteri, and Ovarian Cancers in Lebanon With Emphasis on Breast Cancer Incidence Trends and Risk Factors Compared to Regional and Global Rates. Cancer Control 2024, 31, 10732748241236266. [Google Scholar] [CrossRef] [PubMed]
- Desai, M.; Shah, M. An anatomization on breast cancer detection and diagnosis employing multi-layer perceptron neural network (MLP) and Convolutional neural network (CNN). Clin. eHealth 2021, 4, 1–11. [Google Scholar] [CrossRef]
- Alshehri, A.; AlSaeed, D. Breast Cancer Detection in Thermography Using Convolutional Neural Networks (CNNs) with Deep Attention Mechanisms. Appl. Sci. 2022, 12, 12922. [Google Scholar] [CrossRef]
- Wen, X.; Guo, X.; Wang, S.; Lu, Z.; Zhang, Y. Breast cancer diagnosis: A systematic review. Biocybern. Biomed. Eng. 2024, 44, 119–148. [Google Scholar] [CrossRef]
- Agughasi, V.I. The Superiority of Fine-tuning over Full-training for the Efficient Diagnosis of COPD from CXR Images. Intel. Artif. 2024, 27, 62–79. [Google Scholar] [CrossRef]
- Agughasi, V.I. Leveraging Transfer Learning for Efficient Diagnosis of COPD Using CXR Images and Explainable AI Tech-niques. Intel. Artif. 2024, 27, 133–151. [Google Scholar] [CrossRef]
- Sharafaddini, A.M.; Esfahani, K.K.; Mansouri, N. Deep learning approaches to detect breast cancer: A comprehensive review. Multimed. Tools Appl. 2024. [Google Scholar] [CrossRef]
- Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Con-nections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31, pp. 4278–4284. [Google Scholar] [CrossRef]
- Ekici, S.; Jawzal, H. Breast cancer diagnosis using thermography and convolutional neural networks. Med. Hypotheses 2020, 137, 109542. [Google Scholar] [CrossRef]
- Chowdhury, N.A.; Wang, L.; Gu, L.; Kaya, M. Machine Learning for Early Breast Cancer Detection. J. Eng. Sci. Med. Diagn. Ther. 2024, 8, 010801. [Google Scholar] [CrossRef]
- Abed, A.H.; Shaaban, E.M. Modeling Deep Neural Networks For Breast Cancer Thermography Classification: A Review Study. Int. J. Adv. Netw. Appl. 2021, 13, 4939–4946. [Google Scholar] [CrossRef]
- Mambou, S.; Maresova, P.; Krejcar, O.; Selamat, A.; Kuca, K. Breast Cancer Detection Using Infrared Thermal Imaging and a Deep Learning Model. Sensors 2018, 18, 2799. [Google Scholar] [CrossRef]
- Amethiya, Y.; Pipariya, P.; Patel, S.; Shah, M. Comparative analysis of breast cancer detection using machine learning and biosensors. Intell. Med. 2022, 2, 69–81. [Google Scholar] [CrossRef]
- Gonçalves, C.B.; Souza, J.R.; Fernandes, H. CNN architecture optimization using bio-inspired algorithms for breast cancer detection in infrared images. Comput. Biol. Med. 2022, 142, 105205. [Google Scholar] [CrossRef]
- Silva, L.F.; Saade, D.C.M.; Sequeiros, G.O.; Silva, A.C.; Paiva, A.C.; Bravo, R.S.; Conci, A. A New Database for Breast Research with Infrared Image. J. Med. Imaging Health Inform. 2014, 4, 92–100. [Google Scholar] [CrossRef]
- Aarthy, S.L.; Prabu, S. Classification of breast cancer based on thermal image using support vector machine. Int. J. Bioinform. Res. Appl. 2019, 15, 51–67. [Google Scholar] [CrossRef]
- Allugunti, V.R. Breast cancer detection based on thermographic images using machine learning and deep learning algorithms. Int. J. Eng. Comput. Sci. 2022, 4, 49–56. [Google Scholar] [CrossRef]
- Karthiga, R.; Narasimhan, K. Medical imaging technique using curvelet transform and machine learning for the automated diagnosis of breast cancer from thermal image. Pattern Anal. Appl. 2021, 24, 981–991. [Google Scholar] [CrossRef]
- Banco De Imagens Mastológicas. Available online: https://visual.ic.uff.br/dmi/ (accessed on 4 November 2024).
- Nissar, I.; Alam, S.; Masood, S. Computationally efficient LC-SCS deep learning model for breast cancer classification using thermal imaging. Neural Comput. Appl. 2024, 36, 16233–16250. [Google Scholar] [CrossRef]
- Bhowmik, M.K.; Gogoi, U.R.; Majumdar, G.; Bhattacharjee, D.; Datta, D.; Ghosh, A.K. Designing of Ground Truth Annotated DBT-TU-JU Breast Thermogram Database towards Early Abnormality Prediction. IEEE J. Biomed. Health Inform. 2017, 22, 1238–1249. [Google Scholar] [CrossRef]
- Torres-Galvan, J.C.; Guevara, E.; Gonzalez, F.J. Comparison of Deep Learning Architectures for Pre-Screening of Breast Cancer Thermograms. In Proceedings of the 2019 Photonics North (PN), Quebec City, QC, Canada, 21–23 May 2019. [Google Scholar] [CrossRef]
- D’Alessandro, G.; Tavakolian, P.; Sfarra, S. A Review of Techniques and Bio-Heat Transfer Models Supporting Infrared Thermal Imaging for Diagnosis of Malignancy. Appl. Sci. 2024, 14, 1603. [Google Scholar] [CrossRef]
- Zuluaga-Gomez, J.; Masry, Z.A.; Benaggoune, K.; Meraghni, S.; Zerhouni, N. A CNN-based methodology for breast cancer diagnosis using thermal images. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2020, 9, 131–145. [Google Scholar] [CrossRef]
- Abdullakutty, F.; Akbari, Y.; Al-Maadeed, S.; Bouridane, A.; Hamoudi, R. Advancing Histopathology-Based Breast Cancer Diagnosis: Insights into Multi-Modality and Explainability. arXiv 2024, arXiv:2406.12897. [Google Scholar] [CrossRef]
- Ikechukwu, A.V.; Murali, S. CX-Net: An efficient ensemble semantic deep neural network for ROI identification from chest-x-ray images for COPD diagnosis. Mach. Learn. Sci. Technol. 2023, 4, 025021. [Google Scholar] [CrossRef]
- Iyadurai, J.; Chandrasekharan, M.; Muthusamy, S.; Panchal, H. An Extensive Review on Emerging Advancements in Ther-mography and Convolutional Neural Networks for Breast Cancer Detection. Wirel. Pers. Commun. 2024, 137, 1797–1821. [Google Scholar] [CrossRef]
- Martin, P.P.; Graulich, N. Navigating the data frontier in science assessment: Advancing data augmentation strategies for machine learning applications with generative artificial intelligence. Comput. Educ. Artif. Intell. 2024, 7, 100265. [Google Scholar] [CrossRef]
- Schwartz, R.G.; Horner, C.; Kane, R.; Getson, P.; Brioschi, M.; Pittman, J.; Rind, B.; Campbell, J.; Ehle, E.; Mustovoy, A.; et al. Guidelines for Breast Thermology. In Guidelines for Breast Thermology; American Academy of Thermology: Greenville, SC, USA, 2021. [Google Scholar]
- Bezerra, L.; Ribeiro, R.; Lyra, P.; Lima, R. An empirical correlation to estimate thermal properties of the breast and of the breast nodule using thermographic images and optimization techniques. Int. J. Heat Mass Transf. 2020, 149, 119215. [Google Scholar] [CrossRef]
- Gogoi, U.R.; Majumdar, G.; Bhowmik, M.K.; Ghosh, A.K.; Bhattacharjee, D. Breast abnormality detection through statistical feature analysis using infrared thermograms. In Proceedings of the 2015 International Symposium on Advanced Computing and Communication (ISACC), Silchar, India, 14–15 September 2015; pp. 258–265. [Google Scholar] [CrossRef]
- Resmini, R.; Conci, A.; da Silva, L.F.; Sequeiros, G.O.; Araújo, F.; de Araújo, C.; dos Santos Araújo, A. Application of Infrared Images to Diagnosis and Modeling of Breast. In Application of Infrared to Biomedical Sciences; Ng, E.Y., Etehadtavakol, M., Eds.; Springer: Singapore, 2017; pp. 159–173. [Google Scholar] [CrossRef]
- Husaini, M.A.S.A.; Habaebi, M.H.; Hameed, S.A.; Islam, M.R.; Gunawan, T.S. A Systematic Review of Breast Cancer Detection Using Thermography and Neural Networks. IEEE Access 2020, 8, 208922–208937. [Google Scholar] [CrossRef]
- Omranipour, R.; Kazemian, A.; Alipour, S.; Najafi, M.; Alidoosti, M. Comparison of the Accuracy of Thermography and Mammography in the Detection of Breast Cancer. Breast Care 2016, 11, 260–264. [Google Scholar] [CrossRef]
- Dharani, N.P.; Immadi, I.G.; Narayana, M.V. Enhanced deep learning model for diagnosing breast cancer using thermal im-ages. Soft Comput. 2024, 28, 8423–8434. [Google Scholar] [CrossRef]
- Yadav, S.S.; Jadhav, S.M. Thermal infrared imaging-based breast cancer diagnosis using machine learning tech-niques. Multimed. Tools Appl. 2020, 81, 13139–13157. [Google Scholar] [CrossRef]
- Abraham, A.; Ohsawa, Y.; Gandhi, N.; Jabbar, M.; Haqiq, A.; McLoone, S.; Issac, B. Advances in Intelligent systems and computing. In Proceedings of the 12th International Conference on Soft Computing and Pattern Recognition (SoCPaR 2020), online, India, 15–18 December 2020; Springer Nature: Dordrecht, The Netherlands, 2021. [Google Scholar] [CrossRef]
- Al Husaini, M.A.S.; Habaebi, M.H.; Islam, M.R. Real-time thermography for breast cancer detection with deep learning. Discov. Artif. Intell. 2024, 4, 57. [Google Scholar] [CrossRef]
- Masekela, A.; Zhao, Y.; Ng, E.Y.; Zarikas, V.; Fok, S.C.; Mukhmetov, O. Early detection of the breast cancer using infrared technology—A comprehensive review. Therm. Sci. Eng. Prog. 2022, 27, 101142. [Google Scholar] [CrossRef]
- Rodriguez-Guerrero, S.; Correa, H.L.; Restrepo-Girón, A.D.; Reyes, L.A.; Olave, L.A.; Diaz, S. Breast Thermography. Mendeley Data 2024, 3. [Google Scholar] [CrossRef]
- Omondiagbe, D.A.; Veeramani, S.; Sidhu, A.S. Machine Learning Classification Techniques for Breast Cancer Diagnosis. IOP Conf. Ser. Mater. Sci. Eng. 2019, 495, 012033. [Google Scholar] [CrossRef]
- Ikechukwu, A.V.; Murali, S. xAI: An Explainable AI Model for the Diagnosis of COPD from CXR Images. In Proceedings of the 2023 IEEE 2nd International Conference on Data, Decision and Systems (ICDDS), Mangaluru, India, 1–2 December 2023; pp. 1–6. [Google Scholar] [CrossRef]
Ref. | Approach | Dataset | Imaging Type | Feature Extraction | Results (Accuracy) |
---|---|---|---|---|---|
Silva et al. [29] | SVM | Not specified | Not specified | Standard SVM features | 90% |
Aarthy et al. [30] | SVM | 83 images (34 normal, 49 abnormal) | Static | Custom feature engineering | 97.6% |
Allugunti et al. [31] | SVM, Random Forest | DMR-IR | Static | Standard SVM and RF | SVM: 89.84%, RF: 90.55% |
Karthiga et al. [32] | SVM | DMR-IR | Dynamic | Custom feature extraction | 93.3% |
Bancos et al. [33] | ANN, Decision Tree, Bayesian | DMR-IR | Dynamic | Standard ANN, DT, Bayesian | ANN: 73.38%, DT: 78%, Bayesian: 88% |
Nissar et al. [34] | SVM | DMR-IR | Static | Standard SVM features | Not reported |
Ref. | Approach | Dataset | Imaging Type | Feature Extraction | Results (Accuracy) |
---|---|---|---|---|---|
Bhowmik et al. [35] | Multilayer Perceptron (MLP) | DMR | Static | Traditional MLP | 95% |
Torres et al. [36] | ResNet101, MobileNetV2, DenseNet201 | DMR/IR | Static | CNN-based | MobileNetV2: 99.6% |
D’Alessandro et al. [37] | SVM and DNN | 67 patients (43 healthy, 24 sick) | Static | CNN feature extraction | 94% |
Zuluaga-Gomez et al. [38] | CNNs | DMR/IR | Dynamic | CNN-based | 92% |
Abdullakutty et al. [39] | VGG16, InceptionV3 | DMR-IR | Static | Transfer Learning | VGG16: 87.3% |
Agughasi et al. [40] | ResNet50, InceptionV3 | 57 Thermal Images | Static | CNN-based | ResNet50: 92%, InceptionV3: 90% |
Ref. | Source | SIT | DIT | No. of Images | Clinical Validation | Public Access | Camera Used |
---|---|---|---|---|---|---|---|
Bezerra et al. [44] | Clinical Hospital of the federal University of Pernambuco (HC/UFPE), Brazil | Yes | No | 336 (120 benign, 74 cysts, 76 malignant, 66 without lesion) | No | Yes | Not specified |
Gogoi et al. [45] | Agartala Government Medical College (AGMC) of Govind Ballav Pant (GBP) Hospital, Agartala | Yes | No | 49 abnormal, 45 normal, 6 unknown | Yes | No | FLIR-T650sc |
Resmini et al. [46] | University Hospital Antônio Pedro (HUAP) of Federal Fluminense University, Brazil | Yes | Yes | 311 (267 healthy, 44 sick) | Yes | Yes | FLIR-SC620 |
Metric | Real Data | Augmented Data | Insight |
---|---|---|---|
Kolmogorov–Smirnov Test | KS Statistic = 0.09 | p-Value = 0.0006 | Indicates a statistically significant difference between the real and augmented pixel intensity distributions, suggesting augmentation slightly altered data. |
Entropy | 7.997 | 7.979 | Entropy values are very close, implying similar texture complexity between the real and augmented images. |
Haralick Features | Various (see above) | Various (see above) | Most Haralick texture metrics, such as energy and contrast, are very similar, demonstrating minimal changes in texture properties after augmentation. |
SSIM | - | 0.9969 | The high SSIM value (close to 1) indicates the augmented images retain structural similarity to the real-world data. |
Max Temperature (°F) | 157.79 | 149.90 | Both maximum temperature values are outside the typical clinical range (85–110 °F), indicating potential data generation issues or unrealistic augmentation. |
Min Temperature (°F) | 51.38 | 57.89 | Minimum temperature values are also outside the expected range, highlighting a need to validate or constrain augmentation methods. |
Temperature Validity | Outside realistic range | Outside realistic range | The temperature range in both datasets fails to meet clinical expectations, signaling a need for stricter data preprocessing or augmentation constraints. |
Model | Accuracy | Precision | Recall | Specificity | F1 Score | AUC | Latency (s) | CPU Utilization (%) |
---|---|---|---|---|---|---|---|---|
ResNet152 + SVM | 97.62% | 95.79% | 98.53% | 94.52% | 97.16% | 99% | 0.06 | 88.66 |
DenseNet121 + SVM | 97.00% | 94.85% | 97.45% | 93.60% | 96.12% | 98% | 0.07 | 86.23 |
MobileNetV2 + SVM | 94.00% | 92.35% | 94.56% | 90.33% | 93.43% | 97% | 0.08 | 80.50 |
InceptionV3 + SVM | 94.00% | 98.00% | 92.00% | 97.00% | 94.00% | 98% | 0.10 | 82.81 |
VGG16 + SVM | 93.00% | 91.70% | 93.45% | 89.50% | 92.56% | 95% | 0.08 | 94.93 |
Xception + SVM | 90.50% | 90.10% | 90.80% | 89.20% | 90.44% | 94% | 0.12 | 76.45 |
Random Forest Classifier | Validation Accuracy | Test Accuracy | Precision | Recall | F1-Score | Specificity | AUC | False Positive Rate | Measure Latency | Measure CPU |
---|---|---|---|---|---|---|---|---|---|---|
Vgg16 + RF | 0.96 | 0.96 | 0.96 | 0.96 | 0.96 | 0.95 | 0.99 | 0.05 | 0.12 s | 96% |
Inceptiomv3 + RF | 0.91 | 0.91 | 0.89 | 0.95 | 0.92 | 0.8 | 0.98 | 0.14 | 0.12 s | 99% |
Resnet151 + RF | 0.94 | 0.95 | 0.95 | 0.96 | 0.96 | 0.94 | 0.99 | 0.07 | 0.16 s | 111% |
mobilenetv2 + RF | 0.93 | 0.94 | 0.95 | 0.95 | 0.95 | 0.94 | 0.99 | 0.06 | 0.16 | 111% |
Densenet121 + RF | 0.92 | 0.95 | 0.97 | 0.95 | 0.96 | 0.96 | 0.99 | 0.04 | 0.13 s | 98% |
Xception + RF | 0.93 | 0.94 | 0.92 | 0.96 | 0.94 | 0.91 | 0.99 | 0.09 | 0.10 s | 96% |
Study | Model(s) Used | Dataset | Test Accuracy | Precision | Recall | AUC | Latency |
---|---|---|---|---|---|---|---|
Proposed | ResNet152 + SVM | DMR-IR | 95% | 98% | 94% | 99% | 0.06 s |
Dabhade et al. [8] | Random Forest, SVM | Proprietary | 98.4% (SVM) | - | - | - | - |
Tiwari et al. [14] | VGG16 | Proprietary | 99% | - | - | - | - |
Mambou et al. [26] | SVM, DNN | Proprietary | 94% | - | - | - | - |
Bezerra et al. [27] | Naive Bayes, SVM | DMR-IR | 95% (SVM) | 92% | 93% | 97% | - |
Zuluaga-Gomez et al. [40] | CNN | DMR-IR | 92% | - | - | 92% | - |
Karthiga and Narasimhan [51] | VGG16, InceptionV3 | DMR-IR | 87.3% (VGG16) | - | - | - | - |
Proposed Model Alternative | DenseNet 121 + SVM | DMR-IR | 97% | 95% | 97% | 98% | 0.07 s |
DT Classifier | Validation Accuracy | Test Accuracy | Precision | Recall | F1-Score | Specificity | AUC | False Positive Rate | Measure Latency | Measure CPU |
---|---|---|---|---|---|---|---|---|---|---|
vgg16 + DT | 0.87 | 0.87 | 0.89 | 0.87 | 0.88 | 0.87 | 0.87 | 0.13 | 0.11 s | 99% |
Inceptionv3 + DT | 0.85 | 0.83 | 0.85 | 0.83 | 0.84 | 0.84 | 0.83 | 0.17 | 0.12 s | 98% |
Resnet151 + DT | 0.87 | 0.87 | 0.87 | 0.91 | 0.86 | 0.88 | 0.89 | 0.11 | 0.11 s | 99% |
mobilenetv2 + DT | 0.83 | 0.87 | 0.87 | 0.9 | 0.88 | 0.84 | 0.87 | 0.16 | 0.9 s | 93% |
Desnet121 + DT | 0.91 | 0.87 | 0.84 | 0.85 | 0.85 | 0.85 | 0.84 | 0.15 | 0.10 s | 99% |
Xception + DT | 0.89 | 0.86 | 0.87 | 0.88 | 0.88 | 0.84 | 0.86 | 0.16 | 0.11 s | 96% |
KNN Classifier | Validation Accuracy | Test Accuracy | Precision | Recall | F1-Score | Specificity | AUC | False Positive Rate | Measure Latency | Measure CPU |
---|---|---|---|---|---|---|---|---|---|---|
Vgg16 + KNN | 0.95 | 0.93 | 0.97 | 0.91 | 0.94 | 0.97 | 0.97 | 0.03 | 0.05 s | 77.00% |
Inceptionv3 + KNN | 0.91 | 0.91 | 0.94 | 0.89 | 0.92 | 0.93 | 0.96 | 0.07 | 0.06 s | 96.40% |
Resnet151 + KNN | 0.92 | 0.95 | 0.98 | 0.93 | 0.96 | 0.98 | 0.99 | 0.02 | 0.10 s | 98.30% |
mobilev2 + KNN | 0.93 | 0.93 | 0.97 | 0.9 | 0.93 | 0.96 | 0.97 | 0.04 | 0.18 s | 89.60% |
Desnet121 + KNN | 0.96 | 0.94 | 0.99 | 0.91 | 0.95 | 0.99 | 0.99 | 0.01 | 0.06 s | 93% |
xception + KNN | 0.93 | 0.92 | 0.95 | 0.91 | 0.92 | 0.94 | 0.98 | 0.06 | 021 s | 99.80% |
DNN Classifier | Validation Accuracy | Test Accuracy | Precision | Recall | F1-Score | Specificity | AUC | False Positive Rate | Measure Latency | Measure CPU |
---|---|---|---|---|---|---|---|---|---|---|
Vgg16 + DNN | 0.96 | 0.93 | 0.98 | 0.88 | 0.93 | 0.98 | 0.99 | 0.02 | 0.16 s | 87.60% |
Inceptionv3 + DNN | 0.92 | 0.94 | 0.96 | 0.93 | 0.95 | 0.95 | 0.99 | 0.05 | 0.18 s | 79.60% |
Resnet152 + DNN | 0.96 | 0.94 | 0.95 | 0.95 | 0.95 | 0.94 | 0.99 | 0.06 | 0.13 s | 94.30% |
mobilev2 + DNN | 0.96 | 0.95 | 0.96 | 0.94 | 0.95 | 0.95 | 0.99 | 0.05 | 0.18 s | 89.60% |
Desnet121 + DNN | 0.95 | 0.94 | 0.96 | 0.94 | 0.95 | 0.95 | 0.98 | 0.05 | 0.10 s | 91% |
xception + DNN | 0.93 | 0.93 | 0.95 | 0.92 | 0.94 | 0.94 | 0.98 | 0.06 | 0.12 s | 106.10% |
Naive Bayes Classifier | Validation Accuracy | Test Accuracy | Precision | Recall | F1-Score | Specificity | AUC | False Positive Rate | Measure Latency | Measure CPU |
---|---|---|---|---|---|---|---|---|---|---|
Vgg16 + NB | 0.68 | 0.6 | 0.98 | 0.74 | 0.78 | 0.21 | 0.78 | 0.7 | 0.20 s | 119% |
Inceptionv3 + NB | 0.91 | 0.91 | 0.89 | 0.95 | 0.92 | 0.86 | 0.98 | 0.14 | 0.18 s | 119% |
Resnet151 + NB | 0.85 | 0.83 | 0.85 | 0.83 | 0.84 | 0.83 | 0.83 | 0.17 | 0.18 s | 116% |
mobilev2 + NB | 0.81 | 0.79 | 0.83 | 0.78 | 0.8 | 0.8 | 0.87 | 0.17 | 0.20 s | 119% |
Desnet121 + NB | 0.75 | 0.78 | 0.78 | 0.84 | 0.81 | 0.71 | 0.83 | 0.23 | 0.15 s | 111% |
xception + NB | 0.79 | 0.81 | 0.84 | 0.8 | 0.82 | 0.82 | 0.84 | 0.18 | 0.21 s | 121% |
Author(s) | Model(s) | Dataset | Accuracy (%) | Precision (%) | Recall (%) | AUC (%) |
---|---|---|---|---|---|---|
Silva et al. [36] | SVM | Not specified | 90 | 85 | 88 | 89 |
Bhowmik et al. [37] | Multilayer Perceptron | DMR | 95 | 94 | 92 | 97 |
D’Alessandro et al. [39] | SVM | DMR | 94 | 93 | 91 | 95 |
Abdullakutty et al. [48] | VGG16, InceptionV3 | DMR-IR | 87.3 | 86 | 85 | 89 |
Aarthy et al. [49] | SVM | DMR | 97.6 | 95 | 93 | 96 |
Allugunti et al. [50] | SVM, Random Forest | DMR-IR | 90.55 | 89 | 88 | 92 |
Proposed Model | ResNet152 + SVM | DMR | 95 | 98 | 94 | 99 |
Bancos et al. [53] | ANN, Decision Tree, Bayesian | DMR-IR | 78 | 75 | 72 | 80 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jalloul, R.; Krishnappa, C.H.; Agughasi, V.I.; Alkhatib, R. Enhancing Early Breast Cancer Detection with Infrared Thermography: A Comparative Evaluation of Deep Learning and Machine Learning Models. Technologies 2025, 13, 7. https://doi.org/10.3390/technologies13010007
Jalloul R, Krishnappa CH, Agughasi VI, Alkhatib R. Enhancing Early Breast Cancer Detection with Infrared Thermography: A Comparative Evaluation of Deep Learning and Machine Learning Models. Technologies. 2025; 13(1):7. https://doi.org/10.3390/technologies13010007
Chicago/Turabian StyleJalloul, Reem, Chethan Hasigala Krishnappa, Victor Ikechukwu Agughasi, and Ramez Alkhatib. 2025. "Enhancing Early Breast Cancer Detection with Infrared Thermography: A Comparative Evaluation of Deep Learning and Machine Learning Models" Technologies 13, no. 1: 7. https://doi.org/10.3390/technologies13010007
APA StyleJalloul, R., Krishnappa, C. H., Agughasi, V. I., & Alkhatib, R. (2025). Enhancing Early Breast Cancer Detection with Infrared Thermography: A Comparative Evaluation of Deep Learning and Machine Learning Models. Technologies, 13(1), 7. https://doi.org/10.3390/technologies13010007