Semi-Supervised Deep Subspace Embedding for Binary Classification of Sella Turcica
<p>Sample images of pre-defined Sella Turcica (ST) shapes: (<b>A</b>) Oval ST, (<b>B</b>) Circular ST, (<b>C</b>) Flat ST, and (<b>D</b>) Bridging ST. This study classified Circular ST as non-bridging, and Bridging ST was used for binary classification.</p> "> Figure 2
<p>The schematic representation of a Hybrid Database (<span class="html-italic">L</span>) and Hybrid Case Base (<math display="inline"><semantics> <mrow> <mi>S</mi> <mi>L</mi> </mrow> </semantics></math>) from labeled (<math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>L</mi> <mi>i</mi> </msub> </mrow> </semantics></math>) and unlabeled (<math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>L</mi> <mover accent="true"> <mi>i</mi> <mo>^</mo> </mover> </msub> </mrow> </semantics></math>) case data. Feature extraction using KL divergence, mean (<math display="inline"><semantics> <mi>μ</mi> </semantics></math>), and standard deviation (<math display="inline"><semantics> <mi>σ</mi> </semantics></math>) is applied to both databases. Labeled data form a featured database with labels (<math display="inline"><semantics> <msub> <mi>L</mi> <mi>i</mi> </msub> </semantics></math>), while unlabeled data create a featured database without labels (<math display="inline"><semantics> <msub> <mi>L</mi> <mover accent="true"> <mi>i</mi> <mo>^</mo> </mover> </msub> </semantics></math>). A dynamic responsive data and label mechanism integrates both, resulting in (1) a Hybrid Database (<span class="html-italic">L</span>) and (2) a Hybrid Case Base (<math display="inline"><semantics> <mrow> <mi>S</mi> <mi>L</mi> </mrow> </semantics></math>) for further analysis.</p> "> Figure 3
<p>Process flow diagram of the proposed SSLDSE framework.</p> "> Figure 4
<p>The figure illustrates a comprehensive framework of the proposed SSLDSE that integrates labeled (<math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>L</mi> <mi>i</mi> </msub> </mrow> </semantics></math>) and unlabeled case databases (<math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>L</mi> <mover accent="true"> <mi>i</mi> <mo>^</mo> </mover> </msub> </mrow> </semantics></math>). Features are extracted using Kullback–Leibler divergence, mean (<math display="inline"><semantics> <mi>μ</mi> </semantics></math>), and standard deviation (<math display="inline"><semantics> <mi>σ</mi> </semantics></math>), forming a Hybrid Database. The data undergo stochastic augmentation and are processed through an Inception-ResNet-V2 model. A deep subspace descriptor with t-SNE refines the feature representations, and the outputs are classified by a zero-shot classifier (<math display="inline"><semantics> <mrow> <mi>Z</mi> <mi>s</mi> <mi>C</mi> </mrow> </semantics></math>) with KL divergence loss, enabling the model to handle unseen or unlabeled ST structures.</p> "> Figure 5
<p>The illustrated SSLDSE architectural framework processes labeled (L) and semi-labeled (<math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>L</mi> <mi>i</mi> </msub> </mrow> </semantics></math>) data using Inception-ResNet-V2 as the CNN backbone to extract features (P, Q) and estimate pairwise probability densities (<math display="inline"><semantics> <msub> <mi mathvariant="normal">P</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> </mrow> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi mathvariant="normal">Q</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> </mrow> </msub> </semantics></math>). KL divergence (<math display="inline"><semantics> <mrow> <msub> <mi>D</mi> <mi>KL</mi> </msub> <mrow> <mo>(</mo> <mi>P</mi> <mo>‖</mo> <mi>Q</mi> <mo>)</mo> </mrow> </mrow> </semantics></math>) minimizes divergence through the optimization of (Y). Manifold learning maps feature matrices (X) to t-SNE representations (Y) while preserving structural relationships. The SSL framework employs deep embedding and clustering (mean: <math display="inline"><semantics> <msub> <mi>μ</mi> <mi>j</mi> </msub> </semantics></math>, covariance: <math display="inline"><semantics> <msub> <mi>Σ</mi> <mi>j</mi> </msub> </semantics></math>) for feature representation. The zero-shot classifier constructs semantic vectors and applies KL divergence loss for output prediction (<math display="inline"><semantics> <msub> <mi>O</mi> <mi>i</mi> </msub> </semantics></math>).</p> "> Figure 6
<p>Confusion matrix and ROC curve showcasing the validation results of the binary classifier, highlighting the proposed model’s classification performance through true positive/negative rates and the AUC-ROC score.</p> "> Figure 7
<p>t-SNE plots visualizing the quantitative assessment of the proposed SSLDSE method. The plots illustrate the effective separation between bridging and non-bridging labels from our proprietary and IEEE ISBI 2015 datasets, demonstrating a clear class distinction.</p> "> Figure 8
<p>Boxplots illustrate a detailed comparison of the classification error rates for the proposed SSLDSE method, showing the distribution of error rates across (<b>a</b>) the proprietary dataset and (<b>b</b>) the IEEE ISBI 2015 dataset, highlighting the variability and consistency in classification accuracy.</p> "> Figure 9
<p>Boxplots comparing classification error rates across utilized SSL approaches and the proposed SSLDSE method, illustrating performance differences and the effectiveness of SSLDSE in reducing classification errors.</p> "> Figure 10
<p>Visual interpretation of errors in ST-binary classification predictions, illustrating misclassified instances and highlighting the areas where the model’s predictions diverge from the true labels.</p> ">
Abstract
:1. Introduction
- The method proposed in this study introduces a novel approach for conducting a non-linear embedding with KL divergence for semi-supervised binary classification.
- The proposed semi-supervised deep subspace embedding (SSLDSE) methodology creates a shift-invariant image representation by first producing an augmented image set for each input. This approach addresses the limitations of existing models, such as Active SSL and Contrastive SSL, which struggle with dataset variability and shifts. SSLDSE improves feature representation and robustness to transformations by mapping each image set to a non-linear subspace defined by a KL divergence point.
- The model employs t-SNE for feature representations, utilizing manifold learning to capture non-linear structures that previous methods, such as GAN SSL and Contrastive SSL, may not effectively recognize. This enables SSLDSE to handle complex variations in medical images more effectively, improving classification performance by preserving essential relationships in the data.
- SSLDSE also integrates a zero-shot classifier (ZsC) that leverages KL divergence loss to classify previously unseen or unlabeled ST structures. This is particularly important in medical imaging, where acquiring labeled data is challenging. Unlike GAN SSL and Active SSL, which require retraining to adapt to new data, SSLDSE generalizes to unseen classes without additional labeling, enhancing model adaptability and efficiency.
- By combining real-time stochastic augmentation with KL divergence, SSLDSE generates shift-invariant features and generalizes better across different datasets, mitigating the risk of overfitting. In contrast, models like Active SSL, which may overfit with limited labeled data, show less consistency, while SSLDSE demonstrates accurate performance across diverse datasets.
- The SSLDSE model is benchmarked against state-of-the-art semi-supervised learning models, including Active SSL, GAN SSL, and Contrastive SSL. Through experiments on both our proprietary dataset and the IEEE ISBI 2015 dataset, SSLDSE achieves superior performance in terms of accuracy, precision, recall, and F1 score. It outperforms these models by delivering robust classification with fewer labeled data, demonstrating its effectiveness in real-world applications.
2. Materials and Methods
2.1. Dataset and Pre-Processing
2.2. Semi-Supervised Methods
2.2.1. Active Learning
2.2.2. Generative Adversarial Networks
2.2.3. Contrastive Learning
2.3. Proposed Semi-Supervised Deep Subspace Embedding (SSLDSE) Architecture
2.3.1. Exploring Non-Linear Embeddings with Kullback–Leibler Divergence
- [CNN-Based Feature Extraction and Probability Distribution Estimation] The rich representational power of intermediate CNN outputs, specifically from Inception-ResNet-v2 [59,60,61], allows for the utilization of these features as the primary inputs for the KL divergence calculations [62]. These CNN-extracted features encapsulate both low-level and high-level morphological traits in a reduced dimensional space, making them particularly suitable for this form of probabilistic analysis. The process begins with extracting deep features from each image, followed by converting these features into probability distributions. To estimate these distributions for a pair of images, a binning technique is employed, where features are binned into fixed intervals. The number of bins as well as the binning strategy (e.g., equal width or equal frequency) are optimized for the dataset. To ensure numerical stability and to avoid zero probabilities in the binning process, a smoothing constant is added to each bin count prior to normalization. Given two sets of features P (from image 1) and Q (from image 2), the KL divergence between the distributions of their CNN outputs is calculated as
- [Efficient Computation and Manifold Learning Integration] Computing the KL divergence for all image pairs in large datasets can be computationally prohibitive. Therefore, an optimized sampling method is applied to select a representative subset of image pairs for divergence calculation. This reduces the computational load significantly while maintaining analytical precision. Parallel processing across multiple cores further enhances the efficiency of these calculations, allowing for the rapid processing of large datasets without sacrificing detail. To capture the complex, non-linear relationships inherent in these features, manifold learning techniques, particularly t-distributed Stochastic Neighbor Embedding (t-SNE), are incorporated [63,64]. t-SNE helps embed high-dimensional divergence values into a lower-dimensional space while preserving both the local and global structures of the data. This allows the model to generalize effectively across variations in ST. The embedding reveals clusters and patterns that are not immediately obvious in the original high-dimensional feature space.Let and denote the sets of images represented by bridging and non-bridging ST forms, respectively. For each image I in and , a feature representation is extracted. The KL divergence is computed for each image pair in and in , resulting in a matrix of divergence values. A non-linear embedding technique is then used to visualize these relationships in a lower-dimensional space. and are the sets of feature representations for the bridging and non-bridging ST images to formalize this process. The embedding of D in a lower-dimensional space is given by
- [Normalized KL Divergence and Information Distinguishability] While the KL divergence offers valuable insight into the separability of feature distributions, its unbounded nature can make comparisons across different datasets or conditions challenging. To address this, the Information Distinguishability measure proposed by Soffi et al. [65] is adopted, which normalizes the KL divergence into a bounded metric between 0 and 1:
2.3.2. Semi-Supervised Learning with Deep Subspace Descriptor
- [Algorithmic Framework] The framed algorithm follows this structure by applying stochastic augmentation to dataset SL, then processing it through a transfer learning approach using the Inception-ResNet-v2 network with fine-tuning and dropout [59,60,61]. The output from this network is then passed through a deep subspace descriptor, specifically, the deep embedding clustering algorithm [69,70], which further reduces the dimensionality and generates embeddings as illustrated in Figure 4.To effectively integrate labeled and unlabeled data, our approach employs a sophisticated semi-supervised learning framework that maximizes the utility of both datasets. This integration occurs through several key steps:
- Data Augmentation: Both labeled and unlabeled datasets undergo stochastic augmentation. This step enhances the diversity and robustness of the training samples by introducing variations in the data, such as rotations, translations, and noise, which helps the model generalize better to unseen data.
- Feature Extraction: Features from both labeled and unlabeled datasets are extracted using the Inception-ResNet-v2 network. This deep learning model is fine-tuned to adapt to the specific task at hand, ensuring that the extracted features are relevant and informative for distinguishing between bridging and non-bridging ST shapes.
- Joint Training: The extracted features from both datasets are jointly utilized in the training process. The model performs supervised learning for the labeled data , using the ground truth labels to guide the learning process. This helps the model to learn the direct mappings from input features to the desired output labels. Simultaneously, the unlabeled data are used in unsupervised learning, where the model learns to identify and group similar patterns within the data. This unsupervised learning process aids in discovering the underlying structure of the data, which might not have been apparent from the labeled data alone.
- Deep Embedding Clustering: The deep subspace descriptor is employed to combine the strengths of supervised and unsupervised learning. This involves using a deep embedding clustering algorithm, which reduces the dimensionality of the feature space while preserving its essential characteristics. The algorithm generates embeddings representing the data in a lower-dimensional space, facilitating the clustering of labeled and unlabeled data. This clustering helps form a more coherent and discriminative feature space, improving the overall classification performance.
- [Deep Embedding and Clustering Process] The deep embedding layer operates similarly to t-SNE, as it seeks to find a lower-dimensional representation that preserved the clustering structure of the data. It accomplishes this by employing a clustering loss that refines the embeddings for better clustering as measured by the similarity between an embedding and the mean of a cluster :The use of non-linear embeddings, analyzed with KL divergence, ensures that the essential data structures are preserved while making the embeddings robust to shifts and variations. Non-linear embeddings capture complex patterns in the data, effectively reducing dimensionality and preserving important relationships. KL divergence, used as a clustering loss, helps maintain the structural integrity of the data by aligning the learned embeddings with the true data distribution. This combination of techniques enhances the clustering performance and ensures robust, shift-invariant embeddings. In all the experiments, the hyperparameter is set to 1.
2.3.3. Enhancing Model Robustness with Zero-Shot Classifier
- [Zero-Shot Classifier Integrated with KL Divergence] To execute using KL divergence loss, a CNN is initially trained on the existing data to extract significant features from the input examples. This pre-trained CNN produces robust feature representations , which serve as the foundation for subsequent classification tasks. Each class, including both known and unknown categories, is characterized by a semantic vector. These vectors are constructed using deep visual features extracted from pre-trained CNNs and enhanced with supplementary information such as textual descriptions or hierarchical attributes. The semantic vectors encapsulate the characteristics and relationships among various classes within a high-dimensional semantic space, thereby enabling the model to deduce the existence of unseen classes based on their semantic resemblance to familiar categories.The KL divergence loss plays a crucial role in this process by assessing the difference between the predicted probability distribution of input data embeddings and the target distribution defined by semantic vectors of known classes. As outlined in Equation (2), the KL divergence measures the disparity between the predicted distribution and the target distribution , and is defined asThrough iterative minimization of the KL divergence between predicted and target distributions, the model develops the ability to generalize effectively beyond the training data. This enables the classifier to map embeddings to appropriate class labels , even without explicit examples of those classes. The capacity to identify unseen classes is achieved by utilizing the semantic relationships encoded in the model, allowing it to deduce the existence of new classes based on their resemblance to known classes. This approach enhances the model’s resilience and adaptability, enabling it to function effectively in varied and dynamic settings where new classes may continually emerge. By aligning the learned embeddings with semantic distributions, the framework ensures enhanced performance in recognizing novel and unobserved classes, offering a scalable solution for real-world applications where data for certain classes may be limited or unavailable.
2.4. Experimental Settings
3. Results
3.1. Experimental Results of Binary Classification
3.2. Experimental Validation Results of Classification
3.3. Error Analysis
4. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
ST | Sella Turcica |
SSL | Semi-Supervised Learning |
SSLDSE | Semi-Supervised Deep Subspace Embedding |
KL divergence | Kullback–Leibler Divergence |
t-SNE | t-Distributed Stochastic Neighbor Embedding |
CNN | Convolutional Neural Network |
AL | Active Learning |
GAN | Generative Adversarial Network |
CL | Contrastive Learning |
Probability Density Function | |
ID | Information Distinguishability |
ZsC | Zero-Shot Classifier |
ROC | Receiver Operating Characteristic |
References
- Khouw, F.; Proffit, W.; White, R. Cephalometric evaluation of patients with dentofacial disharmonies requiring surgical correction. Oral Surgery Oral Med. Oral Pathol. 1970, 29, 789–798. [Google Scholar] [CrossRef] [PubMed]
- Alkofide, E.A. The shape and size of the sella turcica in skeletal Class I, Class II, and Class III Saudi subjects. Eur. J. Orthod. 2007, 29, 457–463. [Google Scholar] [CrossRef] [PubMed]
- Tekiner, H.; Acer, N.; Kelestimur, F. Sella turcica: An anatomical, endocrinological, and historical perspective. Pituitary 2015, 18, 575–578. [Google Scholar] [CrossRef] [PubMed]
- Shakya, K.S.; Jaiswal, M.; Priti, K.; Alavi, A.; Kumar, V.; Li, M.; Laddi, A. A novel SM-Net model to assess the morphological types of Sella Turcica using Lateral Cephalogram 2022. Available online: https://www.researchsquare.com/article/rs-2046354/v1 (accessed on 13 June 2024).
- Sathyanarayana, H.P.; Kailasam, V.; Chitharanjan, A.B. Sella turcica-Its importance in orthodontics and craniofacial morphology. Dent. Res. J. 2013, 10, 571. [Google Scholar]
- Shakya, K.S.; Laddi, A.; Jaiswal, M. Automated methods for sella turcica segmentation on cephalometric radiographic data using deep learning (CNN) techniques. Oral Radiol. 2023, 39, 248–265. [Google Scholar] [CrossRef]
- Teal, J. Radiology of the adult sella turcica. Bull. Los Angeles Neurol. Soc. 1977, 42, 111–174. [Google Scholar]
- Camp, J.D. The normal and pathologic anatomy of the sella turcica as revealed by roentgenograms. Am. J. Roentgenol. Radium Ther. 1924, 12, 143–156. [Google Scholar]
- Shakya, K.S.; Jaiswal, M.; Porteous, J.; K, P.; Kumar, V.; Alavi, A.; Laddi, A. SellaMorph-Net: A Novel Machine Learning Approach for Precise Segmentation of Sella Turcica Complex Structures in Full Lateral Cephalometric Images. Appl. Sci. 2023, 13, 9114. [Google Scholar] [CrossRef]
- Leonardi, R.; Barbato, E.; Vichi, M.; Caltabiano, M. A sella turcica bridge in subjects with dental anomalies. Eur. J. Orthod. 2006, 28, 580–585. [Google Scholar] [CrossRef]
- Khaitan, T.; Vishal; Gupta, P.; Naik, S.R.; Shukla, A.K. Morphometric Analysis of Sella Turcica and a Proposed Novel Sella Turcica Index–A Digital Lateral Cephalometric Study. Indian J. Otolaryngol. Head Neck Surg. 2024, 76, 73–77. [Google Scholar] [CrossRef]
- Kucharczyk, W. The sella turcica and parasellar region. In Magnetic Resonance Imaging of the Brain and Spine. 1996. Available online: https://archive.org/details/magneticresonanc0002unse/page/870/mode/2up (accessed on 25 June 2024).
- Shakya, K.S.; Priti, K.; Jaiswal, M.; Laddi, A. Segmentation of Sella Turcica in X-ray Image based on U-Net Architecture. Procedia Comput. Sci. 2023, 218, 828–835. [Google Scholar] [CrossRef]
- Ghasedi Dizaji, K.; Herandi, A.; Deng, C.; Cai, W.; Huang, H. Deep clustering via joint convolutional autoencoder embedding and relative entropy minimization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5736–5745. [Google Scholar]
- Van Engelen, J.E.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef]
- Shakya, K.S.; Alavi, A.; Porteous, J.; K, P.; Laddi, A.; Jaiswal, M. A Critical Analysis of Deep Semi-Supervised Learning Approaches for Enhanced Medical Image Classification. Information 2024, 15, 246. [Google Scholar] [CrossRef]
- Bennett, K.; Demiriz, A. Semi-supervised support vector machines. Adv. Neural Inf. Process. Syst. 1998, 11, 369–374. [Google Scholar]
- Seeger, M. Learning with Labeled and Unlabeled Data 2000. Available online: http://www.cs.columbia.edu/~dplewis/candidacy/seeger01learning.pdf (accessed on 25 June 2024).
- Ouali, Y.; Hudelot, C.; Tami, M. An overview of deep semi-supervised learning. arXiv 2020, arXiv:2006.05278. [Google Scholar]
- Taha, K. Semi-supervised and un-supervised clustering: A review and experimental evaluation. Inf. Syst. 2023, 114, 102178. [Google Scholar] [CrossRef]
- Li, Q.; Han, Z.; Wu, X.M. Deeper insights into graph convolutional networks for semi-supervised learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
- Ponzio, F.; Urgese, G.; Ficarra, E.; Di Cataldo, S. Dealing with lack of training data for convolutional neural networks: The case of digital pathology. Electronics 2019, 8, 256. [Google Scholar] [CrossRef]
- Abdelhafiz, D.; Yang, C.; Ammar, R.; Nabavi, S. Deep convolutional neural networks for mammography: Advances, challenges and applications. BMC Bioinform. 2019, 20, 281. [Google Scholar] [CrossRef]
- Chougrad, H.; Zouaki, H.; Alheyane, O. Deep convolutional neural networks for breast cancer screening. Comput. Methods Programs Biomed. 2018, 157, 19–30. [Google Scholar] [CrossRef]
- Kim, H.; Shim, E.; Park, J.; Kim, Y.J.; Lee, U.; Kim, Y. Web-based fully automated cephalometric analysis by deep learning. Comput. Methods Programs Biomed. 2020, 194, 105513. [Google Scholar] [CrossRef]
- Wang, C.W.; Huang, C.T.; Hsieh, M.C.; Li, C.H.; Chang, S.W.; Li, W.C.; Vandaele, R.; Marée, R.; Jodogne, S.; Geurts, P.; et al. Evaluation and comparison of anatomical landmark detection methods for cephalometric x-ray images: A grand challenge. IEEE Trans. Med. Imaging 2015, 34, 1890–1900. [Google Scholar] [CrossRef] [PubMed]
- Golhar, M.; Bobrow, T.L.; Khoshknab, M.P.; Jit, S.; Ngamruengphong, S.; Durr, N.J. Improving colonoscopy lesion classification using semi-supervised deep learning. IEEE Access 2020, 9, 631–640. [Google Scholar] [CrossRef] [PubMed]
- Ha, Y.; Meng, X.; Du, Z.; Tian, J.; Yuan, Y. Semi-supervised graph learning framework for apicomplexan parasite classification. Biomed. Signal Process. Control 2023, 81, 104502. [Google Scholar] [CrossRef]
- Zhang, X.Y.; Shi, H.; Zhu, X.; Li, P. Active semi-supervised learning based on self-expressive correlation with generative adversarial networks. Neurocomputing 2019, 345, 103–113. [Google Scholar] [CrossRef]
- Moradi, E.; Pepe, A.; Gaser, C.; Huttunen, H.; Tohka, J.; Alzheimer’s Disease Neuroimaging Initiative. Machine learning framework for early MRI-based Alzheimer’s conversion prediction in MCI subjects. Neuroimage 2015, 104, 398–412. [Google Scholar] [CrossRef]
- Su, H.; Shi, X.; Cai, J.; Yang, L. Local and global consistency regularized mean teacher for semi-supervised nuclei classification. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 559–567. [Google Scholar]
- Zhou, Y.; Chen, H.; Lin, H.; Heng, P.A. Deep semi-supervised knowledge distillation for overlapping cervical cell instance segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Proceedings, Part I 23. pp. 521–531. [Google Scholar]
- Li, Y.; Luo, L.; Lin, H.; Chen, H.; Heng, P.A. Dual-consistency semi-supervised learning with uncertainty quantification for COVID-19 lesion segmentation from CT images. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part II 24. pp. 199–209. [Google Scholar]
- Li, C.H.; Yuen, P.C. Semi-supervised learning in medical image database. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Hong Kong, China, 16–18 April 2001; pp. 154–160. [Google Scholar]
- Filipovych, R.; Davatzikos, C.; For the Alzheimer’s Disease Neuroimaging Initiative. Semi-supervised pattern classification of medical images: Application to mild cognitive impairment (MCI). NeuroImage 2011, 55, 1109–1119. [Google Scholar] [CrossRef]
- Batmanghelich, K.N.; Dong, H.Y.; Pohl, K.M.; Taskar, B.; Davatzikos, C. Disease classification and prediction via semi-supervised dimensionality reduction. In Proceedings of the 2011 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Chicago, IL, USA, 30 March–2 April 2011; pp. 1086–1090. [Google Scholar]
- Batmanghelich, N.K.; Taskar, B.; Davatzikos, C. Generative-discriminative basis learning for medical imaging. IEEE Trans. Med. Imaging 2011, 31, 51–69. [Google Scholar] [CrossRef]
- Culotta, A.; McCallum, A. Reducing labeling effort for structured prediction tasks. In Proceedings of the AAAI, Pittsburgh, PA, USA, 9–13 July 2005; Volume 5, pp. 746–751. [Google Scholar]
- Settles, B.; Craven, M. An analysis of active learning strategies for sequence labeling tasks. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, Honolulu, HI, USA, 25–27 October 2008; pp. 1070–1079. [Google Scholar]
- Melville, P.; Mooney, R.J. Diverse ensembles for active learning. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004; p. 74. [Google Scholar]
- Zhang, X.Y.; Wang, S.; Zhu, X.; Yun, X.; Wu, G.; Wang, Y. Update vs. upgrade: Modeling with indeterminate multi-class active learning. Neurocomputing 2015, 162, 163–170. [Google Scholar] [CrossRef]
- Zhang, X.Y.; Wang, S.; Yun, X. Bidirectional active learning: A two-way exploration into unlabeled and labeled data set. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 3034–3044. [Google Scholar] [CrossRef]
- Madani, A.; Ong, J.R.; Tibrewal, A.; Mofrad, M.R. Deep echocardiography: Data-efficient supervised and semi-supervised deep learning towards automated diagnosis of cardiac disease. NPJ Digit. Med. 2018, 1, 1–11. [Google Scholar] [CrossRef]
- Mirza, M. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
- Odena, A. Semi-supervised learning with generative adversarial networks. arXiv 2016, arXiv:1606.01583. [Google Scholar]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144. [Google Scholar] [CrossRef]
- Radford, A. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
- Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning. Sydney, NSW, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
- Shrivastava, A.; Pfister, T.; Tuzel, O.; Susskind, J.; Wang, W.; Webb, R. Learning from simulated and unsupervised images through adversarial training. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2107–2116. [Google Scholar]
- Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
- Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent-a new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
- Chen, X.; He, K. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 15750–15758. [Google Scholar]
- Caron, M.; Misra, I.; Mairal, J.; Goyal, P.; Bojanowski, P.; Joulin, A. Unsupervised learning of visual features by contrasting cluster assignments. Adv. Neural Inf. Process. Syst. 2020, 33, 9912–9924. [Google Scholar]
- Lu, M.Y.; Chen, R.J.; Wang, J.; Dillon, D.; Mahmood, F. Semi-supervised histology classification using deep multiple instance learning and contrastive predictive coding. arXiv 2019, arXiv:1910.10825. [Google Scholar]
- Wang, X.; Tang, F.; Chen, H.; Cheung, C.Y.; Heng, P.A. Deep semi-supervised multiple instance learning with self-correction for DME classification from OCT images. Med. Image Anal. 2023, 83, 102673. [Google Scholar] [CrossRef]
- Chen, Y.; Liginlal, D. A maximum entropy approach to feature selection in knowledge-based authentication. Decis. Support Syst. 2008, 46, 388–398. [Google Scholar] [CrossRef]
- Deng, X.; Cai, P.; Cao, Y.; Wang, P. Two-step localized kernel principal component analysis based incipient fault diagnosis for nonlinear industrial processes. Ind. Eng. Chem. Res. 2020, 59, 5956–5968. [Google Scholar] [CrossRef]
- Cappelli, R.; Maltoni, D. Multispace KL for pattern representation and classification. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 977–996. [Google Scholar] [CrossRef]
- Banerjee, A.; Bhattacharya, R.; Bhateja, V.; Singh, P.K.; Sarkar, R. COFE-Net: An ensemble strategy for computer-aided detection for COVID-19. Measurement 2022, 187, 110289. [Google Scholar] [CrossRef] [PubMed]
- Ananda, A.; Ngan, K.H.; Karabağ, C.; Ter-Sarkisov, A.; Alonso, E.; Reyes-Aldasoro, C.C. Classification and visualisation of normal and abnormal radiographs; a comparison between eleven convolutional neural network architectures. Sensors 2021, 21, 5381. [Google Scholar] [CrossRef] [PubMed]
- Demir, A.; Yilmaz, F. Inception-ResNet-v2 with LeakyReLU and averagepooling for more reliable and accurate classification of chest X-ray images. In Proceedings of the 2020 Medical Technologies Congress (TIPTEKNO), Antalya, Turkey, 19–20 November 2020; pp. 1–4. [Google Scholar]
- Alsayed, A.; Arif, M.; Qadah, T.M.; Alotaibi, S. A Systematic Literature Review on Using the Encoder-Decoder Models for Image Captioning in English and Arabic Languages. Appl. Sci. 2023, 13, 10894. [Google Scholar] [CrossRef]
- Shen, F.; Shen, C.; Shi, Q.; Van den Hengel, A.; Tang, Z.; Shen, H.T. Hashing on nonlinear manifolds. IEEE Trans. Image Process. 2015, 24, 1839–1851. [Google Scholar] [CrossRef]
- Sarfraz, S.; Koulakis, M.; Seibold, C.; Stiefelhagen, R. Hierarchical nearest neighbor graph embedding for efficient dimensionality reduction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 336–345. [Google Scholar]
- Soofi, E.S.; Ebrahimi, N.; Habibullah, M. Information distinguishability with application to analysis of failure data. J. Am. Stat. Assoc. 1995, 90, 657–668. [Google Scholar] [CrossRef]
- Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 478–487. [Google Scholar]
- Gan, H.; Sang, N.; Huang, R.; Tong, X.; Dan, Z. Using clustering analysis to improve semi-supervised classification. Neurocomputing 2013, 101, 290–298. [Google Scholar] [CrossRef]
- Bair, E. Semi-supervised clustering methods. Wiley Interdiscip. Rev. Comput. Stat. 2013, 5, 349–361. [Google Scholar] [CrossRef]
- Cao, W.; Zhang, Z.; Liu, C.; Li, R.; Jiao, Q.; Yu, Z.; Wong, H.S. Unsupervised discriminative feature learning via finding a clustering-friendly embedding space. Pattern Recognit. 2022, 129, 108768. [Google Scholar] [CrossRef]
- Hou, C.; Nie, F.; Yi, D.; Tao, D. Discriminative embedded clustering: A framework for grouping high-dimensional data. IEEE Trans. Neural Netw. Learn. Syst. 2014, 26, 1287–1299. [Google Scholar]
- Li, X.; Guo, Y.; Schuurmans, D. Semi-supervised zero-shot classification with label representation learning. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4211–4219. [Google Scholar]
- Shojaee, S.M.; Baghshah, M.S. Semi-supervised zero-shot learning by a clustering-based approach. arXiv 2016, arXiv:1605.09016. [Google Scholar]
- Li, G.Z.; Yang, J.Y.; Lu, W.C.; Li, D.; Yang, M.Q. Improving prediction accuracy of drug activities by utilising unlabelled instances with feature selection. Int. J. Comput. Biol. Drug Des. 2008, 1, 1–13. [Google Scholar] [CrossRef]
Epochs | Precision | Recall | F1 Score | |||
---|---|---|---|---|---|---|
Brd | NBrd | Brd | NBrd | Brd | NBrd | |
50 | 88.03% | 91.45% | 90.00% | 93.25% | 90.12% | 91.73% |
100 | 86.82% | 90.30% | 90.93% | 88.75% | 88.45% | 96.47% |
150 | 87.25% | 85.75% | 92.51% | 89.60% | 89.01% | 92.85% |
200 | 87.79% | 92.10% | 91.42% | 95.40% | 89.29% | 93.29% |
250 | 85.42% | 86.50% | 90.77% | 96.10% | 88.49% | 96.82% |
300 | 87.17% | 90.00% | 92.79% | 91.85% | 89.65% | 94.51% |
Average | 87.08% | 89.62% | 91.37% | 92.73% | 89.17% | 94.14% |
Epochs | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
50 | 98.50% | 97.25% | 96.80% | 98.50% |
100 | 96.75% | 95.40% | 94.20% | 96.30% |
150 | 97.10% | 96.75% | 95.00% | 97.90% |
200 | 97.90% | 95.85% | 94.85% | 97.25% |
250 | 96.25% | 96.00% | 96.10% | 98.10% |
300 | 97.95% | 94.50% | 94.50% | 97.60% |
Average | 97.28% | 96.12% | 95.65% | 97.77% |
Epochs | GAN | Contrastive | Active | MIncepResV2 | SSLDSE | |||||
---|---|---|---|---|---|---|---|---|---|---|
Brd | NBrd | Brd | NBrd | Brd | NBrd | Brd | NBrd | Brd | NBrd | |
50 | 68.95% | 75.25% | 77.85% | 83.65% | 82.45% | 85.45% | 90.35% | 92.15% | 95.85% | 97.80% |
100 | 70.40% | 73.80% | 78.60% | 84.90% | 80.90% | 84.75% | 88.75% | 90.85% | 94.30% | 96.50% |
150 | 69.20% | 74.45% | 79.20% | 82.75% | 81.30% | 86.30% | 89.10% | 91.60% | 95.50% | 98.10% |
200 | 71.05% | 76.10% | 77.30% | 85.20% | 83.15% | 83.85% | 90.85% | 93.25% | 96.10% | 97.25% |
250 | 67.75% | 72.90% | 78.95% | 83.40% | 79.75% | 85.90% | 87.65% | 92.70% | 93.95% | 96.90% |
300 | 71.55% | 75.15% | 78.95% | 84.85% | 82.55% | 84.35% | 88.95% | 90.85% | 95.45% | 98.00% |
Average | 69.82% | 74.77% | 78.29% | 84.13% | 81.67% | 85.10% | 89.27% | 91.92% | 95.19% | 97.37% |
Models | Accuracy | AUC-ROC | F1 Score | |||
---|---|---|---|---|---|---|
IEEE | Our | IEEE | Our | IEEE | Our | |
GAN SSL | 68.61% | 68.93% | 64.88% | 65.73% | 62.86% | 67.01% |
Contrastive SSL | 70.66% | 73.11% | 74.40% | 74.07% | 72.06% | 73.17% |
Active SSL | 84.92% | 85.64% | 83.73% | 83.00% | 83.00% | 85.07% |
MIncepResV2 | 85.07% | 86.12% | 85.97% | 84.03% | 83.02% | 82.00% |
Proposed SSLDSE | 90.06% | 92.72% | 91.02% | 93.57% | 92.08% | 95.37% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Shakya, K.S.; Alavi, A.; Porteous, J.; Khatri, P.; Laddi, A.; Jaiswal, M.; Kumar, V. Semi-Supervised Deep Subspace Embedding for Binary Classification of Sella Turcica. Appl. Sci. 2024, 14, 11154. https://doi.org/10.3390/app142311154
Shakya KS, Alavi A, Porteous J, Khatri P, Laddi A, Jaiswal M, Kumar V. Semi-Supervised Deep Subspace Embedding for Binary Classification of Sella Turcica. Applied Sciences. 2024; 14(23):11154. https://doi.org/10.3390/app142311154
Chicago/Turabian StyleShakya, Kaushlesh Singh, Azadeh Alavi, Julie Porteous, Priti Khatri, Amit Laddi, Manojkumar Jaiswal, and Vinay Kumar. 2024. "Semi-Supervised Deep Subspace Embedding for Binary Classification of Sella Turcica" Applied Sciences 14, no. 23: 11154. https://doi.org/10.3390/app142311154
APA StyleShakya, K. S., Alavi, A., Porteous, J., Khatri, P., Laddi, A., Jaiswal, M., & Kumar, V. (2024). Semi-Supervised Deep Subspace Embedding for Binary Classification of Sella Turcica. Applied Sciences, 14(23), 11154. https://doi.org/10.3390/app142311154