[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Reliability of IMU-Based Gait Assessment in Clinical Stroke Rehabilitation
Next Article in Special Issue
An Effective Skin Cancer Classification Mechanism via Medical Vision Transformer
Previous Article in Journal
Content-Aware SLIC Super-Pixels for Semi-Dark Images (SLIC++)
Previous Article in Special Issue
CAT: Centerness-Aware Anchor-Free Tracker
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Learning Model for Multiscale Texture Features in Polyp Classification via Computed Tomographic Colonography

1
Department of Radiology, Stony Brook University, Stony Brook, NY 11794, USA
2
Department of Biomedical Engineering, Stony Brook University, Stony Brook, NY 11794, USA
3
Department of Computer Science, City University of New York, New York, NY 10314, USA
4
Department of Radiology, School of Medicine, University of Wisconsin, Madison, WI 53792, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(3), 907; https://doi.org/10.3390/s22030907
Submission received: 1 December 2021 / Revised: 14 January 2022 / Accepted: 20 January 2022 / Published: 25 January 2022
(This article belongs to the Special Issue Artificial Intelligence in Medical Imaging and Visual Sensing)
Figure 1
<p>Illustration of co-occurrence matrix (CM) calculation in 2D/3D images: (<b>a</b>) CM parameters in 2D images; (<b>b</b>) CM parameters in 3D images; and (<b>c</b>) A GLCM example of a 2D case when direction is 0° and displacement = 1. The left is a gray image, and the right one is its GLCM.</p> ">
Figure 2
<p>Visualization of information CNN learnt from each subgroups: (<b>a</b>) <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>1</mn> </msub> </mrow> </semantics></math>, (<b>b</b>) <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>2</mn> </msub> </mrow> </semantics></math> and (<b>c</b>) <math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mn>3</mn> </msub> </mrow> </semantics></math>. The first column is the original GLCM. The corresponding label (0 for benign and 1 for malignant) and model score of the malignancy risk are listed on the top. The remaining two columns are the interpretations of model prediction on the two classes. The red cells show the entries push the model’s decision close to that class, while blue pixels pull the prediction results away.</p> ">
Figure 3
<p>The flowchart of multi-level learning model for fusion of multi-scale feature sets.</p> ">
Figure 4
<p>The flowchart of the feature selection step for the baseline and the complement in the multi-layer learning model.</p> ">
Figure 5
<p>Network structure of FSFS-CNN.</p> ">
Figure 6
<p>Flowchart of data acquisition and preparation for these experiments.</p> ">
Figure 7
<p>Three sample CT slices from select polyp masses. Green contour around the polyp show the segmentation. Air voxels from the lumen below −450 HU are removed post-segmentation and are highlighted red in the images. Images show sample polyps with pathologies (<b>a</b>) adenocarcinoma, (<b>b</b>) villous adenoma, and (<b>c</b>) villous adenoma.</p> ">
Figure 8
<p>The trends of three AUC score curves of polyp classification, their maximums and their partitions over 63 polyps via forward step feature selection method: (<b>a</b>) D1, (<b>b</b>) D2, and (<b>c</b>) D3.</p> ">
Figure 9
<p>ROC curves of proposed and comparative methods.</p> ">
Versions Notes

Abstract

:
Objective: As an effective lesion heterogeneity depiction, texture information extracted from computed tomography has become increasingly important in polyp classification. However, variation and redundancy among multiple texture descriptors render a challenging task of integrating them into a general characterization. Considering these two problems, this work proposes an adaptive learning model to integrate multi-scale texture features. Methods: To mitigate feature variation, the whole feature set is geometrically split into several independent subsets that are ranked by a learning evaluation measure after preliminary classifications. To reduce feature redundancy, a bottom-up hierarchical learning framework is proposed to ensure monotonic increase of classification performance while integrating these ranked sets selectively. Two types of classifiers, traditional (random forest + support vector machine)- and convolutional neural network (CNN)-based, are employed to perform the polyp classification under the proposed framework with extended Haralick measures and gray-level co-occurrence matrix (GLCM) as inputs, respectively. Experimental results are based on a retrospective dataset of 63 polyp masses (defined as greater than 3 cm in largest diameter), including 32 adenocarcinomas and 31 benign adenomas, from adult patients undergoing first-time computed tomography colonography and who had corresponding histopathology of the detected masses. Results: We evaluate the performance of the proposed models by the area under the curve (AUC) of the receiver operating characteristic curve. The proposed models show encouraging performances of an AUC score of 0.925 with the traditional classification method and an AUC score of 0.902 with CNN. The proposed adaptive learning framework significantly outperforms nine well-established classification methods, including six traditional methods and three deep learning ones with a large margin. Conclusions: The proposed adaptive learning model can combat the challenges of feature variation through a multiscale grouping of feature inputs, and the feature redundancy through a hierarchal sorting of these feature groups. The improved classification performance against comparative models demonstrated the feasibility and utility of this adaptive learning procedure for feature integration.

1. Introduction

Colorectal cancer (CRC) is one of the top fatal diseases in the United States. American Cancer Society ranks CRC as the third most common cancer and the third leading cause of cancer-related deaths in both men and women [1]. Because most colon cancers are developed from precursor polyps, polyp screening has become the primary means for CRC prevention [2,3]. Computed tomographic colonography (CTC), as a minimally invasive polyp detection tool, has become an important alternative method in polyp screening and management [4]. This CTC technology has shown its potential in physicians’ hands with computer-aided detection tools to localize polyps in practice [5]. For personalized precision medicine, it is very important to know the pathological subtype of (or to diagnose) the detected polyps for optimal treatment. Yet, diagnosis from the CTC images with subtle image contrast within the polyp volume is very challenging in the radiologist experts’ hands with current existing computer aided diagnosis (CADx) tools [6]. Therefore, more advanced CADx tools are needed.
Heterogeneity is a key factor in determining the malignancy of a lesion and its response to intervention [7,8]; CADx models have, therefore, focused on using image contrast patterns to quantify and describe that heterogeneity [9,10,11]. Toward that purpose, many texture patterns and descriptors have been proposed to extract and quantify texture information for CADx of polyps, such as gray level co-occurrence matrix (GLCM) [10], local binary pattern (LBP) [12,13,14], Gabor filter [15,16], wavelets [17,18], and Weber local descriptor (WLD) [14,19]. With the growth of the number of texture descriptors, how to merge them has become an important issue because of the variation in computing these texture descriptors, as well as possible redundant information among them [20,21,22].
The variation problem can occur from the method used to compute the texture patterns or to extract the texture features. For example, the traditional Haralick texture features [10] were computed by the average and range across the 13 neighbor directions through the volumetric voxel array that did not consider geometrical scale variations (or multi-scale nature) among those neighbor directions [11]. The redundancy problem has been under investigation for many years as a feature selection or dimension reduction task. This is a typical NP (non-deterministic polynomial time) hard problem [22] that has generally been approached by three classes of methods: filter methods [23,24], wrapper methods [25,26], and embedded methods [27,28,29,30,31,32,33]. Despite these advancements, how to obtain the best feature subset from all the extracted patterns and features still remains a great challenge [34,35,36,37].
To address the above, the problems of (1) variation in polyp texture descriptor computation and (2) redundancy in multiple computed patterns and features, this work proposes an adaptive fusion model for the polyp classification task. This work uses the GLCM and its measures to demonstrate this adaptive fusion model, where the second-order nature of the GLCM allows for inherent grouping by spatial distance of image voxel pairs. Two different models, using traditional classifiers and a deep learning architecture, are proposed to evaluate this adaptive fusion methodology, where input features are grouped by spatial displacement and weighted by initial classification performance. By adaptively combining these groups in order of their initial weights, the proposed models can select the most important features from the inputs, thereby reducing variation and redundancy in the final model. Both models are evaluated on a dataset of colorectal polyp masses and show significant improvement in classification performance compared to state-of-the-art methods.

2. Materials and Methods

This section begins with a review of the GLCM texture descriptor calculation. Then the multiscale analysis for fusing the calculated texture descriptor sets is discussed. Thereafter, the adaptive learning model is presented and analyzed.

2.1. Multiscale Sampling of GLCMs for Multiscale Features

Gray level co-occurrence matrix or GLCM as a typical texture pattern descriptor is widely used in medical imaging [9,10,11]. Its computation could be referred to according to the following expression in two-dimensional (2D) representation:
C i , j   ( d , θ ) = m = 1 M n = 1 N { 1 I ( m , n ) = i     &     I ( ( m , n ) + d ( θ ) ) = j   0 o t h e r w i s e
where I is the gray level image, (M,N) is the image size, indices i and j represent a pair of image pixel values, and d(θ) is a shifting vector between two concerned points along the direction θ , such that d ( θ ) = d ( c o s θ , s i n θ ) . For 3D volumetric image data, the definition of GLCM is very similar, except that there are two angular variables in contrast to one in the 2D model. Their calculations are shown in Figure 1.
In a digital image array, the first- and second-order neighbors, which comprise the first ring around the center image voxel, are most frequently used for vector calculation. A voxel in 3D volumetric data generally has 26 neighbors, which could produce 26 vectors, including 13 vectors and 13 negative vectors. From Equation (1), it is easy to prove that the GLCM of one vector is equal to the transposed GLCM of its negative vector. Therefore, only 13 directions are preserved, while their negative vectors are all neglected in GLCM calculation due to redundant information, as shown in Figure 1b. Moreover, only the 1st ring neighbor around one concerned voxel is used; the gray level is set to be 32 in the calculation.
Many statistical measures have been proposed to quantify each GLCM for texture features. Haralick et al. proposed 14 measures, which are called Haralick measures (HMs) [10]. Hu et al. then added 16 new measures based on HM, which are donated by extended Haralick measures (eHMs) [11]. In this article, only 28 of the 30 measures from eHM are used to construct the texture descriptors (two of the 30 were proved to have limited new information and are ignored [38]) and are generated using in-house software. Therefore, the GLCM-descriptor contains 364 variables from 28 HMs over 13 directions, expressed by:
D = ( d 1 ,   ,   d 364 )  
Geometrically, the distance between the cubic center (of the first- and second-order voxel array) and the center of one neighbor voxel is not a constant and varies between 1 and 3 in terms of the voxel side unit. For example, d ( .   ) = 1 for the directions along x, y and z axes, d ( . ) = 2 for the diagonal directions in the 2D planes of the 3D x-y-z array coordinates, and d ( . ) = 3 for the diagonal directions in the 3D x-y-z array coordinates. In other words, in the discrete volumetric data, twenty-six neighbors around one voxel could produce three distances of 1, 2 , and 3 , i.e., a multi-scale data sampling nature. The 13 directions used to compute the GLCMs could be divided into 3 subgroups, i.e., D 1 , D 2 , and D 3 , according to their geometric distances. Each direction within the subgroup, therefore, shares the same geometric sampling distance. Figure 1b gives the geometric interpretation. G 1 (green) contains three directions, G 2 (red) contains six directions, and, lastly, G 3 (blue) contains four directions from this subdivision. The three GLCM groups would produce three descriptors, where their corresponding variable numbers are 84 (28*3 eHMs from G 1 ), 168 (28*6 eHMs from G 2 ), and 112 (28*4 eHMs from G 3 ). In this manuscript, the groups of GLCMs will be given the notation G i , and the groups of texture descriptors given the notation D i . These descriptors could further be written by:
{ D 1 = ( d 1 1 , ,   d 84 1 ) D 2 = ( d 1 2 , , d 168 2 ) D 3 = ( d 1 3 , , d 112 3 )
The traditional Haralick texture feature calculation considered these three direction groups as one scale by computing the average and range across all 13 directions for each of the 14 traditional HMs, resulting in a total of 28 traditional Haralick texture features (HFs). For the 28 eHMs, the average and range across all 13 directions result in a total of 56 extended HFs, called eHFs. These Haralick texture features will be used as the baseline reference in this work to show the gain by the consideration of the multi-scale data sampling nature in the following. The GLCMs are then calculated by three different scales, i.e., 1, 2 ≈ 1.414 and 3 ≈ 1.732, as shown in Figure 1b. Essentially, this multi-scaling feature extraction operation is not only a direction subgrouping but also a feature subdivision. Therefore, this method generates three GLCM subgroups and three texture descriptor subdivisions, each with a different scale, as shown in Table 1. In the following, the variables in each direction group are labeled as a set of data sampled from the polyp object and treat all three direction group datasets as three differently sampled data from the same subject. Then, an adaptive machine learning strategy is developed to integrate these different datasets together for improved CADx performance by circumventing the two problems of (1) variation in polyp texture descriptor computation and (2) redundancy in multi-scale computed features.

2.2. Analyze Group-Specific Information

To analyze and compare the differences among the three data subsets or multi-scale groups, the information provided by each group is then investigated. To understand these differences, the information that can be learnt by CNN on each individual group is first visually analyzed. Next, CNN models based on three GLCM subgroups are trained. Then, features learnt by CNN are understood via interpreting how the final decision is made given an input.
To accomplish this, a game theory based model called SHAP was adapted to explain the output of the machine learning models [39]. Each model was trained by the polyps’ corresponding GLCM subgroup and is similar to GLCM-CNN, with network design optimized to the subgroups [40]. After the CNN model was trained, the decision criteria was visualized on the testing dataset using SHAP. Figure 2 demonstrates the learnt feature from the three subgroups by explaining the decision result of one representative polyp. The first column is the original GLCM. The corresponding label (0 for benign and 1 for malignant) and model score of the malignancy risk are listed on the top. The remaining two columns show the interpretation of model prediction on the two classes. Given a class, the red cells showed that the entries pushed the model’s decision close to that class, while blue pixels pulled the prediction results away. Based on this visualization, it can be observed that the information provided by the three subgroups had both shared patterns and unique patterns. The visualization results of these patterns from deep learning showed the potential for the proposed adaptive learning model to learn these group specific and groupwise shared features.

2.3. Adaptive Learning Model for Fusing Multi-Scale Features

As the variable number grows, simply combining all the input variables for classification can increase a high risk of clustering degradation, which is caused by counteractions of their variations [20,22]. In practice, not all variables of the descriptor will be useful for classification; lots of redundant information remains in the three scales. Inspired by [38], an adaptive learning model is designed to hierarchically circumvent the variation and reduce the redundant information from the multi-scale feature sets.
Problem Formulation: The problem is formulated as follows: Given a set S = { D i   |   i   ϵ   [ 1 ,   n ] } containing n feature groups D i , the task is to find an optimal set S ^   S that maximizes the polyp classification performance in terms of AUC. Actually, this is a famous problem of the curse of dimensionality, which is always NP-hard [41]. To avoid this problem, the greedy algorithm as the suboptimal scheme is introduced.
As shown in Figure 3, the proposed adaptive learning method works in two stages: baseline selection and hierarchical feature integration. The goal of the baseline is to select the best individual group that achieves the highest performance. After ranking the rest feature groups in a descending order based on its individual performance, the multi-level integration method integrates new group one by one following the forward step feature selection (FSFS) method. Given a new feature group D j , FSFS is designed to add new variables from the most significant to the least and to only keep the ones that have performance improvement.
Two models for the adaptive learning method are proposed. The first one is a traditional hybrid method; the second is a deep learning-based method. They are detailed below.
Multigroup hybrid Method: The multigroup hybrid model (MGHM) was designed with random forest for priority calculations and a support vector machine (SVM) for final classification.
For the baseline selection, as each group contained several descriptors, each group was compared by its best performance after feature selection. Separate random forest models were trained on each group; the importance of each feature was based on the GINI index [42], meaning that the information gain it could provide for each involved splitting. Then, in each group, an optimal subset that had the highest performance by AUC was found via SVM, while, naturally, the left-over variables built the complimentary set. D i 0 was used to denote the baseline set and D i 1 to denote the left-over set for group D i . The optimal set that had the highest AUC was selected as the initial baseline; then, the proposed multi-level feature integration was performed on the rest of the groups. The integration sequence was in a descending order of the pre-evaluated AUC on the whole group level. This ranked set of descriptor groups was hereafter referred to as the descriptor pool (DP).
Since there were three descriptor groups, the proposed hierarchical feature integration contained 4 levels. FSFS was performed on each level to find the optimal feature subset as output with support vector machine (SVM) as the classifier and the AUC as the metric, for which cross-validation evaluation was performed. Level i in the hierarchy model is denoted as L i , the current baseline is denoted as B a s e l i n e i , and the next candidate descriptor group in L i is denoted as C a n d i d a t e i . The output of L i , denoted as B a s e l i n e i + 1 , served as the baseline of L i + 1 . Its flow chart is plotted in Figure 4. After all candidate sets were integrated, FSFS was run to integrate the complementary set of the initial baseline.
As this method was designed to iteratively evaluate every variable, it served as the upper-bound of the performance that can be achieved on the dataset.
Multi-group CNN: In the second model, CNN was adapted and performed adaptive learning by each group, as shown in Figure 5. For the baseline selection, the CNN was designed to take the whole GLCM group as input and select the one with the highest AUC. Then, the integration was performed by iteratively adding a group with the next highest AUC following FSFS. The entire evaluation was based on a CNN network, where its detailed structure is listed in Table 2 and the structure of the backbone is plotted in Figure 5. For each level, the input size of the network had 32 × 32 × c, where 32 is the grayscale and c is the number of channels/GLCMs of the input. The convolution network contained two convolution layers, each followed by a batch normalization layer, a max-pooling layer with stride 2 and ReLU as activation function. After the convolution part, three fully connected layers were designed to make a final prediction. For different group combinations, the number of input channels were modified to fit the current input data. This multi-group CNN method is denoted as MG-CNN in the rest of the paper.

3. Results

In this section, the polyp mass dataset used for all experimental results is discussed in detail. The classification results of the multi-scale descriptor sets are presented with the proposed multi-level adaptive learning model. Finally, the proposed models are compared to similar classification methods which input all the multi-scale descriptor sets at once and ignore the differences among the data sets.

3.1. Polyp Dataset

The polyp dataset used for these experiments consisted of 59 patients with a total number of 63 polyp masses found through virtual colonoscopy and confirmed by clinical colonoscopy. A flowchart of the dataset acquisition and preparation is shown in Figure 6 and described below. The polyp dataset used for these experiments was obtained from a retrospective study carried out at the University of Wisconsin Hospital and Clinics, Madison, WI, USA. Over 8000 patients were screened via CTC with the inclusion criteria that the patients were at least 50 years of age (normal screening age without family history of colorectal cancer), a polyp with a size of at least 30 mm in largest diameter was detected during CTC, and corresponding histopathology was available for those polyps. The CTC imaging was carried out according to the procedures described within [43]. Of those screened patients, only 59 patients, with a total of 63 polyp masses, fit the inclusion criteria. For classification discussed below, the dataset was divided into binary categories of 32 malignant adenocarcinomas, and 31 benign polyps including 3 serrated adenomas, 2 tubular adenomas, 21 tubulovillous adenomas, and 5 villous adenomas. All polyps had bulky mass morphology, except for six (four tubulovillous and two villous adenomas), which were designated as flat or carpet polyps. The patient demographics for this polyp dataset are presented in Table 3.
The clinical value of CADx models on CTC polyp mass images is due to their requirement for surgical removal from their size. Unlike endoscopic colonoscopy, CTC is noninvasive and cannot resect polyps during the procedure. Polyp masses that are 30 mm or larger in size, however, require surgical removal and are not treated via colonoscopy. Therefore, the clinical value of examining this dataset is to provide physicians with diagnostic information on the polyp masses before their surgical removal without requiring expensive biopsy procedures. For example, surgeons may decide to be more aggressive in how much tissue they remove if the mass is malignant to ensure that any microscopic disease which may have invaded surrounding tissues can also be removed.

3.1.1. Regions of Interest

The area around the polyp region was manually selected and segmented on each CTC image slice containing the polyp. For each polyp, a volume was constructed by combining the segmentations on each slice to form the region of interest (ROI), which was confirmed by radiologists to ensure accuracy of the manual procedure. It is noted that a cleansing step was used to discard all voxels below −450 HU within these ROIs as being predominately air from the lumen of the colon [44]. The information encoded in these voxels from partial volume effects (above the range of pure air HU values) is minimal, if any, and contributes more noise to the features for classification. The ROIs were used to compute the multi-scale texture features described above. Sample polyp CT slices and their contours are shown in Figure 7.

3.1.2. Dataset Evaluation

A cross-validation strategy was used to evaluate the model performance. The leave-one-out and two-fold methods were adopted in this study to provide the two bounds of the classification performance, where the two evaluation methods were two extremes of the k-fold cross validation. The leave-one-out method tests only on one subject but trains on all the other subjects. The two-fold method trains on half the subjects and tests on the other half, which trains the model with the least data samples. This strategy is particularly attractive for small sized datasets. Results from both methods together will provide a fairer evaluation to consider the overfitting that might happen in the leave-one-out method and the lower amount of training that might happen in the two-fold method. Due to the paper length limit, only the two-fold testing results are used to show the advantage of the proposed model under the toughest conditions. The polyps were randomly divided into training and testing sets for classification with 31 polyps in the training set (15 benign and 16 malignant) and 32 polyps in the testing set (16 benign and 16 malignant). Repeating this random sampling method, 100 training and testing groups were generated to increase statistical confidence and to minimize bias. The 100 classification outcomes were averaged for the results and standard deviation (STD) served as the performance variation measurement.

3.1.3. Settings

For the traditional method, three multi-scale descriptors were calculated using the three groups in Table 1 relevant to the three scales. Then, these descriptors were used to generate 100 training and testing datasets due to the observation splitting schemes.
The Random Forest classifier contains 5000 trees with GINI index as the importance metric. The SVM classifier adapts a kernel function of cubic polynomial, with Gamma as 1/(variable number), coef0 as 0, tolerance as 0.001, and Epsilon as 0.1.
For each learning method, the i-th candidate group is denoted as D i x , where i ϵ [ 1 , 3 ] and x ϵ [ · , b , c ] as the whole group, base group and complementary group. C i , with i ϵ [ 1 , 3 ] , denotes the learned best set from stage i.
The CNN model is trained with Cross-entropy loss between the predicted score and label. Adam [45] was used for optimization. The learning rate was initialized as 0.001 and decayed by 0.01 every 10 epochs. Since the dataset was relatively small, the training ended after 40 epochs to prevent overfitting of the model.

3.2. The Outcomes of the Proposed Method

First, an investigation of how the descriptors from each group contribute to the model trained from all descriptors is analyzed. The statistics summary of the descriptors is listed in Table 1.
After acquiring the optimal subset of descriptors, the contribution of each group is analyzed by comparing how many variables contribute to the best AUC score and the importance of each descriptor. Figure 8 shows the different trends of AUC scores as a function of variable number, where the non-monotonic trend is usually seen due to the redundancy, resulting in parameter overtraining and clustering degradation. In addition, the differences among the multi-scale texture descriptors are also clearly seen.
Based on the observation above, it is necessary to evaluate each descriptor group first before combining them all together in order to avoid deterioration on the overall performance. Besides, this can prove the feasibility of the proposed learning framework.
The performance of the three groups of descriptors using the hybrid model were analyzed first. Among all, as shown in Table 4, the highest AUC is achieved by D 3 where 6 variables were chosen for this preliminary classification result. Following the proposed method, every ranked descriptor was divided into two parts, baseline and complementary set. The six generated subgroups, or the baseline and the compliment for each of the three descriptor groups, are shown in Table 5.
After the first step, based on AUC scores, DP was initialized as { D 3 b , D 3 c , D 2 b , D 2 c , D 1 b , D 1 c }. After selecting D 3 b as the initial baseline, DP became { D 3 b , D 2 , D 1 , D 3 c }. Then, DP was fed into MGHL to remove the redundant variables and to improve classification performance via the proposed bottom-up hierarchical integration. Finally, 17 out of 364 variables were extracted to form the final descriptor. In terms of classification results, the AUC score increased from 0.892 to 0.925, while its standard deviation dropped from 0.098 to 0.035. The changes in AUC score and the chosen variables are listed in Table 6, which illustrates that the hybrid model has a monotonic learning process.
The preliminary classification performances of the MG-CNN are also listed in Table 4. When compared to the results of using the whole 13 directions, the results indicated that multiple directions of GLCM could contribute to the classification performance, which means that GLCM with different directions could provide additional information.
Then, G 1 with 3 GLCMs was chosen as the baseline, with the remaining two groups to be iteratively tested for whether they should be included. Finally, three subgroups were selected and contributed to a final 0.909 AUC score. In addition, classification performance from two-scales already achieved better classification performance than using all the directions without the multiscale concept. The hierarchical learning process is shown in Table 7 and illustrates that the feature integration scheme was indeed useful to further optimize the classification performance.

3.3. Comparisons with State-of-the-Art Models

In addition to the above presentation of the performance details of the adaptive learning model for integration of multiscale texture features, the comparisons to several typical state-of-the-art models are also detailed, including:
  • Extended Haralick Measures (eHM)—this descriptor includes all the 364 variables derived from the 28 HMs over the 13 directions and disregards the multi-scale nature [11];
  • Post-KL Transformation (KLT) eHMs (eHM+KLT)—this method combines eHMs and KLT to address the variation problem due to the multi-scale nature by the KLT [11];
  • The Least Absolute Shrinkage and Selection Operator (LASSO)—a typical method of feature selection containing two steps, i.e., feature regularization and feature selection, for consideration of variation among feature datasets [28,34];
  • The SVM Method with Recursive Feature Elimination (SVM-RFE)—another typical method in feature selection, including feature ranking and feature selection for consideration of variation among feature datasets [37];
  • The Dependence Guided Unsupervised Feature Selection (DGUFS)—a new feature selection method applies the interdependence among original data, features, and labels in a joint learning framework to pick features [28];
  • VGG16—a typical deep learning method, which is fed by 20 salient slices extracted from every polyp, where the feature extraction and selection operations are considered as learning processes [40];
  • GLCM-CNN—the state-of-the-art of texture based deep learning model on the task of polyp diagnosis. It takes the whole 13-directional GLCM as input, ignoring the correlations among different groups to make decisions [40]. The network structure is optimized to fit the polyp dataset used.
Table 8 lists the classification performance of all the methods on the polyp mass dataset, where the AUC, accuracy, sensitivity, and specificity of each model is reported. The AUC score and accuracy of the proposed method exceeds that of the post-KLT eHMs (the best result of the six typical methods) by 2% and 3%, respectively. Against VGG-16, the proposed model improves the AUC score by 10%. Moreover, all ROC curves are also plotted in Figure 9, where the proposed model’s ROC curve is the top one among the seven. These ROC curves further demonstrate the advantage of the proposed method over the others. Based on the graphical judgement in Figure 9 and the quantitative measurements in Table 8, both results demonstrate the advantages of the two adaptive learning models over the rest of the methods by a large margin. Moreover, a significance test was performed, as shown in Table 9, by comparing their prediction probabilities with eight state-of-the-art methods. All the p-values are less than 0.05, which indicates that the proposed methods have significant differences from the comparative methods.

4. Discussion

In this paper, a multi-layer adaptive learning model architecture is proposed. Instead of simply concatenating all the multi-scale texture features together for classification, the proposed architecture not only integrates multi-scale texture descriptors in an adaptive manner to consider the associated variation among multiple datasets, but also provides an effective solution for information redundancy. The primary novelty of this proposed work was in the weighted grouping of the texture patterns and assigning greater contributions to those higher weighted groups, instead of using all features entered into the classifier at the same time. Two schemes, i.e., traditional machine learning-based and CNN-based, were designed to demonstrate this idea. The proposed design contained two stages. In the first stage, GLCM was divided into three groups by their individual scales. A baseline was selected, with the remaining groups ordered by their individual performance. In the second stage, the three group were integrated into one enhanced descriptor in a hierarchical architecture by a multi-layer learning scheme. On each layer, a forward stepwise feature selection method was introduced to selectively add some patterns or variables from complemental subgroups into the baseline to produce better performances. The greedy procedure guarantees a monotonically increasing AUC score from the initial descriptor groups at the first layer and reduces redundant information. Due to the variation among multiple datasets or multiscale descriptors, the proposed adaptive learning model increased the AUC score from 0.886 to 0.925 via MGHM and from 0.895 to 0.909 via MG-CNN.
When comparing against the deep learning state-of-the-art methods, the following observations were noted. The VGG16 and AlexNet models performed quite poorly, with AUC values of 0.823 and 0.779, respectively. These results were expected because deep learning methods tend to have much higher data requirements to fully train the high-level features from that methodology, and the dataset used for these experiments is relatively small. However, the proposed MG-CNN model still attained a significantly higher AUC value of 0.909. This showed that the GLCM input for the model already provided some higher-level texture information, so that the deep learning architecture did not have the same steep data requirements as the other methods. On a much larger dataset, it is expected that the VGG16 and AlexNet models will provide closer comparisons to the proposed models. Against the GLCM-CNN method, which was originally used on the same dataset as these experiments [40], the value of the proposed weighted grouping was demonstrated by the higher AUC value. Since the GLCM-CNN model similarly outperformed the VGG16 and AlexNet models, this further reinforced the value of the GLCM as inputs.
When comparing against the other state-of-the-art methods using traditional features and classifiers, the proposed MGHM still outperforms them significantly. In this category, the post-KLT eHMs obtained the best classification performance of the comparative methods likely because the KL transform provides a measure of reducing redundancy of the texture features through the change of basis representation. Against the other traditional feature selection methods, the value of the proposed model in further reducing variation and redundancy to achieve greater classification is even more significantly demonstrated by AUC values.
Although the presented adaptive learning model is implemented for integration of multiscale texture features, the integration strategy can be applied to fuse multimodal datasets, such as the polyp intensity images, the first derivative gradient image and the second order curvature images that were investigated in Song et al. [6] and Hu et al. [11]. While this work investigated spatial variations through the GLCM, this method may help expand upon those other models that integrated multiple feature sets. Future studies will look to expand on the multi-scale texture descriptors to include other types of descriptors and patterns into a study with a larger dataset.

Author Contributions

Conceptualization, Z.L.; Data curation, A.F.A. and P.J.P.; Formal analysis, W.C. and M.J.P.; Funding acquisition, Z.L.; Investigation, W.C. and Z.L.; Methodology, W.C., M.J.P., S.Z., J.T. and Y.G.; Project administration, Z.L.; Resources, Z.L.; Software, W.C., M.J.P., S.Z. and Y.G.; Supervision, Z.L.; Validation, W.C., M.J.P., S.Z., J.T. and Y.G.; Visualization, Z.L.; Writing—original draft, W.C.; Writing—review & editing, M.J.P., J.T. and Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the NIH/NCI grants #CA206171 and #CA220004.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board. The most recent approval date is 18 May 2021. The study was assigned an ID number 93995_MODCR005 and has the title of “Integrating virtual and optical colonoscopies with pathological analysis to map the highly heterogeneous features of colorectal polyp biomarkers”.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

All data used for these experiments can be made available from the contact author upon reasonable request.

Acknowledgments

The authors thank Hongbing Lu for her work in preparing the ROIs for the polyp dataset used in these experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cancer Facts & Figures 2018; American Cancer Society: Atlanta, GA, USA, 2018.
  2. Levin, B.; Lieberman, D.A.; McFarland, B.; Andrews, K.S.; Brooks, D.; Bond, J.; Dash, C.; Giardiello, F.M.; Glick, S.; Johnson, D. Screening and surveillance for the early detection of colorectal cancer and adenomatous polyps, 2008: A joint guideline from the American Cancer Society, the US Multi-Society Task Force on Colorectal Cancer, and the American College of Radiology. Gastroenterology 2008, 134, 1570–1595. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Lieberman, D.A.; Rex, D.K.; Winawer, S.J.; Giardiello, F.M.; Johnson, D.A.; Levin, T.R. Guidelines for colonoscopy surveillance after screening and polypectomy: A consensus update by the US Multi-Society Task Force on Colorectal Cancer. Gastroenterology 2012, 143, 844–857. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Liang, Z.; Richards, R.J. Virtual colonoscopy versus optical colonoscopy. Expert Opin. Med. Diagn. 2010, 4, 159–169. [Google Scholar] [CrossRef] [Green Version]
  5. Goyal, H.; Mann, R.; Gandhi, Z.; Perisetti, A.; Ali, A.; Aman Ali, K.; Sharma, N.; Saligram, S.; Tharian, B.; Inamdar, S. Scope of artificial intelligence in screening and diagnosis of colorectal cancer. J. Clin. Med. 2020, 9, 3313. [Google Scholar] [CrossRef]
  6. Song, B.; Zhang, G.; Lu, H.; Wang, H.; Zhu, W.; Pickhardt, P.J.; Liang, Z. Volumetric texture features from higher-order images for diagnosis of colon lesions via CT colonography. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 1021–1031. [Google Scholar] [CrossRef]
  7. Buvat, I.; Orlhac, F.; Soussan, M. Tumor texture analysis in PET: Where do we stand? J. Nucl. Med. 2015, 56, 1642–1644. [Google Scholar] [CrossRef] [Green Version]
  8. Chicklore, S.; Goh, V.; Siddique, M.; Roy, A.; Marsden, P.K.; Cook, G.J. Quantifying tumour heterogeneity in 18 F-FDG PET/CT imaging by texture analysis. Eur. J. Nucl. Med. Mol. Imaging 2013, 40, 133–140. [Google Scholar] [CrossRef]
  9. Cao, W.; Pomeroy, M.J.; Pickhardt, P.J.; Barish, M.A.; Stanly, S., III; Liang, Z. A local geometrical metric-based model for polyp classification. In Proceedings of the Medical Imaging 2019: Computer-Aided Diagnosis, San Diego, CA, USA, 16–21 February 2019; p. 1095014. [Google Scholar]
  10. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, 3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  11. Hu, Y.; Liang, Z.; Song, B.; Han, H.; Pickhardt, P.J.; Zhu, W.; Duan, C.; Zhang, H.; Barish, M.A.; Lascarides, C.E. Texture feature extraction and analysis for polyp differentiation via computed tomography colonography. IEEE Trans. Med. Imaging 2016, 35, 1522–1531. [Google Scholar] [CrossRef] [Green Version]
  12. Li, W.; Coats, M.; Zhang, J.; McKenna, S.J. Discriminating dysplasia: Optical tomographic texture analysis of colorectal polyps. Med. Image Anal. 2015, 26, 57–69. [Google Scholar] [CrossRef] [Green Version]
  13. Pietikäinen, M.; Hadid, A.; Zhao, G.; Ahonen, T. Computer Vision Using Local Binary Patterns; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011; Volume 40. [Google Scholar]
  14. Wang, Y.; Pomeroy, M.; Cao, W.; Gao, Y.; Sun, E.; Stanley, S., III; Bucobo, J.C.; Liang, Z. Polyp classification by Weber’s Law as texture descriptor for clinical colonoscopy. In Proceedings of the Medical Imaging 2019: Computer-Aided Diagnosis, San Diego, CA, USA, 16–21 February 2019; p. 109502V. [Google Scholar]
  15. Onizawa, N.; Katagiri, D.; Matsumiya, K.; Gross, W.J.; Hanyu, T. Gabor filter based on stochastic computation. IEEE Signal Processing Lett. 2015, 22, 1224–1228. [Google Scholar] [CrossRef]
  16. Wimmer, G.; Gadermayr, M.; Kwitt, R.; Häfner, M.; Tamaki, T.; Yoshida, S.; Tanaka, S.; Merhof, D.; Uhl, A. Training of polyp staging systems using mixed imaging modalities. Comput. Biol. Med. 2018, 102, 251–259. [Google Scholar] [CrossRef] [PubMed]
  17. Wimmer, G.; Tamaki, T.; Tischendorf, J.J.; Häfner, M.; Yoshida, S.; Tanaka, S.; Uhl, A. Directional wavelet based features for colonic polyp classification. Med. Image Anal. 2016, 31, 16–36. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, R.; Zheng, Y.; Poon, C.C.; Shen, D.; Lau, J.Y. Polyp detection during colonoscopy using a regression-based convolutional neural network with a tracker. Pattern Recognit. 2018, 83, 209–219. [Google Scholar] [CrossRef] [PubMed]
  19. Chen, J.; Shan, S.; He, C.; Zhao, G.; Pietikäinen, M.; Chen, X.; Gao, W. WLD: A robust local image descriptor. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1705–1720. [Google Scholar] [CrossRef] [PubMed]
  20. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  21. Hira, Z.M.; Gillies, D.F. A review of feature selection and feature extraction methods applied on microarray data. Adv. Bioinform. 2015, 2015, 198363. [Google Scholar] [CrossRef]
  22. Fong, S.; Deb, S.; Yang, X.-S.; Li, J. Metaheuristic swarm search for feature selection in life science classification. IEEE IT Prof. Mag. 2014, 16, 24–29. [Google Scholar] [CrossRef]
  23. Ma, S.; Huang, J. Penalized feature selection and classification in bioinformatics. Brief. Bioinform. 2008, 9, 392–403. [Google Scholar] [CrossRef] [Green Version]
  24. Cekik, R.; Uysal, A.K. A novel filter feature selection method using rough set for short text data. Expert Syst. Appl. 2020, 160, 113691. [Google Scholar] [CrossRef]
  25. Roffo, G.; Melzi, S.; Cristani, M. Infinite feature selection. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4202–4210. [Google Scholar]
  26. Solorio-Fernández, S.; Carrasco-Ochoa, J.A.; Martínez-Trinidad, J.F. A new hybrid filter–wrapper feature selection method for clustering based on ranking. Neurocomputing 2016, 214, 866–880. [Google Scholar] [CrossRef]
  27. Cai, D.; Zhang, C.; He, X. Unsupervised feature selection for multi-cluster data. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 24–28 July 2010; pp. 333–342. [Google Scholar]
  28. Guo, J.; Zhu, W. Dependence guided unsupervised feature selection. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  29. Li, J.; Dong, W.; Meng, D. Grouped gene selection of cancer via adaptive sparse group lasso based on conditional mutual information. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017, 15, 2028–2038. [Google Scholar] [CrossRef] [PubMed]
  30. Roffo, G.; Melzi, S.; Castellani, U.; Vinciarelli, A. Infinite latent feature selection: A probabilistic latent graph-based ranking approach. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1398–1406. [Google Scholar]
  31. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
  32. Zeng, H.; Cheung, Y.-M. Feature selection and kernel learning for local learning-based clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1532–1547. [Google Scholar] [CrossRef] [Green Version]
  33. Liu, H.; Zhou, M.; Liu, Q. An embedded feature selection method for imbalanced data classification. IEEE/CAA J. Autom. Sin. 2019, 6, 703–715. [Google Scholar] [CrossRef]
  34. Cai, J.; Luo, J.; Wang, S.; Yang, S. Feature selection in machine learning: A new perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
  35. Gui, J.; Sun, Z.; Ji, S.; Tao, D.; Tan, T. Feature selection based on structured sparsity: A comprehensive study. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 1490–1507. [Google Scholar] [CrossRef]
  36. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput. 2015, 20, 606–626. [Google Scholar] [CrossRef] [Green Version]
  37. Sharma, M.; Kaur, P. A Comprehensive Analysis of Nature-Inspired Meta-Heuristic Techniques for Feature Selection Problem. Arch. Comput. Methods Eng. 2021, 28, 1103–1127. [Google Scholar] [CrossRef]
  38. Cao, W.; Liang, Z.J.; Pomeroy, M.J.; Ng, K.; Zhang, S.; Gao, Y.; Pickhardt, P.J.; Barish, M.A.; Abbasi, A.F.; Lu, H. Multilayer feature selection method for polyp classification via computed tomographic colonography. J. Med. Imaging 2019, 6, 044503. [Google Scholar] [CrossRef]
  39. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4768–4777. [Google Scholar]
  40. Tan, J.; Gao, Y.; Liang, Z.; Cao, W.; Pomeroy, M.J.; Huo, Y.; Li, L.; Barish, M.A.; Abbasi, A.F.; Pickhardt, P.J. 3D-GLCM CNN: A 3-dimensional gray-level Co-occurrence matrix-based CNN model for polyp classification via CT colonography. IEEE Trans. Med. Imaging 2019, 39, 2013–2024. [Google Scholar] [CrossRef] [PubMed]
  41. Pestov, V. Is the k-NN classifier in high dimensions affected by the curse of dimensionality? Comput. Math. Appl. 2013, 65, 1427–1437. [Google Scholar] [CrossRef]
  42. Nembrini, S.; König, I.R.; Wright, M.N. The revival of the Gini importance? Bioinformatics 2018, 34, 3711–3718. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Pickhardt, P.J. Screening CT colonography: How I do it. Am. J. Roentgenol. 2007, 189, 290–298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Lu, L.; Zhao, J. An improved method of automatic colon segmentation for virtual colon unfolding. Comput. Methods Programs Biomed. 2013, 109, 1–12. [Google Scholar] [CrossRef] [PubMed]
  45. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Illustration of co-occurrence matrix (CM) calculation in 2D/3D images: (a) CM parameters in 2D images; (b) CM parameters in 3D images; and (c) A GLCM example of a 2D case when direction is 0° and displacement = 1. The left is a gray image, and the right one is its GLCM.
Figure 1. Illustration of co-occurrence matrix (CM) calculation in 2D/3D images: (a) CM parameters in 2D images; (b) CM parameters in 3D images; and (c) A GLCM example of a 2D case when direction is 0° and displacement = 1. The left is a gray image, and the right one is its GLCM.
Sensors 22 00907 g001
Figure 2. Visualization of information CNN learnt from each subgroups: (a) G 1 , (b) G 2 and (c) G 3 . The first column is the original GLCM. The corresponding label (0 for benign and 1 for malignant) and model score of the malignancy risk are listed on the top. The remaining two columns are the interpretations of model prediction on the two classes. The red cells show the entries push the model’s decision close to that class, while blue pixels pull the prediction results away.
Figure 2. Visualization of information CNN learnt from each subgroups: (a) G 1 , (b) G 2 and (c) G 3 . The first column is the original GLCM. The corresponding label (0 for benign and 1 for malignant) and model score of the malignancy risk are listed on the top. The remaining two columns are the interpretations of model prediction on the two classes. The red cells show the entries push the model’s decision close to that class, while blue pixels pull the prediction results away.
Sensors 22 00907 g002
Figure 3. The flowchart of multi-level learning model for fusion of multi-scale feature sets.
Figure 3. The flowchart of multi-level learning model for fusion of multi-scale feature sets.
Sensors 22 00907 g003
Figure 4. The flowchart of the feature selection step for the baseline and the complement in the multi-layer learning model.
Figure 4. The flowchart of the feature selection step for the baseline and the complement in the multi-layer learning model.
Sensors 22 00907 g004
Figure 5. Network structure of FSFS-CNN.
Figure 5. Network structure of FSFS-CNN.
Sensors 22 00907 g005
Figure 6. Flowchart of data acquisition and preparation for these experiments.
Figure 6. Flowchart of data acquisition and preparation for these experiments.
Sensors 22 00907 g006
Figure 7. Three sample CT slices from select polyp masses. Green contour around the polyp show the segmentation. Air voxels from the lumen below −450 HU are removed post-segmentation and are highlighted red in the images. Images show sample polyps with pathologies (a) adenocarcinoma, (b) villous adenoma, and (c) villous adenoma.
Figure 7. Three sample CT slices from select polyp masses. Green contour around the polyp show the segmentation. Air voxels from the lumen below −450 HU are removed post-segmentation and are highlighted red in the images. Images show sample polyps with pathologies (a) adenocarcinoma, (b) villous adenoma, and (c) villous adenoma.
Sensors 22 00907 g007
Figure 8. The trends of three AUC score curves of polyp classification, their maximums and their partitions over 63 polyps via forward step feature selection method: (a) D1, (b) D2, and (c) D3.
Figure 8. The trends of three AUC score curves of polyp classification, their maximums and their partitions over 63 polyps via forward step feature selection method: (a) D1, (b) D2, and (c) D3.
Sensors 22 00907 g008
Figure 9. ROC curves of proposed and comparative methods.
Figure 9. ROC curves of proposed and comparative methods.
Sensors 22 00907 g009
Table 1. Digital direction subdivision by their voxel distances from one voxel to the concerned center voxel.
Table 1. Digital direction subdivision by their voxel distances from one voxel to the concerned center voxel.
Radius = 1 Radius = 2 Radius = 3
Direction group ID G 1 G 2 G 3
Number of GLCM Directions364
Descriptor group ID D 1 D 2 D 3
Number of variables84168112
Table 2. Detailed network design for MG-CNN.
Table 2. Detailed network design for MG-CNN.
StructureTypeKernel Size# of Kernels/
Channels/Neurons/Strides
Activation
Layer 12D
Convolution
3 × 364 (stride 1)ReLU
Layer 2Batch
Normalization
64
Layer 3Maxpool2 × 2(stride 2)
Layer 42D
Convolution
3 × 364 (stride 1)ReLU
Layer 5Batch
Normalization
64
Layer 6Maxpool2 × 2(stride 2)
Layer 7Dense 1000ReLU
Layer 8Dense 1000ReLU
Layer 10Dense 2softmax
Table 3. Patient demographics of polyp mass dataset.
Table 3. Patient demographics of polyp mass dataset.
PathologyCountClassMale:
Female
Average Age (yrs)Average Size (mm)
Tubular Adenoma202:069.835.0
Serrated Adenoma302:155.234.3
Tubulovillous
Adenoma
21011:1064.436.9
Villous Adenoma504:167.455
Adenocarcinoma32112:2069.943.9
Table 4. The preliminary classification results of the two proposed models.
Table 4. The preliminary classification results of the two proposed models.
Group IDGLCM DirectionsMGHM AUC
(Mean ± STD)
MG-CNN AUC (Mean ± STD)
G 1 30.846 ± 0.0980.895 ± 0.064
G 2 60.875 ± 0.1010.889 ± 0.061
G 3 40.892 ± 0.0980.871 ± 0.074
Table 5. Two parts of each descriptor divided by forward step feature selection method via SVM classifier.
Table 5. Two parts of each descriptor divided by forward step feature selection method via SVM classifier.
Descriptor-ID D 1 D 2 D 3
AUC Score0.8540.8750.892
Sub-ID D 1 b D 1 c D 2 b D 2 c D 3 b D 3 c
Number of
Variables
651931656106
Table 6. All results of the MGHM over the polyp dataset. Descriptor pool represents the current candidates and its sequence in each layer. B a s e l i n e 2 and B a s e l i n e 3 are two new descriptors generated by the baselines and the complements of their previous layers.
Table 6. All results of the MGHM over the polyp dataset. Descriptor pool represents the current candidates and its sequence in each layer. B a s e l i n e 2 and B a s e l i n e 3 are two new descriptors generated by the baselines and the complements of their previous layers.
LayerBaselineCandidateDescriptor Pool (DP)
SourceVariablesAUC (Mean ± STD)SourceSelected
Variables
L 1 B a s e l i n e 1   ( D 3 b ) 60.892 ± 0.098 D 2 4 D 2 ,   D 1 ,   D 3 c
L 2 B a s e l i n e 2 100.916 ± 0.038 D 1 3 D 1 ,   D 3 c
L 3 B a s e l i n e 3 130.919 ± 0.036 D 3 c 4 D 3 c
L 4 Final
Descriptor
170.925 ± 0.035 -
Table 7. The results of MG-CNN over the polyp dataset. Descriptor pool represents the current candidates and its sequence in each layer. B a s e l i n e 2 is a new descriptor generated by the baselines and the complements of their previous layers.
Table 7. The results of MG-CNN over the polyp dataset. Descriptor pool represents the current candidates and its sequence in each layer. B a s e l i n e 2 is a new descriptor generated by the baselines and the complements of their previous layers.
LayerBaselineCandidateDescriptor Pool (DP)
SourceGLCMsAUC (Mean ± STD)SourceGLCMs
L 1 B a s e l i n e 1   ( G 1 ) 30.895 ± 0.064 G 2 4 G 2 ,   G 3
L 2 B a s e l i n e 2 90.904 ± 0.047 G 3 3 G 3  
L 3 Final
Results
130.909 ± 0.051
Table 8. Four evaluation measurements of proposed and comparative methods.
Table 8. Four evaluation measurements of proposed and comparative methods.
MethodAUC AccuracySensitivitySpecificity
eHM0.8860.7970.8680.726
eHM+KLT0.9070.8140.7810.848
AlexNet (image)0.7790.7780.8310.726
VGG16 (image)0.8230.7410.7140.769
LASSO0.8360.7480.7910.706
SVM-RFE0.8560.7830.7750.791
DGUFS0.8660.8060.8360.776
GLCM CNN0.9000.8560.8430.868
MG-CNN0.9090.8640.8660.862
MGHM0.9250.8840.8910.878
Table 9. p-values from statistical significance analysis over the ten methods using Wilcoxon Signed-rank Test between the predicted probabilities of these methods.
Table 9. p-values from statistical significance analysis over the ten methods using Wilcoxon Signed-rank Test between the predicted probabilities of these methods.
MethodeHMeHM+KLT AlexNetVGG16LASSOSVM-RFEDGUFSGLCM CNN
MGHM<<0.050.0179<<0.05<<0.05<<0.05<<0.05<<0.050.0204
MG-CNN<<0.050.0411<<0.05<<0.05<<0.05<<0.05<<0.050.0398
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cao, W.; Pomeroy, M.J.; Zhang, S.; Tan, J.; Liang, Z.; Gao, Y.; Abbasi, A.F.; Pickhardt, P.J. An Adaptive Learning Model for Multiscale Texture Features in Polyp Classification via Computed Tomographic Colonography. Sensors 2022, 22, 907. https://doi.org/10.3390/s22030907

AMA Style

Cao W, Pomeroy MJ, Zhang S, Tan J, Liang Z, Gao Y, Abbasi AF, Pickhardt PJ. An Adaptive Learning Model for Multiscale Texture Features in Polyp Classification via Computed Tomographic Colonography. Sensors. 2022; 22(3):907. https://doi.org/10.3390/s22030907

Chicago/Turabian Style

Cao, Weiguo, Marc J. Pomeroy, Shu Zhang, Jiaxing Tan, Zhengrong Liang, Yongfeng Gao, Almas F. Abbasi, and Perry J. Pickhardt. 2022. "An Adaptive Learning Model for Multiscale Texture Features in Polyp Classification via Computed Tomographic Colonography" Sensors 22, no. 3: 907. https://doi.org/10.3390/s22030907

APA Style

Cao, W., Pomeroy, M. J., Zhang, S., Tan, J., Liang, Z., Gao, Y., Abbasi, A. F., & Pickhardt, P. J. (2022). An Adaptive Learning Model for Multiscale Texture Features in Polyp Classification via Computed Tomographic Colonography. Sensors, 22(3), 907. https://doi.org/10.3390/s22030907

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop