[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
The Power of Words from the 2024 United States Presidential Debates: A Natural Language Processing Approach
Next Article in Special Issue
The Use of Voice Control in 3D Medical Data Visualization Implementation, Legal, and Ethical Issues
Previous Article in Journal
Correction: Alawadh et al. Semantic Features-Based Discourse Analysis Using Deceptive and Real Text Reviews. Information 2023, 14, 34
Previous Article in Special Issue
Leveraging Social Media Data for Enhanced Forecasting of International Student Arrivals in Australia
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Approach for Detecting Parkinson’s Disease by Integrating Optimal Feature Selection Strategies with Dense Multiscale Sample Entropy

by
Minh Tai Pham Nguyen
1,
Minh Khue Phan Tran
2,
Tadashi Nakano
3,
Thi Hong Tran
3 and
Quoc Duy Nam Nguyen
3,*
1
Faculty of Advanced Program, Ho Chi Minh City Open University, Ho Chi Minh City 700000, Vietnam
2
Faculty of Information Technology, Ho Chi Minh City Open University, Ho Chi Minh City 700000, Vietnam
3
Department of Core Informatics, Graduate School of Informatics, Osaka Metropolitan University, Osaka 558-8585, Japan
*
Author to whom correspondence should be addressed.
Information 2025, 16(1), 1; https://doi.org/10.3390/info16010001
Submission received: 5 November 2024 / Revised: 9 December 2024 / Accepted: 17 December 2024 / Published: 24 December 2024
(This article belongs to the Special Issue Feature Papers in Artificial Intelligence 2024)
Graphical abstract
">
Figure 1
<p>(<b>a</b>) Gender and label distribution across each class in severity classification task; (<b>b</b>) gender and label distribution across each class in PD classification task.</p> ">
Figure 2
<p>The methodology procedure has four stages: preprocessing, feature extraction, feature selection, and classification.</p> ">
Figure 3
<p>The process of signal division using the time-slicing window method and outlier removal through the quartile approach and histogram analysis.</p> ">
Figure 4
<p>Boxplot analysis of model consistency for PD and severity classification tasks.</p> ">
Figure 5
<p>(<b>a</b>) Adjusted <span class="html-italic">p</span>-value matrix of paired T-tests comparing accuracies between classifier–feature selection method pairs (M-SamEn); (<b>b</b>) adjusted <span class="html-italic">p</span>-value matrix of paired <span class="html-italic">t</span>-tests comparing accuracies between classifier–feature selection method pairs (DM-SamEn).</p> ">
Figure 6
<p>(<b>a</b>) Correlation matrix of the feature set extracted using the M-SamEn method; (<b>b</b>) correlation matrix of the feature set extracted using the DM-SamEn method.</p> ">
Figure 7
<p>(<b>a</b>) Distribution of selected features by signal source (original and computed signals from Equations (<a href="#FD1-information-16-00001" class="html-disp-formula">1</a>)–(<a href="#FD3-information-16-00001" class="html-disp-formula">3</a>) after the feature selection stage; (<b>b</b>) distribution of selected features by feature extraction method after the feature selection stage.</p> ">
Figure 8
<p>Comparison of feature count across feature selection methods (*: <span class="html-italic">p</span>-value &lt; 0.05; **: <span class="html-italic">p</span>-value &lt; 0.01).</p> ">
Versions Notes

Abstract

:
Parkinson’s disease (PD) is a neurological disorder that severely affects motor function, especially gait, requiring accurate diagnosis and assessment instruments. This study presents Dense Multiscale Sample Entropy (DM-SamEn) as an innovative method for diminishing feature dimensions while maintaining the uniqueness of signal features. DM-SamEn employs a weighting mechanism that considers the dynamic properties of the signal, thereby reducing redundancy and improving the distinctiveness of features extracted from vertical ground reaction force (VGRF) signals in patients with Parkinson’s disease. Subsequent to the extraction process, correlation-based feature selection (CFS) and sequential backward selection (SBS) refine feature sets, improving algorithmic accuracy. To validate the feature extraction and selection stage, three classifiers—Adaptive Weighted K-Nearest Neighbors (AW-KNN), Radial Basis Function Support Vector Machine (RBF-SVM), and Multilayer Perceptron (MLP)—were employed to evaluate classification efficacy and ascertain optimal performance across selection strategies, including CFS, SBS, and the hybrid SBS-CFS approach. K-fold cross-validation was employed to provide improved evaluation of model performance by assessing the model on various data subsets, thereby mitigating the risk of overfitting and augmenting the robustness of the results. As a result, the model demonstrated a significant ability to differentiate between PD patients and healthy controls, with classification accuracy reported as ACC [CI 95%: 97.82–98.5%] for disease identification and ACC [CI 95%: 96.3–97.3%] for severity assessment. Optimal performance was primarily achieved through feature sets chosen using SBS and the integrated SBS-CFS methods. The findings highlight the model’s potential as an effective instrument for diagnosing PD and assessing its severity, contributing to advancements in clinical management of the condition.

Graphical Abstract">
Graphical Abstract

1. Introduction

Neurodegenerative diseases (NDDs) include a variety of diseases characterized by a gradual decline in nervous system structure and function. Parkinson’s disease (PD), a significant movement disorder, affects primarily older people, with an estimated 4 to 6.5 million affected individuals worldwide, including approximately 1 million instances in the United States, accounting for almost 1% of those over 60 [1]. Annually, the United States documents 60,000 new diagnoses of Parkinson’s disease. PD is diagnosed with tremors, rigidity, poor balance, and slow movement, which can cause major problems with walking, a higher risk of falling, and less independence. Although most cases seem sporadic, roughly 15% of PD occurrences exhibit familial patterns, with around 10% associated with certain genetic variations [2].
Gait analysis is important for finding biomechanical problems related to Parkinson’s disease (PD). Furthermore, metrics such as vertical ground reaction force (VGRF) signals, stride lengths, and swing timings can be beneficial in early diagnosis. Such evaluations are essential for the implementation of prompt therapies to potentially slow the progression of PD. Interpreting these measurements typically requires specialized knowledge, which presents difficulties for less experienced physicians. This problem indicates the urgency of algorithms that directly employ raw VGRF data to improve clinical evaluations, standardize diagnostic procedures, and increase accessibility for physicians and engineers [3].
Early research has focused on developing effective methods for the diagnosis and assessment of the severity of Parkinson’s disease. For instance, Aşurolu et al. attained an accuracy of 99.5%, a sensitivity of 98.7%, and a specificity of 99.5% by applying a CNN combined with a locally weighted random forest [4]. Another study by Zhao et al. and El Maachi et al. used CNN-based models to tell the difference between people with Parkinson’s disease and healthy people, and both studies obtained precision rates of more than 90% [5,6]. In addition, in another work [7], Fast Fourier Transform (FFT) was used in conjunction with an Artificial Neural Network (ANN) to classify the severity, resulting in an accuracy of 97%. Another approach proposes a deep learning architecture, “NDDNet”, aimed at resolving this issue, achieving an average accuracy of 96.75% in identifying three categories of neurodegenerative diseases [8].
The Dense Multiscale Sample Entropy (DM-SamEn) method builds upon the conventional Multiscale Sample Entropy (M-SamEn) technique by mitigating its fundamental limitations, especially the issues of feature redundancy and collinearity across scales. In M-SamEn, the entropy values are computed at various scales, frequently demonstrating a high correlation in this study, potentially resulting in less biased and redundant feature representations. Additionally, DM-SamEn addresses this issue by implementing a weighted aggregation mechanism that employs the exponential function of signal values as scale-specific weights. This guarantees that scales with enhanced diagnostic relevance have a more substantial impact on the cumulative entropy value, leading to a more unique feature set. By emphasizing scales with significant variations, DM-SamEn minimizes feature redundancy, enhances interpretability, and strengthens the robustness of the extracted features.
In response to the above problems, this study proposes a three-point strategy:
-
Utilize the DM-SamEn method to reduce feature dimensionality while ensuring feature uniqueness and robustness.
-
Utilize two feature selection methods, sequential backward selection (SBS) and correlation-based feature selection (CFS), to identify the most significant features from the initial set. A hybrid SBS-CFS approach is also employed.
-
Validate the feature extraction (DM-SamEn) and selection (CFS and SBS) stages with three classifiers: adaptive weighted K-nearest neighbors (AW-KNN), radial basis function support vector machine (RBF-SVM), and multilayer perceptron (MLP). These classifiers evaluate this classification performance, which also determines the optimal results in the various feature selection strategies.

2. Material

This study utilized an extensive dataset from the PhysioNet platform [9], mostly derived from three significant investigations: those by Yogev et al. (Ga), Hausdorff et al. (Ju), and Frenkel-Toledo et al. (Si) [10,11,12,13]. Table 1 encapsulates the dataset, comprising 208 records of vertical ground reaction force (VGRF) signals that illustrate gait patterns in two distinct groups. The initial group consists of 93 individuals diagnosed with idiopathic Parkinson’s disease (PD) (mean age: 66.3 years; 63% male), whereas the control group comprises 73 healthy individuals with comparable demographics (mean age: 63.65 years; 55% male). Participants walked at a self-determined velocity on a flat surface for roughly two min, while vertical ground reaction force (VGRF) data were recorded using eight sensors positioned beneath each foot insole, measuring force in Newtons over time. All 16 sensor outputs were digitized at a frequency of 100 Hz, along with supplementary signals representing the combined outputs of the eight sensors per foot.
The study’s database comprises demographic information, serving two primary purposes: differentiating the Control (Co) group from the Parkinson’s Disease (PD) group while rating PD severity via the Hoehn and Yahr (HY) scale (notations 2; 2.5; 3) [14]. A gender propensity was noted, with males exhibiting a higher likelihood of developing Parkinson’s disease (59% for classification tests and 62% for severity tasks). Age influences the beginning of Parkinson’s disease, occurring between 63.65 and 70.8 years, while gait speed diminishes with severity, ranging from 0.79 to 1.03 m/s. A minor imbalance was present in the classification data (Co 44%, PD 56%), whereas a significant imbalance was observed in severity, with “2” as the predominant class at 59%, followed by “2.5” at 30%, and “3” as the least prevalent at 11% (Figure 1).

3. Methodology

Figure 2 depicts a four-step approach that includes preprocessing, feature extraction, feature selection, and classification stages. The preprocessing phase comprises time slicing, outlier elimination, and signal manipulation to enhance signal quality and augment dataset size while reducing noise and artifacts that may hinder classification. We utilize both conventional features and dense multiscale sample entropy (DM-SamEn) features in the feature extraction process to efficiently minimize feature dimensionality. The feature selection step encompasses correlation-based feature selection (CFS), sequential backward selection (SBS), and the hybrid SBS-CFS strategy to retain the most essential features. Adaptive weighted K-Nearest Neighbors (AW-KNN), radial basis function support vector machine (RBF-SVM), and multilayer perceptron (MLP) classifiers assess classification efficacy in the concluding classification phase.

3.1. Preprocessing Stage

“Mean foot” signals (MS) were introduced by aggregating left (LS) and right (RS) foot signals of raw vertical ground reaction force (VGRF) signals for feature extraction. These signals are represented as (MS, LS, RS) for each sample, denoted as X. Initially captured at 100 Hz for almost two minutes, these X signals were divided into non-overlapping 10-s segments to augment sample diversity for training. However, this segmentation produced certain anomalous subsets; hence, we employed the outlier signal detection approach in Theorem 1 to exclude outliers (Figure 3), thereby refining the dataset for further phases.
Theorem 1. 
Outlier Detection Based on Second-order Differences
Let { S i } i = 1 n represent a set of n signals, where each signal ( S i ) is a sequence of real-valued observations ( S i = { s i 1 , s i 2 , , s i m } ) of length m. The second-order difference of a signal ( S i ) is as follows:
Δ 2 S i [ j ] = s i , j + 2 2 s i , j + 1 + s i j , j = 1 , , m 2 .
Then, the **variability score** of the signal ( S i ) is computed as follows:
Score ( S i ) = j = 1 m 2 Δ 2 S i [ j ] .
Let { x i } i = 1 n be the variability score values of all signals, where x i = Score ( S i ) . The distribution of { x i } i = 1 n defines the lower quartile ( Q 1 ), the upper quartile ( Q 3 ), and the interquartile range ( IQR = Q 3 Q 1 ).
A signal ( S i ) is classified as an outlier if
x i < Q 1 k · IQR or x i > Q 3 + k · IQR ,
where k > 0 is a scaling factor (typically k = 1.5 ).
Following preprocessing, the dataset was expanded, with X represented as subsets (sX). Due to the limited number of subjects in this study, each sX signal was regarded as an independent sample. Furthermore, noise was incorporated into the sX signals (Theorem 2) to reduce the risk of data leakage, which might reduce the reliability of the trained models and the overall validity of the study results.
Theorem 2. 
Frequency-Specific Noise Addition for Data Leakage Prevention
Let s = { s 1 , s 2 , , s n } be a discrete signal of length n.
Let S be the discrete Fourier transform (DFT) of signal s :
S = FFT ( s ) ,
To prevent data leakage by introducing controlled perturbations with a noise-level parameter of α [ 0 , 1 ] ,
noise = α · ( max ( s ) min ( s ) ) · R ,
where R = { R 1 , R 2 , , R n } is a vector of independent random variables drawn from a standard normal distribution, i.e., R i N ( 0 , 1 ) for i = 1 , , n .
The noisy frequency-domain representation is then given by
S noisy = S + noise .
Finally, the noisy signal ( s noisy ) is obtained by applying the inverse Fourier transform to S noisy :
s noisy = IFFT ( S noisy ) ,
This procedure introduces noise in the frequency domain to obscure patterns that could provoke data leakage, while maintaining essential signal properties within the defined noise level (α).
Subsequently, the sX signals were transformed by utilizing the first derivative, second derivative, and cumulative sum transformations, as delineated in Equations (1)–(3), to extract more profound information from the original signals. Each sX signal consists of four derived signals, yielding a total of 12 signals per subset. This approach is influenced by Nam et al.’s approach [15] that augments the feature set, equipping classifiers with greater data depth for higher performance during the training, validation, and testing stages.
x ( t ) = d x ( t ) d t
x ( t ) = d 2 x d t 2
y [ n ] = k = 0 n x [ k ]

3.2. Feature Extraction Stage

3.2.1. Conventional Features

Before feature extraction, signals are standardized by rescaling their range to the [0,1] interval, as seen in Equation (4). Subsequent to this normalization process, conventional features, including mean and standard deviation, are determined for each signal, as illustrated in Equations (5) and (6).
x norm = x min ( x ) max ( x ) min ( x )
μ = 1 N i = 1 N x [ i ]
σ = 1 N i = 1 N ( x [ i ] μ ) 2

3.2.2. Dense Multiscale Sample Entropy (DM-SamEn) Features

Dense Multiscale sample entropy (DM-SamEn) comprises three primary terms: sample entropy, multiscale analysis, and dense. This approach predominantly utilizes sample entropy to assess the predictability and complexity of time-series data at several scales, finally aggregating these entropy values into an overall representative value for the input series.
Sample entropy (SamEn) is a statistical instrument utilized for evaluating the complexity or irregularity of time-series data [16]. SamEn mitigates the influences of sequence length and sample size biases, establishing a dependable method for comparing entropy across datasets of varying lengths. Elevated entropy values indicate a more complex dataset, whilst diminished values signify a simpler dataset. This method is very effective for analyzing physiological data to identify situations via fluctuations in complexity [17]. The SamEn calculations are contingent upon the embedding dimension (m), tolerance (r), and signal length (N), with ideal values for m and r established at 3 and 0.15, respectively, in this work. Theorem 3 graphically demonstrates the procedure of the sample entropy method.
Theorem 3. 
Sample Entropy (SamEn)
For a time series (U = { u 1 , u 2 , u 3 , . . . , u N } ),
1. 
Generate X:
X = { x 1 , x 2 , x 3 , . . . , x N m + 1 } , x i = { u i , x i + 1 , . . . , x i + m 1 } , i N m + 1 ;
2. 
Construct C n m :
C n m = j = 1 N m + 1 C i j m N m + 1 ,   i N m + 1
with
C i j m = 1 if d | x i , x j | r 0 if d | x i , x j | > r ,   d | x i , x j | = m a x | x i x j | ;
3. 
Define ϕ m ( r , N ) and ϕ m + 1 ( r , N ) :
Φ m ( r , N ) = 1 N m + 1 i = 1 N m + 1 ( C i m ) ,   Φ m + 1 ( r , N ) = 1 N m i = 1 N m ( C i m + 1 ) ;
4. 
Compute SamEn:
S a m E n ( m , r , N ) = ln ( Φ m + 1 ( r , N ) Φ m ( r , N ) )
Multiscale sample entropy (M-SamEn) extends SamEn by analyzing the complexity of time series over multiple scales [18]. This provides a comprehensive understanding of signal complexity across an extensive array of time periods. M-SamEn provides a coarser analysis by incrementally downsampling the original time series to provide multiple temporal scales, in contrast to SamEn, which assesses complexity at a singular scale. SamEn is computed at each scale, providing insight into the signal’s structural dynamics across several time intervals. M-SamEn can identify nuanced, scale-dependent fluctuations in signal complexity by measuring entropy levels across various scales [19]. In this investigation, M-SamEn was utilized with scales ranging from 1 to 6, with Equation (7) elucidating the downsampling procedure for the original time series.
y j s = 1 s i = k j j s x i , k j = ( j 1 ) s + 1 ,   1 j N s
In this study, the term “dense” denotes an aggregated value that encapsulates all M-SamEn scale values [20]. Aan analysis of M-SamEn values obtained from the identical time-series data demonstrated robust correlation coefficients (>0.7), suggesting a possible collinearity issue that results in model instability and overfitting by intensifying the effect of linked predictors, thereby hiding the distinct effects of individual variables [21]. High collinearity complicates the interpretation of model coefficients, weakens prediction reliability, and skews the individual contributions of associated variables. Hence, to resolve these challenges and improve classifier stability, DM-SamEn is introduced as an aggregated form for M-SamEn values (Equation (8)). In addition, the DM-SamEn value can serve as an indicator of the overall predictability and complexity of the input time series across various temporal scales.
D M S a m E n = N i = 1 exp x i S a m E n i N i = 1 exp x i

3.3. Feature Selection Stage

Post-feature extraction, high correlations (>0.7) among features still remained in the feature set. In order to resolve this issue, we propose three strategies for feature selection: (1) Correlation-Based Feature Selection (CFS), (2) Sequential Backward Selection (SBS), and (3) a hybrid approach incorporating SBS and CFS.

3.3.1. Correlation-Based Feature Selection (CFS)

The correlation-based feature selection (CFS) method, developed by K. Michalak et al. in 2010, aims to improve the effectiveness of feature selection while maintaining classification accuracy [22]. CFS is based on the principle that an optimal feature subset should consist of features that are strongly correlated with the class label while demonstrating low correlations among themselves, minimizing redundancy and enhancing model interpretability [23]. By incorporating a threshold t, CFS enhances the feature selection process to ensure that the chosen features offer distinct information while avoiding collinearity. This threshold-based method facilitates the management of model complexity and enhances stability by prioritizing non-redundant features that have significant predictive value. This study implemented a threshold t of 0.7, as outlined in Theorem 4.
CFS was implemented as the primary filter in strategy 1 to identify key features from the initial set and as a secondary filter following sequential backward selection (SBS) in strategy 3.
Theorem 4. 
Correlation-based Feature Selection (CFS)
  • Input: X-Input Features Spaces ( m × n ), t-Threshold
  • Output: Selected-Feature-List of indices
  • Procedure:
1 
Initialize Selected-Feature as an empty list.
2 
For each pair of features ( f i , f j ) in X (where i < j )
a 
Compute the absolute correlation coefficient between f i and f j )
b 
If the absolute correlation coefficient is less than or equal to the threshold t
Add f i and f j to Selected-Feature
3 
Remove any duplicate entries from Selected-Feature (keeping only unique feature indices)
  • Return Selected-Feature

3.3.2. Sequential Backward Selection (SBS)

Sequential backward selection (SBS) is an optimization technique designed to enhance the feature space and boost the performance of machine learning models via dimensionality reduction. SBS operates by systematically removing the least significant features from the entire feature set, with the objective of maintaining or improving classification accuracy [24]. The method begins with the full set of features and, at each iteration, eliminates the feature whose removal results in the smallest decrease (or greatest increase) in model performance. This method aims to minimize redundancy and computational cost by identifying a subset of features that maintains the predictive capability of the original set. SBS is particularly advantageous for handling high-dimensional data, as reducing the number of features helps mitigate overfitting and improve model stability [25]. The SBS algorithm is outlined in Theorem 5 as a pseudocode.
In strategies 2 and 3, the SBS serves as the primary filter for identifying significant features. In strategy 3, SBS operates through a sequential feature elimination process that continues as long as model performance is stable. Conversely, CFS relies on a correlation threshold (t), which may unintentionally remove elements essential for model effectiveness. Therefore, SBS operates as the primary filter, whereas CFS acts as the secondary filter.
Theorem 5. 
Sequential Backward Selection (SBS)
  • Input:
  • X-Input Features Spaces ( m × n ), y-vector ( n × 1 )
  • model-classifier, k-number of features to select (stopping criterion)
  • Output:Selected-Feature-List of indices
  • Procedure:
1 
Initialize Selected-Feature as an empty list.
2 
While the number of Selected-Feature > k:
a 
Set min-performance = ∞ and candidate-feature = None
b 
For each f in Selected-Feature:
i 
Temporarily remove f from Selected-Feature
ii 
Train the model using X[:, Selected-Feature] and y
iii 
Evaluate the model performance
iv 
If model-performance < min-performance:
set min-performance = model-performance
add f in candidate-feature
c 
Remove candidate-feature from Selected-Feature
  • Return Selected-Feature

3.4. Classification Stage

This study employed three classification models to train the feature sets identified in the preceding step: radial basis function support vector machine (RBF-SVM), adaptive weighted K-nearest neighbors (AW-KNN), and multilayer perceptron (MLP). The selection of these models was guided by three primary objectives. The primary aim is to identify the model that exhibits optimal performance when trained with the selected features. The second aim is to evaluate the alignment of the observed training results with the initial research predictions. Finally, to illustrate the effectiveness of the feature extraction and selection processes, it is reasonable to utilize simpler classifiers or models with moderate training capabilities.

3.4.1. Adaptive Weighted K-Nearest Neighbors (AW-KNN)

Adaptive weighted k-nearest neighbors (AW-KNN) improves upon the primary K-nearest neighbors (KNN) algorithm. This approach strengthens classification accuracy by assigning weights to each neighbor according to their distance from the query point. The weighted method employed by AW-KNN facilitates a ranking of nearer neighbors, reducing the impact of more distant points that could introduce noise or ambiguity [26]. AW-KNN differs from standard KNN by dynamically adjusting weights for neighbors, enhancing its effectiveness in datasets characterized by overlapping classes or non-uniform distributions. The algorithm computes the distance from the query point (x) to each neighboring point ( x i ) among the chosen k [27]. The algorithm allocates increased influence to nearer neighbors, with each neighbor’s weight inversely correlated to its distance from the query point. This study utilized k = 10 for Manhattan distance (Equation (9)) and squared inverse weighting (Equation (10)) for the calculations of distance and weight, respectively.
d ( x , x i ) = j = 1 n | x j x i , j |
w i = 1 d ( x , x i ) 2 + ϵ

3.4.2. Radial Basis Function Support Vector Machine (RBF-SVM)

The radial basis function support vector machine (RBF-SVM) extends support vector machine (SVM) by utilizing the radial basis function (RBF) kernel (Equation (11)), commonly referred to as the Gaussian kernel, to manage non-linear data through transformation into a higher-dimensional feature space [28]. The RBF kernel is well-suited for datasets exhibiting non-linear relationships, as it transforms the input space into a feature space that facilitates linear separation [29]. RBF-SVM effectively classifies complex patterns by identifying a hyperplane that optimally separates classes within the transformed space (Equation (12)).
RBF-SVM is fundamentally designed for binary classification; however, for the severity problem in this study, which encompasses three classes, the “one-vs-all” (OvA) approach is implemented [30]. The OvA method is employed to adapt binary classifiers for multi-class classification by training an individual classifier for each class, framing it as a binary problem of differentiating that class from the others. The OvA approach generates k separate binary classifiers for a dataset including k classes. Each classifier is trained with one class designated as the positive class ( + 1 ) and all other classes as the negative class ( 1 ), resulting in a distinct decision boundary for differentiating that one class from the others. This approach is direct and successful, particularly when combined with a strong kernel such as RBF, which produces optimal decision boundaries for intricate datasets.
K ( x , x ) = exp γ x x 2
min w , b , ξ 1 2 w 2 + C i = 1 N ξ i

3.4.3. Multilayer Perceptron (MLP)

A Multilayer Perceptron (MLP) is a type of artificial neural network (ANN) including several sequential layers of neurons, which include an input layer, one or more hidden layers, and an output layer, with each layer being fully interconnected to the subsequent layer. Multilayer perceptrons (MLPs) are extensively employed for supervised learning tasks, including classification and regression, due to their ability to represent intricate, non-linear relationships of data [31]. The use of non-linear activation functions in the hidden layers allows MLPs to proficiently extract non-linear features, facilitating complex data representation and the mapping of data to higher-dimensional space. The training procedure for a multilayer perceptron employs the backpropagation algorithm, which refines a loss function by modifying weights according to gradients derived from error propagation (Equation (13)). This study utilized the Sigmoid function (Equation (14)) and the Mean Squared Error (MSE) loss function (Equation (15)), employing a network topology consisting of three hidden layers, each containing 128 nodes.
w w η L w
σ ( x ) = 1 1 + e x
L = 1 N i = 1 N y i log ( y ^ i ) + ( 1 y i ) log ( 1 y ^ i )

3.4.4. Model Validation

The stability of the training process was demonstrated via two main methods and one supplementary sub-strategy. The main strategies examine critical aspects affecting results and are discussed in following sections, while the sub-strategy aims to reinforce the study’s efficiency, emphasize its limitations, and drive future research directions.
First, the dataset was divided into two segments: a training/validation set comprising 80% and a test set comprising 20%. The test set was consistent during the training phase, whereas the training/validation set was divided, with 80% allocated for training and 20% for validation. The training process was conducted 1000 times, utilizing randomized selection of training and validation samples from the subset to ensure robustness. All trained models underwent evaluation with a fixed test set, and the results are presented as a box plot in Figure 4, demonstrating the stability of model performance across iterations.
Second, the dataset was evaluated using the k-fold cross-validation (k-foldCV) strategy with k = 10. In this approach, the dataset was divided into 10 folds, and the model was trained and tested 10 times, each time using a different fold as the test set, while the remaining 9 folds formed the training set. This ensured the complete separation of every data point for training and testing, thereby providing a comprehensive assessment of the model’s performance. The results of the k-fold cross-validation provide a comprehensive estimate of model performance and reduce the risk of overfitting by testing the model across varied data splits.
Finally, the sub-strategy required Leave-One-Subject-Out Cross-Validation (LOSOCV). This method divided the dataset of 166 subjects into 166 parts, executing training and testing of the model 166 times. In each iteration, data from a single subject served as the test set, while the remaining 165 subjects formed the training set, preventing any data leakage from the test subject into the training process. In contrast to the other two procedures, LOSOCV evaluates each subject independently, allowing it to be particularly resilient to overfitting. However, it differs from current methods and the existing study outlined in Section 4.2.4, which interpret signals post noise addition as independent units. In LOSOCV, the limited dataset of 166 patients could inherently limit expected model accuracy to approximately 75–80%, which is considered acceptable. This strategy stresses a realistic evaluation by ensuring that signals from test subjects do not affect training, emphasizing generalizability above exaggerated results.

4. Results and Discussion

The post-training results and the pertinent outcomes are both illustrated and discussed in this section.

4.1. Results

4.1.1. PD Classification Task Results

Table 2 indicates that all three training models attained high accuracy levels, surpassing 85%, validating the research methodology and fulfilling the objectives of the PD recognition tests. When comparing different ways to extract features, sets that used M-SamEn did slightly better than those that used DM-SamEn, with a 2~2.5% advantage over similar feature selection methods. From the viewpoint of feature selection, models exhibited optimal performance with the SBS technique, followed by the SBS-CFS hybrid method and the CFS method. Regarding classifier performance, RBF-SVM attained the highest results, somewhat exceeding AW-KNN, with a significant performance disparity noted between RBF-SVM and MLP. The maximum attained model accuracy was 98.38% (DM-SamEn, SBS, and RBF-SVM), succeeded by 98.25% (M-SamEn, SBS, and RBF-SVM), whereas the minimum recorded accuracy was 85.16% (DM-SamEn, SBS-CFS, and MLP).
Figure 4 presents stable mean accuracy results, consistently exceeding 85%, with a few outliers observed. The method–model pair that achieved the highest mean accuracy in the PD classification task (DM-SamEn) is (SBS, AW-KNN), followed by (SBS-CFS, AW-KNN), and the method–model pair that achieved the lowest mean accuracy is (CFS, RBF-SVM). Table 3 shows the results of subject-level validation utilizing the LOSOCV strategy, with all models reaching an accuracy over 77%. The maximum recorded accuracy was 85.67% (CFS, RBF-SVM), succeeded by 85.47% (SBS-CFS, RBF-SVM), while the lowest achieved accuracy was 77.65% (CFS, MLP). The results in Table 3 reveal greater standard deviations than those in Table 2, which is linked to the restricted training data in the subject-level validation strategy.

4.1.2. Severity Classification Task Results

Table 4 indicates that the performance of all three models in the PD severity recognition task surpasses 85%. Consistent with the results in Table 2, feature sets employing M-SamEn marginally surpassed those utilizing DM-SamEn across identical feature selection methods. In terms of feature selection, the models exhibited optimal performance with the SBS technique, followed by the SBS-CFS hybrid and, finally, the CFS method. Among the classifiers, RBF-SVM yielded the most favorable results, marginally exceeding AW-KNN by less than 0.5%; however, a substantial performance disparity persisted between RBF-SVM and MLP, indicating MLP’s probable unsuitability for this process. The maximum recorded accuracy was 96.89% (M-SamEn, SBS, and RBF-SVM), followed by 96.80% (DM-SamEn, SBS, and RBF-SVM), while the minimum accuracy was 85.75% (DM-SamEn, SBS-CFS, and MLP).
In the severity classification task shown in Figure 4, the average accuracy results demonstrate high stability, with all values higher than 85%, similar to those observed in the PD classification task. The PD classification task results show that SBS (AW-KNN) is the method–model pair with the best average accuracy, followed by SBS-CFS (AW-KNN). The task records the lowest average accuracy for (CFS, RBF-SVM) and (CFS, MLP). In the subject-level validation shown in Table 5, the minimum recorded performance was 77.23% (SBS-CFS, MLP), while the maximum reached approximately 86.52% (SBS, RBF-SVM). In comparison to Table 4, the standard deviations in Table 5 are substantially elevated, revealing that the size of the LOSOCV input data strongly impacts the results. The performance variations between the signal-level validation (Table 2 and Table 4) and the subject-level validation (Table 3 and Table 5) are roughly 10~15%, emphasizing the effect of the validation strategy on model results.

4.1.3. Throughput Performance and CO2 Emission Results

Table 6 illustrates the throughput performance of the classification models across different tasks, feature extraction techniques, and feature selection techniques, supported by Table 7, which provides power consumption results in terms of CO2 emissions. The studies were performed using a device equipped with 128 GB RAM, an Intel(R) Core(TM) i9-10980XE CPU working at 3.00 GHz, and a Ubuntu 22.04 operating system.
Initially, when comparing the two feature extraction methods, models employing the M-SamEn feature set showed reduced throughput compared to those utilizing DM-SamEn. The disparity is due to the total number of features—M-SamEn produces 96 features, while DM-SamEn decreases this value to 36, improving computing efficiency. Moreover, the smaller feature set of DM-SamEn leads to reduced CO2 emissions relative to M-SamEn, emphasizing the computational and ecological benefits.
Secondly, among the feature selection techniques, models employing SBS typically demonstrate poorer throughput relative to those applying CFS and SBS-CFS. SBS maintains a greater quantity of features, eliminating only those that substantially affect model performance, preserving elevated computational complexity. Inversely, both CFS and SBS-CFS emphasize the elimination of highly correlated features, resulting in decreased processing time. SBS-CFS generally attains better throughput compared to CFS. The similarity can be seen in the CO2 emission results, with SBS exhibiting the greatest emissions, followed by SBS-CFS, and CFS showing the lowest emissions.
Finally, among the classification models, the MLP demonstrated the highest throughput performance, reaching 2.05 × 10 5 obj/s, while concurrently obtaining the lowest CO2 emissions among the three models. In the PD classification task, RBF-SVM showed an insignificant advantage over AW-KNN in terms of speed, while AW-KNN showed a slightly higher throughput in the severity classification task. The difference can be linked to the quantity of labels—PD classification includes two labels, whereas severity classification comprises three. The increased label issues impact SVM more substantially than KNN due to the former’s dependence on complex decision boundaries, but KNN’s lazy learning technique is less impacted.

4.1.4. Paired t-Test Results

The accuracy results derived from training and validation in Section 3.4.4 were further examined using a paired t-test, with p-values adjusted via the Benjamini–Hochberg procedure for each combination of classification models and feature selection methods. Figure 5 displays the p-value matrix for various pairings. The analysis found that the majority of p-values are <0.05 for both M-SamEn and DM-SamEn, suggesting significant differences in the mean accuracy values among the classification model–selection method combinations. This demonstrates that each combination has unique performance traits. However, several p-values surpass 0.05, indicating the absence of statistically significant differences for these combinations.
According to Table 2 and Table 4, high p-values tend to indicate a crucial accuracy barrier beyond which further increases are challenging to overcome. The criterion for the PD classification task is between 98% and 98.5%, but for the severity classification task, the threshold is between 96% and 97%. These thresholds may signify performance saturation, wherein the models and feature selection techniques achieve their optimum accuracy under the given constraints.

4.2. Discussion

The results (Section 4.1) demonstrate that the proposed procedure (Figure 2), which integrates DM-SamEn for feature extraction with CFS, SBS, and the hybrid SBS-CFS for feature selection and utilizes AW-KNN, RBF-SVM, and MLP classifiers, attained high classification accuracy (>85%) and exhibited robust model stability. Further analysis is necessary to determine whether DM-SamEn features derived from signals generated via Equations (1)–(3) primarily contribute to model training or if feature selection methods identify DM-SamEn features as the predominant factors for optimal model performance. Furthermore, it is crucial to assess the quantity of feature reduction necessary to sustain the model’s average accuracy relative to the original feature set.

4.2.1. Analysis of Multicollinearity and Redundancy Problems in Feature Sets

Multicollinearity is an important issue considered in this study due to the fundamental properties of the feature extraction techniques, M-SamEn and the proposed DM-SamEn. The biggest challenge is the clinical interpretability of these features: entropy values quantify fluctuation patterns in the VGRF signal and are used as features for classification tasks. M-SamEn produces multiple entropy values across multiple scales, resulting in feature redundancy and elevated correlation coefficients among features. Multicollinearity can lead to inadequate post-trained models, since excessive feature dependency can cause overfitting, indicated by unnaturally boosted performance metrics. In order to address this issue, DM-SamEn consolidates entropy values into a single representative value, thereby drastically reducing redundancy. Figure 6 depicts the correlation matrix for the feature sets derived from M-SamEn (96 features) and DM-SamEn (36 features). The color gradient, ranging from white to dark blue, depicts correlation values ranging from 0 to 1. Despite a reduction in feature size using DM-SamEn, a significant quantity of highly correlated features (>0.7) exists.
Three feature selection strategies—CFS, SBS, and SBS-CFS—were implemented to further reduce multicollinearity and enhance performance. These strategies aim to minimize the quantity of correlated features while preserving crucial predictive information. This strategy enhances the reliability of the trained models, increases inference throughput, and reduces model size. The condensed models are beneficial for implementation on portable devices, allowing for potential trials in clinical contexts.

4.2.2. Analysis of Dominant Feature Types in Selected Feature Sets

This study assessed the impact of secondary signals generated in Section 3.1 on the classification results for both PD and severity classification tasks, concentrating on the feature sets derived from the selected pairs of high-performing classification models and selection methods. Figure 7a demonstrates that secondary signals derived from Equations (1)–(3) comprised 65% and 57% of the features employed in PD and severity classification tasks, respectively. This major contribution emphasizes the significance of the preprocessing stage, as the crucial factors are derived from these secondary signals, highlighting their essential role in improving the training process and overall classification performance.
The subsequent analysis analyzes features obtained from the DM-SamEn method and evaluates their impact on the selection of high-performing classification model pairs. Figure 7b illustrates that DM-SamEn’s features include an extensive percentage, representing 55% of features in PD classification and 59% in severity classification tasks. The findings suggest that the entropy values generated by the DM-SamEn approach provide the essential information required to enhance classification accuracy and boost classifier performance.

4.2.3. Analysis of Feature Selection Efficiency

Figure 8 illustrates four groups that indicate the number of features retained following the feature selection stage for each method. Each group comprises four columns: the blue and gray columns denote the PD classification task utilizing the M-SamEn and DM-SamEn methods, respectively, whereas the orange and yellow columns signify the same for the severity classification task.
In the “Original” group, feature extraction with M-SamEn yielded 96 features, whereas DM-SamEn diminished this to 36 features. As indicated in Section 4.1, the training results reveal a minor variance in performance between models trained using M-SamEn and DM-SamEn, with DM-SamEn yielding results that are merely 2–2.5% poorer. Within the “CFS” group, M-SamEn shrank the feature-set size from 96 to 64 and 59 for the PD and severity classification tasks, respectively, while DM-SamEn reduced the feature count from 36 to 20 for both tasks. Despite a reduction in features, the performance of the “CFS” models remains elevated, as shown in Table 2 and Table 4, but there is a performance decline beyond 2% when utilizing DM-SamEn instead of M-SamEn.
In the “SBS”, the quantity of features is not noticeably diminished over the selection stage. Applying M-SamEn yields a reduction in feature size from 96 to 90 and 87 for the PD and severity classification tasks, respectively. With DM-SamEn, the features decreased from 36 to 31 and 24 for the two tasks. The result indicate that models from the “SBS” group exhibit superior performance while retaining most of the features. Finally, the “SBS-CFS” group presents marginally poorer results compared to the “SBS” group. This group competently integrated the advantages of both techniques—sequential elimination from SBS and elimination of highly correlated features from CFS—thereby reducing irrelevant and duplicated features while ensuring accurate results. In M-SamEn, the feature-set size was diminished from 96 to 56 for the PD classification task and to 64 for the severity classification task. For DM-SamEn, the feature-set size was reduced from 36 to 17 and 14 for the PD and severity classification tasks, respectively. The results obviously illustrate a trade-off between the size of the feature set provided to the classifiers and the performance of the models post training.

4.2.4. Performance Comparison with Existing Studies

This section evaluates the performance reported in the current study in relation to existing research. Table 8 summarizes that other studies report high accuracy rates, exceeding 93%, consistent with our findings. The maximum accuracy attained in this study was 98.38% for PD classification and 96.80% for severity classification. Most previous studies utilize more sophisticated classification models, indicating increased computational power and energy usage. Our approach, utilizing the DM-SamEn method for feature extraction and the SBS-CFS method for feature selection, effectively identifies key features while minimizing computational demands. The methodology employed in this study attains a competitive accuracy through the use of simpler models, thereby minimizing the requirements for substantial computational power and energy consumption.

4.2.5. Limitations of the Study

Despite the impressive results of this study, some limits must be recognized, as they offer significant guidance for future research goals. This study utilized a reliable and publicly accessible dataset as the primary data source. This reliance presents issues, including data imbalance and dependence on a singular dataset, which may restrict the generalizability of the results. The dataset may poorly represent varied populations of Parkinson’s disease (PD) patients or adjustments to data collection strategies. External dataset validation is essential to strengthen the reliability of the results.
Secondly, this study acknowledged the problem of data imbalance but resisted using augmentation techniques like the Synthetic Minority Oversampling Technique (SMOTE) due to concerns about the higher risks of data leakage that come with the generation of artificial data. Augmentation algorithms may not consistently provide the most efficient option for rebalancing datasets in this case. Moreover, the restricted dataset size made subject-level validation impractical, even when the post-training performance surpassed 77%, as indicated in Table 3 and Table 5. A noise injection approach was presented in Theorem 2 to prevent data leakage. However, this approach requires further evaluation with larger and more diverse datasets to validate its effectiveness.
Thirdly, the study lacked real-world validation and prospective testing, which would have offered a deeper evaluation of the robustness and utility of the proposed study in clinical contexts. Collaboration with clinical Parkinson’s disease research groups and access to real-world datasets are essential to performing external validation. In addition, despite the DM-SamEn approach exhibiting technical efficiency, the entropy values it produces presently lack clinical significance for the early diagnosis of Parkinson’s disease. Future studies should focus on optimizing the feature set to improve its clinical significance and applicability in early detection.
Ultimately, due to the inherent noisiness of VGRF signals, this approach requires significant improvements to minimize noise effects and adapt hardware constraints, particularly in devices with restricted computational capabilities, such as IoT devices.

5. Conclusions

This study outlines a method for classifying Parkinson’s disease and its severity by analyzing vertical ground reaction force signals from patients. We proposed DM-SamEn as an alternative feature extraction method to standard M-SamEn, which effectively shrinks the feature-set size while maintaining training performance, thereby reducing computational demands. We enhanced model reliability and training efficiency by evaluating three feature selection methods: CFS, SBS, and a hybrid SBS-CFS approach. These addressed multicollinearity within the feature set. Ultimately, three classifiers—AW-KNN, RBF-SVM, and MLP—were employed, demonstrating the effectiveness of the feature extraction and selection methods while avoiding the necessity for overly complex models. Despite these advancements, this study admits several limitations, including the lack of external validation in clinical situations, constrained clinical interpretability of features developed by the DM-SamEn method, data imbalance, limited dataset availability, and challenges with adapting the methodology for resource-limited hardware such as IoT devices.

Author Contributions

Methodology and original draft preparation, M.T.P.N.; original draft preparation and supervision, M.K.P.T.; review and editing and supervision, T.N.; funding acquisition, T.H.T.; methodology, formal analysis, original draft preparation, and supervision, Q.D.N.N.All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Japan Science and Technology Agency (JST) under Strategic Basic Research Programs Precursory Research for Embryonic Science and Technology (PRESTO) under Grant JPMJPR20M6.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement:

The dataset used in the study can be found at https://physionet.org/content/gaitpdb/1.0.0/ accessed on 16 December 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Schalkamp, A.K.; Peall, K.J.; Harrison, N.A.; Sandor, C. Wearable movement-tracking data identify Parkinson’s disease years before clinical diagnosis. Nat. Med. 2023, 29, 2048–2056. [Google Scholar] [CrossRef] [PubMed]
  2. Telarović, S. Epidemiology of Parkinson’s Disease. Arch. Psychiatry Res. 2023, 59, 147–148. [Google Scholar] [CrossRef]
  3. Pagano, G.; D’Addio, G.; Campi, M.D.; Donisi, L.; Biancardi, A.; Cesarelli, M. Rehabilitation Outcome in Patients undergone Hip or Knee Replacement Surgery using Inertial Technology for Gait Analysis. In Proceedings of the 2020 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Bari, Italy, 1 June–1 July 2020; pp. 1–5. [Google Scholar]
  4. Aşuroğlu, T.; Oğul, H. A deep learning approach for parkinson’s disease severity assessment. Health Technol. 2022, 12, 943–953. [Google Scholar] [CrossRef]
  5. Zhao, A.; Qi, L.; Li, J.; Dong, J.; Yu, H. A hybrid spatio-temporal model for detection and severity rating of Parkinson’s disease from gait data. Neurocomputing 2018, 315, 1–8. [Google Scholar] [CrossRef]
  6. Maâchi, I.E.; Bilodeau, G.A.; Bouachir, W. Deep 1D-Convnet for accurate Parkinson disease detection and severity prediction from gait. arXiv 2019, arXiv:abs/1910.11509. [Google Scholar]
  7. Suquilanda-Pesántez, J.D.; Zambonino-Soria, M.C.; López-Ramos, D.E.; Pineda-Molina, M.G.; Milán, N.S.; Muñoz, M.C.J.; Villalba-Meneses, G.F.; Almeida-Galárraga, D.A. Prediction of Parkinson’s Disease Severity Based on Gait Signals Using a Neural Network and the Fast Fourier Transform; Artificial Intelligence, Computer and Software Engineering Advances; Springer: Cham, Switzerland, 2020. [Google Scholar]
  8. Faisal, M.A.A.; Chowdhury, M.E.H.; Mahbub, Z.B.; Pedersen, S.; Ahmed, M.U.; Khandakar, A.A.; Alhatou, M.; Nabil, M.; Ara, I.; Bhuiyan, E.; et al. NDDNet: A deep learning model for predicting neurodegenerative diseases from gait pattern. Appl. Intell. 2023, 53, 20034–20046. [Google Scholar] [CrossRef]
  9. Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Havlin, S.; Hausdorg, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; et al. Physionet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [PubMed]
  10. Yogev, G.; Giladi, N.; Peretz, C.; Springer, S.; Simon, E.S.; Hausdorff, J.M. Dual tasking, gait rhythmicity, and Parkinson’s disease: Which aspects of gait are attention demanding? Eur. J. Neurosci. 2005, 22, 1248–1256. [Google Scholar] [CrossRef]
  11. Hausdorff, J.M.; Lowenthal, J.; Herman, T.; Gruendlinger, L.; Peretz, C.; Giladi, N. Rhythmic auditory stimulation modulates gait variability in Parkinson’s disease. Eur. J. Neurosci. 2007, 26, 2369–2375. [Google Scholar] [CrossRef]
  12. Frenkel-Toledo, S.; Giladi, N.; Peretz, C.; Herman, T.; Gruendlinger, L.; Hausdorff, J.M. Treadmill walking as an external pacemaker to improve gait rhythm and stability in Parkinson’s disease. Mov. Disord. 2005, 20, 1109–1114. [Google Scholar] [CrossRef] [PubMed]
  13. Frenkel-Toledo, S.; Giladi, N.; Peretz, C.; Herman, T.; Gruendlinger, L.; Hausdorff, J.M. Effect of gait speed on gait rhythmicity in Parkinson’s disease: Variability of stride time and swing time respond differently. J. Neuroeng. Rehabil. 2005, 2, 23. [Google Scholar] [CrossRef] [PubMed]
  14. Hoehn, M.M.; Yahr, M.D. Parkinsonism. Neurology 1998, 50, 318. [Google Scholar] [CrossRef]
  15. Nguyen, Q.D.N.; Liu, A.; Lin, C.W. Development of a Neurodegenerative Disease Gait Classification Algorithm Using Multiscale Sample Entropy and Machine Learning Classifiers. Entropy 2020, 22, 1340. [Google Scholar] [CrossRef] [PubMed]
  16. Cirugeda-Roldán, E.M.; Molina-Picó, A.; Novák, D.; Cuesta-Frau, D.; Kremen, V. Sample Entropy Analysis of Noisy Atrial Electrograms during Atrial Fibrillation. Comput. Math. Methods Med. 2018, 2018, 1874651. [Google Scholar] [CrossRef] [PubMed]
  17. Delgado-Bonal, A.; Marshak, A. Approximate Entropy and Sample Entropy: A Comprehensive Tutorial. Entropy 2019, 21, 541. [Google Scholar] [CrossRef] [PubMed]
  18. Li, W.; Shen, X.; Li, Y.; Chen, Z. Improved multivariate multiscale sample entropy and its application in multi-channel data. Chaos 2023, 33, 063125. [Google Scholar] [CrossRef] [PubMed]
  19. Jiao, W.; Li, G.; Jiang, Y.; Baim, R.; Tang, C.; Yan, T.; Ding, X.; Yan, Y. Multi-Scale Sample Entropy-Based Energy Moment Features Applied to Fault Classification. IEEE Access 2021, 9, 8444–8454. [Google Scholar] [CrossRef]
  20. Nguyen, Q.D.N.; Tran, T.H.; Nakano, T. Optimizing Parkinson’s Disease Classification and Severity Assessment Using Dense Multiscale Sample Entropy and Hybrid Feature Selection. In Proceedings of the 2023 11th International Conference on Information Technology: IoT and Smart City, Kyoto, Japan, 14–17 December 2023; ICIT ’23. pp. 64–69. [Google Scholar] [CrossRef]
  21. Dormann, C.F.; Elith, J.; Bacher, S.; Buchmann, C.M.; Carl, G.; Carré, G.; Márquez, J.R.G.; Gruber, B.; Lafourcade, B.; Leitão, P.J.; et al. Collinearity: A review of methods to deal with it and a simulation study evaluating their performance. Ecography 2013, 36, 27–46. [Google Scholar] [CrossRef]
  22. Michalak, K.; Kwasnicka, H. Correlation-based feature selection strategy in classification problems. Int. J. Appl. Math. Comput. Sci. 2006, 16, 503–511. [Google Scholar]
  23. Chormunge, S.; Jena, S. Correlation based feature selection with clustering for high dimensional data. J. Electr. Syst. Inf. Technol. 2018, 5, 542–549. [Google Scholar] [CrossRef]
  24. Aregbesola, S.O.; Won, J.; Kim, S.; Byun, Y.H. Sequential backward feature selection for optimizing permanent strain model of unbound aggregates. Case Stud. Constr. Mater. 2023, 19, e02554. [Google Scholar] [CrossRef]
  25. Chaurasia, V.; Chaurasia, A. Novel Method of Characterization of Heart Disease Prediction Using Sequential Feature Selection-Based Ensemble Technique. Biomed. Mater. Devices 2023, 1, 932–941. [Google Scholar] [CrossRef]
  26. Zhang, H.; Wang, Z.; Xia, W.; Ni, Y.; Zhao, H. Weighted Adaptive KNN Algorithm With Historical Information Fusion for Fingerprint Positioning. IEEE Wirel. Commun. Lett. 2022, 11, 1002–1006. [Google Scholar] [CrossRef]
  27. Sun, L.; Zhang, J.; Ding, W.; Xu, J. Feature reduction for imbalanced data classification using similarity-based feature clustering with adaptive weighted K-nearest neighbors. Inf. Sci. 2022, 593, 591–613. [Google Scholar] [CrossRef]
  28. Razaque, A.; Ben Haj Frej, M.; Almi’ani, M.; Alotaibi, M.; Alotaibi, B. Improved Support Vector Machine Enabled Radial Basis Function and Linear Variants for Remote Sensing Image Classification. Sensors 2021, 21, 4431. [Google Scholar] [CrossRef] [PubMed]
  29. Thurnhofer-Hemsi, K.; L’opez-Rubio, E.; Molina-Cabello, M.A.; Najarian, K. Radial basis function kernel optimization for Support Vector Machine classifiers. arXiv 2020, arXiv:abs/2007.08233. [Google Scholar]
  30. Liu, C.L. One-Vs-All Training of Prototype Classifier for Pattern Classification and Retrieval. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Türkiye, 23–26 August 2010; pp. 3328–3331. [Google Scholar]
  31. Taud, H.; Mas, J.F. Multilayer Perceptron (MLP). In Geomatic Approaches for Modeling Land Change Scenarios; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  32. Balaji, E.; Brindha, D.; Kumar, E.V.; Vikrama, R. Automatic and non-invasive Parkinson’s disease diagnosis and severity rating using LSTM network. Appl. Soft Comput. 2021, 108, 107463. [Google Scholar]
  33. Wang, Q.; Zeng, W.; Dai, X. Gait classification for early detection and severity rating of Parkinson’s disease based on hybrid signal processing and machine learning methods. Cogn. Neurodynamics 2022, 18 1, 109–132. [Google Scholar] [CrossRef]
  34. Alharthi, A.S.; Casson, A.J.; Ozanyan, K.B. Gait Spatiotemporal Signal Analysis for Parkinson’s Disease Detection and Severity Rating. IEEE Sensors J. 2021, 21, 1838–1848. [Google Scholar] [CrossRef]
  35. Nguyen, D.M.D.; Miah, M.; Bilodeau, G.A.; Bouachir, W. Transformers for 1D signals in Parkinson’s disease detection from gait. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 5089–5095. [Google Scholar] [CrossRef]
  36. Zeng, W.; Yuan, C.; Wang, Q.; Liu, F.; Wang, Y. Classification of gait patterns between patients with Parkinson’s disease and healthy controls using phase space reconstruction (PSR), empirical mode decomposition (EMD) and neural networks. Neural Netw. 2019, 111, 64–76. [Google Scholar] [CrossRef]
  37. Rehman, R.Z.U.; Din, S.D.; Guan, Y.; Yarnall, A.J.; Shi, J.Q.; Rochester, L. Selecting Clinically Relevant Gait Characteristics for Classification of Early Parkinson’s Disease: A Comprehensive Machine Learning Approach. Sci. Rep. 2019, 9, 17269. [Google Scholar] [CrossRef]
  38. Noella, R.S.N.; Gupta, D.; Priyadarshini, J. Diagnosis of Parkinson’s disease using Gait Dynamics and Images. Procedia Comput. Sci. 2019, 165, 428–434. [Google Scholar] [CrossRef]
Figure 1. (a) Gender and label distribution across each class in severity classification task; (b) gender and label distribution across each class in PD classification task.
Figure 1. (a) Gender and label distribution across each class in severity classification task; (b) gender and label distribution across each class in PD classification task.
Information 16 00001 g001
Figure 2. The methodology procedure has four stages: preprocessing, feature extraction, feature selection, and classification.
Figure 2. The methodology procedure has four stages: preprocessing, feature extraction, feature selection, and classification.
Information 16 00001 g002
Figure 3. The process of signal division using the time-slicing window method and outlier removal through the quartile approach and histogram analysis.
Figure 3. The process of signal division using the time-slicing window method and outlier removal through the quartile approach and histogram analysis.
Information 16 00001 g003
Figure 4. Boxplot analysis of model consistency for PD and severity classification tasks.
Figure 4. Boxplot analysis of model consistency for PD and severity classification tasks.
Information 16 00001 g004
Figure 5. (a) Adjusted p-value matrix of paired T-tests comparing accuracies between classifier–feature selection method pairs (M-SamEn); (b) adjusted p-value matrix of paired t-tests comparing accuracies between classifier–feature selection method pairs (DM-SamEn).
Figure 5. (a) Adjusted p-value matrix of paired T-tests comparing accuracies between classifier–feature selection method pairs (M-SamEn); (b) adjusted p-value matrix of paired t-tests comparing accuracies between classifier–feature selection method pairs (DM-SamEn).
Information 16 00001 g005
Figure 6. (a) Correlation matrix of the feature set extracted using the M-SamEn method; (b) correlation matrix of the feature set extracted using the DM-SamEn method.
Figure 6. (a) Correlation matrix of the feature set extracted using the M-SamEn method; (b) correlation matrix of the feature set extracted using the DM-SamEn method.
Information 16 00001 g006
Figure 7. (a) Distribution of selected features by signal source (original and computed signals from Equations (1)–(3) after the feature selection stage; (b) distribution of selected features by feature extraction method after the feature selection stage.
Figure 7. (a) Distribution of selected features by signal source (original and computed signals from Equations (1)–(3) after the feature selection stage; (b) distribution of selected features by feature extraction method after the feature selection stage.
Information 16 00001 g007
Figure 8. Comparison of feature count across feature selection methods (*: p-value < 0.05; **: p-value < 0.01).
Figure 8. Comparison of feature count across feature selection methods (*: p-value < 0.05; **: p-value < 0.01).
Information 16 00001 g008
Table 1. Demographic information of the PD dataset.
Table 1. Demographic information of the PD dataset.
TaskPD ClassificationSeverity Classification
LabelCoPD22.53
Age63.6566.364.2168.7870.8
Height (m)1.681.671.671.671.64
Weight (kg)72.53472.17273.18272.28966.3
Gait Speed (m/s)1.241.031.0810.79
Table 2. PD classification task performance metrics (10-FoldCV).
Table 2. PD classification task performance metrics (10-FoldCV).
ModelFSLabelM-SamEnDM-SamEn
PreSenSpecF1AccPreSenSpecF1Acc
AW-KNNCFSPDMean98.2898.4196.1698.3497.7294.4696.3287.4795.3893.56
SD0.410.690.810.360.460.801.011.400.660.85
COMean96.4896.1698.4196.3197.7291.7587.4796.3289.3693.56
SD1.320.810.70.660.462.271.41.011.330.85
SBSPDMean98.6998.7697.0498.7298.2498.0498.3695.6098.1997.51
SD0.310.750.750.390.50.360.840.720.440.58
COMean97.3097.0498.7691.1698.2496.3495.6098.3695.9597.51
SD1.40.750.750.750.51.730.720.840.860.58
SBS-CFSPDMean98.3198.4196.1998.3697.7494.3497.2489.5889.3694.11
SD0.300.660.690.370.480.721.041.270.570.74
COMean96.4896.1998.4196.3397.7493.6587.9892.6195.0494.11
SD1.120.690.660.700.482.251.271.041.120.74
RBF-SVMCFSPDMean93.6587.9892.6198.5698.0094.9696.1988.795.5793.87
SD0.500.531.20.430.60.890.641.580.670.84
COMean97.4795.9698.9296.7098.0091.2788.796.1989.9793.87
SD1.271.120.531.060.61.121.580.641.130.84
SBSPDMean98.2799.2196.1198.7398.2598.5999.0796.8198.8398.38
SD0.490.451.040.260.350.460.560.980.20.26
COMean98.1896.1199.2197.1298.2597.9396.8199.0797.3598.38
SD0.981.040.450.570.351.160.980.560.430.26
SBS-CFSPDMean98.4399.0096.4398.7198.2196.0597.2691.1896.6495.36
SD0.420.441.040.280.420.980.781.870.821.06
COMean97.6696.4399.0097.0498.2193.8191.1897.2992.4195.36
SD1.081.040.440.800.421.561.870.781.561.06
MLPCFSPDMean95.3595.3489.5895.3493.5888.5389.8374.1489.1684.98
SD0.780.901.950.520.651.491.322.251.231.38
COMean89.6789.5895.3489.5993.5876.6874.1489.8375.3484.98
SD1.471.950.901.090.651.632.251.321.361.38
SBSPDMean95.7895.8790.6895.8294.2493.7094.4286.0794.0891.83
SD0.870.881.490.570.70.941.151.690.700.84
COMean90.8290.6895.8790.7294.2487.8486.6894.4686.7591.83
SD1.711.490.880.870.72.091.691.151.050.84
SBS-CFSPDMean94.9094.9788.7394.9393.0288.3090.4273.2489.3895.16
SD0.930.711.500.710.881.101.092.830.630.88
COMean88.7588.7394.9788.7293.0277.2773.2590.4075.2195.16
SD1.491.510.711.180.882.882.821.092.240.88
Table 3. PD classification task performance metrics (LOSOCV).
Table 3. PD classification task performance metrics (LOSOCV).
ModelFSLabelM-SamEnDM-SamEn
PreSenSpecF1AccPreSenSpecF1Acc
AW-KNNCFSPDMean88.7883.7686.6283.8185.1982.0481.9879.6980.4380.83
SD2.243.313.042.602.222.362.962.942.352.19
COMean87.3386.6283.7684.6785.1983.7279.6981.9879.9980.83
SD2.383.043.312.222.222.452.942.962.412.19
SBSPDMean87.7184.3985.6783.9385.0387.4383.8485.3583.2783.27
SD2.353.222.942.532.192.373.43.022.672.27
COMean87.7385.6784.3984.5785.0387.5685.3583.8484.1883.27
SD2.382.943.222.382.192.493.023.42.422.27
SBS-CFSPDMean88.3083.9386.5984.4485.2684.9483.3782.2681.9482.81
SD2.363.22.852.412.172.363.213.172.552.22
COMean87.0386.5983.9385.0185.2686.0282.2683.3781.8282.81
SD2.332.853.22.322.172.433.173.212.542.22
RBF-SVMCFSPDMean84.8791.0180.3386.3685.6777.6186.672.1580.7279.37
SD2.372.553.352.152.132.242.53.22.072.14
COMean92.2480.3391.0183.3585.6785.6672.1586.6076.4879.37
SD2.053.352.552.132.132.433.22.52,642.14
SBSPDMean84.3691.0678.9786.0785.0183.4290.9978.0985.2684.54
SD2.522.533.752.252.322.372.723.412.282.19
COMean91.7478.9791.0681.8885.0192.0478.0890.9981.9684.54
SD2.193.752.533.032.322.213.412.722.192.19
SBS-CFSPDMean84.691.279.7486.4385.4780.4088.0274.5682.4081.29
SD2.422.423.442.122.192.422.713.572.202.24
COMean91.5879.7491.2082.9985.4788.7174.5688.0278.4681.29
SD2.233.442.422.722.192.423.572.712.832.24
MLPCFSPDMean81.986.6877.2782.5981.9878.3481.1674.1377.5977.65
SD2.372.633.242.132.082.443.173.32.452.24
COMean87.3777.2786.6879.7981.9882.574.1381.1675.6777.65
SD2.23.242.632.522.082.553.33.172.622.24
SBSPDMean83.2887.7478.9284.0483.3382.8986.3778.7182.9782.54
SD2.42.453.312.072.072.422.933.242.242.09
COMean88.3078.9287.7481.0783.3388.0378.7186.3780.6182.54
SD2.133.312.452.602.072.163.242.932.492.09
SBS-CFSPDMean82.1288.0877.5983.6382.8481.6682.5477.9579.8180.24
SD2.22.413.241.961.932.483.263.372.492.23
COMean88.8577.5988.0880.3182.8485.2077.9582.5478.8680.24
SD1.943.242.412.511.932.493.373.263.422.23
Table 4. Severity Classification Task Performance Metrics (10-FoldCV).
Table 4. Severity Classification Task Performance Metrics (10-FoldCV).
ModelFSLabelM-SamEnDm-SamEn
PreSenSpecF1AccPreSenSpecF1Acc
AW-KNNCFS2Mean96.7196.8196.5196.1596.1594.3594.7694.1594.5693.02
SD0.950.841.070.540.441.081.471.021.031.03
2.5Mean96.5097.1697.9196.8096.1592.4894.4795.3593.4493.02
SD1.190.970.700.410.441.861.301.231.241.03
3Mean92.4089.7999.0490.9296.1587.7980.1598.6483.7593.02
SD3.063.540.431.760.443.123.580.322.941.03
SBS2Mean97.1297.2996.9697.2096.7297.0696.5396.8996.7996.23
SD0.460.730.560.340.660.651.080.790.60.65
2.5Mean97.5097.5698.4697.5196.7296.4097.2597.8296.8196.23
SD1.261.010.780.670.661.350.950.750.520.65
3Mean92.7391.1299.0491.7796.7292.1091.5598.9891.6996.23
SD3.133.160.391.660.662.583.680.471.880.65
SBS-CFS2Mean97.2096.9797.1097.0896.4894.8695.0494.6494.9494.02
SD0.840.810.830.550.531.231.261.281.030.83
2.5Mean96.8597.5498.0797.1896.4894.5095.0996.6394.7994.02
SD1.270.730.790.640.531.051.040.670.910.83
3Mean91.7190.5598.9891.1896.4888.0685.4298.5586.6594.02
SD2.793.420.391.520.532.043.140.292.040.83
RBF-SVMCFS2Mean96.7697.4596.5897.1596.7293.5694.7293.2994.1292.84
SD0.950.960.970.710.671.611.391.431.241.18
2.5Mean97.8297.3698.7197.5996.7293.0093.8695.8293.4092.84
SD0.790.890.460.590.671.541.720.761.181.18
3Mean92.7390.9899.1191.6896.7288.2480.7998.6184.2392.84
SD3.063.880.352.120.672.443.50.391.951.18
SBS2Mean96.9497.9096.8597.2596.8696.9997.8896.9697.3096.80
SD0.670.850.730.490.430.900.730.810.680.82
2.5Mean98.1097.6298.8697.8496.8697.9497.5898.7897.7596.80
SD0.781.050.470.570.430.661.010.380.680.82
3Mean93.9789.6998.2691.5896.8692.0990.9298.8991.3796.80
SD2.984.530.382.360.432.654.120.422.390.82
SBS-CFS2Mean96.7697.6196.5997.1896.8694.4495.9994.1495.2094.18
SD0.951.120.940.780.771.021.600.891.131.1
2.5Mean97.9197.6498.7697.7896.8695.6194.5497.4395.0694.18
SD0.790.970.520.560.771.881.411.041.481.1
3Mean93.9090.7399.2391.9496.8687.7884.4198.4885.8694.18
SD3.854.130.442.420.772.363.320.3451.651.1
MLPCFS2Mean92.9193.4992.3793.1891.2689.9088.3190.1889.7686.78
SD1.241.281.560.740.561.822.451.931.361.65
2.5Mean91.1990.2894.7090.7091.2687.1288.1189.0585.2686.78
SD1.071.730.670.940.563.892.072.522.271.65
3Mean84.0684.1998.1883.9391.2675.9869.3797.3272.0386.78
SD5.455.110.764.270.566.816.270.5725.881.65
SBS2Mean93.2893.9892.9893.5991.7892.9393.0892.6592.9990.77
SD1.421.351.391.060.881.351.081.380.920.71
2.5Mean91.7391.9395.0191.8191.7889.5191.2993.5290.3290.77
SD1.321.330.850.810.882.321.921.320.870.71
3Mean84.8982.0398.1382.8391.7884.7179.3298.1881.7890.77
SD4.774.510.502.760.884.364.580.672.550.71
SBS-CFS2Mean92.0592.6691.6792.3190.1989.7088.1588.5988.9795.75
SD1.401.631.401.341.221.831.981.711.591.36
2.5Mean89.5990.2493.6489.9190.1983.3585.5389.4984.6195.75
SD2.391.891.421.151.222.151.631.401.621.36
3Mean83.1679.0598.1280.7490.1975.9174.6497.6775.3195.75
SD4.565.670.5124.061.226.966.170.695.201.36
Table 5. Severity Classification Task Performance Metrics (LOSOCV).
Table 5. Severity Classification Task Performance Metrics (LOSOCV).
ModelFSLabelM-SamEnDm-SamEn
PreSenSpecF1AccPreSenSpecF1Acc
AW-KNNCFS2Mean87.7481.892.3882.8985.9084.6678.5591.5679.8779.49
SD3.274.452.253.092.663.744.912.253.782.88
2.5Mean87.1181.8793.6682.2685.9079.1881.8987.5778.1679.49
SD3.45.321.714.22.663.764.922.523.952.88
3Mean87.5591.0191.3087.4885.9083.1978.0490.1077.7179.49
SD3.193.412.632.902.663.85.542.614.292.88
SBS2Mean88.2182.3092.1382.4385.3786.3082.2991.4681.8283.77
SD3.554.582.713.692.853.674.642.543.763.05
2.5Mean88.1082.7594.1685.0285.3786.4982.1892.3381.7383.77
SD3.535.31.742.852.853.375.112.134.093.05
3Mean87.5991.0791.7685.3785.3786.686.891.8585.0483.77
SD3.173.982.363.282.853.774.972.393.773.05
SBS-CFS2Mean87.1382.4192.3883.2585.2785.4080.2391.8880.9581.52
SD3.814.662.483.462.833.54.442.183.552.97
2.5Mean88.7582.7794.4983.6485.2780.5181.888.7779.2281.52
SD3.235.191.674.212.833.674.882.523.932.97
3Mean87.2890.6391.0386.8285.2785.8982.5491.6281.3681.52
SD3.243.82.683.022.833.264.942.193.822.97
RBF-SVMCFS2Mean85.5889.6789.4085.7286.2680.6783.6388.6681.1481.35
SD3.783.463.303.252.833.844.472.433.352.76
2.5Mean87.0385.4592.8384.0886.2681.9982.8188.9180.4681.35
SD3.445.042.014.122.833.684.352.693.522.76
3Mean94.8083.6697.1686.3786.2688.6577.694.4480.1481.35
SD2.25.531.384.192.833.075.271.634.082.76
SBS2Mean85.8188.7990.0286.6186.5283.6489.3688.7085.6686.23
SD3.883.582.993.332.774.153.813.083.122.85
2.5Mean87.1985.4492.6384.2686.5286.5684.9592.7083.7486.23
SD3.554.852.143.92.773.374.991.94.032.85
3Mean95.0285.3197.1286.7886.5296.0584.3697.9387.7286.23
SD2.35.341.374.22.771.755.270.923.882.85
SBS-CFS2Mean85.690.4489.4286.1986.4080.8283.3188.3380.4881.31
SD3.733.33.253.092.764.265.012.723.913.07
2.5Mean85.8584.0292.7983.6886.4081.4681.3188.7779.0881.31
SD3.835.271.944.172.763.884.812.713.913.07
3Mean95.2684.7497.3887.3486.4090.6779.2994.8692.1881.31
SD1.985.381.174.052.763.115.482.034.363.07
MLPCFS2Mean82.8783.7789.3481.7381.6778.8380.8487.2977.6377.47
SD3.553.82.763.292.73.74.612.463.612.81
2.5Mean78.8182.2487.4278.6581.6776.6576.7285.6674.1977.47
SD3.483.482.433.612.73.854.622.813.512.81
3Mean92.5179.0195.7481.981.6786.4874.8693.2677.4077.47
SD2.445.241.664.062.73.35.251.764.082.81
SBS2Mean80.7581.7488.1679.3680.3778.6179.8387.8978.5379.03
SD3.974.342.93.722.914.154.62.773.792.84
2.5Mean78.7380.8387.6377.5280.3776.5979.5285.7779.0379.03
SD4.015.152.554.262.913.954.822.713.832.84
3Mean90.5878.5394.7581.4480.3789.877.7394.8780.7879.03
SD3.155.031.963.892.913.135.531.724.22.84
SBS-CFS2Mean81.5483.9388.1580.8681.0377.5680.4385.5276.8477.23
SD3.623.972.893.322.864.254.693.253.973.48
2.5Mean80.0382.0688.1978.9281.0377.0675.3787.1273.4777.23
SD3.734.892.654.042.863.995.462.494.263.48
3Mean91.2677.1095.1980.0281.0386.7675.8993.2078.7677.23
SD2.684.261.674.152.863.765.042.264.203.48
Table 6. Inference throughput of classifiers and feature selection methods in PD and severity classification tasks.
Table 6. Inference throughput of classifiers and feature selection methods in PD and severity classification tasks.
ModelFSInference Throughput ( 10 5 obj/s)
PD ClassificationSeverity Classification
M-SamEnDM-SamEnM-SamEnDM-SamEn
CFS0.240.550.220.52
AW-KNNSBS0.180.410.180.43
SBS-CFS0.240.610.240.57
CFS0.40.510.160.34
RBF-SVMSBS0.340.510.150.31
SBS-CFS0.370.630.160.59
CFS1.341.990.921.7
MLPSBS1.041.70.751.33
SBS-CFS1.272.050.991.76
Table 7. CO2 emissions of classifiers and feature selection methods in PD and severity classification tasks.
Table 7. CO2 emissions of classifiers and feature selection methods in PD and severity classification tasks.
ModelFSCO2 Emissions ( 10 8 kg/obj)
PD ClassificationSeverity Classification
M-SamEnDM-SamEnM-SamEnDM-SamEn
TrainTestTrainTestTrainTestTrainTest
CFS0.230.1940.0830.0810.1320.1360.0560.062
AW-KNNSBS0.2980.2510.120.110.1690.1710.070.077
SBS-CFS0.2110.1820.0730.070.1230.1290.0490.056
CFS0.9150.1171.220.081.0270.18891.060.088
RBF-SVMSBS0.9370.1330.8110.0851.0710.2010.7890.095
SBS-CFS0.8720.1151.040.0681.0010.180.870.055
CFS0.3210.0380.310.0230.280.0350.2540.02
MLPSBS0.3480.0470.310.0270.1710.0430.2530.026
SBS-CFS0.3170.0380.2930.0220.2720.0340.2570.019
Table 8. Performance comparison with existing research methods.
Table 8. Performance comparison with existing research methods.
Existing ResearchAlgorithmTaskAccuracy
[32]Adam-LSTMPD Classification98.60%
Severity Classification96.6%
[33]PSR, SEE, DQSD,
VMD, and SVM
PD Classification98.92%
Severity Classification93.37%
[34]Parallel 2D-DCNNPD Classification95.5%
Severity Classification95.75%
[35]Transformer 1D-STEPD Classification95.2%
Severity ClassificationN/A
[36]PSR, EMD, and NNPD Classification98.8%
Severity ClassificationN/A
[37]RFE, and RFPD Classification94.28%
Severity ClassificationN/A
[38]PARAFAC, and TDPD Classification97%
Severity ClassificationN/A
Proposed AlgorithmAW-KNN, RBF-SVM, MLPPD Classification98.38%
Severity Classification96.80%
DQSD: Factor Signal Decomposition
VMD: Variational Mode Decomposition
PSR: Phase Space Reconstruction
SEE: Shannon Energy Envelope
1D-STE: 1D-Spatial Transformer Encoder
EMD: Empirical Mode Decomposition
RFE: Recursive Feature Elimination
RF: Random Forest
TD: Tucker Decomposition
PARAFAC: Parallel Factor Analysis
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nguyen, M.T.P.; Tran, M.K.P.; Nakano, T.; Tran, T.H.; Nguyen, Q.D.N. An Approach for Detecting Parkinson’s Disease by Integrating Optimal Feature Selection Strategies with Dense Multiscale Sample Entropy. Information 2025, 16, 1. https://doi.org/10.3390/info16010001

AMA Style

Nguyen MTP, Tran MKP, Nakano T, Tran TH, Nguyen QDN. An Approach for Detecting Parkinson’s Disease by Integrating Optimal Feature Selection Strategies with Dense Multiscale Sample Entropy. Information. 2025; 16(1):1. https://doi.org/10.3390/info16010001

Chicago/Turabian Style

Nguyen, Minh Tai Pham, Minh Khue Phan Tran, Tadashi Nakano, Thi Hong Tran, and Quoc Duy Nam Nguyen. 2025. "An Approach for Detecting Parkinson’s Disease by Integrating Optimal Feature Selection Strategies with Dense Multiscale Sample Entropy" Information 16, no. 1: 1. https://doi.org/10.3390/info16010001

APA Style

Nguyen, M. T. P., Tran, M. K. P., Nakano, T., Tran, T. H., & Nguyen, Q. D. N. (2025). An Approach for Detecting Parkinson’s Disease by Integrating Optimal Feature Selection Strategies with Dense Multiscale Sample Entropy. Information, 16(1), 1. https://doi.org/10.3390/info16010001

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop