In the ensuing subsections, the results and discussions pertaining to emerging mean network measures and information-theoretic measures are presented. These analyses are conducted initially from a statistical perspective and subsequently from a deep learning perspective.
4.1. Statistical Perspective
Violin plots in
Figure 2 illustrate the distributions of mean values for four key network metrics—Degree Centrality, Closeness Centrality, Betweenness Centrality, and Eigenvector Centrality—across two brain network conditions: Healthy Control (HC) and Parkinson’s Disease (PD) groups. These metrics provide insights into the topological characteristics of brain networks, highlighting how the brain’s network structure may evolve from a cognitively healthy state to one associated with neurodegenerative changes like those observed in Parkinson’s Disease.
Starting with Degree Centrality, the distributions for both HC and PD groups appear relatively similar, with slight variations in the median values and the spread of the data. This similarity suggests that, on average, the strength of connections—quantified as the sum of the weights of connections at each node—in the networks of HC and PD subjects may not differ substantially. However, the violin plot reveals subtle differences in the density distribution, hinting at potential outliers or slight variations in node connectivity strength within each group. These nuances could be important, as they might reflect individual differences in how brain regions are interconnected in healthy versus diseased states.
In contrast, the Closeness Centrality plots for HC and PD reveal more pronounced differences. The distributions indicate that, although both groups have a central peak in similar regions, the spread is slightly wider for the PD group. This wider spread could imply that individuals with Parkinson’s Disease exhibit more variability in how “close” a node is to all other nodes in terms of the average shortest path lengths. This variability might reflect disruptions in efficient communication pathways within brain networks due to the disease, which lead to alterations in functional integration. The increased spread in the PD group could be indicative of heterogeneous disease progression or compensatory mechanisms attempting to maintain network efficiency despite pathological changes.
The Betweenness Centrality distributions for both groups show an interesting pattern: very narrow plots with minimal density, indicating that this measure is consistently low across both HC and PD subjects. This consistency suggests that the role of nodes as intermediaries in the shortest paths between other nodes is relatively limited in these brain networks. Moreover, the minimal differentiation between the two groups concerning this specific centrality measure suggests that Betweenness Centrality may not be significantly affected by the neurodegenerative processes in Parkinson’s Disease or that it is not a sensitive marker for distinguishing between healthy and diseased brain networks in this context.
The most striking difference is observed in the Eigenvector Centrality plots. Here, the distributions for the HC and PD groups show a marked contrast, with the HC distribution being more concentrated around a central value, while the PD distribution spreads much wider across a range of values. This indicates a significant difference in the influence of nodes within the networks. In the PD group, nodes might exhibit greater variability in how connected they are to other highly connected nodes. The broad distribution for PD suggests that certain regions of the brain may have enhanced or reduced influence in network communication compared to the HC group. This variability could be due to the disease’s effect on neural pathways and functional connectivity, possibly leading to some nodes becoming more central in compensatory networks, while others lose their centrality due to degeneration.
Figure 3 provides violin plots illustrating the distribution of information-theoretic measures—specifically Network Entropy, Network Complexity, and IIT—for both HC and PD groups. These plots allow for a comparative analysis of how these metrics vary between the two groups, offering insights into potential differences in the complexity and informational properties of brain networks in healthy and diseased states.
The plot for Network Entropy demonstrates notable differences between the HC and PD groups. Both distributions have a similar overall shape, but the HC group appears to have a slightly more dispersed distribution compared to the PD group, which exhibits a more concentrated range around the median. This suggests that while the average entropy—reflecting the randomness or unpredictability of the brain network connections—is somewhat consistent across both groups, the HC group shows greater variability. This variability in entropy might indicate that healthy brains have a wider range of complexity levels, potentially related to adaptive or dynamic responses in neural connectivity. The PD group’s relatively more uniform distribution could imply a certain level of homogeneity in brain network entropy, potentially due to reduced flexibility or altered neural connectivity patterns characteristic of the disease.
In terms of Network Complexity, both groups display minimal differences, with only slight variations. The violin plots show that both the HC and PD groups have similarly low values, with minor deviations in their spread. This uniformity could suggest that this specific measure may not strongly differentiate between HC and PD brain networks or that it is less sensitive to the connectivity changes induced by Parkinson’s Disease. The minimal spread also hints at a consistent structural property across both healthy and diseased brain networks, possibly indicating that network complexity remains relatively stable despite neurodegenerative changes.
The distribution for IIT reveals some distinctions between the two groups. The violin plot indicates that while the distributions of IIT values are centered similarly for both HC and PD groups, the HC group exhibits a more condensed shape, suggesting a tighter clustering of IIT values. The PD distribution, on the other hand, shows a slightly broader range, indicating more variability in how information integration and differentiation occur within the PD brain networks. This variability could reflect disruptions in how efficiently information is processed or integrated—a hallmark of Parkinson’s Disease as it affects cognitive and motor functions. The broader distribution for the PD group suggests that IIT could potentially be an informative measure for distinguishing subtle changes in brain network processing and structural integration in PD patients.
Figure 4 presents the H-C Plane plot, mapping the relationship between entropy (H) and complexity (C) for brain networks across HC and PD groups. The HC group is represented in blue, while the PD group is shown in red, allowing for a visual comparison of how these two measures vary between the groups. The H-C Plane is a valuable visualization as it provides insights into how randomness (entropy) and structural complexity coexist in brain networks, informing us about the network’s stability, efficiency, and adaptability.
Figure 4 reveals an intriguing distribution pattern between the HC and PD groups. For both groups, there is a generally negative correlation between entropy and complexity, consistent with the theoretical notion that systems with higher entropy typically exhibit lower structural complexity and vice versa. However, the distributions show that the PD group tends to have higher complexity values than the HC group at equivalent entropy levels. This is reflected by the broader spread and denser clustering of red points (PD group) in the upper complexity region of the plot.
This suggests that, in PD patients, brain networks may exhibit more pronounced structural organization despite having varying levels of randomness. Such an increase in structural complexity in PD could be indicative of compensatory mechanisms or pathological network changes associated with the disease. The PD group’s wider spread in complexity might reflect network reorganization efforts involving changes in both short-term and long-term dependencies.
The distribution of the blue points (HC group) is more concentrated along a lower complexity range with a relatively wider spread in entropy. This implies that healthy brain networks might maintain a balance between randomness and complexity, indicative of an adaptable and efficient system capable of dynamically responding to external stimuli. The higher entropy levels in the HC group suggest greater variability and flexibility in neural connectivity, essential for normal cognitive functioning.
The NARDL model captures both short-run and long-run dependencies between pairs of Regions of Interest (ROIs), aggregated into a robust metric reflecting combined influences over time. The CD measures provide a nuanced view of functional connectivity, capturing directional and asymmetric interactions within the brain. Integrating this dependency information into adjacency matrices and processing it into network features like entropy and complexity enriches the analysis by incorporating temporal dynamics that might otherwise be overlooked.
In the context of the H-C Plane, the higher complexity observed in the PD group likely reflects network reorganization efforts involving changes in both short-term and long-term dependencies. The CD values derived from NARDL analyses contribute to these elevated complexity levels, indicating that PD networks might rely more on long-term compensatory connections or exhibit altered interactions that increase structural organization but at the cost of adaptability. The PD distributions covering a wider range of complexity suggest that disease-related changes in neural dynamics are more variable, possibly due to differences in disease severity, progression rates, or the brain’s heterogeneous response to degeneration.
In what follows in this subsection, statistical tests are presented after applying SMOTE to address the significant class imbalance present in the dataset. The original dataset comprises 1537 Parkinson’s Disease samples compared to only 142 Healthy Control samples, which poses challenges for reliable statistical analysis. Class imbalance can lead to biased results, reduce the statistical power of tests, and increase the likelihood of Type I and Type II errors, thereby obscuring true differences between groups. By implementing SMOTE, synthetic samples are generated for the minority class (HC), effectively balancing the class distribution. This balancing enhances the fairness and sensitivity of statistical tests, ensuring that comparisons between HC and PD groups are not disproportionately influenced by the overwhelming number of PD samples. Consequently, SMOTE facilitates a more accurate and equitable assessment of the underlying differences in network measures between the two groups. Additionally, balancing the classes allows for more robust effect size estimations and clearer insights into the relationships between variables. However, it is important to acknowledge that while SMOTE mitigates class imbalance, it introduces synthetic data, which may not fully capture the natural variability of the minority class, potentially affecting the authenticity of the results. Therefore, the application of SMOTE in this context is carefully considered to enhance the validity of statistical analyses while being mindful of its limitations.
The Shapiro–Wilk test is a statistical method used to assess the normality of a dataset by comparing the data’s distribution to a theoretically normal distribution. The null hypothesis for the Shapiro–Wilk test states that the data are normally distributed. If the p-value obtained from the test is less than the chosen significance level (commonly 0.05), the null hypothesis is rejected, indicating that the data do not follow a normal distribution. This test is particularly useful for verifying the appropriateness of parametric statistical methods, which often assume normality in the data.
The Shapiro–Wilk test for normality conducted on the network metrics for both the HC and PD groups after applying SMOTE demonstrates in
Table 1 that the data for all tested metrics remains significantly non-normal. The test statistics for each metric are consistently below 1, and the
p-values are either extremely small or effectively zero, indicating a strong rejection of the null hypothesis of normality. For example, in the HC group, Degree Centrality has a test statistic of 0.8371 with a
p-value of
, confirming a significant deviation from normality. Similar results are observed for metrics like Closeness Centrality, Betweenness Centrality, and Eigenvector Centrality, where the
p-values are effectively zero, emphasizing highly non-normal distributions.
The results are consistent across both groups, suggesting that the application of SMOTE, which is used to balance the data between HC and PD groups by synthetically generating examples in the minority class, does not alter the underlying distributional characteristics of the features. This finding highlights that while SMOTE is useful for balancing class representation in a dataset, it does not modify the shape or normality of the feature distributions. The persistence of non-normality in all tested metrics after SMOTE has implications for the choice of statistical methods in subsequent analysis. Parametric tests that assume normal distributions, such as t-tests or ANOVA, are not suitable for these data. Instead, non-parametric methods, such as the Mann–Whitney U test, should be employed to ensure valid results when comparing metrics between the HC and PD groups.
Levene’s test for homogeneity of variance is a statistical method used to determine whether the variances of different groups are equal. This test is particularly useful when comparing the assumption of equal variances between groups in preparation for statistical tests that assume homogeneity of variance. The null hypothesis for Levene’s test states that the variances across groups are equal. If the p-value is less than the significance level (typically 0.05), the null hypothesis is rejected, indicating unequal variances between groups.
The results of Levene’s test in
Table 2 for homogeneity of variance after applying SMOTE reveal that for most network metrics, the null hypothesis of equal variances between the HC and PD groups is rejected. Metrics such as Degree Centrality, Closeness Centrality, Betweenness Centrality, Eigenvector Centrality, Network Entropy, and IIT have
p-values that are significantly below 0.05. For example, Degree Centrality shows a test statistic of 15.6362 with a
p-value of
, and IIT exhibits an even more pronounced result with a statistic of 27.7121 and a
p-value of
. These results suggest that the assumption of homogeneity of variance does not hold for these metrics, indicating significant differences in variances between the HC and PD groups.
On the other hand, Network Complexity presents a p-value of 0.1294, which is above the 0.05 threshold, suggesting that the null hypothesis of equal variances cannot be rejected for this metric. This indicates that the variances for Network Complexity are similar between the HC and PD groups, maintaining the homogeneity of variance assumption for this specific feature.
The Mann–Whitney U test is a non-parametric statistical test used to determine whether there are significant differences between the distributions of two independent groups. This test is particularly suitable when the data do not meet the assumptions of normality or homogeneity of variance, as identified in the previous analyses. The null hypothesis of the Mann–Whitney U test states that the distributions of the two groups are the same. A p-value less than the chosen significance level (e.g., 0.05) indicates that the null hypothesis should be rejected, implying a significant difference between the groups.
The Mann–Whitney U test results in
Table 3 for network metrics after applying SMOTE indicate significant differences between the HC and PD groups for specific metrics, while others do not show significant differences. Closeness Centrality and Betweenness Centrality both have
p-values that fall below the standard significance threshold, even after adjustment for multiple comparisons. Closeness Centrality has an adjusted
p-value of 0.0098, while Betweenness Centrality shows a highly significant result with an adjusted
p-value of
. This indicates that there are meaningful distributional differences in these metrics between the HC and PD groups. The effect sizes for these two features, −0.0665 for Closeness Centrality and −0.1021 for Betweenness Centrality, suggest that the differences, while statistically significant, are moderate in magnitude.
For other metrics, such as Degree Centrality, Network Complexity, and IIT, the results are not significant after adjusting for multiple comparisons, as indicated by their adjusted p-values being greater than 0.05. For instance, Degree Centrality has an adjusted p-value of 0.6450, and IIT shows no significant difference with an adjusted p-value of 1.0. The effect sizes for these metrics are also minimal, indicating negligible practical differences between the groups. Eigenvector Centrality, Network Entropy, and IIT, in particular, exhibit very small effect sizes (e.g., −0.0096 for Eigenvector Centrality), reinforcing the conclusion that the distributions of these metrics between HC and PD are similar.
The presence of statistically significant results for some features but not others suggests that certain network metrics are more sensitive to the differences between the HC and PD groups. Betweenness Centrality and Closeness Centrality, for example, may reflect more distinct structural or functional network properties related to brain connectivity in Parkinson’s Disease. In contrast, metrics such as IIT and Eigenvector Centrality do not appear to provide strong differentiation between the two groups in this analysis.
The Bonferroni correction is a statistical method used to adjust p-values for multiple comparisons to control for Type I error (false positives). This correction involves multiplying the original p-value by the number of comparisons made, thus making it more stringent to achieve significance. The aim is to ensure that the probability of making at least one Type I error remains below a chosen significance level, typically 0.05.
After applying the Bonferroni correction, only Closeness Centrality and Betweenness Centrality remain statistically significant in
Table 4. The adjusted
p-values for these metrics are 0.0098 and
, respectively, indicating that the differences between the HC and PD groups for these features are robust and unlikely to be due to random chance. This suggests that Closeness Centrality and Betweenness Centrality may be particularly relevant metrics for differentiating between these two groups, potentially highlighting underlying differences in network properties, such as how central or mediating certain nodes are within the brain network.
For the other metrics, the Bonferroni correction has led to adjusted p-values that exceed the significance threshold of 0.05. For instance, Degree Centrality has an adjusted p-value of 0.6450, and Network Complexity is at 0.2398, both of which indicate non-significant differences between the HC and PD groups after accounting for multiple comparisons. Metrics such as Eigenvector Centrality, Network Entropy, and IIT show even higher adjusted p-values of 1.0, underscoring that these features do not provide significant differentiation between the groups under the stricter criteria of the Bonferroni correction. The correction highlights that while initial p-values might indicate significance, adjusting for multiple tests is crucial for maintaining the reliability of results. The fact that only two features remain significant after correction suggests that Closeness Centrality and Betweenness Centrality could be more sensitive to the changes in brain network structure associated with Parkinson’s Disease. This underscores the importance of these metrics in network analysis and their potential role in characterizing differences in brain connectivity.
The comparison between the tables reveals key insights into the statistical behavior of network metrics derived from the NARDL-based formation of networks for the HC and PD groups. The Shapiro–Wilk test results consistently show that all network metrics for both groups are non-normally distributed, even after SMOTE was applied to balance the class sizes. This outcome suggests that the complex and nuanced network formations obtained using NARDL, which capture both short-term and long-term dependencies between ROIs, inherently produce metrics that do not conform to normal distributions. The non-normality observed across metrics such as Degree Centrality, Closeness Centrality, and Betweenness Centrality reflects the intricate dependencies modeled by NARDL that characterize brain connectivity in both healthy and diseased states. Levene’s test for homogeneity of variance indicated that most network metrics do not meet the assumption of equal variances between the HC and PD groups, with the exception of Network Complexity. The significant p-values for metrics such as Closeness Centrality and Betweenness Centrality highlight the variance differences captured in the NARDL-based network structure. This suggests that these specific metrics are sensitive to the heterogeneity present in brain connectivity, potentially reflecting the effects of Parkinson’s Disease on network stability and interaction strength across different brain regions. The Mann–Whitney U test provided a detailed look at which metrics showed significant distributional differences between HC and PD groups. Initially, Closeness Centrality and Betweenness Centrality were found to be significant, reinforcing their role as key metrics that can differentiate between the two groups based on the network structure derived from NARDL. However, after applying the Bonferroni correction, which adjusts for multiple comparisons to reduce the risk of Type I errors, only these two metrics remained significant. This outcome highlights their robustness and reliability in detecting meaningful differences when stringent significance criteria are applied. The Bonferroni correction results provide a final layer of comparison by showing how the significance of the metrics changes when correcting for multiple comparisons. While the Mann–Whitney U test suggested initial significance for some metrics, only Closeness Centrality and Betweenness Centrality retained their significance post-correction. This adjustment emphasizes that the detected differences in these metrics are not due to chance, reinforcing the value of these specific measures derived from the NARDL-based network formations in understanding brain network alterations in Parkinson’s Disease.
The correlation matrix of network measures after applying SMOTE in
Figure 5 reflects the intricate relationships between the metrics derived from the NARDL-based network formation and sheds light on how these relationships might differ between the HC and PD classes. NARDL, which models both short-term and long-term dependencies between ROIs, contributes to a deeper understanding of functional connectivity and the derived metrics that capture different aspects of this connectivity. The results observed in the correlation matrix suggest that the way these metrics interrelate may highlight structural and functional differences in the brain connectivity of HC and PD groups.
The strong positive correlation between Closeness Centrality and Eigenvector Centrality (0.98) implies that these centrality metrics share a similar role in characterizing the central and influential nodes within the brain’s network. In the context of the NARDL-based formation, this correlation suggests that brain regions that are centrally positioned (with higher Closeness Centrality) are also likely connected to other highly influential regions (indicated by high Eigenvector Centrality). This relationship could be more pronounced or vary in complexity between HC and PD groups, where PD networks may exhibit changes in how these central regions interact due to disruptions in connectivity. The dependency structures captured by NARDL might reveal how centrality shifts from healthy to diseased states, with certain regions losing or gaining prominence in PD, leading to potential changes in this strong correlation.
Network Entropy’s positive correlation with both Degree Centrality (0.67) and Closeness Centrality (0.72) reflects that networks with more extensive node connections and central nodes are associated with higher levels of randomness or connectivity complexity. From a NARDL-based perspective, this relationship highlights how dependencies between brain regions contribute to overall network entropy, potentially indicating how randomness in connectivity patterns might differ in HC and PD groups. In PD, changes in the short-term and long-term relationships between ROIs could alter how entropy correlates with centrality measures. For example, disrupted or reorganized connectivity in PD might lead to a shift in how these metrics align, suggesting that PD networks may have altered the balance between connectivity strength and entropy compared to HC networks.
The negative correlation involving IIT (e.g., −0.78 with Closeness Centrality and −0.79 with Eigenvector Centrality) is particularly intriguing when considering the NARDL-based network formation. IIT captures the network’s integrated and differentiated information processing capabilities, and its inverse relationship with centrality measures suggests that as networks become more integrated in their information processing, they may rely less on central nodes or specific pathways. In PD, this relationship could indicate that as the disease progresses, the brain’s ability to process information in an integrated manner is compromised, potentially leading to a reduction in the importance of central nodes and an increase in distributed network processing. The NARDL-based model, which incorporates both short-term and long-term dynamics, can reveal how PD alters the balance between local node importance and global information integration, providing a unique perspective on how these relationships shift between HC and PD.
Network Complexity’s moderate negative correlation with IIT (−0.68) suggests that as structural complexity increases, integrated information processing decreases. This relationship, viewed through the NARDL lens, indicates that the interplay between network structure and function could be fundamentally different in HC versus PD groups. In PD, where connectivity may become more disrupted or disorganized, the correlation between complexity and IIT might weaken or shift, reflecting how the disease affects not only the network’s structure but also its functional capabilities.
Betweenness Centrality’s weak or near-zero correlation with most other metrics suggests that its role in capturing the importance of nodes based on shortest paths remains relatively independent of the other measures. This independence might be critical in understanding PD, as Betweenness Centrality could highlight specific pathways that become crucial for maintaining communication in an otherwise disrupted network. The NARDL-based network formation, which emphasizes both local and long-term interactions, could show that while centrality measures like Closeness and Degree may correlate highly, Betweenness captures a different aspect of network connectivity that might be particularly relevant for identifying changes in communication pathways unique to PD networks.
4.2. Deep Learning Perspective
In the following subsection section, we present the results of our deep learning analyses focused on understanding feature importance in distinguishing Parkinson’s disease patients from healthy controls. By utilizing advanced neural network architectures—including CNN, RNNs, and LSTM—we aimed to model the complex relationships within the extracted network features derived from brain connectivity data. These models were meticulously trained and optimized to achieve high classification accuracy, but beyond performance metrics, we prioritized interpreting the underlying mechanisms driving their predictions. Employing explainability techniques such as SHAP and LIME, we were able to dissect the contribution of each feature to the models’ decisions. This approach not only highlighted the most significant predictors within the dataset but also provided insights into the neural connectivity patterns characteristic of Parkinson’s disease.
Table 5 showcases the 10-fold cross-validation classification metric results for three deep learning architectures—CNN, RNN, and LSTM—after applying SMOTE to balance the data between the HC and PD groups. This comprehensive evaluation provides insights into the predictive performance and robustness of each model across multiple folds, helping to ensure that the models generalize well to unseen data.
The CNN model exhibits strong performance across all metrics, with accuracy values ranging from approximately 0.8574 to 0.9523. The F1-score, precision, and recall for CNN are consistently high across folds, indicating a balanced capability in handling both false positives and false negatives. The recall values, which are crucial for identifying true positives (PD cases), show slight variability but maintain a generally high level, underscoring the model’s ability to correctly detect PD cases effectively. The overall stability of the CNN’s results across the folds suggests that this architecture can capture the spatial hierarchies and dependencies within the network features derived from the NARDL-based brain connectivity data.
The RNN model also performs well, with accuracy scores between 0.8145 and 0.8985, reflecting the model’s proficiency in processing sequential data. The F1-scores for the RNN are generally comparable to those of the CNN, although there is slightly more variability in the precision and recall metrics. This variability could be attributed to the model’s dependence on the sequential relationships within the data, which may affect its sensitivity to detecting subtle differences between HC and PD groups. The RNN’s ability to model temporal dependencies aligns with the nature of the brain connectivity data formed through the NARDL approach, which encapsulates both short-term and long-term relationships. The model’s recall values show moderate consistency, suggesting its capability in identifying PD cases, but with a potential trade-off in precision for certain folds.
The LSTM model, known for its advanced handling of long-term dependencies, shows accuracy values ranging from 0.7351 to 0.8991. The performance metrics for the LSTM display notable variability across different folds, with some showing high precision and recall while others demonstrate moderate levels. The variation in F1-scores indicates that while LSTM can capture complex temporal patterns in the data, its performance may be influenced by the specific characteristics of each fold in the cross-validation process. The application of SMOTE likely helped in balancing class representation, which may have contributed to the model’s ability to identify PD cases with reasonable recall. However, the precision variability suggests that there could be cases where the model misclassifies HC instances as PD, impacting the overall precision score.
Aggregated confusion matrices are given in
Figure 6.
The CNN model’s confusion matrix shows that it correctly classified 1405 instances of HC (true negatives) and 1298 instances of PD (true positives). However, there were 132 false positives and 239 false negatives. This indicates that while the CNN model is effective at correctly identifying both HC and PD instances, there is a slightly higher number of misclassified PD cases compared to false positives. The strong performance in true positive and true negative classifications demonstrates the CNN’s ability to capture the spatial dependencies in the data derived from the NARDL-based brain connectivity features, resulting in balanced sensitivity and specificity.
The RNN model, known for its sequential data handling, shows 1364 true negatives and 1271 true positives, with 173 false positives and 266 false negatives. Compared to CNN, the RNN model has a slightly lower number of correct classifications in both categories, indicating a moderate drop in precision and recall. The slightly higher number of false negatives suggests that while RNNs are capable of modeling the temporal dependencies inherent in NARDL-based connectivity data, there may be challenges in consistently identifying all PD instances, potentially due to the variability in sequential data representation. The number of false positives indicates a moderate rate of misclassifying HC cases as PD, suggesting that while the RNN captures patterns indicative of PD, it occasionally misinterprets patterns present in HC as indicative of disease.
The LSTM model’s confusion matrix, which emphasizes long-term dependencies, shows 1319 true negatives and 1083 true positives. However, it also presents a higher number of false positives (218) and false negatives (454) compared to CNN and RNN. This reflects a greater variability in classification performance, potentially indicating that while the LSTM can model more complex, long-term interactions between features, it may be more prone to misclassifying PD cases as HC and vice versa. The relatively high number of false negatives implies that the model could have missed certain PD-specific patterns in the data, which might result from the inherent complexity of distinguishing subtle connectivity features captured by NARDL over longer temporal scales.
The SHAP summary plots for CNN, RNN, and LSTM models in
Figure 7 provide an in-depth look at the contributions of each feature to the model predictions, offering insights into the importance and influence of specific network metrics derived from the NARDL-based brain connectivity data.
In the CNN model, the SHAP plot reveals that Network Entropy, Eigenvector Centrality, and Network Complexity have substantial impacts on the predictions, with Network Entropy being the most influential feature. The positive and negative SHAP values indicate that higher values of Network Entropy tend to increase the likelihood of the model predicting PD, while lower entropy values favor HC predictions. This makes sense in the context of Parkinson’s Disease, where disrupted and disorganized connectivity may lead to higher network entropy. Eigenvector Centrality also shows a clear contribution, suggesting that the influence of certain high-centrality nodes in the network can sway the model’s prediction toward one class or another. The moderate influence of Betweenness and Degree Centrality suggests that while these features are relevant, their impact is secondary compared to entropy and centrality measures that capture the prominence of specific nodes within the network. Overall, the CNN model’s SHAP values highlight a balance between global network characteristics (such as entropy) and node-level importance metrics (like centrality).
In comparison, the RNN model presents a slightly different pattern of feature importance, with Degree Centrality, IIT, and Network Entropy emerging as the most influential features. Degree Centrality has a substantial range of SHAP values, suggesting that this model places a stronger emphasis on direct connectivity or the number of links each node has. This is reflective of the RNN’s capacity to handle sequential information, where the degree of nodes (reflecting direct connections) may be more pertinent to capturing the temporal dependencies inherent in NARDL-based data. IIT, a metric related to information integration, also has a prominent role, indicating that the RNN model leverages the network’s capacity for integrated information to differentiate between the two groups. This reliance on IIT reflects the RNN’s strength in modeling interconnected dependencies, and it may suggest that PD networks show altered information integration patterns compared to HC networks. Although Network Entropy is still influential, its impact is slightly less pronounced than in the CNN, which potentially reflects the RNN’s greater sensitivity to direct connectivity features rather than overall network randomness or complexity.
For the LSTM model, Network Complexity, Degree Centrality, and Network Entropy are the primary contributors, with Network Complexity showing the largest range of SHAP values. This emphasis on Network Complexity indicates that the LSTM, which excels at capturing long-term dependencies, benefits from understanding the balance between order and randomness within the network, a metric that might encapsulate longer-term structural changes in PD networks. Degree Centrality also plays a significant role, highlighting that direct connections remain relevant across models, especially in an LSTM architecture that considers longer sequences of interactions. Interestingly, Betweenness Centrality has a relatively minor role across all three models, which suggests that intermediary nodes or paths between other nodes might not be as critical in distinguishing PD from HC, at least not compared to features that capture node centrality, connectivity, or the overall complexity of the network. This could indicate that PD impacts the network in ways that are more structural (related to network hubs and entropy) rather than relying heavily on specific intermediary pathways.
Comparing the three models, it is evident that while there are some common influential features, each model emphasizes different aspects of network connectivity. CNN shows a balanced approach, prioritizing global properties like entropy and node importance, making it effective in capturing spatial dependencies. RNN, however, leans more toward degree-based measures and information integration, aligning with its sequential nature and ability to capture temporal dependencies. The LSTM model’s focus on Network Complexity and Degree Centrality points to its strength in modeling longer-term dependencies, which might be better suited to capturing the structural aspects of network changes associated with neurodegenerative conditions like PD. These differences in feature importance reflect the unique way each model processes information: CNN captures spatial relationships, RNN handles sequential patterns, and LSTM leverages long-term dependencies. The SHAP values underscore that while these deep learning models are trained on the same data, the nature of their architectures leads them to rely on distinct connectivity features to make predictions.
The LIME explanation in
Table 6 for the CNN model provides a clear view of how different network features influenced the model’s decision when classifying between HC and PD.
Betweenness Centrality, with a value of −0.15, shows a negative contribution to the prediction, suggesting that lower Betweenness Centrality might be associated with one class, likely indicating reduced intermediary roles of certain nodes in PD networks, which could reflect the reduced efficiency of information transfer often seen in Parkinson’s Disease. Network Entropy, on the other hand, has a high positive value (0.52) and contributes positively to the model’s output, aligning with the idea that increased entropy, or randomness in network connectivity, is indicative of the disorganization seen in neurodegenerative diseases. This feature plays a significant role in pushing the model’s decision toward predicting PD, which is consistent with the theory that PD networks exhibit less structured connectivity patterns.
Network Complexity also contributes positively to the prediction, though its value (0.32) is lower than Network Entropy. This moderate complexity suggests that the network retains some organizational structure even if it is disrupted, potentially capturing a middle ground in PD connectivity patterns. Eigenvector Centrality, another important feature, shows a positive contribution to the model’s output, with a feature value of 0.32. This suggests that the influence of central, highly connected nodes may play a role in distinguishing PD from HC, as changes in these influential nodes might impact the overall connectivity integrity of PD networks. Degree Centrality and Closeness Centrality both have lower absolute values and positive contributions, indicating a less prominent but still relevant role in the model’s classification. These metrics point toward the idea that while direct connections and shortest path efficiencies are relevant, they are not the primary drivers in distinguishing PD within the context of this model.
Lastly, ITT has a minimal contribution with a feature value of 0.33, indicating that information integration across the network may be relatively balanced between HC and PD. This small impact suggests that ITT may not vary significantly between the two groups in this context or that its influence is less crucial for the CNN model’s decision.
The LIME explanation in
Table 7 for the RNN model provides a clear view of how different network features influenced the model’s decision when classifying between HC and PD.
Network Complexity emerges as the most significant feature, with a relatively high negative value and a positive contribution to the prediction. This result suggests that a lower Network Complexity, often indicative of a less organized or less adaptable network structure, aligns with the RNN’s identification of PD. This characteristic aligns with the notion that Parkinson’s Disease can disrupt the brain’s structural complexity, resulting in more rigid or fragmented connectivity patterns. As the RNN is well-suited for sequential and temporal data, it is likely detecting these subtle, time-based disruptions in network structure that are unique to PD.
Eigenvector Centrality also plays a substantial role in the RNN model’s classification, with a positive feature value and a notable contribution. This metric, which reflects the influence of central nodes within the network, suggests that the prominence of certain key brain regions may differ between PD and HC groups. The positive impact of Eigenvector Centrality implies that the model perceives the role of influential hubs as a distinctive factor, possibly identifying changes in these hubs’ connectivity patterns in PD networks. Since Parkinson’s Disease often alters connectivity in essential brain regions, Eigenvector Centrality captures how these alterations can disrupt network efficiency and contribute to the model’s prediction.
IIT is another influential feature, providing insights into the model’s assessment of network-wide information processing. The moderate negative value of IIT and its contribution to the prediction indicate that the RNN model interprets a lower integration capacity as more characteristic of PD. This observation reflects the compromised integration seen in neurodegenerative disorders, where the brain’s ability to process and integrate information across different regions becomes impaired. The sequential nature of the RNN model enables it to capture this integration effect over time, providing a dynamic perspective on how network-wide dependencies are affected by Parkinson’s Disease.
Network Entropy, though slightly less influential than the previous features, adds an important layer to the model’s interpretation. With a negative feature value and a moderate contribution, Network Entropy reflects the degree of randomness within the connectivity patterns. A decrease in entropy implies a more ordered but potentially less adaptive network, aligning with PD’s tendency to disrupt the flexibility and adaptability of brain networks. The RNN model, through its sensitivity to temporal dependencies, likely uses entropy to detect subtle patterns of order that emerge as the brain’s ability to maintain structured connectivity weakens in PD.
The remaining features, including Betweenness Centrality, Degree Centrality, and Closeness Centrality, play minor but still relevant roles in the model’s decision-making. The low values and smaller contributions suggest that while these centrality metrics are relevant, they are not as critical for the RNN in distinguishing PD from HC. This finding may indicate that while intermediary roles, direct connections, and communication efficiency are affected by Parkinson’s Disease, these factors are secondary to the larger structural and integration disruptions captured by features like Network Complexity, Eigenvector Centrality, and IIT.
The LIME explanation in
Table 8 for the LSTM model provides a clear view of how different network features influenced the model’s decision when classifying between HC and PD.
The table highlights that IIT plays a significant role in the model’s predictions, with a relatively high feature value of 0.26 and a substantial positive contribution of 0.08. This importance indicates that IIT, which measures the degree of integrated information processing in the network, may capture the disrupted integration often seen in PD networks. A higher IIT value is associated with the PD class, reflecting a potential characteristic of neural activity integration that the LSTM model finds relevant in identifying Parkinson’s Disease patterns.
Network Complexity, with a feature value of −0.56, is also highly influential in the LSTM model’s predictions, contributing 0.06 toward the model’s output. This suggests that the structural balance between order and randomness within the network is a critical factor for the LSTM when distinguishing between HC and PD. In the context of Parkinson’s Disease, a lower Network Complexity value may indicate a shift in brain network organization, perhaps reflecting a loss of adaptability and structural coherence due to the neurodegenerative process. This disruption is consistent with the characteristic changes in connectivity patterns in PD, where certain networks may become more rigid or disorganized, impacting overall complexity.
Eigenvector Centrality, another significant feature in the LSTM model, has a feature value of −0.64 and contributes 0.05 to the prediction. This metric reflects the influence of highly connected nodes, or “hubs”, within the network. Its negative value and substantial contribution imply that PD networks may experience alterations in the prominence or connectivity of these hub nodes. Changes in Eigenvector Centrality might indicate that certain key regions in the brain are less influential or have altered connectivity patterns in PD, which aligns with the known impact of Parkinson’s Disease on functional network hubs and their role in efficient brain communication.
Network Entropy, with a positive feature value of 0.74 and a moderate contribution of 0.04, also plays an important role in the model’s prediction. Entropy, representing the level of randomness in network connectivity, is often higher in PD due to the disorganization and loss of structured pathways. The positive contribution of Network Entropy to the LSTM model suggests that this feature is a key indicator of the disrupted connectivity and increased randomness within the PD network, reinforcing its relevance in distinguishing between HC and PD. Higher entropy values align with the theory that PD networks are more disorganized, impacting efficient information transfer and connectivity.
Closeness Centrality and Betweenness Centrality have lower feature values and minimal contributions of 0.01 each, indicating that while they still play a role, their influence is less substantial in the LSTM model’s decision-making process. Closeness Centrality, with a slightly negative feature value of −0.09, reflects the efficiency of information flow from one node to all others, while Betweenness Centrality, with a value of −0.11, captures the role of nodes as intermediaries in the network. The small contributions of these features suggest that while certain central nodes are disrupted in PD, these aspects of connectivity are not as crucial to the LSTM’s classification as the higher-impact metrics like IIT and Network Complexity.
Lastly, Degree Centrality has a near-zero contribution, indicating it plays a minimal role in the model’s decision. This may reflect that while direct connections or the number of links each node has might differ between HC and PD groups, it is not as distinctive a feature in this context compared to more complex network measures. Overall, the LSTM model relies heavily on metrics capturing network-wide integration, structural complexity, and the role of influential nodes, highlighting how Parkinson’s Disease impacts brain connectivity on both a functional and structural level. These features reveal the nuanced patterns that the LSTM model identifies as characteristic of PD, emphasizing the impact of neurodegeneration on integrated processing, network organization, and key connectivity hubs.
Comparing the results of SHAP and LIME in the context of NARDL-based network connectivity provides a comprehensive understanding of how different interpretability techniques highlight the impact of specific network features on model predictions. Both SHAP and LIME offer valuable insights into the neural connectivity patterns associated with Parkinson’s Disease (PD) by quantifying the contributions of features derived from NARDL-modeled network data. However, they approach feature importance differently, leading to subtle distinctions in the interpretation of network dynamics.
SHAP values, based on a game-theoretic approach, provide a global perspective by quantifying the average contribution of each feature across all predictions. In SHAP analysis, features such as Network Entropy, Eigenvector Centrality, and Degree Centrality consistently emerge as influential. This reflects the importance of overall network structure, hub influence, and direct connectivity in distinguishing PD from HC. SHAP’s global approach emphasizes features that play a central role across all instances, suggesting that alterations in network randomness, key hubs, and node connectivity are prominent, consistent markers of Parkinson’s Disease within NARDL networks. This interpretation aligns with the structural disruptions and reduced adaptability in connectivity often observed in neurodegenerative conditions, as SHAP highlights how these large-scale changes contribute to the model’s understanding of PD across the dataset.
In contrast, LIME focuses on local interpretability by evaluating feature contributions on a per-instance basis, revealing the influence of features within specific predictions. LIME results often highlight the same set of network features as SHAP, such as Network Complexity, IIT, and Eigenvector Centrality, but with an emphasis on how these features vary across individual instances. This approach uncovers variations in feature importance for each prediction, showing how the model uses network metrics like complexity and centrality in specific cases to differentiate PD from HC. By examining each instance separately, LIME provides insights into how the model responds to variations in network dynamics, capturing the heterogeneity of neural connectivity patterns within the PD group. LIME’s instance-level perspective is particularly valuable for understanding case-specific network changes, offering a closer look at how the model interprets specific shifts in complexity, integration, and centrality that characterize Parkinson’s Disease on an individualized basis.
Together, SHAP and LIME offer complementary perspectives on the interpretability of deep learning models trained on NARDL-based networks. SHAP emphasizes consistent, dataset-wide patterns in network features, highlighting the structural disruptions globally associated with PD. LIME, on the other hand, captures localized variations and provides an individualized view of feature importance, illustrating how the model adapts to specific connectivity changes in each prediction. Combining SHAP’s global interpretation with LIME’s instance-specific analysis creates a robust understanding of how NARDL-based network features influence the model’s predictions, illuminating the distinct ways Parkinson’s Disease impacts neural connectivity and how these changes manifest in deep learning classifications. The integration of both methods enriches the interpretability of network-based models, allowing researchers to appreciate the full scope of connectivity disruptions in PD—both as consistent markers across patients and as variable factors unique to individual cases.