[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (12)

Search Parameters:
Keywords = Davies–Bouldin Index (DBI)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 7839 KiB  
Article
Normalized Difference Vegetation Index Prediction for Blueberry Plant Health from RGB Images: A Clustering and Deep Learning Approach
by A. G. M. Zaman, Kallol Roy and Jüri Olt
AgriEngineering 2024, 6(4), 4831-4850; https://doi.org/10.3390/agriengineering6040276 - 16 Dec 2024
Viewed by 571
Abstract
In precision agriculture (PA), monitoring individual plant health is crucial for optimizing yields and minimizing resources. The normalized difference vegetation index (NDVI), a widely used health indicator, typically relies on expensive multispectral cameras. This study introduces a method for predicting the NDVI of [...] Read more.
In precision agriculture (PA), monitoring individual plant health is crucial for optimizing yields and minimizing resources. The normalized difference vegetation index (NDVI), a widely used health indicator, typically relies on expensive multispectral cameras. This study introduces a method for predicting the NDVI of blueberry plants using RGB images and deep learning, offering a cost-effective alternative. To identify individual plant bushes, K-means and Gaussian Mixture Model (GMM) clustering were applied. RGB images were transformed into the HSL (hue, saturation, lightness) color space, and the hue channel was constrained using percentiles to exclude extreme values while preserving relevant plant hues. Further refinement was achieved through adaptive pixel-to-pixel distance filtering combined with the Davies–Bouldin Index (DBI) to eliminate pixels deviating from the compact cluster structure. This enhanced clustering accuracy and enabled precise NDVI calculations. A convolutional neural network (CNN) was trained and tested to predict NDVI-based health indices. The model achieved strong performance with mean squared losses of 0.0074, 0.0044, and 0.0021 for training, validation, and test datasets, respectively. The test dataset also yielded a mean absolute error of 0.0369 and a mean percentage error of 4.5851. These results demonstrate the NDVI prediction method’s potential for cost-effective, real-time plant health assessment, particularly in agrobotics. Full article
Show Figures

Figure 1

Figure 1
<p>Prototype agrobotic system for precision fertilization.</p>
Full article ">Figure 2
<p>Technical methods for NDVI prediction from RGB image.</p>
Full article ">Figure 3
<p>Data collection process and sample blueberry image for NDVI analysis: (<b>a</b>) On-field data collection process; (<b>b</b>) Processed blueberry image.</p>
Full article ">Figure 4
<p>Blueberry plant images (500 × 500) and corresponding RGB pixel distributions: (<b>a</b>) Large blueberry plant with RGB distribution shown in (<b>d</b>); (<b>b</b>) Medium blueberry plant with RGB distribution in (<b>e</b>); (<b>c</b>) Small blueberry plant with RGB distribution in (<b>f</b>).</p>
Full article ">Figure 4 Cont.
<p>Blueberry plant images (500 × 500) and corresponding RGB pixel distributions: (<b>a</b>) Large blueberry plant with RGB distribution shown in (<b>d</b>); (<b>b</b>) Medium blueberry plant with RGB distribution in (<b>e</b>); (<b>c</b>) Small blueberry plant with RGB distribution in (<b>f</b>).</p>
Full article ">Figure 5
<p>K-means and GMM clustering of a sample blueberry plant image. (<b>a</b>) Blueberry plant sample. (<b>b</b>) K-means. (<b>c</b>) GMM.</p>
Full article ">Figure 6
<p>Visualization of K-means and GMM clustering for various plant sizes following adaptive hue-based filtering in HSL space: (<b>a</b>,<b>d</b>) show K-means and GMM clustering for the plant in <a href="#agriengineering-06-00276-f004" class="html-fig">Figure 4</a>a; (<b>b</b>,<b>e</b>) depict K-means and GMM clustering for the plant in <a href="#agriengineering-06-00276-f004" class="html-fig">Figure 4</a>b; (<b>c</b>,<b>f</b>) illustrate K-means and GMM clustering for the plant in <a href="#agriengineering-06-00276-f004" class="html-fig">Figure 4</a>c.</p>
Full article ">Figure 7
<p>Identification of plant clusters (K-means) across different plant sizes using adaptive pixel-to-pixel average distance-based percentile filtering based on DBI index, with corresponding images shown in <a href="#agriengineering-06-00276-f006" class="html-fig">Figure 6</a>a–c: (<b>a</b>–<b>c</b>) display histogram distributions for large, medium, and small plant sizes, respectively, with the adaptive percentile cut-off points highlighted; (<b>d</b>–<b>f</b>) show the filtered plant clusters after thresholding; (<b>g</b>–<b>i</b>) illustrate the bounding boxes outlining the extracted clusters within the original images.</p>
Full article ">Figure 8
<p>CNN model architecture for NDVI regression (input: 350 × 350 RGB image).</p>
Full article ">Figure 9
<p>Impact of HSL transformation and adaptive hue-based filtering on DBI and CHI values in K-means and GMM clustering—before and after processing.</p>
Full article ">Figure 10
<p>Progressive refinement in plant bush identification, demonstrating the impact of successive filtering stages on K-means clustering quality: (<b>a</b>) original image; (<b>b</b>) plant bush cluster obtained following HSL transformation and adaptive hue filtering; (<b>c</b>) refined plant bush cluster with non-plant outlier removal through adaptive pixel-to-pixel average distance filtering; (<b>d</b>) final plant cluster with the bounding box on top of the original image.</p>
Full article ">Figure 11
<p>CNN model training and validation loss curves.</p>
Full article ">Figure 12
<p>Model performance evaluation on test data: (<b>a</b>) Bar chart comparing actual and predicted NDVI values for test data points; (<b>b</b>) Percentage prediction error per test data point.</p>
Full article ">
38 pages, 8511 KiB  
Article
Robust Parameter Optimisation of Noise-Tolerant Clustering for DENCLUE Using Differential Evolution
by Omer Ajmal, Humaira Arshad, Muhammad Asad Arshed, Saeed Ahmed and Shahzad Mumtaz
Mathematics 2024, 12(21), 3367; https://doi.org/10.3390/math12213367 - 27 Oct 2024
Viewed by 842
Abstract
Clustering samples based on similarity remains a significant challenge, especially when the goal is to accurately capture the underlying data clusters of complex arbitrary shapes. Existing density-based clustering techniques are known to be best suited for capturing arbitrarily shaped clusters. However, a key [...] Read more.
Clustering samples based on similarity remains a significant challenge, especially when the goal is to accurately capture the underlying data clusters of complex arbitrary shapes. Existing density-based clustering techniques are known to be best suited for capturing arbitrarily shaped clusters. However, a key limitation of these methods is the difficulty in automatically finding the optimal set of parameters adapted to dataset characteristics, which becomes even more challenging when the data contain inherent noise. In our recent work, we proposed a Differential Evolution-based DENsity CLUstEring (DE-DENCLUE) to optimise DENCLUE parameters. This study evaluates DE-DENCLUE for its robustness in finding accurate clusters in the presence of noise in the data. DE-DENCLUE performance is compared against three other density-based clustering algorithms—DPC based on weighted local density sequence and nearest neighbour assignment (DPCSA), Density-Based Spatial Clustering of Applications with Noise (DBSCAN), and Variable Kernel Density Estimation–based DENCLUE (VDENCLUE)—across several datasets (i.e., synthetic and real). The study has consistently shown superior results for DE-DENCLUE compared to other models for most datasets with different noise levels. Clustering quality metrics such as the Silhouette Index (SI), Davies–Bouldin Index (DBI), Adjusted Rand Index (ARI), and Adjusted Mutual Information (AMI) consistently show superior SI, ARI, and AMI values across most datasets at different noise levels. However, in some cases regarding DBI, the DPCSA performed better. In conclusion, the proposed method offers a reliable and noise-resilient clustering solution for complex datasets. Full article
(This article belongs to the Special Issue Optimization Models and Algorithms in Data Science)
Show Figures

Figure 1

Figure 1
<p>Architecture diagram for evaluating noise-tolerant clustering methods.</p>
Full article ">Figure 2
<p>Proposed method (DE-based DENCLUE parameter estimation).</p>
Full article ">Figure 3
<p>Metric variability across noise levels for synthetic datasets: (<b>a</b>) Aggregation; (<b>b</b>) Two Moons; (<b>c</b>) Path-based; (<b>d</b>) Shapes; (<b>e</b>) Spiral; and (<b>f</b>) Zahn’s Compound.</p>
Full article ">Figure 4
<p>Metric variability across noise levels for real datasets: (<b>a</b>) IRIS; (<b>b</b>) Heart Disease; (<b>c</b>) Seeds; and (<b>d</b>) Wine.</p>
Full article ">Figure 5
<p>Comparison of clustering results with the original (<b>left</b>) and two runs (<b>middle</b> and <b>right</b>). Each cluster is shown with unique color and symbol. Although the run in the middle has a higher ARI, it compromises in DBCV, Silhouette Index, and True Clusters, demonstrating the trade-offs between different clustering validation criteria.</p>
Full article ">Figure A1
<p>Comparison of different DE control parameters—synthetic datasets. (<b>a</b>) Aggregation; (<b>b</b>) Two Moons; (<b>c</b>) Path-based; (<b>d</b>) Shapes; (<b>e</b>) Spiral; (<b>f</b>) Zahn’s Compound; (<b>g</b>) S; (<b>h</b>) A3. The x-axis in the subplots indicates specific combinations of DE control parameters (F and CR), where the first value in the pair represents F, and the second value represents CR. The y-axis indicates the mean generations to reach the threshold objective value.</p>
Full article ">Figure A2
<p>Comparison of different DE control parameters—real datasets. (<b>a</b>) IRIS; (<b>b</b>) Heart Disease; (<b>c</b>) Seeds; (<b>d</b>) Wine. The x-axis in the subplots indicates specific combinations of DE control parameters (F and CR), where the first value in the pair represents F, and the second value represents CR. The y-axis indicates the mean generations to reach the threshold objective value.</p>
Full article ">Figure A3
<p>Distribution of metric values for Aggregation dataset across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A4
<p>Distribution of metric values for Two Moons dataset across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A5
<p>Distribution of metric values for Path-based dataset across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A6
<p>Distribution of metric values for Shapes dataset across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A7
<p>Distribution of metric values for Spiral dataset across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A8
<p>Distribution of metric values for Zahn’s Compound dataset across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A9
<p>Distribution of metric values—IRIS across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A10
<p>Distribution of metric values—Heart Disease across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A11
<p>Distribution of metric values—Seeds across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A12
<p>Distribution of metric values—Wine across noise levels. The outliers are depicted as isolated points.</p>
Full article ">Figure A13
<p>Benchmark vs. proposed method clustering heatmaps for Aggregation dataset.</p>
Full article ">Figure A14
<p>Benchmark vs. proposed method clustering heatmaps for Two Moons dataset.</p>
Full article ">Figure A15
<p>Benchmark vs. proposed method clustering heatmaps for Path-based dataset.</p>
Full article ">Figure A16
<p>Benchmark vs. proposed method clustering heatmaps for Shapes dataset.</p>
Full article ">Figure A17
<p>Benchmark vs. proposed method clustering heatmaps for Spiral dataset.</p>
Full article ">Figure A18
<p>Benchmark vs. proposed method clustering heatmaps for Zahn’s Compound dataset.</p>
Full article ">Figure A19
<p>Benchmark vs. proposed method clustering heatmaps for Iris dataset.</p>
Full article ">Figure A20
<p>Benchmark vs. proposed method clustering heatmaps for Heart Disease dataset.</p>
Full article ">Figure A21
<p>Benchmark vs. proposed method clustering heatmaps for Seeds dataset.</p>
Full article ">Figure A22
<p>Benchmark vs. proposed method clustering heatmaps for Wine dataset.</p>
Full article ">
15 pages, 4416 KiB  
Article
Optimization of Temperature Modulation for Gas Classification Based on Bayesian Optimization
by Tatsuya Iwata, Yuki Okura, Maaki Saeki and Takefumi Yoshikawa
Sensors 2024, 24(9), 2941; https://doi.org/10.3390/s24092941 - 6 May 2024
Viewed by 2959
Abstract
This study proposes an optimization method for temperature modulation in chemiresistor-type gas sensors based on Bayesian optimization (BO), and its applicability was investigated. As voltage for a sensor heater, our previously proposed waveform was employed, and the parameters determining the voltage range were [...] Read more.
This study proposes an optimization method for temperature modulation in chemiresistor-type gas sensors based on Bayesian optimization (BO), and its applicability was investigated. As voltage for a sensor heater, our previously proposed waveform was employed, and the parameters determining the voltage range were optimized. Employing the Bouldin–Davies index (DBI) as an objective function (OBJ), BO was utilized to minimize the DBI calculated from a feature matrix built from the collected data followed by pre-processing. The sensor responses were measured using five test gases with five concentrations, amounting to 2500 data points per parameter set. After seven trials with four initial parameter sets (ten parameter sets were tested in total), the DBI was successfully reduced from 2.1 to 1.5. The classification accuracy for the test gases based on the support vector machine tends to increase with decreasing the DBI, indicating that the DBI acts as a good OBJ. Additionally, the accuracy itself increased from 85.4% to 93.2% through optimization. The deviation from the tendency that the accuracy increases with decreasing the DBI for some parameter sets was also discussed. Consequently, it was demonstrated that the proposed optimization method based on BO is promising for temperature modulation. Full article
(This article belongs to the Special Issue Recent Advancements in Olfaction and Electronic Nose)
Show Figures

Figure 1

Figure 1
<p><math display="inline"><semantics> <msub> <mi>V</mi> <mi mathvariant="normal">H</mi> </msub> </semantics></math> waveform employed in this study. <math display="inline"><semantics> <msub> <mi>V</mi> <mn>0</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>V</mi> <mi>Offset</mi> </msub> </semantics></math> determine the peak and bottom of the wave, and hence, the range of <math display="inline"><semantics> <msub> <mi>V</mi> <mi mathvariant="normal">H</mi> </msub> </semantics></math>.</p>
Full article ">Figure 2
<p>Schematic illustration of the concept of Bayesian optimization.</p>
Full article ">Figure 3
<p>Schematic illustration of the optimization procedure based on Bayesian optimization.</p>
Full article ">Figure 4
<p>Schematic illustration of the flow-control system.</p>
Full article ">Figure 5
<p>Schematic electrical circuit for the measurements. The current for heater resistor (<math display="inline"><semantics> <msub> <mi>R</mi> <mi mathvariant="normal">H</mi> </msub> </semantics></math>) was amplified by a voltage follower, while the sensor conductance (<math display="inline"><semantics> <mrow> <msub> <mi>G</mi> <mi mathvariant="normal">S</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>/</mo> <msub> <mi>R</mi> <mi mathvariant="normal">S</mi> </msub> </mrow> </semantics></math>) was converted to the output voltage (<math display="inline"><semantics> <msub> <mi>V</mi> <mi>Out</mi> </msub> </semantics></math>) by an inverting amplifier.</p>
Full article ">Figure 6
<p>(<b>a</b>) Heater voltage with one of the initial parameter sets (<math display="inline"><semantics> <msub> <mi>V</mi> <mn>0</mn> </msub> </semantics></math>: 0.35 V, <math display="inline"><semantics> <msub> <mi>V</mi> <mi>Offset</mi> </msub> </semantics></math>: 0.75 V) and one of the corresponding measurement results for each of the test gases: (<b>b</b>) <math display="inline"><semantics> <msub> <mi>G</mi> <mi mathvariant="normal">S</mi> </msub> </semantics></math>, (<b>c</b>) <math display="inline"><semantics> <msub> <mi>G</mi> <mrow> <mi mathvariant="normal">S</mi> <mo>,</mo> <mi mathvariant="normal">n</mi> </mrow> </msub> </semantics></math>, and (<b>d</b>) frequency spectra of <math display="inline"><semantics> <msub> <mi>G</mi> <mrow> <mi mathvariant="normal">S</mi> <mo>,</mo> <mi mathvariant="normal">n</mi> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 7
<p>PC plot of the data obtained using the heater voltage shown in <a href="#sensors-24-02941-f006" class="html-fig">Figure 6</a>a.</p>
Full article ">Figure 8
<p>GPR results after (<b>a</b>) first, (<b>b</b>) third, (<b>c</b>), fifth, and (<b>d</b>) seventh trials. The blue circles indicate the predicted mean obtained by the regression, while the red crosses the experimental results.</p>
Full article ">Figure 9
<p>The observed minimum DBI plotted as a function of the number of trials.</p>
Full article ">Figure 10
<p>Classification accuracy plotted as a function of the DBI. The markers and the error bars indicate mean and max./min. accuracies, respectively.</p>
Full article ">
31 pages, 13580 KiB  
Article
Multi-Dimensional Validation of the Integration of Syntactic and Semantic Distance Measures for Clustering Fibromyalgia Patients in the Rheumatic Monitor Big Data Study
by Ayelet Goldstein, Yuval Shahar, Michal Weisman Raymond, Hagit Peleg, Eldad Ben-Chetrit, Arie Ben-Yehuda, Erez Shalom, Chen Goldstein, Shmuel Shay Shiloh and Galit Almoznino
Bioengineering 2024, 11(1), 97; https://doi.org/10.3390/bioengineering11010097 - 19 Jan 2024
Viewed by 2516
Abstract
This study primarily aimed at developing a novel multi-dimensional methodology to discover and validate the optimal number of clusters. The secondary objective was to deploy it for the task of clustering fibromyalgia patients. We present a comprehensive methodology that includes the use of [...] Read more.
This study primarily aimed at developing a novel multi-dimensional methodology to discover and validate the optimal number of clusters. The secondary objective was to deploy it for the task of clustering fibromyalgia patients. We present a comprehensive methodology that includes the use of several different clustering algorithms, quality assessment using several syntactic distance measures (the Silhouette Index (SI), Calinski–Harabasz index (CHI), and Davies–Bouldin index (DBI)), stability assessment using the adjusted Rand index (ARI), and the validation of the internal semantic consistency of each clustering option via the performance of multiple clustering iterations after the repeated bagging of the data to select multiple partial data sets. Then, we perform a statistical analysis of the (clinical) semantics of the most stable clustering options using the full data set. Finally, the results are validated through a supervised machine learning (ML) model that classifies the patients back into the discovered clusters and is interpreted by calculating the Shapley additive explanations (SHAP) values of the model. Thus, we refer to our methodology as the clustering, distance measures and iterative statistical and semantic validation (CDI-SSV) methodology. We applied our method to the analysis of a comprehensive data set acquired from 1370 fibromyalgia patients. The results demonstrate that the K-means was highly robust in the syntactic and the internal consistent semantics analysis phases and was therefore followed by a semantic assessment to determine the optimal number of clusters (k), which suggested k = 3 as a more clinically meaningful solution, representing three distinct severity levels. the random forest model validated the results by classification into the discovered clusters with high accuracy (AUC: 0.994; accuracy: 0.946). SHAP analysis emphasized the clinical relevance of "functional problems" in distinguishing the most severe condition. In conclusion, the CDI-SSV methodology offers significant potential for improving the classification of complex patients. Our findings suggest a classification system for different profiles of fibromyalgia patients, which has the potential to improve clinical care, by providing clinical markers for the evidence-based personalized diagnosis, management, and prognosis of fibromyalgia patients. Full article
Show Figures

Figure 1

Figure 1
<p><b>The CDI-SSV Methodology</b>: An Integrated Approach for Clustering Validation. The figure provides an overview of the Clustering, Distance measures, and Iterative Statistical and Semantic Validation (CDI-SSV) methodology. The CDI phase serves as the initial step, involving the evaluation of cluster quality, the impact of different starting seeds, and the consistency of clusters across various algorithms and pre-defined values of k. Within- and between-consistency checks, along with evaluations of internal semantic consistency, are performed to assess the optimal algorithm and values of k. In the subsequent SSV phase, an external semantic analysis of the results is conducted, with a particular focus on the clinical context, thus enhancing the validation process. Finally, machine learning techniques are employed to validate the results, and their interpretation is facilitated by SHAP (Shapley additive explanations) values.</p>
Full article ">Figure 2
<p>Evaluation of clustering algorithms using evaluation metrics. (<b>A</b>) Silhouette index (SI): the SI scores for different values of k indicate that K-means with Gower’s distance metric achieved the highest score for k = 2, 3, and 5. (<b>B</b>) Calinski–Harabasz index (CHI): K-means consistently outperformed other algorithms, achieving the best score across all values of k. (<b>C</b>) Davies–Bouldin index (DBI): K-means demonstrated superior results for all values of k.</p>
Full article ">Figure 3
<p>Demographics and smoking habits across the clusters (k = 3) (likelihood ratio).</p>
Full article ">Figure 4
<p>Comorbidities and history of trauma across the clusters (likelihood ratio).</p>
Full article ">Figure 5
<p>Symptoms, sleep and functional problems and treatment modalities across the clusters (likelihood ratio).</p>
Full article ">Figure 6
<p>Years with fibromyalgia, sleep, quality of life, treatment effectiveness, and pain level and locations (analysis of variance (ANOVA) corrected with Bonferroni test for multiple comparisons).</p>
Full article ">Figure 7
<p>SHAP (Shapley additive explanations) model to predict clusters, k = 3. (<b>A</b>) Bar plot for mean SHAP values of k = 3; (<b>B</b>) dot plot for Cluster 0; (<b>C</b>) dot plot for Cluster 1; (<b>D</b>) dot plot for Cluster 2.</p>
Full article ">Figure 7 Cont.
<p>SHAP (Shapley additive explanations) model to predict clusters, k = 3. (<b>A</b>) Bar plot for mean SHAP values of k = 3; (<b>B</b>) dot plot for Cluster 0; (<b>C</b>) dot plot for Cluster 1; (<b>D</b>) dot plot for Cluster 2.</p>
Full article ">Figure 8
<p><b>Comparative visualization of k = 2 and k = 3 clustering solutions in fibromyalgia patient analysis.</b> The top-left panel displays k = 2 clustering using PCA components, clearly delineating Clusters 0 and 1. The top-right panel presents k = 3 clustering, offering a detailed view of Clusters 0, 1, and 2. The bottom panel includes bar plots that highlight the five most significant attributes for each cluster, with the left side pertaining to the k = 2 solution and the right side pertaining to the k = 3 solution.</p>
Full article ">Figure A1
<p>Visualization of different algorithms for different values of k. Each color represents a different cluster. The figure illustrates the clustering results obtained using different algorithms and values of k. Notably, the K-means and Gaussian clustering methods demonstrate a higher degree of similarity in their cluster assignments compared to the agglomerative clustering method utilizing the Ward linkage criterion. Also, it can be noted that when alternative linkage criteria were employed with the agglomerative clustering method, the majority of data points were assigned to a single cluster. This observation suggests that utilizing linkage criteria other than Ward may lead to less meaningful cluster assignments.</p>
Full article ">Figure A2
<p>Dendrograms of hierarchical clustering using different linkage criteria. Each vertical line represents a merge between clusters. The height of the vertical lines represents the distance (or dissimilarity) at which clusters are merged. The colors signify different clusters formed, based on the standard threshold (70% of the maximum linkage distance).</p>
Full article ">Figure A3
<p>The SI scores of each algorithm in the different values of k. The agglomerative algorithm with complete, average and single linkage had the best scores for every k. The agglomerative algorithm with ward linkage was the worst for almost every k (except k = 5). K-means and Gaussian mixture have very similar scores, both when using Gower’s distance metric and when not using it.</p>
Full article ">Figure A4
<p>The CHI score of K-means was superior for every k, followed by Gaussian mixture and the agglomerative algorithm with ward linkage.</p>
Full article ">Figure A5
<p>On the left, the DBI scores of all algorithms; on the right, after the removal of the three worst (highest) ones. The agglomerative algorithm with complete, average, and single linkage had the lowest (best) results for all ks. K-means performed better than Gaussian mixture did for all ks. Gaussian mixture performed better than did the agglomerative algorithm with Ward linkage for k = 2 and 3 but not for k = 4 and 5.</p>
Full article ">Figure A6
<p>Distribution of SI scores for K-means and Gaussian with and without Gower’s distance metric for the different values of k, sampling 100% of the data, when using different starting seeds.</p>
Full article ">Figure A7
<p>Distribution of CHI scores for K-means and Gaussian mixture with and without Gower’s distance metric for the different values of k, sampling 100% of the data, when using different starting seeds.</p>
Full article ">Figure A8
<p>Distribution of DBI scores for K-means and Gaussian mixture with and without Gower’s distance metric for the different values of k, sampling 100% of the data, when using different starting seeds.</p>
Full article ">Figure A9
<p>Comparison of silhouette index (SI), Calinski–Harabasz index (CHI), and Davies–Bouldin index (DBI) scores using 100% (<b>left</b>) and 70% (<b>right</b>) of the data. The figure shows consistent performance between 70% and 100% of the data, as mean scores did not significantly vary between them.</p>
Full article ">Figure A10
<p>SHAP (SHapley Additive exPlanations) Model to predict clusters for K = 2.</p>
Full article ">Figure A10 Cont.
<p>SHAP (SHapley Additive exPlanations) Model to predict clusters for K = 2.</p>
Full article ">
22 pages, 3828 KiB  
Article
Automation of Cluster Extraction in Fundus Autofluorescence Images of Geographic Atrophy
by Janan Arslan and Kurt Benke
Appl. Biosci. 2023, 2(3), 384-405; https://doi.org/10.3390/applbiosci2030025 - 21 Jul 2023
Cited by 2 | Viewed by 1225
Abstract
The build-up of lipofuscin—an age-associated biomarker referred to as hyperfluorescence—is considered a precursor in the progression of geographic atrophy (GA). Prior studies have attempted to classify hyperfluorescent regions to explain varying rates of GA progression. In this study, digital image processing and unsupervised [...] Read more.
The build-up of lipofuscin—an age-associated biomarker referred to as hyperfluorescence—is considered a precursor in the progression of geographic atrophy (GA). Prior studies have attempted to classify hyperfluorescent regions to explain varying rates of GA progression. In this study, digital image processing and unsupervised learning were used to (1) completely automate the extraction of hyperfluorescent regions from images, and (2) evaluate prospective patterns and groupings of hyperfluorescent areas associated with varying levels of GA progression. Patterns were determined by clustering methods, such as k-Means, and performance was evaluated using metrics such as the Silhouette Coefficient (SC), the Davies–Bouldin Index (DBI), and the Calinski–Harabasz Index (CHI). Automated extraction of hyperfluorescent regions was carried out using pseudocoloring techniques. The approach revealed three distinct types of hyperfluorescence based on color intensity changes: early-stage hyperfluorescence, intermediate-stage hyperfluorescence, and late-stage hyperfluorescence, with the early and late stages having three additional subclassifications that could explain varying levels of GA progression. The performance metrics for early-stage hyperfluorescence were SC = 0.597, DBI = 0.915, and CHI = 186.989. For late-stage hyperfluorescence, SC = 0.593, DBI = 1.013, and CHI = 217.325. No meaningful subclusters were identified for the intermediate-stage hyperfluorescence, possibly because it is a transitional phase of hyperfluorescence progression. Full article
Show Figures

Figure 1

Figure 1
<p>Application of pseudocoloring (before and after pre-processing the image). If a colormap is applied to a FAF image before pre-processing, as shown in (<b>A</b>), the colormap produces a uniformly colored image which lacks differentiation between the features of interest. Pre-processing before pseudocoloring significantly improves image contrast and enhances feature discrimination (<b>B</b>). FAF: Fundus autofluorescence.</p>
Full article ">Figure 2
<p>Separating foreground and background information from FAF images. Image (<b>A</b>) is an example of an original image, while row (<b>B</b>) shows three segmentation outputs using the pseudocoloring segmentation method for hyperfluorescent regions. Row (<b>C</b>) shows the foreground images extracted from the images using the masks in (<b>B</b>). Row (<b>D</b>) shows the background images. FAF: Fundus autofluorescence.</p>
Full article ">Figure 3
<p>Hyperfluorescent segmentation using pseudocoloring. (<b>A</b>) Pre-processed image. (<b>B</b>) Application of the JET colormap, which shows different stages of hyperfluorescence development. (<b>C</b>) Extraction all stages combined. (<b>D</b>) Early-stage hyperfluorescence development. (<b>E</b>) Intermediate-stage hyperfluorescence development. (<b>F</b>) Late-stage hyperfluorescence development. This late stage appears to be the precursor to lesion formation, as intensity levels (i.e., build-up of lipofuscin) reach a peak before retinal cellular death occurs.</p>
Full article ">Figure 4
<p>Scatter and silhouette plots for lesions and hyperfluorescent regions using <span class="html-italic">k</span>-Means clustering. (<b>A</b>) early-stage hyperfluorescence, (<b>B</b>) intermediate-stage hyperfluorescence, and (<b>C</b>) late-stage hyperfluorescence.</p>
Full article ">Figure 5
<p>Early-stage hyperfluorescence shape and pattern clustering using <span class="html-italic">k</span>-Means. (<b>A</b>) Cluster group 1, (<b>B</b>) cluster group 2, and (<b>C</b>) cluster group 3. There appears to be three groups of early-stage hyperfluorescence: one in which early-stage covers the majority of the retina (i.e., early-stage hyperfluorescence complete coverage [EHCC]) (<b>A</b>), another with partial coverage (i.e., early-stage hyperfluorescence partial coverage [EHPaC]) (<b>B</b>), and finally, one that simply surrounds the proximity of the lesions (i.e., early-stage hyperfluorescence proximal coverage [EHPrC]) (<b>C</b>).</p>
Full article ">Figure 6
<p>Intermediate-stage hyperfluorescence shape and pattern clustering using <span class="html-italic">k</span>-Means. (<b>A</b>) Cluster group 1 and (<b>B</b>) Cluster group 2. Unlike the early- and late-stage hyperfluorescence groups, the intermediate-stage grouping is not as prominent.</p>
Full article ">Figure 7
<p>Late-stage hyperfluorescence shape and pattern clustering using <span class="html-italic">k</span>-Means. (<b>A</b>) Cluster group 1, (<b>B</b>) cluster group 2, and (<b>C</b>) cluster group 3. There appears to be three groups of late-stage hyperfluorescence: one in which there is a small scatter of hyperfluorescence regions (i.e., late-stage hyperfluorescence droplet scatter [LHDS])) (<b>A</b>), another combines the scatter with a halo ring around the lesion (i.e., late-stage hyperfluorescence halo scatter [LHHS] (<b>B</b>), and finally, one that is simply a halo around the lesions (i.e., late-stage hyperfluorescence halo [LHH]) (<b>C</b>).</p>
Full article ">
14 pages, 9721 KiB  
Article
Performance Optimization of a Developed Near-Infrared Spectrometer Using Calibration Transfer with a Variety of Transfer Samples for Geographical Origin Identification of Coffee Beans
by Nutthatida Phuangsaijai, Parichat Theanjumpol and Sila Kittiwachana
Molecules 2022, 27(23), 8208; https://doi.org/10.3390/molecules27238208 - 25 Nov 2022
Cited by 7 | Viewed by 2390
Abstract
This research aimed to improve the classification performance of a developed near-infrared (NIR) spectrometer when applied to the geographical origin identification of coffee bean samples. The modification was based on the utilization of a collection of spectral databases from several different agricultural samples, [...] Read more.
This research aimed to improve the classification performance of a developed near-infrared (NIR) spectrometer when applied to the geographical origin identification of coffee bean samples. The modification was based on the utilization of a collection of spectral databases from several different agricultural samples, including corn, red beans, mung beans, black beans, soybeans, green and roasted coffee, adzuki beans, and paddy and white rice. These databases were established using a reference NIR instrument and the piecewise direct standardization (PDS) calibration transfer method. To evaluate the suitability of the transfer samples, the Davies–Bouldin index (DBI) was calculated. The outcomes that resulted in low DBI values were likely to produce better classification rates. The classification of coffee origins was based on the use of a supervised self-organizing map (SSOM). Without the spectral modification, SSOM classification using the developed NIR instrument resulted in predictive ability (% PA), model stability (% MS), and correctly classified instances (% CC) values of 61%, 58%, and 64%, respectively. After the transformation process was completed with the corn, red bean, mung bean, white rice, and green coffee NIR spectral data, the predictive performance of the SSOM models was found to have improved (67–79% CC). The best classification performance was observed with the use of corn, producing improved % PA, % MS, and % CC values at 71%, 67%, and 79%, respectively. Full article
(This article belongs to the Special Issue Development of Chemometrics: Now and Future)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>NIR spectra of the coffee samples obtained from: (<b>a</b>) NIR Foss 6500 and (<b>c</b>) the developed NIR spectrometer. PCA score plots of the NIR spectra obtained from (<b>b</b>) NIR Foss 6500 and (<b>d</b>) the developed NIR spectrometer.</p>
Full article ">Figure 2
<p>NIR spectra (<b>a</b>,<b>c</b>) and corresponding PCA (<b>b</b>,<b>d</b>) of the transfer samples. Data were established from both the reference and the homemade instruments, respectively.</p>
Full article ">Figure 3
<p>NIR spectra of green coffee beans after PDS transformation using green coffee beans (MT), corn, white rice, red beans, and mung beans, and the corresponding PCA score plots compared with the NIR spectra of the transfer samples.</p>
Full article ">Figure 4
<p>NIR spectra of green coffee beans after the PDS transformation using roasted coffee, black beans, paddy rice, soybeans, and azuki beans and the corresponding PCA score plots compared with the NIR spectra of the transfer samples.</p>
Full article ">Figure 5
<p>Correlation between the −log(DBI) and the % CC values. The vertical dotted line indicates the DBI values of the predictive results from the classification model using the homemade NIR instrument, without transformation.</p>
Full article ">Figure 6
<p>Homemade NIR spectrometer; (<b>a</b>) front side view and (<b>b</b>) diagram showing the setting of the NIR sensor.</p>
Full article ">Figure 7
<p>Overview of the experimental procedure of this research. Transformations using green coffee beans (MT) (left) and various agricultural samples (right). <b><span class="html-italic">X</span></b><sub>SL,MT</sub> and <b><span class="html-italic">X</span></b><sub>MS,MT</sub> represent the NIR spectra of the coffee samples (MT), respectively, which were used as the slave and master data for the PDS transformation. <b><span class="html-italic">X</span></b><sub>SL</sub> and <b><span class="html-italic">X</span></b><sub>MS</sub> represent the NIR spectra of the other agricultural samples obtained from the homemade and NIRSystem 6500 spectrometers. <b><span class="html-italic">X</span></b><sub>SK</sub>:<b><span class="html-italic">X</span></b><sub>YP</sub>:<b><span class="html-italic">X</span></b><sub>NK</sub> represent the NIR spectra of the coffee samples (SK, YP, and NK) recorded using the homemade instrument.</p>
Full article ">
15 pages, 2257 KiB  
Article
Comparative Analysis of Three Methods for HYSPLIT Atmospheric Trajectories Clustering
by Likai Cui, Xiaoquan Song and Guoqiang Zhong
Atmosphere 2021, 12(6), 698; https://doi.org/10.3390/atmos12060698 - 30 May 2021
Cited by 28 | Viewed by 6178
Abstract
Using the Hybrid Single-Particle Lagrangian Integrated Trajectory (HYSPLIT) model to obtain backward trajectories and then conduct clustering analysis is a common method to analyze potential sources and transmission paths of atmospheric particulate pollutants. Taking Qingdao (N36 E120) as an example, the global data [...] Read more.
Using the Hybrid Single-Particle Lagrangian Integrated Trajectory (HYSPLIT) model to obtain backward trajectories and then conduct clustering analysis is a common method to analyze potential sources and transmission paths of atmospheric particulate pollutants. Taking Qingdao (N36 E120) as an example, the global data assimilation system (GDAS 1°) of days from 2015 to 2018 provided by National Centers for Environmental Prediction (NCEP) is used to process the backward 72 h trajectory data of 3 arrival heights (10 m, 100 m, 500 m) through the HYSPLIT model with a data interval of 6 h (UTC 0:00, 6:00, 12:00, and 18:00 per day). Three common clustering methods of trajectory data, i.e., K-means, Hierarchical clustering (Hier), and Self-organizing maps (SOM), are used to conduct clustering analysis of trajectory data, and the results are compared with those of the HYSPLIT model released by National Oceanic and Atmospheric Administration (NOAA). Principal Component Analysis (PCA) is used to analyze the original trajectory data. The internal evaluation indexes of Davies–Bouldin Index (DBI), Silhouette Coefficient (SC), Calinski Harabasz Index (CH), and I index are used to quantitatively evaluate the three clustering algorithms. The results show that there is little information in the height data, and thus only two-dimensional plane data are used for clustering. From the results of clustering indexes, the clustering results of SOM and K-means are better than the Hier and HYSPLIT model. In addition, it is found that DBI and I index can help to select the number of clusters, of which DBI is preferred for cluster analysis. Full article
(This article belongs to the Special Issue Air Quality Management)
Show Figures

Figure 1

Figure 1
<p>The structure of SOM.</p>
Full article ">Figure 2
<p>The process of hierarchical clustering.</p>
Full article ">Figure 3
<p>The relationship between adjacent longitude distances and latitude.</p>
Full article ">Figure 4
<p>The clustering results of latitude and longitude coordinates and plane coordinates: (<b>a</b>) Latitude and longitude coordinates; (<b>b</b>) Plane coordinates.</p>
Full article ">Figure 5
<p>Results of four clustering algorithms: (<b>a</b>) HYSPLIT; (<b>b</b>) SOM; (<b>c</b>) Hier; (<b>d</b>) K-means.</p>
Full article ">Figure 6
<p>The result of the different characteristics. (<b>a</b>) HYSPLIT; (<b>b</b>) Hier; (<b>c</b>) Hier (0.1% characteristics loss).</p>
Full article ">Figure 7
<p>The selection of cluster number. (<b>a</b>) DBI; (<b>b</b>) I.</p>
Full article ">
13 pages, 5562 KiB  
Article
Adaptive Graph Convolution Using Heat Kernel for Attributed Graph Clustering
by Danyang Zhu, Shudong Chen, Xiuhui Ma and Rong Du
Appl. Sci. 2020, 10(4), 1473; https://doi.org/10.3390/app10041473 - 21 Feb 2020
Cited by 3 | Viewed by 2998
Abstract
Attributed graphs contain a lot of node features and structural relationships, and how to utilize their inherent information sufficiently to improve graph clustering performance has attracted much attention. Although existing advanced methods exploit graph convolution to capture the global structure of an attributed [...] Read more.
Attributed graphs contain a lot of node features and structural relationships, and how to utilize their inherent information sufficiently to improve graph clustering performance has attracted much attention. Although existing advanced methods exploit graph convolution to capture the global structure of an attributed graph and achieve obvious improvements for clustering results, they cannot determine the optimal neighborhood that reflects the relevant information of connected nodes in a graph. To address this limitation, we propose a novel adaptive graph convolution using a heat kernel model for attributed graph clustering (AGCHK), which exploits the similarity among nodes under heat diffusion to flexibly restrict the neighborhood of the center node and enforce the graph smoothness. Additionally, we take the Davies–Bouldin index (DBI) instead of the intra-cluster distance individually as the selection criterion to adaptively determine the order of graph convolution. The clustering results of AGCHK on three benchmark datasets—Cora, Citeseer, and Pubmed—are all more than 1% higher than the current advanced model AGC, and 12% on the Wiki dataset especially, which obtains a state-of-the-art result in the task of attributed graph clustering. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Heat diffusion comparison as <math display="inline"><semantics> <mi>s</mi> </semantics></math> increases on an example graph. The center node is represented in yellow, and the similarity between the neighbor nodes and the center node decreases as the node color becomes darker. The range of heat diffusion increases from a small scale (<b>a</b>) to a large scale (<b>b</b>).</p>
Full article ">Figure 2
<p>The linear low-pass filters in adaptive graph convolution (AGC) [<a href="#B7-applsci-10-01473" class="html-bibr">7</a>] and heat kernel comparison.</p>
Full article ">Figure 3
<p>Spectral clustering visualization of <math display="inline"><semantics> <mi>k</mi> </semantics></math>-order graph convolution for dataset Cora. Colors of different nodes represent various labels, and the representation of nodes with the same label is closer as the value of <math display="inline"><semantics> <mi>k</mi> </semantics></math> increases. Low-order graph convolution makes node features indistinguishable, while high-order graph convolution might cause node features over-smoothing.</p>
Full article ">Figure 4
<p>The architecture of proposed adaptive graph convolution using heat kernel model for attributed graph clustering (AGCHK).</p>
Full article ">Figure 5
<p><math display="inline"><semantics> <mrow> <mi>d</mi> <mo>_</mo> <mi>D</mi> <mi>B</mi> <mi>I</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>i</mi> <mi>n</mi> <mi>t</mi> <mi>r</mi> <mi>a</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>i</mi> <mi>n</mi> <mi>t</mi> <mi>e</mi> <mi>r</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </mrow> </semantics></math> and clustering performance Acc, NMI, and F1 w.r.t. <math display="inline"><semantics> <mi>k</mi> </semantics></math> on datasets Cora and Wiki.</p>
Full article ">Figure 6
<p>The Acc scores of clustering performance using the different scaling parameter <math display="inline"><semantics> <mi>s</mi> </semantics></math> and threshold <math display="inline"><semantics> <mi>ε</mi> </semantics></math> on Cora.</p>
Full article ">
18 pages, 3212 KiB  
Article
Automatic and Fast Recognition of On-Road High-Emitting Vehicles Using an Optical Remote Sensing System
by Hao Xie, Yujun Zhang, Ying He, Kun You, Boqiang Fan, Dongqi Yu and Mengqi Li
Sensors 2019, 19(16), 3540; https://doi.org/10.3390/s19163540 - 13 Aug 2019
Cited by 15 | Viewed by 3408
Abstract
Optical remote sensing systems (RSSs) for monitoring vehicle emissions can be installed on any road and provide non-contact on-road measurements, that allow law enforcement departments to monitor emissions of a large number of on-road vehicles. Although many studies in different research fields have [...] Read more.
Optical remote sensing systems (RSSs) for monitoring vehicle emissions can be installed on any road and provide non-contact on-road measurements, that allow law enforcement departments to monitor emissions of a large number of on-road vehicles. Although many studies in different research fields have been performed using RSSs, there has been little research on the automatic recognition of on-road high-emitting vehicles. In general, high-emitting vehicles and low-emitting vehicles are classified by fixed emission concentration cut-points, that lack a strict scientific basis, and the actual cut-points are sensitive to environmental factors, such as wind speed and direction, outdoor temperature, relative humidity, atmospheric pressure, and so on. Besides this issue, single instantaneous monitoring results from RSSs are easily affected by systematic and random errors, leading to unreliable results. This paper proposes a method to solve the above problems. The automatic and fast-recognition method for on-road high-emitting vehicles (AFR-OHV) is the first application of machine learning, combined with big data analysis for remote sensing monitoring of on-road high-emitting vehicles. The method constructs adaptively updates a clustering database using real-time collections of emission datasets from an RSS. Then, new vehicles, that pass through the RSS, are recognized rapidly by the nearest neighbor classifier, which is guided by a real-time updated clustering database. Experimental results, based on real data, including the Davies-Bouldin Index (DBI) and Dunn Validity Index (DVI), show that AFR-OHV provides faster convergence speed and better performance. Furthermore, it is not easily disturbed by outliers. Our classifier obtains high scores for Precision (PRE), Recall (REC), the Receiver Operator Characteristic (ROC), and the Area Under the Curve (AUC). The rates of different classifications of excessive emissions and self-adaptive cut-points are calculated automatically in order to provide references for law enforcement departments to establish evaluation criterion for on-road high-emitting vehicles, detected by the RSS. Full article
(This article belongs to the Special Issue Computational Intelligence in Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>The optical remote sensing system for detecting on-road high-emitting vehicles.</p>
Full article ">Figure 2
<p>The concentration distribution of three main types of emissions collected in 192,097 datasets: (<b>a</b>) 3D front view; (<b>b</b>) 3D side view.</p>
Full article ">Figure 3
<p>The histogram of vehicle specific power (VSP) and three main types of emissions collected in 192,097 datasets: (<b>a</b>) VSP; (<b>b</b>) NO; (<b>c</b>) HC; (<b>d</b>) CO.</p>
Full article ">Figure 4
<p>The architecture of the method for the automatic and fast recognition of on-road high-emitting vehicles.</p>
Full article ">Figure 5
<p>The results of experiments to compare clustering methods: (<b>a</b>) K-means; (<b>b</b>) K-medoids; (<b>c</b>) ADIK+K-means; (<b>d</b>) proposed method.</p>
Full article ">Figure 6
<p>The comparison results of clustering experiments: (<b>a</b>) testing sets; (<b>b</b>) validation sets.</p>
Full article ">Figure 7
<p>The receiver operator characteristic (ROC) of different testing datasets: (<b>a</b>) dataset of Day 1; (<b>b</b>) dataset of Day 2; (<b>c</b>) dataset of Day 3; (<b>d</b>) dataset of Day 4.</p>
Full article ">
14 pages, 2866 KiB  
Article
Feature-Level Fusion of Surface Electromyography for Activity Monitoring
by Xugang Xi, Minyan Tang and Zhizeng Luo
Sensors 2018, 18(2), 614; https://doi.org/10.3390/s18020614 - 17 Feb 2018
Cited by 27 | Viewed by 4685
Abstract
Surface electromyography (sEMG) signals are commonly used in activity monitoring and rehabilitation applications as they reflect effectively the motor intentions of users. However, real-time sEMG signals are non-stationary and vary to a large extent within the time frame of signals. Although previous studies [...] Read more.
Surface electromyography (sEMG) signals are commonly used in activity monitoring and rehabilitation applications as they reflect effectively the motor intentions of users. However, real-time sEMG signals are non-stationary and vary to a large extent within the time frame of signals. Although previous studies have focused on the issues, their results have not been satisfactory. Therefore, we present a new method of conducting feature-level fusion to obtain a new feature space for sEMG signals. Eight activities of daily life (ADLs), including falls, were performed to obtain raw data from EMG signals from the lower limb. A feature set combining the time domain, time–frequency domain, and entropy domain was applied to the raw data to establish an initial feature space. A new projection method, the weighting genetic algorithm for GCCA (WGA-GCCA), was introduced to obtain the final feature space. Different tests were carried out to evaluate the performance of the new feature space. The new feature space created with the WGA-GCCA effectively reduced the dimensions and selected the best feature vectors dynamically while improving monotonicity. The Davies–Bouldin index (DBI) based on fuzzy c-means algorithms of the space obtained the lowest value compared with several fusion methods. It also achieved the highest accuracy when applied to support vector machine classifier. Full article
(This article belongs to the Special Issue Sensors for Health Monitoring and Disease Diagnosis)
Show Figures

Figure 1

Figure 1
<p>Muscles in Schematic and real placement.</p>
Full article ">Figure 2
<p>Eight activities of daily living. (<b>a</b>) stand-to-squat; (<b>b</b>) squat-to-stand; (<b>c</b>) stand-to-sit; (<b>d</b>) sit-to-stand; (<b>e</b>) walking, stair-ascending and stair-descending; (<b>f</b>) trip-fall.</p>
Full article ">Figure 3
<p>Delsys Full Wireless Surface Electromyography Test System (Trigno™ Wireless EMG).</p>
Full article ">Figure 4
<p>The SEN and SPE of WGA-GCCA with trial times in two examples. (<b>a</b>) One example result. (<b>b</b>) Another example result.</p>
Full article ">Figure 5
<p>Three-dimensional projection figure of gastrocnemius after WGA-GCCA.</p>
Full article ">Figure 6
<p>DBI with various numbers of clusters of CCA, GCCA, GA-GCCA, and WGA-GCCA.</p>
Full article ">Figure 7
<p>Dimension of feature sets after GA-GCCA fusion compared with GCCA. U<span class="html-italic">n</span> (<span class="html-italic">n</span> = x, y, z, o) Denotes the feature samples; x, y, z, o presents the four channels of data inputs.</p>
Full article ">Figure 8
<p>Increasing rate of recognition as the input increased from <span class="html-italic">n</span> to <span class="html-italic">n</span> + 1 (<span class="html-italic">n</span> = 1, 2, 3, 4, 5).</p>
Full article ">
3842 KiB  
Article
Performance Evaluation of Cluster Validity Indices (CVIs) on Multi/Hyperspectral Remote Sensing Datasets
by Huapeng Li, Shuqing Zhang, Xiaohui Ding, Ce Zhang and Patricia Dale
Remote Sens. 2016, 8(4), 295; https://doi.org/10.3390/rs8040295 - 30 Mar 2016
Cited by 31 | Viewed by 7109
Abstract
The number of clusters (i.e., the number of classes) for unsupervised classification has been recognized as an important part of remote sensing image clustering analysis. The number of classes is usually determined by cluster validity indices (CVIs). Although many CVIs have [...] Read more.
The number of clusters (i.e., the number of classes) for unsupervised classification has been recognized as an important part of remote sensing image clustering analysis. The number of classes is usually determined by cluster validity indices (CVIs). Although many CVIs have been proposed, few studies have compared and evaluated their effectiveness on remote sensing datasets. In this paper, the performance of 16 representative and commonly-used CVIs was comprehensively tested by applying the fuzzy c-means (FCM) algorithm to cluster nine types of remote sensing datasets, including multispectral (QuickBird, Landsat TM, Landsat ETM+, FLC1, and GaoFen-1) and hyperspectral datasets (Hyperion, HYDICE, ROSIS, and AVIRIS). The preliminary experimental results showed that most CVIs, including the commonly used DBI (Davies-Bouldin index) and XBI (Xie-Beni index), were not suitable for remote sensing images (especially for hyperspectral images) due to significant between-cluster overlaps; the only effective index for both multispectral and hyperspectral data sets was the WSJ index (WSJI). Such important conclusions can serve as a guideline for future remote sensing image clustering applications. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The multispectral images. (<b>a</b>–<b>c</b>) the true/false color map, the ground reference map and the corresponding spectral curves of ground truth classes of QuickBird datasets; (<b>d</b>–<b>f</b>) the corresponding maps of Landsat TM datasets; (<b>g</b>–<b>i</b>) the corresponding maps of Landsat ETM+ datasets; (<b>j</b>–<b>l</b>) the corresponding maps of GaoFen-1 datasets; (<b>m</b>–<b>o</b>) the corresponding maps of FLC1 datasets. (<b>a</b>) True color map; (<b>b</b>) false color map (7, 5, 3); (<b>c</b>) false color map (7, 5, 3); (<b>d</b>) true color map; and (<b>e</b>) false color map (bands 12, 9, and 1).</p>
Full article ">Figure 1 Cont.
<p>The multispectral images. (<b>a</b>–<b>c</b>) the true/false color map, the ground reference map and the corresponding spectral curves of ground truth classes of QuickBird datasets; (<b>d</b>–<b>f</b>) the corresponding maps of Landsat TM datasets; (<b>g</b>–<b>i</b>) the corresponding maps of Landsat ETM+ datasets; (<b>j</b>–<b>l</b>) the corresponding maps of GaoFen-1 datasets; (<b>m</b>–<b>o</b>) the corresponding maps of FLC1 datasets. (<b>a</b>) True color map; (<b>b</b>) false color map (7, 5, 3); (<b>c</b>) false color map (7, 5, 3); (<b>d</b>) true color map; and (<b>e</b>) false color map (bands 12, 9, and 1).</p>
Full article ">Figure 2
<p>The hyperspectral images. (<b>a</b>–<b>c</b>) the false color (FC) map, the corresponding ground reference map and the corresponding spectral curves of ground truth classes of Hyperion data sets; (<b>d</b>–<b>f</b>) the corresponding maps of HYDICE datasets; (<b>g</b>–<b>i</b>) the corresponding maps of ROSIS datasets; (<b>j</b>–<b>l</b>) the corresponding maps of AVIRIS datasets. (<b>a</b>) FC map (bands 93, 60, 10); (<b>b</b>) FC map (bands 120, 90, 10); (<b>c</b>) FC map (bands 90, 60, 10); and (<b>d</b>) FC map (bands 111, 90, 12).</p>
Full article ">Figure 3
<p>Clustering results of the multispectral images (each color represents a cluster). (<b>a</b>) QuickBird, <span class="html-italic">K</span> = 2; (<b>b</b>) <span class="underline">QuickBird, <span class="html-italic">K</span> = 3</span>; (<b>c</b>) QuickBird, <span class="html-italic">K</span> = 4; (<b>d</b>) QuickBird, <span class="html-italic">K</span> = 5; (<b>e</b>) <b>Landsat TM, <span class="html-italic">K</span> = 2</b>; (<b>f</b>) Landsat TM, <span class="html-italic">K</span> = 3; (<b>g</b>) <span class="underline">Landsat TM, <span class="html-italic">K</span> = 4</span>; (<b>h</b>) Landsat TM, <span class="html-italic">K</span> = 5; (<b>i</b>) <b>Landsat ETM+, <span class="html-italic">K</span> = 2;</b> (<b>j</b>) Landsat ETM+, <span class="html-italic">K</span> = 3; (<b>k</b>) <span class="underline">Landsat ETM+, <span class="html-italic">K</span> = 4</span>; (<b>l</b>) Landsat ETM+, <span class="html-italic">K</span> = 5; (<b>m</b>) <b>GaoFen-1, <span class="html-italic">K</span> = 2</b>; (<b>n</b>) GaoFen-1, <span class="html-italic">K</span> = 4; (<b>o</b>) <span class="underline">GaoFen-1, <span class="html-italic">K</span> = 5</span>; (<b>p</b>) GaoFen-1, <span class="html-italic">K</span> = 6; (<b>q</b>) <b>FLC1, <span class="html-italic">K</span> = 2</b>; (<b>r</b>) <b>FLC1, <span class="html-italic">K</span> = 3</b>; (<b>s</b>) <span class="underline">FLC1, <span class="html-italic">K</span> = 5</span>; and (<b>t</b>) FLC1, <span class="html-italic">K</span> = 6.</p>
Full article ">Figure 4
<p>Clustering results of the hyperspectral images by FCM (each color represents a cluster). (<b>a</b>) <b>Hyperion, <span class="html-italic">K</span> = 2</b>; (<b>b</b>) Hyperion, <span class="html-italic">K</span> = 3; (<b>c</b>) <span class="underline">Hyperion, <span class="html-italic">K</span> = 4</span>; (<b>d</b>) Hyperion, <span class="html-italic">K</span> = 5; (<b>e</b>) <b>HYDICE, <span class="html-italic">K</span> = 3</b>; (<b>f</b>) <span class="underline">HYDICE, <span class="html-italic">K</span> = 4</span>; (<b>g</b>) HYDICE, <span class="html-italic">K</span> = 5; (<b>h</b>) HYDICE, <span class="html-italic">K</span> = 6; (<b>i</b>) <b>ROSIS, <span class="html-italic">K</span> = 3</b>; (<b>j</b>) ROSIS, <span class="html-italic">K</span> = 4; (<b>k</b>) <span class="underline">ROSIS, <span class="html-italic">K</span> = 5</span>; (<b>l</b>) ROSIS, <span class="html-italic">K</span> = 6; (<b>m</b>) AVIRIS, <span class="html-italic">K</span> = 3; (<b>n</b>) <b>AVIRIS, <span class="html-italic">K</span> = 4</b>; (<b>o</b>) <span class="underline">AVIRIS, <span class="html-italic">K</span> = 5</span>; and (<b>p</b>) AVIRIS, <span class="html-italic">K</span> = 6.</p>
Full article ">Figure 5
<p>The overall performance of CVIs by applying FCM algorithm to cluster nine types of remote sensing datasets.</p>
Full article ">
284 KiB  
Article
Optimal Elbow Angle for Extracting sEMG Signals During Fatiguing Dynamic Contraction
by Mohamed R. Al-Mulla, Francisco Sepulveda and Bader Al-Bader
Computers 2015, 4(3), 251-264; https://doi.org/10.3390/computers4030251 - 10 Sep 2015
Cited by 4 | Viewed by 6706
Abstract
Surface electromyographic (sEMG) activity of the biceps muscle was recorded from 13 subjects. Data was recorded while subjects performed dynamic contraction until fatigue and the signals were segmented into two parts (Non-Fatigue and Fatigue). An evolutionary algorithm was used to determine the elbow [...] Read more.
Surface electromyographic (sEMG) activity of the biceps muscle was recorded from 13 subjects. Data was recorded while subjects performed dynamic contraction until fatigue and the signals were segmented into two parts (Non-Fatigue and Fatigue). An evolutionary algorithm was used to determine the elbow angles that best separate (using Davies-Bouldin Index, DBI) both Non-Fatigue and Fatigue segments of the sEMG signal. Establishing the optimal elbow angle for feature extraction used in the evolutionary process was based on 70% of the conducted sEMG trials. After completing 26 independent evolution runs, the best run containing the optimal elbow angles for separation (Non-Fatigue and Fatigue) was selected and then tested on the remaining 30% of the data to measure the classification performance. Testing the performance of the optimal angle was undertaken on nine features extracted from each of the two classes (Non-Fatigue and Fatigue) to quantify the performance. Results showed that the optimal elbow angles can be used for fatigue classification, showing 87.90% highest correct classification for one of the features and on average of all eight features (including worst performing features) giving 78.45%. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Elbow angles during Non-Fatigue rep; (<b>b</b>) Elbow angles during Fatiguing rep. Optimal window of joint angles selected by the GA showing where the Surface electromyographic (sEMG) signal was used for classification .</p>
Full article ">Figure 2
<p>(<b>a</b>) A scatter plot of elbow angles selected by the GA. (larger dots indicate better separation); (<b>b</b>) 3D histogram of elbow joint angles selected by all the GA runs. Elbow joint angles selected by 26 evolutionary runs displayed in a scatter plot and a 3D histogram.</p>
Full article ">Figure 3
<p>Graphical representation of the classification performance.</p>
Full article ">
Back to TopTop