[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
A Global Model for Estimating Tropospheric Delay and Weighted Mean Temperature Developed with Atmospheric Reanalysis Data from 1979 to 2017
Previous Article in Journal
Crop Monitoring Using Sentinel-1 Data: A Case Study from The Netherlands
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Selection on Sentinel-2 Multispectral Imagery for Mapping a Landscape Infested by Parthenium Weed

School of Agricultural, Earth and Environmental Sciences, University of KwaZulu Natal, P. Bag X01 Scottsville, Pietermaritzburg 3209, South Africa
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(16), 1892; https://doi.org/10.3390/rs11161892
Submission received: 27 May 2019 / Revised: 5 August 2019 / Accepted: 6 August 2019 / Published: 13 August 2019

Abstract

:
In the recent past, the volume of spatial datasets has significantly increased. This is attributed to, among other factors, higher sensor temporal resolutions of the recently launched satellites. The increased data, combined with the computation and possible derivation of a large number of indices, may lead to high multi-collinearity and redundant features that compromise the performance of classifiers. Using dimension reduction algorithms, a subset of these features can be selected, hence increasing their predictive potential. In this regard, an investigation into the application of feature selection techniques on multi-temporal multispectral datasets such as Sentinel-2 is valuable in vegetation mapping. In this study, ten feature selection methods belonging to five groups (Similarity-based, statistical-based, Sparse learning based, Information theoretical based, and wrappers methods) were compared based on f-score and data size for mapping a landscape infested by the Parthenium weed (Parthenium hysterophorus). Overall, results showed that ReliefF (a Similarity-based approach) was the best performing feature selection method as demonstrated by the high f-score values of Parthenium weed and a small size of optimal features selected. Although svm-b (a wrapper method) yielded the highest accuracies, the size of optimal subset of selected features was quite large. Results also showed that data size affects the performance of feature selection algorithms, except for statistically-based methods such as Gini-index and F-score and svm-b. Findings in this study provide a guidance on the application of feature selection methods for accurate mapping of invasive plant species in general and Parthenium weed, in particular, using new multispectral imagery with high temporal resolution.

1. Introduction

The dimension space of variables given as input to a classifier can be reduced without an important loss of information, while decreasing its processing time and improving the quality of its output [1]. To date, studies on dimension reduction in remote sensing have mostly focused on hyperspectral datasets [2,3,4] and high spatial resolution multispectral imagery using Object-Based Image Analysis (OBIA) [5,6,7]. Generally, multispectral images have received less attention comparatively, likely due to having a limited number of bands, which do not require dimension reduction. However, with the launch of high temporal resolution sensors such as Sentinel-2, the amount of image data that can be acquired within a short period has considerably increased [8]. This is due to the sensor’s improved spectral resolution (13 bands) and a five day temporal resolution [9].
Generally, high-dimensional remotely sensed datasets contain irrelevant information and highly redundant features. Such dimensionality deteriorates quantitative (e.g., leaf area index and biomass) and qualitative (e.g., land-cover) performance of statistical algorithms by overfitting data [10]. High dimensional data are often associated with the Hughes effects or the curse of dimensionality, a phenomenon that occurs when the number of features in a dataset is greater than the number of samples [11,12]. Hughes effects affects the performance of algorithms previously designed for low-dimensional data. Whereas high dimensionality can lead to poor generalization of learning algorithms during the classification process [12], it can also embed features that are crucial for classification enhancement. Hence, when using dimension reduction algorithms, a subset of those features can be selected from the high dimensional data, increasing their predictive potential [13].
There are two main components of dimension reduction strategies: feature extraction or construction and feature selection or feature ranking. Feature extraction (e.g., Principle Component Analysis (PCA)), constructs a new and low dimensional feature space using linear or non-linear combinations of the original high-dimensional feature space [14] while feature selection (e.g., Fisher Score and Information Gain) extracts subsets from existing features [10]. Although feature extraction methods produce higher classification accuracies, the interpretation of generated results is often challenging [2]. However, feature selection methods do not change the original information of features, thus giving models better interpretability and readability. Feature selection techniques have been applied in text mining and genetic analysis [14]. Hence, in this study, they were preferred over feature extraction methods.
Traditional feature selection techniques are typically grouped into three approaches namely; filter, embedded and wrapper methods [15]. In earth observation related studies, feature selection algorithms have generally been compared based on this grouping [5,16]. However, in the advent of big data, this grouping can be regarded to be very broad, necessitating development of new feature selections algorithms. For instance, within filter feature selection methods, there are some that evaluate the importance of features based on the ability to preserve data similarity (e.g., Fisher Score, ReliefF), while others use a heuristic filter criterion (e.g., Mutual Information Maximization). Therefore, it is crucial to re-evaluate the comparison of feature selection algorithms in a data specific perspective.
Li, et al. [17] reclassified traditional feature algorithms for generic data into five groups: similarity-based feature selection, information theoretical-based feature selection, statistical-based feature selection, sparse learning-based feature selection and wrappers. They devised an open-source feature selection repository, named scikit-feature that provides 40 feature selection algorithms (including unsupervised feature selection approaches). Some selections, such as Joint Mutual Information and decision tree forward, are relatively new in earth observation applications [18].
Over the last few decades, the number of vegetation indices has significantly increased. For instance, Henrich, et al. [19] gathered 250 vegetation indices derivable from Sentinel-2. Computation of vegetation indices from a higher temporal resolution imagery like Sentinel-2 would lead to data with very increased multi-collinearity, a higher number of derived variables, and increased dimensions. Hence, the need for accurate and efficient feature selection techniques when dealing with new generation multispectral imagers such as Sentinel-2 is becoming increasingly valuable in vegetation mapping [8]. To the best of our knowledge, no earth observation related study has endeavored to undertake empirical evaluation of feature selection methods as provided in the scikit-feature repository for generic data.
In this study, feature selection algorithms based on the Li, et al. [17]‘s classification were compared for mapping a landscape infested by Parthenium weed using Sentinel-2. The Parthenium weed is an alien invasive herb of tropical American origin that has infested over thirty countries. It has been identified as one of the seven most devastating and hazardous weeds worldwide [20]. A number of studies [21,22] have reported its adverse impacts on ecosystem functioning, biodiversity, agricultural productivity and human health. A detailed comparison of feature selection algorithms on Sentinel-2 spectral bands combined with vegetation indices and with respect to Li, et al. [17]’s classification, would therefore: (a) improve mapping accuracy, valuable for designing mitigation approaches; (b) shed light on the most valuable feature selection group and (c) identify the most suitable feature selection method for accurately mapping a Parthenium weed infested landscape. Unlike previous studies that evaluated features selection methods on the basis of overall classification accuracy [5,23,24], this study investigated their performance on a specific landscape phenomenon’s (Parthenium weed) mapping accuracy, as high overall classification accuracy does not always mean a reliable accuracy for a specific class [25]. In this study, we sought to provide a detailed comparison of feature selection algorithms on higher temporal resolution satellite images with high data volume. Specifically, we looked at (a) their performance on Parthenium weed using specific class-related accuracies as an evaluation criterion, and (b) the impact of data size on their accuracy.

2. Material and Method

2.1. Study Area

This study was conducted within the Mtubatuba municipality on the North-East Coast of the KwaZulu-Natal province, South Africa (Figure 1). The study area covers 129 km2, and is characterized by heavy Parthenium infestation. The area is predominantly underlain by Basalt, Sand and Mudstone geological formations [26]. Annual average rainfall ranges from 600 mm to 1250 mm and temperatures vary around 21 °C, respectively. Summers are generally warm to hot, while winters are cool to mild [27]. The sampling area is characterized by a mosaic of several land use/land cover types that include commercial agriculture (e.g., forestry plantations and sugarcane farming), subsistence farming (beans, bananas, potatoes, and cattle), mining, and high and low density residential areas [27].

2.2. Reference Data

Using a high resolution (50 cm) color orthophotograph [28] of the study area, conspicuous patches of Parthenium weed infestations were randomly selected, stratified equally across different land cover/use types. The selected sites were then surveyed using a differentially corrected Trimble GeoXT handheld GPS receiver with about 50 cm accuracy. The ground truth campaign was conducted during summer, between the 12th of January and 2nd of February 2017. At each Parthenium weed site, at least 10 m × 10 m quadrats were demarcated. The quadrats were located mostly in the middle of a large patch (greater than 10 × 10 m) of Pathernium in order to cater for any possible mismatches with the sentinel pixels [29]. In total, 90 quadrats were randomly selected across different land-cover types to account for variability in different ecological conditions of the study area. GPS points of surrounding land cover types such as forest, grassland, built-up and water bodies were also collected. Supplementary X-Y coordinates of these land-cover types were also created from the color orthophotograph to increase the number of training samples. The aforementioned land-cover classes were the most predominant in the study area and were therefore used to evaluate the discriminatory power of different models developed in this study for mapping Parthenium weed infested areas. In total, 447 reference points for mapping Parthenium weed and its surrounding land cover classes were obtained. To determine the optimal feature selection methods, and to test the effect of data sizes, these ground reference data were randomly split into training and test sets in three different ratios: 1:3; 1:1; 3:1 as shown in Table 1 [30]. The random split was undertaken using the function “train_test_split” of the Sklearn python library. “Random-state” and “stratify” parameters were included in the function to respectively allow reproducibility and obtain the same proportions of class labels as the input dataset.
The data design also allowed evaluating the investigated feature selection methods with respect to Hughes effects.

2.3. Acquisition of Multi-Temporal Sentinel-2 Images and Pre-Processing

Three Level 1C Sentinel-2A images were acquired on 19 January of 2017 under cloudless conditions. The Semi-Automatic Classification Plugin [31] within the QGIS (version 2.14.11) software was used to correct the acquired images for atmospheric effects. The Semi-Automatic Classification Plugin uses the Dark Object Subtraction in order to convert Top Of Atmosphere reflectance (TOA) to Bottom Of Atmosphere reflectance (BOA). Bands (band 1, band 9 and band 10) with 60 m-resolution were omitted in this study. Moreover, bands with 20 m resolution were resampled to 10 m using ArcMap (version 10.3) to allow layer stacking with 10 m bands.

2.4. Data Analysis

2.4.1. Feature Section Methods

In this section, the five groups of feature selection methods are briefly discussed. Two representative methods were randomly chosen from each group to achieve the comparison.

(A) Similarity-Based Feature Selection Methods

Similarity-based feature selection methods evaluate the importance of features by determining their ability to preserve data similarity using some performance criterion. The two selected feature selection algorithms in this study were Trace ratio and Relief. Trace ratio [32] maximizes data similarity for samples of the same class or those that are close to each other while minimizing data similarity for the sample of different classes or those that are far away from each other. More important features also have a larger score. ReliefF [33] assigns a weight to each feature of a dataset and feature values which are above a predefined threshold are then selected. The rationale behind ReliefF is to select features randomly, and based on nearest neighbors, the quality of features is estimated according to how well their values distinguish among the instances of the same and different classes nearing each other. The larger the weight value of a feature, the higher the relevance [34].

(B) Statistical-Based Feature Selection Methods

Statistical-based feature selection methods rely on statistical measures in order to the estimate relevance of features. Some examples of statistical-based feature selection methods include the Gini index and the F-score. The Gini index [35] is a statistical measure that quantitatively evaluates the ability of a feature to separate instances from different classes [14]. It was earlier used in decision tree for splitting attributes. The rationale behind Gini index is as follows: Suppose S is the set of s samples with m different classes ( C i , i = 1 ,   ,   m ) . According to the differences of classes, S can be divided into m subset ( S i , i = 1 ,   ,   m ) . Given S i is the sample set which belongs to class C i , s i is the sample number of set Si, Gini index of set S can be computed according to the equation below [36]:
G i n i ( S ) = 1 i   =   1 m P i 2
where P i denotes the probability for any sample to belong to C i and to estimate with s i / s .
The F-score [37] is calculated as follows: Given feature f i , n j , µ, u j , σ j represent the number of instances from class j, the mean feature value, the mean feature value on class j, the standard deviation of feature value on class j, respectively, the F-score of a feature fi can be determined as follows:
F _ s c o r e ( f i ) = j n j c 1 ( u j u ) 2 1 n c j ( n j 1 ) σ j 2

(C) Sparse Learning Based Methods

Sparse learning based methods is a group of embedded approaches. They aim at reducing the fitting errors along with some sparse regularization terms, which make feature coefficients small or equal to zero. To make a selection, corresponding features are discarded [14]. Feature selection algorithms belonging to this group have been recognized to produce good performance and interpretability. In this study, sparse learning based methods with the following sparse regularization terms: ℓ1-norm Regularizer (LS-121) [38] and ℓ2,1-norm Regularizer (LL-121) [39] were implemented.

(D) Information Theoretical Based Methods

Information theoretical-based methods apply some heuristic filter criteria in order to estimate the relevance of features. Some feature selection algorithms that belong to this family include Joint Mutual Information (JMI) and Mutual Information Maximization (MIM) or Information Gain. The JMI seeks to incorporate new unselected features that are complementary to existing features given the class labels in the feature selection process [14] while the MIM measures the importance of a feature by its correlation with the class labels. MIM assumes that features with strong correlations would achieve a good classification performance [14].

(E) Wrapper

Wrapper methods use a predefined learning algorithm, which acts like a black box to assess the importance measures of selected features based on their predictive performance. Two steps are involved in selecting features. First, a subset of features is searched and then selected features are evaluated repeatedly until the highest learning performance is reached. Features regarded as being relevant are the ones that yield the highest learning performance [14]. These two steps are implemented using the forward or backward selection strategies. In forward selection strategy, the search of relevant features starts with an empty set of features, then features are progressively added into larger subsets, whereas in backward elimination, it starts with the full set of features and then progressively eliminates the least relevant ones [40].
However, the implementation of wrapper methods is limited in practice for high-dimensional dataset due to the large size of the search space. Some examples of wrapper methods that were used in this study include decision tree forward (dt-f) [41] and Support vector machine backward (svm-b) [41].

2.4.2. Vegetation Indices Computation

In total, 75 vegetation indices (VI) derived from Sentinel-2 wavebands were computed using the online indices-database (IDB) developed by Henrich, et al. [19]. The IDB provides over 261 parametric and non-parametric indices that can be used for over 99 sensors and allows the viewing of all available VI for specific sensors and applications [19]. VI in this study were selected because of their usefulness for vegetation mapping and in order to increase the dimensionality of Sentinel-2 data.

2.4.3. Classification Algorithm: Random Forest (RF)

RF classifier was used to infer models from different selected features using the investigated feature selection methods. Random forest (RF) is a combination of decision tree classifiers where each classifier casts a single vote for the most frequent class to classify an input vector [42]. RF grows trees from random subsets drawn from the input dataset using methods such as bagging or bootstrap aggregation. Split of input dataset is typically performed using attribute selection measures (Information Gain, Gini-Index). Attribute selection measures are useful in maximizing dissimilarity between classes and therefore determine the best split selection in creating subsets [43]. In the process of a RF model training, the user defines the number of features at each node in order to generate a tree and the number of trees to be grown. The classification of a new dataset is done by passing down each case of the datasets to each of the grown trees, then the forest chooses a class having the most votes of the trees for that case [44]. More details on RF can be found in Breiman [45]. RF was chosen in this study, as it can efficiently handle large and highly dimensional datasets [46,47].

2.4.4. Model Assessment

To assess the classification accuracy of models on test datasets, estimated classes in different models developed in this study were cross-tabulated against the ground-sampled classes for corresponding pixels in a confusion matrix. The performance of developed models was assessed on test data sets using performance measures such as user’s (UA), and producer’s accuracies (PA) and f-score of Parthenium weed class. Supplementary information including PA, UA of other classes, and Kappa coefficient was added. The UA refers to the probability that a pixel labeled as a certain class on the map represents that class on the ground. PA represents pixels that belong to a ground-sampled class but fail to be classified into the correct class. F-score is the harmonic mean of UA and PA. F-score is typically used to assess accuracy per-class as it represents a true outcome for specific classes [23,25]. Kappa coefficient represents the extent to which classes on the ground are correct representations of classes on the map. Their formulae are as follows:
P A = T P T P + F N
U A = T P T P + F P
F = 2 × P A × U A P A + U A
K a p p a   c o e f f i c i e n t = P o P e 1 P e
where:
TP (true positive) represents the number of correctly labeled positive samples;
FP (false positive) represents to the number of negative samples that were incorrectly labeled as positive samples;
FN (false negative) represents the number of positive samples that were incorrectly labeled as negative samples;
Po: relative observed agreement among classes;
Pe: hypothetical probability of chance agreement.

2.4.5. Software and Feature Selection

All investigated feature selection methods were applied on the three training datasets aforementioned using Scikit-feature library, a package of the Python (version 3.6) programming language. This library was developed by Li, et al. [14] and provides more than 40 feature selection algorithms. To obtain the f-score accuracy of selected features for each feature selection method, we first created a range of numbers that started from 1 to 85 (which is the total of variables) instead of specifying the number of selected features as prescribed by Li, et al. [14]. Each number in the range corresponded to the size of selected feature subsets. Then, for each dataset, a RF model was trained and evaluated on test dataset using selected feature subset through a loop iteration. RF was run in python (version 3.6) using the Sklearn library [48]. As RF was only used in this study for evaluating the performance of optimal variables selected by different feature selection methods, its hyperparameters were kept in default (e.g., number of trees in the forest equals to 10; criterion sets to “gini”). Default hyperparameters often yield excellent results [49]. Additionally, according to Du, et al. [50], larger number of trees do not influence classification results. This procedure was repeated ten times by reshuffling samples of training and test sets and the mean of f-score was computed for each select feature subset. This was to ensure the reliability of the results of investigated feature selection methods. The feature subset with the highest mean f-score was considered the most optimal. A code that automates the whole procedure including deriving VI, was written in Python (version 3.6).

3. Results

3.1. Comparison Among Investigated Features Algorithms

Figure 2 shows that the size of the feature subset and training and test sets determine the f-score accuracy of Parthenium weed using RF. In general, f-score increases with an increase in the size of feature subset until it reaches a plateau around 10 features, showing the insensitivity of RF in face of the noisy or redundant variables. However, it is noticeable that f-score accuracy of some feature selection methods such as Gini-index and ReliefF, which belong to the statistical-based feature selection methods, and LL-121 were found at smaller feature subsets. With respect to the size of the training set, f-score of Parthenium weed increased as the ratio between training and test got larger. As rule of thumb, this shows that when the ratio between training and test dataset is large (for example 3:1), a learning curve of higher F-score would be produced. Nevertheless, some feature selection methods such as svm-b, Gini-index and F-score seemed to yield similar f-score accuracies, regardless of the size of the dataset.

3.2. Comparison of Performance Between Peak Accuracy and Accuracy Derived From Full Feature Subsets

3.2.1. 1st Training Set

As per Table 2, all the investigated feature selection methods yielded similar f-score of Parthenium weed for optimal feature subset. F-score accuracies varied from 71% to 72%. Svm-b produced the highest f-score accuracy (72.5%). However, in terms of the size of optimal feature subset, Relief and Gini-index were the best in reducing dimensionality of full-dataset. The size of feature subsets was 6 and 13 respectively. F-score method was the lowest at this point of view. For this dataset, the ReliefF method can be recommended as its f-score is among the highest and the size of optimal feature subset among the smallest. Computational time and accuracies of other classes were also low and high, respectively, for ReliefF method (Table 3).

3.2.2. 2nd Training Set

As shown in Table 4, apart from LS_121 and F-score, all feature selection methods could reduce the number of features with higher f-score accuracy than the full dataset. As for the first dataset, Svm-b was the best performing feature selection method. Its PA, UA and f-score of Parthenium weed and Kappa coefficient (Table 5) were the highest. ReliefF selected the least number (4) of optimal features and was among the highest top performing feature selection methods after svm_b in terms of f-score and PA of Parthenium weed. It was followed by Gini-index, LL_121 with respect to f-score and size of feature subsets. Once more, F-score method turned out to perform poorly because of a large number of selected features and no improvement of f-score accuracy of the Parthenium weed.

3.2.3. Third Training Set

Table 6 shows that LL_121 and ReliefF, respectively were among the feature selection methods that selected a small subset of features with high f-score of Parthenium weed, with PA and UA accuracies above 3% difference from the full dataset. ReliefF, for example, yielded 77.2% of f-score, 80% of PA and 75% of UA with only 7 optimal features. Full dataset yielded 72.6% of f-score, 75.2% of PA and 71.4% of UA without any feature selection method applied. Svm_b outperformed all the feature selection methods with the highest PA (82.3%) and f-score of Parthenium weed (78.1%), and kappa coefficient (0.83) (Table 7). However, the number of optimal features selected was quite large (33). As for previous datasets, the performance of F-score method was the worst with the lowest PA (78.5%), UA (73.6%) and f-score (75.6%) of Parthenium weed and largest feature subset (82).
Figure 3 illustrates the spatial distribution of Parthenium weed and surrounding land-cover using full dataset and optimal features from ReliefF.

4. Discussion

This study sought to compare ten feature selection algorithms with two of those belonging to the following feature selection method groups: Similarity based feature selection methods, Statistical-based feature selection methods, sparse learning based methods, Information theoretical based methods and Wrapper methods. These feature selection algorithms were applied on Sentinel-2 spectral bands combined with 75 vegetation indices for mapping a landscape infested by Parthenium weed. The comparison was based on the f-score of Parthenium weed using Random forest (RF). We also tested the effect of training and test sets sizes on the performance of the investigated feature selection algorithm.

4.1. Comparison of Feature Selection Methods

The results showed that feature selection algorithms could reduce the dimensionality of Sentinel-2 spectral bands combined with vegetation indices. The algorithms could increase classification accuracies of Parthenium weed using the random forest classifier by up to 4%, depending on the adopted feature selection method and size of the dataset. Previous studies found similar results [34,51]. For example, Colkesen and Kavzoglu [34] who applied filter-based feature selection algorithms and three machine learning techniques on WorldView-2 image for determining the most effective object features in object-based image analysis achieved a significant improvement (about 4%) by applying feature selection methods. Overall, ReliefF was the best performing feature selection method because it could bring down the number of features from 85 to 6, 4 and 7 on the first, second and third dataset respectively. Its f-score, PA and UA accuracies for Parthenium weed were also among the highest (Table 2, for example). According to Vergara and Estévez [52], the purpose of feature selection is to determine the smallest feature subset that can produce the minimum classification error. This finding concurs with studies that endeavored to compare ReliefF with other feature selection methods. For example, Kira and Rendell [53] found that subset of features selected by ReliefF tend to be small compared to other feature selection methods as only statistically relevant features are retained during the selection process. Studies that compared ReliefF with other feature selection methods reported that mapping accuracy with selected features from ReliefF was similar to the best feature selection methods [34,54]. However, our findings are opposed to [34] who found that ReliefF selected the highest number of features in comparison with Chi -square and information gain algorithms, but slightly lower classification accuracies than information gain using random forest, support vector machines and neural network. We suggest that repeated classifications on reshuffled training and test data should be investigated to confirm their findings. In terms of f-score, PA and UA of Parthenium weed, svm-b, which belongs to the wrapper group outperformed all the feature selection methods for the three datasets (Table 2, Table 4 and Table 6). Although computationally expensive, several authors have noted that wrappers outperform filter methods [55,56,57]. However, in this study svm-b was found to be less computational intense than some of filter methods (Table 2, Table 4 and Table 6). F-score method did not perform well for mapping Parthenium weed because of low accuracies and large subset of selected features. To the best of our knowledge, its use is very limited in earth observation related studies. Further investigations are therefore necessary. Concerning investigated feature selection groups, not a single group performed well on all the datasets. This supports the recommendation that there is no universal ‘best’ method for all the learning tasks [58].

4.2. Impact of Training Sizes on Feature Selection Performance

The results show that the performance of feature selection algorithm depends on the ratio between the training and test dataset. It can be noticed that smaller difference of f-score accuracies of Parthenium weed between optimal features and full dataset were obtained when the ratio between training and test was 1:3 or 1:1 (Table 2 and Table 3). In consistency with Jain and Zongker [59], a small sample size and a large number of features impair the performance of feature selection methods due to the curse of dimensionality. All the investigated feature selection algorithms positively influenced f-score accuracies of dataset when the ratio between training and test was large (1:3 or 70% training and 30% test) (Table 6). This concurs with reference [51], who demonstrated that the higher the training size was, the better the classification accuracy. They also highlighted the necessity of selecting the appropriate feature selection for improving classification accuracies. In this study, we found that some feature selection algorithms such as Gini-index and F-score, which belong to the Statistical-based Feature selection methods, and svm-b did not seem to be affected by the curse of dimensionality (Figure 2). We did not come across studies showing this finding, hence further investigations should be carried out for corroboration.

4.3. Implications of Findings in Parthenium Weed Management

On invaded landscapes, Parthenium weed expands more rapidly than native plants [60]. Spectral bands alone are not enough to achieve reliable mapping accuracies [61]. Increasing data dimensionality by combining, among others, Sentinel-2 image bands, vegetation indices and other variable types and applying an appropriate feature selection approach, higher Pathenium mapping accuracy can be achieved. This study provides a guidance on how newly developed feature selection methods based on Li, et al. [17]‘s classification should be used to reduce the dimension of high temporal resolution imagery such as Sentinel-2 in mapping Parthenium weed. Accurate spatial distribution of Pathenium weed would enhance the decision-making for appropriate mitigation measures.

5. Conclusions

The following conclusions can be drawn from the findings:
(1)
Wrappers methods such as svm-b yield higher accuracies on classifying Parthenium weed using the Random forest classifier;
(2)
ReliefF was the best performing feature selection method in terms of f-score and the size of optimal features selected;
(3)
To achieve better performance with feature selection methods, the ratio of 3:1 between the training and test set size turned out to be better than ratios of 1.1 and 1:3;
(4)
Gini-index, F-score and svm-b, were slightly affected by the curse of dimensionality;
(5)
None of feature selection method groups seemed to perform the best for all the datasets.
The findings of this study are critical for reducing the computational complexity of processing large volume of Sentinel-2 image data. With the advent of Sentienl-2A and B, an increased volume of data is available, necessitating feature selection. This offers possibilities to derive useful information from them, hence accurate classification maps of Parthenium weed. Further research should look at comparing other feature selection methods with different classifiers. Moreover, a combination of feature selection methods such ReliefF and svm-b should be considered, as they select a small number of features and yield a high f-score accuracy, respectively.

Author Contributions

Z.K. was responsible for the conceptualization, methodological development, analysis and write-up. O.M., J.O. and K.P. were responsible for conceptualization, methodological development, reviewing and editing the paper. In addition, O.M. was responsible for acquiring funding.

Funding

This study was supported by the UKZN funded Big data for Science society (BDSS) programme and the DST/NRF funded SARChI chair in land use planning and management (Grant Number: 84157).

Acknowledgments

We would like to thank the two anonymous reviewers for providing constructive comments, which greatly improved the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Serpico, S.B.; Bruzzone, L. A new search algorithm for feature selection in hyperspectral remote sensing images. IEEE Trans. Geosci. Remote Sens. 2001, 39, 1360–1367. [Google Scholar] [CrossRef]
  2. Zheng, X.; Yuan, Y.; Lu, X. Dimensionality reduction by spatial–spectral preservation in selected bands. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5185–5197. [Google Scholar] [CrossRef]
  3. Adam, E.; Mutanga, O. Spectral discrimination of papyrus vegetation (Cyperus papyrus L.) in swamp wetlands using field spectrometry. ISPRS J. Photogramm. Remote Sens. 2009, 64, 612–620. [Google Scholar] [CrossRef]
  4. Xie, L.; Li, G.Y.; Peng, L.; Chen, Q.C.; Tan, Y.L.; Xiao, M. Band selection algorithm based on information entropy for hyperspectral image classification. J. Appl. Remote Sens. 2017, 11, 17. [Google Scholar] [CrossRef]
  5. Ma, L.; Fu, T.; Blaschke, T.; Li, M.; Tiede, D.; Zhou, Z.; Ma, X.; Chen, D. Evaluation of feature selection methods for object-based land cover mapping of unmanned aerial vehicle imagery using random forest and support vector machine classifiers. ISPRS Int. J. Geo Inf. 2017, 6, 51. [Google Scholar] [CrossRef]
  6. Yu, Q.; Gong, P.; Clinton, N.; Biging, G.; Kelly, M.; Schirokauer, D. Object-based detailed vegetation classification with airborne high spatial resolution remote sensing imagery. Photogramm. Eng. Remote Sens. 2006, 72, 799–811. [Google Scholar] [CrossRef]
  7. Waser, L.T.; Küchler, M.; Jütte, K.; Stampfer, T. Evaluating the potential of WorldView-2 data to classify tree species and different levels of ash mortality. Remote Sens. 2014, 6, 4515–4545. [Google Scholar] [CrossRef]
  8. Aires, F.; Pellet, V.; Prigent, C.; Moncet, J.L. Dimension reduction of satellite observations for remote sensing. Part 1: A comparison of compression, channel selection and bottleneck channel approaches. Q. J. R. Meteorol. Soc. 2016, 142, 2658–2669. [Google Scholar] [CrossRef]
  9. Immitzer, M.; Vuolo, F.; Atzberger, C. First Experience with Sentinel-2 Data for Crop and Tree Species Classifications in Central Europe. Remote Sens. 2016, 8, 166. [Google Scholar] [CrossRef]
  10. Gnana, D.A.A.; Balamurugan, S.A.A.; Leavline, E.J. Literature review on feature selection methods for high-dimensional data. Int. J. Comput. Appl. 2016, 136, 8887. [Google Scholar] [CrossRef]
  11. Kavzoglu, T.; Mather, P. The role of feature selection in artificial neural network applications. Int. J. Remote Sens. 2002, 23, 2919–2937. [Google Scholar] [CrossRef]
  12. Taşkın, G.; Kaya, H.; Bruzzone, L. Feature selection based on high dimensional model representation for hyperspectral images. IEEE Trans. Image Process. 2017, 26, 2918–2928. [Google Scholar] [CrossRef] [PubMed]
  13. Lagrange, A.; Fauvel, M.; Grizonnet, M. Large-scale feature selection with Gaussian mixture models for the classification of high dimensional remote sensing images. IEEE Trans. Comput. Imaging 2017, 3, 230–242. [Google Scholar] [CrossRef]
  14. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. CSUR 2017, 50, 94. [Google Scholar] [CrossRef]
  15. Cao, X.; Wei, C.; Han, J.; Jiao, L. Hyperspectral Band Selection Using Improved Classification Map. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2147–2151. [Google Scholar] [CrossRef] [Green Version]
  16. Novack, T.; Esch, T.; Kux, H.; Stilla, U. Machine learning comparison between WorldView-2 and QuickBird-2-simulated imagery regarding object-based urban land cover classification. Remote Sens. 2011, 3, 2263–2282. [Google Scholar] [CrossRef]
  17. Li, J.; Tang, J.; Liu, H. Reconstruction-based unsupervised feature selection: An embedded approach. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, IJCAI/AAAI, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  18. Chen, H.M.; Varshney, P.K.; Arora, M.K. Performance of mutual information similarity measure for registration of multitemporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2445–2454. [Google Scholar] [CrossRef]
  19. Henrich, V.; Götze, E.; Jung, A.; Sandow, C.; Thürkow, D.; Gläßer, C. Development of an Online indices-database: Motivation, concept and implementation. In Proceedings of the 6th EARSeL Imaging Spectroscopy SIG Workshop Innovative Tool for Scientific and Commercial Environment Applications, Tel-Aviv, Israel, 16–19 March 2009. [Google Scholar]
  20. Adkins, S.; Shabbir, A. Biology, ecology and management of the invasive parthenium weed (Parthenium hysterophorus L.). Pest Manag. Sci. 2014, 70, 1023–1029. [Google Scholar] [CrossRef]
  21. Dhileepan, K. Biological control of parthenium (Parthenium hysterophorus) in Australian rangeland translates to improved grass production. Weed Sci. 2007, 55, 497–501. [Google Scholar] [CrossRef]
  22. McConnachie, A.J.; Strathie, L.W.; Mersie, W.; Gebrehiwot, L.; Zewdie, K.; Abdurehim, A.; Abrha, B.; Araya, T.; Asaregew, F.; Assefa, F.; et al. Current and potential geographical distribution of the invasive plant Parthenium hysterophorus (Asteraceae) in eastern and southern Africa. Weed Res. 2011, 51, 71–84. [Google Scholar] [CrossRef]
  23. Georganos, S.; Grippa, T.; Vanhuysse, S.; Lennert, M.; Shimoni, M.; Kalogirou, S.; Wolff, E. Less is more: Optimizing classification performance through feature selection in a very-high-resolution remote sensing object-based urban application. GISci. Remote Sens. 2018, 55, 221–242. [Google Scholar] [CrossRef]
  24. Kganyago, M.; Odindi, J.; Adjorlolo, C.; Mhangara, P. Selecting a subset of spectral bands for mapping invasive alien plants: A case of discriminating Parthenium hysterophorus using field spectroscopy data. Int. J. Remote Sens. 2017, 38, 5608–5625. [Google Scholar] [CrossRef]
  25. Ao, Z.; Su, Y.; Li, W.; Guo, Q.; Zhang, J. One-class classification of airborne LiDAR data in urban areas using a presence and background learning algorithm. Remote Sens. 2017, 9, 1001. [Google Scholar] [CrossRef]
  26. Norman, N.; Whitfield, G. Geological Journeys: A Traveller’s Guide to South Africa’s Rocks and Landforms; Struik: Cape Town, South Africa, 2006. [Google Scholar]
  27. Municipality, M.L. Integrated Development Plan; Prepared by the Councillors and Officials of the Msunduzi Municipality; Mtubatuba Municipality: Mtubatuba, South Africa, 2002. [Google Scholar]
  28. National Geo-Spatial Information. NGI, Pietermaritzburg (Air Photo); National Geo-Spatial Information: Cape Town, South Africa, 2008. [Google Scholar]
  29. Carter, G.A.; Lucas, K.L.; Blossom, G.A.; Lassitter, C.L.; Holiday, D.M.; Mooneyhan, D.S.; Holcombe, T.R.; Griffith, J. Remote sensing and mapping of tamarisk along the Colorado river, USA: A comparative use of summer-acquired Hyperion, Thematic Mapper and QuickBird data. Remote Sens. 2009, 1, 318–329. [Google Scholar] [CrossRef]
  30. Pal, M.; Foody, G.M. Feature selection for classification of hyperspectral data by SVM. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2297–2307. [Google Scholar] [CrossRef]
  31. Congedo, L. Semi-automatic classification plugin documentation. Release 2016, 4, 29. [Google Scholar]
  32. Nie, F.; Xiang, S.; Jia, Y.; Zhang, C.; Yan, S. Trace ratio criterion for feature selection. In Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (2008), Chicago, IL, USA, 13–17 July 2008; pp. 671–676. [Google Scholar]
  33. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef]
  34. Colkesen, I.; Kavzoglu, T. Selection of Optimal Object Features in Object-Based Image Analysis Using Filter-Based Algorithms. J. Indian Soc. Remote Sens. 2018, 46, 1233–1242. [Google Scholar] [CrossRef]
  35. Gini, C. Variability and mutability, contribution to the study of statistical distribution and relaitons. Studi Economico-Giuricici della R 1912. reviewed in: Light, rj, margolin, bh: An analysis of variance for categorical data. J. Am. Stat. Assoc. 1971, 66, 534–544. [Google Scholar]
  36. Shang, W.; Huang, H.; Zhu, H.; Lin, Y.; Qu, Y.; Wang, Z. A novel feature selection algorithm for text categorization. Expert Syst. Appl. 2007, 33, 1–5. [Google Scholar] [CrossRef]
  37. Wright, S. The interpretation of population structure by F-statistics with special regard to systems of mating. Evolution 1965, 19, 395–420. [Google Scholar] [CrossRef]
  38. Hastie, T.; Tibshirani, R.; Wainwright, M. Statistical Learning with Sparsity: The Lasso and Generalizations; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  39. Liu, J.; Ji, S.; Ye, J. Multi-task feature learning via efficient 12, 1-norm minimization. In Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, Montreal, QC, Canada, 18–21 June 2009; pp. 339–348. [Google Scholar]
  40. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  41. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  42. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  43. Rodriguez-Galiano, V.F.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J.P. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar] [CrossRef]
  44. Pal, M. Random forest classifier for remote sensing classification. Int. J. Remote Sens. 2005, 26, 217–222. [Google Scholar] [CrossRef]
  45. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  46. Díaz-Uriarte, R.; de Andres, S.A. Gene selection and classification of microarray data using random forest. BMC Bioinform. 2006, 7, 3. [Google Scholar] [CrossRef]
  47. Archer, K.J.; Kimes, R.V. Empirical characterization of random forest variable importance measures. Comput. Stat. Data Anal. 2008, 52, 2249–2260. [Google Scholar] [CrossRef]
  48. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  49. Ahmad, M.W.; Mourshed, M.; Rezgui, Y. Trees vs. Neurons: Comparison between random forest and ANN for high-resolution prediction of building energy consumption. Energy Build. 2017, 147, 77–89. [Google Scholar] [CrossRef]
  50. Du, P.; Samat, A.; Waske, B.; Liu, S.; Li, Z. Random forest and rotation forest for fully polarized SAR image classification using polarimetric and spatial features. ISPRS J. Photogramm. Remote Sens. 2015, 105, 38–53. [Google Scholar] [CrossRef]
  51. Chu, C.; Hsu, A.-L.; Chou, K.-H.; Bandettini, P.; Lin, C.; Alzheimer’s Disease Neuroimaging Initiative. Does feature selection improve classification accuracy? Impact of sample size and feature selection on classification using anatomical magnetic resonance images. Neuroimage 2012, 60, 59–70. [Google Scholar] [CrossRef] [PubMed]
  52. Vergara, J.R.; Estévez, P.A. A review of feature selection methods based on mutual information. Neural Comput. Appl. 2014, 24, 175–186. [Google Scholar] [CrossRef]
  53. Kira, K.; Rendell, L.A. A practical approach to feature selection. In Machine Learning Proceedings 1992; Elsevier: Amsterdam, The Netherlands, 1992; pp. 249–256. [Google Scholar]
  54. Li, S.; Zhu, Y.; Feng, J.; Ai, P.; Chen, X. Comparative study of three feature selection methods for regional land cover classification using modis data. In Proceedings of the 2008 Congress on Image and Signal Processing, Sanya, China, 27–30 May 2008; pp. 565–569. [Google Scholar]
  55. Talavera, L. An evaluation of filter and wrapper methods for feature selection in categorical clustering. In Proceedings of the International Symposium on Intelligent Data Analysis, Madrid, Spain, 8–10 September 2005; pp. 440–451. [Google Scholar]
  56. Chrysostomou, K. Wrapper feature selection. In Encyclopedia of Data Warehousing and Mining, 3nd ed.; IGI Global: Hershey, PA, USA, 2009; pp. 2103–2108. [Google Scholar]
  57. Hall, M.A.; Smith, L.A. Feature selection for machine learning: Comparing a correlation-based filter approach to the wrapper. In Proceedings of the Twelfth International FLAIRS Conference, Hamilton, New Zealand, 1–5 May 1999; pp. 235–239. [Google Scholar]
  58. Urbanowicz, R.J.; Meeker, M.; la Cava, W.; Olson, R.S.; Moore, J.H. Relief-based feature selection: Introduction and review. J. Biomed. Inform. 2018, 85, 189–203. [Google Scholar] [CrossRef]
  59. Jain, A.; Zongker, D. Feature selection: Evaluation, application, and small sample performance. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 153–158. [Google Scholar] [CrossRef]
  60. Terblanche, C.; Nänni, I.; Kaplan, H.; Strathie, L.W.; McConnachie, A.J.; Goodall, J.; van Wilgen, B.W. An approach to the development of a national strategy for controlling invasive alien plant species: The case of Parthenium hysterophorus in South Africa. Bothalia 2016, 46, 1–11. [Google Scholar] [CrossRef]
  61. Sonobe, R.; Yamaya, Y.; Tani, H.; Wang, X.; Kobayashi, N.; Mochizuki, K.I. Crop classification from Sentinel-2-derived vegetation indices using ensemble learning. J. Appl. Remote Sens. 2018, 12, 026019. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Location of the study area.
Figure 1. Location of the study area.
Remotesensing 11 01892 g001
Figure 2. Mean f-score learning curve of trace ratio (a), ReliefF (b), Gini-index (c), F-score (d), LS_121(e), LL_121 (f), JMI (g), MIM (h), svm-b (i), dt-f (j) for different feature subsets (Features are made of 75 VIs and 10 Sentinel-2 bands).
Figure 2. Mean f-score learning curve of trace ratio (a), ReliefF (b), Gini-index (c), F-score (d), LS_121(e), LL_121 (f), JMI (g), MIM (h), svm-b (i), dt-f (j) for different feature subsets (Features are made of 75 VIs and 10 Sentinel-2 bands).
Remotesensing 11 01892 g002aRemotesensing 11 01892 g002b
Figure 3. Spatial distribution of Parthenium weed and surrounding land-cover with optimal features from ReliefF on first (a), second (b) and third (c) training set and from full dataset on the first (d), second (e) and third (f) training set.
Figure 3. Spatial distribution of Parthenium weed and surrounding land-cover with optimal features from ReliefF on first (a), second (b) and third (c) training set and from full dataset on the first (d), second (e) and third (f) training set.
Remotesensing 11 01892 g003
Table 1. Training and test dataset combinations for land cover classes.
Table 1. Training and test dataset combinations for land cover classes.
Land-Cover ClassesTraining Set 3 (70%)Test Set 3 (30%)Training Set 2 (50%)Test Set 2 (50%)Training Set 1 (30%)Test Set 1 (70%)Total
Forest703050503070100
Water Body49213535214970
Parthenium Weed63274545276390
Grassland64284646286492
Settlement66294848296695
Table 2. F-score, PA and UA of Parthenium weed using optimal feature subsets yielded by investigated feature selection methods for first training set.
Table 2. F-score, PA and UA of Parthenium weed using optimal feature subsets yielded by investigated feature selection methods for first training set.
Feature Selection MethodPeak Accuracy
PA (%)UA (%)F-score (%)Number of Features*Comput. Time
Trace Ratio7469.871.6560.58
ReliefF74.570.17261.05
Gini-Index74.269.371.31318.54
F-Score74.969.572630.25
LS_12173.471.271.84120.01
LL_12175.270.272.4290.37
JMI74.869.671.93838.58
MIM74.470.172.0 3939.45
SVM-B74.771.172.5260.50
DT-F73.268.570.338573.44
None72.7 69.1 70.485-
*Comput. Time (s): Computational time in seconds.
Table 3. Classification accuracies of other classes using optimal feature subsets yielded by investigated feature selection methods for first training set.
Table 3. Classification accuracies of other classes using optimal feature subsets yielded by investigated feature selection methods for first training set.
ForestWater BodyGrasslandSettlements
*F.S.MPA (%)UA (%)PA (%)UA (%)PA (%)UA (%)PA (%)UA (%)**Kappa Coef.
Trace Ratio89.488.8999758.264.585.881.30.78
ReliefF91.19010097.658.567.685.6790.75
Gini-Index90.890.499.295.457.964.585.482.40.74
F-Score
LS_121
90.388.510096.858.666.685.982.50.76
91.888.81009655.862.88581.70.75
LL_12190.790.810097.858.364.283.880.30.76
JMI91.890.799.496.859.567.285.681.90.77
MIM93.390.899.29559.766.684.683.10.78
SVM-b90.490.397.41006064.484.482.30.75
DT-F91.889.110096.858.764.883.380.80.73
None91.489.710097.458.866.287.181.80.78
*F.S.M.: Feature selection method; **Kappa Coef.: Kappa coefficient.
Table 4. F-score, PA and UA of Parthenium weed using optimal feature subsets yielded by investigated feature selection methods for the second training set.
Table 4. F-score, PA and UA of Parthenium weed using optimal feature subsets yielded by investigated feature selection methods for the second training set.
Feature Selection MethodPeak Accuracy
PA (%)UA (%)F-score (%)Number of FeaturesComput. Time (s)
Trace Ratio7770.373160.60
ReliefF77.169.673.142.14
Gini-Index75.671.173.11031.29
F-score75.768.672570.26
LS_1217668.771.94120.12
LL_12176.67073.190.43
JMI75.571.573.34059.46
MIM75.372.774.0 2561.07
SVM-b77.17274.1410.54
DT-F78.369.673.339934.08
None75.9 69.772.685-
Table 5. Classification accuracies of other classes using optimal feature subsets yielded by investigated feature selection methods for the second training set.
Table 5. Classification accuracies of other classes using optimal feature subsets yielded by investigated feature selection methods for the second training set.
ForestWater BodyGrasslandSettlements
F.S.MPA (%)UA (%)PA (%)UA (%)PA (%)UA (%)PA (%)UA (%)Kappa Coef.
Trace Ratio91.690.299.295.261.372.886.7810.76
ReliefF8991.298.597.661.668.985.681.40.78
Gini-Index91.591.698.995.560.366.886.883.10.79
F-Score90.492.210096.463.571.887.784.10.79
LS_12190.791.498.89559.96986.581.70.78
LL_12190.291.498.896.760.668.885.180.70.80
JMI91.99198.896.162.370.38882.10.77
MIM9091.898.995.564.869.186.682.90.82
SVM-b91.492.8100976470.786.982.70.83
DT-F91.086.010096606883.380.878.1
None87.282.599.797.362.268.687.282.50.79
Table 6. F-score, PA and UA of Parthenium weed using optimal feature subsets yielded by investigated feature selection methods for third training set.
Table 6. F-score, PA and UA of Parthenium weed using optimal feature subsets yielded by investigated feature selection methods for third training set.
Feature Selection MethodPeak Accuracy
PA (%) UA (%)F-score (%)Number of FeaturesComput. Time (s)
Trace Ratio77.274.575.7110.70
Relief807577.273.56
Gini-Index78.574.275.91243.97
F-Score78.573.675.6820.30
LS_12178.676.176.81020.64
LL_12 179.278.278.3102.20
MIM76.972.674.15087.71
SVM-b82.37578.1330.54
DT-F76.672.474.1261346.43
None75.2 71.4 72.685-
Table 7. Classification accuracies of other classes using optimal feature subsets yielded by investigated feature selection methods for third training set.
Table 7. Classification accuracies of other classes using optimal feature subsets yielded by investigated feature selection methods for third training set.
ForestWater BodyGrasslandSettlements
F.S.MPA (%)UA (%)PA (%)UA (%)PA (%)UA (%)PA (%)UA (%)Kappa Coef.
Trace Ratio93.58198.58161.88188.3810.75
ReliefF9281.410081.466.281.489.981.40.77
Gini-Index92.183.199.583.16483.186.383.10.78
F-Score92.295.299.59864.471.291.484.40.79
LS_12192.581.799.581.762.981.78981.70.82
LL_12192.368.898.568.863.468.89068.80.77
JMI90.682.199.582.168.482.190.982.10.80
MIM76.972.69982.963.782.984.882.90.75
SVM-b91.682.799.582.767.282.789.782.70.83
DT-F93.1731007363.77387.9730.82
None90.882.510082.564.382.589.182.50.74

Share and Cite

MDPI and ACS Style

Kiala, Z.; Mutanga, O.; Odindi, J.; Peerbhay, K. Feature Selection on Sentinel-2 Multispectral Imagery for Mapping a Landscape Infested by Parthenium Weed. Remote Sens. 2019, 11, 1892. https://doi.org/10.3390/rs11161892

AMA Style

Kiala Z, Mutanga O, Odindi J, Peerbhay K. Feature Selection on Sentinel-2 Multispectral Imagery for Mapping a Landscape Infested by Parthenium Weed. Remote Sensing. 2019; 11(16):1892. https://doi.org/10.3390/rs11161892

Chicago/Turabian Style

Kiala, Zolo, Onisimo Mutanga, John Odindi, and Kabir Peerbhay. 2019. "Feature Selection on Sentinel-2 Multispectral Imagery for Mapping a Landscape Infested by Parthenium Weed" Remote Sensing 11, no. 16: 1892. https://doi.org/10.3390/rs11161892

APA Style

Kiala, Z., Mutanga, O., Odindi, J., & Peerbhay, K. (2019). Feature Selection on Sentinel-2 Multispectral Imagery for Mapping a Landscape Infested by Parthenium Weed. Remote Sensing, 11(16), 1892. https://doi.org/10.3390/rs11161892

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop