Next Article in Journal
Deep Learning-Based Automatic Detection of Ships: An Experimental Study Using Satellite Images
Previous Article in Journal
Document Liveness Challenge Dataset (DLC-2021)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Clustering Method with a Filter Feature Selection for Hyperspectral Image Classification

1
Department of Atmospheric and Oceanic Sciences, University of California, Los Angeles, CA 90095, USA
2
College of Resources Science and Technology, Beijing Normal University, Beijing 100875, China
J. Imaging 2022, 8(7), 180; https://doi.org/10.3390/jimaging8070180
Submission received: 18 May 2022 / Revised: 21 June 2022 / Accepted: 23 June 2022 / Published: 28 June 2022
(This article belongs to the Section Color, Multi-spectral, and Hyperspectral Imaging)

Abstract

:
Hyperspectral images (HSI) provide ample spectral information of land cover. The hybrid classification method works well for HSI; however, how to select the suitable similarity measures as kernels with the appropriate weights of hybrid classification for HSI is still under investigation. In this paper, a filter feature selection was designed to select the most representative features based on similarity measures. Then, the weights of applicable similarity measures were computed based on coefficients of variation (CVs) of similarity measures. Implementing the similarity measures as the kernels with weights into the K-means algorithm, a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC) was developed. Standard spectral libraries, operative modular imaging spectrometer (OMIS) airborne HSI, airborne visible/infrared imaging spectrometer (AVIRIS) HSI, and Hyperion satellite HSI were selected to inspect the HCW-SSC method. The results showed that the HCW-SSC method has the highest overall accuracy and kappa coefficient (or F1 score) in all experiments (97.5% and 0.974 for standard spectral libraries, 93.21% and 0.9245 for OMIS, 79.24% and 0.8044 for AVIRIS, and 81.23% and 0.7234 for Hyperion) compared to the classification methods (93.75% and 0.958 for standard spectral libraries, 88.27% and 0.8698 for OMIS, 73.12% and 0.7225 for AVIRIS, and 56.34% and 0.3623 for Hyperion) without feature selection and the machine-learning method (68.27% and 0.6628 for AVIRIS, and 51.21% and 0.4255 for Hyperion). The experimental results demonstrate that the new hybrid method performs more effectively than the traditional hybrid method. This also shed a light on the importance of feature selection in HSI classification.

1. Introduction

Remote sensing is a very useful means to monitor and detect land cover/use in a short time frame [1,2,3,4,5]. With the rapid technological development of spectroradiometer and aviation, hyperspectral images (HSI) have been frequently employed to monitor and detect land cover/use change, such as urban mapping, crop health monitoring, and mineral detection [6,7,8]. The primary advantage of HSI is the continuum of the spectrum [9,10,11], which derives continuous spectral curves in the frequency domain or integrated spectral vectors in the reflectance domain.
Classification is a method to differentiate the objects of remote-sensing imagery into different classes, which is an important step to provide a secondary product for mapping, monitoring, and detecting land cover/use change [9,12,13,14,15]. Since HSIs usually have a high spectral and also spatial resolution and the aim of classification of an HSI is to precisely map, monitor, and detect valuable land cover/use change, the classification method of HSIs is different from the classification method for the multispectral image. There are three research topics for the classification method of HSIs: clustering based on graph theory [16,17], clustering by using a machine-learning algorithm [18,19,20], and clustering with the hybrid kernels [21,22,23,24] The clustering based on the graph theory heavily depends on auxiliary space to hold the cache, which needs the graphics processing unit (GPU) and extra random-access memory (RAM). The clustering by using a machine-learning algorithm usually has nonlinear time complexity. When the size of the HSI is large, the performance of this method is slow. A hybrid classification method, which combines two or more similarity measures as the kernels, has the balance of time complexity and space complexity. For example, a dual-clustering-based method was developed to filter the band on HSI [25]. Spectral angle cosine–Euclidean distance (SAC-ED) simply combines two similarity measures [24] The advantage of the hybrid method is that it can analyze the differences in both reflectance and frequency domains. The unknown parts of the hybrid method are the criteria for selecting the kernels and the weights of selected kernels.
Feature selection is a method to find the minimally valuable features that are necessary and sufficient for classification from the raw images [26]. Feature selection can partially or entirely remove the irrelevant or redundant features from images [27]. The aims of feature selection include training classification models faster, reducing the complexity of classification models, improving the accuracy of classification models, and decreasing overfitting. Therefore, feature selection is a good means to provide prior knowledge of HSI and to select the useful similarity measure with the estimated weights for the classification method.
The objective of this paper is to design a feature selection method that can choose the most useful spectral similarity measures with appropriate weights for a hybrid classification method (HCW-SSC) to achieve a better classification result. In Section 2.1, the datasets used in this paper are introduced. In Section 2.2, the workflow of the feature selection and the classification is displayed. In Section 2.3, the indicators used to evaluate the classification results are shown. In Section 2.4, the implementation of this method by using Python is presented. In Section 3, the classification result of this method is compared to the classification methods with no feature selection and machine-learning method. In Section 4, the importance of feature selection in the HCW-SSC classification method and the time complexity of HCW-SSC are discussed. Moreover, the results of the HCW-SSC method from this paper were compared to other papers. In Section 5, the conclusions and contributions of this paper are exhibited. The contribution of this paper is that a hybrid clustering method with a filter feature selection was developed for HSI. The filter feature selection is designed to select the most representative features based on similarity measures and calculate the suitable similarity measures as kernels with the appropriate weights of hybrid classification for hyperspectral imaging.

2. Data and Method

2.1. Data

Four datasets were employed to test the HCW-SSC method, including standard spectral libraries, operative modular imaging spectrometer (OMIS) (CSA, Shanghai, China) airborne hyperspectral image, airborne visible/infrared imaging spectrometer (AVIRIS) (JPL, Pasadena, CA, USA) hyperspectral image, and Hyperion satellite (JPL, Pasadena, CA, USA) hyperspectral image. The details of the data used in this paper are shown in Table 1. Four preprocessing steps were applied to the HSIs before using them in the classification methods: radiometric calibration was applied; bad and noisy bands were deleted; a minimum noise fraction rotation (MNF) was used to remove the smile/frown effect [28], which causes a significant cross-track and nonlinear disturbances on spectral curves and spectral vectors [29]; and principal component analysis (PCA) was employed to reduce the high-correlated bands [30].

2.1.1. Standard Spectral Libraries

Four standard spectral libraries (Table 2) were used to test the accuracy of the HCW-SSC, including United States Geological Survey (USGS) vegetation and mineral libraries [31], and Chris Elvidge green and dry vegetation libraries [32].
In order to evaluate and validate the accuracy of this method, 80 groups of spectra combinations were selected. For one group of test data, there are three spectral profiles. Two of them belong to the same category and another belongs to a different category based on the description from the spectral library. For example, two spectra of Sporobolus and one spectrum of Andropogon virginicus were selected in one group based on the description of USGS vegetation library. The spectra of Sporobolus are similar and different from Andropogon virginicus.

2.1.2. OMIS Hyperspectral Image

OMIS is an airborne spectroradiometer, which is developed by the Shanghai Institute of Technical Physics (SITP), Chinese Academy of Sciences (CAS), China [33]. OMIS uses a whiskbroom system to cover the spectral region of visible, infrared, and thermal infrared (400–12,500 nm). It has selectable 64 or 128 bands, 2.8 m spatial resolution, 10 nm spectral resolution, and an IFOV (instantaneous field of view) of 0.003 rad. The OMIS data can be found at http://www.scidb.cn/en (accessed on 8 June 2012).
The study area is on Xiaotangshan, Beijing, China. The hyperspectral imagery was obtained on 11 April 2010 (Figure 1). Fifty-one bands between 455.7 nm to 1000.4 nm were selected. The study area is a precision agriculture experimental field, which contains different kinds of wheat [24]. The baseline map was made by field measurement on the experimental field, which was recorded as a shapefile.

2.1.3. AVIRIS Hyperspectral Image

AVIRIS is an airborne spectroradiometer, which is developed by the Jet Propulsion Laboratory (JPL), USA. It has 224 bands ranging from 400 to 2500 nm with a 10 nm bandwidth. AVIRIS is a pushbroom instrument with an 11km-wide swath perpendicular to the satellite motion. The spatial resolution is 20 m [34]. The AVIRIS data can be found at https://aviris.jpl.nasa.gov/dataportal/ (accessed on 9 May 2015).
The study area is in Doña Ana County, NM, USA. The hyperspectral imagery was obtained on 23 May 2011 (Figure 2). One hundred and seventy bands were selected. The study area is in the northern part of San Ysidro city, which contains tree nuts, pepper, grains, and corn [35,36]. The baseline map was made by the high-resolution classification result of the National Agriculture Imagery Program (NAIP) on 25 July 2011. The data can be found at https://nrcs.app.box.com/v/naip (accessed on 9 May 2015).

2.1.4. Hyperion Satellite Hyperspectral Image

Hyperion is a satellite-based spectroradiometer on Earth Observing One (EO-1), which was developed by the National Aeronautics and Space Administration (NASA), USA. It has 242 bands ranging from 357 nm to 2576 nm with a 10 nm bandwidth. Hyperion is a pushbroom instrument with a 7.5 km-wide swath perpendicular to the satellite motion. The spatial resolution is 30 m [37]. The EO-1 data can be found at https://search.earthdata.nasa.gov/search (accessed on 9 February 2015).
The study area is in Qinghai Lake basin, Qinghai province, China. The hyperspectral imagery was obtained on 4 September 2013 (Figure 3). One hundred and seventy-five bands were selected. The study area is in the northern part of the Qinghai Lake basin, which contains winter wheat and rape [38,39]. The baseline map was made by the high-resolution classification result of the Rapideye image on 19 July 2013. The data can be found at https://www.planet.com/products/planet-imagery/ (accessed on 9 May 2015).

2.2. Method

The general workflow has four parts (Figure 4): to extract the features from preprocessed HSI by using Euclidean distance (ED), spectral angle cosine (SAC), spectral correlation coefficient (SCC), and spectral information divergence (SID) similarity measures; to find the minimum absolute Pearson correlation on any of the two (the lines in the same color represents one pair, Figure 4) out of four features; to select the related similarity measures of the features of the minimum absolute Pearson correlation and calculate the coefficients of variation (CVs) of these two similarity measures; and to build a K-means clustering method that contains these similarity measures as the kernels with their weights.

2.2.1. Similarity Measure

A similarity measure is a function to quantify the similarity between two objects [40]. The rule of similarity measure is “the higher similarity measure value is calculated, the smaller similarity these two objects are”. For example, as shown in Figure 5, D1 (the similarity measure between object A1 and object A2) is smaller than D2 (the similarity measure between object A2 and object B), which means object A1 and object A2 have a high probability to belong to the same category but object B is different from them. The commonly used similarity measures include distance measure (ED) to quantify the difference of brightness between pixel Xn and pixel Xn+1 in the frequency domain [41]; consistency measure (SCC) to compare the angle of spectral vectors from pixel Xn and pixel Xn+1 in the reflectance domain [42]; dependence measure to measure the difference of shape of spectral curves (SAC) between pixel Xn and pixel Xn+1 in the frequency domain [43]; and information measure (SID) to compute the information gain from pixel Xn and pixel Xn+1 in the reflectance domain [44].
The similarity measure is used in two places in this workflow: extracting features from HSI and using them as kernels in the K-means clustering. The difference between these two usages is that extracting feature uses the mean value (μ) of the whole HSI as the competitor but using them as kernels in the K-means clustering uses the mean value of jth centroids (μj) of the HSI as the competitor.

2.2.2. Feature Selection

A feature selection technique aims to remove irrelevant or redundant features and keep relevant features for a dataset [26]. There are three ways to carry out the feature selection: filter, wrapper, and embedded [45]. In this study, filter feature selection was used for the classification method. The Pearson correlation was used to check the correlation between any of the two features,
ρ = ( a i a ¯ ) ( b i b ¯ ) ( a i a ¯ ) 2 ( b i b ¯ ) 2 ,
a and b represent the vector belonging to the different features. Then, the minimum of the absolute Pearson correlation ( min ( | ( ρ 1 ρ n ) | ) ) was found. It represents that the features have the most noncorrelation, which means they have the least duplicated information. After finding the two similarity measures, a CV was used to compute the variation in each similarity measure, which is similar to ANOVA. If a feature has a large variation, that means this feature is easy to be recognized in this dataset. The equation to calculate CV is
C V S = 0 m σ ( S ( X n , X n + 1 ) ) μ ( S ( X n , X n + 1 ) ) ¯ ,
CVS is the CV for a similarity measure. σ ( S ( X n , X n + 1 ) ) is the standard deviation of a similarity measure between two pixels (Xn and Xn+1). μ ( S ( X n , X n + 1 ) ) is the mean of a similarity measure between two pixels (Xn and Xn+1). While calculating CVs for different features, the ratio between each CV was also computed because the CV is dimensionless. The ratio was used as the weight of each similarity measure in the K-means clustering method in the next section. The ratio is calculated by one restriction as follows
{ C V S 1 C V S 2 = w 1 w 2 w 1 + w 2 = 1 ,
where CVS1 and CVS2 represent the CV of two similarity measures and w1 and w2 are the weights of these two.

2.2.3. Hybrid Classification Method

A hybrid classification method is a classification method that contains two or more kernels with different similarity measures [21,23]. In this paper, K-means was used as the clustering algorithm for the hybrid classification method [46,47]. The algorithm is
  • Specify the number clusters as K;
  • Randomly select K centroids among samples;
  • Keep iterating the following equations until no change to the centroids.
For one sample x i , calculating the belonging cluster of this sample,
C i = { arg min X i μ j 2 0 , otherwise ,   if   j   belongs   to   the   cluster ,
j is the jth centroids and it is [0, K]. I is the ith sample and it is [0, the number of samples]. xi represents a sample. ci represents a cluster that xi is the most likely belonging to.   μ j   represents the centroid of a cluster. arg min is an argument of the minimum, which means a function attains its minimum.
For one cluster μ j , recalculating the centroid of this cluster,
μ j = 1 | c i | x i ,     x i c i .
For the HCW-SSC method, the calculation of the belonging cluster of one point (Equation (4)) is changed to:
arg m i n   | | w 1 × similarity   measure 1 + w 2 × similarity   measure 2 | | 2  
w 1 = ( C V S 1 C V S 1 + C V S 2 ) / μ S 1 ¯ ,
w 2 = ( C V S 2 C V S 1 + C V S 2 ) / μ S 2 ¯ ,
w1 and w2 were calculated from feature selection in the previous step. μ S ¯ and μ S + 1 ¯ are the means of similarity measures to remove the dimension. For example, ED and SAC were selected as the most useful features, the equation will be
arg m i n   | | w 1 X i X j 2 + w 2 ( 1 X i X j X i 2 X j 2 ) | | 2 .

2.3. Evaluation Indicators

For the test of standard spectral libraries, the overall accuracy and F1 score were used as the evaluation indicator [48]. For the tests of HSIs, the overall accuracy and kappa coefficient were used as the evaluation indicator [49,50]. In a confusion matrix, columns represent a true number of pixels in each class, and rows represent the predicted number of pixels in each class. The matrix is square and all numbers of the correct classified pixels are along the upper-left to lower-right diagonal. The overall accuracy is the ratio of the number of correctly classified pixels to the total number of pixels [2]. The equation is
O v e r a l l A c c u r a c y = t h e   s u m   o f   t h e   n u m b e r s   o n   t h e   d i a g o n a l T o t a l ,
Kappa is designed to compare the accuracy of a classification method to the accuracy of a random selection [48]. Kappa is dimensionless and the value is from −1 to 1. The equation is
K a p p a = t o t a l A c c u r a c y r a n d o m A c c u r a c y 1 r a n d o m A c c u r a c y ,
t o t a l A c c u r a c y = T P + T N T o t a l ,
r a n d o m A c c u r a c y = ( T N + F P ) × ( T N + F N ) + ( F N + T P ) × ( F P + T P ) T o t a l × T o t a l .
TP is true positive, which is the number of pixels in a given class that were classified correctly. TN is true negative, which is the number of pixels in a given class that were not classified correctly. FP is false positive, which is the number of pixels that were predicted to be in a given class but do not belong to that class. FN is false negative, which is the number of pixels that were not predicted to be in a given class but do belong to that class.
F1 score is designed to check the balance of precision and recall. The equation is
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l ,
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N .  

2.4. Implementation

The number of clusters relies on the elbow method, which calculated distortion and inertia per number of clusters. For testing of standard spectral libraries, the number of clusters is 2; the maximum iteration of the K-means clustering method with ED, SAC, ED-SAC, and HCW-SSC similarity measure is 2; and the minimum change threshold is 2%. For testing of OMIS HSI, the number of clusters is 7; the maximum iteration of K-means clustering method with SAC, SID, SID-SAC, and HCW-SSC similarity measure is 6; and the minimum change threshold is 5%. For testing of AVIRIS HSI, the number of clusters is 5; the maximum iteration of K-means clustering method of ED, SAC, ED-SAC, and HCW-SSC similarity measure is 5; and the minimum change threshold is 5%. For testing of Hyperion HSI, the number of clusters is 11; the maximum iteration of K-means clustering method of SID, SCC, SID-SCC, and HCW-SSC similarity measure is 9; and the minimum change threshold is 5%. For random forest (RF) classification in both AVIRIS and Hyperion imagery, the number of trees in the forest is 100 and the maximum depth of a tree is 5 (the RF classifier was directly applied on the HSI, which was not applied on feature selection).
The preprocess of HSI was made by ENVI 5.3 (L3 Harris Technologies, Boulder, CO, USA). The codes of feature selection and hybrid classification were written in Python 3.8 with GDAL and scikit-learn packages [51,52].

3. Results

3.1. Test Based on Standard Spectral Libraries

ED and SAC were selected as the kernels of K-means in the HCW-SSC method. The single-kernel ED, single-kernel SAC, and the unweighted ED and SAC kernels were also implemented in the K-means as the comparisons. The HCW-SSC method resulted in the highest overall accuracy, followed by ED-SAC (Table 3). The single ED produced the lowest overall accuracy.

3.2. Test Based on OMIS HSI

SAC and SID were selected as the kernels of K-means in the HCW-SSC method. The single-kernel SAC, the single-kernel SID, and the unweighted SID-SAC kernels were also implemented in the K-means as the comparisons. The overall accuracy and kappa coefficient were calculated to assess the four types of spectral similarity measures (Table 4). The HCW-SSC was the highest, whether in overall accuracy or kappa coefficient. The overall accuracy was 93.21% and the kappa coefficient was 0.9245.
Compared to the classification maps of field measurement, the HCW-SSC method produced the best classification result, followed by SID-SAC (Figure 6). Single SAC and SID cannot effectively classify several intercropping or interbreeding wheat areas, because the spectra of winter wheat in different colonies were similar. Therefore, only using a single similarity measure of spectral curves and spectral vectors cannot distinguish the winter wheat of different colonies. SID-SAC reflected a better classification effect as a whole, but an “island problem” about classification exists in some centralized winter wheat areas. It did not comply with the actual growing conditions of the wheat. The HCW-SSC method has the less-misclassified pixels and the cleanest boundary.

3.3. Test Based on AVIRIS HSI

ED and SAC were selected as the kernels of K-means in the HCW-SSC method. The single kernel ED, the single kernel SAC, and the unweighted ED-SAC kernels were also implemented in the K-means as the comparisons. The total classification accuracy and kappa coefficient were calculated to assess the four types of spectral similarity measures (Table 5). The HCW-SSC was the highest, whether in overall accuracy or kappa coefficient. Overall accuracy was 79.24% and the kappa coefficient was 0.8044.
Compared to the baseline classification maps, the HCW-SSC method produced the best classification result, followed by ED-SAC (Figure 7). Single ED or SAC has misclassified some land covers such as pepper and pecan, especially on the mixing area of pepper and pecan. RF has less misclassification since it is an ensemble model, which can select the best solution among all solutions.

3.4. Test Based on Hyperion HSI

SID and SCC were selected as the kernels of K-means in the HCW-SSC method. The single-kernel SID, the single-kernel SCC, and the unweighted SID-SCC kernels were also implemented in the K-means as the comparisons. The total classification accuracy and kappa coefficient were calculated to assess the four types of spectral similarity measures (Table 6). The HCW-SSC was the highest, whether in overall accuracy or kappa coefficient. Overall accuracy was 81.23% and kappa coefficient was 0.7234.
Compared to the baseline classification maps, the HCW-SSC method produced the best classification result, followed by SCC (Figure 8). Single SID or SCC has an obvious “island problem”, which means the spectrum of the adjunct pixel surrounding the “island” is similar and it is hard to use only one similarity measure to classify. RF and SID-SCC almost cannot classify the rape from winter wheat and rice, which was caused by the overfitting by the algorithm.

4. Discussion

In this paper, the HCW-SSC method only chose features from four basic similarity measures and the weight restriction is naive. In the future, more similarity measures can be used to extract the features from HSI and then be implemented into K-means as the kernels. In this section, the importance of feature selection in HCW-SSC and the performance of HCW-SSC are discussed.

4.1. The Importance of Feature Selection in HCW-SSC

4.1.1. Select the Suitable Similarity Measures as Kernels

HCW-SSC was compared to the unweighted hybrid kernel that is composed of selected similarity measures from filter feature selection. However, the classification results of these hybrid kernels that are composed of non-selected similarity measures were not shown. In Table 7, these classification results are shown. The unweighted SID-SAC has the highest accuracy among all unweighted hybrid kernels (ED-SCC, ED-SAC, SCC-SID, SAC-SCC, and ED-SID). That means if two or more similarity measures were randomly implemented into a kernel, the classification result of it may not become more accurate. Therefore, choosing appropriate kernels by feature selection for a hybrid classification method is important.

4.1.2. Calculate the Weights for the Kernels

The unweighted hybrid method directly multiplies two similarity features. A different similarity measure has a different unit, so it led to the unbalanced usage of the two kernels. HCW-SSC took the weight from CV and used CV divided by mean to make the similarity measure unitless. The weight helps balance the contribution of two similarity measures for the K-means clustering. Therefore, the hybrid classification method (HCW-SSC) with a feature selection technique generates a better result than the one without it. The weights of the HCW-SSC method based on the standard spectral libraries are shown in Table 8. The mean weight of Euclidean distance fluctuated at 0.31, whereas the mean weight of spectral angle cosine fluctuated at 0.69. SAC contributes more than ED in this dataset for the classification. However, in the ED-SAC method, the absolute value of ED is much larger than SAC, so ED takes more contribution than SAC, which leads to more misclassifications.

4.2. The Time Complexity of HCW-SSC Method

Another advantage of this method is its performance. Based on Figure 9, HCW-SSC has a similar time consumption to the traditional method (ED-SAC). Although HCW-SSC has the filter feature selection, the filter feature selection takes linear time to visit all pixels; therefore, the total time complexity is still linear (O(n), the operation of multiply and sum has linear time complexity). However, the time complexity of a random forest is n-square (O(n2)) because each new tree in the forest will bring n more leaves. When the image becomes larger, the time consumption of random forest becomes much larger.

4.3. Compared to Other Methods

There are three research topics for the classification method of HSIs: clustering based on graph theory [16,17], clustering by using a machine-learning algorithm [18,19], and clustering with the hybrid kernels [21,22,23,24]. The HCW-SSC represents the clustering with the hybrid kernels. Compared to the non-filter-selection method, the new method performs much better (Table 3, Table 4, Table 5 and Table 6). In order to compare the method of clustering based on graph theory and clustering by using a machine-learning algorithm, three papers were selected. Although the HSI data in these three papers are different from this paper, they indirectly show the accuracy of the HCW-SSC method.
In the paper by Meng et al., 2017 [16], they used a semisupervised kernel based on graph theory with K-means clustering to classify three HSIs. The result is in Table 9. For AVIRIS and airborne sensor HSIs, the HCW-SSC method has a similar overall accuracy (79.24% and 93.21%) with this semisupervised kernel (80.37% and 99.17%). The semisupervised kernel is based on the graph theory, which highly depends on the performance of the GPU and the storage of RAM. However, the HCW-SSC method does not need GPU and extra RAM.
In the paper by Li et al., 2019 [20], they used four machine-learning algorithms and six convolutional neural network (CNN)-related deep-learning algorithms to classify three HSIs. The result is in Table 10. For AVIRIS and airborne sensor HSIs, the HCW-SSC method has a similar overall accuracy (79.24% and 93.21%) to the mean overall accuracy (78.79% and 95.80%) of the ten methods in Li’s paper. The HCW-SSC has linear time complexity but the machine-learning method, especially the deep-learning method, has polynomial or exponential time complexity.
In the paper by Yuan et al., 2016 [25], they used four machine-learning algorithms to classify three HSIs. The result is in Table 11. For AVIRIS HSIs, the HCW-SSC method has a similar overall accuracy (79.24%) to the mean overall accuracy (74.94%) of the four methods in Yuan’s paper.

5. Conclusions

This paper raised a hybrid classification method that can utilize the features chosen from filter feature selection on hyperspectral images, which is called HCW-SSC. Standard spectral libraries, OMIS airborne hyperspectral image, AVIRIS hyperspectral image, and Hyperion satellite hyperspectral image were used to inspect the accuracy of the HCW-SSC classification method. The results showed that the HCW-SSC method has the highest overall accuracy and kappa coefficient (or F1 score) in all experiments (97.5% and 0.974 for standard spectral libraries, 93.21% and 0.9245 for OMIS, 79.24% and 0.8044 for AVIRIS, and 81.23% and 0.7234 for Hyperion). The HCW-SSC method exhibits a better spectral-recognition capacity compared to the classification method with one type of spectral characteristic or simply combines two spectral characteristics and a machine-learning method. Compared to the other two research directions of HSI classification, HCW-SSC has a balance of time complexity and space usage but obtains similar accuracy. This paper can be a useful reference for how feature selection optimizes the traditional hybrid classification methods and improves the accuracy of classification methods for hyperspectral remote-sensing data. This also sheds a light on the importance of feature selection in HSI classification. In the future, this method may be applied to time-series remote-sensing data, which has the time-series curves of land cover rather than spectral curves.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

United States Geological Survey (USGS) vegetation and mineral libraries can be found in Clark et al., 1993 and Chris Elvidge green and dry vegetation libraries can be found in Elvidge 1990. The OMIS data can be found at http://www.scidb.cn/en (accessed on 8 June 2012). The AVIRIS data can be found at https://aviris.jpl.nasa.gov/dataportal/ (accessed on 9 May 2015). The EO-1 data can be found at https://www.planet.com/products/planet-imagery/ (accessed on 9 February 2015). NAIP data can be found at https://nrcs.app.box.com/v/naip (accessed on 9 May 2015). Rapideye data can be found at https://www.planet.com/products/planet-imagery/ (accessed on 9 May 2015).

Acknowledgments

We thank Beijing Normal University for providing the field data on Xiaotangshan, Beijing, China.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Peters, D.P.; Okin, G.S.; Herrick, J.E.; Savoy, H.M.; Anderson, J.P.; Scroggs, S.L.; Zhang, J. Modifying connectivity to promote state change reversal: The importance of geomorphic context and plant–soil feedbacks. Ecology 2020, 101, e03069. [Google Scholar] [CrossRef]
  2. Warner, T.A.; Foody, G.M.; Nellis, M.D. The SAGE Handbook of Remote Sensing; SAGE Publications: Newbury Park, CA, USA, 2009; ISBN 978–1-4129–3616–3. [Google Scholar]
  3. Zhang, J. A new ecological-wind erosion model to simulate the impacts of aeolian transport on dryland vegetation patterns. Acta Ecol. Sin. 2020, 41, 304–317. [Google Scholar] [CrossRef]
  4. Zhang, J.; Okin, G.S.; Zhou, B. Assimilating optical satellite remote sensing images and field data to predict surface indicators in the Western US: Assessing error in satellite predictions based on large geographical datasets with the use of ma-chine learning. Remote Sens. Environ. 2019, 233, 111382. [Google Scholar] [CrossRef]
  5. Zhou, B.; Okin, G.S.; Zhang, J. Leveraging Google Earth Engine (GEE) and machine learning algorithms to incorporate in situ measurement from different times for rangelands monitoring. Remote Sens. Environ. 2020, 236, 111521. [Google Scholar] [CrossRef]
  6. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  7. Borengasser, M.; Hungate, W.S.; Watkins, R. Hyperspectral Remote Sensing: Principles and Applications; CRC Press: Los Angeles, CA, USA, 2007. [Google Scholar]
  8. Okin, G.S.; Sala, O.E.; Vivoni, E.R.; Zhang, J.; Bhattachan, A. The interactive role of wind and water in functioning of drylands: What does the future hold? Bioscience 2018, 68, 670–677. [Google Scholar] [CrossRef]
  9. Bhattachan, A.; Okin, G.S.; Zhang, J.; Vimal, S.; Lettenmaier, D.P. Characterizing the Role of Wind and Dust in Traffic Accidents in California. GeoHealth 2019, 3, 328–336. [Google Scholar] [CrossRef] [Green Version]
  10. Chang, C.-I.; Chiang, S.-S. Anomaly detection and classification for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2002, 40, 1314–1325. [Google Scholar] [CrossRef] [Green Version]
  11. Dundar, M.; Landgrebe, D. Toward an Optimal Supervised Classifier for the Analysis of Hyperspectral Data. IEEE Trans. Geosci. Remote Sens. 2004, 42, 271–277. [Google Scholar] [CrossRef]
  12. Jiang, N.; Zhu, W.; Mou, M.; Wang, L.; Zhang, J. A phenology-preserving filtering method to reduce noise in NDVI time series. IEEE Int. Geosci. Remote Sens. Symp. 2012, 2384–2387. [Google Scholar] [CrossRef]
  13. Li, X.; Guo, W.; Li, S.; Zhang, J.; Ni, X. The different impacts of the daytime and nighttime land surface temperatures on the alpine grassland phenology. Ecosphere 2021, 12, e03578. [Google Scholar] [CrossRef]
  14. Liu, J.; Zhu, W.; Sun, G.; Zhang, J.; Jiang, N. Endmember abundance calibration method for paddy rice area extrac-tion from MODIS data based on independent component analysis. Trans. Chin. Soc. Agric. Eng. 2012, 28, 103–108. [Google Scholar]
  15. Running, S.W.; Loveland, T.R.; Pierce, L.L.; Nemani, R.; Hunt, E. A remote sensing based vegetation classification logic for global land cover analysis. Remote Sens. Environ. 1995, 51, 39–48. [Google Scholar] [CrossRef]
  16. Meng, Z.; Merkurjev, E.; Koniges, A.; Bertozzi, A.L. Hyperspectral image classification using graph clustering methods. Image Processing Line 2017, 7, 218–245. [Google Scholar] [CrossRef] [Green Version]
  17. Hossam, M. High Performance Hyperspectral Image Classification using Graphics Processing Units. arXiv 2021, arXiv:2106.12942. [Google Scholar]
  18. Yao, W.; Lian, C.; Bruzzone, L. ClusterCNN: Clustering-Based Feature Learning for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1991–1995. [Google Scholar] [CrossRef]
  19. Fang, B.; Li, Y.; Zhang, H.; Chan, J.C.-W. Collaborative learning of lightweight convolutional neural network and deep clustering for hyperspectral image semi-supervised classification with limited training samples. ISPRS J. Photogramm. Remote Sens. 2020, 161, 164–178. [Google Scholar] [CrossRef]
  20. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  21. Aci, M.; Inan, C.; Avci, M. A hybrid classification method of k nearest neighbor, Bayesian methods and genetic algorithm. Expert Syst. Appl. 2010, 37, 5061–5067. [Google Scholar] [CrossRef]
  22. Chan, J.C.-W.; Paelinckx, D. Evaluation of Random Forest and Adaboost tree-based ensemble classification and spec-tral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sens. Environ. 2008, 112, 2999–3011. [Google Scholar] [CrossRef]
  23. Perronnin, F.; Larlus, D. Fisher vectors meet neural networks: A hybrid classification architecture. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3743–3752. [Google Scholar]
  24. Zhang, J.; Zhu, W.; Dong, Y.; Jiang, N.; Pan, Y. A spectral similarity measure based on Changing-Weight Combination Method. Acta Geod. Cartogr. Sin. 2013, 42, 418–424. [Google Scholar]
  25. Yuan, Y.; Lin, J.; Wang, Q. Dual-Clustering-Based Hyperspectral Band Selection by Contextual Analysis. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1431–1445. [Google Scholar] [CrossRef]
  26. Dash, M.; Liu, H. Feature selection for classification. Intell. Data Anal. 1997, 1, 131–156. [Google Scholar] [CrossRef]
  27. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 1–45. [Google Scholar] [CrossRef]
  28. Kruse, F.; Boardman, J.; Huntington, J. Comparison of airborne hyperspectral data and eo-1 hyperion for mineral mapping. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1388–1400. [Google Scholar] [CrossRef] [Green Version]
  29. Goodenough, D.G.; Dyk, A.; Niemann, K.O.; Pearlman, J.S.; Chen, H.; Han, T.; Murdoch, M.; West, C. Processing hyperion and ali for forest classification. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1321–1331. [Google Scholar] [CrossRef]
  30. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  31. Clark, R.N.; Swayze, G.A.; King, T.V.; Gallagher, A.J.; Calvin, W.M. The US Geological Survey, Digital Spectral Reflectance Library. Version 1: 0.2 to 3.0 Microns. 1993. Available online: https://pubs.usgs.gov/of/1993/0592/report.pdf (accessed on 9 May 2022).
  32. Elvidge, C.D. Visible and near infrared reflectance characteristics of dry plant materials. Int. J. Remote Sens. 1990, 11, 1775–1795. [Google Scholar] [CrossRef]
  33. Du, P.-J.; Kun, T.; Su, H.-J. Feature extraction for target identification and image classification of OMIS hyperspectral image. Min. Sci. Technol. 2009, 19, 835–841. [Google Scholar] [CrossRef]
  34. Macenka, S.A.; Chrisp, M.P. Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) Spectrometer Design and Per-Formance. In IMAGING Spectroscopy II; International Society for Optics and Photonics: San Francisco, CA, USA, 1987; pp. 32–43. [Google Scholar]
  35. Dentzman, K.; Pilgeram, R.; Lewin, P.; Conley, K. Queer Farmers in the 2017 US Census of Agriculture. Soc. Nat. Resour. 2020, 34, 227–247. [Google Scholar] [CrossRef]
  36. Zhang, J.; Okin, G.S.; Zhou, B.; Karl, J.W. UAV-derived imagery for vegetation structure estimation in rangelands: Validation and application. Ecosphere 2021, 12, e03830. [Google Scholar] [CrossRef]
  37. Beck, R. EO-1 User Guide-Version 2.3. Satellite Systems Branch; USGS Earth Resources Observation Systems Data Center (EDC): Sioux Falls, SD, USA, 2003. [Google Scholar]
  38. Rhode, D.; Haiying, Z.; Madsen, D.B.; Xing, G.; Brantingham, P.J.; Haizhou, M.; Olsen, J.W. Epipaleolithic/early neo-lithic settlements at Qinghai Lake, western China. J. Archaeol. Sci. 2007, 34, 600–612. [Google Scholar] [CrossRef]
  39. Zhang, J.; Zhu, W.; Wang, L.; Jiang, N. Evaluation of similarity measure methods for hyperspectral remote sensing data. In Proceedings of the 2012 IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012; pp. 4138–4141. [Google Scholar] [CrossRef]
  40. Santini, S.; Jain, R. Similarity measures. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 871–883. [Google Scholar] [CrossRef] [Green Version]
  41. Danielsson, P.-E. Euclidean distance mapping. Comput. Graph. Image Processing 1980, 14, 227–248. [Google Scholar]
  42. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Proceedings of the Summaries 3rd Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 23–26 January 1992; pp. 147–149. [Google Scholar]
  43. De Carvalho, O.A.; Meneses, P.R. Spectral correlation mapper (SCM): An improvement on the spectral angle mapper (SAM). In Proceedings of the Summaries of the 9th JPL Airborne Earth Science Workshop, JPL Publication 00–18, Pasadena, CA, USA, 23–25 February 2000; JPL Publication: Pasadena, CA, USA, 2000. [Google Scholar]
  44. Chang, C.-I. An information-theoretic approach to spectral variability, similarity, and discrimination for hyperspectral image analysis. IEEE Trans. Inf. Theory 2000, 46, 1927–1932. [Google Scholar] [CrossRef] [Green Version]
  45. Kira, K.; Rendell, L.A. A Practical approach to feature aelection. In Machine Learning Proceedings; Elsevier: Amsterdam, The Netherlands, 1992; pp. 249–256. [Google Scholar] [CrossRef]
  46. Hamerly, G.; Elkan, C. Learning the k in k-means. Adv. Neural Inf. Processing Syst. 2004, 16, 281–288. [Google Scholar]
  47. Likas, A.; Vlassis, N.; Verbeek, J.J. The global k-means clustering algorithm. Pattern Recognit. 2003, 36, 451–461. [Google Scholar] [CrossRef] [Green Version]
  48. Jones, H.G.; Vaughan, R.A. Remote Sensing of Vegetation: Principles, Techniques, and Applications; Oxford University Press: Oxford, UK, 2010. [Google Scholar]
  49. Campbell, J.B.; Wynne, R.H. Introduction to Remote Sensing; Guilford Press: New York, NY, USA, 2011. [Google Scholar]
  50. Jensen, J.R. Remote Sensing of the Environment: An Earth Resource Perspective; Pearson Education India: New Delhi, India, 2009. [Google Scholar]
  51. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Du-bourg, V. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  52. Warmerdam, F. The Geospatial Data Abstraction Library. In Open Source Approaches in Spatial Data Handling; Springer: Berlin/Heildelberg, Germany, 2008; pp. 87–104. [Google Scholar]
Figure 1. The study area at Xiaotangshan, Beijing, China (OMIS true color synthesis image, R = 699.2 nm, G = 565.4 nm, B = 465.0 nm).
Figure 1. The study area at Xiaotangshan, Beijing, China (OMIS true color synthesis image, R = 699.2 nm, G = 565.4 nm, B = 465.0 nm).
Jimaging 08 00180 g001
Figure 2. The study area at Doña Ana County, NM, USA.
Figure 2. The study area at Doña Ana County, NM, USA.
Jimaging 08 00180 g002
Figure 3. The study area at Qinghai Lake basin, Qinghai Province, China.
Figure 3. The study area at Qinghai Lake basin, Qinghai Province, China.
Jimaging 08 00180 g003
Figure 4. The evaluation scheme for the HCW-SSC method (Euclidean distance (ED), spectral angle cosine (SAC), spectral correlation coefficient (SCC), and spectral information divergence (SID)).
Figure 4. The evaluation scheme for the HCW-SSC method (Euclidean distance (ED), spectral angle cosine (SAC), spectral correlation coefficient (SCC), and spectral information divergence (SID)).
Jimaging 08 00180 g004
Figure 5. A schematic diagram showing the rule of the similarity measure.
Figure 5. A schematic diagram showing the rule of the similarity measure.
Jimaging 08 00180 g005
Figure 6. Classification maps of four different classification methods and the field measurement (spectral angle cosine (SAC), spectral information divergence (SID), spectral information divergence–spectral angle cosine (SID-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Figure 6. Classification maps of four different classification methods and the field measurement (spectral angle cosine (SAC), spectral information divergence (SID), spectral information divergence–spectral angle cosine (SID-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Jimaging 08 00180 g006
Figure 7. Classification maps of five different classification methods and baseline map (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Figure 7. Classification maps of five different classification methods and baseline map (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Jimaging 08 00180 g007
Figure 8. Classification maps of five different classification methods and baseline map (spectral information divergence (SID), spectral correlation coefficient (SCC), spectral information divergence–spectral correlation coefficient (SID-SCC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Figure 8. Classification maps of five different classification methods and baseline map (spectral information divergence (SID), spectral correlation coefficient (SCC), spectral information divergence–spectral correlation coefficient (SID-SCC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Jimaging 08 00180 g008
Figure 9. The time consumption of different kernels is based on different image sizes (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC), and random forest (RF)).
Figure 9. The time consumption of different kernels is based on different image sizes (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC), and random forest (RF)).
Jimaging 08 00180 g009
Table 1. The details of data used in this paper.
Table 1. The details of data used in this paper.
NameSpatial ResolutionSpectral ResolutionSpectral DomainLocation (Based on WGS 84)
Standard spectral librariesNA (fieldwork)0.2 nm or 1 nm400–2500 nmPasadena, CA, USA
OMIS2.8 m10 nm400–12,500 nmBeijing, China (40°10′57″ N, 116°26′36″ E)
AVIRIS20 m10 nm357–2576 nmNM, USA (32°28′16.6″ N, 106°54′23.7″ W)
Hyperion30 m10 nm426–2356 nmQianhai, China (37°25′54″ N, 100°11′44″ E)
Table 2. The details of spectral libraries.
Table 2. The details of spectral libraries.
Spectral LibraryThe Extent of WavelengthSpectral Resolution in the Visible RegionSpectral Resolution in the Infrared Region
USGS vegetation0.4–2.5 μm0.2 nm0.5 nm
USGS mineral0.4–2.5 μm0.2 nm0.5 nm
Chris Elvidge green0.4–2.5 μm1 nm4 nm
Chris Elvidge dry0.4–2.5 μm1 nm4 nm
Table 3. Overall accuracy of each similarity measure based on standard spectral libraries (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Table 3. Overall accuracy of each similarity measure based on standard spectral libraries (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Similarity Measure EDSACED-SACHCW-SSC
Overall accuracy (%)87.5091.2593.7597.50
F1 score0.8880.9430.9580.974
Table 4. Overall accuracy and kappa coefficient of each similarity measure based on OMIS HSI (spectral angle cosine (SAC), spectral information divergence (SID), spectral information divergence–spectral angle cosine (SID-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Table 4. Overall accuracy and kappa coefficient of each similarity measure based on OMIS HSI (spectral angle cosine (SAC), spectral information divergence (SID), spectral information divergence–spectral angle cosine (SID-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Similarity MeasureSACSIDSID-SACHCW-SSC
Overall accuracy (%)62.6975.6988.2793.21
Kappa coefficient0.59890.75630.86980.9245
Table 5. Overall accuracy and kappa coefficient of each similarity measure based on AVIRIS HSI (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Table 5. Overall accuracy and kappa coefficient of each similarity measure based on AVIRIS HSI (Euclidean distance (ED), spectral angle cosine (SAC), Euclidean distance–spectral angle cosine (ED-SAC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Similarity MeasureEDSACED-SACRFHCW-SSC
Overall accuracy (%)51.2363.7673.1268.2779.24
Kappa coefficient0.42490.53230.72250.66280.8044
Table 6. Overall accuracy and kappa coefficient of each similarity measure based on Hyperion HSI (spectral information divergence (SID), spectral correlation coefficient (SCC), spectral information divergence–spectral correlation coefficient (SID-SCC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Table 6. Overall accuracy and kappa coefficient of each similarity measure based on Hyperion HSI (spectral information divergence (SID), spectral correlation coefficient (SCC), spectral information divergence–spectral correlation coefficient (SID-SCC), random forest (RF), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)).
Similarity MeasureSIDSCCSID-SCCRFHCW-SSC
Overall accuracy (%)54.8262.3956.3451.2181.23
Kappa coefficient0.39340.43630.36230.42550.7234
Table 7. Overall accuracy and kappa coefficient of each hybrid kernel based on OMIS HSI (the gray area has the results from Table 4, Euclidean distance–spectral correlation coefficient (ED-SCC), Euclidean distance–spectral angle cosine (ED-SAC), spectral correlation coefficient–spectral information divergence (SCC-SID), spectral angle cosine–spectral correlation coefficient (SAC-SCC), Euclidean distance–spectral information divergence (ED-SID), spectral information divergence–spectral angle cosine (SID-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)). The numbers in bold mean the most important results.
Table 7. Overall accuracy and kappa coefficient of each hybrid kernel based on OMIS HSI (the gray area has the results from Table 4, Euclidean distance–spectral correlation coefficient (ED-SCC), Euclidean distance–spectral angle cosine (ED-SAC), spectral correlation coefficient–spectral information divergence (SCC-SID), spectral angle cosine–spectral correlation coefficient (SAC-SCC), Euclidean distance–spectral information divergence (ED-SID), spectral information divergence–spectral angle cosine (SID-SAC), and a new hybrid changing-weight classification method with a filter feature selection (HCW-SSC)). The numbers in bold mean the most important results.
Similarity MeasureED-SCCED-SACSCC-SIDSAC-SCCED-SIDSID-SACHCW-SSC
Overall accuracy (%)63.2258.8261.3969.2466.1788.2793.21
Kappa coefficient0.62150.49320.43420.70140.61380.86980.9245
Table 8. Weight of selected similarity measures based on standard spectral libraries (Euclidean distance (ED) and spectral angle cosine (SAC)).
Table 8. Weight of selected similarity measures based on standard spectral libraries (Euclidean distance (ED) and spectral angle cosine (SAC)).
WeightTypes of the Standard Spectral Library
USGS VegetationChris Elvidge DryChris Elvidge GreenUSGS Mineral
ED0.3120.3290.3430.246
SAC0.6880.6710.6570.754
Table 9. Overall accuracy of a semisupervised kernel with K-means on three different datasets.
Table 9. Overall accuracy of a semisupervised kernel with K-means on three different datasets.
Urban (Hypercube)Kennedy Space Center (AVIRIS)DC Mall (Airborne Sensor)
Number of bands210176210
Number of classes4137
Overall Accuracy
Semisupervised kernel93.48%80.37%99.17%
Table 10. Overall accuracy of support vector machines (SVM), extended morphological profiles (EMPs), joint spare representation (JSR), edge-preserving filtering (EPF), 3D-CNN, CNN with pixel–pair features (CNN-PPF), Gabor-CNN, Siamese CNN (S-CNN), 3D-generative adversarial network (3D-GAN), and the deep feature fusion network (DFFN) on three different datasets.
Table 10. Overall accuracy of support vector machines (SVM), extended morphological profiles (EMPs), joint spare representation (JSR), edge-preserving filtering (EPF), 3D-CNN, CNN with pixel–pair features (CNN-PPF), Gabor-CNN, Siamese CNN (S-CNN), 3D-generative adversarial network (3D-GAN), and the deep feature fusion network (DFFN) on three different datasets.
Salinas (AVIRIS)University of Pavia (ROSIS-03 Sensor)Houston (Airborne Sensor)
Number of bands204115144
Number of classes16915
Overall Accuracy
SVM0.75130.88130.8985
EMP0.78150.95040.9707
JSR0.75280.93490.9307
EPF0.78240.96880.9700
3D-CNN0.80130.95020.9695
CNN-PPE0.79910.9690.9404
Gabor-CNN0.81140.96620.9734
S-CNN0.80520.97430.9510
3D-GAN0.76160.96970.9793
DFFN0.83280.98080.9967
Table 11. Overall accuracy of support vector machines (SVM), k-nearest neighbors (kNN), classification and regression trees (Cart), and naïve Bayes on three different datasets.
Table 11. Overall accuracy of support vector machines (SVM), k-nearest neighbors (kNN), classification and regression trees (Cart), and naïve Bayes on three different datasets.
Indian Pines (AVIRIS)Salinas (AVIRIS)University of Pavia (ROSIS-03 Sensor)
Number of bands224224103
Number of classes9169
Overall Accuracy
SVM0.78750.89940.8992
kNN0.67330.85020.7942
Cart0.63010.83890.7406
Naïve Bayes0.52920.78730.6776
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, J. A Hybrid Clustering Method with a Filter Feature Selection for Hyperspectral Image Classification. J. Imaging 2022, 8, 180. https://doi.org/10.3390/jimaging8070180

AMA Style

Zhang J. A Hybrid Clustering Method with a Filter Feature Selection for Hyperspectral Image Classification. Journal of Imaging. 2022; 8(7):180. https://doi.org/10.3390/jimaging8070180

Chicago/Turabian Style

Zhang, Junzhe. 2022. "A Hybrid Clustering Method with a Filter Feature Selection for Hyperspectral Image Classification" Journal of Imaging 8, no. 7: 180. https://doi.org/10.3390/jimaging8070180

APA Style

Zhang, J. (2022). A Hybrid Clustering Method with a Filter Feature Selection for Hyperspectral Image Classification. Journal of Imaging, 8(7), 180. https://doi.org/10.3390/jimaging8070180

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop