Next Article in Journal
The Effect of Key Anthropometric and Biomechanics Variables Affecting the Lower Back Forces of Healthcare Workers
Next Article in Special Issue
Detection of Plastic Granules and Their Mixtures
Previous Article in Journal
A Self-Decoupling Technique to Realize Dense Packing of Antenna Elements in MIMO Arrays for Wideband Sub-6 GHz Communication Systems
Previous Article in Special Issue
Investigation of the Hue–Wavelength Response of a CMOS RGB-Based Image Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mutual Information-Driven Feature Reduction for Hyperspectral Image Classification

1
Department of Computer Science and Engineering, Hajee Mohammad Danesh Science and Technology University, Dinajpur 5200, Bangladesh
2
Department of Computer Science and Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
3
School of Information Technology, Deakin University, Geelong, VIC 3220, Australia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(2), 657; https://doi.org/10.3390/s23020657
Submission received: 16 November 2022 / Revised: 28 December 2022 / Accepted: 5 January 2023 / Published: 6 January 2023
(This article belongs to the Special Issue Optical Spectral Sensing and Imaging Technology)

Abstract

:
A hyperspectral image (HSI), which contains a number of contiguous and narrow spectral wavelength bands, is a valuable source of data for ground cover examinations. Classification using the entire original HSI suffers from the “curse of dimensionality” problem because (i) the image bands are highly correlated both spectrally and spatially, (ii) not every band can carry equal information, (iii) there is a lack of enough training samples for some classes, and (iv) the overall computational cost is high. Therefore, effective feature (band) reduction is necessary through feature extraction (FE) and/or feature selection (FS) for improving the classification in a cost-effective manner. Principal component analysis (PCA) is a frequently adopted unsupervised FE method in HSI classification. Nevertheless, its performance worsens when the dataset is noisy, and the computational cost becomes high. Consequently, this study first proposed an efficient FE approach using a normalized mutual information (NMI)-based band grouping strategy, where the classical PCA was applied to each band subgroup for intrinsic FE. Finally, the subspace of the most effective features was generated by the NMI-based minimum redundancy and maximum relevance (mRMR) FS criteria. The subspace of features was then classified using the kernel support vector machine. Two real HSIs collected by the AVIRIS and HYDICE sensors were used in an experiment. The experimental results demonstrated that the proposed feature reduction approach significantly improved the classification performance. It achieved the highest overall classification accuracy of 94.93% for the AVIRIS dataset and 99.026% for the HYDICE dataset. Moreover, the proposed approach reduced the computational cost compared with the studied methods.

1. Introduction

A hyperspectral image (HSI), which is acquired at a contiguous spectral wavelength of the electromagnetic spectrum (EM), is a rich data source for a wide range of real-world remote sensing applications, including agriculture, geology, mining, military surveillance, and others [1,2]. Moreover, an HSI is set up as a hypercube and often has hundreds of contiguous, narrow bands in the spectral image [3,4]. Due to the fact that each of these image bands contains varying intensities for the ground cover, they are each referred to as individual features [5,6,7].
There are two dimensions of spatial information and one dimension of spectral information in an HSI, which comprise the three dimensions of spectral-spatial information in the HSI (see Figure S1 in the Supplementary Files) [5,6]. Each spectral image is referred to as a feature for classification in this context, since it contains the distinct responses of the ground surface [7]. Four essential obstacles to a successful classification task are present in a high-dimensional HSI (i.e., an HSI with hundreds of image bands or features). First, because the hyperspectral sensor collects the images in continuous and contiguous spectral ranges, the neighboring image bands are highly correlated and certain image bands carry less discriminating information [5,8]. Secondly, the spectral bands are not equally important, as the bands are captured in different wavelengths of the EM spectrum [9]. Thirdly, there is a significant lack of training samples for some classes [10], which, in turn, creates the Hughes phenomenon or curse of dimensionality problem [11]. The Hughes phenomenon describes the fact that classification accuracy initially rises steadily as the number of spectral bands or dimensions rises, but falls sharply after the number of bands reaches a certain level. Finally, the computational cost of using the entire original HSI is highly expensive [7].
Effective feature (band) reduction is necessary to lower high-dimensional HSIs and create a suitable subspace of the features in order to improve the classification results [12,13,14,15,16] in order to address the aforementioned issues. For the accurate classification of HSIs, feature reduction (FR) techniques using feature extraction (FE) and/or feature selection (FS) might be used. FE maps the original HSI into a new space with a dimensionality of K from the original space with a dimensionality of P , where K     P , using nonlinear or linear conversion [3]. Unsupervised and supervised procedures are the two methods of reducing dimensionality that are used most frequently. While unsupervised procedures do not make any assumptions about the existing knowledge, supervised methods are intended to preserve previously known information (ground truth). The most widely used unsupervised linear FE approach is principal component analysis (PCA) [17,18,19]. It is based on the idea that adjacent bands are highly correlated, and uses global statistics to eliminate the connections between bands [20,21]. It is often claimed that PCA is better for data compression purposes but is not suitable for extracting the most informative feature in the classification task [22,23,24,25,26,27,28,29]. The reasons for this are: (i) PCA may not catch the detailed local statistics, as it determines the overall characteristics of the entire HSI; (ii) the top principal components (PCs) or transformed features may not always contain the informative structure of the entire HSI (i.e., the tasks are biased in PCs with high variance); and (iii) PCA requires a high computational cost for high-volume hyperspectral data, as it considers the global statistics [30,31,32].
To address the pitfalls of the classical PCA, correlation-based segmented PCA (SPCA) was presented in [22], which applies conventional PCA to the bands’ subgroups. The entire dataset is divided into multiple segments using the image’s band-to-band correlation matrix. For a subgroup’s dataset, the contiguous strongly correlated bands are often assigned. However, this correlation-based segmentation strategy can only sufficiently reflect the linear relationships of the bands for making the subgroups. As such, correlation-based segmentation might not be feasible for performing classical PCA on large-volume HSIs with a huge number of bands for effective FE. Comparatively, mutual information (MI) is a dependence metric that has a built-in ability to manage the HSI in both linear and nonlinear connections [9,33]. With this motivation, we proposed a band grouping method of partitioning the spectral bands using a band-to-band normalized MI (NMI) matrix for effective FE, which is called band grouping-based PCA (BgPCA). The suggested FE method, BgPCA, first uses the NMI measure to divide the original bands into multiple groups and then applies conventional PCA separately to each subgroup of the original image bands at a minimum computing cost.
As segmented PCA is applied to the complete dataset, there is a need to apply feature selection to select the optimal number of features. For FS, the subspace of effective features extracted by our BgPCA transformation for classification is selected using the NMI values of the transformed features to a specific range, thus meeting the minimum redundancy and maximum relevance (mRMR) criteria. Accordingly, the complete FR approach is known as BgPCA-NMI, which significantly enhances the classification performance and minimizes the computational costs as well. Although the proposed method shows outstanding performance in terms of different performance measure metrics, it has some limitations. A user-defined threshold is used to effectively partition the complete HSI. It can be optimized adaptively, and our future goal is to use a network model that automatically selects the threshold value from the dataset. On the other hand, the proposed method only addresses the spectral features. However, data redundancy exists in the spatial domain of the HSI. As such, in the future, a deep learning-based approach could be used to extract the spectral-spatial information [34,35] alongside our proposed FR technique for further improving the classification outcome. To this end, the main contributions of this study are listed below.
  • We propose an MI-driven efficient FR approach for the effective classification of HSI.
  • We introduce an NMI-based band grouping strategy for intrinsic FE by applying classical PCA transformation to each group of bands independently for effective FE from HSI.
  • We propose an NMI-based mRMR FS method using the extracted features through our proposed transformation.
  • We performed extensive experiments on two widely used benchmark HSI datasets captured by the AVIRIS and HYDICE sensors to validate the superiority of our proposed FR approach.
We have organized the rest of the article as follows. In Section 2, we first describe the insights of the proposed NMI-based band grouping strategy for applying classical PCA in a segmented manner. Next, the proposed FE called BgPCA is elaborately presented. After that, we discuss the NMI-based mRMR FS criteria on top of our BgPCA transformation. Lastly, we present the complete FR method called BgPCA-NMI at the end of Section 2. In Section 3, we intricately analyze the experiments conducted on two real HSI datasets using the proposed BgPCA-NMI FR approach and the state-of-the-art methods. Finally, Section 4 summarizes the outcomes and concludes the article.

2. Methodology

The proposed FR approach, called BgPCA-NMI, encompasses three main steps: (i) band grouping based on the band-to-band NMI matrix of the entire original HSI; (ii) FE through implementation of classical PCA on each subgroup dataset independently; and (iii) FS through measurement of the NMI-based mRMR criteria of the transformed features of the HSI. Figure 1 illustrates the working steps of our BgPCA-NMI.

2.1. Proposed Band Grouping Strategy Based on NMI

The MI is a popular information-theoretic metric used to measure the general dependency of two random variables, say two image bands A and B . The MI, denoted I ( A , B ) , is defined as
I ( A , B ) = b B a A p ( a , b ) log p ( a , b ) p ( a )   p ( b )
where p ( . ) denotes the marginal probability and p ( . ,   . ) is the distribution of the joint probability. In the context of HSI, the MI assesses the information that is shared among the spectral image bands and intuitively determines the interdependence among the image bands. A higher MI indicates more dependency between them and vice versa. Note that although correlation has been successfully applied as a similarity measure tool in many studies [22,36,37], it suffers when nonlinearity exists in the image bands. Unlike the correlation metric, the MI finds the dependency of the image bands both in linear and nonlinear ways. However, the MI value is not bound to a precise range, which creates difficulty when measuring the actual relationships of the relevant image bands of the HSI. Therefore, the MI value can be mapped to a specific range for normalizing the measure as follows [38]
  I ^ ( A , B ) = I ( A , B ) H ( A ) H ( B )
In Equation (2),   I ^ ( A , B ) is the NMI between the image bands A and B , and H ( . ) denotes the marginal entropy. Figure 2 represents how the NMI measure is more influential compared with the correlation measure and traditional MI. Figure 2a depicts the scatterplots of two random variables showing a perfect linear relationship, and Figure 2b illustrates the scatterplots expressing a point-to-point but nonlinear relationship between the variables. It can be observed that the correlation is strong when the variables are fully linear and low when the variables are nonlinear. Contrariwise, the NMI is high (i.e., 1.0) for both cases, which reemphasizes the superiority of the NMI-based similarity measure.
We determined the band-to-band NMI matrix for the dataset by computing the NMI between every pair of HSI bands in order to segment the complete HSI. As an illustration, Figure 3a shows the baseline correlation matrix image notation and Figure 3b shows the NMI matrix in image notation of the AVIRIS Indian Pines dataset. The NMI values of the HSI bands that are close to one another are greater than those of the bands that are farther apart. The HSI bands from the original data cube were therefore divided and grouped using the NMI matrix image. The details of the subgroups using this proposed NMI-based segmentation strategy and the baseline correlation-based segmentation strategy of SPCA for the AVIRIS and HYDICE datasets are shown in Table 1 and Table 2, respectively.

2.2. PCA

In order to extract meaningful information, PCA was used to calculate the relationships between the spectral image bands in HSI. This depended on the fact that the HSI’s neighboring bands were strongly linked and frequently communicate information about the ground entities that were similar to one another [28,29]. Let the spectral vector, denoted as X n , in X be defined as X n = [ X n 1 X n 2 X n P ] T , where n [ 1   S all ] . The mean adjusted spectral vector, I n , can now be obtained as follows
I n = X n M
where the mean image vector, M = 1 S all n = 1 S all X n . The zero-mean image, denoted I , is thus obtained as I = [ I 1 I 2 I n ] . After that, the covariance matrix, C is computed as follows
C = 1 S all I   I T
Eigenvalues E = [ E 1 E 2 E P ] and eigenvectors V = [ V 1 V 2 V P ] are obtained by disintegrating the covariance matrix C as C = V E V T . The orthonormal matrix, Z , is collected by picking K eigenvectors after reorganizing the eigenvectors with the peak eigenvalues, where K < P and often K P . Finally, the transformed or projected data matrix, Y , is calculated as
Y = Z T I

2.3. Proposed BgPCA

When the multispectral dataset has a small number of distinct bands, it has been observed that conventional PCA is practical and delivers satisfactory results when extracting appropriate features [19]. Applying traditional PCA to the complete dataset for hyperspectral images may provide biased outcomes in addition to an exponential rise in processing time and computing expense [22]. The highly connected image bands of the HSI, however, appear in blocks and were shown to be highly associated with respect to the bands that are closer together. Additionally, PCA retrieves the HSI data while taking the overall HSI features into account and fails to extract the local information. In order to eliminate the less correlated bands between the highly correlated blocks, SPCA changes the use of traditional PCA.
Because it cannot handle the nonlinear interactions between the bands during the segmentation phase, the correlation-based segmentation of HSI utilized in SPCA may still be impractical. As such, we proposed an NMI-based band grouping mechanism for more efficient band segmentation to handle both linearity and nonlinearity in the image bands, which was discussed in Section 2.1. The improvement allowed by the proposed BgPCA over conventional PCA is that it extracts the local characteristics of data in an efficient way rather than considering the global statistics of the HSI. Moreover, the computational cost of conventional PCA can be significantly reduced with BgPCA and, consequently, the total computational cost of HSI classification is decreased.
In the implementation of BgPCA, the complete HSI dataset is separated into k subgroup datasets based on the NMI-based segmentation scheme. Next, the covariance matrix is computed for each subgroup’s dataset. Afterward, each estimated covariance matrix of the distinct subgroups is subjected to the eigen-decomposition procedure independently. Consequently, the final projection matrix of the entire dataset is found by merging the individual projection matrices consecutively. We pictorially illustrate the working principle of the proposed BgPCA in Figure 4, and the pseudocode is given in Algorithm 1.
Algorithm 1. BgPCA
  • Start {X: the 2D dataset of the HSI}
  • Calculate the band-to-band NMI matrix to produce the subgroups of X
  • Based on the NMI matrix, divide X into k subgroups
  • For each subgroup, do
  • Compute the projection matrix via PCA
  • Combine all individual projection matrices successively to build the complete projection matrix

2.4. Proposed BgPCA-NMI

For selecting the effective subspace of the extracted features, we first measured the MI between each new extracted feature, Y i , from our BgPCA transformation and the ground truth image, T . Accordingly, the most informative feature could be calculated using Equation (6) and assigned to the feature subspace, S .
V = max i P   I ( Y i , T ) .
The K features selected by using Equation (6) can have potential redundancy, which would affect the classification performance. As such, the redundancy between the selected features needed to be minimized for efficient classification. To address this, we applied the mRMR criteria for selecting K effective features as follows
G ( Y i , K ) = I ( Y i , T ) 1 K s j S I ( Y i , s j ) , Y i S
Since it is impacted by the entropy of two variables and is not constrained to a certain range, the MI value, I ( Y i , T ) from the equation above is challenging to utilize directly. To measure how good an MI value is, we normalized it to the range [0, 1] using Equation (8), while the proposed informative FS approach is defined in Equation (9).
  I ^ ( Y i , T ) = I ( Y i , T ) H ( Y i ) H ( T )
  G ^ ( Y i , K ) = I ( Y i , T ) 1 K s j S   I ^ ( Y i , s j ) , Y i S
Equation (9), when applied, might result in a difference between the selected characteristics and those that have previously been chosen, which is undesirable. Therefore,   G ^ ( Y i , K ) was considered to be positive, i.e.,   G ^ ( Y i , K ) > 0 , in our analysis. The complete pseudocode is given in Algorithm 2.
Algorithm 2. BgPCA-NMI
  • Start {Y: the projected data matrix via BgPCA}
  • Instantiate the feature subspace to null, S0 = {Φ}
  • Choose the first feature, Yj using Equation (6) via utilizing Equation (8) and execute S 1 = S 0     Y j
  • For choosing the rest of the features, do
  • Apply Equation (9) and update S
  • Take S as the subspace of the top effective features

3. Experiment and Analysis of the Results

3.1. Description of the Dataset

In the experiment, the classification task used two different benchmark HSI datasets. We leveraged the dataset from an urban mall in Washington, DC, and the mixed agricultural Indian Pines dataset, which were captured by the AVIRIS and HYDICE sensors, respectively. The Indian Pines HSI has 220 spectral imaging bands in total, each with a spatial resolution of 145 × 145 pixels [39]. However, due to the impact of atmospheric phenomena, its 20 water absorption bands ([104, 108], [150, 163], and 220) have been disregarded in this analysis. For classification purposes, there w 16 classes in the ground truth map. The “grass/mown pasture” and “oats” classes were not utilized in this experiment, since there were insufficient data for them. A high-volume dataset with a spatial resolution of 1280 × 307 pixels, the urban Washington DC Mall dataset, comprises 191 spectral image bands [40]. The ground truth map has seven classes available. The “paths” class was not utilized for classification in the experimental analysis, as there were insufficient samples. We display some sample band images and the ground truth image of the Indian Pines HSI and the false color image of the Washington DC Mall HSI with its ground truth information in Figure S2 in the Supplementary Files, while Table 3 illustrates the key properties of these datasets.

3.2. Results of FE and FS

The SPCA transformation partitioned the original HSI into several subgroups using the correlation matrix; however, the suggested FE method BgPCA divided the image bands using the NMI matrix. The results of BgPCA segmentation shown in Table 1 are based on the findings obtained by accounting for NMI values greater than the threshold of 0.3 and looking for edges in the image of the NMI matrix (Figure 3a) together with the diagonal direction. The total number of subgroup datasets that will be created is indicated by the user-defined threshold when the threshold is low, the number of subgroups increases, and uncorrelated bands may be clustered together. We chose the threshold value so that the image bands with high correlations were grouped together. To carry this out, we looked at the dataset and the NMI matrix image, and used the trial-and-error method to choose the threshold value. The entire AVIRIS dataset was split into three subgroups in Table 1, which each contained 102, 41, and 57 bands. However, the correlation-based SPCA also partitioned the entire dataset into three subgroups consisting of 35, 68, and 97 bands. Similarly, we partitioned the whole HYDICE dataset into four subgroup datasets comprising 58, 50, 51, and 33 bands for BgPCA, as illustrated in Table 2. Nevertheless, SPCA divided the entire HYDICE dataset into three highly correlated subgroups consisting of 56, 46, and 89 bands. PCA was applied separately to each individual subgroup for both BgPCA and SPCA with both datasets. For both BgPCA and SPCA, the informative features were selected using the NMI-based mRMR criteria for effective subspace detection. Table S1 in the Supplementary Files presents the acronyms associated with the proposed method and the different methods studied, while the order of ranked features that were used for classification is listed in Tables S2 and S3 in the Supplementary Files for the Indian Pines and Washington DC datasets, respectively. In Tables S2 and S3, the segmentation number is utilized first, followed by the PC number of this segment (Segment:PC) for the segmented PCA (SPCA) technique. The proposed BgPCA approach initially gives the group number, followed by the number of PCs in this group (Group:PC).

3.3. Performance Evaluation Metrics

The overall accuracy (OA), average accuracy (AA), Kappa coefficient, and F1 score are widely used quality indices applied in this study to evaluate how well the proposed technique performed. The percentage of all correctly identified pixels is known as the OA, and it can be calculated as follows:
OA = i = 1 C A i i B
In Equation (11), C stands for the number of classes and A for the confusion matrix, which is determined by contrasting the classification map with the ground truth image. The number of samples belonging to Class i and labeled as Class i (i.e., values found along the diagonal of the confusion matrix) is represented by Aii, whereas the total number of test samples is represented by B.
AA stands for the average accuracy, which is the average proportion of correctly classified pixels for each class, which is determined as follows
AA = i = 1 C ( A i i / i = 1 C A i + ) C
where Aii stands for total number of samples belonging to Class i and classified as Class i (i.e., values found along the diagonal of the confusion matrix), and Ai+ represents the total number of samples as classified as Class i.
The Kappa coefficient computes the proportion of classified pixels adjusted for the number of agreements predicted only by chance. The Kappa statistic indicates how much better the categorization performs than the likelihood of randomly assigning pixels to their correct categories and can be calculated using the notation used in Equations (10) and (11) as
Kappa = ( B i = 1 C A i i i = 1 C ( A i + ) ( A + i ) ) ( B 2 i = 1 C i = 1 C ( A i + ) ( A + i ) )
where A+i represents the total number of actual samples in Class i. The F1 score can now be calculated as follows
F 1   score = 2 × Precision × Recall Precision + Recall
where the precision and recall can be calculated as follows:
Precision = TP TP + FP   and   Recall = TP TP + FN
Here, TP, FP, and FN denote the number of true positive, false positive, and false negative classifications of the testing samples of multiple classes, respectively.

3.4. Classification Results and Evaluation

The performance of the proposed FR method, BgPCA-NMI, was assessed in terms of the following classical performance measure metrics: OA, AA, Kappa, precision, recall, and F1 score. We also considered a scatterplot-based feature space analysis scheme and the computation cost for better expressing the robustness of our BgPCA-NMI FR method. After FE, the first few ranked features selected by the FS approach, as illustrated in Tables S2 and S3 in the Supplementary Files, were used to calculate the abovementioned performance measure metrics using the kernel support vector machine (KSVM) and the radial basis kernel function (RBF) [41] to tackle any nonlinearity in the final feature set. For the task of efficient training, 10-fold cross-validation with a grid search strategy was used for selecting the best cost parameter (C_cost) and kernel width (γ) associated with the RBF-KSVM. The proposed FR approach was compared with conventional PCA and correlation-based SPCA with variance, and with NMI-based mRMR FS and BgPCA with variance-based feature ranking. The kernel parameters C_cost = 8 and γ = 1.33 for the AVIRIS dataset and C_cost = 6 and γ = 1.23 for the HYDICE dataset were tuned by using 15 features of the AVIRIS dataset and 8 features of the HYDICE dataset. Table 4 shows all the parameter tuning results for the KSVM classifier of the proposed and other algorithms studied for the two HSI datasets. In total, 2127 pixels from 14 different classes from the AVIRIS dataset were used for the classification, where around 50% were used for training and 50% were used for testing, as shown in Table S4 in the Supplementary Files. On the other hand, 4464 pixels in total from six classes from the HYDICE dataset were used for classification, where 30% were used for training and 70% were used for testing, as illustrated in Table S5 in the Supplementary Files.
For the AVIRIS Indian Pines dataset, the OAs produced by the studied methods (PCA, SPCA, and SPCA-NMI) and the proposed approach (BgPCA-NMI along with plain BgPCA) are depicted in Figure 5. In this case, a line graph was utilized to assess the relevance of the ranking attributes at each stage. To begin, just the top-ranked feature was utilized to calculate the classification accuracy. The top two features were then used and calculated. Following that, the classification accuracy was computed using the first three characteristics. In this manner, the total classification accuracy was determined and shown in the line graph using the selected features. The traditional PCA and correlation-based SPCA achieved 92.45% and 83.40% OA, respectively. On the other hand, the proposed FR method, BgPCA-NMI, had the highest OA of 94.93% using the same number of features, which clearly indicates the advantage of using the proposed approach. For the HYDICE Washington DC Mall dataset, the OAs for the different methods are illustrated in Figure 6. The OAs produced by PCA and SPCA were 92.8% and 94.81%, respectively, whereas the proposed FR method, BgPCA-NMI, had the highest OA of 99.026% using the same number of features. This also led to the superiority of using the proposed feature space identification over the existing methods.
Table 5 and Table 6 show all the classification performance metrics (AA, OA, Kappa, and F1 score) for the proposed FR method and each of the other studied methods. The proposed method demonstrated an improvement in terms of all these metrics as compared with the studied methods. The robustness of the proposed method, BgPCA-nMI, for multiclass classification was also evaluated using the error matrices, as shown in Tables S6 and S7 for the AVIRIS and HYDICE datasets, respectively, in the Supplementary Files. From both error matrices, it can be seen that almost all classes were correctly predicted, except for very few of them.
By utilizing the feature space analysis framework, we evaluated the robustness of the suggested method, BgPCA-NMI. Figure 7 depicts the 2D feature space for the AVIRIS Indian Pines HSI using the first two ranked features for each method to show the effects of feature selection and feature extraction. For simplicity, eight classes were plotted in the feature space. Visually, it is clear that the proposed FR method separated the classes better than the tested methods. Similarly, Figure 8 shows the 2D feature space for the Washington DC Mall dataset using the first two ranked features. The outcome further proved that the suggested FR strategy separated the classes more effectively than the approaches under investigation. It was therefore observed that the scatterplots represented how the classes were separated from one another. For the proposed method, we can see that the classes are more separated compared with the studied methods. If the classes are separated well, the method will classify the samples more accurately. Note that we could not find any other additional statistics to calculate these classes regarding the transformed spaces from the literature on hyperspectral imagery.
The effectiveness of the proposed approach, BgPCA-NMI, was finally assessed using the total computational time in different stages, as given in Table 7, for both datasets. The proposed method was tested on a personal computer equipped with an Intel Core i5 3.2 GHz CPU and 8 GB of RAM, running on the Microsoft Windows 10 operating system. It was evident that the suggested approach, BgPCA-NMI, took less time overall to compute than conventional PCA, indicating an increase in computational efficiency.

4. Conclusions and Future Work

Because an HSI is a high-dimensional data cube, effective FE is necessary to provide outstanding classification performance while decreasing the computing costs. In this study, we used the NMI measure because of its appropriate treatment of nonlinearity in partitioning the original HSI bands efficiently instead of using the correlation for the segmentation, as in the case of SPCA. For successful FE, PCA was performed on each subgroup of bands after the band-to-band NMI matrix of the HSI had been utilized to divide all the spectral bands into a number of groups. As a result, the proposed FE approach extracted useful features while taking the HSI dataset’s local characteristics into account, and the computational cost of extracting the features decreased greatly. After that, the NMI between each transformed feature and the ground truth was used for selecting the subspace of informative features using the mRMR scheme. In comparison with traditional PCA and correlation-based SPCA, BgPCA-NMI increased the classification accuracy, as shown by the classification performance and analysis of the results on two actual HSI datasets, Indian Pines and Washington DC Mall. Ultimately, the proposed method, BgPCA-NMI, effectively reduced the computational cost.
Effective partitioning of the whole HSI was achieved by using a user-defined threshold. We want to utilize a network model that automatically chooses the threshold value from the dataset in the future. It may be optimized adaptively. On the other hand, the proposed method just takes the spectral characteristics into account. However, there is data redundancy in the HSI’s spatial domain. As a result, in the future, our suggested FR technique, as well as a deep learning-based strategy, will be used to extract the spectral and spatial information to further improve the classification results. Finally, as well as our feature space analysis, other distance metrics or statistics, such as the Bhattacharyya distance, class compactness, etc., within the PC space and BgPCA space could be used in the future to quantify the separation better.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s23020657/s1, Figure S1: Structure of an HSI; Table S1: Acronyms of different studied and proposed methods; Figure S2: HSI datasets; Table S2: The rank of selected features for the classification of Indian Pines dataset; Table S3: The rank of selected features for the classification of DC Mall dataset; Table S4: Training and testing samples for AVIRIS dataset; Table S5: Training and testing samples for HYDICE dataset; Table S6: Error matrix of BgPCA-NMI for Indian Pines dataset; Table S7: Error matrix of BgPCA-NMI for DC Mall dataset.

Author Contributions

Conceptualization, M.R.I., B.A. and M.P.U.; methodology, M.R.I., M.P.U. and M.A.H.; software M.R.I.; validation, M.R.I. and M.P.U.; formal analysis, M.R.I.; investigation, M.R.I., B.A. and M.P.U.; resources, M.R.I.; data curation, M.R.I.; writing—original draft preparation, M.R.I.; writing—review and editing, B.A., M.A.H. and M.P.U.; visualization, M.R.I. and M.P.U.; supervision, B.A., M.A.H. and M.P.U.; funding acquisition, M.P.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The AVIRIS Indian Pines data [39] are available at https://purr.purdue.edu/publications/1947/1 (accessed on 3 March 2022), while the HYDICE Washington DC Mall data [40] are available at https://engineering.purdue.edu/~biehl/MultiSpec/hyperspectral.html (accessed on 3 March 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mohan, B.K.; Porwal, A. Hyperspectral image processing and analysis. Curr. Sci. 2015, 108, 833–841. [Google Scholar]
  2. Wang, N.; Zeng, X.; Duan, Y.; Deng, B.; Mo, Y.; Xie, Z.; Duan, P. Multi-Scale Superpixel-Guided Structural Profiles for Hyperspectral Image Classification. Sensors 2022, 22, 8502. [Google Scholar] [CrossRef] [PubMed]
  3. Richards, J.A.; Jia, X. Remote Sensing Digital Image Analysis; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  4. Tinega, H.; Chen, E.; Ma, L.; Mariita, R.M.; Nyasaka, D. Hyperspectral Image Classification Using Deep Genome Graph-Based Approach. Sensors 2021, 21, 6467. [Google Scholar] [CrossRef]
  5. Guo, B.; Gunn, S.R.; Damper, R.I.; Nelson, J.D.B. Band Selection for Hyperspectral Image Classification Using Mutual Information. IEEE Geosci. Remote Sens. Lett. 2006, 3, 522–526. [Google Scholar] [CrossRef] [Green Version]
  6. Manian, V.; Alfaro-Mejía, E.; Tokars, R.P. Hyperspectral Image Labeling and Classification Using an Ensemble Semi-Supervised Machine Learning Approach. Sensors 2022, 22, 1623. [Google Scholar] [CrossRef]
  7. Zabalza, J.; Ren, J.; Yang, M.; Zhang, Y.; Wang, J.; Marshall, S.; Han, J. Novel folded-PCA for improved feature extraction and data reduction with hyperspectral imaging and SAR in remote sensing. ELSEVIER ISPRS J. Photogramm. Remote Sens. 2014, 93, 112–122. [Google Scholar] [CrossRef] [Green Version]
  8. Rasti, B.; Scheunders, P.; Ghamisi, P.; Licciardi, G.; Chanussot, J. Noise Reduction in Hyperspectral Imagery: Overview and Application. Remote Sens. 2018, 10, 482. [Google Scholar] [CrossRef] [Green Version]
  9. Hossain, M.A.; Jia, X.; Pickering, M. Subspace Detection Using a Mutual Information Measure for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2014, 11, 424–428. [Google Scholar] [CrossRef]
  10. Du, W.; Qiang, W.; Lv, M.; Hou, Q.; Zhen, L.; Jing, L. Semi-supervised dimension reduction based on hypergraph embedding for hyperspectral images. Int. J. Remote. Sens. 2018, 39, 1696–1712. [Google Scholar] [CrossRef]
  11. Hughes, G. On the Mean Accuracy of Statistical Pattern Recognizers. IEEE Trans. Inf. Theory 1968, IT-14, 55–63. [Google Scholar] [CrossRef] [Green Version]
  12. Hossain, M.A.; Jia, X.; Pickering, M. Subspace detection based on the combination of nonlinear feature extraction and feature selection. In Proceedings of the IEEE Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, Gainesville, FL, USA, 26–28 June 2013. [Google Scholar]
  13. Islam, M.R.; Ahmed, B.; Hossain, M.A. Feature Reduction Based on Segmented Principal Component Analysis for Hyperspectral Images Classification. In Proceedings of the International Conference on Electrical, Computer and Communication Engineering (ECCE), Cox’s Bazar, Bangladesh, 7–9 February 2019. [Google Scholar]
  14. Haque, M.R.; Mishu, S.Z.; Uddin, M.P.; Mamun, M.A. A lightweight 3D-2D convolutional neural network for spectral-spatial classification of hyperspectral images. J. Intell. Fuzzy Syst. 2022, 43, 1241–1258. [Google Scholar] [CrossRef]
  15. Islam, M.R.; Hossain, M.A.; Ahmed, B. Improved Subspace Detection Based on Minimum Noise Fraction and Mutual Information for Hyperspectral Image Classification. In Proceedings of the International Joint Conference on Computational Intelligence, Dhaka, Bangladesh, 14–15 December 2018. [Google Scholar]
  16. Siddiqa, A.; Afzal, M.I.; Islam, M.R.; Nitu, A.M. Spectral Subset Detection for Hyperspectral Image Classification. In Proceedings of the International Conference on Innovation in Engineering and Technology (ICIET), Dhaka, Bangladesh, 26–27 November 2020. [Google Scholar]
  17. Ying, L.; Yanfeng, G.; Ye, Z. Hyperspectral feature extraction using selective PCA based on genetic algorithm with subgroups. In Proceedings of the International Conference on Innovative Computing, Information and Control, Beijing, China, 30 August–1 September 2006. [Google Scholar]
  18. Munishamaiaha, K.; Rajagopal, G.; Venkatesan, D.K.; Arif, M.; Vicoveanu, D.; Chiuchisan, I.; Izdrui, D.; Geman, O. Robust Spatial–Spectral Squeeze–Excitation AdaBound Dense Network (SE-AB-Densenet) for Hyperspectral Image Classification. Sensors 2022, 22, 3229. [Google Scholar] [CrossRef] [PubMed]
  19. Rodarmel, C.; Shan, J. Principal Component analysis for hyper-spectral image classification. ACM Surv. Land Inf. Syst. 2002, 62, 115–122. [Google Scholar]
  20. Arslan, O.; Akyürek, O.; Kaya, S.; Şeker, D.Z. Dimension reduction methods applied to coastline extraction on hyperspectral imagery. Geocarto Int. 2018, 35, 376–390. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Li, T.; Tang, X.; Hu, X.; Peng, Y. CAEVT: Convolutional Autoencoder Meets Lightweight Vision Transformer for Hyperspectral Image Classification. Sensors 2022, 22, 3902. [Google Scholar] [CrossRef]
  22. Jia, X.; Richards, J.A. Segmented Principal Components Transformation for Efficient Hyperspectral Remote-Sensing Image Display and Classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 538–542. [Google Scholar]
  23. Chen, G.; Qian, S.; Gleason, S. Denoising of hyperspectral imagery by combining PCA with block-matching 3-D filtering. Can. J. Remote Sens. 2012, 37, 590–595. [Google Scholar] [CrossRef]
  24. Ibarrola-Ulzurrun, E.; Marcello-Ruiz, J.; Gonzalo-Martín, C. Assessment of Component Selection Strategies in Hyperspectral Imagery. Entropy 2017, 19, 666. [Google Scholar] [CrossRef]
  25. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. Feature Extraction for Hyperspectral Image Classification. In Proceedings of the IEEE 5th Region 10 Humanitarian Technology Conference (R10-HTC), Dhaka, Bangladesh, 21–23 December 2017. [Google Scholar]
  26. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. Segmented FPCA for Hyperspectral Image Classification. In Proceedings of the IEEE 3rd International Conference on Electrical Information and Communication Technology (EICT), Khulna, Bangladesh, 7–9 December 2017. [Google Scholar]
  27. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. Improved Feature Extraction Using Segmented FPCA for Hyperspectral Image Classification. In Proceedings of the International Conference on Electrical & Electronic Engineering, Rajshahi, Bangladesh, 27–29 December 2017. [Google Scholar]
  28. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. Effective feature extraction through segmentation-based folded-PCA for hyperspectral image classification. Int. J. Remote Sens. 2019, 40, 7190–7220. [Google Scholar] [CrossRef]
  29. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. PCA-based Feature Reduction for Efficient Hyperspectral Image Classification. IETE Tech. Rev. 2020, 38, 1740615. [Google Scholar]
  30. Wang, L.; Wang, X. Dual-Coupled CNN-GCN-Based Classification for Hyperspectral and LiDAR Data. Sensors 2022, 22, 5735. [Google Scholar] [CrossRef] [PubMed]
  31. Qasim, M.; Khan, S.D. Detection and Relative Quantification of Neodymium in Sillai Patti Carbonatite Using Decision Tree Classification of the Hyperspectral Data. Sensors 2022, 22, 7537. [Google Scholar] [CrossRef] [PubMed]
  32. Mukundan, A.; Huang, C.-C.; Men, T.-C.; Lin, F.-C.; Wang, H.-C. Air Pollution Detection Using a Novel Snap-Shot Hyperspectral Imaging Technique. Sensors 2022, 22, 6231. [Google Scholar] [CrossRef] [PubMed]
  33. Dionisio, A.; Menezes, R.; Mendes, D.A. Mutual information: A measure of dependency for nonlinear time series. Phys. A: Stat. Mech. Its Appl. 2004, 344, 326–329. [Google Scholar] [CrossRef]
  34. Lixin, G.; Weixin, X.; Jihong, P. Segmented minimum noise fraction transformation for efficient feature extraction of hyperspectral images. Pattern Recognit. 2015, 48, 3216–3226. [Google Scholar] [CrossRef]
  35. Kuo, B.; Chang, W.; Li, C.; Hung, C. Correlation matrix feature extraction based on spectral clustering for hyperspectral image segmentation. In Proceedings of the Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Shanghai, China, 4–7 June 2012. [Google Scholar]
  36. Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J. Spectral-spatial classification of hyperspectral imagery Based on partitional clustering techniques. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2973–2987. [Google Scholar] [CrossRef]
  37. Tsai, F.; Lin, E.-K.; Yoshino, K. Spectrally segmented principal component analysis of hyperspectral imagery for mapping invasive plant species. Int. J. Remote Sens. 2007, 28, 1023–1039. [Google Scholar] [CrossRef]
  38. Uddin, M.P.; Mamun, M.A.; Afjal, M.I.; Hossain, M.A. Information-theoretic feature selection with segmentation-based folded principal component analysis (PCA) for hyperspectral image classification. Int. J. Remote Sens. 2020, 42, 286–321. [Google Scholar] [CrossRef]
  39. Baumgartner, M.F.; Labial, L.; Landgrebe, D.A. 220 Band AVIRIS Hyperspectral Image Data Set: June 12, 1992 Indian Pine Test Site 3; Purdue University Research Repository: West Lafayette, IN, USA, 2015. [Google Scholar]
  40. Landgrebe, D.A. Available online: https://engineering.purdue.edu/~biehl/MultiSpec/hyperspectral.html (accessed on 3 March 2022).
  41. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
Figure 1. The workflow of the proposed BgPCA-NMI.
Figure 1. The workflow of the proposed BgPCA-NMI.
Sensors 23 00657 g001
Figure 2. Scatterplots illustrating two variables in (a) a perfect linear connection and (b) a nonlinear point-to-point relationship.
Figure 2. Scatterplots illustrating two variables in (a) a perfect linear connection and (b) a nonlinear point-to-point relationship.
Sensors 23 00657 g002
Figure 3. Spectral band-to-band measurement used for segmentation of the Indian Pines HSI: (a) NMI matrix in the image and (b) the correlation matrix in image form.
Figure 3. Spectral band-to-band measurement used for segmentation of the Indian Pines HSI: (a) NMI matrix in the image and (b) the correlation matrix in image form.
Sensors 23 00657 g003
Figure 4. Working stages of BgPCA, P* denotes the number of transformed features.
Figure 4. Working stages of BgPCA, P* denotes the number of transformed features.
Sensors 23 00657 g004
Figure 5. Overall classification accuracy on the AVIRIS dataset.
Figure 5. Overall classification accuracy on the AVIRIS dataset.
Sensors 23 00657 g005
Figure 6. Overall classification accuracy on the HYDICE dataset.
Figure 6. Overall classification accuracy on the HYDICE dataset.
Sensors 23 00657 g006
Figure 7. Scatterplots for different methods for the Indian Pines dataset: (a) PCA, (b) SPCA, (c) BgPCA, and (d) BgPCA-NMI.
Figure 7. Scatterplots for different methods for the Indian Pines dataset: (a) PCA, (b) SPCA, (c) BgPCA, and (d) BgPCA-NMI.
Sensors 23 00657 g007aSensors 23 00657 g007b
Figure 8. Scatterplots for different methods for the Washington DC Mall dataset: (a) PCA, (b) SPCA, (c) BgPCA, and (d) BgPCA-NMI.
Figure 8. Scatterplots for different methods for the Washington DC Mall dataset: (a) PCA, (b) SPCA, (c) BgPCA, and (d) BgPCA-NMI.
Sensors 23 00657 g008
Table 1. Band grouping information of the Indian Pines (AVIRIS) HSI.
Table 1. Band grouping information of the Indian Pines (AVIRIS) HSI.
SPCA (Baseline Approach)BgPCA (Proposed Approach)
GroupRange of Bands# of BandsAverage CorrelationRange of Bands# of BandsAverage NMI
11–35350.87701–1021020.3400
236–103680.7171103–143410.7171
3104–200970.6950144–200670.6950
Table 2. Band grouping information of the Washington DC (HYDICE) HSI.
Table 2. Band grouping information of the Washington DC (HYDICE) HSI.
SPCA (Baseline Approach)BgPCA (Proposed Approach)
GroupRange of Bands# of BandsAverage CorrelationRange of Bands# of BandsAverage NMI
11–56560.94431–58580.4500
257–102460.884259–108500.5460
3103–191890.9813109–159510.5460
4---159–191330.5600
Table 3. Summary of the datasets.
Table 3. Summary of the datasets.
Name of the DatasetCapturing SensorPWavelength
Range (nm)
HWGround
Classes
Ground Sampling
Distance (m)
Indian PinesAVIRIS220400–25001451451620
Washington DC MallHYDICE191400–2400128030773
Table 4. Parameter tuning using 10-fold cross-validation.
Table 4. Parameter tuning using 10-fold cross-validation.
Method NameBest CBest γTraining Accuracy
AVIRISPCA10398.55
SPCA3.52.894.65
SPCA-NMI5298.50
BgPCA1.83.796.85
BgPCA-NMI71.298.88
HYDICEPCA10397.55
SPCA42.195.83
SPCA-NMI2.53.997.68
BgPCA31.598.15
BgPCA-NMI71.298.95
Table 5. Detailed classification outcomes for the Indian Pines dataset.
Table 5. Detailed classification outcomes for the Indian Pines dataset.
ClassPCASPCASPCA-NMIBgPCABgPCA-NMI
Hay—windrowed97.6690.6090.6094.4196.43
Soybean—no till80.4176.6794.8184.5484.69
Woods94.9692.2791.2796.1996.71
Wheat100.0098.4110094.03100.00
Grass—trees100.00100.00100100100.00
Soybean– min. till89.6770.2192.6494.3494.94
Grass—pasture94.7460.0090.0077.4285.71
Corn –no till96.6795.2495.12100100.00
Corn88.1092.6876.4788.0097.78
Corn—min. till100.00100.0068.7587.50100.00
Stone, steel, towers100.00100.0077.27100100.00
Alfalfa54.55100.00100100100.00
Soybean—clean100.0069.2383.33100100.00
Buildings, grass, trees, roads80.0075.0083.3388.8981.82
AA91.2087.1788.8393.2495.58
OA92.4583.4091.3592.5494.93
Kappa91.2480.7790.0091.4194.16
Precision80.2370.4582.2282.2391.87
Recall91.2087.1788.8393.2495.58
F1 score85.3677.9285.4087.9593.69
Table 6. Detailed classification outcomes for the Washington DC Mall dataset.
Table 6. Detailed classification outcomes for the Washington DC Mall dataset.
ClassPCASPCASPCA-NMIBgPCABgPCA-NMI
Shadow34.0457.1459.2672.7388.89
Tree99.2799.889999.7999.81
Roof10099.0599.08100100
Water100100100100100
Street93.6381.5189.4183.8697.62
Grass69.1286.4687.4295.8298.26
AA95.0587.3489.0392.0397.43
OA92.8092.0793.5795.9799.03
Kappa90.2289.3291.3094.5498.67
Precision93.6295.4996.7997.7799.51
Recall95.0587.3489.0292.0397.43
F1 score94.3391.2392.7594.8198.46
Table 7. Computational time in seconds (s) of the proposed and studied methods.
Table 7. Computational time in seconds (s) of the proposed and studied methods.
StageAVIRISHYDICE
PCABgPCA-NMIPCABgPCA-NMI
FE0.098 s0.017 s0.120 s0.067 s
FS1.200 s0.980 s1.100 s0.670 s
Total cost1.298 s0.997 s1.220 s0.737 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Islam, M.R.; Ahmed, B.; Hossain, M.A.; Uddin, M.P. Mutual Information-Driven Feature Reduction for Hyperspectral Image Classification. Sensors 2023, 23, 657. https://doi.org/10.3390/s23020657

AMA Style

Islam MR, Ahmed B, Hossain MA, Uddin MP. Mutual Information-Driven Feature Reduction for Hyperspectral Image Classification. Sensors. 2023; 23(2):657. https://doi.org/10.3390/s23020657

Chicago/Turabian Style

Islam, Md Rashedul, Boshir Ahmed, Md Ali Hossain, and Md Palash Uddin. 2023. "Mutual Information-Driven Feature Reduction for Hyperspectral Image Classification" Sensors 23, no. 2: 657. https://doi.org/10.3390/s23020657

APA Style

Islam, M. R., Ahmed, B., Hossain, M. A., & Uddin, M. P. (2023). Mutual Information-Driven Feature Reduction for Hyperspectral Image Classification. Sensors, 23(2), 657. https://doi.org/10.3390/s23020657

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop