Next Article in Journal
Learning-Based Clutter Mitigation with Subspace Projection and Sparse Representation in Holographic Subsurface Radar Imaging
Next Article in Special Issue
Fusion Classification of HSI and MSI Using a Spatial-Spectral Vision Transformer for Wetland Biodiversity Estimation
Previous Article in Journal
Modification of Temperature Lapse Rates and Cloud Properties during a Spatiotemporally Extended Dust Aerosol Episode (16–18 June 2016) over the Mediterranean Basin Based on Satellite and Reanalysis Data
Previous Article in Special Issue
Remote Sensing Scene Image Classification Based on Self-Compensating Convolution Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multiple Superpixel Graphs Learning Based on Adaptive Multiscale Segmentation for Hyperspectral Image Classification

1
College of Information and Communication Engineering, Harbin Engineering University, Harbin 150001, China
2
Key Laboratory of Advanced Marine Communication and Information Technology, Ministry of Industry and Information Technology, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(3), 681; https://doi.org/10.3390/rs14030681
Submission received: 3 January 2022 / Revised: 25 January 2022 / Accepted: 28 January 2022 / Published: 31 January 2022

Abstract

:
Hyperspectral image classification (HSIC) methods usually require more training samples for better classification performance. However, a large number of labeled samples are difficult to obtain because it is cost- and time-consuming to label an HSI in a pixel-wise way. Therefore, how to overcome the problem of insufficient accuracy and stability under the condition of small labeled training sample size (SLTSS) is still a challenge for HSIC. In this paper, we proposed a novel multiple superpixel graphs learning method based on adaptive multiscale segmentation (MSGLAMS) for HSI classification to address this problem. First, the multiscale-superpixel-based framework can reduce the adverse effect of improper selection of a superpixel segmentation scale on the classification accuracy while saving the cost to manually seek a suitable segmentation scale. To make full use of the superpixel-level spatial information of different segmentation scales, a novel two-steps multiscale selection strategy is designed to adaptively select a group of complementary scales (multiscale). To fix the bias and instability of a single model, multiple superpixel-based graphical models obatined by constructing superpixel contracted graph of fusion scales are developed to jointly predict the final results via a pixel-level fusion strategy. Experimental results show that the proposed MSGLAMS has better performance when compared with other state-of-the-art algorithms. Specifically, its overall accuracy achieves 94.312%, 99.217%, 98.373% and 92.693% on Indian Pines, Salinas and University of Pavia, and the more challenging dataset Houston2013, respectively.

1. Introduction

Hyperspectral images (HSIs) contain hundreds of bands [1,2,3], which provide rich spectral information and spatial information [4,5]. The HSI classification (HSIC) has always been an activate research topic. HSIC is widely used in the fields of agriculture [6,7], environment monitoring [8,9] and smart city design [10]. The representations of HSIs are spatially sparse while spectrally correlated; unlike the object-level images classification, HSIC needs to perform dense classification or regression over the whole image, hence how to extract pixel-level spectral features while utilizing local spatial information of HSIs is the core to process HSIs. Recently, computer vision has become a mainstream research direction and has been applied in many fields, such as agronomy [11,12], plant science [13], remote sensing and so on. After deep learning (DL) is applied to HSI classification, the detailed spectral–spatial features can be extracted [14,15,16,17]. The authors in [18] proposed a 1D+2D HSIC method, which uses a one-dimensional (1D) convolution kernel to extract spectral features and a two-dimensional (2D) convolution kernel to extract spatial features. Hamida et al. [19] directly used the three dimensional (3D) convolution neural networks (3D-CNN) to extract spatial–spectral features simultaneously. Considering the local details of HSIs, a framework based on kernel-guide deformable convolution and double-window joint bilateral filter is proposed to capture spatial–spectral features flexibly [20]. These DL-based methods achieved excellent performance but often require a large number of training samples. The limited number of training samples will obviously reduce the classification accuracy of these DL-based methods. However, manually annotating a large number of accurate training samples is extremely labor-intensive and time-consuming in a real scenario [21,22,23], and thus, the small labeled training sample size (STLSS) problem is always a challenge for HSIC [24,25].
Recently, superpixel (SP) segmentation has been widely applied to HSIC [26,27]. SP-based methods can generate an irregular homogeneous region with similar spatial information [28,29]. The local spatial similarity within an SP can easily enlarge the size of training samples, so the SP segmentation can be utilized to tackle the SLTSS problem [30,31]. The most commonly used SP segmentation methods are simple linear iterative clustering (SLIC) [32,33] and entropy rate SP (ERS) [34]. Initially, the support vector machine (SVM) and SP segmentation was first combined together for HSIC by [35]. Then, Zheng et al. [36] proposed an SP-guided training sample enlargement and the distance weighted linear regression-based classification (STSE_DWLR) to solve STLSS problem. Liu et al. [37] applied SP segmentation to deep learning and employed the graph convolutional networks (GCN) to extract global features. They proposed the CNN-enhanced GCN (CEGCN) to integrate the complementary advantages of CNN and GCN, and achieved the pixel- and superpixel-level feature fusion. Sellars et al. [38] introduced a superpixel contracted graph based transductive learning framework, named superpixel graph learning (SGL).The SGL applies the advantages of SP segmentation to HSIC and has good classification performance. These SP-based methods have usually applied the SP segmentation as a pretreating technique and have achieved excellent performance, however, how to select a suitable segmentation scale (the number of SPs), which will affect the results of subsequent methods, still need to be fixed. Some researches have demonstrated that different segmentation scales will seriously affect the classification accuracy of SP-based methods [39,40,41]. To tackle with this issue, Jia et al. [42] proposed an effective framework to calculate the optimal single segmentation scale and select multiscale, but this method did not consider the spatial complementarity between different segmentation scales. Although multiscale fusion can cope with the manual dependence dilemma [43,44,45,46], these non-adaptive multiscale selection methods will select the unwanted fusion scale and cause a negative contribution to the classification accuracy. Later, a novel adaptive multiscale segmentations (AMSs) framework which can select scales through an intra-inter discrimination index f ( s i ) was provided by [47]. AMSs adopted a two-step framework that first selects the basic scale and then selects the fusion scales according to the basic scale. The intra-inter discrimination index f ( s i ) has satisfactory performance in small scale intervals, however, for hyperspectral data sets with complex spatial distribution, such as that of the University of Pavia which requires large scale to segment, this index may lose its effect.
To reduce the bias and variance of classification results when the number of training samples is very small and select a group of multi-scale superpixel maps with complementary information, a multiple superpixel graphs learning method based on adaptive multiscale segmentation (MSGLAMS) is proposed for HSIC in this paper. In the proposed method, the multiscale superpixel segmentation maps through SLIC are generated first, and an optimal reference scale (ORS) selection algorithm (ORSSA) is used to select an ORS map as the basic scale. Then, a multiscale selection algorithm based on sparse representation is used to select fusion scale maps that have positive contribution to supplement the spatial information of the ORS. Finally, the classification results of a different fusion scale is obtained by SGL and the these results are fused by voting to obtain the final result. The main contributions of this paper are:
(1)
The proposed MSGLAMS adopts the multiscale-superpixel-based framework, which can reduce the adverse effect of improper selection of a superpixel segmentation scale on the classification accuracy while saving the cost to manually seek a suitable segmentation scale.
(2)
To make full use of the spatial information of different segmentation scales, a novel two-steps multiscale selection strategy is designed to adaptively select a set of complementary multiscales. Specifically, an optimal reference scale selection algorithm (ORSSA) which can select the single basic scale suitable for different data sets is proposed to select an optimal reference scale (ORS). Then, the multiscale selection method base on sparse representation (MSSR) is proposed to select fusion scales that have positive contribution to supplement the spatial information of ORS.
(3)
Multiple superpixel-based graphical models (SGL-based model), which are created via constructed superpixel contracted graph of determined scales (fusion scales pool), are adopted to jointly predict the final classification results, that is, pixel-level labels are determined via the voting results of these different models. This Boosting-like fusion strategy can significantly reduce the bias and instability of the final results, and keep the similar inductive bias of models.

2. Proposed Method

In this section, the proposed HSIC framework is explained in detail. As shown in Figure 1, for an HSI, the principal component analysis (PCA) [30] is first applied to extract top three principal components (PCs), and the candidate scale pool is constructed according to the size of the HSI. Then, the SLIC algorithm was used to segment HSI with different scales to obtain multiscale segmentation maps. A novel optimal reference scale (ORS) selection algorithm (ORSSA) is proposed to select the ORS suitable for different datasets as the basic scale, and the classification results of the ORS map are used to construct a sparse dictionary. Next, a multiscale selection method base on sparse representation (MSSR) is used to select fusion scales that can supplement the spatial information of enlarged samples of ORS map. Finally, multiple SGL models are obtained via constructed superpixel contracted graph with different scales (fusion scales), and the final results are jointly predicted via the voting results of these models.

2.1. Construction of the Candidate Scale Pool

Different hyperspectral datasets usually reflect different ground distribution. It is necessary to construct candidate scale pools [ S l o w e r , , S u p p e r ] suitable for different datasets, because the meaningless SP segmentation scale is not helpful to improve the classification accuracy and will waste computing time. As shown in Figure 2, when the scales of SP exceed the appropriate range, it is a meaningless segmentation. Small-scale segmentation maps cannot preserve the local detailed information of the HSI and can’t ensure that the pixels within the SP have similar land-cover categories (undersegmentation). The large-scale segmentation maps will lose large amount of spatial information (oversegmentation) [20,48].
Therefore, it is important to construct a candidate scale pool that can exclude the meaningless segmentation scale and include the indispensable segmentation scales. A candidate scale pool [ S l o w e r , , S u p p e r ] in [42] was defined as
Interval ( S ) = [ C C / R ]
where C is the real number of land cover classes and R is the texture ratio of the HSI. C is usually much smaller than the lower boundary of meaningful segmentation scale, especially for datasets such as that of Pavia University with large image sizes and a large number of edge pixels. In addition, the texture ratio R is easily affected by algorithm of texture detectors. The ideal interval is determined only by the dataset, so the lower boundary and upper boundary are defined as
S l o w e r = floor ( L × W K 1 ) S u p p e r = S l o w e r × C
where L and W are the length and width of the HSI. K 1 = arg min ( L , W ) . The S l o w e r and S u p p e r defined by this method will not miss meaningful segmentation scales.
If the scale interval is defined as [ S l o w e r : k : S u p p e r ] , the definition of S l o w e r and S u p p e r will only affect the runtime of the algorithm, but the setting of k will affect the selection of fusion scale. As shown in Figure 3, the number of SP cluster centers and the segmentation scale are not linear. If the setting of k is too small, the final fusion scale pool will contain many scales with same segmentation (i.e., non-contribution scales). If k is too large, it is easy to lose the important fusion scale (i.e., positive-contribution scales). So the candidate scale pool is divided into small scale pool [ S l o w e r : k s : S M ] , middle scale pool [ S M : k m : S L ] and large scale pool [ S L : k l : S u p p e r ] . Assuming Δ = S u p p e r S l o w e r , S M is defined as
S M = S l o w e r + Δ 6
S L is defined as
S L = S M + Δ 3 .
In addition, k s = k / 2 , k m = 2 k and k m = 3 k , which k = Δ / N S , N S is the number of multiscale segmentation maps (generally set to 30). Such a candidate scale pool can retain positive-contribution scales and exclude the meaningless segmentation maps.

2.2. Optimal Reference Scale Selection

To select multiple fusion scales, an ORS should be selected as the basic scale first. As shown in Figure 4, the influence of the SP scale on the overall accuracy (OA) is not linear. The small scale SP contains more similar spectral features, while the large scale SP contains richer spatial information. It is difficult to choose a single segmentation scale with good classification performance. Therefore, the optimal reference scale selection algorithm (ORSSA) is designed to choose the optimal single scale suitable for different datasets by calculating the segmentation evaluation indexes of different scale maps.
Small scale maps that contain pure spectral features can obtain high-quality enlarged samples. Large scale maps that contain better spatial information can obtain more enlarged samples. The optimal single scale generally contains similar spectral features and rich spatial spatial features, so the optimal single scale can be only sought with balanced spatial–spectral features in the middle scale pool.
Assuming that the middle scale pool is S = S 1 , S 2 , , S K , S S M : k m : S L . The kth scale has M SP regions, S k = { S 1 k , S 2 k , , S M k } . Then, p i S m k is the ith pixel within the mth SP of kth scale, p c is the corresponding cluster center of S m k , ( x i , y i ) and ( x c , y c ) are the spatial coordinates of p i and p c , respectively. In order to ensure the similar spectral features of the enlarged sample, the optimal single scale should contain small spectral differences between pixels within an SP, so the spectral evaluation index of the mth SP is defined as
f spectral ( S m k ) = i = 1 Im cos 1 p i T p c p i · p c I m
where I m is the number of pixel within the mth SP. Then, the overall spectral evaluation index of kth scale can be expressed as
F spectral S k = m = 1 M f spectral S m k M .
In addition, each SP should contain a larger number of pixels, so that SPs with poor spatial information will not easily be randomly selected to enlarge training samples. In order to ensure the size of enlarged samples, we define the spatial evaluation index of mth SP as
f spatial ( S m k ) = i = 1 Im ( x i x c 2 + y i y c 2 / d ) I m
where d is the length of the diagonal of mth SP S m k . The overall spatial evaluation index of kth scale can be expressed as
F spatial S k = m = 1 M f spatial S m k M .
The final segmentation evaluation index of kth scale is presented as
F ( S k ) = ( 1 λ ) F spectral ( S k ) + λ F spatial ( S k )
where F spectral ( S k ) and F spatial ( S k ) are the normalized results of (6) and (8), and λ is a spatial–spectral balance factor. The optimal reference scale ORS = arg max ( F ( S k ) ) . The pseudocode of the ORSSA is given in Algorithm 1.
Algorithm 1 ORSSA
Require: 
The middle scale pool, S = S 1 , S 2 , , S K , S S M : k m : S L ; Number of SPs in each scale, M = { M 1 , M 2 , , M K } ; parameter λ ;
Ensure: 
The optimal reference scale ORS;
1:
Obtain a series of middle scale maps through SLIC;
2:
for k = 1 to K do
3:
    for m = 1 to M k  do
4:
         f spectral ( S m k ) can be calculated by (5);
5:
         f spatial ( S m k ) can be calculated by (7);
6:
    end for
7:
     F spectral S k can be calculated by (6);
8:
     F spatial S k can be calculated by (8);
9:
    Compute the segmentation evaluation index of kth scale F spectral ( S k ) by (9);
10:
end for
11:
ORS = arg max ( F ( S k ) ) ;
12:
return ORS;

2.3. Superpixel Graph Learning

The superpixel graph learning (SGL) proposed by [38] is an SP contracted graph-based transductive learning framework for semi-supervised HSI classification. The SGL can be split into two major tasks: graph construction and label propagation.

2.3.1. Graph Construction

An HSI is denoted as I R L × W × B and the reduced image after PCA is denoted as I ^ R L × W × A . Assuming that the ith SP S i contains n i connected pixels, S i = p i , 1 , , p i , n i , three meaningful features need to be extracted from each SP ready for graph construction first. The mean features vector of ith SP S i m is defined as
S i m = j = 1 n i I ^ p i , j n i .
The weighted feature vector S i w can be expressed as
S i w = j = 1 J w i , z j S z j m
where the weight between adjacent SPs w i , z j is defined as
w i , z j = exp S z j m S i m 2 2 / h j = 1 J exp S z j m S i m 2 2 / h .
The final centroidal location of each SP S i p is calculated by
S i p = j = 1 n i p i , j n i .
Next, a weighted undirected graph G = ( V , E , W ) can be created from these previously discussed features and SP node set. The weight between two connected SPs S i and S j is constructed based on two Gaussian kernels and is given as
w i j = s i j l i j
where
s i j = exp ( β 1 ) S i ω S j w 2 2 β S i m S j m 2 2 σ s 2
l i j = exp S i p S j p 2 2 σ l 2
where β is the balance factor between the mean and weighted features, and the width of the Gaussian kernels are determined by σ s , σ l . The edge set is constructed using k-nearest neighbors. Therefore, the edge weights are defined as
W i j = w i j , if i is one of the k nearest neighbor of j or vice versa 0 , otherwise

2.3.2. Label Propagation

A set of labeled spectral pixels are first randomly selected from the original HSI. The final labeling can be specified using a matrix F R K × c . The cost function associated with the matrix F is given by
Q ( F ) = 1 2 i , j = 1 n W i j F i D i i F j D j j 2 + μ 2 i = 1 n F i Y i 2
where μ > 0 is a regularization parameter, Y R K × C is a labeling matrix contained the labels of K superpixels. F denotes the set of n × c matrices with non-negative entries. The labeling matrix is given by F * = argmin F F Q ( F ) . The above cost function has a closed form solution which reads: F * = β I α D ( 1 / 2 ) W D ( 1 / 2 ) Y , where β = μ / ( 1 + μ ) and α = 1 β . The final labeling of the nodes is then computed as: y i = argmax j c F i j . The final classification result map of the SP map can be obtained by calculating the label of each SP.

2.4. Multiscale Selection Based on Sparse Representation

As shown in Figure 4, the classification accuracy of a single scale is greatly affected by the selection of SP scale. When the number of samples is limited, the random selection of samples will seriously affect the spectral–spatial features of the enlarged samples after SP segmentation and thus affect the final classification results. The results shown in Figure 5 indicate that choosing a good single scale can improve the accuracy, but the robustness of the single scale is poor. The accuracy of the same single scale under different iterations is also very different because each iteration will randomly selects different samples to obtain corresponding enlarged samples. After using a multiscale fusion strategy, complementary spatial information can improve the classification accuracy under STLSS conditions, and complementary spectral features of enlarged samples can improve the robustness of classification accuracy. Hence, the multiscale fusion strategy has more advantages than the single-scale SP classification algorithm under the condition of STLSS. Hence, the accuracy and robustness of classification can be obviously improved by multiscale fusion.
The sparse representation can select pixels similar to the spectral features of the sample used to construct the dictionary [49,50,51,52], so when the classification result of ORS map is used to construct the dictionary, the fusion scale maps similar to ORS map can be selected by the representation residual. Consequently, the multiscale selection algorithm based on sparse representation (MSSR) can be exploited to select fusion scale maps that have positive contribution to supplement the spatial information of ORS.
The traditional sparse representation assumes that a pixel in HSI is represented as y R B , where B is the number of spectral bands. The structured dictionary is denoted by X = X 1 , X 2 , , X K R B × D , where K is the total number of classes and D is the number of training samples. In this algorithm, we suppose the HSI have K distinct classes, and structured dictionary X = X 1 , X 2 , , X K R B × N , where N is the number of all pixels of the ORS map. Assuming the classification result of ORS maps has N j pixels of jth class, the jth class subdictionary is X j = X 1 j , X 2 j , , X N j j . The superpixel S can be represented by a linear combination of these reference pixels as
S α 1 j X 1 j + α 2 j X 2 j + + α N j j X N j j .
The mth SP in the ith scale map can be represented as
S m i = X α m i + N m i
where α m i is the sparse coefficient matrix. The sparse coefficient matrix α m i can be calculated by
α ^ m i = arg min S m i X α m i F s . t . α m i 0 T
where · F is the Frobenius norm and the OMP [53] is implemented to solve this problem. Once α ^ m i is obtained, and the jth class reconstruction residual error which is the difference between original SP and reconstructed SP can be calculated by
r j ( S m i ) = S m i X j · α ^ m i 2 , j = 1 , 2 , , K .
The residual matrix can be expressed by calculating the minimum residual error of each SP. Mean value of residual matrix of ith scale map can be calculated by
E ( S i ) = m e a n ( ( S i ) ) = m = 1 M arg min ( r j ( S m i ) ) N c
where M is the number of SPs in the S i th scale map. Small mean value of means that the segmentation information is more similar to the ORS map. The finally fusion scale pool is determined by
S F = S 1 , S 2 , , S f s . t . S i S F , E ( S i ) T s
where T s is a threshold and T s can be presented as
T s = E m i n + E m a x E m i n κ , κ = 1 , 2 , , K s
where E m a x and E m i n are the maximum and minimum values of the E ( S i ) , respectively. κ = 1 means that all scales are selected for fusion, and T s = E m i n means that only the ORS map selected by ORSSA is used for fusion. The classification result maps of S F is obtained by the SGL. Subsequently, these result maps are fused to obtain the final classification result.

2.5. A Pixel-Level Fusion Strategy for Multiple Graphical Models

In general, superpixel-based methods assume that the pixels within the same superpixel share the same label. These superpixel-level predictions can be sub-optimal for dense prediction tasks due to the discrepancy between superpixel-level and pixel-level prediction. Moreover, the instability and bias of the single-scale superpixel based graph classifier comes from the insufficient generalization ability of the single graph model, that is, the single graph cannot well represent the distribution of a real HSI. To fix this gap, a pixel-level fusion strategy for multiple graphical models is proposed in this paper. Similar to Boosting [54], multiple graphical models, which are generated via constructed superpixel contracted graphs of determined scales S F = S 1 , S 2 , , S f , are adopted to jointly predict the label of each pixel. Thereby, assume the f constructed graphs as
G 1 = ( V 1 , E 1 , W 1 ) , G 2 = ( V 2 , E 2 , W 2 ) , , G f = ( V f , E f , W f ) .
For the pixel p i , the labels predicted by different models can be presented as
y i 1 = argmax j c F i j 1 , y i 2 = argmax j c F i j 2 , , y i f = argmax j c F i j f
where F i j can be obtain by Equation (18) in Section 2.3.2. The final label can be obtained by the majority voting rule, which presented as
y i = V o t i n g { y i 1 , y i 2 , , y i f } .
After predicting the label of each pixel, the final classification map is generated. The pseudocode of the MSGLAMS is given in Algorithm 2.
Algorithm 2 The proposed MSGLAMS
Require: 
Raw HSI dataset I R L × W × B with B bands; Training set A R B × m with m labeled samples; Parameter λ and κ ;
Ensure: 
A final classification map after multiscale fusion;
1:
Define the interval of candidate scale pool S lower and S upper by Equation (2);
2:
Construct a candidate scale pool composed of the small scale pool, middle scale pool and large scale pool by Equations (3) and (4);
3:
Reduced image I ^ R L × W × A is obtained by PCA.
4:
Apply SLIC algorithm on candidate scale pool to create multiscale segmentation maps.
5:
Select the ORS by Algorithm 1;
6:
Obtain the classification result of ORS ( l 1 , l 2 , , l L × W ) by SGL;
7:
Construct a sparse dictionary X = X 1 , X 2 , , X K R B × D by the classification results of ORS;
8:
for S i = S lower to S upper  do
9:
    Calculate the E ( S i ) through Equation (23);
10:
end for
11:
Determine the final fusion scale pool S F by Equation (24);
12:
Create multiple graphical models of S F by Equation (26).
13:
Fuse multiple scale maps to obtain the final classification result by Equations (27) and (28).
14:
return Final classification result;

3. Experimental Results and Analysis

In this section, the details of three experimental datasets and experimental settings are first introduced. Then, the results of the proposed method and other state-of-the-art methods on three HSI datasets are compared. Finally, the effect of the ORASS algorithm and the MSSR algorithm, and the parameter analysis of the proposed MSGLAMS is discussed.

3.1. Datasets

Three HSI datasets are selected for experiments, which are those of the Indian Pines, Salinas and University of Pavia. These datasets are collected by different sensors over different land covers, and have been selected for experiments by many related papers. The details of these datasets are given as follows:
  • The dataset was collected by an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor over an agricultural site in Indiana. The dataset consists of 145 × 145 pixels, 200 spectral channels in the wavelength range of 0.4–2.45 μ m. The spatial resolution of the data is 20 m. The ground truth map utilizes 10,249 labeled pixels with 16 different surface classes.
  • This image was also collected by the AVIRIS sensor over Salinas Valley, California. It contains 512 × 217 pixels with a spatial resolution of 3.7 m and 200 spectral channels over 0.4–2.5 μ m. The dataset has a spatial resolution of 3.7 m/pixel. Its ground truth contains 54,129 labeled pixels and 16 classes.
  • This dataset was acquired by the Reflective Optics System Imaging Spectrometer (ROSIS) sensor over the urban area surrounding the University of Pavia, Italy. The spatial size of the data is 610 × 340, and the spatial resolution is as high as 1.3 m/pixel. The image contains 115 spectral channels from 0.43 to 0.86 μm and has a spatial resolution of 1.3 m. Its ground truth contains 42,776 labeled pixels and 9 classes.
  • The Houston 2013 dataset was captured by an airborne spectrographic image sensor, which covers the area of University of Houston and its neighboring urban area. Houston 2013 dataset include 349 × 1905 pixels with a spatial resolution of 2.5 m and consists of 144 spectral channels ranging from 0.38 to 1.05 μ m, and the ground truth map utilizes 15,029 labeled pixels with 15 different land-cover classes. It adopted the standard training and testing sets given by the 2013 GRSS Data Fusion Contest.
Detailed information of different land cover classes and the corresponding false colors of Indian Pines, Salinas, and University of Pavia are listed in Table 1, Table 2, Table 3 and Table 4, respectively.

3.2. Experimental Settings

In order to verify the effectiveness of the proposed MSGLAMS comprehensively, the experiment can be divided into three parts. First, the comparison between MSGLAMS and other state-of-the-art methods is conducted. Second, as two key components of the the proposed MSGLAMS, the performances of ORASS and MSSR are also compared with their counterparts, respectively. Finally, the influence of key parameters of the MSGLAMS is analyzed. For the evaluation metrics, overall accuracy (OA), average accuracy (AA) and the kappa coefficient are adopted to quantify the classification performance. All the experiments are repeated ten times to obtain the final average performance. The hardware platform is Intel Xeon Silver 4210 2.4 GHz twelve-core processor CPU, GeForce RTX 3090 GPU and 32G memory.

3.2.1. Experimental Settings for Comparing with Other State of the Arts

In order to validate the performance of our proposed framework MSGLAMS, seven state-of-the-art algorithms are selected for comparison. These methods include three HSI classification methods based on SP segmentation: CEGCN [37], STSE_DWLR [36], and SGL [38]. Two HSI classification methods to solve small sample problems: KDCDWBF [20] and 3D-CNN [19]. Two common HSI classification methods: 1D+2D [18] and SVM. For the Houston, we still can’t provide the memory needed for GCN-M method even if we have adopted a decent hardware platform(CPU: Intel Xeon Silver 4210 2.4 GHz twelve-core processor; GPU: GeForce RTX 3090; Memory: 32G).

3.2.2. Experimental Settings of ORSSA

The purpose of proposed ORASS is to select an ORS map with best classification performance as basic scale. The three mainstream single scale selection algorithms are used to compare with ORSSA. First, the texture ratio was proposed in [55] to calculate the optimal single scale (TRS). The ORS calculated by TRS are 706, 1603 and 2765 for the Indian Pines, Salinas Scene, and University of Pavia, respectively. Second, the resolution of the dataset is used to calculate the optimal single scale (RES) in [42], and the RES are obtained as 1051, 3142 and 2216 for the Indian Pines, Salinas Scene, and University of Pavia. Finally, the ORS can be adaptively selected through inter-intra parameters (ASS) in [47], and the parameter setting is consistent with [47]. The spatial–spectral balance factor λ of ORSSA is set to 0.3 on the three datasets.

3.2.3. Experimental Settings of MSSR

The multi-scale comparison experiment is to compare the impact of different multiscale selection methods on the classification results of the same classifier SGL. The multiscale selection comparison methods include the multiscale selection formula (MSF) in [42], the multiscale SP number selection formula (MSNSF) in [56] and the adaptive multiscale selection (AMS) method proposed by [47]. The number of multiscales K in MSF and MSNSF is the same as the number of fusion scales of MSSN and the basic scale of MSF and MSNSF are both set to the ORS. The parameter κ in Equation (25) of MSSR is set to {4, 4, 7} for the Indian Pines, Salinas Scene, and University of Pavia, respectively. The influence of κ will be analyzed in detail in the Section 3.5.

3.3. Comparison of Results of Different Methods

In this section, seven state of the arts are compared with the proposed MSGLAMS framework. Table 5, Table 6, Table 7 and Table 8 report the classification accuracy of MSGLAMS and seven methods on the Indian Pines, Salinas and University of Pavia dataset, and the cells with a green background represent the highest accuracy. The classification maps obtained by these methods are illustrated in Figure 6, Figure 7, Figure 8 and Figure 9. It is easy to find that the accuracy of MSGLAMS far exceeds the common classification methods SVM and 1D+2D because the common classification methods cannot solve the problem of STLSS. Furthermore, the classification accuracy of MSGLAMS also far exceeds that of the small sample classification methods KDCDWBF and 3D-CNN. By comparing three small sample classification methods based on SP segmentation GNC-M, STSE_DWLR and SGL, the OA of MSGLAMS exceeds the best performing comparison method SGL 5.66%, 4.65% on the university of Pavia and the Indian Pines datasets, respectively. Kappa exceeds 7.23%, 4.57% than SGL on these two datasets. GNC-M, STSE_DWLR, SGL and our MSGLAMS perform well on the Salinas dataset, because the texture of Salinas dataset is not complicated. For the more challenging dataset Houston2013, MSGLAMS yields 92.906% overall accuracy, which outperforms the KDCDWBF and SGL by a large margin, i.e., 3.665% and 6.845%. It can be inferred that MSGLAMS also can achieve excellent performance even if the spatial distribution of HSI is complex. It should be noted that all accuracies except stressed grass and synthetic grass reached the top-1. Overall, these experimental results suggest that the proposed method can overcome the problem of unstable classification performance under small sample conditions and thus effectively improve the classification accuracy.

3.4. Analysis of Experimental Results of the ORASS and MSSR

To better display the effect and function of these two key components of the the proposed MSGLAMS, the performance of the ORASS and MSSR are compared with other single scale and multiscale selection methods, respectively. The experimental results of single-scale and multiscale selection methods are put in the same tables for comparison. Table 9, Table 10 and Table 11 show the classification accuracy OA of different scales on the three datasets. The bolded results represent the highest accuracy in the single-scale selection methods, and the cells with green background represent the highest accuracy in both single-scale selection methods and multiscale selection methods.
It can be seen from the results that among all single-scale selection algorithms, the single scale selected by ORASS has better performance. ASS is also an adaptive single-scale selection method, but because the training sample size is small, the intra-inter indexes f ( s i ) of the SP maps is not must positively related to the OAs. Therefore, the optimal single scale selected by the maximum value of the intra-inter index a r g m a x ( f ( s i ) ) may not have the best performance. TRS and RES do not have the ability to adaptively select scales according to different SP segmentation methods, so the fixed scale chosen by TRS and RES may not be suitable for the SGL classifier.
In addition, the classification accuracy of HSI has been significantly improved through a multiscale fusion strategy. Especially for the dataset of the University of Pavia with complex texture, the classification accuracy of multiscale fusion is much higher than single scale. However, choosing different fusion scales will supplement different spatial information of ORS and effect the final classification accuracy. The negative-contribution scales will destroy the quality and quantity of the enlarged samples and reduce the classification accuracy. For example, the OA of MSF and MSNSF is lower than MSSR, because the fixed fusion scales selected by MSF and MSNSF may contain scales with negative contribution. AMS has the ability to adaptively select multiple scales, but it still relies too much on intra-inter indexes f ( s i ) . The f ( s i ) is not must positively related to the OAs, so AMS does not have the best performance. As shown in Table 9, Table 10 and Table 11, the OA of MSSR is higher than other multiscale selection methods and much higher than single-scale SP segmentation. Because under the condition of STLSS, the fusion scales selected by MSSR have complementary enlarged samples information, which makes a positive contribution to improving the classification accuracy. Finally, with the increase of the number of samples, the OAs of different scales are very close, especially on the dataset Salinas with simple texture.

3.5. Parameter Analysis

In this section, the influence of three key parameters of the MSGLAMS method are analyzed in detail. First, we jointly analyzed the influence of the spatial–spectral balance parameter λ on the ORS selection and the influence of κ in Equation (25) on the classification results. Then, the effect of the number of training samples ({3, 5, 7, 10, 15} randomly selected samples per class) on several classifiers is examined on the sets of the Indian Pines, Salinas, and University of Pavia images.

3.5.1. Parameter Analysis of λ and κ

The setting of the spatial–spectral balance parameter λ determines the selection of ORS which used to construct the sparse dictionary, and ultimately affects the fusion scale pool. The interval of λ is set to [ 0 : 0.1 : 1 ] . When λ = 0 , ORSSA only selects the ORS based on the spectral evaluation index F spectral , i.e., ORS = S u p p e r , because S u p p e r contains the smallest spectral difference between pixels within the same SP. Conversely, λ = 1 means that ORASS only selects ORS based on the spatial evaluation index F spatial , i.e., ORS = S l o w e r , because each SP in S l o w e r scale map contains the largest number of pixels. The interval of κ in the fusion threshold T S is set to [ 1 : 2 : 11 ] . When κ = 1 , T S = E m a x , representing all scales fusion. For κ = 11 or higher, the performance of MSGLAMS is close to ORSSA. Figure 10 shows the influence of λ and κ on the overall accuracy on the three datasets, and deeper color corresponds to higher performance. For the Indian Pines dataset, MSGLAMS suggests the best performance when λ = 0.3 and κ = 5 . In addition, it can be seen from the result of κ = 11 , the ORS selected by ORASS has the best performance when the λ = 0.3 . When the κ value is small, especially κ = 5 , OA of MSGLAMS is higher than other κ values. For the Salinas dataset, MSGLAMS shows the best performance when λ = 0.7 and κ = 5 , and when λ = 0.4 , the ORS selected by ORSSA achieved the best performance. Finally, for the University of Pavia dataset, MSGLAMS has the best performance when λ = 0.3 and κ = 7 , and when λ = 0.3 , ORASS yield the best performance.
Two conclusions can be drawn from the above analysis. First, the small spatial–spectral balance factor λ (i.e., small spectral difference between pixels within the same SP) will get a better performing ORSSA. Consequently, in order to obtain a best-performance ORS to construct a sparse dictionary, λ can be set to 0.3. Second, fusing worse scales can reduce the accuracy (e.g., when κ = 1 or κ = 3 ), and too few fusion scales may not capture complete spatial information of multiscale (e.g., when κ = 9 or κ = 11 ). For normal HSIs, κ can be set to 5, but for datasets with complex textures (such as University of Pavia), setting κ = 7 will achieve better classification results.

3.5.2. Effect of Different Number of Training Samples

The effect of the number of training samples on several classifiers is examined on the Indian Pines, Salinas, and University of Pavia images in this section. The parameters for all the classifiers are kept the same as that in Section 3.3, and randomly select {3, 5, 7, 10, 15} samples per class as training samples. Figure 11 shows the OA of different classifiers when selecting randomly different numbers of training samples. It can be seen from the results in Figure 11 that the OA obtained by the eight classifiers will increase with the increasing of the number of training samples. In addition, MSGLAMS has the best performance under different training sample sizes. When the number of samples selected is smaller, the advantage of MSGLAMS is more obvious. Therefore, when the number of training samples is small (e.g., three or five training samples per class), the OA of MSGLAMS is significantly higher than other comparison methods. This result proves the obvious advantages of our proposed MSGLAMS under the condition of STLSS.

3.5.3. Running Time Comparison

To comprehensively evaluate our proposed method, the corresponding experiments are conducted on four datasets. Table 12 presents the runtime results of different methods. For the small-size datasets, such as Indian Pines, Salinas and University of Pavia, SVM and lightweight networks (1D+2D and 3D-CNN) yield shorter runtime. For the large-size dataset, Houston2013, the results suggest that all methods require more runtime. Compared with the first three methods, STSE_DWLR and SGL require more computation time. The key of GCN-M is the graph convolution network, which needs to feed the whole HSI into network, hence the GCN-M requires much more calculated cost. KDCDWBF also requires more runtime due to its extra post-processing procedure. Our proposed method also requires more computational cost because MSGLAMS needs spend a lot of time to build multiple superpixel contracted graphs. Especially when the size of an HSI is large, it requires more time to construct the large-scale superpixel maps.

4. Conclusions

This paper proposes a multiple superpixel graphs learning method based on adaptive multiscale segmentation for HSI classification (MSGLAMS). The MSGLAMS can be split into two major section: the selection of multiscale, and creating multiple superpixel-based graphical models to jointly predict the classification results via a pixel-level fusion. Aiming at the election of a multiscale, a novel two-steps multiscale selection strategy is designed to adaptively select a set of complementary multiscales. Initially, an optimal reference scale selection algorithm (ORSSA) is proposed to select the optimal reference scale (ORS) as a basic scale from the built candidate scale pool. Then, a multiscale selection algorithm based on sparse representation (MSSR) is developed to select the fusion scales, which has positive contribution to supplementing spatial information of ORS. Finally, multiple superpixel-based graphical models are obatined via a constructed superpixel contracted graph of determined scales (fusion scales), and the final results are jointly predicted via the voting results of these models. The proposed MSGLAMS adopts the multiscale-superpixel-based framework, which can reduce the adverse effect of improper selection of a superpixel segmentation scale on the classification accuracy while saving the cost to manually seek suitable segmentation scale. To make full use of the spatial information of different segmentation scales, a novel two-steps multiscale selection strategy is designed to adaptively select a set of complementary multiscales. To fix the bias and instability of a single model, multiple superpixel-based graphical models obatined by constructing superpixel contracted graph of fusion scales are developed to jointly predict the final results via a pixel-level fusion strategy. This Boosting-like prediction strategy can significantly improve the accuracy and robustness of the classification results, especially when the number of labeled samples is small. The experimental results on the three datasets indicate the superiority of the proposed MSGLAMS over several well-known classifiers in terms of qualitative and quantitative results, especially when the number of selected labeled samples is small.
The proposed method has achieved excellent performance on four datasets, but it needs a higher computational costs for HSIs with larger sizes. In the future, we will try to reduce computational complexity and adopt the CUDA to accelerate MSGLAMS.

Author Contributions

B.Q. and S.F. wrote and edited original draft preparation; C.Z. supervised the work and reviewed the manuscript; B.Q. and S.F. designed and implemented this framework. B.Q. and W.Z. tuned and evaluated this method by designing different experiments. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China Grant 62002083, 61971153, 61801142 and 62071136, and the Heilongjiang Provincial Natural Science Foundation of China Grant LH2021F012, and Heilongjiang Postdoctoral Foundation Grant LBH-Q20085, and the Fundamental Research Funds for the Central Universities Grant 3072021CF0801, 3072021CF0814, 3072021CF0807 and 3072021CF0808. (Corresponding author: Shou Feng).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, W.; Du, Q. Gabor-Filtering-Based Nearest Regularized Subspace for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1012–1022. [Google Scholar] [CrossRef]
  2. Jafarzadeh, H.; Mahdianpari, M.; Gill, E.; Mohammadimanesh, F.; Homayouni, S. Bagging and Boosting Ensemble Classifiers for Classification of Multispectral, Hyperspectral and PolSAR Data: A Comparative Evaluation. Remote Sens. 2021, 13, 4405. [Google Scholar] [CrossRef]
  3. Liu, Q.; Wu, Z.; Jia, X.; Xu, Y.; Wei, Z. From Local to Global: Class Feature Fused Fully Convolutional Network for Hyperspectral Image Classification. Remote Sens. 2021, 13, 5043. [Google Scholar] [CrossRef]
  4. Song, M.; Yu, C.; Xie, H.; Chang, C.I. Progressive Band Selection Processing of Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1762–1766. [Google Scholar] [CrossRef]
  5. Xue, B.; Yu, C.; Wang, Y.; Song, M.; Li, S.; Wang, L.; Chen, H.M.; Chang, C.I. A Subpixel Target Detection Approach to Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5093–5114. [Google Scholar] [CrossRef]
  6. Haboudane, D. Hyperspectral vegetation indices and novel algorithms for predicting green LAI of crop canopies: Modeling and validation in the context of precision agriculture. Remote Sens. Environ. 2004, 90, 337–352. [Google Scholar] [CrossRef]
  7. Merentitis, A.; Debes, C.; Heremans, R. Ensemble Learning in Hyperspectral Image Classification: Toward Selecting a Favorable Bias-Variance Tradeoff. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1089–1102. [Google Scholar] [CrossRef]
  8. Ellis, R.J.; Scott, P.W. Evaluation of hyperspectral remote sensing as a means of environmental monitoring in the St. Austell China clay (kaolin) region, Cornwall, UK. Remote Sens. Environ. 2004, 93, 118–130. [Google Scholar] [CrossRef]
  9. Shao, Z.; Fu, H.; Li, D.; Altan, O.; Cheng, T. Remote sensing monitoring of multi-scale watersheds impermeability for urban hydrological evaluation. Remote Sens. Environ. 2019, 232, 111338. [Google Scholar] [CrossRef]
  10. Schneider, S.; Murphy, R.J.; Melkumyan, A. Evaluating the performance of a new classifier—The GP-OAD: A comparison with existing methods for classifying rock type and mineralogy from hyperspectral imagery. ISPRS J. Photogramm. Remote Sens. 2014, 98, 145–156. [Google Scholar] [CrossRef] [Green Version]
  11. Chen, M.; Tang, Y.; Zou, X.; Huang, Z.; Zhou, H.; Chen, S. 3D global mapping of large-scale unstructured orchard integrating eye-in-hand stereo vision and SLAM. Comput. Electron. Agric. 2021, 187, 106237. [Google Scholar] [CrossRef]
  12. Cao, X.; Yan, H.; Huang, Z.; Ai, S.; Xu, Y.; Fu, R.; Zou, X. A Multi-Objective Particle Swarm Optimization for Trajectory Planning of Fruit Picking Manipulator. Agronomy 2021, 11, 2286. [Google Scholar] [CrossRef]
  13. Wu, F.; Duan, J.; Chen, S.; Ye, Y.; Ai, P.; Yang, Z. Multi-Target Recognition of Bananas and Automatic Positioning for the Inflorescence Axis Cutting Point. Front. Plant Sci. 2021, 12, 705021. [Google Scholar] [CrossRef] [PubMed]
  14. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature Extraction for Hyperspectral Imagery: The Evolution From Shallow to Deep: Overview and Toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  15. Liu, B.; Yu, A.; Yu, X.; Wang, R.; Gao, K.; Guo, W. Deep Multiview Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7758–7772. [Google Scholar] [CrossRef]
  16. Yu, C.; Zhao, M.; Song, M.; Wang, Y.; Li, F.; Han, R.; Chang, C.I. Hyperspectral Image Classification Method Based on CNN Architecture Embedding with Hashing Semantic Feature. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1866–1881. [Google Scholar] [CrossRef]
  17. Liu, D.; Han, G.; Liu, P.; Yang, H.; Sun, X.; Li, Q.; Wu, J. A Novel 2D-3D CNN with Spectral-Spatial Multi-Scale Feature Fusion for Hyperspectral Image Classification. Remote Sens. 2021, 13, 4621. [Google Scholar] [CrossRef]
  18. Audebert, N.; Le Saux, B.; Lefevre, S. Deep Learning for Classification of Hyperspectral Data: A Comparative Review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef] [Green Version]
  19. Ben Hamida, A.; Benoit, A.; Lambert, P.; Ben Amar, C. 3-D Deep Learning Approach for Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  20. Zhao, C.; Zhu, W.; Feng, S. Hyperspectral Image Classification Based on Kernel-Guided Deformable Convolution and Double-Window Joint Bilateral Filter. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Crawford, M.M. A Batch-Mode Regularized Multimetric Active Learning Framework for Classification of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6594–6609. [Google Scholar] [CrossRef]
  22. Tu, B.; Zhang, X.; Kang, X.; Zhang, G.; Wang, J.; Wu, J. Hyperspectral Image Classification via Fusing Correlation Coefficient and Joint Sparse Representation. IEEE Geosci. Remote Sens. Lett. 2018, 15, 340–344. [Google Scholar] [CrossRef]
  23. Hao, Q.; Li, S.; Kang, X. Multilabel Sample Augmentation-Based Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4263–4278. [Google Scholar] [CrossRef]
  24. Zhang, S.; Kang, X.; Duan, P.; Sun, B.; Li, S. Polygon Structure-Guided Hyperspectral Image Classification with Single Sample for Strong Geometric Characteristics Scenes. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  25. Imani, M.; Ghassemian, H. Band Clustering-Based Feature Extraction for Classification of Hyperspectral Images Using Limited Training Samples. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1325–1329. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Chen, Y. Multiscale Weighted Adjacent Superpixel-Based Composite Kernel for Hyperspectral Image Classification. Remote Sens. 2021, 13, 820. [Google Scholar] [CrossRef]
  27. Zhao, Y.; Yan, F. Hyperspectral Image Classification Based on Sparse Superpixel Graph. Remote Sens. 2021, 13, 3592. [Google Scholar] [CrossRef]
  28. Yang, X.; Zhao, M.; Gao, T.; Chen, J.; Zhang, J. Multiscale-Superpixel-Based SparseCEM for Hyperspectral Target Detection. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  29. Yu, H.; Gao, L.; Liao, W.; Zhang, B.; Pizurica, A.; Philips, W. Multiscale Superpixel-Level Subspace-Based Support Vector Machines for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2142–2146. [Google Scholar] [CrossRef]
  30. Shi, C.; Pun, C.M. Multiscale Superpixel-Based Hyperspectral Image Classification Using Recurrent Neural Networks With Stacked Autoencoders. IEEE Trans. Multimed. 2020, 22, 487–501. [Google Scholar] [CrossRef]
  31. Yang, C.; Bruzzone, L.; Zhao, H.; Tan, Y.; Guan, R. Superpixel-Based Unsupervised Band Selection for Classification of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 7230–7245. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Liu, K.; Dong, Y.; Wu, K.; Hu, X. Semisupervised Classification Based on SLIC Segmentation for Hyperspectral Image. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1440–1444. [Google Scholar] [CrossRef]
  33. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Susstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [Green Version]
  34. Liu, M.Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar] [CrossRef]
  35. He, Z.; Shen, Y.; Zhang, M.; Wang, Q.; Wang, Y.; Yu, R. Spectral-spatial hyperspectral image classification via SVM and superpixel segmentation. In Proceedings of the 2014 IEEE International Instrumentation and Measurement Technology Conference (I2MTC) Proceedings, Montevideo, Uruguay, 12–15 May 2014; pp. 422–427. [Google Scholar] [CrossRef]
  36. Zheng, C.; Wang, N.; Cui, J. Hyperspectral Image Classification with Small Training Sample Size Using Superpixel-Guided Training Sample Enlargement. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7307–7316. [Google Scholar] [CrossRef]
  37. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. CNN-Enhanced Graph Convolutional Network with Pixel- and Superpixel-Level Feature Fusion for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 8657–8671. [Google Scholar] [CrossRef]
  38. Sellars, P.; Aviles-Rivero, A.I.; Schonlieb, C.B. Superpixel Contracted Graph-Based Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4180–4193. [Google Scholar] [CrossRef] [Green Version]
  39. Li, H.; Feng, R.; Wang, L.; Zhong, Y.; Zhang, L. Superpixel-Based Reweighted Low-Rank and Total Variation Sparse Unmixing for Hyperspectral Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2021, 59, 629–647. [Google Scholar] [CrossRef]
  40. Cui, B.; Xie, X.; Ma, X.; Ren, G.; Ma, Y. Superpixel-Based Extended Random Walker for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3233–3243. [Google Scholar] [CrossRef]
  41. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of Hyperspectral Images by Exploiting Spectral–Spatial Information of Superpixel via Multiple Kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  42. Jia, S.; Deng, X.; Zhu, J.; Xu, M.; Zhou, J.; Jia, X. Collaborative Representation-Based Multiscale Superpixel Fusion for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7770–7784. [Google Scholar] [CrossRef]
  43. Liu, T.; Gu, Y.; Chanussot, J.; Dalla Mura, M. Multimorphological Superpixel Model for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6950–6963. [Google Scholar] [CrossRef]
  44. Lu, Q.; Wei, L. Multiscale Superpixel-Based Active Learning for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  45. Lu, T.; Li, S.; Fang, L.; Jia, X.; Benediktsson, J.A. From Subpixel to Superpixel: A Novel Fusion Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4398–4411. [Google Scholar] [CrossRef]
  46. Jia, S.; Deng, B.; Zhu, J.; Jia, X.; Li, Q. Superpixel-Based Multitask Learning Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2575–2588. [Google Scholar] [CrossRef]
  47. Leng, Q.; Yang, H.; Jiang, J.; Tian, Q. Adaptive MultiScale Segmentations for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5847–5860. [Google Scholar] [CrossRef]
  48. Jia, S.; Deng, B.; Zhu, J.; Jia, X.; Li, Q. Local Binary Pattern-Based Hyperspectral Image Classification with Superpixel Guidance. IEEE Trans. Geosci. Remote Sens. 2018, 56, 749–759. [Google Scholar] [CrossRef]
  49. Li, J.; Du, Q.; Li, Y.; Li, W. Hyperspectral Image Classification with Imbalanced Data Based on Orthogonal Complement Subspace Projection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3838–3851. [Google Scholar] [CrossRef]
  50. Prasad, S.; Labate, D.; Cui, M.; Zhang, Y. Morphologically Decoupled Structured Sparsity for Rotation-Invariant Hyperspectral Image Analysis. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4355–4366. [Google Scholar] [CrossRef]
  51. Dundar, T.; Ince, T. Sparse Representation-Based Hyperspectral Image Classification Using Multiscale Superpixels and Guided Filter. IEEE Geosci. Remote Sens. Lett. 2019, 16, 246–250. [Google Scholar] [CrossRef]
  52. Yu, H.; Shang, X.; Song, M.; Hu, J.; Jiao, T.; Guo, Q.; Zhang, B. Union of Class-Dependent Collaborative Representation Based on Maximum Margin Projection for Hyperspectral Imagery Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 553–566. [Google Scholar] [CrossRef]
  53. Tropp, J.; Gilbert, A. Signal Recovery From Random Measurements via Orthogonal Matching Pursuit. IEEE Trans. Inf. Theory 2008, 53, 4655–4666. [Google Scholar] [CrossRef] [Green Version]
  54. Su, H.; Yu, Y.; Du, Q.; Du, P. Ensemble Learning for Hyperspectral Image Classification Using Tangent Collaborative Representation. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3778–3790. [Google Scholar] [CrossRef]
  55. Jiang, J.; Ma, J.; Wang, Z.; Chen, C.; Liu, X. Hyperspectral Image Classification in the Presence of Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 851–865. [Google Scholar] [CrossRef] [Green Version]
  56. Sun, L.; Ma, C.; Chen, Y.; Shim, H.J.; Wu, Z.; Jeon, B. Adjacent Superpixel-Based Multiscale spatial–spectral Kernel for Hyperspectral Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1905–1919. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed MSGLAMS method. PCA: principal component (PC) analysis; SLIC: simple linear iterative clustering; ORSSA: optimal reference scale (ORS) selection algorithm; SGL: superpixel graph learning.
Figure 1. The flowchart of the proposed MSGLAMS method. PCA: principal component (PC) analysis; SLIC: simple linear iterative clustering; ORSSA: optimal reference scale (ORS) selection algorithm; SGL: superpixel graph learning.
Remotesensing 14 00681 g001
Figure 2. Segmatation maps with different superpixel scale S on University of Pavia. (a) S = 91. (b) S = 1376. (c) S = 6856. (b) is a meaningful segmentation scale map. (a) is an undersegmentation map. (c) is an oversegmentation map.
Figure 2. Segmatation maps with different superpixel scale S on University of Pavia. (a) S = 91. (b) S = 1376. (c) S = 6856. (b) is a meaningful segmentation scale map. (a) is an undersegmentation map. (c) is an oversegmentation map.
Remotesensing 14 00681 g002
Figure 3. The number of cluster centers corresponding to different SP scale maps obtained by SLIC on University of Pavia. The same number of cluster centers means the same segmentation map.
Figure 3. The number of cluster centers corresponding to different SP scale maps obtained by SLIC on University of Pavia. The same number of cluster centers means the same segmentation map.
Remotesensing 14 00681 g003
Figure 4. OAs of the SGL in different scales on three HSI datasets. (a) Indian Pines. (b) Salinas (c) University of Pavia.
Figure 4. OAs of the SGL in different scales on three HSI datasets. (a) Indian Pines. (b) Salinas (c) University of Pavia.
Remotesensing 14 00681 g004
Figure 5. OAs of different scales for 200 iterations on three HSI datasets. The ORS selected by ORSSA, worse single scale, and randomly selected multiscale fusion are labeled in orange ‘*’, blue ‘*’ and green ‘*’, respectively. The three horizontal lines are the mean values of 200 iterations, that is, the bias of different manners. The “*” curves can reflect the variances of different manners. The number of training samples are 3 per class. (a) Indian Pines. (b) Salinas (c) University of Pavia.
Figure 5. OAs of different scales for 200 iterations on three HSI datasets. The ORS selected by ORSSA, worse single scale, and randomly selected multiscale fusion are labeled in orange ‘*’, blue ‘*’ and green ‘*’, respectively. The three horizontal lines are the mean values of 200 iterations, that is, the bias of different manners. The “*” curves can reflect the variances of different manners. The number of training samples are 3 per class. (a) Indian Pines. (b) Salinas (c) University of Pavia.
Remotesensing 14 00681 g005
Figure 6. Classification maps of different methods on Indian Pines dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) GCN-M; (g) KDCDWBF; (h) STSE_DWLR; (i) SGL; (j) MSGLAMS (Proposed).
Figure 6. Classification maps of different methods on Indian Pines dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) GCN-M; (g) KDCDWBF; (h) STSE_DWLR; (i) SGL; (j) MSGLAMS (Proposed).
Remotesensing 14 00681 g006
Figure 7. Classification maps of different methods on Salinas dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) GCN-M; (g) KDCDWBF; (h) STSE_DWLR; (i) SGL; (j) MSGLAMS (Proposed).
Figure 7. Classification maps of different methods on Salinas dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) GCN-M; (g) KDCDWBF; (h) STSE_DWLR; (i) SGL; (j) MSGLAMS (Proposed).
Remotesensing 14 00681 g007
Figure 8. Classification maps of different methods on University of Pavia dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) GCN-M; (g) KDCDWBF; (h) STSE_DWLR; (i) SGL; (j) MSGLAMS (Proposed).
Figure 8. Classification maps of different methods on University of Pavia dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) GCN-M; (g) KDCDWBF; (h) STSE_DWLR; (i) SGL; (j) MSGLAMS (Proposed).
Remotesensing 14 00681 g008
Figure 9. Classification maps of different methods on Houston2013 dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) KDCDWBF; (g) STSE_DWLR; (h) SGL; (i) MSGLAMS (Proposed).
Figure 9. Classification maps of different methods on Houston2013 dataset; (a) Composite RGB image. (b) Groundtruth map. (c) SVM; (d) 3D-CNN; (e) 1D+2D; (f) KDCDWBF; (g) STSE_DWLR; (h) SGL; (i) MSGLAMS (Proposed).
Remotesensing 14 00681 g009
Figure 10. Influence of two parameters on classification accuracy on the three HSI data sets. (a) Indian Pines. (b) Salinas. (c) University of Pavia. λ is the spatial–spectral balance parameter, which determines the selection of the optimal reference scale (ORS). κ is the fusion threshold parameter, which determines the number of the multiscale maps. Deeper colors represent better performance. Generally, MSGLAMS can achieve excellent performance when the λ = 0.3 and κ = 5 .
Figure 10. Influence of two parameters on classification accuracy on the three HSI data sets. (a) Indian Pines. (b) Salinas. (c) University of Pavia. λ is the spatial–spectral balance parameter, which determines the selection of the optimal reference scale (ORS). κ is the fusion threshold parameter, which determines the number of the multiscale maps. Deeper colors represent better performance. Generally, MSGLAMS can achieve excellent performance when the λ = 0.3 and κ = 5 .
Remotesensing 14 00681 g010
Figure 11. Classification accuracy values of different numbers of labeled samples on three datasets. (a) Indian Pines; (b) University of Pavia; (c) Salinas.
Figure 11. Classification accuracy values of different numbers of labeled samples on three datasets. (a) Indian Pines; (b) University of Pavia; (c) Salinas.
Remotesensing 14 00681 g011
Table 1. Background color and number of land cover classes on indian pines.
Table 1. Background color and number of land cover classes on indian pines.
ColorLand Cover TypeNumbers
Stone-Steel-Towers93
Hay-windrowed478
Corn-mintill830
Soybean-notill972
Alfalfa46
Soybean-clean593
Grass-pasture483
Woods1265
Buildings-Grass-Tree-Drives386
Grass-pasture-mowed28
Corn237
Oats20
Corn-notill1428
Soybean-mintill1257
Grass-trees1257
Wheat205
Total Number10,249
Table 2. Background color and number of land cover classes on salinas.
Table 2. Background color and number of land cover classes on salinas.
ColorLand Cover TypeNumbers
Brocoli_green_weeds_12009
Brocoli_green_weeds_23726
Fallow1976
Fallow_rough_plow1394
Fallow_smooth2678
Stubble3959
Celery3579
Grapes_untrained11,271
Soil_vinyard_develop6203
Corn_senesced_green_weeds3278
Lettuce_romaine_4wk1068
Lettuce_romaine_5wk1927
Lettuce_romaine_6wk916
Lettuce_romaine_7wk1070
Vinyard_untrained7268
Vinyard_vertical_trellis1807
Total Number54,129
Table 3. Background color and number of land cover classes on Pavia university.
Table 3. Background color and number of land cover classes on Pavia university.
ColorLand Cover TypeNumbers
Asphalt6631
Meadows18,649
Gravel2099
Trees3064
Painted metal sheets1345
Bare Soil5029
Bitumen1330
Self-Blocking Bricks3682
Shadows947
Total Number42,776
Table 4. Background color and number of land cover classes on Houston2013.
Table 4. Background color and number of land cover classes on Houston2013.
ColorLand Cover TypeNumbers
Healthy grass1251
Stressed grass1254
Synthetic grass697
Trees1244
Soil1242
Water5029
Residential126
Commercial1244
Road1252
Highway122
Railway1235
Parking Lot 11233
Parking Lot 2469
Tennis Court428
Running Track660
Total Number15,029
Table 5. Classification accuracy for the Indian Pines image with ten labeled samples per class.
Table 5. Classification accuracy for the Indian Pines image with ten labeled samples per class.
Class NameSVM3D-CNN1D+2DGCN-MKDCDWBFSTSE_DWLRSGLMSGLAMS
Alfalfa0.495710.659400.504570.640921.000000.442641.000001.00000
Corn-notill0.468330.296840.274240.937930.875930.616500.734840.93067
Corn-mintill0.436150.352150.307600.859500.931720.693430.955950.99578
Corn0.453380.424140.384540.398230.951570.708950.955950.99578
Grass-pasture0.708570.632860.486500.861740.904970.555800.959830.84058
Grass-trees0.844650.840960.706110.990670.990340.862240.981940.99452
Grass-pasture-mowed0.432560.356180.250980.441030.888910.462401.000001.00000
Hay-windrowed0.922440.958330.903420.999141.000000.976391.000001.00000
Oats0.185110.211890.601690.325451.000000.323121.000001.00000
Soybean-notill0.321390.401870.330470.846580.859730.627630.813930.85185
Soybean-mintill0.499230.506660.568320.944200.757160.777890.886300.95723
Soybean-clean0.342180.218640.244350.762090.751330.536530.950260.81619
Wheat0.872830.838300.825770.994941.000000.934170.979491.00000
Woods0.805490.888820.803260.997230.999280.807351.000000.99921
Buildings-Grass-Trees-Drives0.381410.359190.314230.738690.997350.379271.000001.00000
Stone-Steel-Towers0.864950.822750.970960.853391.000000.481411.000001.00000
OA0.546440.534840.499260.898500.883720.705820.896620.94312
AA0.566520.548060.529810.749460.931770.636610.943620.95839
Kappa0.493720.481200.437180.883600.868030.669480.882280.92802
Table 6. Classification accuracy for the Salinas image with ten labeled samples per class.
Table 6. Classification accuracy for the Salinas image with ten labeled samples per class.
Class NameSVM3D-CNN1D+2DGCN-MKDCDWBFSTSE_DWLRSGLMSGLAMS
Brocoli_green_weeds_10.972130.771900.895751.000001.000001.000000.993501.00000
Brocoli_green_weeds_20.979790.837000.882491.000000.998990.985411.000001.00000
Fallow0.922270.690370.383181.000001.000000.999311.000001.00000
Fallow_rough_plow0.981310.960110.993480.999570.999310.993730.997830.99498
Fallow_smooth0.969580.889740.941610.978180.988840.998600.987260.98208
Stubble0.994690.991410.997500.999800.999260.999130.998731.00000
Celery0.988290.895130.925190.998720.997270.942860.989630.99888
Grapes_untrained0.620560.687780.700290.876800.991560.878040.987030.98501
Soil_vinyard_develop0.972190.900530.172411.000001.000000.998401.000001.00000
Corn_senesced_green_weeds0.825600.802950.754800.935110.960250.953720.979800.98383
Lettuce_romaine_4wk0.799530.625830.538600.996590.983630.900000.964080.99906
Lettuce_romaine_5wk0.971460.889200.837331.000001.000000.997151.000000.99637
Lettuce_romaine_6wk0.977380.808320.832150.996680.975780.997300.986750.98908
Lettuce_romaine_7wk0.908630.906560.945500.993960.987750.971580.944340.94393
Vinyard_untrained0.558980.251290.555530.935840.870100.793360.995590.99106
Vinyard_vertical_trellis0.973320.578470.864720.986971.000000.824731.000001.00000
OA0.830180.748080.706060.959950.976650.931390.991750.99217
AA0.900980.780410.763780.981210.984670.952080.989030.99152
Kappa0.812210.719550.688560.955460.973950.923550.990820.99128
Table 7. Classification accuracy for the university of Pavia image with ten labeled samples per class.
Table 7. Classification accuracy for the university of Pavia image with ten labeled samples per class.
Class NameSVM3D-CNN1D+2DGCN-MKDCDWBFSTSE_DWLRSGLMSGLAMS
Asphalt0.700130.843150.822340.973170.840830.877210.851530.97240
Meadows0.791820.743260.603750.898580.878750.847160.871670.99507
Gravel0.339540.815940.601630.941290.997120.779530.855430.92758
Trees0.642750.706430.648110.956900.771470.821800.892930.93570
Painted metal sheets0.982061.000001.000001.000001.000000.988730.995510.99926
Bare Soil0.430170.619360.360570.990161.000000.740311.000001.00000
Bitumen0.462580.648290.624090.999551.000000.682260.995451.00000
Self-Blocking Bricks0.712940.901150.805870.988290.981560.875640.885890.98099
Shadows0.999650.995530.996340.989750.521960.999310.991460.99894
OA0.694450.751160.633560.943190.893840.831380.895910.98373
AA0.673520.808120.718080.970850.887970.845770.926650.97888
Kappa0.603720.691620.552930.926150.862030.788240.865230.97843
Table 8. Classification accuracy for the Houston2013 image with ten labeled samples per class.
Table 8. Classification accuracy for the Houston2013 image with ten labeled samples per class.
Class NameSVM3D-CNN1D+2DKDCDWBFSTSE_DWLRSGLMSGLAMS
Healthy grass0.869780.695030.891930.958220.933280.954140.97161
Stressed grass0.827560.895450.961600.963310.685590.873450.77028
Synthetic grass0.320080.837210.986060.990640.924840.995740.99748
Trees0.906070.800340.925570.952750.472880.785600.89952
Soil0.811660.841340.952730.979530.727481.000001.00000
Water0.667800.485510.916620.867030.933130.893800.98230
Residential0.464020.754790.732280.860800.519780.720180.93766
Commercial0.511500.391130.648770.804510.334070.721560.81905
Road0.591230.185080.618950.836880.260360.705920.84684
Highway0.340810.357460.668460.790760.987270.984551.00000
Railway0.450400.484710.572220.875270.883030.875470.98275
Parking Lot 10.201570.357460.455680.845900.894290.834140.95451
Parking Lot 20.138620.061550.242150.845860.926040.830690.93196
Tennis Court0.761420.595320.881010.946170.982100.996101.00000
Running Track0.971480.946070.967090.964330.963190.953630.99248
OA0.593870.594930.753400.893500.718860.860910.92906
AA0.594640.579040.761410.898800.755860.849850.93905
Kappa0.559170.561170.733490.885410.697440.875200.92339
Table 9. OAs obtained by different single scale and multiscale on Indian Pines dataset with different number of training samples per calss.
Table 9. OAs obtained by different single scale and multiscale on Indian Pines dataset with different number of training samples per calss.
Samples per ClassSingle-Scale Selection MethodsMultiscale Selection Methods
TRSRESASSORASSMSFAMSMSNSFMSSR
30.751980.741200.724340.794130.836370.815290.777930.86604
50.828600.808730.765070.840990.852960.853740.828470.87032
70.856360.846460.764720.868770.873060.867110.898140.92643
100.865390.850940.815040.871350.918430.909450.899750.94214
150.880010.883900.855030.913580.954430.921070.929550.95648
Table 10. OAs of different single scale and multiscale on Salinas dataset with different number of training samples per calss.
Table 10. OAs of different single scale and multiscale on Salinas dataset with different number of training samples per calss.
Samples per ClassSingle-Scale Selection MethodsMultiscale Selection Methods
TRSRESASSORASSMSFAMSMSNSFMSSR
30.945560.954640.953220.967710.981160.964340.962270.99048
50.963580.959960.969660.970080.985010.980130.975720.99188
70.963790.976520.971210.988090.988770.985530.981080.99408
100.972970.984190.981370.990430.991540.986190.984230.99477
150.989730.986970.984410.990610.991610.990120.986830.99547
Table 11. OAs of different single scale and multi-scale fusion on university of Pavia dataset with different number of training samples per calss.
Table 11. OAs of different single scale and multi-scale fusion on university of Pavia dataset with different number of training samples per calss.
Samples per ClassSingle-Scale Selection MethodsMultiscale Selection Methods
TRSRESASSORASSMSFAMSMSNSFMSSR
30.827870.814730.836040.843760.909450.877030.826120.94128
50.831310.837930.848950.861530.917690.900200.888130.94588
70.853370.840370.865870.877980.922020.907980.927760.95404
100.871850.880420.880780.894510.938770.927670.947350.96327
150.913600.914870.916690.933490.955320.932460.958410.96454
Table 12. Runtime result (seconds) comparison on four data sets.
Table 12. Runtime result (seconds) comparison on four data sets.
Data SetSVM3D-CNN1D+2DGCN-MKDCDWBFSTSE_DWLRSGLMSGLAMS
Indian Pines3.3921.8379.16110.49131.3440.31764.6595.033
Salinas20.5550.3120.04750.28672.35103.08163.79569.90
University of Pavia4.1456.5717.61608.22714.3570.05351.76590.53
Houston201334.56101.4242.70-1157.9670.05520.31874.59
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, C.; Qin, B.; Feng, S.; Zhu, W. Multiple Superpixel Graphs Learning Based on Adaptive Multiscale Segmentation for Hyperspectral Image Classification. Remote Sens. 2022, 14, 681. https://doi.org/10.3390/rs14030681

AMA Style

Zhao C, Qin B, Feng S, Zhu W. Multiple Superpixel Graphs Learning Based on Adaptive Multiscale Segmentation for Hyperspectral Image Classification. Remote Sensing. 2022; 14(3):681. https://doi.org/10.3390/rs14030681

Chicago/Turabian Style

Zhao, Chunhui, Boao Qin, Shou Feng, and Wenxiang Zhu. 2022. "Multiple Superpixel Graphs Learning Based on Adaptive Multiscale Segmentation for Hyperspectral Image Classification" Remote Sensing 14, no. 3: 681. https://doi.org/10.3390/rs14030681

APA Style

Zhao, C., Qin, B., Feng, S., & Zhu, W. (2022). Multiple Superpixel Graphs Learning Based on Adaptive Multiscale Segmentation for Hyperspectral Image Classification. Remote Sensing, 14(3), 681. https://doi.org/10.3390/rs14030681

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop