Next Article in Journal
Detecting Morphing Attacks through Face Geometry Features
Next Article in Special Issue
A Siamese Neural Network for Non-Invasive Baggage Re-Identification
Previous Article in Journal
Fully 3D Active Surface with Machine Learning for PET Image Segmentation
Previous Article in Special Issue
Automatic Recognition of Dendritic Solidification Structures: DenMap
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble of ERDTs for Spectral–Spatial Classification of Hyperspectral Images Using MRS Object-Guided Morphological Profiles

1
State Key Laboratory of Desert and Oasis Ecology, Xinjiang Institute of Ecology and Geography, Chinese Academy of Sciences, Urumqi 830011, China
2
Research Center for Ecology and Environment of Central Asia, CAS, Urumqi 830011, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
Department of Geographical Information Science, Jiangsu Normal University, Xuzhou 221100, China
5
College of Surveying and Geoinformatics, Tongji University, Shanghai 200092, China
6
School of Geosciences & Info-Physics, Central South University, Changsha 410012, China
*
Author to whom correspondence should be addressed.
J. Imaging 2020, 6(11), 114; https://doi.org/10.3390/jimaging6110114
Submission received: 9 September 2020 / Revised: 16 October 2020 / Accepted: 21 October 2020 / Published: 26 October 2020
(This article belongs to the Special Issue Advances in Image Feature Extraction and Selection)

Abstract

:
In spectral-spatial classification of hyperspectral image tasks, the performance of conventional morphological profiles (MPs) that use a sequence of structural elements (SEs) with predefined sizes and shapes could be limited by mismatching all the sizes and shapes of real-world objects in an image. To overcome such limitation, this paper proposes the use of object-guided morphological profiles (OMPs) by adopting multiresolution segmentation (MRS)-based objects as SEs for morphological closing and opening by geodesic reconstruction. Additionally, the ExtraTrees, bagging, adaptive boosting (AdaBoost), and MultiBoost ensemble versions of the extremely randomized decision trees (ERDTs) are introduced and comparatively investigated for spectral-spatial classification of hyperspectral images. Two hyperspectral benchmark images are used to validate the proposed approaches in terms of classification accuracy. The experimental results confirm the effectiveness of the proposed spatial feature extractors and ensemble classifiers.

1. Introduction

Due to the technical evolution of optical remote sensors over the last few decades, now the remote sensing (RS) community can obtain diverse data sets with rich spatial, spectral and temporal information. In particular, hyperspectral sensors can provide detailed spectral information with hundreds of spectral wavelengths and can increase the possibility of more accurately discriminating materials of interest. Furthermore, the high (5.0 m ≤ spatial resolution ≤ 10.0 m) and very high (spatial resolution < 5.0 m) spatial resolution (HR, VHR) of some of these sensors enables the analysis of small spatial structures with unprecedented detail. However, the high dimensionality of hyperspectral images may lead to the Hughes phenomenon, in which the classification accuracy will be downgraded in case of the limited number of training samples and the classification method is not capable of handling high-dimensional data [1]. Additionally, while HR and VHR data solve the problem of being able to “see” structure objects and elements, they do not help in focusing on the extraction procedure [2]. Therefore, solutions of spectral dimensionality reduction via feature selection (FS) and feature extraction (FE), and design of specific spectral-spatial classifiers have been identified for hyperspectral image classification in recent years [3,4].
In contrast with the works of using supervised, semi-supervised and unsupervised FS and FE methods for hyperspectral dimensionality reduction, the design of specific spectral-spatial classifiers has gained more interest from the fields of HR/VHR hyperspectral image processing. Especially when considering improvements on classification accuracy brought by spatial features. Hence, spatial contextual information has been widely incorporated into advanced machine learning (ML)-based classifiers such as support vector machines (SVMs) [5], extreme learning machines (ELMs) [6], ensemble learning methods such as AdaBoost [7], random forests (RaF) and rotation forests (RoF) [8,9], ExtraTrees [10], XGBoost [11], and deep neural networks (DNNs) [12], for multispectral, hyperspectral and full polarimetric synthetic aperture radar (PolSAR) image classification.
According to their underlying model, spatial contextual information extraction approaches can be categorized into: (1) structural filter-based spatial processing methods using a fixed or adaptive structural element, such as edge-preserving filtering [13], local harmonic analysis [14,15,16], adaptive multidimensional Wiener filtering [17,18] and superpixel based filtering [19]; (2) random field models using a crisp neighborhood system, such as Markov random fields (MRF) [20,21], conditional random fields (CRF) [22] and discriminative random fields (DRF) [23]; (3) mathematical morphology (MM)-based approaches, such as morphological profiles (MPs) and extended MPs (EMPs) [24], object based MPs (OMPs) [25], attribute profiles (APs) [26], MPs with partial reconstruction (MPPR) [27], maximally stable extreme region guided MPs (MSER_MPs) [10] and its extended version EMSER_MPs [11]; (4) those based on image segmentation techniques [21,28,29]; (5) sparse representation based classification [30,31]; and (6) deep learning (DL)-based approaches [11,32,33,34]. Among these, MM-based approaches are likely the most widely used methods in the last ten years in the context of HR/VHR hyperspectral image processing [3,10,11,27].
Indeed, the advantages of MPs, EMPs, APs and MPPR in extracting spatial information from HR/VHR imagery has been clearly reported in many studies. However, being connected filters, they have the following limitations: (1) structural elements (SEs) with user-specified shape and size are inefficient for objects with diverse characteristics such as size, shape and homogeneity; (2) attribute filters (AFs) still suffer from the problem of leakage; and (3) limited numbers of SEs with specified sizes and shapes are unable to perfectly match all the sizes and shapes of the objects in a given image [3,10,27,35].
Inspired by the works in [25] and our previous work of [35], we adopt multiresolution segmentation (MRS)-based objects [36] as SEs for MPs extraction as an alternative solution. In particular, the original image is first segmented with the MRS technique as we demonstrated in [35]; in this procedure, multiscale values are provided for the purpose of multiscale spatial information extraction. Then, multiscale objects are exploited to extract OMPs by means of the basic principle of MM. Moreover, to avoid possible side effects from unusual minimum or maximum pixel values within objects, the OMPsM approach that contains additional mean pixel values within regions is proposed. Additionally, extended OMPs (EOMPs) and extended OMPsM (EOMPsM) are proposed by applying OMPs on the first three components after a principal component analysis (PCA) transformation executed on the spectral data. In contrast with OMPs in [25], we adopt MRS instead of arbitrary segmentation algorithm (ASA) for the image segmentation process, and shape and context features of objects are not considered in our MRS object-guided MPs. In contrast with our previous work in [35], object profiles such as roundness, compactness, rectangularity, density, asymmetry, border index, shape index, elliptic fit, minimum, maximum and standard deviation vales are not considered as well. At last, both works in [35] and [25] were conducted on multispectral imageries.
Following the quality of input features and a big enough set of training samples, the classifier robustness is the third component that affects the classification performance [8,37]. The extremely randomized decision tree (ERDT), a new tree induction algorithm that selects both attribute and cut-point splits, either totally or partially performed at random, and the ExtraTrees ensemble version were proposed for use in both classification and regression problems [38]. In our previous work, ERDT and ExtraTrees were investigated for their ability to classify three VHR multispectral images acquired over urban areas, and compared against decision tree (DT, C4.5), bagging, RaF, SVM and RoF in terms of classification accuracy and computational efficiency [10]. However, the performance of other ensemble versions (e.g., bagging, AdaBoost and MultiBoost) of ERDT for RS, particularly for hyperspectral image classification tasks using OMP and OMPsM features, has not yet been investigated. Hence, another contribution of this letter is to introduce and investigate the performance of bagging, AdaBoost and MultiBoost versions of ERDT in a hyperspectral image classification task.

2. Methods

2.1. Object-Guided MPs

Generally, morphological operators act on the values of the pixels by considering the neighborhood of the pixels determined by an SE with a predefined size and shape, based on two basic operators: dilation and erosion. In grayscale morphological reconstruction, two images and one SE are involved. One image, the marker f, contains the starting points for the transformation, while the other image, the mask g, constrains the transformation. According to the definitions from MM, morphological opening by reconstruction (OBR) of grayscale images can be obtained by first eroding (returning the minimum values of f contained in the specified SE) the input image and using it as a marker, while closing by reconstruction (CBR) can be obtained by complementing the marker image f, obtaining the OBR, and complementing the subsequent procedure [10,11,24,34,35]. In general, the object-guided morphological OBR can be obtained by first eroding the input image using segmented objects (where J S λ represent the numbers (S) of objects from MRS procedure with scale λ ) in the SE approach and by using the result as a marker in geodesic reconstruction by a dilation phase [35]:
O R O b j ( f ) = R f D [ ( f ( J i S λ J S λ ) ) ]
Similarly, we have
C R O b j ( f ) = R f E [ ( f ( J i S λ J S λ ) ) ]
where the object-guided CBR, obtained by complementing the image, contains the object-guided OBR (OOBR) using J i * J * as SEs and complements the resulting procedure:
C R O b j ( f ) = R f D [ ( f ( J i S λ J S λ ) ) ]
In MM, the erosion of f by b at any location (x, y) is defined as the minimum value of all the pixels in its neighborhood defined by b ( J i S λ J S λ in our case). In contrast, dilation returns the maximum value of the image in the window outlined by b. Then, the erosion and dilation operators can be defined as follows:
[ f ( J i S λ J S λ ) ] ( x , y ) = min ( s , t ) J S λ { f ( x + s , y + t ) } [ f ( J i S λ J S λ ) ] ( x , y ) = max ( s , t ) J S λ { f ( x + s , y + t ) }
Finally, if the structuring elements J i S λ J S λ are specified by objects, the OMPs of an image f can be defined as:
O M P s ( f ) = [ O R O b j ( f ) , f , C R O b j ( f ) ]
To avoid possible side effects from unusual minimum or maximum pixel values within objects, OMPsM are proposed by using extra mean pixel values that are contained within regions in an object-oriented manner:
O M P s M ( f ) = [ O M P s ( f ) , O O m e a n ( f ) ]
Although the use of MPs can help in creating an image feature set that has more discriminative information, the redundancy is still evident in the feature set, particularly for hyperspectral images. Therefore, feature extraction can be used to find the most important features first; then, morphological operators are applied [24]. After PCA is applied to the original feature set, EOMPs and EOMPsM can be obtained by applying the basic principles of OMPs and OMPsM described above to the first few (typically three) features.

2.2. ExtraTrees

The ERDT approach is a new decision tree (DT) induction algorithm that selects attribute and cut-point splits, either completely or partially at random, whereas the ExtraTrees algorithm is an ensemble version of unpruned ERDT, which follows by introducing the random committee-based ensemble criterion [38]. By comparing ExtraTrees with other DT-based ensemble methods such as bagging, boosting and RaF, the main differences can be outlined: (1) this new algorithm splits nodes by choosing cut points (which are responsible for a significant part of the error rates of tree-based methods) fully at random in the tree induction phase, which makes the tree structures independent of the target variable values of the learning samples, and (2) it uses the entire set of learning samples, rather than a bootstrap replica sample (typically adopted by the other DT methods), to grow trees.
Let X = { x τ } τ = 1 l denote a labeled training set with Y = { y τ } τ = 1 l as the labels, K represents the number of attributes randomly selected at each node, and η is the minimum sample size for splitting a node. An ERDT can be built by following the steps described in Algorithm 1.
Algorithm 1 Algorithmic steps to build an extremely randomized decision tree (ERDT) [38].
  • Inputs: labeled training set X , K and η .
  • Build_ERDT( X , Y, K, η ) as follows:
    (1)
    Return a leaf labeled by class frequencies in X if (1) | X l | < η or (2) all candidate attributes are constant in X , or (3) the output variable is constant in X .
    (2)
    Otherwise:
  • Randomly select K attributes { a 1 , , a K } without replacement among all candidates attributes;
  • Generate K splits { s 1 , , s K } | s i = [ a < a c ] , i = 1 , , K , where a is numerical attribute and a c is a cut-point uniformly drawn from [ a min X , a max X ] , which denote the minimal and maximal values of a in X , respectively;
  • Select a split s * = max i = 1 , , K { s i , X } ;
  • Split X into subsets X l and X r according to s * ;
  • Build single ERDT t f E R D T and t r E R D T from subsets X l and X r , respectively;
  • Create a node with the split s * , and attach t f E R D T and t r E R D T as left and right subtrees of this node;
  • Output: return the final resulting t E R D T .
Thereafter, the ExtraTrees ensemble algorithm can be built exploiting the random committee ensemble learning (EL) criterion, i.e., an ensemble of randomizable base classifiers is built using a different random number seed, and the final prediction is the average of the predictions generated by the individual base classifiers [10,38]. Similarly, bagging, AdaBoost and MultiBoost versions of ERDT can be realized following the corresponding ensemble construction criteria. Bagging, also called bootstrap aggregating, trains each model in the ensemble using a randomly drawn subset of the training set, and then votes with equal weight [39]. AdaBoost, an abbreviation for adaptive boosting, incrementally builds an ensemble in the sense that subsequent weak learners are tweaked in favor of those instances misclassified by previous classifiers [40]. MultiBoost can be viewed as a combination of AdaBoost with bagging, which can harness both AdaBoost’s bias and variance reduction with bagging’s superior variance reduction to produce a committee with lower error, also offering, as an advantage over AdaBoost, the suitability to parallel execution [41].

3. Data Sets

The first hyperspectral image was acquired by the Reflective Optics System Imaging Spectrometer (ROSIS) optical sensor, which provides 115 bands with a spectral range coverage ranging from 0.43 μm to 0.86 μm. The main objective of the ROSIS project is the detection of spectral fine structures especially in coastal waters. This task determined the selection of the spectral range, bandwidth, number of channels, radiometric resolution and its tilt capability for sun glint avoidance. However, ROSIS can be used just as well for the monitoring of spectral features above land or within the atmosphere. The image shown in Figure 1a depicts the Engineering School of Pavia University (Pavia, Italy) with the geometric resolution of 1.3 m. The image has 610 × 340 pixels with 103 spectral channels, where 12 very noisy bands were discarded manually after the data acquisition. The validation data refer to nine land cover classes (as shown in Figure 1). This scene was provided by Professor Paolo Gamba from the Telecommunications and Remote Sensing Laboratory, Pavia University (Pavia, Italy).
The second hyperspectral image was acquired at a spatial resolution of 2.5 m by the National Science Foundation (NSF) -funded Center for Airborne Laser Mapping (NCALM) over the University of Houston campus and the neighboring urban area, on 23 June 2012 (Figure 2). The 15 classes of interest selected by the Institute of Electrical and Electronics Engineers Geoscience and Remote Sensing Society (IEEE GRSS) Image Analysis and Data Fusion Technical Committee for organizing the 2013 Data Fusion Contest (DFC) are reported for both the training and validation sets [42]. Originally, this image has 349 × 1905 pixels with 144 spectral bands in the spectral range between 380 and 1050 nm. In our experiment, dense cloud-covered area at the right part and total of nine blank pixel lines at the upper and lower image edges were removed, which result in subset image with the size of 340 × 1350 pixels.

4. Results

4.1. Experimental Configuration

The free parameters of ERDT, where K represents the number of the attributes set, are the same as the default for the C4.5 algorithm used in bagging and RaF. The overall accuracy (OA) and kappa statistic are used to evaluate the classification performances of these methods. In the case of multiclass classification, OA is usually calculated by dividing the sum of diagonal numbers, which represent correctly classified instances, by the total number of reference instances in the confusion matrix:
O A = i N T p i i N T n i
where T p i represents numbers of the correctly classified instances for class i, T n i represents the total number of instances from class i, and there are a total of N classes.
To generate MPs and MPPR, we apply a disk shape SE with n = 10 openings and closings by conventional and partial reconstructions, ranging from one to ten with a step-size increment of one. This choice results in a total of 2163 = 103 + 103 × 10 × 2 and 3024 = 144 + 144 × 10 × 2 dimensional stacked data sets using original spectral bands and a total of 70 = 10 + 3 × 10 × 2 and 67 = 7 + 3 × 10 × 2 dimensional stacked data sets using PCA-transformed features, for Pavia University and GRSS-DFC2013, respectively. Note that only the first ten and seven PCA-transformed features from Pavia University and GRSS-DFC2013, respectively, are considered in the experiments. For fair comparison purposes, we set a total of ten scales for MRS in the image segmentation phase. For instance, the scale parameter is increased from 10 to 55 by a step-size of five to produce a total of 10 scale segmentation results. The segmentation result, which is crucial for guiding MPs, typically relies on the scale parameters that are highly dependent on the spatial resolution and geometrical complexity of the image under consideration. Hence, in the next experiment, we examine the performance of OMPs, OMPsM, EOMPs and EOMPsM with different scale sets. Note that OMPsM and EOMPsM also contain the mean pixel values within objects that produce 3193 = 103 + 103 × 10 × 3, 4464 = 144 + 144 × 10 × 3, 100 = 10 + 3 × 10 × 3 and 97 = 7 + 3 × 10 × 3 dimensional stacked data sets using the original spectral bands and PCA-transformed features for Pavia University and GRSS-DFC2013, respectively.

4.2. Results and Analysis

Figure 3 shows the examples of OBR, opening by partial reconstruction (OBPR), and the proposed OOBR with different parameter sets using the second principal component of the Pavia University data. The range of disk shape SEs in OBR and OBPR were set between 6 to 10, while the scale parameters of MRS and OOBR were set between 60 to 100 empirically. A comparison of the results in the first row indicates that OBPR is more capable of modeling the attributes of different objects than OBR from a sequence of SEs, in accordance with the finding in [12]. However, many large objects and boundaries between different objects that should have appeared were removed at a very small scale after OBPR. In contrast, OOBR maintains the object information between boundaries exactly as in the original by affecting only the brightness or darkness of the objects with different scale parameters. In other words, effects from the scale parameter of OOBR are much smaller than effects from the scale of OBPR.
In Figure 4, we present the results for various spatial feature extractors with different parameter sets using an SVM with an radial bias function (RBF) kernel to evaluate the performance of EOMPs and EOMPsM on the considered data sets. A total of 10 rounds were executed for each experiment for the purpose of an objective evaluation.
The graphs confirm the superiority of the proposed feature extraction methods in contrast to MPs and MPPR, specifically with the best improvements obtained by EOMPsM, and this is valid for both data sets (Figure 4c,d,g,h). However, the effects of the segmentation scale parameter in MRS are different for the considered data sets using the original spectral and the PCA-transferred features. For instance, the best OA curves are achieved by OMPs using the original spectral bands of ROSIS university data with the MRS scale ranges set as 40 to 400 with 40 sequence steps (see Figure 4a). In contrast, the best OA curves are achieved by OMPsM using the original spectral bands with the MRS scale range set from 310 to 400, with 10 sequence steps (see Figure 4b).
Interestingly, the superiority of the larger scale set relative to the smaller one is no longer true when using the original spectral features, because some noise corrupted bands mislead partial reconstruction in MPs and MPPR and the image segmentation procedures in OMPs and OMPsM (see Figure 4a,b,e,f). Additionally, EOMPsM with a larger scale set could limit and even degrade the classification accuracy, whereas a single mean value was assigned to different targets contained in single large segmented objects (see Figure 4c,d,g,h). Summarizing these results, OMPs and EOMPs are more suitable to accommodate the original spectral and PCA-transformed features for a larger MRS scale range set with a larger sequence step, while OMPsM and EOMPsM are more suitable for a larger MRS scale range but with a smaller sequence step.
Figure 5 shows the OA values with respect to the number of trees in bagging, RaF and ExtraTrees, and with respect to the number of iterations in AdaBoost and MultiBoostAB ensemble classifiers. According to these graphs, there are no prominent improvements or decreasing trends for a tree size greater than 100 in most of the cases, a result consistent with the findings in other studies [8,10]. Moreover, it is clear that the bagging ensemble of ERDT (Bag(ERDT)) is uniformly better than the bagging ensemble of the conventional C4.5 approach (Bag(C4.5)) in all classification scenarios in terms of classification accuracy. Instead, the performance of the MultiBoostAB and AdaBoost ensemble of ERDT (MB(ERDT) and AB(ERDT)) are not constantly superior to the MultiBoostAB and AdaBoost ensemble of C4.5 (MB(C4.5) and (C4.5)) using different features on the two data sets. Specifically, MB(C4.5) and AB(C4.5) show better OA values than MB(ERDT) and (ERDT) using MPs and MPPR but show lower OA values using OMPs features for considered data sets, and similar OA values shown by using OMPsM features of Pavia University data set but lower values by using OMPsM features from the GRSS-DFC2013 data set. MB(ERDT) and AB(ERDT) show better results using OMPs and OMPsM features but lower results using MPs and MPPR features, which can be explained by the fact that: (1) fewer, but harder to be correctly classified, instances are those focused on by the MultiBoostAB and AdaBoost criteria, which could further weaken ERDT and lead to an overly abundant diversity that hindered the construction of an improved ensemble scenario, (2) however, this shortage could be overcome by exploiting advanced discriminative features. Other solutions for this limitation could be either (1) early stopping of the ensemble or (2) critical tuning of the parameters of ERDT in each iteration step.
Finally, Figure 6 and Figure 7 present the best classification maps corresponding to the highlighted values in Table 1 and Table 2, with OA and the kappa statistics for MPs, MPPR, OMPs and OMPsM features extracted from the original raw bands or the PCA-transformed features.
Once again, all classifiers uniformly and clearly confirm the effectiveness of the proposed MRS OMPs over the conventional MPs and MPPR approaches. For instance, the best classification results, with the highest OA (98.75%) and kappa statistic (0.98) values, were achieved by MultiBoost(C4.5) using EOMPsM features on the ROSIS university data, and by AdaBoost(ERDT) using OMPsM features on GRSS-DFC2013 data (OA = 96.59%, kappa statistic = 0.96). If we compare the ensemble versions of ERDT, it is clear that ExtraTrees is better than Bag(C4.5) and is comparable to RaF(C4.5), and that the best improvement in OA values is achieved by either AdaBoost or MultiBoost ensemble (see the numbers in bold in Table 1 and Table 2). Additionally, regarding EOMPs, the superiority of EOMPsM over EMPs is clear, its performance is comparable to the one by EMPPR, or even better in some cases.

5. Conclusions

In this study, we propose the concept of OMPs for spatial feature extraction in high-resolution hyperspectral images, by using multiscale objects after multi resolution segmentation as the SEs. Additionally, ExtraTrees, bagging, AdaBoost, and MultiBoost ensemble versions of the ERDT algorithm are introduced and comparatively investigated on two benchmark hyperspectral data sets. The experimental results confirm the effectiveness of the proposed OMPs, OMPs(M) and their extended versions. In addition, the superiority of EOMPsM over the conventional MPs and MPPR is reported. In the evaluation of the adopted classifiers, the bagging ensemble of ERDT is better than the bagging version of C4.5, and ExtraTrees is better than Bag(C4.5) but comparable to RaF(C4.5). The best improvements are reached by the AdaBoost or MultiBoost ensemble of ERDT using OMPsM extracted from the original bands, or EOMPsM extracted from the PCA-transformed features.
Future works will focus on the role of self-adaptive segmentation scale selection for multiscale segmentation in the usefulness of OMPs and EOMPsM. The early steps and self-adaptive parameter tuning of individual ERDT in the AdaBoost and MultiBoost ensemble framework will also be investigated.

Author Contributions

Conceptualization and methodology, A.S.; validation, E.L. and W.W.; formal analysis, A.S.; investigation, A.S. and S.L.; writing—original draft preparation, A.S.; writing—review and editing, A.S., S.L. and Z.M.; visualization, A.S. and W.W.; project administration, A.S.; funding acquisition, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (grants no. 42071424, 41601440), the Youth Innovation Promotion Association Foundation of the Chinese Academy of Sciences (2018476), and West Light Foundation of the Chinese Academy of Sciences (2016-QNXZ-B-11).

Acknowledgments

The authors would like to thank P. Gamba and his team, who freely provided the Pavia University hyperspectral data set with ground truth information. They would also like to thank the Institute of Electrical and Electronics Engineers Geoscience and Remote Sensing Society (IEEE GRSS) Image Analysis and Data Fusion Technical Committee for organizing the 2013 Data Fusion Contest and for freely providing the Houston University hyperspectral data set with ground truth information used in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shahshahani, B.M.; Landgrebe, D.A. The effect of unlabeled samples in reducing the small sample size problem and mitigating the Hughes phenomenon. IEEE Trans. Geosci. Remote Sens. 1994, 32, 1087–1095. [Google Scholar] [CrossRef] [Green Version]
  2. Gamba, P.; Dell’Acqua, F.; Stasolla, M.; Trianni, G.; Lisini, G. Limits and challenges of optical high-resolution satellite remote sensing for urban applications. In Urban Remote Sensing—Monitoring, Synthesis and Modelling in the Urban Environment; Yang, X., Ed.; Wiley: New York, NY, USA, 2011; pp. 35–48. [Google Scholar]
  3. Plaza, A.; Benediktsson, J.A.; Boardman, J.W.; Brazile, J.; Bruzzone, L.; Camps-Valls, G.; Chanussot, J.; Fauvel, M.; Gamba, P.; Gualtieri, A.; et al. Recent advances in techniques for hyperspectral image processing. Remote Sens. Environ. 2009, 113, S110–S122. [Google Scholar] [CrossRef]
  4. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in spectral-spatial classification of hyperspectral images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef] [Green Version]
  5. Mountrakis, G.; Im, J.; Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  6. Samat, A.; Du, P.; Liu, S.; Li, J.; Cheng, L. E2LMs: Ensemble Extreme Learning Machines for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1060–1069. [Google Scholar] [CrossRef]
  7. Chan, J.C.W.; Paelinckx, D. Evaluation of Random Forest and Adaboost tree-based ensemble classification and spectral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sens. Environ. 2008, 112, 2999–3011. [Google Scholar] [CrossRef]
  8. Du, P.; Samat, A.; Waske, B.; Liu, S.; Li, Z. Random forest and rotation forest for fully polarized SAR image classification using polarimetric and spatial features. ISPRS J. Photogramm. Remote Sens. 2015, 105, 38–53. [Google Scholar] [CrossRef]
  9. Xia, J.; Du, P.; He, X.; Chanussot, J. Hyperspectral remote sensing image classification based on rotation forest. IEEE Geosci. Remote Sens. Lett. 2014, 11, 239–243. [Google Scholar] [CrossRef] [Green Version]
  10. Samat, A.; Persello, C.; Liu, S.; Li, E.; Miao, Z.; Abuduwaili, J. Classification of VHR Multispectral Images Using ExtraTrees and Maximally Stable Extremal Region-Guided Morphological Profile. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3179–3195. [Google Scholar] [CrossRef]
  11. Samat, A.; Li, E.; Wang, W.; Liu, S.; Lin, C.; Abuduwaili, J. Meta-XGBoost for hyperspectral image classification using extended MSER-guided morphological profiles. Remote Sens. 2020, 12, 1973. [Google Scholar] [CrossRef]
  12. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  13. Kang, X.; Li, S.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification with edge-preserving filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  14. Shen, L.; Jia, S. Three-dimensional Gabor wavelets for pixel-based hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2011, 49, 5039–5046. [Google Scholar] [CrossRef]
  15. Qian, Y.; Ye, M.; Zhou, J. Hyperspectral image classification based on structured sparse logistic regression and three-dimensional wavelet texture features. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2276–2291. [Google Scholar] [CrossRef] [Green Version]
  16. Zhu, Z.; Jia, S.; He, S.; Sun, Y.; Ji, Z.; Shen, L. Three-dimensional Gabor feature extraction for hyperspectral imagery classification using a memetic framework. Inform. Sci. 2015, 298, 274–287. [Google Scholar] [CrossRef]
  17. Peng, J.; Zhou, Y.; Chen, C.P. Region-kernel-based support vector machines for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4810–4824. [Google Scholar] [CrossRef]
  18. Lu, Z.; He, J. Spectral–spatial hyperspectral image classification with adaptive mean filter and jump regression detection. Electron. Lett. 2015, 51, 1658–1660. [Google Scholar] [CrossRef]
  19. Bourennane, S.; Fossati, C.; Cailly, A. Improvement of classification for hyperspectral images based on tensor modeling. IEEE Geosci. Remote Sens. Lett. 2010, 7, 801–805. [Google Scholar] [CrossRef]
  20. Jackson, Q.; Landgrebe, D.A. Adaptive Bayesian contextual classification based on Markov random fields. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2454–2463. [Google Scholar] [CrossRef] [Green Version]
  21. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  22. Zhong, P.; Wang, R. Learning conditional random fields for classification of hyperspectral images. IEEE Trans. Image Process. 2010, 19, 1890–1907. [Google Scholar] [CrossRef] [PubMed]
  23. Kumar, S. Discriminative random fields: A discriminative framework for contextual interaction in classification. In Proceedings of the 2003 Ninth IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003; pp. 1150–1157. [Google Scholar]
  24. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  25. Geiß, C.; Klotz, M.; Schmitt, A.; Taubenböck, H. Object-based morphological profiles for classification of remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2016, 54, 5952–5963. [Google Scholar] [CrossRef]
  26. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of hyperspectral images by using extended morphological attribute profiles and independent component analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef] [Green Version]
  27. Liao, W.; Chanussot, J.; Dalla Mura, M.; Huang, X.; Bellens, R.; Gautama, S.; Philips, W. Taking Optimal Advantage of Fine Spatial Resolution: Promoting partial image reconstruction for the morphological analysis of very-high-resolution images. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–28. [Google Scholar] [CrossRef] [Green Version]
  28. Ma, L.; Li, M.; Ma, X.; Cheng, L.; Du, P.; Liu, Y. A review of supervised object-based land-cover image classification. ISPRS J. Photogramm. Remote Sens. 2017, 130, 277–293. [Google Scholar] [CrossRef]
  29. Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J. Spectral–spatial classification of hyperspectral imagery based on partitional clustering techniques. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2973–2987. [Google Scholar] [CrossRef]
  30. Liu, J.; Wu, Z.; Wei, Z.; Xiao, L.; Sun, L. Spatial-spectral kernel sparse representation for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2013, 6, 2462–2471. [Google Scholar] [CrossRef]
  31. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification via multiscale adaptive sparse representation. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7738–7749. [Google Scholar] [CrossRef]
  32. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  33. Ma, X.; Wang, H.; Geng, J. Spectral–spatial classification of hyperspectral image based on deep auto-encoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  34. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  35. Samat, A.; Yokoya, N.; Du, P.; Liu, S.; Ma, L.; Ge, Y.; Issanova, G.; Saparov, A.; Abuduwaili, J.; Lin, C. Direct, ECOC, ND and END Frameworks—Which One Is the Best? An Empirical Study of Sentinel-2A MSIL1C Image Classification for Arid-Land Vegetation Mapping in the Ili River Delta, Kazakhstan. Remote Sens. 2019, 11, 1953. [Google Scholar] [CrossRef] [Green Version]
  36. Petrovic, A.; Escoda, O.D.; Vandergheynst, P. Multiresolution segmentation of natural images: From linear to nonlinear scale-space representations. IEEE Trans. Image Process. 2004, 13, 1104–1114. [Google Scholar] [CrossRef] [Green Version]
  37. Olofsson, P.; Foody, G.M.; Herold, M.; Stehman, S.V.; Woodcock, C.E.; Wulder, M.A. Good practices for estimating area and assessing accuracy of land change. Remote Sens. Environ. 2014, 148, 42–57. [Google Scholar] [CrossRef]
  38. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–4. [Google Scholar] [CrossRef] [Green Version]
  39. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  40. Rätsch, G.; Onoda, T.; Müller, K.R. Soft margins for AdaBoost. Mach. Learn. 2001, 42, 287–320. [Google Scholar] [CrossRef]
  41. Benbouzid, D.; Busa-Fekete, R.; Casagrande, N.; Collin, F.D.; Kégl, B. MultiBoost: A multi-purpose boosting package. J. Mach. Learn. Res. 2012, 13, 549–553. [Google Scholar]
  42. Debes, C.; Merentitis, A.; Heremans, R.; Hahn, J.; Frangiadakis, N.; van Kasteren, T.; Liao, W.; Bellens, R.; Pižurica, A.; Gautama, S.; et al. Hyperspectral and LiDAR data fusion: Outcome of the 2013 GRSS data fusion contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2405–2418. [Google Scholar] [CrossRef]
Figure 1. Pavia University data set: (a) color composite of the scene; (b) training set; (c) test set.
Figure 1. Pavia University data set: (a) color composite of the scene; (b) training set; (c) test set.
Jimaging 06 00114 g001
Figure 2. GRSS-DFC2013 data set: (a) color composite of the scene; (b) training set; (c) test set.
Figure 2. GRSS-DFC2013 data set: (a) color composite of the scene; (b) training set; (c) test set.
Jimaging 06 00114 g002
Figure 3. Examples of opening by reconstruction (OBR) (row 1), opening by partial reconstruction (OBPR) (row 2), multiresolution segmentation (MRS) objects (row 3) and object-guided opening by reconstruction (OOBR) (row 4) with different parameter sets from the second principal component of the Pavia University data.
Figure 3. Examples of opening by reconstruction (OBR) (row 1), opening by partial reconstruction (OBPR) (row 2), multiresolution segmentation (MRS) objects (row 3) and object-guided opening by reconstruction (OOBR) (row 4) with different parameter sets from the second principal component of the Pavia University data.
Jimaging 06 00114 g003
Figure 4. Mean overall accuracy (OA) values versus different sizes of structural elements (SEs) and segmentation scale sets in morphological profiles (MPs), morphological profiles with partial reconstruction (MPPR), extended object-guided morphological profiles (EOMPs) and EOMPs with mean values (EOMPsM) on raw spectral (a,b,e,f) and principal component analysis (PCA)-transformed (c,d,g,h) features from Reflective Optics System Imaging Spectrometer (ROSIS) university (row 1) and GRSS-DFC2013 (row 2) data sets using an support vector machine (SVM).
Figure 4. Mean overall accuracy (OA) values versus different sizes of structural elements (SEs) and segmentation scale sets in morphological profiles (MPs), morphological profiles with partial reconstruction (MPPR), extended object-guided morphological profiles (EOMPs) and EOMPs with mean values (EOMPsM) on raw spectral (a,b,e,f) and principal component analysis (PCA)-transformed (c,d,g,h) features from Reflective Optics System Imaging Spectrometer (ROSIS) university (row 1) and GRSS-DFC2013 (row 2) data sets using an support vector machine (SVM).
Jimaging 06 00114 g004
Figure 5. OA values versus the number of trees/iterations in bagging (Bag(C4.5), Bag(ERDT)), AdaBoost (AB(C4.5), AB(ERDT)), MultiBoost (MB(C4.5), MB(ERDT)), RaF(C4.5) and ExtraTrees using MPs (a,e), MPPR (b,f), object-guided morphological profiles (OMPs) (c,g) and OMPsM (d,h) from PCA-transformed features of Pavia University (ad) and GRSS-DFC2013 (eh) data sets.
Figure 5. OA values versus the number of trees/iterations in bagging (Bag(C4.5), Bag(ERDT)), AdaBoost (AB(C4.5), AB(ERDT)), MultiBoost (MB(C4.5), MB(ERDT)), RaF(C4.5) and ExtraTrees using MPs (a,e), MPPR (b,f), object-guided morphological profiles (OMPs) (c,g) and OMPsM (d,h) from PCA-transformed features of Pavia University (ad) and GRSS-DFC2013 (eh) data sets.
Jimaging 06 00114 g005
Figure 6. Classification maps with OA values (aj) corresponding to Table 1 and ground truth map, (k) for comparison with Pavia University test data.
Figure 6. Classification maps with OA values (aj) corresponding to Table 1 and ground truth map, (k) for comparison with Pavia University test data.
Jimaging 06 00114 g006
Figure 7. Classification maps with OA values (aj) corresponding to Table 2 and ground truth map, (k) for comparison with GRSS-DFC2013 test data.
Figure 7. Classification maps with OA values (aj) corresponding to Table 2 and ground truth map, (k) for comparison with GRSS-DFC2013 test data.
Jimaging 06 00114 g007
Table 1. The best OA and kappa statistics for considered classifiers using MPs, MPPR, OMPs and OMPsM features from PCA-transformed features of Pavia University data (F1: MPs, F2: MPPR; F3: OMPs; F4: OMPsM).
Table 1. The best OA and kappa statistics for considered classifiers using MPs, MPPR, OMPs and OMPsM features from PCA-transformed features of Pavia University data (F1: MPs, F2: MPPR; F3: OMPs; F4: OMPsM).
Ensemble MethodClassifierRaw:+F1+F2+F3+F4PC10:+F1+F2+F3+F4
NoneC4.565.76,0.5877.42,0.7182.60,0.7783.51,0.8091.89,0.8973.17,0.6678.45,0.7380.34,0.7492.05,0.9095.84,0.95
ERDT64.66,0.5779.61,0.7481.72,0.7683.62,0.7991.74,0.8962.42,0.5471.83,0.6573.39,0.6690.15,0.8795.06,0.94
SVM80.49,0.7687.31,0.8487.24,0.8396.26,0.9592.24,0.9081.51,0.7784.61,0.8083.79,0.7991.49,0.8991.13,0.89
BaggingC4.572.18,0.6679.90,0.7486.74,0.8295.94,0.9591.15,0.8876.36,0.7079.90,0.7486.74,0.8295.26,0.9497.75,0.97
ERDT71.42,0.6584.63,0.8088.74,0.8587.76,0.8495.01,0.9378.56,0.7380.97,0.7682.42,0.7795.95,0.9598.43,0.98
AdaBoostC4.574.18,0.6892.44,0.9088.94,0.8693.80,0.9291.41,0.8976.99,0.7192.21,0.9087.93,0.8494.48,0.9398.66,0.98
ERDT73.67,0.6888.84,0.8489.35,0.8691.15,0.8897.09,0.9677.40,0.7288.42,0.8586.28,0.8297.05,0.9698.29,0.98
MultiBoostC4.573.91,0.6892.66,0.9088.80,0.8594.89,0.9390.52,0.8777.49,0.7292.01,0.8988.18,0.8494.52,0.9398.75,0.98
ERDT72.79,0.6688.15,0.8489.97,0.8790.56,0.8794.22,0.9277.63,0.7283.96,0.7984.92,0.8096.71,0.9698.51,0.98
Random ForestC4.571.08,0.6485.65,0.8187.88,0.8490.93,0.8892.94,0.9176.31,0.7082.83,0.7885.06,0.8094.29,0.9297.66,0.97
ExtraTreesERDT72.70,0.6686.58,0.8284.09,0.7988.54,0.8594.96,0.9376.09,0.7085.23,0.8184.60,0.8095.24,0.9497.80,0.96
Table 2. The best OA and kappa statistics for considered classifiers using MPs, MPPR, OMPs and OMPsM features from PCA-transformed features of GRSS-DFC2013 data (F1: MPs, F2: MPPR; F3: OMPs; F4: OMPsM).
Table 2. The best OA and kappa statistics for considered classifiers using MPs, MPPR, OMPs and OMPsM features from PCA-transformed features of GRSS-DFC2013 data (F1: MPs, F2: MPPR; F3: OMPs; F4: OMPsM).
Ensemble MethodClassifierRaw:+F1+F2+F3+F4PC10:+F1+F2+F3+F4
NoneC4.580.07,0.7888.93,0.8888.48,0.8789.76,0.8991.43,0.9183.51,0.8286.54,0.8583.97,0.8388.64,0.8891.59,0.91
ERDT78.74,0.7788.35,0.8785.27,0.8489.07,0.8892.39,0.9281.40,0.8086.98,0.8683.90,0.8288.16,0.8789.41,0.88
SVM90.19,0.8994.19,0.9490.42,0.9093.13,0.9396.46,0.9689.72,0.8993.61,0.9391.42,0.9193.51,0.9395.18,0.95
BaggingC4.584.90,0.8490.73,0.9079.44,0.7890.87,0.9094.06,0.9486.59,0.8590.26,0.8985.98,0.8594.18,0.9496.30,0.96
ERDT86.24,0.8593.50,0.9392.06,0.9192.59,0.9294.71,0.9489.78,0.8990.90,0.9091.75,0.9193.91,0.9396.13,0.96
AdaBoostC4.585.94,0.8591.44,0.9189.78,0.8992.75,0.9294.04,0.9388.36,0.8791.72,0.9191.54,0.9191.91,0.9194.61,0.94
ERDT86.50,0.8594.29,0.9492.64,0.9292.91,0.9294.80,0.9489.41,0.8890.81,0.9092.03,0.9194.42,0.9496.59,0.96
MultiBoostC4.586.16,0.8591.54,0.9189.73,0.8992.65,0.9294.34,0.9388.45,0.8791.98,0.9191.27,0.9093.77,0.9396.06,0.96
ERDT86.58,0.8593.68,0.9391.78,0.9193.05,0.9294.86,0.9490.23,0.8991.87,0.9191.99,0.9193.98,0.9395.66,0.95
Random ForestC4.585.25,0.8493.61,0.9391.30,0.9192.93,0.9294.65,0.9488.95,0.8889.86,0.8991.63,0.9192.15,0.9195.76,0.95
ExtraTreesERDT86.91,0.8693.20,0.9392.36,0.9292.68,0.9295.06,0.9589.28,0.8890.19,0.8991.70,0.9193.13,0.9395.26,0.95
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Samat, A.; Li, E.; Liu, S.; Miao, Z.; Wang, W. Ensemble of ERDTs for Spectral–Spatial Classification of Hyperspectral Images Using MRS Object-Guided Morphological Profiles. J. Imaging 2020, 6, 114. https://doi.org/10.3390/jimaging6110114

AMA Style

Samat A, Li E, Liu S, Miao Z, Wang W. Ensemble of ERDTs for Spectral–Spatial Classification of Hyperspectral Images Using MRS Object-Guided Morphological Profiles. Journal of Imaging. 2020; 6(11):114. https://doi.org/10.3390/jimaging6110114

Chicago/Turabian Style

Samat, Alim, Erzhu Li, Sicong Liu, Zelang Miao, and Wei Wang. 2020. "Ensemble of ERDTs for Spectral–Spatial Classification of Hyperspectral Images Using MRS Object-Guided Morphological Profiles" Journal of Imaging 6, no. 11: 114. https://doi.org/10.3390/jimaging6110114

APA Style

Samat, A., Li, E., Liu, S., Miao, Z., & Wang, W. (2020). Ensemble of ERDTs for Spectral–Spatial Classification of Hyperspectral Images Using MRS Object-Guided Morphological Profiles. Journal of Imaging, 6(11), 114. https://doi.org/10.3390/jimaging6110114

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop