Next Article in Journal
Acoustic Emission and K-S Metric Entropy as Methods for Determining Mechanical Properties of Composite Materials
Previous Article in Journal
An Ultra-Short Baseline Underwater Positioning System with Kalman Filtering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Suppression Based Matched Filter for Hyperspertral Imagery Target Detection

College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(1), 144; https://doi.org/10.3390/s21010144
Submission received: 16 November 2020 / Revised: 16 December 2020 / Accepted: 22 December 2020 / Published: 28 December 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Target detection in hyperspectral imagery (HSI) aims at extracting target components of interest from hundreds of narrow contiguous spectral bands, where the prior target information plays a vital role. However, the limitation of the previous methods is that only single-layer detection is carried out, which is not sufficient to discriminate the target parts from complex background spectra accurately. In this paper, we introduce a hierarchical structure to the traditional algorithm matched filter (MF). Because of the advantages of MF in target separation performance, that is, the background components are suppressed while preserving the targets, the detection result of MF is used to further suppress the background components in a cyclic iterative manner. In each iteration, the average output of the previous iteration is used as a suppression criterion to distinguish these pixels judged as backgrounds in the current iteration. To better stand out the target spectra from the background clutter, HSI spectral input and the given target spectrum are whitened and then used to construct the MF in the current iteration. Finally, we provide the corresponding proofs for the convergence of the output and suppression criterion. Experimental results on three classical hyperspectral datasets confirm that the proposed method performs better than some traditional and recently proposed methods.

1. Introduction

Hyperspectral imagery (HSI) is composed of a three-dimensional data structure with spatial-spatial-spectral components, which is viewed as an image cube [1,2]. Typically, the sensor acquires the pixels in the scene in the form of spectral vectors. Hundreds of adjacent spectral bands constitute elements of the vectors, which can be used to distinguish different materials on the basis of their different electromagnetic reflection energy [3,4]. As a result, HSI target detection has huge application prospects, which span military [5,6], agriculture [7,8], prospecting [9], and disaster monitoring fields [3,10,11].
In the past few decades, researchers and scholars have proposed several hyperspectral target detection algorithms. Spectral angle mapper (SAM) [12] is generally considered as the simplest one since it does not require any assumption on data distribution. The constrained energy minimization (CEM) [13,14] minimizes the output energy by constructing a finite impulse response (FIR) filter. Based on CEM, a robust detector uses an inequality constraint on CEM (CEM-IC) to guarantee that the outputs of target spectra are larger than one [15]. Some traditional target detection algorithms are usually based on the statistical information of the target and background spectra [16], such as spectral matched filter (SMF) [3,17], adaptive matched filter (AMF) [18], and adaptive coherence/cosine estimator (ACE) [19,20]. These methods assume that the Gaussian distribution exists on the target and background spectra and adopt the generalized likelihood ratio test (GLRT) [16] to distinguish the target and background pixels. Additionally, the subspace model is applied to describe the difference between the target and background spectra. In this model, the information of the target and background spectra is contained in the target and background subspaces, respectively. By constructing the subspace model, a number of classical target detection algorithms have been developed, such as orthogonal subspace projection (OSP) [21,22], matched subspace detector (MSD) [23], and adaptive subspace detector (ASD) [19,24]. The above subspace-based target detectors extend the dimension of target signatures for overcoming the spectral variability problems [3].
As machine learning methods are introduced to HSI target detection, the classical subspace-based target detectors are also in further development. Instead of exploiting first- and second-order statistics [25], the generalization of kernel-based machine learning extends these HSI target detectors to their corresponding nonlinear (kernel) versions, such as kernel MSD [26], kernel ASD and kernel OSP detectors [4,27]. In addition, sparsity-based target detectors depend on the conception that each pixel in HSI can be represented by as few atoms as possible in a given spectral dictionary. According to the sparse representation of known target signatures, several methods develop into significant target detection strategies such as sparsity-based target detector (STD) [28], simultaneous joint sparsity-based detector [29], a sparse representation-based binary hypothesis (SRBBH) model [30], etc. [31,32,33].
The prior target spectral signature is one key point in hyperspectral target detection. However, the spectral variability is generated in the signal transmission path, which greatly affects the quality of the detection results, including uncompensated errors in the sensor, uncompensated atmospheric and environmental effects [34], etc. In [35], the authors propose an automatic robust iteratively reweighted method to ease the spectral variation. An optimal spectrum is obtained by iterating the target spectrum. In [36], the hyperspectral target detector, constructed by an effective spectral feature combined with the spectral reflectance and spectral derivative, is able to tolerate the spectral variations of the same material.
Recently, detection methods based on the hierarchical structure have been proposed. Rather than directly improving the quality of the target spectrum, their work concentrates on the background suppression through one hierarchical structure in the detection task. The hierarchical CEM (hCEM) [37] detector proposed by Zou et al., consists of different layers of classical CEM detectors. In hCEM, the output of each layer’s spectra is transformed by a nonlinear suppression and then used to update the HSI input for the next iteration. Furthermore, motivated by the spirit of the hierarchical structure, Hao et al., propose an angle distance-based hierarchical background separation method (ADHBS) [38] to enlarge the angle distance between the target and background pixels. Using the ADHBS method, the background spectral vectors gradually move to the direction perpendicular to the target spectral vector.
Since the traditional matched filter (MF) detector only performs single-layer detection on HSI input, it is evident that the background suppression is not sufficient, so the target and the background cannot be completely separated. In this paper, we impose the hierarchical background suppression on the MF detector (HSMF), designed to filter the HSI input several times to strengthen background suppression. In our method, the output of each layer imparts an important effect on the HSI input of the next layer. Since the traditional MF detector has the characteristic of suppressing the background and retaining the target, which is directly reflected in the output of each pixel, we choose the mean value of the output for each layer as a suppression criterion. Compared with hCEM and ADHBS, which also adopt a hierarchical structure, HSMF only needs fewer iterations to complete the output convergence based on this suppression structure. In addition, for hCEM and ADHBS, the whole HSI input will be suppressed in the next iteration, where the background convergence speed is faster than the target pixel, so that the targets can be separated from the background. Our method only suppresses the background, but the target pixels remain unchanged, and the target data are completely preserved in each iteration. The contributions of our work are summarized as follows:
  • Compared with the traditional MF detector, since the hierarchical structure greatly enhances the background suppression, the detection performance of HSMF is significantly improved.
  • The suppression criterion we adopted quickly completes the convergence of the proposed algorithm, so it only takes a few iterations to output the result. In addition, we provide theoretical proofs for the convergences of the suppression criterion and output energy.
  • Comparison experiments on three real hyperspectral images show that the proposed method is superior to the recently proposed algorithms and some classical detectors.
The organization of the rest of this paper is as follows. In Section 2, we briefly introduce the MF detectors and applications of hierarchical structure. In Section 3, our proposed method is explained in detail, and some theoretical analyses are given. Extensive experiments are presented in Section 4. Some discussions are demonstrated in Section 5. Finally, we draw conclusions in Section 6.

2. Related Work

2.1. Brief Introduction to Matched Filter Detectors

A spectral observation sample consisting of B spectral bands can be represented in a vector form as x = [ x 1 , x 2 , , x B ] T . Typically, the input spectral data are mostly only composed of the background clutter. However, when the spectral signature of targets mixes background clutter noise, the mixed input is equivalent to a linear additive model [3]. The model for MF is given by the following two hypotheses:
H 0 : x = n ,   target   absent H 1 : x = a s + n ,   target   present
where s = [ s 1 , s 2 , , s B ] T is the spectral signature of the desired target with B spectra bands, a is the target abundance measure, and n is the background clutter noise.
Generally, the MF detector follows a specific assumption: the background clutter covariance matrix and the spectral signature of targets are known in the form of a linear additive model. According to the GLRT [16], the MF output for a test spectral input vector x is expressed by
y MF ( x ) = s T C ^ 1 x s T C ^ 1 s H 0 H 1 η MF
where C ^ and s represent the estimated covariance matrix and the prior target spectrum, respectively. If the output y MF is larger than the given threshold η MF , the input sample is viewed as a target; otherwise, it is considered as a background pixel.
Instead of using global background clutter statistics, adaptive MF (AMF) [18] designs a dual window construction consisting of two regions. The inner window is framed based on the size of the target, and the outer window consists of the local background clutter, which makes the AMF detector more adaptive to the local background clutter statistics.
Based on the AMF detector, imposing the regularization term constructs a more stable detector [39]. The output function of the regularized AMF is expressed by
y = s T ( C ^ + β I ) 1 X s T ( C ^ + β I ) 1 s
where β is the regularization coefficient, and I is the identity matrix. In the regularized AMF, a quadratic penalization term is beneficial to keep the covariance matrix non-singular and make the regularization coefficient tend to zero.

2.2. Applications of Hierarchical Structure in Hyperspectral Target Detection

Limited by the single-layer structure, the traditional target detectors cannot completely separate the target and background pixels. Hence, some hyperspectral target detection methods make use of a hierarchical structure on the basis of traditional algorithms. Both hCEM and ADHBS introduce a hierarchical structure to the traditional algorithms to improve their background suppression ability.
In hCEM [37], the authors apply one hierarchical structure to the traditional CEM detector. After each iteration, the magnitude of the background spectra is reduced, while the target spectra are preserved. The hCEM output vector of the kth layer is formulated as:
y k = d T R k 1 d T R k 1 d X k
where X k represents the spectra matrix in the kth layer, d is the prior target spectrum and R k is the correlation matrix of the kth layer. Based on the output of the kth layer, each spectral vector of the next layer is updated, expressed as follows:
x n k + 1 = q ( y n k ) x n k
where q ( y n k ) is a nonlinear suppression function. The HSI spectral matrix of the next layer, which is applied to reconstruct a new CEM detector, will be updated continuously with the output of the current layer until the background pixels are entirely suppressed.
In ADHBS [38], the angular distance between the tested spectral vector and the known target spectrum, represented by θ, is used as the metric to separate the target and background pixels. The tested spectral vector gradually moves to the direction perpendicular to the target spectrum according to the metric θ in the spectral space. The angle distance in the kth layer is calculated by:
θ n k = arccos { | ( x n k ) T d k | x n k 2 d k 2 }
where | | means to take absolute value to ensure that the output angle is between 0 degrees and 90 degrees, and 2 is to calculate the Euclidean norm of the vector.
Each spectral vector of the ( k + 1 ) th layer is adjusted by the following function:
x n k + 1 = ( 1 α n k ) x n k + α n k d
where d is an orthogonal vector of d and α n k determines the adjustment degree of the tested spectrum on the ( k + 1 ) th layer. The value of α is obtained by one power control function expressed as follows:
α n k = [ θ n k / 90 ] p
where p controls the change rate of α n k which further controls the rate of the tested spectra moving to d . The change varies considerably from pixel to pixel, which is the key to distinguish target pixels and background pixels. The size of θ corresponding to background pixels is larger than that of the target pixels, so the movement rates of background spectra are faster than those of target spectra, which eventually makes the target pixels separated from the background clutter.

3. Proposed Method

3.1. Algorithm Flow Description

Inspired by related work on the hierarchical structure, we realize that the spectra transformation is beneficial to the separation of target and background, which results in better detection results. For MF, the information about the relationship between the target and the background is reflected in the output at a rough degree. The output value of the target pixels is greater than that of background pixels. In this paper, we utilize the output value of the MF detector to suppress the background spectra through a hierarchical structure in order to obtain one more accurate detection result. Each layer has one MF detector and the corresponding input and output, and the output of the current layer is used to adjust the input of the next layer. At the beginning of the next layer, the HSI input spectrum is multiplied by a constraint when one pixel is judged as a background pixel based on the output in the previous layer. In the whole iterative process, the output of each layer is the concrete embodiment of the convergence trend. In a word, the gradual convergence of backgrounds in the output eventually strips off the background clutter and retains the target pixels.
We determine the whole algorithm flow shown in Figure 1, and the proposed method is summarized as the following steps:
  • Step 1: Obtain the HSI spectral matrix and the target spectrum vector, then whiten both the HSI input and the prior target spectrum.
  • Step 2: Construct the MF detector for the current layer.
  • Step 3: Obtain the output and calculate the mean value of the output.
  • Step 4: Determine whether to continue the iteration. If the termination condition is met, go to step 5; otherwise, update the spectral input matrix of the next layer, and return to step 1.
  • Step 5: Output the final detection result.

3.2. Hierarchical Background Suppression on the MF Detector (HSMF)

Because the detector of each layer is developed from the traditional MF, the statistical models of the target pixels and the background clutter are similarly assumed to obey Gaussian distribution. In the proposed algorithm, we consider an observation spectral vector x ( B × 1 ) under this assumption with the following distribution:
x N ( a s , σ 2   C ^ )
where   C ^ = ( 1 / N ) X X T is the estimate covariance matrix of the background clutter, s is a ( B × 1 ) mean vector, and a and σ 2 are both nonnegative numbers which determine the mean value and the value of the covariance matrix, respectively.
Generally, the distribution of raw HSI data is severely discretized. There is considerable overlap between the distribution of the target and background in the original spectral observation space. By adopting the data whitening process, the data tend to a circular distribution instead of an elliptic distribution. The diagram of the whitening transformation is shown in Figure 2. As a result, the aliasing area between the target and background is reduced, and at the same time, it enlarges the spectral difference between the target and background pixels. When constructing the MF detector of different layers, not only does the HSI spectral matrix need whitening, but the vector d ( B × 1 ) containing the prior information is also whitened to be consistent with the input data. According to the known distribution characteristics, consider the whitening transformation:
M =   C ^ 1 / 2 X ,   s =   C ^ 1 / 2 d
where the new spectral matrix and the new target spectrum vector are denoted by capital letters M and s, respectively. In the different layers of the iteration, the HSI spectral matrix suppresses its background components on the basis of the previous layer so that a better MF detector can be constructed to obtain more accurate detection results than the upper layer.
After the whitening process, the target pixels display the obvious separability to the background clutter. Then, according to the structure of the traditional MF detector, M and s in the whitened space are further centralized. That is, s subtracts a B×1 vector consisting of the average value of all pixels in each band, and M subtracts a B×N matrix obtained by copying the B×1 column vector.
Next, the preprocessed data are further used to construct the MF detector in different layers. The lowercase letter k indicates the number of iterations (layers). For the kth layer, the output of the HSMF is given by:
y k = s k T C ^ k 1 s k T C ^ k 1 s k M k
where s k and C ^ k define the filter coefficients with constraint (the new target spectrum vector) and the estimated covariance matrix of the background clutter for the kth layer, respectively.
Then, the next layer’s spectral input vector x n k + 1 gets a corresponding update. By multiplying a constraint, the equation is formulated as follows:
x n k + 1 = λ ( y n k ) x n k
where λ ( y n k ) is an energy constraint function, and the constraint function is imposed on the raw input X, rather than the whitened M. In other words, the energy constraint function outputs a score to adjust the property of each pixel. When the score meets the suppression criterion, the tested pixel will be judged as a background pixel that needs to be suppressed in the next iteration.
With the update of the input, the MF detector of the next layer will be reconstructed. Herein, the energy constraint function is represented as
λ ( y n k ) = { 1 , y n k 1 N n = 1 N y n k β , y n k < 1 N n = 1 N y n k
where β is a parameter to constrain the backgrounds. β generally uses a positive number that is approximately zero. Such a choice is to make sure that the reconstructed estimate covariance matrix C ^ is not singular. On the other hand, the purpose of the constraint is to decrease the influence of background pixels on constructing the MF detector in the next layer, and the output of the upper layer determines these background pixels. Therefore, weakening the influence of background pixels on the new MF detector by multiplying the constraint value β helps to improve the target detection performance of the next layer.
Typically, the output values of the MF detector objectively reflect the similarity between each HSI input spectrum and the prior target spectrum. Hence, y n k corresponding to the target pixels is larger than y n k corresponding to the background pixels. As the iteration continues, the number of background pixels detected increases. Therefore, with the support of the energy constraint function in different layers, we choose the mean of the score for each layer to determine whether the iteration needs to continue, as follows:
η k = 1 N n = 1 N λ ( y n k ) b r e a k c o n t i n u e ε
where ε is a threshold value. The detailed procedures of the proposed method are exhibited in Algorithm 1.

3.3. Theoretical Proof

Here, we prove that the mean value of a single layer output is suitable as a suppression criterion. We denote the MF detector of the kth layer as a vector k = ( l 1 k , l 2 k , , l B k ) T . The mean value of the kth layer y ¯ k can be divided into targets and backgrounds:
y ¯ k = 1 N n = 1 N y n k = 1 N n = 1 N k T M k = 1 N ( n = 1 N ˜ k T M ˜ k + n = 1 N k T M k ) = 1 N n = 1 N ˜ k T M ˜ k + θ
where N ˜ and N are the number of the background and target pixels, respectively, M ˜ and Μ are the spectral input of the background and target pixels, respectively, and θ is a constant representing the mean value of the output after the background pixels are completely suppressed. From Equations (12) and (15), the known background pixels are suppressed according to the output of the previous layer in the next layer. Now, we try to obtain the minimum value of the output mean value:
lim k k 0 1 N n = 1 N ˜ k T Μ ˜ k = 0 lim k k 0 y ¯ k = θ
where k 0 represents a sufficiently large number of layers so that the background pixels are fully suppressed. The mean value of the output for each layer gradually converges to a constant θ. In addition, when the mean value of the output eventually converges to a constant, the proportion of the suppressed background pixels will increase, which causes the output of Equation (14) to become smaller. Thus, we set the threshold to determine whether the iteration is stopped, as shown in Equation (14).
The following proof analyzes the convergence of the output energy. Here, the MF detector as a vector is expressed in detail as follows:
k = C ^ k 1 ( d k u k ) ( d k u k ) T C ^ k 1 ( d k u k ) = C ^ k 1 s k s k T C ^ k 1 s k
where the vector u represents the average value of all pixel values in each band. Then, we calculate the energy of the kth layer and the ( k + 1 ) th layer as follows:
y k 2 2 = k T Μ k 2 2 = s k T C ^ k 1 Μ k s k T C ^ k 1 s k 2 2 = N s k T C ^ k 1 s k
y k + 1 2 2 = ( k + 1 ) T M k + 1 2 2 = s k + 1 T C ^ k + 1 1 Μ k + 1 s k + 1 T C ^ k + 1 1 s k + 1 2 2 = N s k + 1 T C ^ k + 1 1 s k + 1
Then, we need to compare the size of the two denominators [37]. Firstly, the difference between the covariance matrix of adjacent layers is calculated:
C ^ k + 1 C ^ k = 1 N 1 N ( ( x n k + 1 μ k + 1 ) ( x n k + 1 μ k + 1 ) T ( x n k μ x k ) ( x n k μ x k ) T ) = 1 N 1 N ( λ 2 m n k T m n k T m n k m n k T ) = W W T
where W = [ w 1 , w 2 , , w N ] B × N . According to the Sherman–Morrison formula [40], we suppose that C ^ k is an invertible square matrix and ( C ^ k + w n w n T ) is invertible if and only if 1 + w n T C ^ k 1 w n 0 . On this condition, we obtain
( C ^ k + 1 + w n w n T ) 1 = C ^ k + 1 1 C ^ k + 1 1 w n w n T C ^ k + 1 1 1 + w n T C ^ k + 1 1 w n
Then, we obtain the difference between the inverse covariance matrix of adjacent layers:
C ^ k + 1 1 C ^ k 1 = C ^ k + 1 1 ( C ^ k + 1 + W W T ) 1 = n = 1 N C ^ k + 1 1 w n w n T C ^ k + 1 1 1 + w n T C ^ k + 1 1 w n
Next, we obtain from Equations (20) and (22):
s k T ( C ^ k + 1 1 C ^ k 1 ) s k = n = 1 N s k T C ^ k + 1 1 w n w n T C ^ k + 1 1 s k 1 + w n T C ^ k + 1 1 w n = n = 1 N s k T C ^ k + 1 1 w n 2 2 1 + w n T C ^ k + 1 1 w n
where s k T C ^ k + 1 1 w n 2 2 0 and w n T C ^ k + 1 1 w n 0 . We have the following size relationship:
1 s k T C ^ k + 1 1 s k 1 s k T C ^ k 1 s k
where s k T = ( d u n k ) T . With the detected background pixels constrained, we have:
s k T 2 2 s k + 1 T 2 2
1 s k + 1 T C ^ k + 1 1 s k + 1 1 s k T C ^ k 1 s k
Finally, we compare the output energy of adjacent layers:
1 N y k + 1 2 2 1 N y k 2 2
The above proof suggests that the output energy still converges, although s k and C ^ k 1 keep changing as the HSI spectral matrix is updated. In fact, these backgrounds suppressed in the upper layer have less impact on constructing the next layer detector, and the new detector pays more attention to the unsuppressed background.
Algorithm 1 describes the detailed procedures of the proposed method.
Algorithm 1: HSMF Algorithm
Input:
1. spectral matrix X = [ x 1 , x 2 , , x N ] B × N , target spectrum d B × 1 , set threshold ε,
Initialization:
2.   k = 1 ,   X 1 = X ,   n = 1 , 2 , , N , i = 1 , 2 , , B ,
Hierarchical Suppression:
3.   C ^ k = ( 1 / N ) X k X k T ,
4.   M k = C ^ k 1 / 2 X k , d k = C ^ k 1 / 2 d k ,
5.   for   band   i : u i k = ( 1 / N ) n = 1 N x n k ,
6.   u k B × 1 copy   column   vector U k B × N ,
7.   M k = X k U k , s k = d k u k ,
8.   C ^ k = ( 1 / N ) Μ k Μ k T ,
9.   y k = ( ( s k T C ^ k 1 ) / s k T C ^ k 1 s k ) M k ,
10.   x n k + 1 = λ ( y n k ) × x n k ,
11.   k = k + 1 ,
Stop Iteration:
12. η k = 1 N n = 1 N λ ( y n k ) , if η k > ε , come back to step 3; else, go to step 13,
Output:
13. final outputs: y k = [ y 1 k , y 2 k , y N k ] 1 × N .

4. Experiment

In this section, three different experiments on classical HSI datasets are implemented to verify the effectiveness of HSMF. Here, we compare the proposed algorithms with several methods: SAM, CEM, CEM-IC, SMF, ACE, STD, MSD, SRBBH, hCEM, and ADHBS. Among them, SAM, CEM, SMF, and ACE are classical single-layer algorithms. In the following experiments, we set the spectral signature of the target based on the mean of all target samples of certain material in HSI, which is also adopted in [37,38]. The prior target spectral information of most detection algorithms is consistent so as to ensure fairness of the comparative experiments, except for MSD, STD, and SRBBH. For STD and SRBBH, instead of a feature vector, their prior information is expressed in the form of a matrix by a linear combination of a few atoms from the spectral dictionary. For MSD, the change of the spectral feature vectors is limited in a subspace of the band space.
In the experiment, we set a series of reasonable parameters for algorithms. Single-layer detectors such as SAM, CEM, SMF, and ACE do not need to preset any parameters. For MSD, SRBBH, and STD, the dual window size depends on the size of the target. Generally speaking, the inner window is larger than the target, and the outer window is larger than the inner window. To ensure that three methods have satisfactory results in three datasets, we set the sizes of their inner and outer windows to 13 and 19, 23 and 31, 13 and 19, respectively. For CEM-IC and hCEM, the original codes provided by previous studies are employed with default parameters. For ADHBS, the power function parameter is set to 3, which greatly reduces the number of iterations, and at the same time, it also obtains excellent detection results. For HSMF, we set suppression coefficient β = 0.0001 and threshold value ε = 0.01. The setting of these two parameters is analyzed in Section 5.3 in detail.
The experimental results of different detection algorithms are compared against the groundtruth. To visually illustrate the performance difference between our method and other algorithms, we plot results using receiver operating characteristic (ROC) [41,42]. ROC curves describe the dynamic relationship between the probability of detection (PD) and the false alarm rate (FAR), which is controlled by different thresholds. These two expressions are obtained as follows:
P D = N c N t , F A R = N f N b
where N c , N t represent the number of correct target pixels detected and the total true target pixels in HSI, respectively, and N f , N b represent the number of false alarm pixels and total number of background pixels, respectively. In addition, area under curve (AUC), whose range is between 0.5 and 1, also acts as a performance measure. Moreover, we compare the normalized AUC at lower FAR to exhibit the performance difference between different algorithms. The normalized AUC at lower FAR is obtained by normalizing AUC where FAR ranges from 0 to 0.001. Moreover, we plot the separability map to visually describe the separation of the target and background pixels for different detectors.
Regarding the experimental part of output energy convergence and mean convergence, we discuss this in Section 5.1 of the paper. The simulation software in the experiments is MATLAB 2018b, and the equipment used in all aforementioned methods is on an AMD 2600 personal computer with 16 GB RAM.

4.1. Experiment on San Diego Airport Data

This classical dataset is collected by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) [37,38] from the airport area in San Diego, USA, where the AVIRIS is a hyperspectral instrument designed, built, and operated by National Aeronautics and Space Administration (NASA). The raw spectral data consist of 224 spectral reflectance bands, which span from 0.4 to 2.5 μm. There are 200 × 200 pixels in each band, and the band number of the real hyperspectral image amounts to 189, where the low signal-to-noise ratio (SNR) and water absorption bands have been removed. The groundtruth and the false color image are shown in Figure 3m,n, respectively. The detection results of different detectors are shown in Figure 3, where Figure 3(a1–a7) show the detection result of different layers of HSMF. The ROC curves and the separability map are exhibited in Figure 4. As listed in Table 1, we compare both AUC values and AUC values at lower FAR of all detectors to describe the hyperspectral target detection performance in detail.
The iteration of HSMF stops in the 7th layer, and the final detection results are obtained. Combining the ROC curves with the AUC values in Table 1, we notice that three algorithms using hierarchical structure have better output results than other classical detectors. Comparing the AUC values of hCEM, ADHBS, and HSMF, although the difference between the three detectors is small, we can still find that HSMF has certain advantages. The AUC value of HSMF is second only to that of ADHBS, and the AUC value at lower FAR of HSMF is the maximum compared with the other detectors. From Figure 4b, we can further evaluate the separability between the target and background pixels. For ADHBS, the performance of separation is slightly better than HSMF and hCEM. For HSMF, the backgrounds are fully compressed, and the backgrounds and the targets are sufficiently separated. For SAM, CEM, CEM-IC, SMF, and ACE, the gaps between the backgrounds and the targets are small. For STD, MSD, and SRBBH, the area of the target box is partially overlapped with the background, which illustrates poor separability.

4.2. Experiment on Airport-Beach-Urban Data

The Airport-Beach-Urban dataset is composed of multiple sets of sample images. The original images are collected by the AVIRIS and divided into size 100 × 100 manually. In addition, the spectral data retain 191 spectral reflectance bands after removing the noisy bands [43]. The groundtruth is labeled using the Environment for Visualizing Images (ENVI) software. The above-processed data are provided by [44]. In the experiment, we choose a sample image from the urban scenes in Gainesville, Florida, USA. The captured place of the image is Gainesville, and the spatial resolution is 3.5m. The spectral reflectance bands cover 0.4–2.5 μm. The groundtruth and the false color image are shown in Figure 5m,n, respectively. The detection results of different detectors are shown in Figure 5, where Figure 5(a1–a5) show the detection results of different layers of HSMF. The ROC curves and the separability map are exhibited in Figure 6. In Table 2, we record both AUC values and AUC values at lower FAR of all detectors to fully describe the performance of target detection.
For this dataset, the iteration in HSMF stops in the 5th layer. In Figure 6, the ROC curves of hCEM, ADHBS, and HSMF are at the top of the image, indicating a better detection result. Similarly, the AUC values and the AUC values at lower FAR show that HSMF is better than classical detectors. According to the separability map in Figure 6b, we compare the separation of the targets and the backgrounds in all algorithm results. For HSMF, the gap between the backgrounds and the targets is the largest. For hCEM, the separability is greatly close to HSMF. For CEM, CEM-IC, SMF, ACE, STD, and ADHBS, the targets and the backgrounds have no overlap, but the gaps are small. For SAM, the area of the target box is larger than the others, which demonstrates that the background is not sufficiently suppressed. For MSD, STD, and SRBBH, they have a bad separability because the area of the target boxes is partially or completely overlapped with that of the background boxes.

4.3. Experiment on Urban Data

As one of the most widely used hyperspectral datasets, the Urban dataset is collected by the Hyperspectral Digital Imagery Collection Experiment (HYDICE) sensor over California, USA, where the HYDICE sensor is manufactured by Hughes Danbury Optical Systems. The original image scene consists of 307 × 307 pixels. The raw spectral data contain 210 spectral reflectance bands in the wavelength ranging from 0.4 to 2.5 μm. After removing the water-absorbed and noisy bands, the number of bands is lessened to 162 [45]. The groundtruth of the Urban scene has several classes, mainly including asphalt, grass, tree, and roof. In the experiment, we only choose a subset of 80 × 120 pixels in the upper left corner of the large image, where the roof is chosen as the groundtruth. The groundtruth and the false color image are shown in Figure 7m,n, respectively. The detection results of different detectors are shown in Figure 7, where Figure 7(a1–a6) show the detection results of different layers of HSMF. The ROC curves and the separability map are exhibited in Figure 8. In Table 3, we record both AUC values and AUC values at lower FAR of all detectors to describe the target detection performance exhaustively.
In this experiment, HSMF obtains the detection output after six layers of hierarchical suppression. Figure 8a shows that the ROC curves of hCEM and HSMF are on the top of the other detectors. In other words, hCEM and HSMF outperform other methods. For ADHBS, the AUC value at lower FAR is worse than hCEM and HSMF. In Figure 8b, the separability of hCEM and HSMF still behaves well. For SAM and STD, their backgrounds are not effectively suppressed. For the others, the gaps between the target boxes and the background boxes are relatively small.

5. Discussion

5.1. Energy Output Convergence and Mean Convergence

In this section, we test and verify the convergence performance of HSMF for different layers. Meanwhile, the rationality of the suppression criterion, chosen in HSMF, is discussed. In fact, the mean value of the absolute output value also gradually converges to a constant after multiple iterations. The constant is an ideal value when the background pixels are completely suppressed, and only the target pixels remain.
The square of the Euclidean norm of each layer output, expressed as y k 2 2 , works as the output energy of the corresponding layer. Curves drawn by the output energy of different layers exhibit the energy convergence trend.
As shown in Figure 9, the output energy tends to converge after a few iterations on three hyperspectral images. The sharp drop in energy between the first and second layers is the most obvious. The convergence tends to be flat after the second layer. It is easy to find that the first two layers suppress most of the background pixels, and the remaining layers will suppress the background at a finer level to improve the detection accuracy.
In Figure 10, the blue curves denote the mean value of the absolute output value, and the pink curves represent the mean value of only target-remained output where the background output is considered to be zero. The two output means can be expressed, respectively, as:
y ¯ k = 1 N 1 N | y n k |
y ¯ k = 1 N 1 N | y ˜ n k |
where y ˜ n k represents the target pixels output, and N denotes the number of target pixels. The pink curves fluctuate in a very small range. This is because the target pixels are well preserved. The blue curves converge quickly, and the sharp drop between the first and second layers is the most obvious. In line with the expectation, the blue curves gradually converge to the pink curves. In other words, the mean value of the absolute output value eventually equals the mean value of the only target-remained output, which accords with Equation (16) and means that the target pixels are preserved while the background pixels are suppressed.
Since the process of the background suppression is irreversible, once a pixel is detected as a background pixel, its input in the next layer will be suppressed. Hence, detection performance is determined by the different suppression criteria. Here, we choose several suppression criteria for comparison. Besides mean value, we also choose one-eighth mean, quarter mean, one-half mean, mean, and three-quarter mean.
As exhibited in Figure 11a, different suppression criteria require different iteration layers, and the average energy of each layer is different from each other. When the suppression criterion is the mean value and three-quarter mean, the number of layers is seven. For other suppression criteria, eight iterations are required. According to Figure 11b, although the ROC curves of different suppression criteria remain different, the detection results are relatively close. Combining Figure 11a,b, when the mean value works as the suppression criterion, the detection only requires fewer iterations while the detection performance is also excellent.

5.2. Time Performance

In Table 4, we list the running time of different detection algorithms on the three datasets. Meanwhile, we count the number of iterations and calculate the time consumption of each layer. Some detectors have similar time performance, and the other detectors take a long time to get the output.
For SAM, it only needs to calculate the angular distance between each pixel and the prior spectral information. Thus, its running time is the shortest on the three datasets. For CEM and SMF, they just design a single layer FIR filter and MF, respectively, which are both outstanding in time performance. For STD and SRBBH, both are based on the sparse representation, which often costs considerable time for recovering the sparse vectors. At the same time, it is necessary to take time to construct a local background dictionary and a local target dictionary. As the dual windows move, a tremendous amount of computation needs to be required. MSD is the worst in terms of time performance because it uses singular value decomposition to construct the target and background subspaces. For CEM-IC, it also takes relatively little time to output the detection results. For ADHBS, the number of iterations is the highest on the last two datasets, and the running time is also more than hCEM and HSMF on the three datasets. For the hCEM, the total running time is less than ADHBS and HSMF on three datasets. For HSMF, it costs slightly more time than hCEM, but the number of iterations is the lowest. It means that HSMF has the best convergence performance. Because the MF coefficients with constraint and the estimated covariance matrix need to be recalculated at each layer in the new whitened space, it costs the most time in each layer.

5.3. The Effect of Parameter

In the proposed method, we need to set two parameters, namely suppression coefficient β and threshold value ε manually. These two parameters play an important role in detection performance. The size of β controls the degree of background suppression, and the size of ε controls the stop condition. This section discusses the influence of different parameter values on the experiment, mainly including the number of iterations and detection performance.
According to Figure 12b, the average output energy of different β values is very close in the same layer, which means that the different β values have little effect on the average output energy. In other words, under the suppression criterion in this paper, the output energy convergence is less affected by the parameter β. However, the suppression coefficient β has a relatively large impact on the number of iterations required for final convergence and the detection performance (i.e., AUC). Combining Table 5 and Figure 12a, as the value decreases the number of layers as are necessary also drops to a constant. Until the β value is not greater than 0.1, the number of iterations tends to be stable. In addition, once the β value is between 0.5 and 0.4, the AUC is relatively small. When the β value is smaller than 0.4, AUC tends to be stable, while AUC at lower FAR fluctuates in a small range. When the β value is 0.1 or 0.0001, AUC at lower FAR has better results. On the Airport-Beach-Urban data and Urban data, the parameter β also shows this trend. Herein, we only discuss the San Diego Airport data.
In the part of discussing the threshold value, we set β = 0.0001, which keeps up with the experiment in Section 4. In Figure 13a, the ordinate represents the mean value of all pixels’ suppression coefficient in different layers, as shown in Equation (14). For three different data, their convergence trends are the same, and they all tend to a minimal value in the end, which means that the selection of the threshold value must be greater than the convergence value. Otherwise, the threshold has no ability to constrain the process of the algorithm. In datasets Airport-Beach-Urban data and Urban, they only need six and seven iterations to converge to a constant. Hence, in order to better distinguish the impact of different thresholds on the detection performance, we use San Diego Airport data for comparison experiments.
Table 6 exhibits that as the threshold decreases, the number of iterations gradually increases until the threshold fails to terminate the iteration. With reference to the AUC value in Table 6 and the ROC curves in Figure 13b, the detection performance has a significant disadvantage when the threshold is set relatively large. From Table 6, we also find that the AUC at lower FAR has a trend where the value will increase first and then decrease. Considering the number of iterations and detection performance, we choose an appropriate threshold value, i.e., ε = 0.1, in the experiments.
In the parameter discussion, we conclude that the selection of parameters does not have an optimal value. It is crucial to choose reasonable parameters for different experiments. On the one hand, we must ensure the efficiency and accuracy of detection, and on the other hand, the parameters can be generalized to some extent.

6. Conclusions

In this paper, we are not limited to the single-layer detection of the MF detector but continuously construct a new MF detector for each iteration via the hierarchical structure to pursue better detection results. The mean output criterion adopted in the hierarchical background suppression also speeds up the output energy convergences, so the proposed method has fewer iterations compared with other hierarchical detectors. After the final iteration, the backgrounds are effectively suppressed while the targets are retained. Experimental results on three real hyperspectral images prove the outstanding detection performance of HSMF. From the ROC curves of the three datasets in the experimental part, it is straightforward to judge that HSMF has a better detection performance than classical algorithms. Compared with hCEM and ADHBS, HSMF maintains an excellent detection performance with the least number of iterations.

Author Contributions

Methodology, software, and writing—original draft preparation, C.G.; validation, formal analysis, and data curation, X.H.; writing—review and editing and supervision, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Nature Science Foundation of China under Grant 61573183 and the Open Project Program of the National Laboratory of Pattern Recognition (NLPR) under Grant 201900029.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The authors wish to thank the editor and reviewers for their suggestions and thank Fei Zhou and Kang Ni for their valuable comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.M.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  2. Manolakis, D.; Marden, D.; Shaw, G.A. Hyperspectral Image Processing for Automatic Target Detection Applications. Lincoln Lab. J. 2003, 14, 79–116. [Google Scholar]
  3. Manolakis, D.; Truslow, E.; Pieper, M.; Cooley, T.; Brueggeman, M. Detection algorithms in hyperspectral imaging systems: An overview of practical algorithms. IEEE Signal Process. Mag. 2014, 31, 24–33. [Google Scholar] [CrossRef]
  4. Nasrabadi, N.M. Hyperspectral target detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2014, 31, 34–44. [Google Scholar] [CrossRef]
  5. Eismann, M.T.; Stocker, A.D.; Nasrabadi, N.M. Automated hyperspectral cueing for civilian search and rescue. Proc. IEEE 2009, 97, 1031–1055. [Google Scholar] [CrossRef]
  6. Manolakis, D.G.; Shaw, G. Detection algorithms for hyperspectral imaging applications. IEEE Signal Process. Mag. 2002, 19, 29–43. [Google Scholar] [CrossRef]
  7. Patel, N.K.; Patnaik, C.; Dutta, S.; Shekh, A.M.; Dave, A.J. Study of crop growth parameters using Airborne Imaging Spectrometer data. Int. J. Remote Sens. 2001, 22, 2401–2411. [Google Scholar] [CrossRef]
  8. Datt, B.; McVicar, T.R.; Van Niel, T.G.; Jupp, D.L.B.; Pearlman, J.S. Preprocessing EO-1 Hyperion hyperspectral data to support the application of agricultural indexes. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1246–1259. [Google Scholar] [CrossRef] [Green Version]
  9. Hörig, B.; Kühn, F.; Oschütz, F.; Lehmann, F. HyMap hyperspectral remote sensing to detect hydrocarbons. Int. J. Remote Sens. 2001, 22, 1413–1422. [Google Scholar] [CrossRef]
  10. Camps-valls, G.; Tuia, D.; Bruzzone, L. Advances in hyperspectral image classification. IEEE Signal Process. Mag. 2014, 31, 45–54. [Google Scholar] [CrossRef] [Green Version]
  11. Martín, G.; Plaza, A. Spatial-spectral preprocessing prior to endmember identification and unmixing of remotely sensed hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 380–395. [Google Scholar] [CrossRef]
  12. Kruse, F.A.; Lefkoff, A.B.; Boardman, J.W.; Heidebrecht, K.B.; Shapiro, A.T.; Barloon, P.J.; Goetz, A.F.H. The spectral image processing system (SIPS)-interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
  13. Farrand, W.H.; Harsanyi, J.C. Mapping the distribution of mine tailings in the Coeur d’Alene River Valley, Idaho, through the use of a constrained energy minimization technique. Remote Sens. Environ. 1997, 59, 64–76. [Google Scholar] [CrossRef]
  14. Niu, Y.; Wang, B. Extracting Target Spectrum for Hyperspectral Target Detection: An Adaptive Weighted Learning Method Using a Self-Completed Background Dictionary. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1604–1617. [Google Scholar] [CrossRef]
  15. Yang, S.; Shi, Z.; Tang, W. Robust hyperspectral image target detection using an inequality constraint. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3389–3404. [Google Scholar] [CrossRef]
  16. Reed, I.S.; Yu, X. Adaptive Multiple-Band CFAR Detection of an Optical Pattern with Unknown Spectral Distribution. IEEE Trans. Acoust. 1990, 38, 1760–1770. [Google Scholar] [CrossRef]
  17. Manolakis, D.; Lockwood, R.; Cooley, T.; Jacobson, J. Is there a best hyperspectral detection algorithm? In Proceedings of the Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XV, Orlando, FL, USA, 13–17 April 2009; Volume 7334. [Google Scholar] [CrossRef]
  18. Robey, F.C.; Fuhrmann, D.R.; Kelly, E.J.; Nitzberg, R. A CFAR adaptive matched filter detector. IEEE Trans. Aerosp. Electron. Syst. 1992, 28, 208–216. [Google Scholar] [CrossRef] [Green Version]
  19. Kraut, S.; Scharf, L.L. Adaptive subspace detectors. IEEE Trans. Signal Process. 2001, 49, 1–16. [Google Scholar] [CrossRef] [Green Version]
  20. Kraut, S.; Scharf, L.L.; Butler, R.W. The adaptive coherence estimator: A uniformly most-powerful-invariant adaptive detection statistic. IEEE Trans. Signal Process. 2005, 53, 427–438. [Google Scholar] [CrossRef]
  21. Chang, C.I. Orthogonal Subspace Projection (OSP) revisited: A comprehensive study and analysis. IEEE Trans. Geosci. Remote Sens. 2005, 43, 502–518. [Google Scholar] [CrossRef]
  22. Harsanyi, J.C.; Chang, C.I. Hyperspectral Image Classification and Dimensionality Reduction: An Orthogonal Subspace Projection Approach. IEEE Trans. Geosci. Remote Sens. 1994, 32, 779–785. [Google Scholar] [CrossRef] [Green Version]
  23. Wang, Z.; Zhu, R.; Fukui, K.; Xue, J.H. Matched Shrunken Cone Detector (MSCD): Bayesian Derivations and Case Studies for Hyperspectral Target Detection. IEEE Trans. Image Process. 2017, 26, 5447–5461. [Google Scholar] [CrossRef] [PubMed]
  24. Kraut, S.; Scharf, L.L. The CFAR adaptive subspace detector is a scale-invariant GLRT. IEEE Trans. Signal Process. 1999, 47, 2538–2541. [Google Scholar] [CrossRef] [Green Version]
  25. Kwon, H.; Nasrabadi, N.M. A comparative analysis of kernel subspace target detectors for hyperspectral imagery. EURASIP J. Adv. Signal Process. 2007, 2007. [Google Scholar] [CrossRef] [Green Version]
  26. Kwon, H.; Member, S.; Nasrabadi, N.M. Kernel Matched Subspace Detectors for Hyperspectral Target Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 178–194. [Google Scholar] [CrossRef]
  27. Capobianco, L.; Garzelli, A.; Camps-Valls, G. Target detection with semisupervised kernel orthogonal subspace projection. IEEE Trans. Geosci. Remote Sens. 2009, 47, 3822–3833. [Google Scholar] [CrossRef]
  28. Chen, Y.; Nasser, M. Nasrabadi.; Trac, D. Tran. Sparse Representation for Target Detection in Hyperspectral Imagery. Optik (Stuttg.) 2011, 126, 5633–5640. [Google Scholar] [CrossRef]
  29. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Simultaneous joint sparsity model for target detection in hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2011, 8, 676–680. [Google Scholar] [CrossRef]
  30. Bitar, A.W.; Cheong, L.F.; Ovarlez, J.P. Target and Background Separation in Hyperspectral Imagery for Automatic Target Detection. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1598–1602. [Google Scholar] [CrossRef] [Green Version]
  31. Du, B.; Zhang, Y.; Zhang, L.; Tao, D. Beyond the Sparsity-Based Target Detector: A Hybrid Sparsity and Statistics-Based Detector for Hyperspectral Images. IEEE Trans. Image Process. 2016, 25, 5345–5357. [Google Scholar] [CrossRef]
  32. Li, W.; Du, Q.; Zhang, B. Combined sparse and collaborative representation for hyperspectral target detection. Pattern Recognit. 2015, 48, 3904–3916. [Google Scholar] [CrossRef]
  33. Guo, T.; Luo, F.; Zhang, L.; Member, S.; Tan, X.; Liu, J.; Zhou, X. Target Detection in Hyperspectral Imagery via Sparse and Dense Hybrid Representation. IEEE Geosci. Remote Sens. Lett. 2020, 17, 716–720. [Google Scholar] [CrossRef]
  34. Shaw, G.A.; Burke, H.K. Spectral Imaging for Remote Sensing. Lincoln Lab. J. 2003, 14, 3–28. [Google Scholar] [CrossRef]
  35. Fan, X.; Zhang, Y.; Li, F.; Chen, Y.; Shao, T.; Zhou, S. A robust spectral target recognition method for hyperspectral data based on combined spectral signatures. In Proceedings of the 2011 IEEE International Geoscience and Remote Sensing Symposium, Vancouver, BC, Canada, 24–29 July 2011; pp. 4328–4331. [Google Scholar] [CrossRef]
  36. Wang, T.; Du, B.; Zhang, L. An automatic robust iteratively reweighted unstructured detector for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2367–2382. [Google Scholar] [CrossRef]
  37. Zou, Z.; Shi, Z. Hierarchical Suppression Method for Hyperspectral Target Detection. IEEE Trans. Geosci. Remote Sens. 2016, 54, 330–342. [Google Scholar] [CrossRef]
  38. Hao, X.; Wu, Y.; Wang, P. Angle distance-based hierarchical background separation method for hyperspectral imagery target detection. Remote Sens. 2020, 12, 697. [Google Scholar] [CrossRef] [Green Version]
  39. Nasrabadi, N.M. Regularized spectral matched filter for target recognition in hyperspectral imagery. IEEE Signal Process. Lett. 2008, 15, 317–320. [Google Scholar] [CrossRef]
  40. Sherman, J.; Morrison, W.J. Adjustment of an Inverse matrix corresponding to a change in one element of a given matrix. Ann. Math. Stat. 1950, 21, 124–127. [Google Scholar] [CrossRef]
  41. Chang, C.I. Multiparameter receiver operating characteristic analysis for signal detection and classification. IEEE Sens. J. 2010, 10, 423–442. [Google Scholar] [CrossRef]
  42. Robinson, G.K. Receiver Operating Characteristic Curve Confidence Intervals and Regions. Wiley Encycl. Clin. Trials 2008, 5, 251–255. [Google Scholar] [CrossRef]
  43. Liu, X.; Tanaka, M.; Okutomi, M. Single-image noise level estimation for blind denoising. IEEE Trans. Image Process. 2013, 22, 5226–5237. [Google Scholar] [CrossRef]
  44. Kang, X.; Zhang, X.; Li, S.; Li, K.; Li, J.; Benediktsson, J.A. Hyperspectral Anomaly Detection with Attribute and Edge-Preserving Filters. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5600–5611. [Google Scholar] [CrossRef]
  45. Xu, Y.; Wu, Z.; Li, J.; Plaza, A.; Wei, Z. Anomaly detection in hyperspectral images based on low-rank and sparse representation. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1990–2000. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed hierarchical background suppression on the matched filter detector (HSMF).
Figure 1. Flowchart of the proposed hierarchical background suppression on the matched filter detector (HSMF).
Sensors 21 00144 g001
Figure 2. The original spectral observation space is shown at the left, and the whitened space with spherical covariance, created by the whitening transformation, is shown at the right. μ 0 and μ 1 represent the background and the target in the original spectral observation space, respectively.   μ ˜ 0 and   μ ˜ 1 represent the background and the target in whitened space, respectively. There is an aliasing area in the original spectral observation space. Both the target and background are considered as Gaussian distribution. The green dotted line works as a decision boundary.
Figure 2. The original spectral observation space is shown at the left, and the whitened space with spherical covariance, created by the whitening transformation, is shown at the right. μ 0 and μ 1 represent the background and the target in the original spectral observation space, respectively.   μ ˜ 0 and   μ ˜ 1 represent the background and the target in whitened space, respectively. There is an aliasing area in the original spectral observation space. Both the target and background are considered as Gaussian distribution. The green dotted line works as a decision boundary.
Sensors 21 00144 g002
Figure 3. Detection results of different detectors on the San Diego Airport data. (a1a7): results of HSMF for the first to the seventh layers. (a8): The final result of HSMF. (bk): Results of spectral angle mapper (SAM), constrained energy minimization (CEM), inequality constraint on CEM (CEM-IC), spectral matched filter (SMF), adaptive coherence estimator (ACE), sparsity-based target detector (STD), matched subspace detector (MSD), sparse representation-based binary hypothesis (SRBBH), hierarchical CEM (hCEM), and angle distance-based hierarchical background separation (ADHBS). (m): The groundtruth information. (n) The false color image (bands 11, 40, 88).
Figure 3. Detection results of different detectors on the San Diego Airport data. (a1a7): results of HSMF for the first to the seventh layers. (a8): The final result of HSMF. (bk): Results of spectral angle mapper (SAM), constrained energy minimization (CEM), inequality constraint on CEM (CEM-IC), spectral matched filter (SMF), adaptive coherence estimator (ACE), sparsity-based target detector (STD), matched subspace detector (MSD), sparse representation-based binary hypothesis (SRBBH), hierarchical CEM (hCEM), and angle distance-based hierarchical background separation (ADHBS). (m): The groundtruth information. (n) The false color image (bands 11, 40, 88).
Sensors 21 00144 g003
Figure 4. (a) The receiver operating characteristic (ROC) curves of different detectors on the San Diego Airport data. (b) The separability map of all detectors for the San Diego Airport data.
Figure 4. (a) The receiver operating characteristic (ROC) curves of different detectors on the San Diego Airport data. (b) The separability map of all detectors for the San Diego Airport data.
Sensors 21 00144 g004
Figure 5. Detection results of the Airport-Beach-Urban data. (a1a5): Results of HSMF for the first to the fifth layers. (a6): The final result of HSMF. (bk): Results of SAM, CEM, CEM-IC, SMF, ACE, STD, MSD, SRBBH, hCEM, and ADHBS. (m): The groundtruth information. (n) The false color image (bands 23, 56, 72).
Figure 5. Detection results of the Airport-Beach-Urban data. (a1a5): Results of HSMF for the first to the fifth layers. (a6): The final result of HSMF. (bk): Results of SAM, CEM, CEM-IC, SMF, ACE, STD, MSD, SRBBH, hCEM, and ADHBS. (m): The groundtruth information. (n) The false color image (bands 23, 56, 72).
Sensors 21 00144 g005
Figure 6. (a) ROC curves of different detectors on the Airport-Beach-Urban data. (b) The separability map of all detectors for the Airport-Beach-Urban data.
Figure 6. (a) ROC curves of different detectors on the Airport-Beach-Urban data. (b) The separability map of all detectors for the Airport-Beach-Urban data.
Sensors 21 00144 g006
Figure 7. Detection results of different algorithms on the Urban data. (a1a6): Results of HSMF for the first to the sixth layers. (a7): The final result of HSMF. (bk): Results of SAM, CEM, CEM-IC, SMF, ACE, STD, MSD, SRBBH, hCEM, and ADHBS. (m): The groundtruth information. (n) The false color image (bands 7, 48, 144).
Figure 7. Detection results of different algorithms on the Urban data. (a1a6): Results of HSMF for the first to the sixth layers. (a7): The final result of HSMF. (bk): Results of SAM, CEM, CEM-IC, SMF, ACE, STD, MSD, SRBBH, hCEM, and ADHBS. (m): The groundtruth information. (n) The false color image (bands 7, 48, 144).
Sensors 21 00144 g007
Figure 8. (a) ROC curves of different detectors on the Urban data. (b) The separability map of all detectors for the Urban data.
Figure 8. (a) ROC curves of different detectors on the Urban data. (b) The separability map of all detectors for the Urban data.
Sensors 21 00144 g008
Figure 9. The average energy of HSMF with different layers on three hyperspectral data.
Figure 9. The average energy of HSMF with different layers on three hyperspectral data.
Sensors 21 00144 g009
Figure 10. Output mean value of different layers on three hyperspectral data.
Figure 10. Output mean value of different layers on three hyperspectral data.
Sensors 21 00144 g010
Figure 11. The effects of the different suppression criteria on the San Diego Airport data. The suppression criteria used in the comparative experiments are one-eighth mean, quarter mean, mean, one-half mean, and three-quarter mean. (a) The average output energy of different layers for different suppression criteria. (b) The ROC curves comparison for different suppression criteria.
Figure 11. The effects of the different suppression criteria on the San Diego Airport data. The suppression criteria used in the comparative experiments are one-eighth mean, quarter mean, mean, one-half mean, and three-quarter mean. (a) The average output energy of different layers for different suppression criteria. (b) The ROC curves comparison for different suppression criteria.
Sensors 21 00144 g011
Figure 12. The effects of the different β values on the San Diego Airport data. We choose 10 different values for comparison, including 0.5, 0.4, 0.3, 0.25, 0.2, 0.1, 0.01, 0.005, 0.0005, and 0.0001. (a) The ROC curves comparison for different β values. (b) The average output energy of different layers for different β values.
Figure 12. The effects of the different β values on the San Diego Airport data. We choose 10 different values for comparison, including 0.5, 0.4, 0.3, 0.25, 0.2, 0.1, 0.01, 0.005, 0.0005, and 0.0001. (a) The ROC curves comparison for different β values. (b) The average output energy of different layers for different β values.
Sensors 21 00144 g012
Figure 13. (a) The mean value of all pixels’ suppression coefficient in different layers on three different datasets. (b) The ROC curves comparison for different ε values where ε value contains 0.4, 0.2, 0.05, 0.01, 0.007, 0.006, and 0.005.
Figure 13. (a) The mean value of all pixels’ suppression coefficient in different layers on three different datasets. (b) The ROC curves comparison for different ε values where ε value contains 0.4, 0.2, 0.05, 0.01, 0.007, 0.006, and 0.005.
Sensors 21 00144 g013
Table 1. Area under the receiver operating characteristic (ROC) curves (AUC) values and AUC values at lower false alarm rate (FAR) of all detectors for the San Diego Airport data.
Table 1. Area under the receiver operating characteristic (ROC) curves (AUC) values and AUC values at lower false alarm rate (FAR) of all detectors for the San Diego Airport data.
AlgorithmSAMCEMCEM-ICSMFACESTD
AUC0.98600.99050.99410.99040.97720.8441
AUC (lower FAR)0.17680.49900.44850.50170.54060.0868
AlgorithmMSDSRBBHhCEMADHBSHSMF
AUC0.96870.86490.99240.999910.9925
AUC (lower FAR)0.29030.09410.95620.94700.95871
1 The bold represents the maximum.
Table 2. AUC values and AUC values at lower FAR of all detectors for the Airport-Beach-Urban data.
Table 2. AUC values and AUC values at lower FAR of all detectors for the Airport-Beach-Urban data.
AlgorithmSAMCEMCEM-ICSMFACESTD
AUC0.99720.99940.99680.99940.99950.7695
AUC (lower FAR)0.40810.75640.13250.75850.75600.0321
AlgorithmMSDSRBBHhCEMADHBSHSMF
AUC0.95030.79941.00001.00001.0000
AUC (lower FAR)0.34620.15721.00001.00001.0000
The bold represents the maximum.
Table 3. AUC values and AUC values at lower FAR of all detectors for the Urban data.
Table 3. AUC values and AUC values at lower FAR of all detectors for the Urban data.
AlgorithmSAMCEMCEM-ICSMFACESTD
AUC0.90200.99900.99400.99910.99960.9988
AUC (lower FAR)0.47780.45930.42960.46670.72590.4667
AlgorithmMSDSRBBHhCEMADHBSHSMF
AUC0.97790.97191.00000.99981.0000
AUC (lower FAR)0.45190.42591.00000.85931.0000
The bold represents the maximum.
Table 4. Time performances of different detectors on the three datasets (keep four decimal places in the units of seconds).
Table 4. Time performances of different detectors on the three datasets (keep four decimal places in the units of seconds).
AlgorithmsSan Diego AirportAirport-Beach-UrbanUrban
Total TimeTime/LayerTotal TimeTime/LayerTotal TimeTime/Layer
SAM0.05760.05760.02490.02490.01770.0177
CEM0.16590.16590.04020.04020.02780.0278
CEM_IC0.40600.40600.28550.28550.20160.2016
SMF0.20540.20540.05910.05910.04870.0487
ACE0.59490.59490.08150.08150.09360.0936
STD76.582876.582817.855717.855717.502817.5028
MSD369.7743369.774392.205492.205468.223468.2234
SRBBH59.155859.155810.858510.858514.264314.2643
hCEM3.8807(15) 10.25871.2482(8)0.15601.1822(8)0.1478
ADHBS4.3435(11)0.39492.6310(17)0.15473.3069(22)0.1503
HSMF4.2831(7)0.61191.3298(5)0.26591.2975(6)0.2163
1 The value in parentheses indicates the number of iterations.
Table 5. The number of iterations and detection performance for different β values.
Table 5. The number of iterations and detection performance for different β values.
Suppression coefficient β0.50.40.30.250.2
AUC0.98510.98510.99250.99250.9925
AUC (lower FAR)0.96380.96380.94760.94430.9592
Number of iterations3527262321
Suppression coefficient β0.10.010.0050.00050.0001
AUC0.99250.99250.99250.99250.9925
AUC (lower FAR)0.97420.93100.94090.95620.9599
Number of iterations1616161616
Table 6. The number of iterations and detection performance for different ε values.
Table 6. The number of iterations and detection performance for different ε values.
Threshold value ε0.40.20.050.010.0070.0060.005
AUC0.99040.99060.99220.99250.99250.99250.9925
AUC (lower FAR)0.55990.63590.87180.95870.96980.96700.9599
Layer number1247911NaN 1
1 NaN (Not a Number) indicates that the termination condition cannot be met, and the algorithm will iterate indefinitely.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, C.; Wu, Y.; Hao, X. Hierarchical Suppression Based Matched Filter for Hyperspertral Imagery Target Detection. Sensors 2021, 21, 144. https://doi.org/10.3390/s21010144

AMA Style

Gao C, Wu Y, Hao X. Hierarchical Suppression Based Matched Filter for Hyperspertral Imagery Target Detection. Sensors. 2021; 21(1):144. https://doi.org/10.3390/s21010144

Chicago/Turabian Style

Gao, Ce, Yiquan Wu, and Xiaohui Hao. 2021. "Hierarchical Suppression Based Matched Filter for Hyperspertral Imagery Target Detection" Sensors 21, no. 1: 144. https://doi.org/10.3390/s21010144

APA Style

Gao, C., Wu, Y., & Hao, X. (2021). Hierarchical Suppression Based Matched Filter for Hyperspertral Imagery Target Detection. Sensors, 21(1), 144. https://doi.org/10.3390/s21010144

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop