Next Article in Journal
SG-Det: Shuffle-GhostNet-Based Detector for Real-Time Maritime Object Detection in UAV Images
Next Article in Special Issue
Spectral–Temporal Transformer for Hyperspectral Image Change Detection
Previous Article in Journal
Unveiling Air Pollution in Crimean Mountain Rivers: Analysis of Sentinel-5 Satellite Images Using Google Earth Engine (GEE)
Previous Article in Special Issue
Target Detection in Hyperspectral Remote Sensing Image: Current Status and Challenges
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CRNN: Collaborative Representation Neural Networks for Hyperspectral Anomaly Detection

1
Department of Aerospace Information Engineering, Beihang University, Beijing 100191, China
2
Wuhan Digital Engineering Institute, Wuhan 430205, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2023, 15(13), 3357; https://doi.org/10.3390/rs15133357
Submission received: 15 May 2023 / Revised: 26 June 2023 / Accepted: 29 June 2023 / Published: 30 June 2023
(This article belongs to the Special Issue Advances in Hyperspectral Data Exploitation II)

Abstract

:
Hyperspectral anomaly detection aims to separate anomalies and backgrounds without prior knowledge. The collaborative representation (CR)-based hyperspectral anomaly detection methods have gained significant interest and development because of their interpretability and high detection rate. However, the traditional CR presents a low utilization rate for deep latent features in hyperspectral images, making the dictionary construction and the optimization of weight matrix sub-optimal. Due to the excellent capacity of neural networks for generation, we formulate the deep learning-based method into CR optimization in both global and local streams, and propose a novel hyperspectral anomaly detection method based on collaborative representation neural networks (CRNN) in this paper. In order to gain a complete background dictionary and avoid the pollution of anomalies, the global dictionary is collected in the global stream by optimizing the dictionary atom loss, while the local background dictionary is obtained by using a sliding dual window. Based on the two dictionaries, our two-stream networks are trained to learn the global and local representation of hyperspectral data by optimizing the objective function of CR. The detection result is calculated by the fusion of residual maps of original and represented data in the two streams. In addition, an autoencoder is introduced to obtain the hidden feature considered as the dense expression of the original hyperspectral image, and a feature extraction network is concerned to further learn the comprehensive features. Compared with the shallow learning CR, the proposed CRNN learns the dictionary and the representation weight matrix in neural networks to increase the detection performance, and the fixed network parameters instead of the complex matrix operations in traditional CR bring a high inference efficiency. The experiments on six public hyperspectral datasets prove that our proposed CRNN presents the state-of-the-art performance.

1. Introduction

Hyperspectral remote sensing integrates two sensing technologies of panchromatic imaging and spectrometry to fully exploit the spatial and spectral features [1]. Hyperspectral images record nearly continuous spectral bands and better reflect the physical properties of the objects, which helps to identify subtle differences between various materials. Therefore, they are widely employed in both civil and military fields, such as image classification [2,3,4] and target detection [5,6,7]. As one of the vital research branches of target detection, hyperspectral anomaly detection distinguishes anomalies from background in unsupervised way for lack of prior knowledge [8]. Due to the challenges in establishing a comprehensive and accurate spectrum database for targets, hyperspectral anomaly detection has received a lot of interest and has great practical prospects [9].
In hyperspectral anomaly detection, the anomaly is typically considered as the target with a notably different spectral features from the surrounding background pixels [1,10]. In real hyperspectral images, it is difficult to consider all background clusters since the background occupies most of the pixels and has a complex composition [11]. Moreover, hyperspectral images are highly unbalanced, and the occurrence probability of anomalies is low, which makes it impossible to extract the intrinsic features of anomalies. Therefore, the main challenge in hyperspectral anomaly detection is to effectively suppress the background while distinguishing the anomalies.
The existing anomaly detection methods exploit the property of anomalies with low probability and high deviation from the surrounding background cluster to distinguish anomalies from background [10,12]. The most widely known benchmark algorithm is the Reed–Xiaoli (RX) detector [13]. RX utilizes the whole image to build background model under the presumption that the background is distributed according to a multivariate Gaussian distribution. The probability of anomalies is measured by the Mahalanobis distance between the test pixel and the background. The statistical theory-based hyperspectral anomaly detection methods use a simple mathematical model and have low computational consumption. Subsequently, researchers have proposed many improved algorithms based on RX, such as Kernel RX [14], Local RX [15], Dual-Window RX [16], and Weighted RX [17]. In addition to the generalized likelihood ratio test-based RX, Chang [9] developed dummy variable trick (DVT) to convert hyperspectral target detection to anomaly detection, and also introduced effective anomaly space (EAS) [18] to remove the background while retain the anomaly. Cluster-based hyperspectral anomaly detectors [19,20], the support vector data description (SVDD) [21], and graph-based methods [22,23] take the distance to measure the anomaly score. Ref. [24] introduced a fractional Fourier entropy (FrFE)-based hyperspectral anomaly detection method to distinguish anomalies from background. To utilize the spatial features in the hyperspectral data, an attribute and edge-preserving filtering detector (AED) [25] and other filtering-based methods [26,27] were proposed. Tensor-based methods treat the hyperspectral cube as a third-order tensor and utilize the spectral and spatial information [28]. A prior-based tensor approximation (PTA) method for hyperspectral anomaly detection is proposed in [29], which decomposed the hyperspectral data into a background tensor and an anomaly tensor under a low-rank and a piecewise-smooth prior. Iterative spectral–spatial hyperspectral anomaly detector (ISSHAD) [30] extracted spectral–spatial information via anomaly detection maps and gave feedback to create a new expanded data cube in iteration.
The representation-based methods, including sparse representation (SR) [31,32,33], collaborative representation (CR) [34,35,36,37,38,39,40], and low-rank representation (LRR) [41,42,43,44,45], try to use a linear combination of atoms in the constructed background dictionary to represent the hyperspectral images, and assume that the background pixels will be represented better than anomalous pixels since the clusters of anomalies are far different from the background clusters. Therefore, representation-based methods usually use the residual map of the represented image and the original image to detect the anomaly.
The SR-based approaches assume that only a small subset of atoms in the given over-complete dictionary contribute to the representation under the sparsity constraint. Li et al. [32] adopted principal component analysis (PCA) [46] to learn a robust background dictionary and proposed a SR model based on the re-weighted Laplace prior. The CR-based methods utilize all of the background dictionary atoms to represent each pixel with the  l 2 -norm [47] to constrain the representation coefficients for better background modeling capability [34]. The CR-based methods have received many developments due to their computational simplicity and good performance, but the pollution of anomalous atom caused by the sliding dual window limits the performance of the method. To solve this issue, researchers have devised lots of methods. Vafadar and Ghassemian [35] removed anomalous pixels that deviate noticeably from the distribution of most pixels when constructing the background dictionary. Other methods were also tried in [36,37,38,39] to exclude potential anomalous pixels and obtain a pure background. In contrast to the popular dual window strategy, Wu et al. [40] introduced a practical non-global dictionary and proposed a new relaxed CR hyperspectral anomaly detector. The LRR-based approaches divide the hyperspectral images into background and anomalies with the assumption that the background pixels are low rank because they are widely dispersed and homogeneous, while the anomalies are few and sparse. The LRR theory determines the lowest rank representation of the background based on the background dictionary, and constrains the anomaly with sparsity [41]. Xu et al. [42] combined low-rank and sparse representation (LRASR) and constructed a stable and discriminative dictionary. Zhang et al. [43] implemented low-rank and sparse decomposition (LRaSMD) through the Go Decomposition (GoDec) [48] approach and adopted the RX detector to detect anomalies. Cheng et al. [44] proposed a graph and total variation regularized LRR-based model (GTVLRR) to retain the local geometrical and spatial structure. LRR-based methods are still widely used due to their high detection performance [11,45,49].
With the development of artificial intelligence, researchers have introduced deep learning into hyperspectral anomaly detection [50]. Due to the small proportion of anomalies in hyperspectral data that limits their contribution to model training, and the absence of prior spectral knowledge, most current deep learning-based hyperspectral anomaly detection methods chose unsupervised learning [49,51,52,53,54,55,56,57].
The deep neural networks are introduced to learn the intrinsic expression in the hyperspectral data to take advantage of their capacity on complex data modeling and feature extraction. The traditional method is then performed on the extracted features to detect anomalies. Lei et al. [51] utilized a deep brief network (DBN) [58] to automatically learn discriminative features and high-level representations in the spectral domain and fed them into an RX detector. Based on the generative models of autoencoder (AE) [59] and Generative Adversarial Networks (GAN) [60], the original hyperspectral data are reconstructed and the residuals of the original and the reconstructed data are used for anomaly detection. The AE-based methods have received a lot of attention recently. In [52], manifold learning is introduced to an AE network for hyperspectral anomaly detection. Jiang et al. [49] proposed a low-rank embedded network (LREN) to learn spectral features via an AE, and the LRR-based anomaly detector is used in the deep feature space. Xiang et al. [61] proposed a guided AE (GAE) which introduced a guided image to suppress the feature learning of anomalies and used the reconstruction error as the metric. GAN-based methods introduce adversarial learning [53] into background reconstruction. Jiang et al. [54] proposed GAN for hyperspectral anomaly detection (HADGAN) and reconstructed the hyperspectral data via an adversarial autoencoder (AAE) [62]. The morphological filter and RX were implemented on the residual image. Xie et al. [55] also used the AAE for reconstruction, but chose to use the morphological method on the hidden layer features.
Since the above methods usually detect anomalies in two stages, and deep features are not extracted specially for the hyperspectral anomaly detection tasks, the separate training of the feature extractor and the anomaly detector hinders deep neural networks from attaining their full potential. Therefore, an autonomous hyperspectral anomaly detection network (Auto-AD) is proposed in [56] to avoid the pre-processing and post-processing in other two-stage networks. Xie et al. [57] proposed an end-to-end Spectral Distribution-Aware Estimation network (SDEN) to directly estimate anomalies, which jointly learned the two parts of the feature extraction network and the anomaly detector. Meanwhile, in addition to the above unsupervised methods, weakly supervised methods are gradually becoming an attractive option. Most of the current weakly supervised methods firstly filtered out background pseudo-samples for training based on the estimated background distribution [10,11,63,64].
The traditional hyperspectral anomaly detection methods have a reasonable mathematical foundation and interpretability, but the shallow learning method somehow limits their detection performance. While the deep learning-based methods have a high rate in feature utilization and strong generalization ability, their interpretability is weaker than that of traditional algorithms. Therefore, a new hyperspectral anomaly detection method based on Collaborative Representation Neural Networks (CRNN) is proposed in this paper. Our motivation is to implement traditional representation theory of hyperspectral anomaly detection into the framework of deep neural networks.
To be specific, the fairly complete background dictionaries of global and local streams are constructed, and the representation weight matrices are learned by optimizing the objective function of CR in the two-stream networks. The global background dictionary is a generalization of the global background, which is constructed based on all pixels of the entire image. To avoid the pollution of the anomalous pixels, the global CR networks is trained to obtain the most representative dictionary atoms under the constraint of the dictionary atom loss function. The local background dictionary focuses on the local information surrounding the test pixel, which is obtained by a sliding dual window. The global and local representation images of the test pixel are calculated through the linear combination of atoms in the two dictionaries, respectively. In traditional CR-based methods, the optimal solution of weight coefficients usually involves matrix inversion and other complex operations, which would bring large computational load and time consumption. While, in our proposed method, the two-stream networks learn the weight matrix via iterative optimization during training and the fixed parameters in inference, which leads to a high calculation efficiency. Finally, the anomaly score of each pixel can be measured by the residual input data and its representation through networks based on the jointly considered global and local background dictionary.
In addition, CRNN firstly introduces an AE to reconstruct the original input hyperspectral data to obtain the dimensionality reduction and dense expression of the original image. The global and local CR is then operated on the hidden feature map of the AE. Moreover, we design a feature extraction network to further learn the comprehensive spectral–spatial features in the hidden feature map.
The main contributions of this paper are as follows:
  • We propose a new collaborative representation neural network for hyperspectral anomaly detection that combines the mathematical foundation of CR theory with the effective feature learning capacity of deep neural networks. CRNN optimizes the objective function of CR through iterative training, and improves the detection performance with faster inference speed.
  • We perform both global and local streams of CR, and fuse the detection maps to obtain the final result comprehensively. The global dictionary is collected in the networks by optimizing the dictionary atom loss to refine the background without the pollution of anomalies. While the local dictionary by a sliding dual window reflects the background neighboring the test pixel.
  • We utilize a feature extraction network to generate the comprehensive feature map, including spectral–spatial features from both global and local perspectives. The deep latent feature space helps to learn the representation weigh matrices more effectively.
The remainder of our article is organized as follows. Section 2 reviews some related works. Section 3 introduces the proposed method. Section 4 provides the experiments in detail. Section 5 makes the discussion. Finally, the conclusion is presented in Section 6.

2. Related Works

We briefly review major works on CR-based detector [34] and AE [59] in this section.

2.1. Collaborative Representation-Based Detector

For a given hyperspectral dataset,  X = x 1 , x 2 , , x N , where  x i R d  is the i-th pixel in the data, the CR theory believes that the approximate representation  y ^  can be calculated via the linear combination of background pixels in the surrounding area of the test pixel  y ,
y ^ = X s α
where  y ^  is the collaborative representation of the test pixel.  X s  is the collection of background pixels around the test pixel through a dual window.  α R s × 1  is the weight vector.
The Euclidean distance r between  y  and  y ^  is usually introduced to measure the probability of anomaly
r = y y ^ 2 = y X s α 2
where  · 2  is the  l 2 -norm.
Collaboration means that the constraint on the weight vector  α  is that  y X s α 2 2  is the smallest, while  α 2 2  is the smallest, so the objective function is
min α y X s α 2 2 + λ α 2 2
where  λ  is the Lagrange multiplier.  l 2 -norm provides each element in  α  a tiny value and guarantees that all atoms take part in the representation.
The formula (3) is equal to:
min α α T X s T X s + λ I α 2 α T X s T y
The above formula (4) belongs to the convex function optimization problem and exists as a closed-form solution for  α . Setting the derivative to 0 gains the optimal solution for  α ,
α = X s T X s + λ I 1 X s T y

2.2. Autoencoder

AE is an unsupervised deep learning-based neural network model and consists of the encoder and the decoder. The encoder firstly maps the input vector  x  from input layer to hidden layer, while the decoder is a structure where the vector  z  in hidden layer is mapped to the output vector  x ^ ,
z = f ( x ; θ ) x ^ = f ^ ( z ; θ ^ )
where  f ( · ; θ )  and  f ^ ( · ; θ ^ )  are the mapping functions of the encoder and the decoder.
In order to accurately reconstruct the input vector  x , the divergence between  x  and  x ^  is as small as possible during training, so that the hidden  z  can well reflect the feature of  x . The loss function  L R  is as follows:
L R = x x ^ 2

3. Proposed Method

In this section, a hyperspectral anomaly detection method named CRNN is proposed. The flow chart is presented in Figure 1. Firstly, an AE reconstructs the original input hyperspectral data  X . The hidden feature  Z  during the reconstruction process serves as the basis for subsequent anomaly detection. Secondly, the comprehensive feature extraction network further learns the global and local spectral–spatial feature in  Z  to obtain comprehensive feature map  Z c . The CR learning network includes global stream and local stream. The global background dictionary  D g  and the local background dictionary  D l  are both obtained from  Z , while the global representation weight  α g  and local representation weight  α l  are learned based on  Z c . Finally, the anomaly score map  A  is calculated in the detection fusion module.

3.1. Hidden Feature Map Generation

High spectral resolution and continuous spectral bands in hyperspectral data produce the issue of highly correlated and redundant bands. Therefore, the original hyperspectral data are densely expressed before anomaly detection to improve computational efficiency, specifically through the encoding process in an AE [59].
Supposing the original hyperspectral data as  X R H × W × K , where K is the number of bands of the original data, and N is the total amount of pixels in the image that  N = H × W  (H and W are the height and width, respectively). As shown in Figure 1, the reshaped  X = x 1 , x 2 , , x N R K × N  is firstly input into an AE, and the output reconstructed image is  X ^ = x ^ 1 , x ^ 2 , , x ^ N R K × N . In the encoder E, the original hyperspectral data  X  are firstly encoded into a lower-dimensional and denser hidden variable  Z = z 1 , z 2 , , z N R k × N , where k is the number of bands of the hidden feature map. This hidden variable  Z  is mapped back to the original hyperspectral space in the decoder De for the reconstruction  X ^ . Huber loss  L δ  [65] is chosen as the loss function,
L δ ( x , x ^ ) = 1 2 ( x x ^ ) 2 for | x x ^ | δ δ | x x ^ | 1 2 δ 2 otherwise
where  δ  is the hyperparameters.
The structures of E and De are similar as both are composed of  l a  numbers of 1 × 1 two-dimensional convolutional layers. The hidden feature map  Z  is obtained from the output of the  l a -th layer in E,
X i + 1 = σ G N f i E X i ; Θ i E , i = 1 , , l a 1 Z = f l a E X l a ; Θ l a E
where  X i  is the input datum of the i-th layer,  f i E · ; Θ i E  denotes the two-dimensional convolution with parameters  Θ i E  of the i-th layer in encoder E,  G N ( · )  is the group normalization operation [66], and  σ  is a leakly rectified linear unit (Leakly ReLU) activation function [67].

3.2. Comprehensive Feature Extraction

Since the AE only considers the spectral dimension of each pixel during the reconstruction without the global and local structural information, it is necessary to further extract the features of  Z  to mine its deep potential features.

3.2.1. Non-Local Module

A non-local network [68] is introduced to learn global correlation features.  Z = z 1 , z 2 , , z N R k × N  is the input of the module, and  Y g = y 1 g , y 2 g , , y N g T R N × k  is the global feature calculated by the module,
y i g = 1 C i j f z i , z j g z j
where i and j are the indices of the pixel position; the function  f ( · )  calculates the similarity between the two pixels, and Gaussian function is chosen in our model,  f z i , z j = e z i T z j ; the function  g ( · )  denotes the feature representation, which is implemented by a 1 × 1 convolutional layer;  C i  is the normalization factor,  C i = j f z i , z j .
Therefore, the non-local module is equivalent to the self-attention mechanism [69], normalized in the form of softmax [70]. The global correlation feature  Y g  can be expressed as:
Y g = softmax Z T Z g ( Z )
Finally, the output feature map  Z n l R N × k  of the non-local module with the residual connection is
Z n l = W Y g + Z T
where  W  denotes the weight matrix learned by the network.

3.2.2. Local Feature Learning Module

The local feature learning module consists of two convolutional layers, a two-dimensional one and a three-dimensional one, as shown in Figure 1. In the 1 × 1 two-dimensional convolution layer, the spectral features for each pixel are learned as  V R H × W × k . In the 3 × 3 × 3 three-dimensional convolutional layer, the local spectral–spatial features are extracted on the expanded data  V e R H × W × 1 × k  with an additional dimension. The three-dimensional convolution not only fuses the spatial neighboring pixels to learn the local spatial features, like the 3 × 3 two-dimensional convolution, but also extracts the local spectral features via the convolution on the adjacent spectral bands, since some anomalies typically occur in the local spectral bands.
For the reshaped input feature  Z n l R H × W × k , the comprehensive feature map  Z c R H × W × n k  is obtained as followed
V = σ G N f conv 2 Z n l ; Θ conv 2 Z c = σ G N f conv 3 V e ; Θ conv 3
where  f conv 2 · ; Θ conv 2  denotes the two-dimensional convolution and parameters  Θ conv 2  and  f conv 3 · ; Θ conv 3  denote the three-dimensional convolution with parameters  Θ conv 3 .

3.3. Two-Stream Collaborative Representation Learning Networks

Based on the hidden layer features  Z R k × N , we design a two-stream CR learning network for global and local streams.

3.3.1. Global Collaborative Representation Learning

The calculation of the global collaborative representation image  Z ^ g R k × N  is
Z ^ g = D g α g
where  D g R k × C  is the global background dictionary;  α g R C × N  is the global representation weight matrix; and C is the amount of dictionary atoms.
The global representation weight matrix  α g  is obtained through the global CR learning networks based on the comprehensive feature map  Z c , by  l b  layers of 1 × 1 convolutional layer,
Z i + 1 g = σ G N f i G C Z i g ; Θ i G C , i = 1 , , l b 1 α g = f l b G C Z l b g ; Θ l b G C
where  Z i g  is the input datum of the i-th layer in the global CR learning networks, and  Z 1 g = Z c .
In order to make the error between the global representation  Z ^ g  of the network and the hidden feature map  Z  as small as possible, the loss function  L G l o b a l  of the global CR learning network is designed according to the CR objective function in formula (3)
L G l o b a l = i = 1 N z i z ^ i g 2 2 + λ α g 2 2
where  z i  is the i-th pixel in  Z , and  z ^ i g  is the i-th pixel in  Z ^ g .
Moreover, we design a loss function  L D i c  to constrain the dictionary atoms and make them more representative in  D g
L D i c = j = 1 C i = 1 N d j g z i 2 2
where  d j g  is the j-th atom in  D g . The purpose of this constraint is to minimize the distance from each dictionary atom to the hidden feature  Z .

3.3.2. Local Collaborative Representation Learning

Different from the global CR learning networks, the local background dictionary is obtained via a sliding dual window on the hidden feature  Z , with the sizes of inner and outer windows as  w i n in  and  w i n out , respectively. For the i-th pixel in  Z , the local collaborative representation  Z ^ l = z ^ 1 l , z ^ 2 l , , z ^ N l R k × N  is calculated as
z ^ i l = D i l α i l i = 1 , 2 , 3 N
where  z ^ i l  is the i-th pixel of  Z ^ l D i l R k × S  denotes the local background dictionary of the i-th pixel, which is a collection of the feature vectors between the inner and outer windows centered on the i-th pixel in  Z α i l R S × 1  is the weight coefficient of the i-th pixel in matrix  α l R S × N ; and S represents the number of background pixels between the inner and outer windows that  S = w i n out × w i n out w i n in × w i n in .
Since the weight matrix  α l  depends on the background pixels in the dual window, the convolutional kernel size in local CR learning networks is set to  w i n out × w i n out . The number of layers is still  l b  the same as the global CR learning networks, and the calculation of  α l  is similar with  α g  in Equation (15).
The loss function of local collaborative representation  Z ^ l  is also designed according to the objective function (3):
L L o c a l = i = 1 N z i z ^ i l 2 2 + λ α l 2 2
Therefore, combining the two streams of global and local, the objective function of CR learning networks is obtained as follows
L C R = L D i c + λ 1 L G l o b a l + λ 2 L L o c a l
where  λ 1  and  λ 2  are the factors of the trade-off between the global and local streams.
Due to the introduction of learning layers in the neural network, the convex optimization problem in Formula (3) is transformed into a non-convex loss function in (20). The deep learning-based algorithm seeks multiple possible local optimal solutions during network iteration, instead of being restricted to a unique closed-form solution. In this way, the generalization performance of the detector would be enhanced and noise interference would be more difficult.

3.4. Detection Fusion

The Euclidean distance between the hidden feature  Z  and the global representation  Z ^ g  or the local representation  Z ^ l  is calculated, respectively, to obtain the global detection result  A g R H × W  and the local detection result  A l R H × W
A i g = z i z ^ i g 2 i = 1 , 2 , 3 N
A i l = z i z ^ i l 2 i = 1 , 2 , 3 N
where  A i g  is the i-th pixel in  A g , and  A i l  is the i-th pixel in  A l .
The final anomaly score map  A R H × W  will be obtained by element-wise multiplying of  A g  and  A l  to achieve a higher detection accuracy:
A = A g A l
The proposed CRNN method is described in Algorithm 1. CFL-Net stands for the comprehensive feature extraction networks and CRL-Net stands for the two-stream CR learning networks. AE_pre_num is epochs that the AE is trained individually during the pre-training phase to strengthen the stability of network training. AE_pre_num = 10.
Algorithm 1: Algorithm flow diagram of CRNN
  Input: original Hyperspectral data  X = x 1 , x 2 , , x H × W R H × W × K .
  Output: Anomaly score map  A = a 1 , a 2 , , a H × W R H × W .
Remotesensing 15 03357 i001

4. Experiments

To validate the detection performance of our proposed CRNN, we perform experiments on six real hyperspectral images and compare with eight previous algorithms.

4.1. Experimental Settings

4.1.1. Datasets Description

Six datasets are employed in the experiment, including Beach [25], Urban [25], XiongAn [71], HYDICE [72], Gulfport [25], and SanDiego [42]. Table 1 lists the details of each dataset. The targets in the images of Beach, Urban, XiongAn, and HYDICE are small. While the anomalies in the datasets of Gulfport and SanDiego are relatively large and have specific shape features, which are planar targets.

4.1.2. Evaluation Criteria

To quantitatively evaluate the detection performance, the three-dimensional receiver operating characteristic (3D ROC) curves and the area under the ROC curve (AUC) [73] are introduced. The 3D ROC curves depict the trade-off among the three parameters of the true positive rate  P d , the false positive rate  P f , and the detection threshold  τ . Three types of two-dimensional ROC (2D ROC) curves decomposed from the 3D ROC curve are also applied in the experiment, including  ( P d , P f ) ( P d , τ ) , and  ( P f , τ ) . Correspondingly, three types of AUC values are applied, and their values range in [0, 1]. The  AUC ( D , F )  (AUC of curve  ( P d , P f ) ) value approaching to 1 indicates a perfect effectiveness, the  AUC ( D , τ )  (AUC of curve  ( P d , τ ) ) value closer to 1 evaluates the better detection probability, and the  AUC ( F , τ )  (AUC of curve  ( P f , τ ) ) value approaching 0 indicates an outstanding performance on background suppression. In addition, another five AUC values [74] are introduced to further evaluate the detector of joint anomaly detection (JAD), joint background suppression (JBS), anomaly detection with background suppression (ADBS), overall anomaly detection probability (OADP), and signal-to-noise probability ratio (SNPR). They are denoted as  AUC JAD AUC JBS AUC ADBS AUC OADP , and  AUC SNPR , respectively. The five AUC values are calculated in the following equation, and larger AUC values indicate a better performance. The values of  AUC JAD AUC JBS , and  AUC ADBS  range in [0, 2], while  AUC OADP  value is between [0, 3].
AUC JAD = AUC ( D , F ) + AUC ( D , τ ) AUC JBS = AUC ( D , F ) + 1 AUC ( F , τ ) AUC ADBS = AUC ( D , τ ) + 1 AUC ( F , τ ) AUC OADP = AUC ( D , F ) + AUC ( D , τ ) + 1 AUC ( F , τ ) AUC SNPR = AUC ( D , τ ) AUC ( F , τ )

4.1.3. Compared Method

The comparison methods include traditional anomaly detection methods of the benchmark algorithm such as global RX (GRX) [13], and four representation-based methods, such as the original CRD [34], CRD in EAS (EAS_CR) [18], relaxed CRD (RCRD) [40], and the local summation unsupervised nearest regularized subspace anomaly detector (LSUNRS) [38]. We also experimented with recent deep learning-based anomaly detection methods, such as LREN [49], HADGAN [54], and GAE [61]. The comparison algorithms used in this paper are all verified based on the source code released by the original author. In order to assure the equitable results for the comparison experiment, the parameters of the CRD, EAS_CR, RCRD, LSUNRS, LREN, HADGAN, and GAE in the experiment were all optimal.

4.1.4. Implementation Details

Our approach is implemented in Pytorch 1.8, and all the experiments were tested on a workstation with an Nvidia GeForce GTX 1080-Ti GPU and an Intel i9 CPU. According to the network structure described in Section 3, the specific experiment settings are as follows. Both the encoder and the decoder use the structure of a 4-layer convolutional network, i.e.,  l a = 4 . The channel dimension of each layer is configured as E{K, 100, 50, 20, k} and De{k, 20, 50, 100, K}, where K is the number of input hyperspectral channels, and k is the feature dimension of  Z  as  k = 10 . The comprehensive feature  Z c R H × W × n k  is obtained in the feature extraction networks, where  n = 5 . In CR learning networks, both the global and local streams adopt the 3-layer convolutional network, i.e.,  l b = 3 . The channel number configuration is GC{ n k , k, k, C} and LC{ n k , k, k, S}.  δ  in the Huber loss  L δ  in Equation (8) is set as 1.  λ 1  and  λ 2  in the loss function  L C R  in Equation (20) are set to 0.1 and 0.1. During training, the number of epochs is set as 500 to balance the detection accuracy and the calculation load, and the Adam algorithm is chosen to optimize the parameters of networks. The network learning rate is initially set to  10 4  and an exponential decay would be performed after 100 epochs. The above parameters are the default in the experiment and can be adjusted in other versions of this experiment.

4.2. Detection Performance

The anomaly detection maps of the proposed CRNN and the other eight algorithms of the six datasets are displayed in Figure 2. For a further and more intuitive analysis of the quantitative performance of each algorithm, the 3D ROC curves and the three types of 2D ROC curves, including  ( P d , P f ) ( P d , τ ) , and  ( P f , τ ) , are shown in Figure 3. The 2D ROC curves of  ( P d , P f )  are set in a logarithmic scale for better visualization. The eight AUC performance comparisons for each dataset are shown in Table 2, and the average values of the AUC metrics over six datasets are listed in Table 3.

4.2.1. Detection Performance for Small Targets

The detection maps of the Beach dataset are shown in Figure 2a; the target in the detection map of CRNN is prominent, and the suppression of the background is the best since CRNN can effectively suppress the false alarms in the background of this dataset. In the detection maps of other methods, the beach at the bottom left is a more obvious false alarm, especially in LREN. Figure 2b shows the detection maps of the Urban dataset and the result is that CRNN presents outstanding targets and better background suppression. The results of GAE, HADGAN, LREN, RCRD, EAS_CR, and GRX present obvious background texture and banding noise; in particular, the result of LREN presents obvious false detection. The background and targets are both at a low brightness level in the result maps of CRD and LSUNRS. However, the ROC curves  ( P d , P f )  of CRNN and LSUNRS are quite close when  P d  is reaching 1. It may be due to the disparity in the brightness of different targets in the detection result of CRNN, which means only some anomalous pixels are visually obvious. The detection maps of the XiongAn dataset are shown in Figure 2c; the background suppression is not so effective in the results of the eight compared methods, especially in LREN, and the detection maps of HADGAN, LSUNRS, CRD, and GRX present evident banding noise. Compared with CRD, the two CR-based methods of RCRD and EAS_CR improve the background suppression, but still present some false alarms in detection maps. While the result of CRNN not only presents a perfect suppression on the background but also has a strong anti-interference ability against noise. The detection maps of the HYDICE image are shown in Figure 2d. The detection results of CRNN and LREN are best when dealing with prominent targets and good background suppression, while LREN has some instances of missed detection and the detection score of some anomalous pixels is low. The background suppression in other methods is poor; in addition, the background texture is clear in HADGAN, CRD, and GRX, while some points of false alarm present higher brightness in RCRD and EAS_CR.
For small anomalous targets, the proposed CRNN produces visual results closest to the reference compared with other seven state-of-the-art methods. In Figure 3, the ROC curves of  ( P d , P f )  are all the first to reach 1, and the ROC curves of  ( P f , τ )  are all the closest to the lower left corner. It is clear from the study above that the proposed CRNN can clearly highlight the small anomalous targets in the detection maps while effectively suppressing the background, since it utilizes neural networks to learn features that are more conducive to anomaly detection.

4.2.2. Detection Performance for Planar Targets

As the detection maps of the Gulfport dataset show in Figure 2e, it can be seen that only the larger aircraft is significant in the detection results of CRNN, and the other two smaller ones are difficult to observe directly but still present higher values than the background. However, in the result maps of traditional detectors of GRX, CRD, EAS_CR, and LSUNRS, the saliency of targets is poor, and obvious false alarm points occur. However, in GAE, HADGAN, and LREN, the networks only learn the texture features of the hyperspectral image, causing a poor detection result. Targets in the detection maps of RCRD are obvious, and the ROC curves  ( P d , P f )  of CRNN and RCRD are close, but CRNN is better at background suppression. The detection results of the SanDiego dataset are shown in Figure 2f. The detection map of CRNN presents clear targets and perfect background suppression, but the anomaly scores vary greatly between the three airplanes. In the detection results of GRX, CRD, EAS_CR, LSUNRS, and HADGAN, the targets are lost in the background and obvious noise, causing some false alarms. Although some pixels of the three anomalies have higher brightness in the result of GAE, LREN, and RCRD, the background suppression is not so good that the roof on the lower right turns to be a false alarm target.
For planar targets on the Gulfport and SanDiego datasets, the ROC curves of  ( P d , P f ) , and  ( P f , τ )  are all the closest to the upper left corner and the lower left corner, respectively. It can be concluded that CRNN also presents a perfect detection accuracy with the lowest false alarm rate for planar targets such as airplanes. Although the target texture is not obvious in the detection maps of CRNN, the outline of the aircraft can be clearly seen after performing the logarithmic transformation on the detection maps as show in Figure 4, which proves that the anomalies and the background are still separable. The logarithmic transformation is as follows
R = log v + 1 ( 1 + v · I )
where  ( v + 1 )  is the base of the logarithm. I and R denote the input and output images, respectively, and the values are all in the range of [0, 1].

4.2.3. AUC Performance

According to the comparison of the average AUC values in Table 3, CRNN has the highest  AUC ( D , F )  value of 0.9982, which is 6.79% higher than that of the traditional CRD, indicating that CRNN has perfect detection efficiency and finds the balance between improving the detection rate and reducing the false alarm rate. At the same time, CRNN also shows the smallest  AUC ( F , τ )  score, and the largest  AUC JBS  and  AUC SNPR , which are  7.47 × 10 6 , 1.9982,  1.03 × 10 7 , respectively. It indicates that CRNN has a strong ability to suppress the background. However, the performance of CRNN in  AUC ( D , τ ) AUC JAD AUC ADBS , and  AUC OADP  needs to be improved, and the values of the latter three are mainly related to the former. This may be because the high efficiency of background suppression in CRNN also has an impact on some anomalous pixels, and the detection rate needs to rely on a lower threshold, causing the poor performance on  ( P d , τ )  curves and  AUC ( D , τ ) . However, it does not mean that the background and anomalies are inseparable. The average  AUC SNPR  of the detection maps reaches seven orders of magnitude, and the anomaly scores of background pixels is concentrated around  10 11  to  10 7 . The order of magnitude of the potential anomalous pixels spans a large range that the anomaly scores of some pixels are quite close to 1, while other pixels may reach about  10 4 . As a supplementary illustration, since the orders of magnitude of the background and partial anomalies are small, we tried logarithmic transformation on the result maps of CRNN as the image enhancement strategy to re-evaluate the AUC performance. The logarithmic transformation is in shown in Equation (25) and v is set to  10 2  in all datasets. This version is not included in the comparison of other methods in Table 3 because it is only used as the post-processing method. According to the results of AUC, logarithmic transformation searches a balance between  AUC ( D , τ )  and  AUC ( F , τ ) , and improves the target detectability.
In addition, we repeated 10 experiments for each dataset to measure the uncertainty of network training. The mean and variance of the three AUC values of  AUC ( D , F ) AUC ( D , τ ) , and  AUC ( F , τ )  are shown in Table 4, which verifies the credibility of the experimental results for our proposed model.

4.3. Computing Time

In addition to the accuracy of anomaly detection, computing time is also an important part of algorithm efficiency. The computing time of the proposed CRNN and the compared eight algorithms is listed Table 5. For deep learning-based methods, the training epochs of CRNN, GAE, and HADGAN are all set to 500. We counted their training time on GPU and inference time tested on CPU. While LREN has no module for testing, so we only recorded the entire computing time including network training on GPU. The training epochs of LREN are 1000 according to [49]. For non-deep learning methods, LSUNRS, RCRD, EAS_CR, CRD, and RX, were implemented on MATLAB R2019b.
The comparison of computing times in Table 5 are just for reference due to the differences in the implementation of these methods, but it still shows the high efficiency of our proposed CRNN. Compared with the three deep learning-based method, the inference time of CRNN is the shortest, mainly because of the simple linear computation of matrices in a few network layers. GAE calculates the guided images through a spectral similarity method that takes most of the computation time. HADGAN directly used the traditional detector on the residual image and still retains the computational cost of traditional algorithms. LREN needs to consume part of computing time in solving the optimization of LRR.

5. Discussion

In this section, discussions of the ablation study and parameter analysis are made.

5.1. Ablation Study

To verify the effectiveness of the two-stream CR learning networks, including the fusion strategy and the joint training strategy in the proposed CRNN, we set a series of ablation experiments for research. In this experiment, the following scenarios were compared.
  • CRNN: the proposed CRNN model, whose detection result is  A = A g A l ;
  • CRNN+: the CRNN model in additive fusion version, whose detection result is  A + = A g + A l ;
  • LCRNN: the local stream of CR networks in CRNN, whose detection result is  A l ;
  • GCRNN: the global stream of CR networks in CRNN, whose detection result is  A g ;
  • CRNN_se: the local stream and global stream are separately trained and fused to obtain the detection result. It is a separate training version for CRNN;
  • LCRNN_se: the local stream of CR networks in CRNN_se;
  • GCRNN_se: the global stream of CR networks in CRNN_se.
Table 6 lists the average of the eight AUC values for the ablation study. In addition, the detailed eight AUC values on each dataset are presented in the Appendix A. It is suggested that either the local CR networks or the global CR networks have outstanding detection performance. Their average  AUC ( D , F )  values are all above 0.9940, which are obviously better than the other eight comparison algorithms. Compared with other six scenarios of CRNN+, LCRNN, GCRNN, CRNN_se, LCRNN_se, and GCRNN_se, the average  AUC ( D , F )  scores of CRNN are, respectively, improved by 0.06%, 0.08%, 0.13%, 0.03%, 0.40%, and 0.10%.
(1)
Global CR and Local CR
The anomaly detection maps for six datasets of LCRNN and GCRNN are shown in Figure 5. Compared with the global CR networks, the detection results of the local CR networks present more false alarm targets and worse suppression on the background. This may be, mainly, because the local CR networks are composed of the local background dictionary, which is more likely to highlight specific areas in the background, such as the white area in HYDICE, the runway in Gulfport, and the roof in SanDiego. In addition, the anomaly scores of different targets in the results of local CR networks are relatively close, while the anomaly scores of different targets vary greatly in the results of global CR networks.
(2)
Detection Fusion of Global CR and Local CR
From the quantitative perspective, CRNN obtained the best  AUC ( D , F )  values when compared to LCRNN and GCRNN. The results of separate training versions also presents the similar findings that the average  AUC ( D , F )  values of CRNN_se have, respectively, improved by 0.36% and 0.06% compared with LCRNN_se and GCRNN_se. This demonstrates that the detection results fused with the local and global ones is better than that of the two alone. Similar to the random forest theory [75], it is possible to create a stronger and better detector by comprehensively analyzing the results of several detectors with good detection performance. In addition, the multiplicative fusion strategy used in proposed CRNN is better than additive fusion in CRNN+. As shown in Figure 5, the background suppression in the result maps of LCRNN is not as good as that of GCRNN, so multiplicative fusion can offer the advantages of global result maps and better background suppression compared with additive fusion.
(3)
Joint Training and Separate Training
According to the average  AUC ( D , F )  value, GCRNN is 0.04% smaller compared with GCRNNse, while the score of LCRNN is 0.3% higher than LCRNNse. For the final fusion result, the  AUC ( D , F )  score of CRNN is improved by 0.03% compared with the separate training version. Therefore, the joint training of the local and global streams of the networks is better than separate training in general. In our opinion, this may be because of the comprehensive usage of the global and local streams with a better balance in joint training. It optimizes each part in the networks through the objective function of  L C R , and extracts features more conducive to detection performance.

5.2. Parameter Analysis

Under the hyperparameters configured in Section 4.1.4, the performance of CRNN is mainly related to two key parameters, C as the amount of the global background dictionary atoms and S as the amount of the local background dictionary atoms, where S is mainly determined by the scale of the dual windows.
Since the size of C affects the ability of the global dictionary to generalize the background of the entire image to a certain extent, it is vital to set C at the appropriate value to ensure the comprehensiveness of the background and avoid anomalous pollution. The parameter C is analyzed in the range of {5, 10, 15, 20}. In Figure 6, it is shown by multiple experiments that the anomaly detection performance reaches the optimum when C is 15. In addition, the analysis of parameter C on the eight types of AUC values for the six datasets are listed in the Appendix B.
The detection performance of the local CR networks is mainly affected by the scale of inner and outer windows, which is generally correlated with the targets size. The larger the size of the windows presents a higher computational cost and a slower inference speed for the networks. In the experiments, the range of the inner window size is set in [3, 9], and the range of the outer window size is in [2 +  w i n i n , 8 +  w i n i n ] relatively. As shown in Figure 7, for the Beach, Urban, XiongAn and HYDICE datasets that only contain small targets, the detection results are not sensitive to the scale of the inner and outer windows, and the  AUC ( D , F )  values are all above 0.99. For the Gulfport and SanDiego datasets with planar targets, the  AUC ( D , F )  values remain at a suboptimal level when the scale of inner windows is much smaller than the size of anomalies, and they are significantly reduced when the inner window size is excessively wider than the anomaly. However, the selection of the optimal window size still involves a certain dynamic range. According to the experimental results, it is suggested that the range of the inner window is set in [5, 7], and the range of outer window is [9, 11] for the planar target.

6. Conclusions

In this paper, we proposed a deep learning-based hyperspectral anomaly detection method named CRNN, which implements CR theory into neural networks. CRNN mines the deep features that are helpful for the background construction and representation weight optimization in CR, which enhances the detection performance and efficiency. Firstly, an autoencoder is introduced to reconstruct the original hyperspectral data and obtain a dense expression. Based on the hidden feature map in the AE, the feature extraction network is used to further extract its comprehensive feature. Then, the two-stream global and local CR learning networks are ideally designed to learn the representation data of the hidden features under the objective function of CR. Finally, the anomaly detection result is contributed by fusing the residual maps of original and representation data in the two streams. Compared with state-of-the-art methods, experiments on six real hyperspectral images show that CRNN presents a perfect balance on detection accuracy and false alarm rate with a faster inference speed. In the future, we may try a more reasonable structure for CR learning networks instead of only applying convolutional networks to learn features, and also study a better fusion strategy to balance the global and local detection results.

Author Contributions

Conceptualization, Y.D., T.O. and J.W.; methodology, Y.D. and T.O.; software, Y.D. and T.O.; validation, Y.D., T.O. and J.W.; formal analysis, Y.D.; investigation, Y.D. and T.O.; resources, T.O. and J.W.; data curation, Y.D. and T.O.; writing—original draft preparation, Y.D.; writing—review and editing, Y.D. and J.W.; visualization, Y.D.; supervision, J.W.; project administration, J.W.; funding acquisition, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Fundamental Research Funds for the Central Universities under Grant YWF-21-JC-01.

Data Availability Statement

Publicly available datasets were analyzed in this study. The Beach, Urban, and Gulfport datasets are available at: http://xudongkang.weebly.com/ (accessed on 6 April 2023). The XiongAn dataset is available at: http://www.hrs-cas.com/a/share/shujuchanpin/2019/0501/1049.html (accessed on 6 April 2023). The HYDICE dataset is publicly available at: https://sites.google.com/site/feiyunzhuhomepage/ (accessed on 6 April 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. AUC values for the ablation study on Beach dataset. (The best is in bold).
Table A1. AUC values for the ablation study on Beach dataset. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN1.00000.2803 7.75 × 10 9 1.28032.00001.28032.2803 3.62 × 10 7
CRNN+1.00000.4354 8.35 × 10 6 1.43542.00001.43542.4354 5.21 × 10 4
LCRNN0.99990.2663 5.00 × 10 6 1.26621.99991.26632.2662 5.32 × 10 4
GCRNN1.00000.4141 7.97 × 10 6 1.41412.00001.41412.4141 5.20 × 10 4
CRNN_se1.00000.1791 2.66 × 10 8 1.17912.00001.17912.1791 6.74 × 10 6
LCRNN_se0.99840.2710 6.83 × 10 4 1.26941.99781.27032.2688 3.97 × 10 2
GCRNN_se1.00000.2968 2.72 × 10 6 1.29682.00001.29682.2968 1.09 × 10 5
Table A2. AUC values for the ablation study on Urban dataset. (The best is in bold).
Table A2. AUC values for the ablation study on Urban dataset. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN0.99730.0899 2.45 × 10 5 1.08721.99731.08992.0872 3.66 × 10 3
CRNN+0.99640.1604 1.06 × 10 4 1.15671.99631.16032.1566 1.51 × 10 3
LCRNN0.99600.2255 1.22 × 10 4 1.22151.99581.22542.2214 1.85 × 10 3
GCRNN0.99610.0640 7.01 × 10 5 1.06021.99611.06402.0601 9.14 × 10 2
CRNN_se0.99790.0504 1.30 × 10 6 1.04831.99791.05042.0483 3.88 × 10 4
LCRNN_se0.99350.2495 2.09 × 10 3 1.24301.99141.24742.2410 1.20 × 10 2
GCRNN_se0.99640.0828 1.87 × 10 5 1.07921.99641.08272.0791 4.43 × 10 3
Table A3. AUC values for the ablation study on XiongAn dataset. (The best is in bold).
Table A3. AUC values for the ablation study on XiongAn dataset. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN1.00000.2718 7.48 × 10 6 1.27182.00001.27182.2718 3.63 × 10 4
CRNN+1.00000.5173 5.70 × 10 3 1.51731.99431.51162.511690.7
LCRNN0.99990.4930 8.02 × 10 3 1.49291.99191.48502.484961.5
GCRNN0.99980.2706 1.38 × 10 4 1.27041.99971.27042.2703 1.96 × 10 3
CRNN_se1.00000.3773 4.89 × 10 5 1.37732.00001.37732.3773 7.71 × 10 3
LCRNN_se0.99990.5628 6.30 × 10 3 1.56271.99361.55652.556589.4
GCRNN_se1.00000.4389 3.07 × 10 3 1.43891.99691.43592.4359 1.43 × 10 2
Table A4. AUC values for the ablation study on HYDICE dataset. (The best is in bold).
Table A4. AUC values for the ablation study on HYDICE dataset. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN0.99990.1570 9.75 × 10 7 1.15701.99991.15702.1570 1.61 × 10 5
CRNN+0.99900.3518 1.05 × 10 3 1.35081.99791.35082.3498 3.36 × 10 2
LCRNN0.99980.3038 4.51 × 10 4 1.30371.99941.30342.3032 6.74 × 10 2
GCRNN0.99810.2572 1.19 × 10 3 1.25531.99691.25602.2541 2.15 × 10 2
CRNN_se0.99970.1562 1.62 × 10 6 1.15591.99971.15622.1558 9.63 × 10 4
LCRNN_se0.99480.3070 2.94 × 10 3 1.30181.99181.30412.2989 1.04 × 10 2
GCRNN_se0.99900.2150 2.41 × 10 4 1.21401.99881.21472.2138 8.92 × 10 2
Table A5. AUC values for the ablation study on Gulfport dataset. (The best is in bold).
Table A5. AUC values for the ablation study on Gulfport dataset. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN0.99550.0843 3.32 × 10 9 1.07991.99551.08432.0799 2.54 × 10 7
CRNN+0.99610.1347 2.18 × 10 5 1.13081.99611.13472.1308 6.19 × 10 3
LCRNN0.99530.0934 1.45 × 10 5 1.08871.99531.09332.0887 6.42 × 10 3
GCRNN0.99040.0956 1.64 × 10 5 1.08611.99041.09562.0861 5.81 × 10 3
CRNN_se0.99770.0445 7.68 × 10 8 1.04221.99771.04452.0422 5.79 × 10 5
LCRNN_se0.99300.2548 7.33 × 10 3 1.24781.98561.24752.240434.7
GCRNN_se0.99630.0504 3.90 × 10 6 1.04681.99631.05042.0467 1.29 × 10 4
Table A6. AUC values for the ablation study on SanDiego dataset. (The best is in bold).
Table A6. AUC values for the ablation study on SanDiego dataset. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN0.99650.0368 1.18 × 10 5 1.03331.99651.03682.0333 3.11 × 10 3
CRNN+0.99400.1284 8.51 × 10 4 1.12241.99321.12752.1216 1.51 × 10 2
LCRNN0.99260.1197 1.05 × 10 3 1.11231.99161.11862.1112 1.14 × 10 2
GCRNN0.99600.0513 7.98 × 10 5 1.04731.99591.05122.0472 6.43 × 10 2
CRNN_se0.99150.0322 1.74 × 10 4 1.02371.99131.03212.0236 1.85 × 10 2
LCRNN_se0.98550.0167 2.31 × 10 3 1.00221.98321.01441.99997.22
GCRNN_se0.99150.0634 5.38 × 10 4 1.05491.99091.06292.0544 1.18 × 10 2

Appendix B

Table A7. Analysis of parameter C on the eight AUC values. (The best of each dataset is in bold).
Table A7. Analysis of parameter C on the eight AUC values. (The best of each dataset is in bold).
DatasetC AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
Beach50.99990.1939 2.70 × 10 5 1.19381.99981.19392.1937 7.17 × 10 3
100.99990.2776 5.54 × 10 5 1.27751.99981.27762.2775 5.01 × 10 3
151.00000.2803 7.75 × 10 9 1.28032.00001.28032.2803 3.62 × 10 7
200.99980.2025 3.58 × 10 5 1.20231.99981.20252.2023 5.66 × 10 3
Urban50.99560.0668 7.81 × 10 6 1.06241.99561.06682.0624 8.55 × 10 3
100.99610.0479 2.03 × 10 5 1.04391.99601.04792.0439 2.36 × 10 3
150.99730.0899 2.45 × 10 5 1.08721.99731.08992.0872 3.66 × 10 3
200.99740.0879 7.96 × 10 6 1.08531.99741.08792.0853 1.10 × 10 4
XiongAn51.00000.2931 5.53 × 10 6 1.29312.00001.29312.2931 5.30 × 10 4
100.99990.3401 1.77 × 10 6 1.34011.99991.34012.3400 1.92 × 10 5
151.00000.2718 7.48 × 10 6 1.27182.00001.27182.2718 3.63 × 10 4
201.00000.2789 5.50 × 10 6 1.27892.00001.27892.2788 5.07 × 10 4
HYDICE50.99990.1788 3.61 × 10 7 1.17882.00001.17882.1788 4.95 × 10 5
100.99990.1873 2.22 × 10 6 1.18721.99991.18732.1872 8.43 × 10 4
150.99990.1570 9.75 × 10 7 1.15701.99991.15702.1570 1.61 × 10 5
200.99990.2152 1.34 × 10 6 1.21511.99991.21522.2151 1.60 × 10 5
Gulfport50.95210.0098 2.54 × 10 4 0.96191.95191.00951.961638.4
100.99190.0883 2.99 × 10 5 1.08021.99181.08832.0802 2.95 × 10 3
150.99550.0843 3.32 × 10 9 1.07991.99551.08432.0799 2.54 10 7
200.95340.0077 7.88 × 10 4 0.96111.95261.00691.96049.79
SanDiego50.99280.0249 4.76 × 10 4 1.01761.99231.02442.017252.3
100.98660.0012 3.81 × 10 4 0.98781.98621.00081.98743.12
150.99650.0368 1.18 × 10 5 1.03331.99651.03682.0333 3.11 × 10 3
200.99060.0143 2.85 × 10 4 1.00491.99031.01402.004650.2

References

  1. Eismann, M.T.; Stocker, A.D.; Nasrabadi, N.M. Automated Hyperspectral Cueing for Civilian Search and Rescue. Proc. IEEE 2009, 97, 1031–1055. [Google Scholar] [CrossRef]
  2. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. A new deep convolutional neural network for fast hyperspectral image classification. ISPRS J. Photogramm. Remote Sens. 2018, 145, 120–147. [Google Scholar] [CrossRef]
  3. Tuia, D.; Flamary, R.; Courty, N. Multiclass feature learning for hyperspectral image classification: Sparse and hierarchical solutions. ISPRS J. Photogramm. Remote Sens. 2015, 105, 272–285. [Google Scholar] [CrossRef] [Green Version]
  4. Su, H.; Yu, Y.; Du, Q.; Du, P. Ensemble learning for hyperspectral image classification using tangent collaborative representation. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3778–3790. [Google Scholar] [CrossRef]
  5. Nasrabadi, N.M. Hyperspectral target detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2013, 31, 34–44. [Google Scholar] [CrossRef]
  6. Axelsson, M.; Friman, O.; Haavardsholm, T.V.; Renhorn, I. Target detection in hyperspectral imagery using forward modeling and in-scene information. ISPRS J. Photogramm. Remote Sens. 2016, 119, 124–134. [Google Scholar] [CrossRef]
  7. Jiao, C.; Chen, C.; McGarvey, R.G.; Bohlman, S.; Jiao, L.; Zare, A. Multiple instance hybrid estimator for hyperspectral target characterization and sub-pixel target detection. ISPRS J. Photogramm. Remote Sens. 2018, 146, 235–250. [Google Scholar] [CrossRef] [Green Version]
  8. Chang, C.I. Hyperspectral Anomaly Detection: A Dual Theory of Hyperspectral Target Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–20. [Google Scholar] [CrossRef]
  9. Chang, C.I. Target-to-Anomaly Conversion for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–28. [Google Scholar] [CrossRef]
  10. Li, K.; Ling, Q.; Qin, Y.; Wang, Y.; Cai, Y.; Lin, Z.; An, W. Spectral-Spatial Deep Support Vector Data Description for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  11. Xie, W.; Zhang, X.; Li, Y.; Lei, J.; Li, J.; Du, Q. Weakly Supervised Low-Rank Representation for Hyperspectral Anomaly Detection. IEEE Trans. Cybern. 2021, 51, 3889–3900. [Google Scholar] [CrossRef] [PubMed]
  12. Su, H.; Wu, Z.; Zhang, H.; Du, Q. Hyperspectral anomaly detection: A survey. IEEE Geosci. Remote Sens. Mag. 2021, 10, 64–90. [Google Scholar] [CrossRef]
  13. Reed, I.S.; Yu, X. Adaptive multiple-band CFAR detection of an optical pattern with unknown spectral distribution. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 1760–1770. [Google Scholar] [CrossRef]
  14. Kwon, H.; Nasrabadi, N.M. Kernel RX-algorithm: A nonlinear anomaly detector for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 388–397. [Google Scholar] [CrossRef]
  15. Borghys, D.; Kåsen, I.; Achard, V.; Perneel, C. Comparative evaluation of hyperspectral anomaly detectors in different types of background. In Proceedings of the Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XVIII, Baltimore, MD, USA, 24 May 2012; Volume 8390, pp. 803–814. [Google Scholar]
  16. Li, W.; Du, Q. Decision fusion for dual-window-based hyperspectral anomaly detector. J. Appl. Remote Sens. 2015, 9, 097297. [Google Scholar] [CrossRef] [Green Version]
  17. Guo, Q.; Zhang, B.; Ran, Q.; Gao, L.; Li, J.; Plaza, A. Weighted-RXD and linear filter-based RXD: Improving background statistics estimation for anomaly detection in hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2351–2366. [Google Scholar] [CrossRef]
  18. Chang, C.I. Effective Anomaly Space for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–24. [Google Scholar] [CrossRef]
  19. Carlotto, M.J. A cluster-based approach for detecting human-made objects and changes in imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 374–387. [Google Scholar] [CrossRef]
  20. Hytla, P.C.; Hardie, R.C.; Eismann, M.T.; Meola, J. Anomaly detection in hyperspectral imagery: Comparison of methods using diurnal and seasonal data. J. Appl. Remote Sens. 2009, 3, 033546. [Google Scholar] [CrossRef]
  21. Banerjee, A.; Burlina, P.; Diehl, C. A support vector method for anomaly detection in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2282–2291. [Google Scholar] [CrossRef]
  22. Messinger, D.W.; Albano, J. A graph theoretic approach to anomaly detection in hyperspectral imagery. In Proceedings of the 2011 3rd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lisbon, Portugal, 6–9 June 2011; pp. 1–4. [Google Scholar]
  23. Yuan, Y.; Ma, D.; Wang, Q. Hyperspectral anomaly detection by graph pixel selection. IEEE Trans. Cybern. 2015, 46, 3123–3134. [Google Scholar] [CrossRef] [PubMed]
  24. Tao, R.; Zhao, X.; Li, W.; Li, H.C.; Du, Q. Hyperspectral Anomaly Detection by Fractional Fourier Entropy. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4920–4929. [Google Scholar] [CrossRef]
  25. Kang, X.; Zhang, X.; Li, S.; Li, K.; Li, J.; Benediktsson, J.A. Hyperspectral anomaly detection with attribute and edge-preserving filters. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5600–5611. [Google Scholar] [CrossRef]
  26. Xie, W.; Jiang, T.; Li, Y.; Jia, X.; Lei, J. Structure Tensor and Guided Filtering-Based Algorithm for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4218–4230. [Google Scholar] [CrossRef]
  27. Cheng, X.; Wen, M.; Gao, C.; Wang, Y. Hyperspectral Anomaly Detection Based on Wasserstein Distance and Spatial Filtering. Remote Sens. 2022, 14, 2730. [Google Scholar] [CrossRef]
  28. Shang, W.; Jouni, M.; Wu, Z.; Xu, Y.; Dalla Mura, M.; Wei, Z. Hyperspectral Anomaly Detection Based on Regularized Background Abundance Tensor Decomposition. Remote Sens. 2023, 15, 1679. [Google Scholar] [CrossRef]
  29. Li, L.; Li, W.; Qu, Y.; Zhao, C.; Tao, R.; Du, Q. Prior-Based Tensor Approximation for Anomaly Detection in Hyperspectral Imagery. IEEE Trans. Neural Networks Learn. Syst. 2022, 33, 1037–1050. [Google Scholar] [CrossRef]
  30. Chang, C.I.; Lin, C.Y.; Chung, P.C.; Hu, P.F. Iterative Spectral–Spatial Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–30. [Google Scholar] [CrossRef]
  31. Li, J.; Zhang, H.; Zhang, L.; Ma, L. Hyperspectral anomaly detection by the use of background joint sparse representation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2523–2533. [Google Scholar] [CrossRef]
  32. Li, F.; Zhang, X.; Zhang, L.; Jiang, D.; Zhang, Y. Exploiting structured sparsity for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4050–4064. [Google Scholar] [CrossRef]
  33. Soofbaf, S.R.; Sahebi, M.R.; Mojaradi, B. A sliding window-based joint sparse representation (swjsr) method for hyperspectral anomaly detection. Remote Sens. 2018, 10, 434. [Google Scholar] [CrossRef] [Green Version]
  34. Li, W.; Du, Q. Collaborative representation for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2014, 53, 1463–1474. [Google Scholar] [CrossRef]
  35. Vafadar, M.; Ghassemian, H. Hyperspectral anomaly detection using outlier removal from collaborative representation. In Proceedings of the 2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA), Shahrekord, Iran, 19–20 April 2017; pp. 13–19. [Google Scholar]
  36. Su, H.; Wu, Z.; Du, Q.; Du, P. Hyperspectral anomaly detection using collaborative representation with outlier removal. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 5029–5038. [Google Scholar] [CrossRef]
  37. Tu, B.; Li, N.; Liao, Z.; Ou, X.; Zhang, G. Hyperspectral anomaly detection via spatial density background purification. Remote Sens. 2019, 11, 2618. [Google Scholar] [CrossRef] [Green Version]
  38. Tan, K.; Hou, Z.; Wu, F.; Du, Q.; Chen, Y. Anomaly detection for hyperspectral imagery based on the regularized subspace method and collaborative representation. Remote Sens. 2019, 11, 1318. [Google Scholar] [CrossRef] [Green Version]
  39. Zhao, C.; Li, C.; Feng, S.; Su, N.; Li, W. A Spectral–Spatial Anomaly Target Detection Method Based on Fractional Fourier Transform and Saliency Weighted Collaborative Representation for Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5982–5997. [Google Scholar] [CrossRef]
  40. Wu, Z.; Su, H.; Tao, X.; Han, L.; Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Hyperspectral Anomaly Detection with Relaxed Collaborative Representation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  41. Sun, W.; Liu, C.; Li, J.; Lai, Y.M.; Li, W. Low-rank and sparse matrix decomposition-based anomaly detection for hyperspectral imagery. J. Appl. Remote Sens. 2014, 8, 083641. [Google Scholar] [CrossRef]
  42. Xu, Y.; Wu, Z.; Li, J.; Plaza, A.; Wei, Z. Anomaly detection in hyperspectral images based on low-rank and sparse representation. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1990–2000. [Google Scholar] [CrossRef]
  43. Zhang, Y.; Du, B.; Zhang, L.; Wang, S. A low-rank and sparse matrix decomposition-based Mahalanobis distance method for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1376–1389. [Google Scholar] [CrossRef]
  44. Cheng, T.; Wang, B. Graph and Total Variation Regularized Low-Rank Representation for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 391–406. [Google Scholar] [CrossRef]
  45. Li, L.; Li, W.; Du, Q.; Tao, R. Low-rank and sparse decomposition with mixture of Gaussian for hyperspectral anomaly detection. IEEE Trans. Cybern. 2020, 51, 4363–4372. [Google Scholar] [CrossRef] [PubMed]
  46. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  47. Wu, F.; Yang, X.H.; Packard, A.; Becker, G. Induced L2-norm control for LPV systems with bounded parameter variation rates. Int. J. Robust Nonlinear Control 1996, 6, 983–998. [Google Scholar] [CrossRef]
  48. Zhou, T.; Tao, D. Godec: Randomized low-rank & sparse matrix decomposition in noisy case. In Proceedings of the Proceedings of the 28th International Conference on Machine Learning, ICML 2011, Bellevue, WA, USA, 28 June–2 July 2011. [Google Scholar]
  49. Jiang, K.; Xie, W.; Lei, J.; Jiang, T.; Li, Y. LREN: Low-rank embedded network for sample-free hyperspectral anomaly detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 4139–4146. [Google Scholar]
  50. Hu, X.; Xie, C.; Fan, Z.; Duan, Q.; Zhang, D.; Jiang, L.; Wei, X.; Hong, D.; Li, G.; Zeng, X.; et al. Hyperspectral anomaly detection using deep learning: A review. Remote Sens. 2022, 14, 1973. [Google Scholar] [CrossRef]
  51. Lei, J.; Xie, W.; Yang, J.; Li, Y.; Chang, C.I. Spectral–spatial feature extraction for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8131–8143. [Google Scholar] [CrossRef]
  52. Lu, X.; Zhang, W.; Huang, J. Exploiting Embedding Manifold of Autoencoders for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1527–1537. [Google Scholar] [CrossRef]
  53. Arisoy, S.; Nasrabadi, N.M.; Kayabol, K. GAN-based hyperspectral anomaly detection. In Proceedings of the 2020 28th European Signal Processing Conference (EUSIPCO), Amsterdam, The Netherlands, 18–21 January 2021; pp. 1891–1895. [Google Scholar]
  54. Jiang, T.; Li, Y.; Xie, W.; Du, Q. Discriminative reconstruction constrained generative adversarial network for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4666–4679. [Google Scholar] [CrossRef]
  55. Xie, W.; Liu, B.; Li, Y.; Lei, J.; Chang, C.I.; He, G. Spectral adversarial feature learning for anomaly detection in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2019, 58, 2352–2365. [Google Scholar] [CrossRef]
  56. Wang, S.; Wang, X.; Zhang, L.; Zhong, Y. Auto-AD: Autonomous Hyperspectral Anomaly Detection Network Based on Fully Convolutional Autoencoder. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  57. Xie, W.; Fan, S.; Qu, J.; Wu, X.; Lu, Y.; Du, Q. Spectral Distribution-Aware Estimation Network for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  58. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Hinton, G.E.; Zemel, R. Autoencoders, minimum description length and Helmholtz free energy. Adv. Neural Inf. Process. Syst. 1994, 6, 3–30. [Google Scholar]
  60. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  61. Xiang, P.; Ali, S.; Jung, S.K.; Zhou, H. Hyperspectral Anomaly Detection with Guided Autoencoder. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  62. Makhzani, A.; Shlens, J.; Jaitly, N.; Goodfellow, I.; Frey, B. Adversarial autoencoders. arXiv 2015, arXiv:1511.05644. [Google Scholar]
  63. Xie, W.; Liu, B.; Li, Y.; Lei, J.; Du, Q. Autoencoder and Adversarial-Learning-Based Semisupervised Background Estimation for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5416–5427. [Google Scholar] [CrossRef]
  64. Li, Y.; Jiang, T.; Xie, W.; Lei, J.; Du, Q. Sparse Coding-Inspired GAN for Hyperspectral Anomaly Detection in Weakly Supervised Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  65. Huber, P.J. Robust estimation of a location parameter. In Breakthroughs in Statistics; Springer: Berlin, Germany, 1992; pp. 492–518. [Google Scholar]
  66. Wu, Y.; He, K. Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  67. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the ICML, Atlanta, GA, USA, 16–21 June 2013; Volume 30, p. 3. [Google Scholar]
  68. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  69. Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 16–18 April 2019; pp. 7354–7363. [Google Scholar]
  70. Jang, E.; Gu, S.; Poole, B. Categorical reparameterization with gumbel-softmax. arXiv 2016, arXiv:1611.01144. [Google Scholar]
  71. Senling, C.Y.L.X.Y.W.; Peng, Z. Aerial hyperspectral remote sensing classification dataset of Xiongan New Area (Matiwan Village). J. Remote Sens. 2020, 24, 1299–1306. (In Chinese) [Google Scholar] [CrossRef]
  72. Zhu, F.; Wang, Y.; Xiang, S.; Fan, B.; Pan, C. Structured sparse method for hyperspectral unmixing. ISPRS J. Photogramm. Remote Sens. 2014, 88, 101–118. [Google Scholar] [CrossRef] [Green Version]
  73. Chang, C.I. An Effective Evaluation Tool for Hyperspectral Target Detection: 3D Receiver Operating Characteristic Curve Analysis. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5131–5153. [Google Scholar] [CrossRef]
  74. Chang, C.I. Comprehensive Analysis of Receiver Operating Characteristic (ROC) Curves for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–24. [Google Scholar] [CrossRef]
  75. Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 2016, 114, 24–31. [Google Scholar] [CrossRef]
Figure 1. An overview of the proposed CRNN. It consists of four parts: AE, comprehensive feature extraction, two-stream CR learning network, and detection fusion.
Figure 1. An overview of the proposed CRNN. It consists of four parts: AE, comprehensive feature extraction, two-stream CR learning network, and detection fusion.
Remotesensing 15 03357 g001
Figure 2. Pseudo-color images, references, and detection maps of the proposed CRNN and compared algorithms on six datasets. (a) Beach. (b) Urban. (c) XiongAn. (d) HYDICE. (e) Gulfport. (f) SanDiego.
Figure 2. Pseudo-color images, references, and detection maps of the proposed CRNN and compared algorithms on six datasets. (a) Beach. (b) Urban. (c) XiongAn. (d) HYDICE. (e) Gulfport. (f) SanDiego.
Remotesensing 15 03357 g002
Figure 3. 3D ROC curves and the generated three 2D ROC curves of the proposed CRNN and compared algorithms on six datasets. (a) Beach. (b) Urban. (c) XiongAn. (d) HYDICE. (e) Gulfport. (f) SanDiego.
Figure 3. 3D ROC curves and the generated three 2D ROC curves of the proposed CRNN and compared algorithms on six datasets. (a) Beach. (b) Urban. (c) XiongAn. (d) HYDICE. (e) Gulfport. (f) SanDiego.
Remotesensing 15 03357 g003
Figure 4. Detection maps after the logarithmic transformation. The red boxes are manually added to mark the targets. (a) Gulfport ( v = 10 7 ). (b) SanDiego ( v = 10 4 ).
Figure 4. Detection maps after the logarithmic transformation. The red boxes are manually added to mark the targets. (a) Gulfport ( v = 10 7 ). (b) SanDiego ( v = 10 4 ).
Remotesensing 15 03357 g004
Figure 5. Detection maps of LCRNN and GCRNN. (a) Beach. (b) Urban. (c) XiongAn. (d) HYDICE. (e) Gulfport. (f) SanDiego.
Figure 5. Detection maps of LCRNN and GCRNN. (a) Beach. (b) Urban. (c) XiongAn. (d) HYDICE. (e) Gulfport. (f) SanDiego.
Remotesensing 15 03357 g005
Figure 6. Analysis of parameter C on  AUC ( D , F )  value.
Figure 6. Analysis of parameter C on  AUC ( D , F )  value.
Remotesensing 15 03357 g006
Figure 7. Analysis of parameters  w i n i n  and  w i n o u t  over  AUC ( D , F )  value on six datasets.
Figure 7. Analysis of parameters  w i n i n  and  w i n o u t  over  AUC ( D , F )  value on six datasets.
Remotesensing 15 03357 g007
Table 1. Details of the six hyperspectral datasets in the experiment.
Table 1. Details of the six hyperspectral datasets in the experiment.
DatasetSensorResolutionSpatial SizeBandsAnomaly TypeAnomaly ProportionAnomaly Size
BeachAVIRIS 14.4 m100 × 100188boat0.61%11
UrbanAVIRIS 117.2 m100 × 100204buildings0.67%2–14
HYDICEHYDICE 21 m80 × 100162vehicles0.24%1–4
XiongAnGaofeng 30.5 m100 × 100250vehicle0.10%10
GulfportAVIRIS 13.4 m100 × 1001883 airplanes0.60%10–40
SanDiegoAVIRIS 13.5 m100 × 1001883 airplanes0.34%30–50
1 Airborne Visible Infrared Imaging Spectrometer. 2 Hyperspectral Digital Imagery Collection Experiment. 3 Gaofeng Full Spectrum Multi-modal Imaging Spectrometer.
Table 2. AUC values of the proposed CRNN and compared algorithms on six datasets. (The best of each dataset is in bold).
Table 2. AUC values of the proposed CRNN and compared algorithms on six datasets. (The best of each dataset is in bold).
DatasetMethod AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
BeachCRNN1.00000.2803 7.75 × 10 9 1.28032.00001.28032.2803 3.62 × 10 7
CRNN_t 11.00000.77470.00141.77471.99861.77332.7733553.6985
GAE [61]0.97730.35420.01761.33151.95961.33662.313920.0874
HADGAN [54]1.00000.65980.04481.65971.95521.61502.615014.7397
LREN [49]0.84510.40710.16621.25211.67881.24082.08592.4489
LSUNRS [38]0.99980.44140.00921.44131.99071.43232.432148.1575
RCRD [40]0.99960.40360.01141.40321.98821.39222.391835.3348
EAS_CR [18]0.99860.50200.01491.50051.98371.48712.485733.7063
CRD [34]0.99440.53510.04691.52951.94741.48822.482511.4035
GRX [13]0.99980.53140.02591.53131.97391.50552.505320.4912
UrbanCRNN0.99730.0899 2.45 × 10 5 1.08721.99731.08992.0872 3.66 × 10 3
CRNN_t 10.99730.68960.02771.68691.96961.66192.659224.8633
GAE [61]0.96220.36950.06661.33171.89561.30292.26515.5496
HADGAN [54]0.95840.24830.07631.20671.88221.17202.13043.2555
LREN [49]0.93990.48380.14061.42371.79931.34322.28313.4411
LSUNRS [38]0.99630.13520.01051.13151.98581.12472.121012.8765
RCRD [40]0.99370.31620.01601.30991.97771.30032.293919.8251
EAS_CR [18]0.99110.30820.04961.29931.94141.25862.24966.2098
CRD [34]0.96550.08480.01271.05031.95281.07212.03766.6565
GRX [13]0.99070.31430.05561.30491.93511.25872.24945.6570
XiongAnCRNN1.00000.2718 7.48 × 10 6 1.27182.00001.27182.2718 3.63 × 10 4
CRNN_t 11.00000.87080.24601.87081.75401.62472.62473.5392
GAE [61]0.99990.40140.00791.40131.99201.39352.393450.6621
HADGAN [54]1.00000.46340.07181.46341.92821.39172.39166.4562
LREN [49]1.00000.77510.16411.77511.83591.61092.61094.7220
LSUNRS [38]0.99960.42960.04991.42921.94971.37972.37938.6172
RCRD [40]0.99990.57100.02091.57091.97911.55012.550127.3467
EAS_CR [18]0.98810.19090.01051.17901.97761.18042.168518.1455
CRD [34]0.97830.44130.13151.41961.84691.30982.28813.3569
GRX [13]0.97650.42220.10001.39861.87651.32222.29874.2235
HYDICECRNN0.99990.1570 9.75 × 10 7 1.15701.99991.15702.1570 1.61 × 10 5
CRNN_t 10.99990.81320.02101.81311.97901.79232.792238.8089
GAE [61]0.99820.39730.00991.39551.98831.38742.385640.1662
HADGAN [54]0.99030.50240.05711.49271.93321.44532.43568.7944
LREN [49]0.97840.42710.00931.40551.96911.41792.396346.1746
LSUNRS [38]0.99290.34490.01661.33781.97631.32832.321220.7202
RCRD [40]0.99620.33800.00591.33421.99031.33222.328457.7771
EAS_CR [18]0.92330.05540.00370.97871.91961.05171.975115.0686
CRD [34]0.98680.41170.03511.39851.95171.37652.363411.7149
GRX [13]0.97630.21840.03801.19471.93821.18032.15665.7406
GulfportCRNN0.99550.0843 3.32 × 10 9 1.07991.99551.08432.0799 2.54 × 10 7
CRNN_t 10.99550.48250.01051.47801.98511.47202.467646.1582
GAE [61]0.96900.26750.03141.23641.93761.23612.20518.5239
HADGAN [54]0.96020.33800.14961.29821.81061.18842.14862.2588
LREN [49]0.75210.31040.17391.06261.57821.13651.88871.7851
LSUNRS [38]0.95890.04550.01391.00441.94501.03161.99053.2769
RCRD [40]0.98990.25520.01601.24511.97391.23922.229115.9448
EAS_CR [18]0.57140.00590.00150.57731.56981.00441.57583.8782
CRD [34]0.77620.01120.00690.78741.76931.00431.78051.6246
GRX [13]0.95260.07360.02481.02621.92781.04892.00152.9743
SanDiegoCRNN0.99650.0368 1.18 × 10 5 1.03331.99651.03682.0333 3.11 × 10 3
CRNN_t 10.99650.54810.01781.54461.97871.53032.526830.7612
GAE [61]0.99100.23800.00791.22901.98311.23012.221130.1127
HADGAN [54]0.96440.28980.06531.25421.89911.22442.18884.4353
LREN [49]0.43870.10410.03460.54281.40411.06951.50823.0098
LSUNRS [38]0.97960.27560.07891.25521.90071.19672.17633.4914
RCRD [40]0.98250.15750.00841.13991.97401.14902.131518.6983
EAS_CR [18]0.75150.04640.01490.79791.73661.03161.78313.1202
CRD [34]0.90710.19930.12721.10641.77991.07211.97921.5671
GRX [13]0.94030.17780.05891.11811.88141.11892.05923.0176
1 A logarithmic transformation was performed on the result maps of CRNN, and this method is only used as a supplementary illustration and does not participate in the comparison of other methods.
Table 3. Average AUC values of the proposed CRNN and compared algorithms over six datasets. (The best is in bold).
Table 3. Average AUC values of the proposed CRNN and compared algorithms over six datasets. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN0.99820.1534 7.47 × 10 6 1.15161.99821.15342.1516 1.03 × 10 7
CRNN_t 10.99820.69650.05411.69471.94411.64242.6406116.3049
GAE [61]0.98290.33800.02361.32091.95941.31442.297425.8503
HADGAN [54]0.97890.41690.07751.39581.90141.33952.31836.6567
LREN [49]0.82570.41790.11481.24361.71091.30322.128910.2636
LSUNRS [38]0.98790.27870.02981.26661.95801.24892.236716.1900
RCRD [40]0.99360.34030.01311.33391.98051.32722.320829.1545
EAS_CR [18]0.87060.18480.01591.05551.85481.16902.039613.3548
CRD [34]0.93470.28060.06011.21531.87471.22052.15526.0539
GRX [13]0.97270.28960.05051.26231.92221.23912.21187.0174
1 A logarithmic transformation was performed on the result maps of CRNN, and this method is only used as a supplementary illustration and does not participate in the comparison of other methods.
Table 4. AUC ( D , F )   AUC ( D , τ )  and  AUC ( F , τ )  values of CRNN in repeated experiments for each dataset.
Table 4. AUC ( D , F )   AUC ( D , τ )  and  AUC ( F , τ )  values of CRNN in repeated experiments for each dataset.
  AUC ( D , F ) AUC ( D , τ ) AUC ( F , τ )
 MeanVariance(%)MeanVariance(%)MeanVariance(%)
Beach0.999500.048490.221514.75515 4.25 × 10 5 0.00432
Urban0.996420.108410.071291.24275 1.51 × 10 5 0.00110
XiongAn0.999940.006110.25614.02692 1.90 × 10 5 0.00242
HYDICE0.999880.009530.187993.54861 4.80 × 10 6 0.00070
Gulfport0.985641.369940.082482.59059 2.52 × 10 4 0.02073
SanDiego0.989540.430000.025731.11803 2.70 × 10 4 0.01517
Table 5. Computing time in second of the proposed CRNN and compared algorithms. (The content in brackets in columns 2–4 represents the inference time of the network at test).
Table 5. Computing time in second of the proposed CRNN and compared algorithms. (The content in brackets in columns 2–4 represents the inference time of the network at test).
CRNNGAE [61]HADGAN [54]LREN [49]LSUNRS [38]RCRD [40]EAS_CR [18]CRD [34]GRX [13]
Beach49.8535
(0.0279)
47.7064
(6.4304)
674.0821
(1.8471)
100.224818.782938.25932.59181.53950.0854
Urban60.6503
(0.0259)
48.6605
(6.4776)
677.5559
(2.0005)
109.084419.354740.29363.19431.61190.0896
XiongAn21.0506
(0.0558)
56.7479
(7.9990)
673.0219
(2.6153)
102.559730.990651.70063.30398.83120.1826
HYDICE19.2016
(0.0189)
26.0721
(5.2584)
693.5843
(1.6805)
73.059614.559925.51561.66841.20250.0669
Gulfport70.5033
(0.0488)
47.8685
(6.4286)
676.432
(1.8710)
100.803418.857539.40052.65211.53770.0807
SanDiego31.5995
(0.0229)
48.6626
(6.7521)
676.4849
(1.8861)
101.365719.031736.67092.12321.55730.0840
Table 6. Average AUC values over six datasets for the ablation study. (The best is in bold).
Table 6. Average AUC values over six datasets for the ablation study. (The best is in bold).
AUC ( D , F )
AUC ( D , τ )
AUC ( F , τ )
AUC JAD
AUC JBS
AUC ADBS
AUC OADP
AUC SNPR
CRNN0.99820.1534 7.47 × 10 6 1.15161.99821.15342.1516 1.03 × 10 7
CRNN+0.99760.2880 1.29 × 10 3 1.28561.99631.28672.2843 1.01 × 10 4
LCRNN0.99730.2503 1.61 × 10 3 1.24751.99561.24872.2459 1.04 × 10 4
GCRNN0.99680.1921 2.51 × 10 4 1.18891.99651.19192.1887 1.03 × 10 4
CRNN_se0.99780.1400 3.77 × 10 5 1.13771.99771.13992.1377 1.24 × 10 6
LCRNN_se0.99420.2770 3.61 × 10 3 1.27121.99061.27342.2676 1.25 × 10 2
GCRNN_se0.99720.1912 6.45 × 10 4 1.18841.99661.19062.1878 2.13 × 10 4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Duan, Y.; Ouyang, T.; Wang, J. CRNN: Collaborative Representation Neural Networks for Hyperspectral Anomaly Detection. Remote Sens. 2023, 15, 3357. https://doi.org/10.3390/rs15133357

AMA Style

Duan Y, Ouyang T, Wang J. CRNN: Collaborative Representation Neural Networks for Hyperspectral Anomaly Detection. Remote Sensing. 2023; 15(13):3357. https://doi.org/10.3390/rs15133357

Chicago/Turabian Style

Duan, Yuxiao, Tongbin Ouyang, and Jinshen Wang. 2023. "CRNN: Collaborative Representation Neural Networks for Hyperspectral Anomaly Detection" Remote Sensing 15, no. 13: 3357. https://doi.org/10.3390/rs15133357

APA Style

Duan, Y., Ouyang, T., & Wang, J. (2023). CRNN: Collaborative Representation Neural Networks for Hyperspectral Anomaly Detection. Remote Sensing, 15(13), 3357. https://doi.org/10.3390/rs15133357

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop