Next Article in Journal
A Block-Scale FFT Filter Based on Spatial Autocorrelation Features of Speckle Noise in SAR Image
Next Article in Special Issue
Double Deep Q-Network for Hyperspectral Image Band Selection in Land Cover Classification Applications
Previous Article in Journal
Integrated Geophysical Imaging and Remote Sensing for Enhancing Geological Interpretation of Landslides with Uncertainty Estimation—A Case Study from Cisiec, Poland
Previous Article in Special Issue
The Influence of Image Degradation on Hyperspectral Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parallel Spectral–Spatial Attention Network with Feature Redistribution Loss for Hyperspectral Change Detection

1
Aerospace Information Research Institute, Chinese Academy of Sciences, No. 20 Datun Road, Chaoyang District, Beijing 100101, China
2
University of Chinese Academy of Sciences, No. 3 Datun Road, Chaoyang District, Beijing 100101, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(1), 246; https://doi.org/10.3390/rs15010246
Submission received: 3 December 2022 / Revised: 26 December 2022 / Accepted: 28 December 2022 / Published: 31 December 2022
(This article belongs to the Special Issue Advances in Hyperspectral Remote Sensing Image Processing)

Abstract

:
Change detection methods using hyperspectral remote sensing can precisely identify differences of the same area at different observing times. However, due to massive spectral bands, current change detection methods are vulnerable to unrelatedspectral and spatial information in hyperspectral images with the stagewise calculation of attention maps. Besides, current change methods arrange hidden change features in a random distribution form, which cannot express a class-oriented discrimination in advance. Moreover, existent deep change methods have not fully considered the hierarchical features’ reuse and the fusion of the encoder–decoder framework. To better handle the mentioned existent problems, the parallel spectral–spatial attention network with feature redistribution loss (TFR-PS 2 ANet) is proposed. The contributions of this article are summarized as follows: (1) a parallel spectral–spatial attention module ( PS 2 A ) is introduced to enhance relevant information and suppress irrelevant information in parallel using spectral and spatial attention maps extracted from the original hyperspectral image patches; (2) the feature redistribution loss function (FRL) is introduced to construct the class-oriented feature distribution, which organizes the change features in advance and improves the discriminative abilities; (3) a two-branch encoder–decoder framework is developed to optimize the hierarchical transfer and change features’ fusion; Extensive experiments were carried out on several real datasets. The results show that the proposed PS 2 A can enhance significant information effectively and the FRL can optimize the class-oriented feature distribution. The proposed method outperforms most existent change detection methods.

1. Introduction

Hyperspectral remote sensing can capture subtle changes on the Earth’s surface because of its characteristic of high spectral resolution [1]. At present, technologies of hyperspectral change detection have been extensively applied in various domains, such as vegetation inspection, urban planning, and disaster monitoring [2,3,4,5]. Usually, change detection methods need plenty of training samples for effective and stable model results in different application scenes. The more training samples there are, the more stability the model can obtain. However, current change detection approaches are mostly designed for optical or multi-spectral images, which is a domain in which a model can be fed with sufficient training samples. This situation is in sharp contrast with the predicament of inadequate training samples in hyperspectral imaging. Furthermore, the data redundancy in both spectral and spatial information, mixed pixels caused by low spatial resolution, and the expensive change labeling for hyperspectral datasets make it difficult for multi-spectral change detection technologies to be transferred to hyperspectral images’ analysis [6].
Usually, the procedure of existent hyperspectral change detection methods can be divided into three different steps: the preprocessing of hyperspectral data, the usage of an appropriate change detection method, and the evaluation of the predicted change results:
(1) Data preprocessing: The preprocessing of hyperspectral images is the basic step for later change detection methods and greatly influences the effectiveness of the change detection results. A typical preprocessing includes registering of dual-temporal images and radiation correction to remove the effects of sunlight and the atmosphere.
(2) Change detection methods: An appropriate change detection method plays a critical role in the whole hyperspectral change detection procedure, which can directly influence the final performance indices. The selection or design of change detection methods needs to take the data structure of the preprocessed hyperspectral images into consideration, such as the remaining band number after noisy bands have been removed.
(3) Evaluation of the predicted change results: The prediction evaluation is the last step. Different evaluation indices can describe the performances of change detection methods from different viewpoints, such as the inference ability in the situation of imbalanced change types.
According to the order of appearance of change detection technologies, they can be categorized as traditional change detection methods and deep learning methods. In traditional change detection methods, based on the different procedures of the data processing, they can be further categorized as transformation-based approaches, algebra-based approaches, and independent image classification approaches. Basically, the main idea of transformation-based approaches is transforming temporal variants into another characteristic space to deal with information redundancy, but this makes it difficult for the model to decide on a proper threshold to precisely detect the changes, such as for change vector analysis (CVA) [7,8] and the similarity measure [9]. As for the algebra-based methods, they must handle the great computational cost due to the high-dimensional data structure. Regarding independent image classification, because the detection result is generated by processing two independent classification maps to produce a single change map, the result may be affected by the error propagation of both images’ classification results, which leads to bad performance.
The rapid development of deep learning methods has stimulated the evolution of hyperspectral change detection technologies. Compared to the traditional methods, the successful performance of deep learning models from other domains makes it possible to solve high-dimensional problems and extract extensive information more effectively. From the recent literature, deep learning methods can be roughly categorized into pixel-based methods and spatial–spectral-information-based methods. For pixel-wide scale analysis, researchers have contributed semi-supervised change detection frameworks, which are often composed of an encoder and a decoder to reconstruct the hyperspectral images by pixel spectral mapping. Then, the reconstructed images are distributed to downstream transformation methods and specific threshold segmentation methods, such as PCA and Otsu, to obtain the change map [10]. In this situation, the current methods consider deep learning as a preprocessing and compression step, but studies have not integrated the end-to-end change detection performance into a complete model, which results in the wasting of time and memory resources. Regarding the literature on feature extractors for both spectral and spatial information, existing methods propose the convolutional neural network to aggregate multi-direction information, by simply applying the conventional 2D spatial convolution on the compressed images or a direct 3D network to aggregate the two kinds of information at the same time [11,12,13]. Although these methods acknowledge that hyperspectral images themselves are equipped with various information, they still encounter bottlenecks that result from ignoring the spectral similarity in spatial areas and due to the insufficiency of hierarchical feature propagation. Furthermore, the local information usually can be distinguished by the relation to the global information, but current deep methods lack the ability to focus on the resemblance of spectral and spatial information simultaneously. The encoder–decoder framework in the literature was typically trained in a semi-supervised manner, which cannot hierarchically pass and fuse the features into deeper weight presentation to benefit change classification. Moreover, random features are distributed in fully connected layers, which are optimized by the later classification loss function and cannot be assigned for class judgment in advance. Therefore, these models’ drawbacks deeply hinder the improvement of current hyperspectral change detection methods’ performance.
To address these drawbacks of existing methods, in this paper, a parallel spectral–spatial attention network with feature redistribution loss in a supervised two-branch encoder–decoder framework is proposed for hyperspectral change detection. The proposed model rectifies the inner structure of the encoder–decoder framework so that it can be adopted in the proposed end-to-end model for transfer its original deep feature performance. Moreover, a weighted-in-parallel spectral–spatial attention module is constructed to focus on the long-range spectral and spatial dependencies within a patch’s input. To ensure that the fluent information flow can be fed into the classification function, an intensive module with multiple skip connections is applied. Furthermore, due to the unclear feature distribution’s influence in classification tasks [14], the feature redistribution loss function is developed here to provide better fused features for class-oriented loss computation.
The contributions of this paper are as follows:
(i) A parallel spectral–spatial attention module ( PS 2 A ) is proposed, which can provide the long-range dependenciesin the spectral domain and the local spatial region. Moreover, it fuses the separate stages of weighting the input data into a one-stage procedure, which can preserve the original data dependenciesand provide an enhancement in both the spectral and spatial domains at the same time.
(ii) The feature redistribution loss function (FRL) is developed for better feature arrangement. By maximizing the intra-group correlation and minimizing the extra-group correlation, a rectified loss function is developed, which helps the fused features form an enhanced distribution for class-oriented pattern recognition in advance. To ensure that fluent feature information for redistribution can be received, an intensive connected block (ICB) is applied with multiple skip connections between the convolutional submodules.
(iii) The encoder–decoder framework in a two-branch configuration is constructed in the end-to-end supervised change detection task. Although the decoder was not designed for pixel-level alignment, we found it quite effective to reuse its hierarchically transferred featuresfor change class prediction. This can largely contribute to the feature transfer by the skip connection and the fusion with the expanded feature reconstruction at a different scale.
Extensive experiments prove the effective and efficient performance of the presented method. The rest of this article is arranged as follows: Section 2 presents the related works on hyperspectral change detection methods. Section 3 introduces the proposed method. Section 4 gives the comparison result and ablation analysis of the proposed method. Section 5 is the conclusion of this study.

2. Related Works

2.1. Attention Mechanism

The attention mechanism was initially presented in the natural language processing domain, which is used to address the sequence-to-sequence transformation problem and is able to aggregate all the related information from the entire sequential input [15]. Along with the potential novelty and value of transferring the attention mechanism to computer vision applications, many works have created variants of the attention-based model for better effects in specific natural image analysis tasks. Among them, ViT is a prominent one, which reshapes the image patches into sequences and uses the attention mechanism in image classification [16]. Regarding object detection, DETR is another model that successfully explains the relation between the detected object and the global context to predict the final set of outputs in parallel [17]. Swin Transformer V2 is a model designed for semantic segmentation and uses the residual post-norm method to improve the training stability [18]. In the remote sensing change detection domain, Gong et al. proposed a spectral–spatial attention block to extract relevant information [11]. However, the generation of the attention map does not follow the standard weighting procedure, which may lead to inaccurate similarity computation. Qu et al. composed a dual-domain network to detect the change area from SAR images, which is based on the CNN architecture, and it can extract features at two different scales [19]. Wang et al. created a pyramid self-attention network to concentrate on features from different layers, but the training procedure followed a two-phrase auto-encoder framework and only applied dependency checks in the spatial domain [20]. Moreover, SSA-SiamNet, proposed by Wang et al., is an end-to-end model with a spectral–spatial attention mechanism. It uses a two-branch CNN feature extractor to generate gradually decreasing features for classification with a fully connected layer [21]. Applying the attention mechanism makes the extant models effective by emphasizing the spectral band and location relations while suppressing the unrelated information. Therefore, it is meaningful to take full advantage of the effectiveness of the attention mechanism in both the spectral and spatial domain.

2.2. Encoder–Decoder Framework

The encoder–decoder framework is a general method that generates an output sequence corresponding to an input sequence, and currently, it is being adopted in a variety of tasks such as semantic segmentation [22,23] and natural language processing [24,25]. In hyperspectral image analysis, most extant methods are designed according to the CNN or a dense prediction structure, such as GETNET and DSFANet [26,27], but there are still few studies on the supervised encoder–decoder framework. For hyperspectral classification, Zhu et al. created a spectral–spatial-dependent global learning framework, which utilizes the encoder–decoder structure to construct a pixel-level segmentation map [28]. For hyperspectral image reconstruction, Miao et al. developed a dual-stage generative model to reconstruct the desired 3D signal in snapshot compressive-spectral imaging. By applying U-Net to design a generative model, the 3D spectral cube can be directly reconstructed from the measurements and masks [29]. Huang et al. introduced a 3D filter generator, which can generate the spatially variant filters, into a lightweight U-Net for hyperspectral image reconstruction [30]. With the aim to address the conflicts of the dense sample tokens of the conventional Transformer model and the spatially sparse nature of hyperspectral image signals, Cai et al. developed a sparse Transformer (CST) model embedding HSI sparsity into deep learning for hyperspectral image reconstruction [31]. With respect to the application of the encoder–decoder structure in hyperspectral change detection, Lei et al. used the auto-encoder structure to learn the unsupervised patterns in the spectral pixel information by the definition of the reconstruction loss, then the reconstructed image was input into the later threshold segmentation procedure [10]. Due to the structure of hyperspectral images, the dimensionality reduction and data argumentation are usually used for the initial processing, and Li et al. adopted a two-branch U-Net network with feature fusion to achieve end-to-end change information detection automatically [32]. By adopting the encoder–decoder model in hyperspectral image analysis, current models can extend their potential to every specific research interest area. Therefore, it is necessary to improve the hierarchical feature transfer and fusion in the encoder–decoder networks for better application performances.

3. Proposed Method

To address these problems and challenges mentioned above and strengthen the robustness and accuracy of model performance, a two-branch feature redistribution network based on the parallel spectral–spatial attention mechanism (TFR-PS 2 ANet) is proposed. The overall architecture and module designation are shown in Figure 1.
The hyperspectral image dataset for change detection contains dual-temporal images, which can be described as T1 and T2. First, two parallel spectral–spatial attention modules ( PS 2 A ) are inserted at the start of the whole model to obtain long-range dependencies from both the spectral and spatial domain. Second, weighted feature maps are input into the encoder–decoder framework for a further fused and transformed feature representation. Finally, an intensive connection block is applied to distribute the features into the redistribution loss function and the classification loss function to have a better training procedure.
In this section, the proposed method is presented in four parts to give the detailed information of the design of each, which are the parallel spectral–spatial attention module, the two-branch encoder–decoder framework, the intensive connection block, and the feature redistribution loss function.

3.1. Parallel Spectral–Spatial Attention Module

The parallel spectral–spatial attention module ( PS 2 A ) is inserted in the first stage to aggregate the long-range dependencies from the original hyperspectral images as much as possible. As the module name implies, PS 2 A is composed of spectral attention and spatial attention, but each attention map is computed in parallel. The design details of the proposed PS 2 A module are shown in Figure 2.
For a hyperspectral image input patch x o r i in the shape of R H × W × D , the parallel spectral and spatial attention module is introduced. Regarding the processing of the spectral attention mechanism, the query vector is defined as v q s p e c , which is used as a reassignment factor for the attention map along with the key vector v k s p e c . A convolutional layer with a 1 × 1 kernel size is introduced to decrease the spectral channel of the input patch x o r i , followed by two adjacent fully connected layers with the SoftMax activation function to make the query vector v q s p e c eventually. In reference to the key vector v k s p e c , it is computed by a convolutional layer with 2 dilation rates, which can expand receptive field, followed by a reshape operation to change the features into a size of R H W × D / 2 . After v q s p e c and v k s p e c are generated, the matrix multiplication operation ⊗ is inserted between them to obtain the attention map v q k s p e c along with a fully connected layer and a SoftMax activation. All the process steps of the spectral attention module can be formulated as follows:
v q s p e c = F s o f t m a x ( F f c 2 ( F R ( F c o n v 1 × 1 ( x o r i ) ) ) ) R 1 × H W ,
v k s p e c = F R ( F d i l a t i o n 3 × 3 ( x o r i ) ) R H W × D / 2 ,
v q k s p e c = F s o f t m a x ( F f c ( v q s p e c v k s p e c ) ) R 1 × D ,
where F c o n v 1 × 1 ( · ) denotes the convolution operation with a kernel size of 1 × 1 and F f c ( · ) denotes the fully connected layer used to squeeze and expand the middle hidden features. The SoftMax activation function is represented as F s o f t m a x ( · ) . The dilation convolution operation with a kernel size of 3 × 3 is formulated as F d i l a t i o n 3 × 3 ( · ) , and the reshape operation F R ( · ) is applied in the computation of v q s p e c and v k s p e c . Then, the spectral attention map can be obtained by the matrix multiplication operation ⊗ to have v v weighted, which is also x o r i here.
With reference to spatial feature extraction, it has a similar process as for spectral feature extraction. The query vector of the spatial domain v q s p a is calculated by a convolutional layer with a kernel size of 3 × 3 , followed by an adaptive average pooling operation to compress the spatial size to 1 × 1 pixels. Then, the fully connected layer and the SoftMax activation function are applied to halve the number of features. As for the vector of spatial key vector v k s p a , it is calculated by dilation convolution with a kernel size of 3 × 3 and a dilation rate of 2, followed by the reshape operation to change the feature map size to R D / 2 × H W . Then, the same matrix multiplication operation ⊗ is applied between v q s p a and v k s p a , followed by the fully connected layer and SoftMax activation function. For the spatial weighting process, the attention map v q k s p a is reshaped to the same resolution size of R H × W . The detailed calculations are formulated as follows:
v q s p a = F s o f t m a x ( F f c ( F A A P ( F c o n v 3 × 3 ( x o r i ) ) ) ) R 1 × D / 2 ,
v k s p a = F R ( F d i l a t i o n 3 × 3 ( x o r i ) ) R D / 2 × H W ,
v q k s p a = F R ( F s o f t m a x ( F f c ( v q s p a v k s p a ) ) ) R H × W ,
With the obtained attention maps of v q k s p e c and v q k s p a by Equations (3) and (6), the value vector v v can be parallelly weighted to achieve the spectral and spatial dependency modeling. The local patch-level perception for hyperspectral images can be defined as:
x o u t a t t = v q k s p a v v v q k s p e c R H × W × D ,
where ⊙ indicates pointwise production for every band and pixel and v v represents the value vector, which here indicates the original patch input x o r i . x o u t a t t is the output feature map after weighting.

3.2. Two-Branch Encoder–Decoder Framework

Different from the traditional classification backbone following the pyramid structure, here, the proposed feature extractor is guided by the encoder–decoder structure, which is common in the semantic segmentation of natural or biomedical images. Due to the fact that the dual-temporal images may have different local spectral and spatial information at the same pixel location, the feature extractor based on the encoder–decoder framework is constructed and assembled in the proposed method in a parallel manner. The two-branch design of the rectified encoder–decoder feature extractor is shown in Figure 3.
As the left side of Figure 3 shows, the proposed encoder generates internal feature maps using independent convolutional layers with a kernel size of 3 × 3 . The patch size decreases gradually from 9 × 9 to 1 × 1 . Take the first layer of the encoder network as an example: feature weighted by the attention map in the shape of R 9 × 9 × D is input into a convolutional block with three successive convolutional layers, and then, the output feature map e 0 R 9 × 9 × 256 is calculated and temporarily saved. By a max pooling layer with a kernel size of 2 × 2 , the output features from the last layer are downsampled into the shape of R 7 × 7 × 256 . Therefore, the components of the encoder are made up of the same and stacked convolutions and max pooling operations. As the right side of Figure 3 shows, the decoder is responsible for processing and expanding the output features generated from the encoder components. Take the last output layer in the encoder as an example: the spatial resolution of e 4 R 1 × 1 × 16 is expanded by the upsampling operation, while the channel number is enlarged by a convolutional layer. The built feature map is defined as d 3 R 3 × 3 × 32 , which has the same shape as the corresponding encoder e 3 . Both e 3 and d 3 are concatenated, and hence, the channel dimension is doubled as well. Then, the features are processed by a convolutional block with three successive convolutional layers, which have the same definition as that in the encoder.
Given an input patch x e n i , a single convolutional layer F C L ( · ) can be defined as follows, which is constructed with a convolution, a batch normalization, and a SELU activation function:
F C L ( x e n i ) = F S E L U ( F B N ( F c o n v 3 × 3 ( x e n i ) ) ) , i = 0 , 1 , , 4 ,
In (8), F B N ( · ) indicates batch normalization and F S E L U ( · ) denotes the SELU activation function. The three successive convolutions are represented as F C L 3 ( · ) , and i corresponds to the i-th block of the encoder. Therefore, the relation between the features from the front and back block can be written as follows:
x e n i + 1 = F m a x p o o l ( F C L 3 ( x e n i ) ) ,
where F m a x p o o l ( · ) indicates the max pooling operation used to squeeze the spatial resolution in the encoder part. For the decoder part, the input for every decoder block x d e i is processed by the following formula:
F c o n v u p ( x d e i ) = F R e L U ( F B N ( F c o n v 3 × 3 ( F s a m p l e u p ( x d e i ) ) ) ) ,
x d e i = F C L 3 ( F c o n c a t ( F c o n v u p ( x d e i + 1 ) + x e n i ) ) , i = 0 , 1 , 2 , 3 ,
The definition of the upsampling with the convolutionis represented in (10), where F s a m p l e u p ( · ) indicates the upsampling operation using the nearest interpolation and F R e L U ( · ) indicates the ReLU activation function. Then, the output features of the upsampling with the convolutionis concatenated along with the corresponding copy x e n i from the encoder and computed by sequential convolutional layers F C L 3 ( · ) .

3.3. Intensive Connected Block

To keep the information flow fluent and to pass the hierarchical features, the intensive connected block is constructed. Compared to the residual skip connection, three bottleneck layers with intensive skip connections here were applied to make the model have a low computation cost while keeping a similar time complexity, which was inspired by DenseNet [33].
The input features x I C B i n R H × W × 64 are concatenated by two output feature sets generated from two-branch encoder–decoder framework. Then, it is processed by the stem layer to reduce the channel number to avoid overfittingin the intensive skip connections. When feature maps arrive at the last skip connection, the channel number is expanded by three bottleneck layers. Therefore, a transition layer is introduced to avoid the burden of too many features. The process in the intensive connected block can be formulated as follows:
x 0 = F s t e m ( x I C B i n ) = F c o n v 3 × 3 ( x I C B i n ) R H × W × 32 ,
x l = β ( F c o n c a t ( x 0 + x 1 + x l 1 ) ) , l = 1 , 2 , 3 ,
x I C B o u t = F t r a n s i t i o n ( x 3 ) = F c o n v 1 × 1 ( F R e L U ( F B N ( x 3 ) ) ) R H × W × 16 ,
where F s t e m ( · ) indicates a stem convolutional layer with a 3 × 3 kernel size to decrease feature channel number and β ( · ) indicates the bottleneck layer shown in Figure 4. F t r a n s i t i o n ( · ) denotes the transition layer for the computation of the feature output x I C B o u t .

3.4. Feature Redistribution Loss Function

To obtain a better class-oriented feature expression, the feature redistribution loss function (FRL) was developed to distribute the features more properly. The feature redistribution can provide enhanced and fused features for pattern recognition [14]. Inspired by this work and due to the procedure of the conventional loss computation lacking class distribution information for change detection, the FRL is proposed and can enhance the independence of the feature representation. The structure of the FRL is shown in Figure 5.
After the process of the ICB, the output features x I C B o u t are convoluted to x F R L i n R 9 × 9 × 16 , followed by a fully connected layer to change the features to x G R 1 × 1 × 64 . Then, the features x G are divided into a set of vectors of length m. Therefore, the defined feature number of vectors is calculated as:
n = 64 / m .
The number of divided groups is the same as the number of land cover change categories c, and the covariance matrix C corresponding to the variants n can be calculated as:
C = 1 m i = 1 m ( x i μ ) ( x i μ ) T R n × n ,
μ = 1 m i = 1 m x i ,
where μ R n × 1 indicates the mean vector of vector length m. All defined n features can be represented as V and grouped into c categories { G 1 , G 2 , , G c } . Every group contains s = n / c defined features. With the constructed covariance matrix C , the correlation matrix can be represented as Re , where the correlation coefficient between the indices of i and j is calculated as:
Re i j = | C i j | C i i C j j ,
To maximize the correlation between extra-groups γ e x t r a and minimize the correlation in inter-groups γ i n t e r , both correlations can be defined as follows:
γ e x t r a = k = 1 c i G k , j V G k Re i j i G k , j V Re i j ,
γ i n t e r = k = 1 c i , j G k Re i j i G k , j V Re i j ,
Considering a simple and convenient representation of the formula in (19), the part that represents the correlation for the k-th group can be defined as:
μ k = i G k , j V G k Re i j i G k , j V Re i j ,
According to the correlation matrix Re and the relation of the indices in μ k , the interval of μ k can be calculated and γ e x t r a can be simplified as well:
μ k [ 0 , c 1 c ] ,
γ e x t r a = k = 1 c μ k ,
Therefore, the normalized loss between the extra-groups can be represented as:
L o s s e x t r a = c c ( c 1 ) k = 1 c μ k [ 0 , 1 ] ,
Maximizing the internal correlation γ i n t e r is equivalent to minimizing 1 γ i n t e r . The definition is as follows:
σ k = i G k , j V Re i j i , j G k Re i j [ 1 , c ] ,
Therefore, the normalized loss of the internal correlation can be calculated with:
L o s s i n t e r = 1 c ( c 1 ) k = 1 c σ k 1 [ 0 , 1 ] ,
With the calculated extra-loss and inter-loss, it can be assembled together by the weights’ distribution. The whole loss function can be represented as below:
L o s s = λ 2 ( λ 1 L o s s e x t r a + ( 1 λ 1 ) L o s s i n t e r ) + ( 1 λ 2 ) C E L o s s ,
where λ 1 , λ 2 ( 0 , 1 ) indicate the weight coefficients to adjust the sizeof every loss. C E L o s s indicates the cross-entropy loss. The effectiveness of the proposed loss function is discussed in the later ablation experiments, and the impacts of the hyperparameters are analyzed as well.
After the redistribution loss is applied, the features are sent to the fully connected layer, followed by SoftMax activation function, which links the cross-entropy loss function. This prediction procedure can be formulated as:
y p r e d = F s o f t m a x ( F f c ( x G ) ) .

4. Experiments

4.1. Description of Dataset

In this paper, three public dual-temporal datasets for hyperspectral change detection were chosen for model testing, which were all captured by the Earth Observing-1 (EO-1) satellite with the Hyperion sensor. It provides a spectral range of 0.4–2.5 μm with 242 spectral bands and a spectral resolution of 10 nm approximately, as well as a spatial resolution of 30 m. In the experiments, spectral bands with a low signal-to-noise ratio (SNR) were removed. To distinguish whether the land cover area had changed or not, the binary ground truth maps were obtained by visually analyzing extensive studies and on-the-spot investigation. The first hyperspectral image dataset is the Irrigated Agriculture Dataset [34] captured on 1 May 2004 and 8 May 2007, which illustrates an irrigated agricultural area of Hermiston City in Umatilla County, Oregon, USA. It contains 307 × 241 pixels and 156 bands after omitting no-data bands and removing the noise. The training rate followed the original work, which was 9.7% approximately. This dataset is shown in Figure 6. The second dataset is the Wetland Agriculture Dataset [34], which contains images captured on 3 May 2006 and 23 April 2007. This dataset illustrates a farmland area of Yuncheng City, Zhejiang Province, China. It contains 450 × 140 pixels and 156 bands after removing noise. Referring to the training rate of the Irrigated Agriculture Dataset, the training samples for the Wetland Agriculture Dataset were further decreased compared to the original work, which was set to about 9.7% as well. The T1 image, T2 image, and ground truth are shown Figure 7. The last change detection dataset is the River Dataset [26], which contains images captured on 3 May 2013 and 31 December 2013. This dataset illustrates a river area in Jiangsu Province, China. It contains 463 × 241 pixels and 198 bands after noise removal. The false color image for the River Dataset is shown in Figure 8. Every dataset was divided into a training set, validation set, and test set. The training samples of the Irrigated Agriculture Dataset and Wetland Agriculture Dataset occupied 9.7% of the total samples approximately using stratified random sampling. For the River Dataset, to consider the problem of class imbalance, the proportion of the unchanged samples to the changed samples was 2:1, which followed the rule of original work. Eventually, the training rate for the River Dataset was 4.03%. For all three public datasets, stratified random sampling was used to generate the random training samples. The details of every dataset are shown in Table 1.

4.2. Experimental Setup

All experiments were performed on an NVIDIA RTX 3060 with 12G of video memory. Due to the model structure, it can handle different patch input sizes without any spatial resolution restriction. In the training set procedure, the batch size was chosen as 64 patch inputs by default and the learning rate as 10 5 with a corresponding 10 4 weight decay applied to exhibit better convergence. The Adagrad optimizer was used here to train the proposed TFR-PS 2 ANet. The united loss function with the cross-entropy loss and the feature redistribution loss were chosen for the model training procedure, and the formula of the cross-entropy loss is shown below:
L = 1 N i L i = 1 N i c = 1 M y i c l o g ( p i c ) ,
where M indicates the number of categories and y i c is an indicator for which the value is equal to 1 or 0 depending on whether it is the true category of sample i. The prediction probability belonging to c of sample i is expressed as p i c .

4.3. Evaluation Metrics

For a fair and convenient performance comparison among the three hyperspectral change detection datasets, the accuracy (ACC), kappa coefficient (kappa), F1-score, precision, and recall were selected to evaluate and quantized model’s performance. Every index was calculated based on a confusion matrix, and the larger the value, the better the performance is.
Accuracy (ACC): Regarding pixel-level classification tasks, accuracy is a relatively simple, but effective metric to weigh the model’s performance. The formula for the calculation of the accuracy is:
A C C = i = 0 c T P i i = 0 c ( T P i + F P i ) .
Kappa coefficient (kappa): The kappa coefficient is another metric usually used for pixel-level classification. According to the formula of the kappa coefficient, it takes the class imbalance into consideration and can measure the model’s performance on different datasets fairly. The formula for the calculation of the kappa coefficient is:
k = p o p e 1 p e .
F1-score: The F1-score (also known as the F1-measure) is designed to evaluate the performance of a pixel-level binary classification model. It can be considered as the harmonic mean of the precision and recall. The F1-score is calculated as:
F 1 = 2 · p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l .
Precision and recall: The precision indicates the true positives in the sum of the true positive and false positive samples. The recall indicates the true positives in the sum of the true positive and false negative samples. The formulae are, respectively:
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N .

4.4. Experimental Results

In this section, five methods in the change detection domain were chosen to make comparisons with the proposed TFR-PS 2 ANet. Two of them are traditional methods based on supervised or unsupervised learning, and the rest are SOTA deep learning methods. CVA is a traditional method using difference maps and unsupervised segmentation, which was Otsu here [35]. The second traditional method for pattern recognition is SVM, which is extensively used in land cover change detection [36]. As for the deep learning methods, GETNET [26] based on the LSConvolution architecture and DSFANet [27] following the principle of semi-supervised learning were chosen as the comparison methods. For the CNN architecture in two-domain learning, the patch-based WCRN model [19] was chosen as the comparison model. For the mentioned deep-learning-based methods, the same default parameter settings introduced in the corresponding works were selected. Extensive experiments were conducted on all three hyperspectral change detection datasets. The detailed model comparison results are described below. The overall accuracy and F1-score comparison is shown in Figure 9, which indicates that the proposed TFR-PS 2 ANet is superior to most of the other methods.

4.4.1. Experiments on the Irrigated Agriculture Dataset

The model comparison results can be seen in Table 2. The proposed TFR-PS 2 ANet was able to achieve the best results on the ACC, kappa coefficient, F1-score, and recall. The best precision was achieved by CVA, while this model had the worst recall result, which is less than about 0.29 compared to that of TFR-PS 2 ANet. The best competitor was the SVM method. As a traditional method, it outperformed most of the selected deep learning methods. Among them, the unsupervised DSFANet showed the worst change detection results. GETNET showed the best recall in comparison to the other methods, but its precision was lower than that of TFR-PS 2 ANet, which resulted in an inferior F1-score result. DSFANet is an unsupervised method just like CVA, but it detects too many unchanged pixels wrongly according to its poor precision performance, which may be a result of the complex scenes of the agriculture images.
In the last three rows of Table 2, the comparisons of TFR-PS 2 ANet, TFR-PS 2 ANet without redistribution loss, and TFR-PS 2 ANet without PS 2 A are given. The accuracy and kappa coefficient performance of TFR-PS 2 ANet without redistribution loss showed a clear drop compared to TFR-PS 2 ANet. TFR-PS 2 ANet without PS 2 A showed a slight accuracy decrease, while the kappa coefficient had an obvious disparity. Although the two ablation studies of the models demonstrated weaker results than TFR-PS 2 ANet, they still outperformed the other competitors, which implies that the encoder–decoder framework can effectively extract the change features. In conclusion, the proposed TFR-PS 2 ANet had the best performance on most of the metrics compared to the other methods.
The change detection maps for the Irrigated Agriculture Dataset are shown in Figure 10. Using the ground truth map as a reference, the proposed TFR-PS 2 ANet showed the closest visual effects, and the border pixels were visually distinct like the ground truth map, while the change maps detected by CVA, GETNET, and WCRN failed to distinguish the subtle changes in the border pixels. Although SVM showed great visual effects, it detected unrelated subtle changes of the border pixels mistakenly, which led to inferior accuracy and precision. Regarding DSFANet, it showed scattered points in the change result map, which caused a false detection and led to lower accuracy and kappa indices. The reason for this situation could be that the unsupervised DSFANet is not suitable for complex land cover spectral information. Regarding the ablation results of the models TFR-PS 2 ANet without redistribution loss or PS 2 A , it can be seen from the visual effects that TFR-PS 2 ANet without redistribution loss would detect some border pixels mistakenly, while TFR-PS 2 ANet without PS 2 A would detect border pixels as a large spot, which would affect the detection result.

4.4.2. Experiments on the Wetland Agriculture Dataset

The comparison results of the models on the Wetland Agriculture Dataset are shown in Table 3. For this dataset, the proposed method TFR-PS 2 ANet outperformed the other methods in all indices as well. The best competitor was GETNET, which achieved the highest accuracy in comparison the other models, but it was inferior to TFR-PS 2 ANet by a slight worse accuracy and kappa coefficient. CVA and SVM were still stable models, achieving very similar accuracy and kappa coefficient performance. DSFANet showed the worst performance and almost could not detect the main farm change region, which can be concluded from the kappa coefficient result. WCRN could achieve an acceptable accuracy result, but the kappa coefficient remained low.
The last three rows of Table 3 show the comparison among TFR-PS 2 ANet, TFR-PS 2 ANet without redistribution loss, and TFR-PS 2 ANet without PS 2 A . Compared to TFR-PS 2 ANet, the ablation results of the model TFR-PS 2 ANet without redistribution loss were inferior on every metric. The performance of TFR-PS 2 ANet without PS 2 A was worse than TFR-PS 2 ANet without redistribution loss, which may indicate that long-range dependencies are very necessary in areas with much crop spectral information. In conclusion, the proposed TFR-PS 2 ANet showed the best performance on the Wetland Agriculture Dataset.
The detected change maps of these methods for the Wetland Agriculture Dataset are shown in Figure 11. The proposed TFR-PS 2 ANet showed the most similarity to the ground truth map in visual effects, which kept the detailed change information while suppressing unrelated subtle information. The two traditional methods, CVA and SVM, could detect the main agriculture change area, but too many scattered points were also detected mistakenly, which influenced the accuracy and precision indices. GETNET successfully detected most of the change area, but there was a loss of subtle farmland change information as well. DSFANet showed the worst performance, which basically only detected the border of change areas. This may have resulted from its insufficient capacity to process complex scenes. WCRN could detect most of the change area, but the main drawbacks were that it lost the detailed information and farmland edge change information. Furthermore, in the top area of the WCRN change map, it detected some unrelated points, which decreased the accuracy and precision results. Regarding the last three change maps, it can be observed that they basically had similar visual effects except several wrong pixels detected by mistake, and they all showed the best similarity to the ground truth map compared to the other methods.

4.4.3. Experiments on the River Dataset

The comparison results of different methods on the River Dataset are shown in Table 4. Our proposed TFR-PS 2 ANet achieved the best ACC, kappa, F1-score, and precision. WCRN was the best competitor, which was designed in the original work on the River Dataset, and showed the best accuracy performance (0.9284) among the compared deep learning methods, followed by the deep leaning GETNET method, having a 0.9260 accuracy. SVM and DSFANet had a 0.9198 and 0.8883 accuracy, respectively. The unsupervised DSFANet did not perform well, which could be a result of its insufficient feature learning under class imbalance.
In the same way, the last three rows give the ablation results of the models. Compared to TFR-PS 2 ANet, the model TFR-PS 2 ANet without redistribution loss showed a considerable accuracy drop by around 0.01, while the model without PS 2 A showed a slight accuracy decrease. Therefore, it is necessary for change detection to adopt the class-oriented feature redistribution in the second to last layerfor the River Dataset. In conclusion, the proposed TFR-PS 2 ANet can perform best among these methods while handling with class imbalance.
The change detection maps for the River Dataset are shown in Figure 12. It can be observed that TFR-PS 2 ANet showed the closest similarity to the ground truth map from the visual effects. The first result map generated by the unsupervised CVA showed many unrelated pixels detectedto the left. On the contrary, the second SVM result map showed a failure to detect change pixels to the left. GETNET, as the best competitor, could detect distinct change areas, but they were expanded slightly. DSFANet showed many points and gaps due to it insufficient unsupervised post-processing. For the visual effects, the change map from WCRN showed rough boundaries, which caused a low-precision result. With reference to the change map generated by TFR-PS 2 ANet without redistribution loss or PS 2 A , both could detect some noise-like pixels shown ot the top of the change map, which influenced the final accuracy and F1-score, while they still outperformed the other compared methods.

4.4.4. Impact of Hyperparameters

In this section, the influence by both hyperparameters for the redistribution loss and the cross-entropy loss, which are the weights λ 1 , λ 2 [ 0.2 , 0.4 , 0.5 , 0.6 , 0.8 ] in Formula (27), were analyzed. The hyperparameter λ 1 is responsible for the adjustment for the extra-loss in the redistribution loss function, while 1 λ 1 is responsible for the inter-loss. With the hyperparameter λ 1 selected, the influence of the distribution of the features can be analyzed. The hyperparameter λ 2 was utilized to adjust the influence of redistribution loss, while 1 λ 2 worked as the weight for the cross-entropy loss. As Figure 13 shows, the performance impacts for the accuracy, Kappa, and F1-score were analyzed on all three datasets.
The first row shows the accuracy change influenced by hyperparameters λ 1 and λ 2 . When λ 1 = 0.2 , the best accuracy performance on the River Dataset was achieved at λ 2 = 0.6 , while that on the Irrigated Agriculture Dataset was achieved at λ 2 = 0.8 . The best accuracy performance on the Wetland Agriculture Dataset was achieved at λ 2 = 0.6 as well. When λ 1 = 0.4 , the two best accuracies on Wetland Agriculture and River Datasets were both achieved at λ 2 = 0.6 , while on the Irrigated Agriculture Dataset, the proposed method achieved the best performance at λ 2 = 0.5 . Along with the increase of λ 1 , the proposed TFR-PS 2 ANet could achieve the best accuracy on the three datasets at λ 2 = 0.6 .
The influence on the kappa index is shown in the second row of Figure 13. It had a similar trend to the accuracy change. From the first figure of the second row, it can be observed that the proposed method could achieve the highest kappa results at λ 2 = 0.6 on the Wetland Agriculture Dataset and the River Dataset, while on the Irrigated Agriculture Dataset, it performed best at λ 2 = 0.8 . When λ 1 = 0.4 , the best results of the kappa index on the Wetland Agriculture and River Datasets were achieved at λ 2 = 0.6 as well. However, the highest performance on the Irrigated Agriculture Dataset was achieved at λ 2 = 0.5 . According to the last three figures, the proposed TFR-PS 2 ANet achieved the highest kappa results when λ 2 = 0.6 at the same time.
The last row shows the influence on the F1-score index. From the overall visual effects, TFR-PS 2 ANet performed stably on both the Irrigated and Wetland Agriculture Datasets. Regarding the analysis of the Irrigated Agriculture Dataset, except the results at λ 1 = 0.2 , 0.4 , the rest of the figures show that the best F1-score results were achieved at λ 2 = 0.6 as well. The results on the Wetland Agriculture Dataset showed similar stable trends as those on the Irrigated Agriculture Dataset, which mostly achieved the highest F1-score results at λ 2 = 0.6 . The trends on the River Dataset increased their fluctuation and stopped at λ 2 = 0.6 to obtain the best performance.
In other words, from the optimal analysis results of the accuracy, kappa coefficient, and F1-score on all three datasets, the proposed model could obtain the best performance when λ 1 = 0.5 , 0.6 , 0.8 and λ 2 = 0.6 . Due to the hyperparameter λ 1 being able to influence the extra- and inter-loss balance, the model hyperparameters were set as λ 1 = 0.5 and λ 2 = 0.6 finally.

4.5. Discussion

The advantage in terms of the accuracy of the proposed TFR-PS 2 ANet method over the benchmark methods was due mainly to the application of the PS 2 A module, which contains the parallel spectral–spatial attention mechanism. Moreover, the proposed TFR-PS 2 ANet method is assisted by the FRL, which can reassign the features to a class-oriented distribution. According to the extensive experiments conducted above, CVA, as the conventional unsupervised method, performed stably on the three datasets. However, the change detection results largely relied on the determination of a threshold. The inadequate extraction of spectral information led to inferior detection results compared to our proposed method. As a conventional supervised method, SVM needs training samples. Practically, it can be difficult for SVM to handle local spatial information while finding a proper discriminative property for spectral information. On the contrary, the proposed method extracts features from both the spectral and spatial domain, which led to better discriminative abilities of change detection.
For the deep learning methods, GETNET is based on the supervised LSConvolution architecture. It extracts features from the input patch without considering internal spectral and spatial dependencies. Our proposed method can adaptively obtain long-range dependencies using the parallel attention mechanism, while the FRL assists in arranging the features properly. DSFANet is performed in a semi-supervised manner. However, this method showed the worst results on all the change detection datasets, which implies its weakness in dealing with complex spectral information using unsupervised post-processing. In terms of the comparison between WCRN and TFR-PS 2 ANet, the proposed method using the attention mechanism showed the better effectiveness of the adaptive two-domain feature learning compared to the CNN-based architecture.
However, there are also some limitations of the proposed method. First, the three change detection datasets have different time intervals. This leads to the problem of the continuous change amount during different periods being hard to compare. The design of the proposed method regarding the dual-temporal images as a pair of independent moments indicates that the change monitoring needs the support of data with the same or higher time resolution. To address this issue, it is worthwhile to extend the current TFR-PS 2 ANet to time series analysis. Nevertheless, in most situations, a time series change detection method can only be driven by dual-temporal images from satellites with different spectral and spatial resolutions. Therefore, reliable change monitoring results would largely depend on the accurate match between image pairs with different spatial and spectral resolutions. Furthermore, anthropogenic effects may influence the agriculture change results. However, the proposed supervised method based on the binary change ground truth detects changes in the agricultural area without distinguishing whether these changes are caused by human activities. The performance of the supervised change method is largely influenced by the available ground reference labels. Therefore, to detect changes from one labelto another using the change method, more attention should be paid to constructing multiple change ground truths with more informative on-the-spot investigation.

5. Conclusions

In this article, a general network named TFR-PS 2 ANet was proposed to detect land cover changes in dual-temporal hyperspectral images. First, the proposed method, which integrates the P S 2 A module, adaptively extracts spectral and spatial features from input patch pairs. The extracted features enhance the relevant long-range dependencies and suppress the irrelevant information in both the spectral and spatial domain simultaneously. Second, the FRL was added to reassign hidden features to a class-oriented distribution, which can enhance the discriminative ability for changed and unchanged pixels. Moreover, a general two-branch encoder–decoder framework was designed to transform and fuse the high-dimensional information to another characteristic space while keeping the hierarchically transferred features.
We implemented our algorithm and performed experiments on three public hyperspectral change detection datasets. The visual and quantitative results both showed that the proposed method outperformed most of the state-of-the-art methods including conventional and deep network algorithms.
When dealing with different change detection tasks, the proposed TFR-PS 2 ANet can be seen as a benchmark method with an adaptive one-stage spectral–spatial feature extraction module. The FRL can provide a reference for various loss functions that accelerate the convergence of the network, as well. In the future, using representation information in self-supervised learning will be considered to improve the performance.

Author Contributions

Conceptualization, Y.H.; methodology, Y.H.; software, Y.H.; validation, Y.H.; writing, Y.H.; writing—review and editing, C.H., W.Q. and R.S.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Key Research and Development Projects (grant no. 2022YFF0904400) and the National Natural Science Foundation of China (grant no. 41830108).

Data Availability Statement

The data presented in this study are openly available in Remote Sensing Datasets at https://rslab.ut.ac.ir/data, accessed on 1 December 2022, and River Dataset at http://crabwq.github.io, accessed on 1 December 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in Hyperspectral Image and Signal Processing: A Comprehensive Overview of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef] [Green Version]
  2. Du, P.; Liu, S.; Bruzzone, L.; Bovolo, F. Target-Driven Change Detection Based on Data Transformation and Similarity Measures. In Proceedings of the 2012 IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012; pp. 2016–2019. [Google Scholar]
  3. Kaldane, H.; Turkar, V.; De, S.; Shitole, S.; Deo, R. Land Cover Change Detection for Fully Polarimetric SAR Images. In Proceedings of the 2019 URSI Asia-Pacific Radio Science Conference (AP-RASC), New Delhi, India, 9–15 March 2019; pp. 1–4. [Google Scholar]
  4. Xiao, P.; Sheng, G.; Zhang, X.; Liu, H.; Guo, R. Direction-Dominated Change Vector Analysis for Forest Change Detection. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102492. [Google Scholar] [CrossRef]
  5. Washaya, P.; Balz, T.; Mohamadi, B. Coherence Change-Detection with Sentinel-1 for Natural and Anthropogenic Disaster Monitoring in Urban Areas. Remote Sens. 2018, 10, 1026. [Google Scholar] [CrossRef] [Green Version]
  6. Liu, S.; Marinelli, D.; Bruzzone, L.; Bovolo, F. A Review of Change Detection in Multitemporal Hyperspectral Images: Current Techniques, Applications, and Challenges. IEEE Geosci. Remote Sens. Mag. 2019, 7, 140–158. [Google Scholar] [CrossRef]
  7. Liu, S.; Bruzzone, L.; Bovolo, F. Peijun Du Hierarchical Unsupervised Change Detection in Multitemporal Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 244–260. [Google Scholar] [CrossRef]
  8. Bovolo, F.; Marchesi, S.; Bruzzone, L. A Framework for Automatic and Unsupervised Detection of Multiple Changes in Multitemporal Images. IEEE Trans. Geosci. Remote Sens. 2012, 50, 2196–2212. [Google Scholar] [CrossRef]
  9. Seydi, S.T.; Hasanlou, M. A New Land-Cover Match-Based Change Detection for Hyperspectral Imagery. Eur. J. Remote Sens. 2017, 50, 517–533. [Google Scholar] [CrossRef]
  10. Lei, J.; Li, M.; Xie, W.; Li, Y.; Jia, X. Spectral Mapping with Adversarial Learning for Unsupervised Hyperspectral Change Detection. Neurocomputing 2021, 465, 71–83. [Google Scholar] [CrossRef]
  11. Gong, M.; Jiang, F.; Qin, A.K.; Liu, T.; Zhan, T.; Lu, D.; Zheng, H.; Zhang, M. A Spectral and Spatial Attention Network for Change Detection in Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  12. Ou, X.; Liu, L.; Tu, B.; Zhang, G.; Xu, Z. A CNN Framework With Slow-Fast Band Selection and Feature Fusion Grouping for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  13. Seydi, S.T.; Hasanlou, M. A New Structure for Binary and Multiple Hyperspectral Change Detection Based on Spectral Unmixing and Convolutional Neural Network. Measurement 2021, 186, 110137. [Google Scholar] [CrossRef]
  14. Zuobin, W.; Kezhi, M.; Ng, G.-W. Feature Regrouping for CCA-Based Feature Fusion and Extraction Through Normalized Cut. In Proceedings of the 2018 21st International Conference on Information Fusion, Cambridge, UK, 10–13 July 2018. [Google Scholar]
  15. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:170603762. [Google Scholar]
  16. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition At Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  17. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. arXiv 2020, arXiv:200512872. [Google Scholar]
  18. Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin Transformer V2: Scaling Up Capacity and Resolution. arXiv 2022, arXiv:211109883. [Google Scholar]
  19. Qu, X.; Gao, F.; Dong, J.; Du, Q.; Li, H.-C. Change Detection in Synthetic Aperture Radar Images Using a Dual-Domain Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  20. Wang, G.; Peng, Y.; Zhang, S.; Wang, G.; Zhang, T.; Qi, J.; Zheng, S.; Liu, Y. Pyramid Self-Attention Mechanism-Based Change Detection in Hyperspectral Imagery. J. Appl. Remote Sens. 2021, 15. [Google Scholar] [CrossRef]
  21. Wang, L.; Wang, L.; Wang, Q.; Atkinson, P.M. SSA-SiamNet: Spectral–Spatial-Wise Attention-Based Siamese Network for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 18. [Google Scholar] [CrossRef]
  22. Bao, H.; Dong, L.; Wei, F. BEiT: BERT Pre-Training of Image Transformers. arXiv 2021, arXiv:2106.08254. [Google Scholar]
  23. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  24. Takase, S.; Kiyono, S. Rethinking Perturbations in Encoder-Decoders for Fast Training. arXiv 2021, arXiv:210401853. [Google Scholar]
  25. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to Sequence Learning with Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  26. Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A General End-to-End 2-D CNN Framework for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3–13. [Google Scholar] [CrossRef] [Green Version]
  27. Du, B.; Ru, L.; Wu, C.; Zhang, L. Unsupervised Deep Slow Feature Analysis for Change Detection in Multi-Temporal Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9976–9992. [Google Scholar] [CrossRef] [Green Version]
  28. Zhu, Q.; Deng, W.; Zheng, Z.; Zhong, Y.; Guan, Q.; Lin, W.; Zhang, L.; Li, D. A Spectral-Spatial-Dependent Global Learning Framework for Insufficient and Imbalanced Hyperspectral Image Classification. IEEE Trans. Cybern. 2021, 1–15. [Google Scholar] [CrossRef]
  29. Miao, X.; Yuan, X.; Pu, Y.; Athitsos, V. Lambda-Net: Reconstruct Hyperspectral Images From a Snapshot Measurement. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4058–4068. [Google Scholar]
  30. Huang, T.; Dong, W.; Yuan, X.; Wu, J.; Shi, G. Deep Gaussian Scale Mixture Prior for Spectral Compressive Imaging. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 16211–16220. [Google Scholar]
  31. Cai, Y.; Lin, J.; Hu, X.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Van Gool, L. Coarse-to-Fine Sparse Transformer for Hyperspectral Image Reconstruction. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  32. Li, Q.; Mu, T.; Feng, Y.; Gong, H.; Han, F.; Tuniyazi, A.; Li, H.; Wang, W.; Li, C.; He, Z.; et al. Hyperspectral Image Change Detection Using Two-Branch Unet Network with Feature Fusion. In Proceedings of the Fourth International Conference on Photonics and Optical Engineering, Xi’an, China, 15 January 2021; p. 82. [Google Scholar]
  33. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  34. Hasanlou, M.; Seydi, S.T. Hyperspectral Change Detection: An Experimental Comparative Study. Int. J. Remote Sens. 2018, 39, 7029–7083. [Google Scholar] [CrossRef]
  35. Bovolo, F.; Bruzzone, L. A Theoretical Framework for Unsupervised Change Detection Based on Change Vector Analysis in the Polar Domain. IEEE Trans. Geosci. Remote Sens. 2007, 45, 218–236. [Google Scholar] [CrossRef] [Green Version]
  36. Nemmour, H.; Chibani, Y. Multiple Support Vector Machines for Land Cover Change Detection: An Application for Mapping Urban Extensions. ISPRS J. Photogramm. Remote Sens. 2006, 61, 125–133. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the proposed TFR-PS 2 ANet. TFR-PS 2 ANet consists of four components: PS 2 A module, two-branch encoder–decoder architecture, ICB module, and feature redistribution loss function.
Figure 1. The overall architecture of the proposed TFR-PS 2 ANet. TFR-PS 2 ANet consists of four components: PS 2 A module, two-branch encoder–decoder architecture, ICB module, and feature redistribution loss function.
Remotesensing 15 00246 g001
Figure 2. The detailed design of the proposed PS 2 A module. The top part denotes the generation procedure of the spectral attention map, and the bottom part denotes the generation procedure of the spatial attention map.
Figure 2. The detailed design of the proposed PS 2 A module. The top part denotes the generation procedure of the spectral attention map, and the bottom part denotes the generation procedure of the spatial attention map.
Remotesensing 15 00246 g002
Figure 3. The two-branch backbone based on the rectified encoder–decoder structure. The encoder is on the left side, and the decoder is on the right side.
Figure 3. The two-branch backbone based on the rectified encoder–decoder structure. The encoder is on the left side, and the decoder is on the right side.
Remotesensing 15 00246 g003
Figure 4. The structure of the proposed intensive connected block.
Figure 4. The structure of the proposed intensive connected block.
Remotesensing 15 00246 g004
Figure 5. Structure of the proposed feature redistribution loss function using the maximized inter-group relation and minimized extra-group relation. Hidden features are divided into c groups.
Figure 5. Structure of the proposed feature redistribution loss function using the maximized inter-group relation and minimized extra-group relation. Hidden features are divided into c groups.
Remotesensing 15 00246 g005
Figure 6. Irrigated Agriculture Dataset with false color map (Bands 134, 90m and 75 as RGB). (a) U.S. farmland image captured on 1 May 2004. (b) U.S. farmland image captured on 8 May 2007. (c) Binary ground truth for the Irrigated Agriculture Dataset.
Figure 6. Irrigated Agriculture Dataset with false color map (Bands 134, 90m and 75 as RGB). (a) U.S. farmland image captured on 1 May 2004. (b) U.S. farmland image captured on 8 May 2007. (c) Binary ground truth for the Irrigated Agriculture Dataset.
Remotesensing 15 00246 g006
Figure 7. Wetland Agriculture Dataset with false color map (Bands 134, 90, and 75 as RGB). (a) Chinese farmland image captured on 3 May 2006. (b) Chinese farmland image captured on 23 April 2007. (c) Binary ground truth for the Wetland Agriculture Dataset.
Figure 7. Wetland Agriculture Dataset with false color map (Bands 134, 90, and 75 as RGB). (a) Chinese farmland image captured on 3 May 2006. (b) Chinese farmland image captured on 23 April 2007. (c) Binary ground truth for the Wetland Agriculture Dataset.
Remotesensing 15 00246 g007
Figure 8. River Dataset with false color map (Band 134, 90, and 75 as RGB). (a) Chinese river image captured on 3 May 2013. (b) Chinese river image captured on 31 December 2013. (c) Binary ground truth for the River Dataset.
Figure 8. River Dataset with false color map (Band 134, 90, and 75 as RGB). (a) Chinese river image captured on 3 May 2013. (b) Chinese river image captured on 31 December 2013. (c) Binary ground truth for the River Dataset.
Remotesensing 15 00246 g008
Figure 9. Overall accuracy (a) and F1-score (b) comparison.
Figure 9. Overall accuracy (a) and F1-score (b) comparison.
Remotesensing 15 00246 g009
Figure 10. The change map results of different methods on the Irrigated Agriculture Dataset. (a) CVA. (b) SVM. (c) GETNET. (d) DSFANet. (e) WCRN. (f) Without FRL. (g) Without PS 2 A . (h) TFR-PS 2 ANet. (i) Ground truth.
Figure 10. The change map results of different methods on the Irrigated Agriculture Dataset. (a) CVA. (b) SVM. (c) GETNET. (d) DSFANet. (e) WCRN. (f) Without FRL. (g) Without PS 2 A . (h) TFR-PS 2 ANet. (i) Ground truth.
Remotesensing 15 00246 g010
Figure 11. The change map results of different methods on the Irrigated Agriculture Dataset. (a) CVA. (b) SVM. (c) GETNET. (d) DSFANet. (e) WCRN. (f) Without FRL. (g) Without PS 2 A . (h) TFR-PS 2 ANet. (i) Ground truth.
Figure 11. The change map results of different methods on the Irrigated Agriculture Dataset. (a) CVA. (b) SVM. (c) GETNET. (d) DSFANet. (e) WCRN. (f) Without FRL. (g) Without PS 2 A . (h) TFR-PS 2 ANet. (i) Ground truth.
Remotesensing 15 00246 g011
Figure 12. The change map results of different methods on the Irrigated Agriculture Dataset. (a) CVA. (b) SVM. (c) GETNET. (d) DSFANet. (e) WCRN. (f) Without FRL. (g) Without PS 2 A . (h) TFR-PS 2 ANet. (i) Ground truth.
Figure 12. The change map results of different methods on the Irrigated Agriculture Dataset. (a) CVA. (b) SVM. (c) GETNET. (d) DSFANet. (e) WCRN. (f) Without FRL. (g) Without PS 2 A . (h) TFR-PS 2 ANet. (i) Ground truth.
Remotesensing 15 00246 g012
Figure 13. Impacts of hyperparameters λ 1 and λ 2 on all three datasets. The first row indicates the influence on the accuracy, the second row the influence on the kappa coefficient, and the last row the influence on the F1-score.
Figure 13. Impacts of hyperparameters λ 1 and λ 2 on all three datasets. The first row indicates the influence on the accuracy, the second row the influence on the kappa coefficient, and the last row the influence on the F1-score.
Remotesensing 15 00246 g013
Table 1. Details of the selected datasets.
Table 1. Details of the selected datasets.
DatasetSpatial SizeBandDate 1Date 2Training RateTraining Samples
Irrigated 307 × 241 156May 1st, 2004May 8th, 20079.7%7250
Wetland 450 × 140 156May 3rd, 2006Apr. 23rd, 20079.7%6173
River 463 × 241 198May 3rd, 2013Dec. 31st, 20134.03%4500
Table 2. Model comparison results and module ablation study results on the Irrigated Agriculture Dataset (repeated 3 times).
Table 2. Model comparison results and module ablation study results on the Irrigated Agriculture Dataset (repeated 3 times).
ModelsACCKappaF1-ScorePrecisionRecall
CVA0.92860.77040.81270.99530.6867
SVM0.9614 ± 0.00210.8868 ± 0.01170.9754 ± 0.00080.9657 ± 0.01930.9854 ± 0.0184
GETNET0.9456 ± 0.00800.8466 ± 0.01720.9646 ± 0.00570.9690 ± 0.00910.9608 ± 0.0203
DSFANet0.8208 ± 0.00730.5449 ± 0.00230.6631 ± 0.01540.5754 ± 0.01140.7823 ± 0.0219
WCRN0.9113 ± 0.00600.7516 ± 0.01640.9422 ± 0.00390.9509 ± 0.00330.9336 ± 0.0045
Without FRL0.9652 ± 0.00540.9039 ± 0.01330.9772 ± 0.00370.9920 ± 0.00510.9629 ± 0.0122
Without PS 2 A 0.9709 ± 0.00140.9159 ± 0.00500.9813 ± 0.00090.9765 ± 0.00500.9862 ± 0.0033
TFR-PS 2 ANet0.9763 ± 0.00090.9324 ± 0.00280.9846 ± 0.00050.9862 ± 0.00290.9831 ± 0.0019
Table 3. Model comparison results and module ablation study results on the Wetland Agriculture Dataset (repeated 3 times).
Table 3. Model comparison results and module ablation study results on the Wetland Agriculture Dataset (repeated 3 times).
ModelsACCKappaF1-ScorePrecisionRecall
CVA0.95250.88590.91960.90320.9366
SVM0.9525 ± 0.00150.8851 ± 0.00450.9185 ± 0.00350.9150 ± 0.00720.9223 ± 0.0144
GETNET0.9543 ± 0.01280.8926 ± 0.02740.9253 ± 0.01770.8926 ± 0.05480.9644 ± 0.0255
DSFANet0.6043 ± 0.0025−0.1127 ± 0.00600.7452 ± 0.00320.6861 ± 0.00070.8157 ± 0.0088
WCRN0.9003 ± 0.03620.7643 ± 0.08020.8355 ± 0.05350.8170 ± 0.03540.8579 ± 0.0179
Without FRL0.9760 ± 0.00250.9419 ± 0.00630.9589 ± 0.00450.9538 ± 0.00260.9640 ± 0.0119
Without PS 2 A 0.9743 ± 0.00440.9380 ± 0.00160.9561 ± 0.00750.9490 ± 0.00750.9633 ± 0.0075
TFR-PS 2 ANet0.9827 ± 0.00040.9580 ± 0.00080.9701 ± 0.00050.9754 ± 0.00810.9648 ± 0.0072
Table 4. Model comparison results and module ablation study results on the River Dataset (repeated 3 times).
Table 4. Model comparison results and module ablation study results on the River Dataset (repeated 3 times).
ModelsACCKappaF1-ScorePrecisionRecall
CVA0.92800.66170.69920.54920.9617
SVM0.9198 ± 0.00080.5850 ± 0.04770.6278 ± 0.04690.5266 ± 0.00920.7876 ± 0.0624
GETNET0.9260 ± 0.01040.6508 ± 0.03140.6893 ± 0.02670.5467 ± 0.03900.9368 ± 0.0164
DSFANet0.8883 ± 0.00670.4610 ± 0.02530.5196 ± 0.00220.4154 ± 0.02120.6940 ± 0.0192
WCRN0.9284 ± 0.01110.6544 ± 0.03460.6919 ± 0.02950.5579 ± 0.02460.9158 ± 0.0175
Without FRL0.9519 ± 0.00190.7439 ± 0.00500.7699 ± 0.00410.6594 ± 0.03100.9250 ± 0.0143
Without PS 2 A 0.9566 ± 0.00870.7649 ± 0.00520.7884 ± 0.00370.6850 ± 0.04890.9286 ± 0.0081
TFR-PS 2 ANet0.9602 ± 0.00030.7803 ± 0.00820.8019 ± 0.00740.7068 ± 0.00380.9267 ± 0.0134
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Y.; Zhang, L.; Huang, C.; Qi, W.; Song, R. Parallel Spectral–Spatial Attention Network with Feature Redistribution Loss for Hyperspectral Change Detection. Remote Sens. 2023, 15, 246. https://doi.org/10.3390/rs15010246

AMA Style

Huang Y, Zhang L, Huang C, Qi W, Song R. Parallel Spectral–Spatial Attention Network with Feature Redistribution Loss for Hyperspectral Change Detection. Remote Sensing. 2023; 15(1):246. https://doi.org/10.3390/rs15010246

Chicago/Turabian Style

Huang, Yixiang, Lifu Zhang, Changping Huang, Wenchao Qi, and Ruoxi Song. 2023. "Parallel Spectral–Spatial Attention Network with Feature Redistribution Loss for Hyperspectral Change Detection" Remote Sensing 15, no. 1: 246. https://doi.org/10.3390/rs15010246

APA Style

Huang, Y., Zhang, L., Huang, C., Qi, W., & Song, R. (2023). Parallel Spectral–Spatial Attention Network with Feature Redistribution Loss for Hyperspectral Change Detection. Remote Sensing, 15(1), 246. https://doi.org/10.3390/rs15010246

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop