Next Article in Journal
Designing an Intelligent Virtual Educational System to Improve the Efficiency of Primary Education in Developing Countries
Next Article in Special Issue
Graph-Embedded Online Learning for Cell Detection and Tumour Proportion Score Estimation
Previous Article in Journal
Design of the Buck Converter without Inductor Current Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised and Self-Supervised Tensor Train for Change Detection in Multitemporal Hyperspectral Images

School of Computer Science and Technology, Donghua University, Shanghai 201620, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(9), 1486; https://doi.org/10.3390/electronics11091486
Submission received: 1 April 2022 / Revised: 2 May 2022 / Accepted: 3 May 2022 / Published: 6 May 2022
(This article belongs to the Special Issue Deep Learning for Big Data Processing)

Abstract

:
Remote sensing change detection (CD) using multitemporal hyperspectral images (HSIs) provides detailed information on spectral–spatial changes and is useful in a variety of applications such as environmental monitoring, urban planning, and disaster detection. However, the high dimensionality and low spatial resolution of HSIs do not only lead to expensive computation but also bring about inter-class homogeneity and inner-class heterogeneity. Meanwhile, labeled samples are difficult to obtain in reality as field investigation is expensive, which limits the application of supervised CD methods. In this paper, two algorithms for CD based on the tensor train (TT) decomposition are proposed and are called the unsupervised tensor train (UTT) and self-supervised tensor train (STT). TT uses a well-balanced matricization strategy to capture global correlations from tensors and can therefore effectively extract low-rank discriminative features, so the curse of the dimensionality and spectral variability of HSIs can be overcome. In addition, the two proposed methods are based on unsupervised and self-supervised learning, where no manual annotations are needed. Meanwhile, the ket-augmentation (KA) scheme is used to transform the low-order tensor into a high-order tensor while keeping the total number of entries the same. Therefore, high-order features with richer texture can be extracted without increasing computational complexity. Experimental results on four benchmark datasets show that the proposed methods outperformed their tensor counterpart, the tucker decomposition (TD), the higher-order singular value decomposition (HOSVD), and some other state-of-the-art approaches. For the Yancheng dataset, OA and KAPPA of UTT reached as high as 98.11% and 0.9536, respectively, while OA and KAPPA of STT were at 98.20% and 0.9561, respectively.

1. Introduction

The ability to comprehend global change in its entirety is essential for giving timely and accurate information about the Earth [1,2,3,4,5]. Multitemporal remote sensing images are currently used in a wide range of change detection applications, including ecological change studies [6], natural disaster investigations [7], and especially in the urban development race [8]. Meanwhile, the high spectral resolution of hyperspectral images (HSIs) has attracted the attention of researchers in the CD community as compared to synthetic aperture radar (SAR) images and multispectral images. It is because subtler variations can be detected via HSIs with the detailed composition of various reflected objects. The abundant spectral information contained in HSIs have been made full use of in a lot of HSI CD methods [9,10]. Moreover, CD can also be regarded as a classification task where different types of changes, including changes in the background, are considered as different classes [11,12]. With the rapid developments in deep learning [13,14,15,16], satisfactory results in HSI CD have also been achieved.
However, there still exist some challenging problems in HSI CD. Apart from the useful change information, there are also a lot of redundant information contained in HSIs which may lead to pseudo changes. Meanwhile, the high dimensionality of HSIs increases the computational complexity, and the low spatial resolution can bring about inter-class homogeneity and inner-class heterogeneity. Therefore, some unsupervised methods such as K-Means [17] cannot precisely differentiate between a region that has undergone change and one which has not. Tensor-based methods [18,19,20] treat HSIs as three-order tensors, with two dimensions for spatial and one for spectral, and make use of tensor algebra [21,22] to extract low-rank discriminative features from the original HSIs. As the density matrices [23] that represent the quantum states are two-order tensors in nature, the effectiveness of tensor algebra in extracting discriminative features can be confirmed by virtue of matricization and quantum information theory [23]. Another challenge in HSI CD comes from the expensive cost of obtaining manual annotations, which not only requires high-level geographical expertise but also takes a lot of time. Therefore, supervised learning methods are not practical in the field of HSI CD, where accurate ground truth is necessary.
To tackle the above-mentioned problems, this study proposes two novel tensor train (TT)-based techniques for multitemporal HSI CD. The first one is an unsupervised TT-based technique called UTT. The second one is a self-supervised TT-based technique named STT. The motivations are as follows. First, tensor algebra is well-suited for HSI processing as HSIs are three-order tensors in nature. TT decomposition is one type of low-rank tensor decomposition [24] that can effectively remove the redundancy and overcome the curse of dimensionality. Second, TT can capture more global information between temporal images than TD as TT is based on a well-balanced matricization scheme [25] (k-modes versus the rest), while TD is based on an unbalanced matricization scheme (one mode versus the rest). Therefore, the features extracted by TT contain more changed information and are thus more discriminative. Third, manual annotations are difficult to obtain as field investigation is expensive and high-level geographical expertise is necessary. The proposed UTT and STT are based on unsupervised and self-supervised learning, respectively, which indicates that laborious manual annotations are not needed. Additionally, the ket-augmentation (KA) [26] scheme can obtain higher-order tensor representations of change features while retaining the total number of entries. This means that high-order rich texture features can be obtained without increasing computational complexity. In addition, TT decomposition is more efficient for the tensor augmented by KA because the local structure of the data can be exploited effectively in terms of computational resources [25].
The proposed UTT and STT work as follows. High-order difference tensors are firstly produced using KA and substraction for both UTT and STT. Then, UTT directly employs low TT-rank optimization to reconstruct the difference tensor, while STT combines clustering and classification in a self-supervised manner. Classification is used as the pretext task in order to make the features learnt friendlier to binary clustering, which is the main task. Clustering centroids and the classification network are optimized jointly in STT. For STT, clustering is used to produce the initial pseudo labels and divide the learnt features into the categories of changed or unchanged, while TT is used to extract low-rank features and perform pseudo classification. Both the UTT and STT approaches are followed by binary clustering in order to categorize the changed and unchanged features.
The novelty and contributions of this study can be summarized as follows:
(1).
Inspired by the knowledge from quantum information theory, this work theoretically proves that TT decomposition exhibits greater ability than the traditional TD in capturing global correlations between changed and unchanged tensor entries. Thus, TT is used to extract spectral–spatial low-rank features for multitemporal HSI CD, which decomposes a high-order tensor into a set of low-order tensors by approximating the optimal TT rank.
(2).
KA is used to obtain higher-order tensor representations of changed features. This technique leverages the representation of changed features and provides discriminative information for CD while retaining the total number of entries.
(3).
Two novel TT models that do not require manual annotations are proposed for CD. In the first model, UTT bypasses SVD, which is a usual but computationally expensive algorithm for optimization, in order to extract changed and unchanged features. In the second one, STT leverages pseudo clustering labels to train an accurate change classifier built on TT. Experimental results show that STT is more accurate than UTT, while UTT is more efficient. Moreover, they both outperform state-of-the-art models upon comparison.
The remainder of this work is structured in the following manner. Closely related works are briefly reviewed in Section 2. The background knowledge is introduced in Section 3. In Section 4, proof of the superiority of TT and details of the algorithm for the suggested approaches as well as their implementation are presented. Experiments on four real-world datasets are conducted in Section 5, and the conclusion is discussed in Section 6.

2. Related Works

This section gives a brief review of some closely related works, including change detection in multitemporal hyperspectral images, self-supervision for image analysis, and tensor analysis.

2.1. Change Detection in Multitemporal Hyperspectral Images

Change detection methods can be mainly divided into four categories according to the ways of extracting the changed information.
Firstly, algebra-based methods [18,27,28,29] leverage algebra operations to extract the changed information. These methods are based on an assumption that changes can be reflected in the difference between corresponding pixel values. Absolute distance (AD) and Euclidian distance (ED) [27] consider each pixel as a vector and calculate the absolute distance and the Euclidian distance between each pixel as the changed information. For these methods, the detection accuracy highly relies on the accuracy of radiometric and geometric correction results.
Secondly, transformation-based methods [30,31,32] convert the original remote sensing data into another feature space. Then, the CD result is obtained in the new space. In local subspace-based change detection (LSCD) and adaptive subspace-based change detection (ASCD) [30], background subspace is constructed, and the change detection result can be achieved by calculating the subspace distance. The new feature space is more discriminative and is able to differentiate between changed and unchanged. However, some spectrum information is unavoidably lost during the process of transformation in the methods of the second category.
Thirdly, classification-based methods [11,12] categorize the HSIs obtained at different stages separately, and the change detection results can be achieved by comparing the classification results. In [11], the support vector machine (SVM) was used to perform classification independently, and the final CD map could be achieved via post-classification fusion. Since two classification results were obtained separately, environmental factors during the process of HSI acquisition, such as the atmosphere, could be eliminated. However, CD results rely highly on the classification results in this category.
Fourthly, deep learning-based methods [13,14,15,16] have become increasingly prosperous with the development of deep learning. Dual attentive fully convolutional Siamese networks (DASNet) [13] and the change detection generative adversarial network (CDGAN) [14], which use the dual attention mechanism and GAN, respectively, have both achieved superior change detection results. Deep models have strong abilities in feature representation and are therefore suitable for solving complex tasks such as HSI CD.
Other methods [10,33,34,35,36,37] such as spectral unmixing [10,34,35,36,37] have also achieved great change detection results in HSI CD. In [34], subpixel information was fully leveraged for HSI CD via multi-level spectral unmixing. However, all the mentioned methods above only leverage the spectral information instead of spectral and spatial information together due to the flattening operation they use. This means that the inherent spatial structure information is inevitably lost.

2.2. Self-Supervision for Image Analysis

Self-supervised learning methods, as a subset of unsupervised learning methods, aim to avoid the extensive costs of annotating large-scale datasets. It includes a variety of pretext tasks such as image classification [38], colorizing grayscale images [39], image inpainting [40], and image jigsaw puzzle [41]. Features can be learnt in pretext tasks, where pseudo labels are generated automatically; thus, the learnt features perform better on the main task where no annotations are available. Additionally, the features learnt can also be evaluated in a variety of downstream tasks, including semantic segmentation [42], object detection [43], and so on.
In view of the difficulty of acquiring a lot of annotated samples and the requirements for more discriminative features, the self-supervised have won researchers’ attention in the field of CD. For example, in [31], deep neural networks and unsupervised K-Means were combined to learn Gaussian-distributed difference representations, and then the learnt representations were used to detect multiple types of changes. In [44], a multiscale self-attention deep clustering technique was proposed, which combined the convolutional neural network (CNN) with K-Means for remote sensing images CD. In [38], a self-supervised tensor network SSTN composed of pre-train and fine-tune stages was proposed for HSI CD. In [35], an image differencing algorithm and spectral unmixing manner were combined to generate pseudo training data; consequently, a more accurate HSI change detector could be obtained. These self-supervised learning approaches demonstrate how the data structure is used to give supervisory signals in order to learn changed features from temporal images. The self-supervised methods use temporal prediction to represent change features that are more consistent and discriminative than when temporal differences are directly computed. This results in better changed features, with altered areas considerably enhanced and unchanged areas suppressed [45]. This also makes the analysis for the final change map less complex.

2.3. Tensor Analysis

As is already known, HSIs are three-order tensors in nature, with two orders for spatial and the rest for spectral. Tensor analysis and multilinear algebra [22], which can capture the structural characteristics of high dimensional tensors, are therefore suitable for hyperspectral image processing. Tensor decomposition [21], e.g., Tucker decomposition, tensor train, tensor ring, is one of the most significant methods in tensor analysis and has already found applications in dimension reduction [46], target detection [47], anomaly detection [48], hyperspectral images classification [49], etc.
There have also been attempts to approach CD in multitemporal HSIs based on tensor decomposition [18,19,20]. In [20], the tensor-based approach using 4-D HOSVD was proposed, in which bitemporal HSIs were stacked to extract changed features using Tucker decomposition (TD). In [18], a three-order Tucker decomposition and reconstruction detector (TDRD) was proposed for HSI CD, where unpurified bitemporal HSIs were reconstructed using TD before the changed features were extracted. Nevertheless, TD based on SVD provides stable (i.e., non-iterative) decomposition and optimal approximations regarding their matrix versions due to the orthonormal and diagonal characteristics of the factor matrices. However, the non-diagonality of the tensor core confining the tucker decomposition to ensure optimality at the tensor level results in a lack of changed feature extraction in higher-order tensors [50]. Similarly, components of the Tucker rank are ranks of matrices generated using an unbalanced matricization approach, which is a conceptual problem (one mode versus the rest), ignores the global correlation between changed/unchanged features in bitemporal images. Furthermore, matrix rank approximation is only effective when the matrix is more balanced [25]. TT is another type of tensor decomposition and can capture the global correlation due to its well-balanced matricization scheme. Therefore, TT is promising for the acquisition of more discriminative low-rank features and thus achieves greater change detection results in HSI CD.

3. Background Knowledge

This section gives some background knowledge necessary to comprehend the following contents, including tensor operations, tensor train (TT) decomposition, and quantum information theory.

3.1. Tensor Operations

Basic notations and operations in tensor algebra that are closely related to our work are introduced here. To better characterize the following framework, graphical representation is accepted, as shown in Figure 1. Basic blocks such as scalar (0-order tensor), vector (1-order tensor), matrix (2-order tensor), and high-order tensors are shown in Figure 1a.
Let G ¯ I 1 × I 2 × × I N be an N-order tensor, where N represents the number of dimensions or modes and I n n = 1 , 2 , , N represents the dimension of the n-th mode. An entry or element of G ¯ can be denoted by ( G ¯ ) i 1 , i 2 , , i N , where i n = 1 , 2 , , I n and n = 1 , 2 , , N .
Matricization is also termed as the unfolding or flattening of a tensor. It has two different forms, namely mode-k and mode-(1, 2, …, k) matricization for TD and TT, respectively. Mode-k and mode-(1, 2, …, k) matricization transforms a tensor G ¯ I 1 × I 2 × × I N into X ( k ) m 1 × n 1 and X [ k ] m 2 × n 2 , where m 1 = I k , n 1 = j = 1 , j k N I j , m 2 = j = 1 k I j , and n 2 = j = k + 1 N I j . The correspondence of elements between the original tensor and the matrix after matricization can be expressed as follows:
X ( k ) i k , i 1 i k 1 i k + 1 i N ¯ = G ¯ i 1 , i 2 , , i N ,
X [ k ] i 1 i k ¯ , i k + 1 i N ¯ = G ¯ i 1 , i 2 , , i N .
A multi-index is an index that combines all values of indices following a specific order and can be calculated as follows:
i 1 i N ¯ = i N + ( i N 1 1 ) I N + ( i N 2 1 ) I N I N 1 + + ( i 1 1 ) i = N 2 I i .
The vectorization of the tensor G ¯ is represented by vec( G ¯ ), which is operated on the mode-1 unfolded matrix X ( 1 ) as:
b = vec G ¯ = vec X ( 1 ) k = 1 N I k .
The graphical representations of mode-k matricization, mode-(1, 2, …, k) matricization, and vectorization can be clearly seen in Figure 1b.

3.2. Tensor Train (TT) Decomposition

A tensor train decomposition represents an N-order tensor G ¯ I 1 × I 2 × × I N with N 3-order tensors [24], and each element of G ¯ can be expressed as follows:
G ¯ i 1 , , i N = X ¯ 1 : , i 1 , : X ¯ 2 : , i 2 , : X ¯ N : , i N , : ,
where X ¯ k r [ k 1 ] × I k × r [ k ] is the k-th core tensor with r [ 0 ] = r [ N ] = 1 . r [ 0 ] , r [ 1 ] , , r [ N ] are called TT ranks.
In index form, Equation (5) can be written as follows:
G ¯ i 1 , , i N = α 0 , , α N ( X ¯ 1 ) α 0 , i 1 , α 1 ( X ¯ 2 ) α 1 , i 2 , α 2 ( X ¯ N ) α N 1 , i N , α N .
Figure 2 shows a typical representation of a 5-order TT decomposition in graphical notations. There exist some applications, such as the low-rank tensor completion (LRTC) [25], which make full use of the ability of low-rank features extraction of TT. Their process of extracting low-rank features can be summarized as a low TT-rank optimization problem. One solution to solve this problem is by using singular value decomposition (SVD) [21], which can be called TT-SVD. Another way that uses the multilinear matrix factorization model [25] to approximate the TT rank of a tensor can bypass the computationally expensive SVD; this is called a TT-noSVD.

3.3. Quantum Information Theory

Quantum states can be represented using density matrices [23], which are naturally 2-order tensors. Every tensor corresponds to a quantum state if the matricization operation is used. Therefore, it is promising to validate the effectiveness of tensor analysis using the knowledge in the field of quantum information theory as it can help to analyze the ability of different tensor methods in capturing useful information.
The information in the quantum field is measured by the von Neumann entropy [23], which is an extension of the Shannon entropy in the classical field. The von Neumann entropy is defined as follows:
S ( ρ ) = tr ( ρ lo g 2 ( ρ ) ) ,
where ρ and tr ( ) represent the quantum state and the trace operation, respectively.
The correlation between two subsystems, H A m and H B n , of the composite system H A B m × n can be quantified by quantum mutual information [23]:
I ( A : B ) = S ( ρ A ) + S ( ρ B ) S ( ρ A B ) ,
where ρ A , ρ B , and ρ A B represent the quantum states of the corresponding systems H A , H B , and H A B . The quantum mutual information can also be considered as a measure of global correlation [51].

4. Methodology

In this section, the proof for the advantages of TT over TD is firstly shown by virtue of quantum information theory. Then, two TT-based methods, UTT and STT are proposed for CD. The first one performs unsupervised CD using multilinear matrix factorization to approximate the TT rank, as shown in Figure 3, and the second one performs self-supervised CD using TT decomposition, as shown in Figure 4. The proposed methods can yield accurate CD results without manual annotations.

4.1. Ability of Tensor Train in Capturing the Global Correlation

A core challenge in the field of HSI CD is that the features extracted cannot be discriminative enough. It is because the features may not contain enough changed information. Therefore, it is crucial to ensure the ability of the CD methods in capturing changed information (or correlation in quantum information theory). As density matrices representing the quantum states are 2-order tensors in nature, matricization operation and quantum information theory are combined here to compare the ability of TT and TD in capturing the changed information of the tensors; the complete proof is shown below.
For any given matrix X m × n , if SVD is applied, then X = W V T can be produced, where W m × m , m × n , and V n × n . Matrices W and V consist of eigenvectors of X X T m × m and X T X n × n , respectively. Therefore, X X T and X T X contain significant information from the dimensions m and n of X , respectively. In the field of quantum theory, X X T and X T X can represent density matrices [23] in two subspaces H A m and H B n , respectively. The density matrix is one of the forms representing a quantum state. To correspond to the notation in quantum physics, ρ A and ρ B are used to replace X X T and X T X in the following statements.
According to the Schmidt decomposition [23] and the fact that X X T and X T X have the same nonvanishing eigenvalues, there exist orthonormal bases { i A } in H A and { i B } in H B , such that:
ρ A = i = 1 r k λ i 2 i A i A ,
ρ B = i = 1 r k λ i 2 i B i B ,
where r k is the rank of X , λ i i = 1 , , r k are nonvanishing singular values satisfying i = 1 r k λ i 2 = 1 . Then, the density matrix of a composite system AB in the space H A B m × n can be represented as follows:
ρ A B = i = 1 r k λ i 2 i A i A i B i B ,
where represents the tensor product [23] and H A B is the tensor product of two subspaces H A m and H B n .
The global correlation between two subsystems A and B can be studied via quantum mutual information, which is defined in Equation (8). Substituting Equation (9) into Equation (7), the von Neumann entropy of ρ A can be calculated as follows:
S ( ρ A ) = tr ( ρ A log 2 ( ρ A ) ) = i = 1 r k λ i 2 log 2 λ i 2 .
Similarly, it can be obtained that:
S ( ρ B ) = S ( ρ A B ) = i = 1 r k λ i 2 log 2 λ i 2 ,
Simply, S ( ρ A ) = S ( ρ B ) = S ( ρ A B ) = S , where:
S = i = 1 r k λ i 2 log 2 λ i 2 .
With S substituted into Equation (8), it can be achieved that the global correlation between two subsystems H A m and H B n is equal to S. According to the quantum relative entropy and Klein inequality [23], it is shown in Equation (15) that S is bounded by 0 S log 2 r k , where I represents the identity matrix:
S ( ρ | | I / r k ) = tr ( ρ log 2 ( ρ ) ) tr ( ρ log 2 ( I / r k ) ) = S i = 1 r k λ l i 2 og 2 ( 1 / r k ) = S + log 2 r k 0 .
Here, it is assumed that I 1 = I 2 = = I N = I for simplicity. For the mode-k matricization of G ¯ I 1 × I 2 × × I N from TD, X ( k ) m 1 × n 1 can be obtained, where m 1 = I k and n 1 = j = 1 , j k N I j . Therefore, r ( k ) , which represents the rank of X ( k ) , satisfies the following:
r ( k ) min ( I k , j = 1 , j k N I j ) = I .
Then, for the mode-(1, 2, …, k) matricization from TT, X [ k ] m 2 × n 2 can be obtained, where m 2 = j = 1 k I j and n 2 = j = k + 1 N I j . Therefore, r [ k ] , which represents the rank of X [ k ] , satisfies the following equation:
r [ k ] min ( j = 1 k I j , j = k + 1 N I j ) = I min ( k , N k ) .
From the analysis above, it can be seen that the upper bound of r [ k ] in TT is far greater than its counterpart r ( k ) in TD for k = 2 , 3 , , N 2 and is equivalent to it only when k is equal to 1 or N-1. Therefore, the conclusion that the upper bound of the global correlation that TT can capture is far greater than its counterpart TD in the vast majority of cases can be drawn from Equation (15). Because TT can capture more changed information which TD cannot, the superiority of TT in HSI CD is confirmed.

4.2. Unsupervised Tensor Train for Change Detection

Due to the effective changed information extraction ability of TT, it is used in the proposed algorithms to extract discriminative features. The first proposed algorithm leverages unsupervised learning to avoid the complex and time-consuming manual annotation process. The proposed UTT is composed of three parts: (1) difference image generation, (2) TT decomposition for spectral–spatial low-rank feature extraction, and (3) binary clustering to distinguish between changed and unchanged regions via the features extracted by the TT decomposition.
Difference image generation consists of two steps, KA and substraction. Let Y ¯ 1 J 1 × J 2 × J 3 and Y ¯ 2 J 1 × J 2 × J 3 be a set of bitemporal HSIs acquired at two different times, t 1 and t 2 , respectively, where J 1 is the number of rows, J 2 is the number of columns, and J 3 is the number of bands. They can be represented by a set of orthonormal bases { y j k ( k ) J k } k = 1 , 2 , 3 , as follows:
Y ¯ t = j 1 , j 2 , j 3 ( Y ¯ t ) j 1 j 2 j 3 y j 1 ( 1 ) y j 2 ( 2 ) y j 3 ( 3 ) ,
where t = 1 , 2 and represents outer product. To obtain high-order rich texture features [26] and make full use of the local structure of the data in terms of computational resources [25], KA is applied to three-dimensional HSIs Y ¯ 1 and Y ¯ 2 . This process does not increase the computational complexity as the total number of elements does not change. Then, two five-dimensional tensors, A ¯ 1 I 1 × I 2 × I 3 × I 4 × I 5 and A ¯ 2 I 1 × I 2 × I 3 × I 4 × I 5 , are obtained:
A ¯ t = i 1 , i 2 , i 3 , i 4 , i 5 ( A ¯ t ) i 1 i 2 i 3 i 4 i 5 e i 1 ( 1 ) e i 2 ( 2 ) e i 3 ( 3 ) e i 4 ( 4 ) e i 5 ( 5 ) ,
where t = 1 , 2 , I 1 I 2 I 3 I 4 I 5 = J 1 J 2 J 3 , and { e i k ( k ) I k } k = 1 , 2 , 3 , 4 , 5 represents a set of orthonormal bases in the corresponding Hilbert space. Then, substraction is conducted between A ¯ 1 and A ¯ 2 to get the high-order difference tensor:
G ¯ = A ¯ 1 A ¯ 2 s . t G ¯ , A ¯ 1 , A ¯ 2 I 1 × I 2 × I 3 × I 4 × I 5 ,
where I k is the corresponding mode-k dimension for k = 1 , , 5 .
Taking into account the intrinsic complexity and structure of HSI, TT is used for change detection in HSI as an unsupervised technique that extracts changed and unchanged features efficiently. As proved in Section 4.1, more global correlations can be captured with TT as compared to its tensor counterpart TD. This means that TT can extract changed information more effectively. The process of the spectral–spatial low-rank feature extraction of the generated high-order difference tensor G ¯ can be conducted via the following process of TT rank optimization:
min X [ k ] k = 1 N 1 α k rank ( X [ k ] ) s . t X ¯ Ω = G ¯ Ω ,
where α k represents the weight that the TT rank of the matrix X [ k ] contributes to, with the constraint k = 1 N 1 α k = 1 , and Ω represents a given index set. After the process of low TT rank feature extraction, the reconstructed tensor X ¯ obtained by Equation (21) is sent to K-Means to obtain the final CD results T J 1 × J 2 . The pseudocode and framework of the proposed UTT are summarized in Algorithm 1 and Figure 3, respectively. Different groups of variables are alternatively optimized using the block coordinate descent (BCD) method [25], and the update of X ¯ in each loop is given by Equation (22).
Algorithm 1. Pseudocode of the proposed UTT for multitemporal HSI change detection.
Input: Observed data G ¯ I 1 × I 2 × I 3 × I 4 × I 5 , index set Ω
Parameters: α i , r i , i = 1 , , N 1
Initialization: Initialize U 0 , V 0 , X ¯ 0 , with X ¯ Ω 0 = G ¯ Ω , l = 0
While not converged do:
   for k = 1 to N-1 do
      Unfold tensor G ¯ l to get X [ k ] l
       U k l + 1 = X [ k ] l ( V k l ) T ( V k l ( V k l ) T )
       V k l + 1 = ( ( U k l + 1 ) T U k l + 1 ) ( U k l + 1 ) T X [ k ] l
       X [ k ] l + 1 = U k l + 1 V k l + 1
   end for
   Update tensor X ¯ l + 1 using Equation (22)
End while
Apply K-Means to the reconstructed tensor X ¯
Output: Change detection results T
X ¯ i 1 i 2 i 3 i 4 i 5 l + 1 = ( k = 1 N 1 α k fold ( X [ k ] l + 1 ) ) i 1 i 2 i 3 i 4 i 5 , ( i 1 , i 2 , i 3 , i 4 , i 5 ) Ω G ¯ i 1 i 2 i 3 i 4 i 5 ,   ( i 1 , i 2 , i 3 , i 4 , i 5 ) Ω ,
where folding is the inverse process of unfolding, i.e., matricization in Equation (2).
The specific process of optimizing Equation (21) in UTT is described as follows. Factorization model X [ k ] = U V is leveraged to minimize the Frobenius norm, where X [ k ] m × n is a matrix of rank r [ k ] , U m × r [ k ] and V r [ k ] × n .
Equation (21) can be re-modeled as follows:
min U k , V k , X [ k ] k = 1 N 1 α k 2 U k V k X [ k ] F 2 s . t . X ¯ Ω = G ¯ Ω ,
Where U k j = 1 k I j × r [ k ] and V k r [ k ] × j = k + 1 N I j . This problem is now convex when each variable U k , V k , and X [ k ] is modified while keeping the other two fixed. Therefore, the variables are updated alternatively following Equations (24)–(26):
U k l + 1 = X [ k ] l ( V k l ) T ( V k l ( V k l ) T ) ,
V k l + 1 = ( ( U k l + 1 ) T U k l + 1 ) ( U k l + 1 ) T X [ k ] l ,
X [ k ] l + 1 = U k l + 1 V k l + 1 ,
where denotes the Moore–Penrose pseudoinverse. After updating U k , V k , and X [ k ] , elements of the tensor X ¯ l + 1 can be computed with Equation (22).
The pseudocode is summarized in Algorithm 1. The main benefit of this approach is that it bypasses the SVD, which saves a lot of time in the computation. However, since Equation (21) can also be optimized using SVD [21,25], UTT-SVD and UTT-noSVD are used here to differentiate whether SVD is used in the optimization process.

4.3. Self-Supervised Tensor Train for Change Detection

STT uses self-supervised learning to avoid the complex and time-consuming manual annotation process. Pseudo classification is used as the pretext task to learn features that are friendly to binary clustering. K-Means is used both to generate the initial pseudo labels and divide the learnt features into the categories of changed or unchanged, while the classification network is typically composed of two kinds of layers, i.e., the TT decomposition (TTD) layers and the TT output (TTO) layers.
The input of STT is denoted by an N-order difference tensor G ¯ = Y ¯ ( 0 ) I 1 × I 2 × × I N , I i = I i ( 0 ) ( i = 1 , 2 , , N ) . It can be obtained by difference image generation defined in UTT. For TTD layers, features extracted from the input samples can be defined as follows:
Y ¯ j 1 , , j N ( l ) = σ i 1 , , i N I 1 ( l 1 ) , , I N ( l 1 ) W ¯ j 1 , , j N , i 1 , , i N ( l ) Y ¯ i 1 , , i N ( l 1 ) + B ¯ j 1 , , j N ( l ) ,
where Y ¯ ( l ) I 1 ( l ) × I 2 ( l ) × × I N ( l ) ( l = 1 , 2 , , L ) is the output of the l-th TTD layer, and it can be treated as a primary feature tensor, L is the number of TTD layers, σ is the ReLU activation function, and θ ( l ) = { W ¯ ( l ) , B ¯ ( l ) } is the parameter set of the l-th TTD layer, where W ¯ ( l ) I 1 ( l ) × × I N ( l ) × I 1 ( l 1 ) × × I N ( l 1 ) and B ¯ ( l ) I 1 ( l ) × × I N ( l ) .
The coefficient tensor W ¯ ( l ) of Equation (27) in the l-th TTD layer is decomposed into N three-order tensors X ¯ k ( l ) r [ k 1 ] ( l ) × I k ( l 1 ) I k ( l ) × r [ k ] ( l ) [52] for k = 1 , 2 , , N , while { r [ k ] ( l ) } k = 0 , 1 , , N represents the set of TT ranks in the l-th TTD layer with r [ 0 ] ( l ) = r [ N ] ( l ) = 1 , as shown in Equation (5) and Figure 4b. This is so that low-rank features can be efficiently extracted, and the number of parameters can be significantly compressed. Each element in W ¯ ( l ) can be represented equivalently by the following matrix product:
W ¯ ( l ) j 1 , , j N , i 1 , , i N = X ¯ 1 ( l ) : , j 1 i 1 ¯ , : X ¯ 2 ( l ) : , j 2 i 2 ¯ , : X ¯ N ( l ) : , j N i N ¯ , : ,
where X ¯ k ( l ) : , j k i k ¯ , : r [ k 1 ] ( l ) × r [ k ] ( l ) is the lateral slice of the tensor X ¯ k ( l ) for k = 1 , 2 , , N .
Applying the TT format for the weight subtensor in Equation (27), the l-th TTD layer can be redefined as follows:
Y ¯ j 1 , , j N ( l ) = σ i 1 , , i N I 1 ( l 1 ) , , I N ( l 1 ) X ¯ 1 ( l ) : , j 1 i 1 ¯ , : X ¯ 2 ( l ) : , j 2 i 2 ¯ , : X ¯ N ( l ) : , j N i N ¯ , : Y ¯ i 1 , , i N ( l 1 ) + B ¯ j 1 , , j N ( l ) .
In our STT model, N = 5 is used so that the coefficient tensors are decomposed into five three-order tensors.
For the TTO layers, the output of the last TTD layer Y ¯ ( L ) is firstly vectorized according to Equation (4), and the output features are calculated as follows:
z = f ϑ ( G ¯ ) = v e c ( Y ¯ ( L ) ) d ,
where d = I 1 ( L ) I 2 ( L ) I N ( L ) is the dimension of the features. This can be seen clearly in Figure 4c. Afterward, a fully connected layer is employed to f ϑ ( G ¯ ) to compute the posterior output y C 1 for the classification problem over C 1 pseudo class labels corresponding to the input tensor G ¯ as follows:
y = s ( f ϑ ( G ¯ ) × 1 ω ) = s ω ( f ϑ ( G ¯ ) ) ,
where s ( ) represents the softmax activation function, and ω d × C 1 is the weight of the fully connected layer.
For the training set { G ¯ i } i = 1 , 2 , , P , which represents the high-order difference tensor generated from each patch, P is the number of the training samples. STT achieves self-learning through the joint training of TTD/TTO layers for pseudo-label classification and clustering layers for binary clustering. Pseudo-labels are generated by implementing an unsupervised method, i.e., K-Means, to the set of high-order difference tensors { G ¯ i } i = 1 , 2 , , P . The cross-entropy classification loss function is defined with regard to the parameter set ζ = θ ( l ) , ω as:
l c l f ( ζ ; { G ¯ i } ) = 1 P i = 1 P y ^ i c l f log s ω ( f ϑ ( G ¯ i ) ) ,
where { y ^ i clf C 1 } i = 1 , 2 , , P represents the reference pseudo labels.
Then, with respect to the part of clustering, the features computed by Equation (30) in the TTO layer are clustered by the binary clustering module to update the original centroid matrix M p t c d × C 2 for the changed cluster and the unchanged one, where C 2 = 2 , as illustrated in Figure 4. The difference between the output features of STT in Equation (30) and the clustering results is used as the other part of the self-learning loss for the optimization of STT, which is computed as follows:
l c l t = 1 P i = 1 P f ϑ ( G ¯ i ) M o p t c y ^ i o p t F 2 s . t y ^ j , i o p t 0 , 1 , 1 T y ^ i o p t = 1 , i , j ,
where M o p t c d × C 2 is an updated centroid matrix with initial values M o p t c = M p t c and y ^ o p t = y ^ p t . Afterwards, the new centroid can be updated as [17]:
M k o p t c = M k o p t c M k o p t c f ϑ ( G ¯ i ) n k , i y ^ k , i o p t s . t k = 1 , 2 , C 2 ,
where n k , i is the number of times the algorithm allocated a sample to cluster k before handling the incoming sample G ¯ i , and 1 / n k , i can be used to regulate the learning rate as a gradient step size.
In the STT model, the pseudo classification and the binary clustering process are performed alternatively. The classification loss and cluster loss are combined as the total network loss:
L ( ζ ; { G ¯ i } ) = l c l t + μ l c l f ,
where μ is a hyper-parameter to constrain the balance between classification and clustering. The STT learns low-rank changed features in a self-supervised manner using the pre-clustering pseudo labels and the binary clustering layers throughout the training process.
For STT training, the Adam optimizer [53] is used to minimize the loss function with regard to the parameter set ζ = { W ¯ ( l ) , B ¯ ( l ) , ω } . The STT parameters are learnt in a better way by computing the gradient of loss function L ( ζ ; { G ¯ i } ) directly, with reference to the cores of the TT-representation of { W ¯ ( l ) } l = 1 , 2 , , L [52].
The backpropagation (BP) method is used to update the whole parameter set ζ = { θ ( l ) , ω } during the training process. Firstly, BP starts to calculate the gradients of the loss function with reference to the parameters in the output layer, which is ω in our case. Then, the BP method is applied to calculate the gradients of the parameter set from the TTO layer to the first TTD layer using the chain rule. After the gradients are calculated, the Adam optimizer is used to update them at the same time.
In the testing phase of the STT model, the optimized and updated parameters ζ = { W ¯ ( l ) , B ¯ ( l ) , ω } are used on the samples generated from every pixel to obtain the binary change detection map T J 1 × J 2 . The optimization algorithm for STT is summarized in Algorithm 2. The maximum number of epochs is specified so that iterations can be halted for practical reasons. The framework of the proposed STT is summarized in Figure 4.
TT decomposition overcomes the curse of dimensionality by reducing the number of network parameters significantly as compared to the tucker decomposition. Let I = max I 1 ( l ) , I 2 ( l ) , , I N ( l ) be the maximum dimension and r = max r [ 0 ] ( l ) , r [ 1 ] ( l ) , , r [ N ] ( l ) be the maximum TT rank, then the storage complexity of each weight subtensor is O ( r N + N I r ) in the Tucker format, and the storage complexity is O ( N I r 2 ) if each subtensor is stored in the TT format. This reveals that the number of parameters in the TT format scales linearly in N, whereas it scales exponentially in the Tucker format. Another advantage of the TT decomposition is its simple practical implementations. TT decomposition achieves the optimal low TT-ranks due to its balanced matricization scheme using current algorithms such as the SVD-based or the alternative low-rank matrix approximation, whereas tucker decomposition is uncertain to achieve the optimal ranks because of its unbalanced matricization scheme.

5. Experiments

To demonstrate the efficacy and efficiency of the proposed model, experiments on four hyperspectral datasets are conducted. The detailed analysis is as follows.

5.1. Datasets

The proposed algorithms are tested on four real-world bitemporal HSI datasets. Each dataset consists of three images, two sets of HSIs, and a ground-truth map. The first dataset, “Yancheng”, is shown in Figure 5a,e (bands 20, 100, and 10 as RGB). It has 450 × 140 pixels and depicts the countryside in Yancheng in Jiangsu Province, China [30]. The two HSIs were acquired on 3 May 2006 and 23 April 2007. After noise removal, 155 bands were chosen for the change detection. The second dataset, “Bay Area”, is shown in Figure 5b,f (bands 50, 60, and 70 as RGB), with the AVIRIS sensor surrounding the city of Patterson (California), whose spatial dimensions are 600 × 500 pixels and includes 224 spectral bands [54]. The third dataset, “River”, is shown in Figure 5c,g (bands 20, 40, and 60 as RGB). The two River HSI datasets were acquired in Jiangsu Province, China, on 3 May 2013 and 31 December 2013. After removing the noisy bands, this dataset has a size of 463 × 241 pixels and 198 bands [15]. The fourth dataset, “Hermiston city”, is shown in Figure 5d,h (bands 20, 40, and 60 as RGB). In the years 2004 and 2007, the two HSIs were taken with the HYPERION sensor over the Hermiston city area (Oregon), whose spatial dimensions are 390 × 200 pixels and includes 242 spectral bands [32].
Algorithm 2. Pseudocode of the proposed STT for multitemporal HSI change detection.
Input: High-order difference tensors { G ¯ i I 1 × I 2 × I 3 × I 4 × I 5 } i = 1 , 2 , , P , pseudo-labels y ^ c l f , maximum epochs
Parameters: ζ = { W ¯ ( l ) , B ¯ ( l ) , ω } , M o p t c , y ^ o p t , C 1 , C 2
Initialization: Initialize ζ , M p t c , y ^ p t , C 1 , C 2 , y ^ c l f . Maximum epochs
While not converged or maximum epochs not reached, do:
   Compute l c l f and l c l t with Equations (32) and (33)
   Update network loss L with Equation (35)
   Update parameter set ζ using Adam optimizer
   Update centroids M o p t c with Equation (34)
End while
 Obtain features z with Equation (30)
Perform binary clustering on z
Output: Change detection results T

5.2. Setup

For UTT, KA is applied to the HSI tensors Y ¯ 1 and Y ¯ 2 to represent low-order tensors with higher-order tensors A ¯ 1 and A ¯ 2 . Therefore, the dimensions of high-order tensors after KA should be firstly determined, with I i i = 1 , 2 , , 5 in our case. Different combinations satisfying i = 1 5 I i = j = 1 3 J j are attempted to achieve the best CD results, and the settings are different for different datasets, i.e., for the Yancheng dataset, the original tensor size of 450 × 140 × 155 becomes 15 × 30 × 7 × 20 × 155 after the KA scheme. The TT ranks for the UTT model were initialized as ( r [ i ] , i = 1 , 2 , , N 1 ) by the matrix product state (MPS) [55], and the weight parameter α k was initialized as follows:
α k = β k k = 1 N 1 β k ,
where β k = min ( i = 1 k I i , i = k + 1 N I i ) and k = 1 , 2 , , N 1 . Hence, larger weights can be applied to the more balanced matrices in this fashion. To get the initial TT ranks for UTT, each rank r [ i ] is constrained by keeping only the singular values that satisfy the following:
λ j [ i ] λ 1 [ i ] > t h ,
where j = 1 , 2 , , r [ i ] and { λ j [ i ] } is arranged in descending order. The threshold t h is selected empirically. This criterion is used so that more singular values will be truncated for low-rank matricization.
For STT, the backbone comprises two TTD layers and one TTO layer. Each TTD layer is followed by rectified linear units (ReLU) for nonlinear mapping. The properties of the TTD and TTO layers are investigated, and different methods are compared to determine their parameters, such as the number of TTD/TTO layers, dimensions of the tensors after KA, and the TT-ranks of the compressed weight matrix W ¯ ( l ) . The dimensions of the tensors after the KA scheme and TT ranks ( r [ i ] ( l ) , i = 1 , 2 , 3 , 4 , 5 , l = 1 , 2 ) are configured separately for each dataset, i.e., for the Yancheng dataset, a patch size of 30 × 30 × 9 becomes 5 × 6 × 5 × 6 × 9 after the KA scheme, and their ranks for the corresponding dimensions are set as ( r [ i ] ( l ) = 3 , i = 1 , 2 , 3 , 4 , 5 , l = 1 , 2 ) to control the compression factor. A similar procedure is followed for the remaining datasets with different KA and rank settings in order to achieve the best change detection results.
To initialize the pseudo labels of the classification network in STT, the difference of the bitemporal images is computed after KA, K-Means is used as a classical unsupervised method on the extracted difference image, and C 1 is set as 4. The best parameter settings for μ in Equation (35), which balances the classification loss and clustering loss to achieve the best results in the four datasets mentioned above, are μ = [ 0.35 , 0.25 , 0.015 , 0.34 ] , and C 2 is set to 2. The batch size is set to 20 for all the datasets. The default parameter settings may not give the best performance all the time, but they generally give satisfactory results. In the training phase of the proposed self-supervised method, 80% of the samples are chosen randomly for each bitemporal HSI, and the remaining 20% of the random samples are used for validation. During the testing phase, all samples of each bitemporal HSI dataset are used to generate binary change maps.
The proposed model is compared with two classical methods, LSCD and ASCD [30], and six state-of-the-art techniques, HOSVD [20], TDRD [18], PCANet [12], DSFANet [28], HI-DRL [31], and the tucker decomposition-based self-supervised tensor network technique SSTN [38]. For all unsupervised methods, K-Means is used to segment output features into changed areas and unchanged areas. For visual evaluation, white pixels represent the changed areas, and the black pixels represent the unchanged areas.
For numerical evaluation, overall accuracy (OA), class accuracy of unchanged area (CA_UN), class accuracy of changed area (CA_CH), KAPPA coefficient (KAPPA), and Area Under Receiver Operating Characteristic (ROC) Curve (AUC) are used, whereas OA, CA_UN, and CA_CH are presented in percentages, and KAPPA and AUC range from 0 to 1. Each of the indexes indicates better results when it shows higher values. All of the algorithms were developed using TensorFlow in a Python environment, on an NVIDIA GeForce GTX 1080 Ti with 256 GB memory.

5.3. Results

5.3.1. Efficacy

As shown in Figure 6, the visual CD maps produced for the Yancheng dataset by the UTT and STT methods are better than those produced by the classical methods and the counterpart tensor techniques. This viewpoint is supported by the numerical data in the first row of Table 1. It can be seen from Table 1 that although the CA_UN of the proposed method, STT, is not the best, its OA and KAPPA are both the best, which confirms that TT can capture more global correlations. It also proves that TT can detect changed and unchanged regions in a more balanced way (CA_UN and CA_CH of STT are both relatively higher compared to other methods). Though the CA_UN of ASCD is the best, its OA and KAPPA are nearly the worst, which shows that ASCD cannot capture enough global correlations to detect changed regions in a balanced way. The same conclusion can be drawn from the visual CD maps of ASCD and STT in Figure 6c,l respectively. ASCD smooths out the changed features, and the white (changed) regions in the CD map of ASCD in Figure 6c are significantly fewer than STT in Figure 6l, which results in a slightly better CA_UN but a much worse OA. It can also be seen from the white speckles on the top-right and bottom-left corner of Figure 6d,g that results produced by HOSVD and DSFANet have more false alarm rates. As the feature extraction process of HOSVD is based on TD, it can be concluded that the ability of TD in capturing changed information is worse than TT.
Similar conclusions can be drawn for the River dataset, as shown in the visual CD maps in Figure 7 and the numerical data in the second row of Table 1. It can be seen from Table 1 that though CA_UN and CA_CH of STT are not the best, its OA and KAPPA turn out to be the best. On the contrary, although TDRD and HI-DRL have advantages in detecting changed and unchanged regions, respectively, they cannot capture enough global correlations such that they cannot detect changed and unchanged regions in a balanced way. That is why they have the best CA_CH and CA_UN, respectively, but their OA and KAPPA are relatively lower than STT. The visual CD maps in Figure 7 can further validate the advantages of the proposed TT-based CD methods. Figure 7b,d,e show that the bottom-left corner has visible white speckles as compared to UTT and STT. PCANet detects more white regions (pseudo changes), while HI-DRL suppresses the changed features excessively so that it ignores real changes, as shown in Figure 7f,h, respectively. ASCD in Figure 7c smooths out the changed features and results in the lower prediction of changed and unchanged features, which greatly affect the accuracy. The visual results demonstrate that the proposed methods do not only reduce the white noise, but they also discriminate the changed and unchanged features effectively.
The visual CD results and numerical data of the Bay Area and Hermiston datasets are shown in Figure 8 and the third row of Table 1, and in Figure 9 and the fourth row of Table 1, respectively. For the Bay Area dataset, although the OA of HI-DRL is the best, its CA_CH is nearly the worst, which limits the practical application of this method in HSI CD. Similarly, ASCD and DSFANet cannot detect changed and unchanged regions in a balanced way despite the highest CA_UN and CA_CH that they respectively have, which brings out relatively lower OA and KAPPA compared to STT. As for the Hermiston dataset, although the CA_UN of STT is slightly lower than HI-DRL, its CA_CH is significantly better than HI-DRL and all other methods, which makes STT obtain the best OA and KAPPA compared to all other classical and state-of-the-art methods. The visual results of CD maps in Figure 8 and Figure 9 can further validate the mentioned conclusions.
A comparison of AUC values between different methods is also made, and the results are shown in Figure 10 and Table 2, respectively. It can be seen from Figure 10a,b that the ROC curve of STT in the Yancheng and Hermiston datasets is almost at the upper left corner as compared to other methods. In Figure 10c of the River dataset, UTT without SVD outperforms the other methods, which shows that TT-based methods outperform other methods not only in OA and KAPPA, but also in AUC. The numerical results in Table 2 also validate our conclusion. In the Bay Area dataset shown in Figure 10d, although all methods have a lower AUC compared to other datasets, STT still obtained the highest AUC. However, as an evaluation index, the AUC of ROC has its own limits. For example, different ROC curves may have similar AUC. This can be observed in Table 2 and Figure 10. For the Bay Area dataset, the AUC of ASCD is 0.6927 and the AUC of UTT-noSVD is 0.6907, which are very close. Meanwhile, the indexes such as OA and KAPPA in Table 1 show that UTT-noSVD largely outperforms ASCD. Therefore, an evaluation of the CD methods by different indexes is recommended.
The above analysis shows that the proposed technique surpasses the current clustering unsupervised algorithms in terms of OA, CA_UN, CA_CH, KAPPA, and AUC accuracy. This is because the characteristics learned using these techniques are not discriminative enough for clustering. Unlike unsupervised techniques such as subspace projection, HOSVD, and SSTN, our TT-based clustering methodology uses a balanced matricization scheme to capture the global correlation of tensor elements for change detection and can thus suppress the unchanged features and enhance changed features at the same time. In Figure 11, the discriminative efficacy of STT is further highlighted. Because of the overlapping samples from multiple classes, the initial high-order difference tensors are obviously unsuitable for clustering, as shown in Figure 11a,c for the Yancheng and Hermiston datasets, respectively. Meanwhile, as shown in Figure 11b,d, the features of different categories learned by STT are considerably separated. The accuracy of clustering and change detection increases due to the discriminative features having high intra-class similarity and inter-class dissimilarity.

5.3.2. Ablation Study

To validate the effectiveness of the proposed TT methods for CD, an ablation study is conducted for unsupervised TT CD. The experiments are carried out using UTT with SVD, named UTT-SVD, and are compared with the proposed UTT without SVD, named UTT-noSVD. As shown in Figure 6j,k, the proposed UTT-noSVD provides better visual results. Moreover, the numerical evaluation listed in Table 1 validates our point of view for the Yancheng dataset. The ablation experiments for the other datasets for UTT-SVD and the UTT-noSVD are shown in Figure 7j,k for the River dataset, Figure 8j,k for the Bay Area dataset, and in Figure 9j,k for the Hermiston dataset. All the results show that UTT-noSVD is more efficient and provide better results as compared to the UTT-SVD. Their corresponding OA and KAAPA accuracy values, listed in Table 1, also support our point of view.
Similarly, for STT, an ablation study is conducted by lower-dimensional input patch size without using the KA scheme. Lower three-dimensional input patches are compared with the proposed higher five-dimensional input patch sizes. In Figure 12, the visual change detection maps of the River and Bay Area datasets clearly show that the proposed STT model performs better on a higher dimension (5-dimensional input patches), as seen in Figure 12a,c, as compared to a lower dimension (3-dimensional input patches), as seen in Figure 12b,d. This is because high-order rich texture features can be extracted after using KA. Table 3 demonstrates OA and KAPPA accuracy by using different input dimensions for the River and Bay Area datasets. The visual results and numerical evaluation show that STT using a higher dimension performs better than that with low-dimension input patches.

5.3.3. Efficiency

Table 4 lists the time cost per iteration and the total time cost of convergence of comparable techniques. Each time cost is the average value observed from all four HSI datasets. While the same strategy and the same computation platform is adopted, LSCD and ASCD are three times slower than UTT-noSVD and HOSVD, and four times slower than UTT-noSVD for the Yancheng dataset. For the Bay Area dataset, LSCD is 40 times slower than the proposed UTT-noSVD, ASCD is 25 times slower, and HOSVD is 8 times slower than UTT-noSVD. For the River dataset, LSCD is 38 times slower, ASCD is 19 times, and HSOVD is 11 times slower than the proposed UTT-noSVD. Similarly, for the Hermiston city dataset, UTT-noSVD outperforms the LSCD, ASCD, and HOSVD methods with an 18, 7, and 8-fold difference, respectively. The efficiency of UTT-noSVD is benefited from bypassing SVD to compute the TT-ranks and applying KA to the hyperspectral data.
The convergence curve of STT for the Yancheng dataset is shown in Figure 13a. It shows that network loss (combined loss) converges more smoothly and reduces the error efficiently as compared to the use of individual classification and cluster loss. For the Hermiston dataset, the STT convergence curves for the cluster, classification, and network losses are shown in Figure 13b. It shows that the combination of cluster loss with classification loss produces better convergence. The convergence graphs in Figure 13c,d also validate the efficiency of the proposed self-supervised TT technique for the River and Bay Area dataset, respectively.

5.3.4. Discussions

The experiments above demonstrate the capability of TT-based CD in multitemporal HSIs. The strength of UTT and STT can be summarized as follows:
(1).
The inter-class homogeneity and inner-class heterogeneity of HSIs are addressed by UTT and STT effectively by exploiting the ability of TT in capturing global correlations. To be specific, UTT and STT can detect changed and unchanged regions in a more balanced way due to the correlations that TT captures between the changed and unchanged information contained in the original HSIs. This can be validated by the better OA, KAPPA, and AUC values of UTT and STT as compared to the TD-based methods HOSVD [20], TDRD [18], and SSTN [38], whose low-rank features are extracted in an unbalanced way. The T-SNE results in Figure 11 also indicate that the features extracted by the TT-based methods are discriminative enough to differentiate between the changed and unchanged regions in HSI CD.
(2).
Both UTT and STT successfully handle the high dimensionality of HSIs by TT decomposition, which decomposes N-order weight tensors into small three-order tensor cores by approximating the low-order optimal TT ranks. Hence, the dimensionality can be reduced and the redundant information can be removed. At the same time, the execution time of UTT-noSVD is obviously lower than other existing unsupervised HSI CD methods such as LSCD [30]. Costly manual annotations can also be removed as unsupervised learning and self-learning are introduced into UTT and STT, respectively.
(3).
Tensor augmentation is achieved through the KA scheme, which involves replacing a low-order tensor with a higher-order tensor without changing the number of tensor entries. Therefore, a high-order tensor with richer texture features can be achieved without increasing computation complexity. It can be seen in Figure 12 and Table 3 that KA indeed works in our proposed methods.
Despite all of the benefits of the TT-based change detection methods UTT/STT, there is still potential for improvement. Here are few points that could be further studied. First, UTT/STT both use K-Means for binary classification as well as during the pre-clustering stage to cluster the changed and unchanged features. The cluster number specification in the advance non-handling of outliers and their dependency on convex data cause the K-Means to be inadequate for complicated datasets. Similarly, during pre-clustering, pseudo-labels can be generated by more advanced methods to improve the accuracy of STT. Second, although STT and UTT are used for change and background classification for bitemporal images, they could be further applied to multiple change detection with multitemporal images.

6. Conclusions

Inspired by knowledge from the quantum information theory, this paper proves that TT decomposition is capable of capturing more global correlations between the changed/unchanged information than TD. Based on this, two novel approaches are proposed for HSI CD. The unsupervised UTT technique is based on multilinear matrix factorization and is more efficient since it bypasses the expensive SVD. STT, the second one, is a self-supervised technique. STT performs much better than the direct unsupervised technique, UTT. TT-based CD methods use a well-balanced matricization strategy so that they can detect changed and unchanged regions in a more balanced way. Tensors are augmented from low-order to higher-order through a KA scheme while not changing the total number of entries in order to extract changed features efficiently without increasing computational complexity. Then, the proposed TT-based CD methods are tested on four real-world bitemporal HSI datasets. Experiments have demonstrated that the proposed TT-based CD methods outperform their tensor counterparts TDRD, HOSVD, SSTN, and other state-of-the-art techniques in OA, CA_UN, CA_CH, KAPPA, and AUC. With such high accuracy and no need for manual annotations, the proposed methods exhibit the potential of applications to deal with emergencies such as landslide detection and earthquake damage estimation. For future work, more real-world multitemporal HSI datasets will be acquired to further investigate the proposed methods and update the UTT/STT so that it could be adapted for change detection in multitemporal HSIs.

Author Contributions

Conceptualization, Z.C.; methodology, Z.C. (all), M.S. (part of Subsection 4.2 and Subsection 4.3), and H.W (part of Subsection 4.1); validation, M.S.; writing—original draft preparation, M.S.; writing—review and editing, H.W.; visualization, H.W.; supervision, Z.C. and G.L.; funding acquisition, Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the “Chenguang Program”, supported by the Shanghai Education Development Foundation and the Shanghai Municipal Education Commission under Grant 18CG38, the National Natural Science Foundation of China under Grant 61702094, and the Science and Technology Commission of Shanghai Municipal under Grant 17YF1427400.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bovolo, F.; Marchesi, S.; Bruzzone, L. A framework for automatic and unsupervised detection of multiple changes in multitemporal images. IEEE Trans. Geosci. Remote Sens. 2011, 50, 2196–2212. [Google Scholar] [CrossRef]
  2. Liu, S.; Bruzzone, L.; Bovolo, F.; Du, P. Hierarchical unsupervised change detection in multitemporal hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2014, 53, 244–260. [Google Scholar]
  3. Deng, J.S.; Wang, K.; Deng, Y.H.; Qi, G.J. PCA based land use change detection and analysis using multitemporal and multisensor satellite data. Int. J. Remote Sens. 2008, 29, 4823–4838. [Google Scholar] [CrossRef]
  4. Lu, D.; Mausel, P.; Brondizio, E.; Moran, E. Change detection techniques. Int. J. Remote Sens. 2004, 25, 2365–2401. [Google Scholar] [CrossRef]
  5. Kennedy, R.E.; Townsend, P.A.; Gross, J.E.; Cohen, W.B.; Bolstad, P.; Wang, Y.Q.; Adams, P. Remote sensing change detection tools for natural resource managers: Understanding concepts and tradeoffs in the design of landscape monitoring projects. Remote Sens. Environ. 2009, 113, 1382–1396. [Google Scholar] [CrossRef]
  6. Coppin, P.R.; Bauer, M.E. Digital change detection in forest ecosystems with remote sensing imagery. Remote Sens. Rev. 1996, 13, 207–234. [Google Scholar] [CrossRef]
  7. Yang, X.; Chen, L. Using multi-temporal remote sensor imagery to detect earthquake-triggered landslides. Int. J. Appl. Earth Obs. Geoinf. 2010, 12, 487–495. [Google Scholar] [CrossRef]
  8. Ma, L.; Li, M.; Blaschke, T.; Ma, X.; Tiede, D.; Cheng, L.; Chen, Z.; Chen, D. Object-based change detection in urban areas: The effects of segmentation strategy, scale, and feature space on unsupervised methods. Remote Sens. 2016, 8, 761. [Google Scholar] [CrossRef] [Green Version]
  9. Hasanlou, M.; Seydi, S.T. Hyperspectral change detection: An experimental comparative study. Int. J. Remote Sens. 2018, 39, 7029–7083. [Google Scholar] [CrossRef]
  10. Guo, Q.; Zhang, J.; Zhong, C.; Zhang, Y. Change Detection for Hyperspectral Images via Convolutional Sparse Analysis and Temporal Spectral Unmixing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4417–4426. [Google Scholar] [CrossRef]
  11. Wu, C.; Zhang, L.; Du, B. Kernel slow feature analysis for scene change detection. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2367–2384. [Google Scholar] [CrossRef]
  12. Gao, F.; Dong, J.; Li, B.; Xu, Q. Automatic change detection in synthetic aperture radar images based on PCANet. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1792–1796. [Google Scholar] [CrossRef]
  13. Chen, J.; Yuan, Z.; Peng, J.; Chen, L.; Huang, H.; Zhu, J.; Liu, Y.; Li, H. Dasnet: Dual attentive fully convolutional siamese networks for change detection of high resolution satellite images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 1194–1206. [Google Scholar] [CrossRef]
  14. Hou, B.; Liu, Q.; Wang, H.; Wang, Y. From W-Net to CDGAN: Bitemporal change detection via deep learning techniques. IEEE Trans. Geosci. Remote Sens. 2019, 58, 1790–1802. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A general end-to-end 2-D CNN framework for hyperspectral image change detection. IEEE Trans. Geosci. Remote Sens. 2018, 57, 3–13. [Google Scholar] [CrossRef] [Green Version]
  16. Shi, Q.; Liu, M.; Li, S.; Liu, X.; Wang, F.; Zhang, L. A Deeply Supervised Attention Metric-Based Network and an Open Aerial Image Dataset for Remote Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5604816. [Google Scholar] [CrossRef]
  17. Yang, B.; Fu, X.; Sidiropoulos, N.D.; Hong, M. Towards k-means-friendly spaces: Simultaneous deep learning and clustering. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3861–3870. [Google Scholar]
  18. Hou, Z.; Li, W.; Tao, R.; Du, Q. Three-Order Tucker Decomposition and Reconstruction Detector for Unsupervised Hyperspectral Change Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6194–6205. [Google Scholar] [CrossRef]
  19. Huang, F.; Yu, Y.; Feng, T. Hyperspectral remote sensing image change detection based on tensor and deep learning. J. Vis. Commun. Image Represent. 2019, 58, 233–244. [Google Scholar] [CrossRef]
  20. Chen, Z.; Wang, B.; Niu, Y.; Xia, W.; Zhang, J.Q.; Hu, B. Change detection for hyperspectral images based on tensor analysis. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Milan, Italy, 26–31 July 2015; pp. 1662–1665. [Google Scholar]
  21. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  22. Cichocki, A. Era of big data processing: A new approach via tensor networks and tensor decompositions. arXiv 2014, arXiv:1403.2048. [Google Scholar]
  23. Nielsen, M.A.; Chuang, I. Quantum Computation and Quantum Information; Cambridge University Press: Cambridge, UK, 2002. [Google Scholar]
  24. Oseledets, I.V. Tensor-train decomposition. SIAM J. Sci. Comput. 2011, 33, 2295–2317. [Google Scholar] [CrossRef]
  25. Bengua, J.A.; Phien, H.N.; Tuan, H.D.; Do, M.N. Efficient tensor completion for color image and video recovery: Low-rank tensor train. IEEE Trans. Image Process. 2017, 26, 2466–2479. [Google Scholar] [CrossRef] [Green Version]
  26. Latorre, J.I. Image compression and entanglement. arXiv 2005, arXiv:quant-ph/0510031. [Google Scholar]
  27. Du, P.; Liu, S.; Gamba, P.; Tan, K.; Xia, J. Fusion of difference images for change detection over urban areas. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1076–1086. [Google Scholar] [CrossRef]
  28. Du, B.; Ru, L.; Wu, C.; Zhang, L. Unsupervised deep slow feature analysis for change detection in multi-temporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9976–9992. [Google Scholar] [CrossRef] [Green Version]
  29. Li, Q.; Gong, H.; Dai, H.; Li, C.; He, Z.; Wang, W.; Mu, T. Unsupervised Hyperspectral Image Change Detection via Deep Learning Self-Generated Credible Labels. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9012–9024. [Google Scholar] [CrossRef]
  30. Wu, C.; Du, B.; Zhang, L. A subspace-based change detection method for hyperspectral images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 815–830. [Google Scholar] [CrossRef]
  31. Zhang, P.; Gong, M.; Zhang, H.; Liu, J.; Ban, Y. Unsupervised difference representation learning for detecting multiple types of changes in multitemporal remote sensing images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 2277–2289. [Google Scholar] [CrossRef]
  32. López-Fandiño, J.; Garea, A.S.; Heras, D.B.; Arguello, F. Stacked autoencoders for multiclass change detection in hyperspectral images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 1906–1909. [Google Scholar]
  33. Hou, Z.; Li, W.; Li, L.; Tao, R.; Du, Q. Hyperspectral change detection based on multiple morphological profiles. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5507312. [Google Scholar] [CrossRef]
  34. Seydi, S.T.; Shah-Hosseini, R.; Hasanlou, M. New framework for hyperspectral change detection based on multi-level spectral unmixing. Appl. Geomat. 2021, 13, 763–780. [Google Scholar] [CrossRef]
  35. Seydi, S.T.; Hasanlou, M. A New Structure for Binary and Multiple Hyperspectral Change Detection Based on Spectral Unmixing and Convolutional Neural Network. Measurement 2021, 186, 110137. [Google Scholar] [CrossRef]
  36. Hasanlou, M.; Seydi, S.T.; Shah-Hosseini, R. A sub-pixel multiple change detection approach for hyperspectral imagery. Can. J. Remote Sens. 2018, 44, 601–615. [Google Scholar] [CrossRef]
  37. Jafarzadeh, H.; Hasanlou, M. An unsupervised binary and multiple change detection approach for hyperspectral imagery based on spectral unmixing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4888–4906. [Google Scholar] [CrossRef]
  38. Zhou, F.; Chen, Z. Hyperspectral Image Change Detection by Self-Supervised Tensor Network. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 19–24 July 2020; pp. 2527–2530. [Google Scholar]
  39. Zhang, R.; Isola, P.; Efros, A.A. Colorful image colorization. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 649–666. [Google Scholar]
  40. Pathak, D.; Krahenbuhl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2536–2544. [Google Scholar]
  41. Noroozi, M.; Favaro, P. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 69–84. [Google Scholar]
  42. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  43. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  44. Dong, H.; Ma, W.; Jiao, L.; Liu, F.; Li, L. A Multiscale Self-Attention Deep Clustering for Change Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  45. Chen, Y.; Bruzzone, L. Self-supervised Change Detection in Multi-view Remote Sensing Images. arXiv 2021, arXiv:2103.05969. [Google Scholar]
  46. Deng, Y.J.; Li, H.C.; Fu, K.; Du, Q.; Emery, W.J. Tensor low-rank discriminant embedding for hyperspectral image dimensionality reduction. IEEE Trans. Geosci. Remote Sens. 2018, 56, 7183–7194. [Google Scholar] [CrossRef]
  47. Zhao, M.; Li, W.; Li, L.; Ma, P.; Tao, R. Three-order tensor creation and tucker decomposition for infrared small-target detection. IEEE Trans. Geosci. Remote Sens. 2021, 99, 1–16. [Google Scholar] [CrossRef]
  48. Li, L.; Li, W.; Qu, Y.; Zhao, C.; Tao, R.; Du, Q. Prior-based tensor approximation for anomaly detection in hyperspectral imagery. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 1037–1050. [Google Scholar] [CrossRef]
  49. Velasco-Forero, S.; Angulo, J. Classification of hyperspectral images by tensor modeling and additive morphological decomposition. Pattern Recognit. 2013, 46, 566–577. [Google Scholar] [CrossRef] [Green Version]
  50. Zniyed, Y. Breaking the Curse of Dimensionality Based on Tensor Train: Models and Algorithms. Ph.D. Thesis, Paris-Saclay, Paris, France, 2019. [Google Scholar]
  51. Henderson, L.; Vedral, V. Classical, quantum and total correlations. J. Phys. A. Math. Gen. 2001, 34, 6899–6905. [Google Scholar] [CrossRef]
  52. Novikov, A.; Podoprikhin, D.; Osokin, A.; Vetrov, D. Tensorizing neural network. arXiv 2015, arXiv:1509.06569. [Google Scholar]
  53. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  54. López-Fandiño, J.; B Heras, D.; Arguello, F.; Dalla, M.M. GPU framework for change detection in multitemporal hyperspectral images. Int. J. Parallel Program. 2019, 47, 272–292. [Google Scholar] [CrossRef]
  55. Perez-Garcia, D.; Verstraete, F.; Wolf, M.M.; Cirac, J.I. Matrix product state representations. arXiv 2006, arXiv:quant-ph/0608197. [Google Scholar] [CrossRef]
Figure 1. Graphical representation of tensor manipulations. (a) Basic building blocks for tensor network diagrams; (b) mode-2 matricization (left), mode-(1,2) matricization (middle), and vectorization (right) of a 3-order tensor G ¯ I 1 × I 2 × I 3 .
Figure 1. Graphical representation of tensor manipulations. (a) Basic building blocks for tensor network diagrams; (b) mode-2 matricization (left), mode-(1,2) matricization (middle), and vectorization (right) of a 3-order tensor G ¯ I 1 × I 2 × I 3 .
Electronics 11 01486 g001
Figure 2. Visualization of a 5-order tensor train decomposition in graphical notations.
Figure 2. Visualization of a 5-order tensor train decomposition in graphical notations.
Electronics 11 01486 g002
Figure 3. Framework of the proposed UTT for multitemporal HSI change detection.
Figure 3. Framework of the proposed UTT for multitemporal HSI change detection.
Electronics 11 01486 g003
Figure 4. Framework of the proposed STT for multitemporal HSI change detection. (a) Overall framework of STT, (b) tensor train decomposition (TTD) layer, (c) tensor train output (TTO) layer.
Figure 4. Framework of the proposed STT for multitemporal HSI change detection. (a) Overall framework of STT, (b) tensor train decomposition (TTD) layer, (c) tensor train output (TTO) layer.
Electronics 11 01486 g004
Figure 5. False-color composition of (a,e) Yancheng, (b,f) Bay Area, (c,g) River, and (d,h) Hermiston city.
Figure 5. False-color composition of (a,e) Yancheng, (b,f) Bay Area, (c,g) River, and (d,h) Hermiston city.
Electronics 11 01486 g005
Figure 6. Binary detection results in the Yancheng dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, I TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Figure 6. Binary detection results in the Yancheng dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, I TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Electronics 11 01486 g006
Figure 7. Binary detection results in the River dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, (e) TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Figure 7. Binary detection results in the River dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, (e) TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Electronics 11 01486 g007
Figure 8. Binary detection results in the Bay Area dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, (e) TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Figure 8. Binary detection results in the Bay Area dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, (e) TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Electronics 11 01486 g008
Figure 9. Binary detection results in the Hermiston dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, (e) TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Figure 9. Binary detection results in the Hermiston dataset: (a) ground truth, (b) LSCD, (c) ASCD, (d) HOSVD, (e) TDRD, (f) PCANet, (g) DSFANet, (h) HI-DRL, (i) SSTN, (j) UTT-SVD, (k) UTT-noSVD, and (l) STT.
Electronics 11 01486 g009
Figure 10. ROCs of different methods for the (a) Yancheng dataset (b) Hermiston dataset (c) River dataset (d) Bay Area dataset.
Figure 10. ROCs of different methods for the (a) Yancheng dataset (b) Hermiston dataset (c) River dataset (d) Bay Area dataset.
Electronics 11 01486 g010
Figure 11. Scatter plots produced by t-SNE of all samples from the Yancheng dataset (a,b) and Hermiston dataset (c,d). Red circles represent unchanged samples, and blue circles represent changed samples. Yancheng (a) original difference tensor, (b) features learnt by STT, Hermiston (c) original difference tensor, (d) features learnt by STT.
Figure 11. Scatter plots produced by t-SNE of all samples from the Yancheng dataset (a,b) and Hermiston dataset (c,d). Red circles represent unchanged samples, and blue circles represent changed samples. Yancheng (a) original difference tensor, (b) features learnt by STT, Hermiston (c) original difference tensor, (d) features learnt by STT.
Electronics 11 01486 g011
Figure 12. Ablation study of STT. River dataset (a) ground truth, (b) with 5-dimensional input patches, (c) with 3-dimensional input patches; Bay Area dataset (d) ground truth, (e) with 5-dimensional input patches, and (f) with 3-dimensional input patches.
Figure 12. Ablation study of STT. River dataset (a) ground truth, (b) with 5-dimensional input patches, (c) with 3-dimensional input patches; Bay Area dataset (d) ground truth, (e) with 5-dimensional input patches, and (f) with 3-dimensional input patches.
Electronics 11 01486 g012
Figure 13. Convergence curve of STT for different datasets. (a) Yancheng dataset, (b) Hermiston dataset, (c) River dataset, (d) Bay Area dataset.
Figure 13. Convergence curve of STT for different datasets. (a) Yancheng dataset, (b) Hermiston dataset, (c) River dataset, (d) Bay Area dataset.
Electronics 11 01486 g013
Table 1. Evaluation of change detection results in different datasets.
Table 1. Evaluation of change detection results in different datasets.
Data
Set
MetricLS
CD
AS
CD
HO
SVD
TD
RD
PCA-
Net
DSFA-NetHI-
DRL
SSTNUTT-SVDUTT-noSVDSTT
YanchengCA_UN(%)94.3399.9497.7995.6794.1197.0199.1098.1998.1798.1597.79
CA_CH(%)85.1375.8998.6995.0998.3592.4092.0096.2597.7898.0398.69
OA(%)91.7393.1398.0497.3195.3195.7197.0997.6498.0798.1198.20
KAPPA0.79590.81760.95240.93450.88900.89420.92700.94200.95280.95360.9561
RiverCA_UN(%)92.7197.9292.3792.0383.1797.0698.7697.3492.1693.7198.44
CA_CH(%)44.1340.3290.7294.0368.5566.2158.4872.8491.0185.9569.24
OA(%)88.4892.9392.2392.2181.5294.2594.2394.5892.3093.1095.90
KAPPA0.33640.47680.62830.63650.35900.67680.66400.72100.62850.64620.7237
Bay AreaCA_UN(%)86.9099.5487.3085.9689.3682.1997.5993.1790.7291.8194.29
CA_CH(%)31.4411.0242.9645.2838.5048.6425.0236.6042.7339.5438.72
OA(%)73.3277.8676.4475.3778.2373.9781.7180.7978.9778.9881.02
KAPPA0.20280.15000.32200.30060.30400.30460.29700.34600.37080.36520.3841
HermistonCA_UN(%)94.7799.7498.1476.8285.4898.7699.7699.3398.2098.4899.02
CA_CH(%)80.8870.1393.3655.5567.7891.3583.8693.8893.2992.9595.02
OA(%)92.9995.9597.5374.0983.1997.8197.7298.6397.5797.7698.83
KAPPA0.70680.79380.89220.21810.41400.90180.89100.93800.89360.90100.9384
Table 2. Evaluation of AUC values for different methods.
Table 2. Evaluation of AUC values for different methods.
MethodsYanchengHermistonRiverBay Area
LSCD0.94000.92240.86290.5968
ASCD0.97960.94200.80820.6927
HOSVD0.99600.99080.97190.6685
TDRD0.99680.80220.98070.6699
DSFANet0.98870.98410.92730.6595
UTT-SVD0.99640.99100.98130.6863
UTT-noSVD0.99670.99190.98170.6907
STT0.99700.99510.95310.6978
Table 3. Ablation study: STT with 5-dimensional vs. 3-dimensional input patches.
Table 3. Ablation study: STT with 5-dimensional vs. 3-dimensional input patches.
MethodsRiverBay Area
OA (%)KAPPAOA (%)KAPPA
STT(3-Dimensional)95.620.708080.360.3820
STT(5-Dimensional)95.900.723781.020.3841
Table 4. Execution time of various methods using different experimental datasets (seconds).
Table 4. Execution time of various methods using different experimental datasets (seconds).
MethodsYanchengBay AreaRiverHermiston
LSCD3.945217.66238.80418.838
ASCD3.866129.60920.4877.421
HOSVD4.90843.48011.8309.316
UTT-SVD8.992262.7475.70921.018
UTT-noSVD1.4765.4410.9340.678
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sohail, M.; Wu, H.; Chen, Z.; Liu, G. Unsupervised and Self-Supervised Tensor Train for Change Detection in Multitemporal Hyperspectral Images. Electronics 2022, 11, 1486. https://doi.org/10.3390/electronics11091486

AMA Style

Sohail M, Wu H, Chen Z, Liu G. Unsupervised and Self-Supervised Tensor Train for Change Detection in Multitemporal Hyperspectral Images. Electronics. 2022; 11(9):1486. https://doi.org/10.3390/electronics11091486

Chicago/Turabian Style

Sohail, Muhammad, Haonan Wu, Zhao Chen, and Guohua Liu. 2022. "Unsupervised and Self-Supervised Tensor Train for Change Detection in Multitemporal Hyperspectral Images" Electronics 11, no. 9: 1486. https://doi.org/10.3390/electronics11091486

APA Style

Sohail, M., Wu, H., Chen, Z., & Liu, G. (2022). Unsupervised and Self-Supervised Tensor Train for Change Detection in Multitemporal Hyperspectral Images. Electronics, 11(9), 1486. https://doi.org/10.3390/electronics11091486

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop