Next Article in Journal
Snow Depth Estimation with GNSS-R Dual Receiver Observation
Next Article in Special Issue
A Deep Hashing Technique for Remote Sensing Image-Sound Retrieval
Previous Article in Journal
Development and Intercomparison Study of an Atmospheric Motion Vector Retrieval Algorithm for GEO-KOMPSAT-2A
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Large-Scale Remote Sensing Image Retrieval Based on Semi-Supervised Adversarial Hashing

1
Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, International Research Center for Intelligent Perception and Computation, Joint International Research Laboratory of Intelligent Perception and Computation, School of Artificial Intelligence, Xidian University, Xi’an 710071, China
2
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(17), 2055; https://doi.org/10.3390/rs11172055
Submission received: 4 July 2019 / Revised: 11 August 2019 / Accepted: 29 August 2019 / Published: 1 September 2019
(This article belongs to the Special Issue Content-Based Remote Sensing Image Retrieval)

Abstract

:
Remote sensing image retrieval (RSIR), a superior content organization technique, plays an important role in the remote sensing (RS) community. With the number of RS images increases explosively, not only the retrieval precision but also the retrieval efficiency is emphasized in the large-scale RSIR scenario. Therefore, the approximate nearest neighborhood (ANN) search attracts the researchers’ attention increasingly. In this paper, we propose a new hash learning method, named semi-supervised deep adversarial hashing (SDAH), to accomplish the ANN for the large-scale RSIR task. The assumption of our model is that the RS images have been represented by the proper visual features. First, a residual auto-encoder (RAE) is developed to generate the class variable and hash code. Second, two multi-layer networks are constructed to regularize the obtained latent vectors using the prior distribution. These two modules mentioned are integrated under the generator adversarial framework. Through the minimax learning, the class variable would be a one-hot-like vector while the hash code would be the binary-like vector. Finally, a specific hashing function is formulated to enhance the quality of the generated hash code. The effectiveness of the hash codes learned by our SDAH model was proved by the positive experimental results counted on three public RS image archives. Compared with the existing hash learning methods, the proposed method reaches improved performance.

Graphical Abstract

1. Introduction

With the development of Earth observation (EO) techniques, remote sensing (RS) has entered the big data era. The number of RS images collected by the EO satellites every day is increased explosively. How to manage these massive amounts of RS images in the content level becomes an open and tough task in the RS community [1]. As a useful management tool, remote sensing image retrieval (RSIR) is always adopted to organize the RS images according to their contents [2]. The main target of RSIR is finding similar target images from the archive according to the query image provided by users. RSIR is a complicated technology that consists of a series of image processing methods [3], such as feature extraction, similarity matching, etc. It is also a comprehensive technology that covers a large number of techniques [4], such as feature representation, metric learning, image annotation, image caption, etc.
The basic framework of RSIR is summarized in Figure 1. Feature learning focuses on obtaining useful image representation while similarity matching aims at measuring the resemblance between images. The process of RSIR can be formulated as Υ = s o r t d i s t q , I , where q indicates the query RS image, I means a set of RS images within archive (target images), d i s t ( · ) denotes a certain distance measure, s o r t ( · ) illustrates the sort function, and Υ is the retrieval results. Two assessment criteria, i.e., precision and efficiency, are always selected to assess the retrieval performance [2,5]. The ideal scenario is that the developed retrieval algorithms can search for more correct results with the minimal time cost. In the very beginning, researchers always paid close attention to develop effective feature learning methods [1,2,6]. The similarity matching was always accomplished in an exhausting manner using the simple or dedicated distance measures. From the precision aspect, this kind of retrieval methods performs well since the continuous and dense image features learned by the specific algorithms are useful enough [7,8,9,10]. From the efficiency aspect, this kind of approaches is feasible as the volume of RS images within the archive is not large. Nevertheless, in the current RS big data era, the volume of RS images within the archive becomes large. The traditional exhausting search mechanism cannot reach the demands of the timeliness in this large-scale scenario. Thus, the approximate nearest neighbor (ANN) search [11] draws researchers’ attention. Note that the large-scale mentioned in this paper means there are many RS images within the archive.
ANN, as an approximate searching strategy, aims at finding the samples that are near to the query under certain distances [12]. For a query image q , ANN can be mathematically formulated as ANN q = I d i s t q , I δ , I I , where δ is a distance threshold. Many ANN approaches are proposed in the last decades, including time-constrained search methods [13,14], and error-constrained neighbor search algorithms [15,16,17]. During the search, the time-constrained ANNs aim to reduce the time-, cost-, and error-constrained ANNs from pouring attention into the accuracy. Among the diverse ANN methods, hashing is a successful solution for many applications, which focuses on mapping the images from the original feature space to the hash space. Through transforming an image item into compact and short binary codes, not only the room for storing the images could be reduced but also the similarity matching speed could be enhanced [18]. For an image I , hashing can be defined as b = h 1 I , h 2 I , , h K I , where h i · , i = 1 , , K indicates the K hash functions, while b means the K-dimensional binary hash code with the value 0 , 1 or - 1 , 1 . The crucial problem in hashing is formulating or learning the hash functions.
In the beginning, locality sensitive hashing (LSH) [16,19] and its variations [20,21,22,23,24] play a dominant role in the hashing community. By introducing the random projection, the different hashing functions are generated first. Then, the hash codes of images can be obtained using those functions. Accompanying with the proper searching schemes, the LSH family achieves successes in the large-scale retrieval community. However, this kind of data-independent hashing methods needs long bits to ensure retrieval precision [18]. To overcome the mentioned limitation, data-dependent hashing becomes popular recently, which aims at learning the hash functions from the specific dataset so that the retrieval results based on the obtained hash codes are as similar as the results based on the original features [25]. In general, there are three key points in learning to hash [26], including similarity preserving, low quantization loss, and bit balance. The similarity preserving means the resemblance relationships between images in the original feature space should remain to the binary code space. The low quantization loss indicates that the performance gap between the hash code before and after the binarization should be minimized. The bit balance denotes that the bits within a hash code should have approximate 50% chance to be 0 or 1. There are many hash learning methods [27,28,29] that have been proposed during the last decades, and they achieve successes in their own applications. However, their performance is still limited by the original low-level features.
Due to the strong capacity of feature representation, deep learning [30], especially the convolutional neural network (CNN) [31], is bringing a technical revolution for image processing. Hash learning, of course, also benefits from the CNN model. Since the features obtained by CNN (which is recorded deep features in this paper) can capture more useful information from images, the behavior of hash codes learned by those deep features is superior to that of the binary codes learned by handcrafted features. Many deep hash learning methods are proposed [32,33,34,35]. Most of them embed the hash learning into the CNN framework, i.e., a hash layer is added in the top of the CNN model. When the network is trained, the hash codes are the outputs of the hash layer. Although deep hashing achieves astounding results in large-scale image retrieval, there are still some issues that could be improved. First, due to the characteristic of the CNN model, many labeled data should be used to complete the network training. This is tough work for many practical applications, such as the RS related tasks. Second, the hash codes are the discrete binary vectors. Since they do not have derivatives, the CNN model cannot be trained using the conventional optimization methods (e.g., stochastic gradient descent) directly. Some researchers adopt the algebraic relaxing scheme to avoid this limitation [35,36,37,38,39]. Nevertheless, this scheme always reduces the final hash codes’ performance. Third, bit balance, one of the key points in hash learning, is always accomplished by the penalty term within the objective function. This would influence the other two points (i.e., similarity preserving and quantization loss) as the optimization results are the trade-off between different penalty terms.
In the RS community, learning to hash is also an important technique to mine the contents of RS images in the large-scale scenario [5,40]. However, due to the specific properties of RS images, it is improper to directly use the hash learning algorithms proposed in the natural image community to deal with the large-scale RSIR task. For example, the number of labeled RS images is not enough for many successful deep hash learning methods. In addition, the contents within an RS image are more complex compared with a natural image. The diverse objects with multi-scale information enhance the difficulty of hash learning. Consequently, there is a need for the effective deep hash learning method to address the large-scale RS image retrieval task. In this paper, considering the limitations (for the existing deep hashing techniques) and difficulties (for the RS images) discussed above, we propose a new deep hashing method to accomplish the large-scale RSIR, named semi-supervised deep adversarial hashing (SDAH). First, we assume that the proper visual features of RS images have been obtained, so that our main work is mapping the continuous and high dimensional visual features into the discrete and short binary vectors. This can increase the practicability of our method. Second, we introduce adversarial auto-encoder (AAE) [41] to be the backbone of our SDAH. The residual auto-encoder (RAE) is developed to generate hash codes with minimal information loss. The uniform discriminator is used to impose the uniform binary distribution on the generated hash codes, which ensures that the obtained hash codes are bit balanced. Third, to decrease the number of labeled data, a category discriminator is added to guarantee that our hashing can be accomplished in a semi-supervised manner. Through defining the specific objective function, the learned hash codes are similarity preserving, low quantization loss and discriminated. Finally, the normal hamming distance is used to get the retrieval results rapidly. The proposed hash learning model can be formulated as b = H f = H M I , where b and f indicate the hash code and visual feature of the image I , M · denotes the visual feature extraction, H · means the nonlinear hash functions learned from a certain number of RS images using the SDAH network. Based on the hashing model, the retrieval process can be defined as Υ = s o r t d i s t h H M q , H M I , where d i s t h · illustrates the hamming distance.
Our main contributions can be summarized as follows.
  • To ensure our hashing method is practical, we adopt a two-stage scheme to replace the end-to-end learning framework. Thus, any useful RS feature learning methods can be used to learn effective visual features.
  • To get the bit balanced hash codes, we embed our hash learning in the generative adversarial framework. Through the adversarial learning, the prior binary uniform distribution can be imposed on the generated codes. Thus, SDAH can ensure the coding balance intuitively.
  • To learn the effective binary code with minimal costs, we expand SDAH to the semi-supervised framework. In addition, the hashing objective function is developed to ensure the binary vectors are not only similarity preserving and low quantization loss but also discriminative.
The rest of the paper is organized as follows. Section 2 reviews some published literature related to RSIR and hash learning. Both preliminaries (e.g., AAE model) and our SDAH are introduced in Section 3. Two RS image archives used to testify our method are presented in Section 4. The experimental results and related discussion are displayed in Section 5. Section 6 provides the conclusion.

2. Related Work

2.1. Remote Sensing Image Retrieval

RSIR is a popular research topic in the RS community, and there are many successful methods that were proposed in the last decades. Here, we divide them into two groups for review. In the first group, the proposed methods aim to develop effective content descriptors or similarity measures for RSIR. Datcu et al. [42] presented a knowledge-driven information mining method in 2003. In this method, the stochastic model (Gibbs Markov random field) was used to extract the RS images’ texture features from the content aspect. Meanwhile, the unsupervised clustering scheme was developed to reduce the feature vectors’ dimension. In addition, Bayesian networks were introduced to capture the users’ interests from the semantic level. Combining them together, satisfactory retrieval results could be acquired. Yang et al. [43] presented an RSIR method based on the bag-of-words (BOW) [44] features. Several key issues for constructing the BOW features are discussed in detail for RS images, such as scale invariant feature transform (SIFT) [45], codebook size, etc. Based on the BOW features, the retrieval results could be obtained using the common distance measures. With the help of CNN, Tang et al. [46] proposed an unsupervised deep feature learning method for RSIR. Both the complex contents and the multi-scale information of the RS images were taken into account in this method. Apart from the mentioned feature extraction/learning methods, there are still many algorithms that focus on similarity measures. A region-based distance was proposed for Synthetic Aperture Radar (SAR) image content retrieval [47]. In this method, SAR image was segmented into several regions using the texture, brightness and shape features first. Then, the distances between SAR images were formulated as the integration of the dissimilarities of multiple regions. Tang et al. [48] introduced a fuzzy similarity measure to complete the RSIR. Similar to the approach presented in [47], the SAR image was divided into different regions first. Then, the fuzzy theory was chosen to represent different regions, which could decrease the negative influence of segmentation uncertainty. Finally, the formulation of resemblance between SAR images was transformed into calculating the similarities between fuzzy sets. Li et al. [49] developed a collaborative affinity metric fusion method to measure the similarities between high-resolution RS images for RSIR. First, the contributions of handcrafted and deep features were combined under the graph theory. Then, the resemblance between RS images was defined as the affinity values between the nodes within the graph.
In the second group, the researchers concentrate their efforts on improving the initial retrieval results. Some of them use the supervised relevance feedback (RF) framework to enhance retrieval performance. Ferecatu and Boujemaa [50] proposed an active learning (AL) driven RF method to improve retrieval performance. To ease the users’ burden of sample selection, AL was introduced in the RF iteration. Meanwhile, the authors proposed the most ambiguous and orthogonal (MAO) principle under the support vector machine (SVM) paradigm to ensure the selected samples are useful. Finally, the initial retrieval results were reranked using the SVM classifier. Another AL driven RF method was introduced in [51]. Besides the uncertainty and diversity (which correspond to the MAO principle), the density of samples was also taken into account during the sample selection. To overcome the limitation of a single AL method, a multiple AL-based RF method was proposed in [48]. Diverse AL algorithms were embedded in the RF framework to generate different reranked results. Then, those results were fused in the relevance score level to obtain the final retrieval results. Apart from the RF methods, other researchers select the unsupervised reranking techniques to improve the initial retrieval results. Tang et al. [52] developed an image reranking approach for SAR images based on the multiple graphs fusion. First, diverse SAR-oriented visual features were extracted, and the relevant scores were estimated in multiple feature spaces. Then, a model-image matrix was constructed based on the estimated scores for calculating the similarities between SAR images. Finally, a fully connected graph was established using the obtained similarities to accomplish the image reranking. Another image reranking method was developed for RS images [53], in which the retrieval improvement was achieved in a coarse-to-fine manner.

2.2. Learning to Hash

We roughly divide the existing learning to hash methods into two groups according to if the deep neural network is added in hash learning or not. The methods within the first group aim to develop effective hashing functions without the help of deep learning, and we name them non-deep hashing. Weiss et al. [26] proposed the spectral hashing to obtain the binary codes with the consideration of the semantic relationships between data-points. The authors transformed the hash learning into the graph partitioning and developed the spectral algorithm to complete the coding problem with the relaxation. Based on the spectral hashing, the kernelized version was proposed [54]. Through adding the kernel functions, the image contents represented by the original visual features could be fully mapped to the hash codes, which enhances the hashing performance. Heo et al. [55,56] proposed the spherical hashing to learn the hash functions for mapping more spatially coherent images into similar binary codes. In addition, the spherical hamming distance was developed to match up the spherical hashing. Shen et al. [57] developed the supervised discrete hashing to simplify the hash optimization (NP-hard problem in general) by reformulating the objective function using an auxiliary variable. Through dividing the NP-hard hashing into several sub-problems, the normal regularization algorithm (such as cyclic coordinate descent) could be used to optimize the binary codes. A two-step hash learning algorithm was introduced in [58]. The first step aimed to define a unified formulation for both supervised and unsupervised hashing. The second step focused on transforming the bit learning into the binary quadratic problem. In the RS community, Demir and Bruzzone [5] selected the kernel-based nonlinear hashing method [59] to complete the large-scale RSIR. The comprehensive experiments proved the feasibility of hash learning for RS images. Although the methods discussed above are reasonable and feasible, their performance is still limited due to the input data which is the handcrafted visual feature.
In the second group, the deep neural network is incorporated to accomplish hash learning, we name them deep hashing. The earliest deep hash learning method might be semantic hashing [60], in which the restricted Boltzmann machine (RBM) was introduced to map the documents from the original feature space to the binary space. The learning process was completed by two steps, i.e., pre-training and fine-tuning, which aim to generate the low-dimensional vectors and binary-like codes, respectively. Xia et al. [36] proposed another two-stage deep hashing method named supervised hashing. In the first stage, the approximate hash codes of the training images were obtained by decomposing the pairwise similarity matrix without considering the visual contents. Then, a CNN model was developed to learn the deep features and final hash codes with the guidance of semantic labels and obtained approximate hash codes. An end-to-end deep hashing approach was presented in [34] to deal with multi-label image retrieval. Both the semantic ranking and CNN model were combined to learn hash functions so that the obtained hash codes could remain the multiple similarities between the images in the semantic space. Do et al. [61] developed a binary deep neural network to learn the hash codes in an unsupervised manner. Considering the key points within the hash learning, the authors designed the specific loss function and utilized the alternating optimization algorithm with algebraic relaxation to solve it. Then, they expanded the proposed network to the supervised version to reach the stronger performance. To reduce the time cost of hash learning, Jiang and Li [62] introduced an asymmetric deep supervised hashing model. In this model, only the query images’ hash functions were learned via CNN. The target images’ hash codes were learned by the asymmetric pairwise loss function directly. From this point, the whole hash learning process could be speeded up obviously. In the RS community, Li et al. [40,63] proposed a set of deep hashing methods to deal with the single source and multi-source RSIR tasks. In [40], with the consideration of the characteristics of single source RS images, the authors developed the deep CNN model with the pairwise loss function to learn the hash codes. Two scenarios, i.e., the labeled images were limited or sufficient, were taken into account during the network training. In [63], the source-invariant deep hashing CNN was proposed to cope with the cross-source RSIR task. Besides the normal intra-source pairwise constraints, authors developed the specific loss term for inter-source. Then, the alternating learning strategy was proposed to train the network. The experimental results verified the effectiveness of the proposed method.

3. Methodology

The proposed SDAH network is discussed in this section. Before explaining our SDAH, the adversarial auto-encoder (AAE) is introduced first.

3.1. Adversarial Autoencoder

AAE model [41] is an expansion of auto-encoder (AE). Through the adversarial learning (which is proposed in GAN [64]), AAE can not only learn the latent feature but also impose a certain distribution on the learned feature. From this point, AAE can be regarded as the probabilistic AE. The basic framework of AAE is shown in Figure 2.
There are two networks within the AAE model, including the generator G and discriminator D. The generator G is actually an AE, which aims at learning the latent feature z from the input data x according to the reconstruction error. From the view of distribution, the encoding can be formulated as
q z = x q z x p d x d x ,
where q z indicates the aggregated posterior distribution of the latent feature z , q z x means the encoding distribution, and p d x denotes the data distribution. The decoder uses the latent feature z to generate the reconstructed data x ^ . The generator G can be trained by minimizing the error between x and x ^ , which is often measured by the mean square error (MSE). The discriminator D is a normal multi-layer network, which is added on the top of the learned feature z for guiding q z to match an arbitrary prior distribution p z . The goal of D is distinguishing if the input vector follows the prior distribution p z or not. The AAE model can be trained in the minimax optimization manner which is proposed in [64]. The solution of the minimax optimization is formulated as
min G max D V D , G = E x ˜ p d a t a X log D x + E x ˜ p z z log 1 - D G z .
Apart from the basic model, there are still other two models in the AAE framework. One is supervised AAE, and the other is semi-supervised AAE. For supervised AAE, the label information is incorporated in the decoding stage. First, the label is converted into the one-hot vector. Then, both the label vector and the latent feature z are fed to the decoder of G to reconstruct the input data x . For semi-supervised AAE, the label information is embedded through the adversarial learning. Suppose there are two prior distributions, i.e., categorical and expected distributions. For generator G, after the encoding, two vectors are generated, including the discrete class variable y and the continuous latent feature z . Then, they are used to reconstruct the input data by the decoder. For discriminator D 1 , the prior categorical distribution is imposed to regularize y for ensuring the discrete class variable do not carry any information of latent feature z . For discriminator D 2 , the expected distribution is imposed to force the the aggregated posterior distribution of latent feature z matches the predefined distribution. The reconstructed error, regularization loss and the semi-supervised classification error should be considered under the adversarial learning framework.
Due to the characteristic of the AAE model, especially the semi-supervised version, it can be used to learn the hash code inherently. First, the hash code can be adversarial leaned through imposing a prior uniform distribution with binary values. Second, under the semi-supervised learning framework, the semantic information (labels) of samples can be embedded, which can guarantee the learned binary features are discriminative. In addition, the volume of the labeled data is not large, which leads the hashing method feasible in practice.

3.2. Proposed Deep Hashing Network

Based on the semi-supervised AAE model, we propose our SDAH in this section. Its framework is exhibited in Figure 3, which consists of three components, they are, residual auto-encoder (RAE) based generator G, category discriminator D c , and uniform discriminator D u . Now, we introduce them in detail.
To illustrate the generator G clearly, we first introduce RAE network. Similar to the AE model, there are three components in our RAE, including encoder, hidden layer and decoder. Nevertheless, the following modification is carried out for developing RAE. First, to reduce the difficulty of training and enrich the information of the latent features (i.e., the outputs of the encoder), the residual module [65] is introduced into our RAE. Through adding the residual module, not only the problems of gradient vanishing and degradation can be mitigated but also the layer-wise information loss would be decreased. Second, we divide the hidden layer into two parts, including the class variable y and the hash code b . The class variable y is used to accomplish the semi-supervised learning, while the hash code b is our final target. The work flow of RAE can be summarized as follows. When users input the visual feature v , the encoder with the residual module maps it into y and b . Then, both y and b are fed into the residual decoder for the reconstructed data v ^ . The generator G of our SDAH model is the encoder of RAE, which focuses on generating the class and hash variables.
Generally speaking, the class information is always described by the one-hot vector (single class scenario). However, the generated class variable y in our SDAH is continuous. Thus, the category discriminator D c is developed to impose the categorical distribution [66] on y . Note that the categorical distribution is a discrete probability distribution, which describes the results of a random variable that belongs one of c categories. The category discriminator D c is a normal multi-layer neural network with the softmax output. Its main function is distinguishing if a sample generated by G or not. To this end, we construct a one-hot matrix with the categorical distribution as the “true” input of D c . Apart from regularizing y to follow the categorical distribution, D c can also ensure the generated y only contains the class information. Thus, these generated label representations can be used to accomplish the semi-supervised learning directly.
Besides D c , another multi-layer network is also developed to regularize the generated latent hash code b . Here, we record it uniform discriminator D u . Due to the characteristic of the back propagation algorithm, the generated b is actually a continuous variable rather than a discrete binary code with value 0/1. To obtain the expected binary code, we impose the binary uniform distribution on b using D u . First, a binary matrix with uniform distribution is constructed to be the “true” data for D u . Second, D u forces the generated b to follow the prior binary uniform distribution by the adversarial learning. Through adding the uniform discriminator D u , we can not only learn the pseudo-binary codes but also ensure the codes keep bit balanced. Here, the pseudo-binary code means the values of each bit in b are concentrated around 0 and 1.

3.3. Learning Strategy of Proposed Hashing Network

After introducing the components of our SDAH model, the learning strategy is discussed in this section. The network training can be divided into three parts, including unsupervised reconstruction, adversarial regularization and semi-supervised learning. Before describing them in detail, we first define some important notations for clarity. Suppose there is an image dataset I = I 1 , I N with N RS images. The visual features V = v 1 , , v N of those images have been obtained. The semantic labels s 1 , , s n of a small number of RS images are provided, where n N .

3.3.1. Unsupervised Reconstruction

The main target of unsupervised reconstruction is to train the generator G using all of N samples, and this step is accomplished by the RAE model.
Similar to the traditional AE model, we assume that the encoder of RAE is represented by f · , while the decoder of RAE is denoted by g · . Thus, for an input sample v , the RAE model can be defined as
y , b = f v ; Θ r , v ^ = g y , b ; Θ r ,
where v ^ means the reconstructed sample, and Θ r indicates the weights of RAE. To estimate the encoding and decoding functions, we minimize the mean squared error between v and v ^ . The optimization can be formulated as
L R A E = arg min Θ r i = 1 N v i - v ^ i 2 2 = arg min Θ r i = 1 N v i - g f v ; Θ r ; Θ r 2 2 ,
Different from the conventional AE model, the basic unit of RAE is the residual block rather than the normal fully connection block. Suppose the input of the lth residual block is ϕ l , and the weights corresponding to the lth residual block is Θ r l = Θ r l , i , 1 i n l , where n l denotes the number of layers in the residual block. Therefore, the output of the lth residual block is
φ l = ϕ l + F ϕ l ; Θ r l ,
where F · means the residual function that completed by several fully connection layers. After the element-wise addition, the input of the ( l + 1 ) th residual block can be defined as ϕ l + 1 = a c t i v e φ l , where a c t i v e · denotes the activation function. If we select the linear activation function, the input of the ( l + 1 ) th residual block can be written as
ϕ l + 1 = ϕ l + F ϕ l ; Θ r l .
Consequently, the recurrence formula of the L residual blocks can be defined as
ϕ L = ϕ l + i = l L - 1 F ϕ i ; Θ r i .
When we use the backward propagation algorithm to optimize L R A E , the chain rule of derivative should be adopted, i.e.,
L R A E ϕ l = L R A E ϕ L ϕ L ϕ l = L R A E ϕ L 1 + ϕ l i = l L - 1 F ϕ i ; Θ r i .
To further explain the residual block, we exhibit a comparison between the fully connection block and the residual block in Figure 4. Figure 4a shows a normal fully connection block. The output of the ( l ) th block φ l can be obtained using mapping function F · directly. In contrast, the output of the ( l ) th residual block (displayed in Figure 4b) should be acquired by the summation of ϕ l and results of F · .
After training the RAE, we can obtain the generator G of our SDAH model, which is actually the encoder of RAE.

3.3.2. Adversarial Regularization

When we learn the generator G, two latent variables y and b can also be obtained. As mentioned in Section 3.2, we wish that y could reflect the semantic label information and b could be the expected binary code. To this end, two discriminators D c and D u are developed to impose the certain distributions on y and b .
Now, the problem is how to construct the “true” input data with the proper distribution for two discriminators. For D c , we construct the “true” input data as follows. First, n c a one-hot column vectors m c a R c × 1 are generated randomly, where c means the number of semantic class. Then, we combine them together to construct a matrix M c a R c × n c a . It is not difficult to find that the constructed M c a follows the category distribution naturally. For D u , we construct the “true” input data under the following conditions. First, we define a matrix M u n R K × n u n with the uniform distribution, where K indicates the length of expected hash code. In addition, the elements within M u n are 0 or 1. Then, we further let M u n · M u n T = I · K / 2 .
When the “true” input data for discriminators is decided, the next step is imposing the prior distributions on the generated latent variables through the adversarial learning. For clarity, we separate the generator G into G c and G u , which correspond to y and b , respectively. Then, the adversarial regularization procedure can be formulated by Equations (9) and (10).
min G c max D c E x p M c a log D c x + E y p y log 1 - D c G c y
min G u max D u E x p M u n log D u x + E b p b log 1 - D u G u b
There is another point we want to explain further: the value of n c a and n u n . The data used for training the deep neural network are usually divided into several batches. For different batches, we would construct different “true” data with the same distribution for D c and D u . Thus, the values of n c a and n u n are equal to the batch size n b .

3.3.3. Semi-Supervised Learning

After the unsupervised reconstruction and the adversarial regularization, we can obtain the latent variable y with the category distribution and the binary-like variable b with the uniform distribution. Nevertheless, since there are no other specific operations, the current code b does not have any discriminative information, which is not proper for the retrieval task. Thus, we develop the semi-supervised learning part to improve b .
First, to ensure b is discriminative, we adopt the classification loss function to embed the semantic information. As mentioned at the beginning of this section, there are n samples with the semantic labels s 1 , , s n . Assume that the class variables y 1 , , y n corresponding to those samples have been obtained. Then, the cross entropy function is selected to measure the classification error, and the formula is
L c l s = - i n s i log y i .
Note that there are two common formats of the cross entropy function [66,67,68], and we select the current version (Equation (11)) since the softmax function is selected to accomplish the classification.
Second, to remain the similarity relationships between samples from the original feature space to hash code space, and further introduce the semantic label information, we design the similarity preserving loss function as follow,
L s i m = i = 1 n j = 1 n r i j b i - b j 2 2 + 1 - r i j max m - b i - b j 2 2 , 0 ,
where r i j indicates the semantic relationships between two samples, b i - b j 2 2 measures the distance between two samples, and m > 0 denotes the margin parameter. For r i j , when two samples belong to the same semantic class r i j = 1 and otherwise r i j = 0 . This function constrains the similar samples should be mapped into similar hash codes.
To solve the proposed similarity preserving loss function using the back propagation algorithm, some characteristics of L s i m should be analyzed, including the convexity and the differentiability. For convexity, there are two terms within the objective function L s i m , and both of them are the convex functions. In addition, since the value of r i j is equal to 0 or 1, the coefficient 1 - r i j is nonnegative. Thus, the linear combination of two terms, i.e., L s i m , is also a convex function. For differentiability, due to the max operation, the object function is non-differentiable at some certain points [35]. To optimize Equation (12) smoothly, the sub-gradients is selected to replace the normal gradients. In addition, we define the sub-gradients to be 1 at such differentiable points. In detail, the two terms within L s i m are recorded L s i m 1 and L s i m 2 for short. Then, their sub-gradients can be formulated by Equations (13) and (14). After calculating the sub-gradients over the batches, the back propagation algorithm can be carried out normally.
L s i m 1 b i = 2 r i j b i - b j , L s i m 2 b j = - 2 r i j b i - b j .
L s i m 2 b i = - 2 1 - r i j b i - b j , b i - b j 2 2 < m 0 , o t h e r w i s e , L s i m 2 b j = 2 1 - r i j b i - b j , b i - b j 2 2 < m 0 , o t h e r w i s e .
Third, to decrease the quantization error, we introduce the following loss function in bit level [69],
L q u a n = - i = 1 n k = 1 K b i k log b i k + 1 - b i k log 1 - b i k ,
where b i k indicates the kth bit in the hash code. Through minimizing this function, the hash bits can be pushed toward 0 or 1. This can help to reduce the loss of quantization.
Finally, the discussed loss function is linearly combined together and the semi-supervised learning can be formulated as
L h a s h = arg min y , b λ c L c l s + λ s L s i m + λ q L q u a n ,
where λ c , λ s and λ q are the free parameters for adjusting the contribution of each term.

3.3.4. Flow of Learning Strategy

We adopt the stochastic learning method to train our SDAH model. The three parts discussed above are trained alternatively. In particular, the unsupervised reconstruction loss function L R A E is optimized first to initial the parameters of the generator G. Then, the parameters of the generator G and discriminators D c and D u are updated by Equations (9) and (10). Third, the generated latent variables y and b are optimized using L h a s h . The first two steps are accomplished using all N samples, while the third step is achieved using n labeled samples. It is note that the learned variable b is a binary-like vector that the elements within the vector are near 1 or 0. To obtain the real discrete binary code, we set a threshold t. When b i > t , b i = 1 , and otherwise b i = 0 . In this paper, we set t = 0.5 unless otherwise stated.

4. Dataset Introduction

To testify our SDAH model is effective for RSIR task, we selected three published RS image sets. The first one is a high resolution land-use image set constructed by University of California Merced [43,70]. We refer to it as UCM in this paper. There are 2100 RS images with the pixel resolution of one foot in UCM. Their size is 256 × 256 . All of those images are equally divided into 21 semantic categories, including “agricultural”, “airplane”, etc. The examples of RS images corresponding to each category are exhibited in Figure 5, and the names and volume of different scenes are summarized in Table 1. The second one is an aerial image dataset constructed by Wuhan University and HuaZhong University of Science and Technology [71]. Here, we name it AID for short. In AID, there are 10,000 aerial images with the fixed size of 600 × 600 , and their pixel resolution covers from 0.5 m to 8 m. All of the images are partitioned into 30 semantic scenes, including “Airport”, “Bare Land”, etc. The number of images within each scene is not balanced. The scene with the minimal images (220) is “Baseball Field”, while the categories with maximum images (420) are “Pond” and “Viaduct”. Some examples of different scenes are exhibited in Figure 6, and their names and volume are displayed in Table 2. The last one is a 45-scene-class RS image dataset published by Northwestern Polytechnical University [72], and we name it NWPU for short in this paper. There are 700 images for each scene category, and the spatial resolution of those images ranges from 0.2 m to 30 m. The size of images within NWPU is 256 × 256 . Some examples of different classes are displayed in Figure 7, and the names and volume of these scenes can be found in Table 3.

5. Experiments

5.1. Experimental Settings

To generate the hash code using our SDAH, some preparatory work should be completed in advance. First, the structure of SDAH should be decided. Suppose the dimension of input data is d, the number of semantic classes is c, and the bit of expected hash codes is K. Then, our hashing network structure is summarized in Table 4. Second, the training data for SDAH need to be prepared. We randomly selected 20% of images from the original dataset to be the training data. The influence of training data with different proportions is discussed in Section 5.4. Third, the visual features of the RS images need to be extracted. Here, to prove the effectiveness of SDAH, we adopted five common visual features. The details of the performance based on these visual features are discussed in Section 5.2. Fourth, the values of some parameters need to be decided, including the margin parameter m (Equation (12)), and the weighting parameters λ c , λ s and λ q (Equation (16)). All of the parameters were tuned by the k-fold cross-validation in this study. For various RS image archives and diverse visual features, the optimal parameters are different. The details are displayed in Section 5.2. In addition, we used the Adam algorithm [73] to optimize our adversarial hashing network. The learning rate and iteration were set to 1 × 10 - 4 and 50, respectively. All experiments were accomplished using an HP 840 high-performance computer with GeForce GTX Titan X GPU.
We selected precision–recall (P-R curve) and mean average precision (MAP) to verify SDAH’s retrieval behavior numerically. Assume that we obtain n r retrieval results for a query image q . The number of correct retrieval results is n r c , and the number of the correct target images within the archive is n r t . Thus, the precision and recall can formulated as n r c / n r and n r c / n r t . Here, we define that a retrieved/target image is correct if it belongs to the same semantic category as query. Furthermore, let us impose the labels l 1 , l 2 , , l n r on the results. If the retrieved image is correct, l i = 1 , and otherwise l i = 0 . Then, MAP can be defined as
1 Q Q 1 n r i = 1 n r l i i j = 1 i l i ,
where Q means the number of query images.

5.2. Retrieval Performance Based on Different Visual Features

The performance of our SDAH model based on different visual features was studied. Considering the characteristics of RS images, we selected five visual features to represent RS images:, SIFT-based BOW features [70], the outputs of the seventh fully connected layer of Alexnet [31] and VGG16 [74], and the outputs of the seventh fully connected layer of fine-tuned Alexnet and VGG16. These features are common in the RS community [43,75,76]. Here, we refer to them as BOW, AlexFC7, VGG16FC7, AlexFineFC7, and VGG16FineFC7, respectively, for short. Among those features, BOW is mid-level and theothers are high-level features. To be fair, their dimension was set to 4096. In addition, 80% of images from RS archives were selected randomly to fine tune two deep networks for feature extraction. As mentioned above, the parameters (m, λ c , λ s , and λ q ) of our model depend on the different input visual features. After the k-fold cross-validation, the optimal parameters for different features and different RS image archives are displayed in Table 5.
The experimental results are exhibited in Table 6, which are from three datasets using the Top 50 retrieval results. The “Baseline” in the table indicates the retrieval performance of visual features. For measuring the similarities between visual feature vectors, we selected Cosine (BOW) and L2-norm (AlexFC7, VGG16FC7, AlexFineFC7, and VGG16FineFC7) distances. The retrieval results were obtained according to the distance orders. For our hash codes, we set their length to be 32, 64, 128, 256, and 512, respectively, to study their retrieval behavior. Moreover, we selected the hamming distance to weigh the resemblance between hash codes. The retrieval results were also obtained in accordance with the distance orders.
In Table 6, we can easily find the following points. First, our SDAH model can generate useful hash codes based on different visual features. The strongest hash codes were obtained from the VGG16FineFC7 features, while the weakest binary vectors were learned using BOW features. This illustrates that the original visual features are important to hash learning. In other words, we could obtain the satisfactory hash codes if the original visual features can represent the complex contents within RS images well. The current visual features, from the mid-level to the high-level, capture the RS images’ contents from different aspects. Nevertheless, it is obvious that the fine-tuned VGG16 explores more useful information from images, which can be proved by the “Baseline” scores. Thus, the hash codes based on VGG16FineFC7 features achieve the best retrieval performance. Second, the retrieval performance of hash codes is better than that of the original visual features. The reason is that we introduce the semantic information into our hashing network, which can enhance the discrimination of the hash codes. Third, the hash codes’ retrieval behavior becomes stronger with the increase of the bit, which demonstrates that longer codes perform better. This is owing to our adversarial learning framework and the specific hashing function, which ensures the obtained binary codes are bit-balanced and low quantization loss. Thus, the longer codes bring richer information so that better performance can be reached. The encouraging experimental results displayed in Table 6 prove that our SDAH is useful to hash learning. For clarity, we fix AlexFineFC7 to be the input data for the rest of the experiments.

5.3. Retrieval Behavior Compared with Diverse Hashing Methods

To study the performance of our SDAH deeply, we selected six popular hash learning methods for comparison:
  • Kernel-based supervised hashing (KSH) [59]. KSH is a classical and successful hash learning method, which aims to map images into the compact binary codes by minimizing/maximizing the hamming distances between similar/dissimilar data pairs. The target hash functions and their algebra relaxation are formulated in the kernel version.
  • Bootstrap sequential projection learning based hashing (BTNSPLH) [77]. BTNSPLH develops a nonlinear function for hash learning, which can also explore the latent relationships between images. Meanwhile, a semi-supervised optimization method based on the bootstrap sequential projection learning is proposed to obtain the binary vectors with the lowest errors during the hash coding.
  • Semi-supervised deep hashing (SSDH) [78]. SSDH proposes a semi-supervised deep neural network to accomplish the hash leaning in the end-to-end fashion. The developed hashing function minimizes the empirical error on both labeled and unlabeled data, which can preserve the semantic similarities and capture underlying data structures simultaneously.
  • Deep quantization network (DQN) based hashing [79]. DQN embeds a hash layer on the top of the normal CNN model to learn the image’s representation and hash codes at the same time. To obtain the useful hash codes, the pairwise cosine loss is developed to remain the similarity relationships between images while the product quantization loss is introduced to reduce the quantization errors.
  • Deep hashing network (DHN) [38]. Similar to DQN, another deep neural network with the hashing layer is developed. DHN also develops the specific loss functions to deal with the issues of similarity preserving and quantization loss.
  • Deep supervised hashing (DSH) [35]. Based on the usual CNN framework, DSH devises a new hashing network to generate the high discriminative hash codes. Besides preserving the similarity relationships using the supervised information, the designed hash function can also reduce the information loss in the binarization stage by imposing the regularization on the real-valued outputs.
Among the comparison methods, KSH and BTNSPLH are the non-deep hashing techniques, while others are the deep hashing methods. In addition, BTNSPLH and SSDH are semi-supervised models and others are supervised algorithms. All of the approaches were accomplished by the codes published by the authors except SSDH. The parameters were set according to the original literature. Note that, to be fair, the proportion of the training to testing set was 2:8 for all methods. Moreover, the training iteration was set to 50. The experimental results based on different hashing methods are exhibited in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.
For UCM dataset, the P-R curves of different hashing methods are exhibited in Figure 8, and the MAPs of diverse hash learning approaches counted by different number of retrieval results are shown in Figure 9. From the observation of figures, we can find that the performance of all hashing models is acceptable, and the hash codes generated by our SDHA outperform others in all scenarios. When the bit of the hash codes is less than 128, the performance of DSH is the weakest. This is because the training data for DSH is limited. As a deep supervised hashing method, DSH needs abundant samples to guarantee its performance. Nevertheless, we only have 420 labeled images (20% of total data), which are not enough for the DSH model, thus its behavior is not as good as expected. The same phenomenon also appears on DHN and DQN models. Their performance is weaker than other compared methods, especially when the bits increases. In contrast, another supervised method, KSH, performs better in most cases. The reason for this is that KSH is a non-deep model, which does not need a large number of training samples to ensure its behavior. The performance of KSH is even better than that of two semi-supervised comparison methods when the length of binary codes equals 512, which illustrates its effectiveness. For BTNSPLH and SSDH, their behavior is better than that of the supervised ones in most situations. This is because not only the supervised information but also the data structure is used to train the models. In addition, the performance of SSDH is stronger than that of BTNSPLH, which demonstrates that the deep hashing is more suitable for UCM dataset. Although the comparison mehods’ behavior is positive, our SDAH can achieve better performance. Taking 128 bits hash codes as the examples, the highest improvements of retrieval precision resulted from our model are 4.64% (SSDH), 8.87% (BTNSPLH), 8.95% (KSH), 13.47% (DQN), 19.12% (DHN), and 26.98% (DSH), while the largest enhancements of retrieval recall achieved by SDAH are 3.46% (SSDH), 10.40% (BTNSPLH), 8.26% (KSH), 12.57% (DQN), 17.97% (DHN), and 26.00% (DSH). Meanwhile, the biggest promotions of retrieval MAP achieved by SDAH are 4.56% (SSDH), 8.79% (BTNSPLH), 9.42% (KSH), 13.73% (DQN), 21.11% (DHN), and 28.26% (DSH).
For the AID archive, the P-R curves of diverse hashing algorithms are shown in Figure 10, and their MAPs counted on a different number of retrieval results are shown displayed in Figure 11. Among the comparison methods, the behavior of DHN and DQN is weaker than others. The performance of DSH and SSDH is close. When the bit length less than 64, SSDH outperforms DSH. Otherwise, the behavior of DSH is better than that of SSDH. A noticed observation is that the performance of DSH increases dramatically when the binary codes get longer, which demonstrates that this deep supervised hashing method needs long bit hash code to ensure its behavior for the AID dataset. Two non-deep hashing approaches, BTNSPLH and KSH, outperform other comparison methods. This is an interesting observation. The reason for this issue is that the iteration times for the deep models are not enough. We believe that the performance of deep models can be enhanced with the increasing iterations. However, time costs may be too large to accept. Compared with the hashing methods discussed above, our model achieves the best performance in all cases. Taking the hash codes with 128 bits as the examples, the largest enhancements of retrieval precision caused by SDAH are 25.83% (SSDH), 14.57% (BTNSPLH), 9.06% (KSH), 35.54% (DQN), 42.15% (DHN), and 20.64% (DSH), while the biggest growth of retrieval recall achieved by SDAH is 26.27% (SSDH), 14.12% (BTNSPLH), 9.51% (KSH), 35.55% (DQN), 41.08% (DHN), and 21.72% (DSH). Meanwhile, the highest promotions of retrieval MAP achieved by SDAH are 24.38% (SSDH), 13.11% (BTNSPLH), 5.75% (KSH), 32.93% (DQN), 38.95% (DHN), and 19.45% (DSH).
For the NWPU archive, the P-R curves of different hashing methods are exhibited in Figure 12, and the MAPs of diverse hash learning approaches counted by different number of retrieval results are shown in Figure 13. We can find a similar conclusion, i.e., our SDAH model achieves the best performance compared with all comparison methods. For non-deep comparison methods, the behavior of BTNSPLH is stronger than that of KSH when the length of hash codes is 32, 64, 128, and 256 bits. Furthermore, the performance of KSH is improved as the code length increases. When the length of hash codes is 512, KSH outperforms BTNSPLH. This indicates that the KSH can achieve positive results as long as the length of generated codes is large enough. For deep comparison methods, the behavior of SSDH is stronger than that of DQH and DHN and is a little weaker than that of DSH. This demonstrates that both of the supervised knowledge and the data structure can help the hashing model to learn the useful hash code. Similar to the AID dataset, the non-deep models outperform the deep models in most cases. The reason for this is that the volume of images within AID and NWPU is relatively large. Deep models should be trained further to reach better performance. However, this would result in much more time costs. Compared with the other methods, the best retrieval results can be obtained using the hash codes learned by our SDAH model. In addition, let us take the hash codes with 128 bits as the examples; the highest enhancements of retrieval precision resulted from our model are 21.07% (SSDH), 13.86% (BTNSPLH), 18.78% (KSH), 44.27% (DQN), 33.05% (DHN), and 36.44% (DSH), while the largest improvements of retrieval recall achieved by SDAH are 20.26% (SSDH), 9.08% (BTNSPLH), 17.85% (KSH), 41.63% (DQN), 30.87% (DHN), and 34.83% (DSH). Meanwhile, the biggest promotions of retrieval MAP achieved by SDAH are 17.30% (SSDH), 8.93% (BTNSPLH), 10.93% (KSH), 39.09% (DQN), 27.38% (DHN), and 32.84% (DSH).
The encouraging results discussed above prove that our method is effective for RS image retrieval. To further study the SDAH model, we count the MAP values across the diverse semantic categories within three datasets using the Top 50 retrievals, and the results are summarized in Table 7, Table 8 and Table 9. From the observation of tables, we can easily find that the proposed method performs best in most of the categories. For the UCM archive, compared with other hashing methods, the highest enhancements achieved by SDAH appear in “Medium-density Residential” (SSDH, DHN), “Parking Lot” (BTNSPLH), “Buildings” (KSH, DQN), and “Freeway” (DSH). For the AID archive, the largest improvements of our model appear in “Commercial” (SSDH), “Park” (BTNSPLH), “Square” (KSH), “Resort” (DQN, DHN), and “School” (DSH). For the NWPU dataset, the biggest superiority of our method can be found in “Tennis Court” (SSDH, DHN), “Railway Station” (BTNSPLH), “Palace” (KSH), “Basketball Court” (DQN), and “Mountain” (DSH). Moreover, an encouraging observation is that SDAH remains relatively high performance in some categories which other comparison methods’ behavior is unsatisfactory, such as “Dense Residential” (UCM), “Resort” (AID) and “River” (NWPU). It is worth noting that our SDAH cannot achieve the best retrieval results on a few of categories, such as “Baseball Diamond” (UCM), “Pond” (AID) and “Bridge” (NWPU). Nevertheless, its behavior remains in the Top 2 position, which demonstrates the proposed method performs steadily. These favorable results illustrate the usefulness of the proposed method again.
Apart from the numerical assessment, we also study the structure of the learned hash codes using the t-distributed stochastic neighbor embedding (t-SNE) algorithm [80] in this part. In addition, the structure of the hash codes obtained by the comparison methods is provided for reference. Here, the hamming distance is selected to complete the t-SNE algorithm, and the hash codes’ dimension is reduced from 128 to 2. The visual results of different hash codes are exhibited in Figure 14, Figure 15 and Figure 16. From the observation of the figures, we can easily find that the structure of the hash codes learned by our SDAH model (Figure 14a and Figure 15a) is the clearest among all of the results, in which the clusters are obviously separable and the distances between different clusters are distinct. For the comparison methods, similar to the numerical analysis, the results of non-deep hashing (KSH and BTNSPLH) are better than those of deep hashing (DQH, DHN, and DSH), and the results of the semi-supervised hashing (SSDH and BTNSPLH) is better than those of the supervised hashing (DQH, DHN, and DSH) in general. These positive results prove the effectiveness of our model again.
We also display some retrieval examples obtained by different hashing methods in Figure 17, Figure 18 and Figure 19. Here, we adopted the 128 bits hash codes learned by diverse hashing methods and hamming distance metric to obtain the retrieval results. Three RS images were randomly picked up from three archives, which belong to “Storage Tanks” (UCM), “Square” (AID) and “Forest” (NWPU), respectively. Due to the space limitation, only the Top 10 retrieval images are exhibited. The first images within each row are the queries, and then the retrieval results are listed according to the hamming distances’ order. The incorrect results are tagged in red for clarity. In addition, the number of correct results within the Top 50 retrieval images is also provided for reference. It is obvious that our SDAH performs best. For the “Storage Tanks” query, the Top 10 results obtained by SDAH are totally correct, and there are only nine incorrect results within the Top 50 results. For the “Square” and the “Forest” queries, the Top 50 retrieval results obtained by our model are totally correct.

5.4. Influence of Different Parameters

The influence of different parameters on our SDAH model is discussed in this section. As mentioned in Section 5.2, there are four free parameters within our hashing function (Equation (16)), including three weighting variate λ c , λ s and λ q , and one margin parameter m. Besides those parameters, there are still two factors that impact our model: the threshold t for the binarization (introduced in Section 3.3.4) and the percentage of the samples T for training SDAH network. To study their influence on our method, we first fixed the length of hash codes to be 128 bits here. Then, we tuned the values of different parameters with certain limits to observe the variation of our method. For the UCM dataset, we set λ c 0.50 , 0.70 , λ s 0.80 , 1.00 , λ q 0.10 , 0.30 , and m 2.2 , 3.0 . For the AID dataset, we set λ c 0.20 , 0.40 , λ s 0.70 , 0.90 , λ q 0.01 , 0.20 , and m 2.0 , 2.8 . For the NWPU archive, we set λ c 0.15 , 0.35 , λ s 0.80 , 1.00 , λ q 0.01 , 0.20 , and m 2.6 , 3.4 . Moreover, the range of the other two parameters was unified as t 0.1 , 0.9 and T 10 , 50 % for the three image databases. Note that only one parameter was changed at one time, and the others were fixed according to Table 5.
The results on the UCM image database using the Top 50 retrieval images are exhibited in Figure 20. From the observation, we found that the free parameters within the hash function ( λ c , λ s , λ q and m) impact our SDAH model slightly, which demonstrates that our method is robust to the UCM dataset. In detail, for λ c , when λ c 0.60 SDAH’s performance is enhanced with the value of λ c grows. When λ c > 0.60 , the behavior of SDAH is reduced slightly. For λ s , the whole trend of SDAH’s behavior is upward with the increased λ s . The peak value appears at λ s = 0.95 . For λ q , the best performance of SDAH can be obtained when λ q = 0.20 , and the behavior is affected with small degrees when the λ s value is larger or less than 0.20. The results of different m are similar to λ s . SDAH can reach the peak value when m is equal to 2.6, and the performance is decreased in other cases. For the other two parameters, i.e., the binarization threshold t and the training data percentage T, the following conclusion can be summarized. The performance of SDAH almost remains with the varied t, which indicates that our method is not sensitive to t. Furthermore, it also means that the elements within vectors before the binarization are close to 0 or 1, which demonstrates the superiority of our model. For training data percentage T, it is obvious that better performance can be achieved with more training data. However, more training data would result in larger time costs. Thus, in this study, we kept the percentage of the training data at 20%.
The results counted on the AID dataset using the Top 50 retrieval images are shown in Figure 21. For λ c , the peak value appears at λ c = 0.3 . When λ c < 0.3 , the SDAH’s performance is increasing. When λ c > 0.3 , the performance of SDAH first goes down and then goes up. For λ s , the best behavior can be obtained when λ s = 0.8 . Otherwise, SDAH’s performance is decreased more or less. For λ q , the performance of SDAH is decreasing when the value of λ q grows. For m, when m < 2.4 , the model’s performance is ascending and the peak value appears at m = 2.4 . When m > 2.4 , the behavior of SDAH falls slightly. Although the above four parameters impact our model at different levels, their influence remains in an acceptable range. Similar to the UCM dataset, the performance of SDAH varies within a narrow range with different t. This means that the outputs of our model are the binary-like vectors, which proves the usefulness of our SDAH for hash learning. For T, it is obvious that more training data leads to better performance. Taking the time consumption into account, we kept T = 20 % .
The results on the NWPU image archive using the Top 50 retrieval images are exhibited in Figure 22. Similar to the results of other two datasets, the impact of λ c , λ s , λ q and m on SDAH is not drastic. The SDAH’s performance peak values appear at λ c = 0.25 , λ s = 0.60 , λ q = 0.01 , and m = 3.0 . When their values change, the behavior of SDAH fluctuates in a small range. In addition, for training data percentage T, we can find the same results, i.e., the larger training data the better performance. To keep the trade-off between training time costs and performance, we chose T = 20 % for NWPU. An interesting observation is that our model is sensitive to the binarization threshold t for NWPU. The weakest performance is 0.8257 when t = 0.9 while the strongest behavior is 0.8856 when t = 0.5 . The reason for this is that the elements of the vectors before the binarization are not close enough to 0 or 1, which indicates that our model still has the space for the enhancement.

5.5. Computational Cost

In this section, we discuss the computational cost of our model from the following aspects, including the time cost of hash learning and the retrieval efficiency.
The time cost of hash learning can be divided into two parts, i.e., SDAH network training and hash codes generation. For model training, the time costs depend on the volume of the training set directly. Suppose we still select 20% images from the archive to construct the training set. Thus, the numbers of training data corresponding to different archives are 420 (UCM), 2000 (AID), and 6300 (NWPU). In addition, as mentioned in Section 5.2, we chooe AlexFineFC7 as the input visual feature. Therefore, the dimension of the training data was fixed to 4096. Under these conditions, we need around 6 min (UCM), 15 min (AID), and 25 min (NWPU) to accomplish the network training for the 128-bit hash codes. It is noted that the time consumption of model training is similar when the length of hash codes varies between 32 and 512. When the SDAH model is trained, the hash codes generation is fast, needing approximately 0.1 ms for input data.
To deeply study the retrieval speed of SDAH, we changed the archive size and calculated the similarity and ranking time costs under the different lengths of hash codes. In detail, based on the NWPU image dataset, we constructed the target archives with the size of 1000, 5000, 10,000, 15,000, 20,000, 25,000, and 31,500. Then, 300 images were randomly selected to be the queries, and the final retrieval speed was the average value counted on these query images. Meanwhile, the retrieval times based on the original features (AlexFineFC7) are also provided as a reference. Note that the similarities between hash codes were calculated by hamming distance while the resemblance between AlexFineFC7 features was computed by Euclidean distance.
The results are summarized in Table 10. It is obvious that the retrieval based on the hash codes is much faster than the search on the basis of AlexFineFC7 features, especially when the size of the dataset is large. The reasons can be summarized as follows. The AlexFineFC7 features are continuous and dense. To obtain the similarities between these vectors, the distance metric should calculate the differences between every element pairs, which is a time-consuming procedure. Moreover, the dimension of AlexFineFC7 features is 4096, which increases the time costs of similarity computation. In contrast, the hash codes are discrete and their elements are 0 or 1. To measure the resemblance between them, the hamming distance metric only needs to count the number of different element pairs. Furthermore, the dimension of hash codes is low. Consequently, the time costs of similarity matching based on the hash codes are small. The retrieval efficiency results demonstrate that the hash codes are more suitable for the large-scale scenario.

6. Conclusions

Under the paradigm of generative adversarial learning, a new semi-supervised deep hashing method named SDAH is presented in this paper. The generator G of SDAH is the encoder of RAE, which focuses on learning the class variable y and hash code b using the input visual features. Two discriminators ( D c and D u ) aim to impose the categorical distribution and binary uniform distribution on y and b , respectively. Through the regularization, the class variable y could be the approximate one-hot vector that reflects the class information, while the hash code b could be the bit-balanced vector with the binary-like values. The generator and the discriminators can be trained using the total data with the reconstruction loss function and minimax loss function under the generative adversarial learning framework. Furthermore, to improve the performance of hash code, a specific hashing function is developed. Only a small number of labeled data should be used to optimize the hashing function. Integrating the generator, discriminators, and hashing function optimization together, the discriminative, similarity preserving and low quantization error hash code can be obtained.
We selected three published RS archives (UCM, AID, and NWPU) to verify the usefulness of our SDAH. First, our model can learn the hash codes from different visual features, including the mid-level BOW vectors and the high-level deep vectors (AlexFC7, VGG16FC7, AlexFineFC7, and VGG16FineFC7). In addition, the retrieval performance of learned hash codes is better than that of the original visual features. In general, enhancements of around 10% can be achieved. Second, we compared our SDAH with six other popular hash learning methods (SSDH, BTNSPLH, KSH, DQN, DHN, and DSH). Fixing the input feature (AlexFineFC7) and the length hash codes (128 bits), the MAP values of our SDAH counted on Top 50 retrieval results are 0.8286 (UCM), 0.9394 (AID), and 0.8855 (NWPU), which are higher than those of the comparison methods. Third, the learned hash codes can obviously speed up the retrieval efficiency compared with the original visual features. The promising experimental results counted on different RS image datasets prove that our SDAH is effective to the RSIR task.

Author Contributions

X.T. designed the project, oversaw the analysis, and wrote the manuscript; C.L. completed the programming; J.M., X.Z. and F.L. analyzed the data; and L.J. improved the manuscript.

Funding

This work was funded in part by the National Natural Science Foundation of China (Nos. 61801351, 61802190, and 61772400), the National Science Basic Research Plan in Shaanxi Province of China (No. 2018JQ6018), Key Laboratory of National Defense Science and Technology Foundation Project (No. 6142113180302), the China Postdoctoral Science Foundation Funded Project (No. 2017M620441), and the Xidian University New Teacher Innovation Fund Project (No. XJS18032).

Acknowledgments

The authors would like to show their gratitude to the editors and the anonymous reviewers for their comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, W.; Newsam, S.; Li, C.; Shao, Z. PatternNet: A benchmark dataset for performance evaluation of remote sensing image retrieval. ISPRS J. Photogramm. Remote Sens. 2018, 145, 197–209. [Google Scholar] [CrossRef] [Green Version]
  2. Quartulli, M.; Olaizola, I.G. A review of EO image information mining. ISPRS J. Photogramm. Remote Sens. 2013, 75, 11–28. [Google Scholar] [CrossRef] [Green Version]
  3. Shyu, C.R.; Klaric, M.; Scott, G.J.; Barb, A.S.; Davis, C.H.; Palaniappan, K. GeoIRIS: Geospatial information retrieval and indexing system—Content mining, semantics modeling, and complex queries. IEEE Trans. Geosci. Remote Sens. 2007, 45, 839–852. [Google Scholar] [CrossRef] [PubMed]
  4. Aptoula, E. Remote sensing image retrieval with global morphological texture descriptors. IEEE Trans. Geosci. Remote Sens. 2013, 52, 3023–3034. [Google Scholar] [CrossRef]
  5. Demir, B.; Bruzzone, L. Hashing-based scalable remote sensing image search and retrieval in large archives. IEEE Trans. Geosci. Remote Sens. 2015, 54, 892–904. [Google Scholar] [CrossRef]
  6. Gu, Y.; Wang, Y.; Li, Y. A Survey on Deep Learning-Driven Remote Sensing Image Scene Understanding: Scene Classification, Scene Retrieval and Scene-Guided Object Detection. Appl. Sci. 2019, 9, 2110. [Google Scholar] [CrossRef]
  7. Wang, Q.; Chen, M.L.; Nie, F.P.; Li, X.L. Detecting coherent groups in crowd scenes by multiview clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018. [Google Scholar] [CrossRef]
  8. Wang, Q.; Qin, Z.Q.; Nie, F.P.; Li, X.L. Spectral embedded adaptive neighbors clustering. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1265–1271. [Google Scholar] [CrossRef]
  9. Wang, Q.; Wan, J.; Li, X.L. Hierarchical feature selection for random projection. IEEE Trans. Neural Networks and Learning Systems 2018, 30, 1581–1586. [Google Scholar] [CrossRef]
  10. Wang, Q.; Wan, J.; Nie, F.P.; Liu, B. Robust hierarchical deep learning for vehicular management. IEEE Trans. Veh. Technol. 2018, 68, 4148–4156. [Google Scholar] [CrossRef]
  11. Wang, J.; Shen, H.T.; Song, J.; Ji, J. Hashing for similarity search: A survey. arXiv 2014, arXiv:1408.2927. [Google Scholar]
  12. Wang, J.; Zhang, T.; Sebe, N.; Shen, H.T. A survey on learning to hash. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 769–790. [Google Scholar] [CrossRef] [PubMed]
  13. Muja, M.; Lowe, D.G. Fast approximate nearest neighbors with automatic algorithm configuration. VISAPP 2009, 2, 2. [Google Scholar]
  14. Muja, M.; Lowe, D.G. Scalable nearest neighbor algorithms for high dimensional data. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2227–2240. [Google Scholar] [CrossRef] [PubMed]
  15. Indyk, P.; Motwani, R. Approximate nearest neighbors: Towards removing the curse of dimensionality. In Proceedings of the Thirtieth Annual ACM Symposium on Theory of Computing, Dallas, TX, USA, 24–26 May 1998; pp. 604–613. [Google Scholar]
  16. Charikar, M.S. Similarity estimation techniques from rounding algorithms. In Proceedings of the Thiry-Fourth Annual ACM Symposium on Theory of Computing, Montreal, QC, Canada, 19–21 May 2002; pp. 380–388. [Google Scholar]
  17. Andoni, A.; Indyk, P. Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions. Commun. ACM 2008, 51, 117. [Google Scholar] [CrossRef]
  18. Chi, L.; Zhu, X. Hashing techniques: A survey and taxonomy. ACM Comput. Surv. (CSUR) 2017, 50, 11. [Google Scholar] [CrossRef]
  19. Gionis, A.; Indyk, P.; Motwani, R. Similarity search in high dimensions via hashing. In Proceedings of the 25rd International Conference on Very Large Data, Edinburgh, Scotland, UK, 7–10 September 1999; pp. 518–529. [Google Scholar]
  20. Datar, M.; Immorlica, N.; Indyk, P.; Mirrokni, V.S. Locality-sensitive hashing scheme based on p-stable distributions. In Proceedings of the Twentieth Annual Symposium on Computational Geometry, Brooklyn, NY, USA, 8–11 June 2004; pp. 253–262. [Google Scholar]
  21. Lv, Q.; Josephson, W.; Wang, Z.; Charikar, M.; Li, K. Multi-probe LSH: Efficient indexing for high-dimensional similarity search. In Proceedings of the 33rd International Conference on Very Large Data Bases, Vienna, Austria, 23–27 September 2007; pp. 950–961. [Google Scholar]
  22. Li, P.; König, C. b-Bit minwise hashing. In Proceedings of the 19th International Conference on World Wide Web, Raleigh, NC, USA, 26–30 April 2010; pp. 671–680. [Google Scholar]
  23. Li, P.; Konig, A.; Gui, W. b-Bit minwise hashing for estimating three-way similarities. In Proceedings of the Advances in Neural Information Processing Systems 2010, Vancouver, BC, Canada, 6–11 December 2010; Curran Associates, Inc.: Red Hook, NY, USA, 2010; pp. 1387–1395. [Google Scholar]
  24. Gan, J.; Feng, J.; Fang, Q.; Ng, W. Locality-sensitive hashing scheme based on dynamic collision counting. In Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data, Scottsdale, AZ, USA, 20–24 May 2012; pp. 541–552. [Google Scholar]
  25. Cao, Y.; Qi, H.; Zhou, W.; Kato, J.; Li, K.; Liu, X.; Gui, J. Binary hashing for approximate nearest neighbor search on big data: A survey. IEEE Access 2017, 6, 2039–2054. [Google Scholar] [CrossRef]
  26. Weiss, Y.; Torralba, A.; Fergus, R. Spectral hashing. In Proceedings of the Advances in Neural Information Processing Systems 2009, Vancouver, BC, Canada, 7–10 December 2009; Curran Associates, Inc.: Red Hook, NY, USA, 2009; pp. 1753–1760. [Google Scholar]
  27. Liu, W.; Mu, C.; Kumar, S.; Chang, S.F. Discrete graph hashing. In Proceedings of the Advances in Neural Information Processing Systems 2014, Montreal, QC, Canada, 8–13 December 2014; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 3419–3427. [Google Scholar]
  28. Shi, X.; Xing, F.; Cai, J.; Zhang, Z.; Xie, Y.; Yang, L. Kernel-based supervised discrete hashing for image retrieval. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 419–433. [Google Scholar]
  29. Gui, J.; Liu, T.; Sun, Z.; Tao, D.; Tan, T. Fast supervised discrete hashing. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 490–496. [Google Scholar] [CrossRef]
  30. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 2012, Lake Tahoe, NV, USA, 3–8 December 2012; Curran Associates, Inc.: Red Hook, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  32. Erin Liong, V.; Lu, J.; Wang, G.; Moulin, P.; Zhou, J. Deep hashing for compact binary codes learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2475–2483. [Google Scholar]
  33. Lai, H.; Pan, Y.; Liu, Y.; Yan, S. Simultaneous feature learning and hash coding with deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3270–3278. [Google Scholar]
  34. Zhao, F.; Huang, Y.; Wang, L.; Tan, T. Deep semantic ranking based hashing for multi-label image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1556–1564. [Google Scholar]
  35. Liu, H.; Wang, R.; Shan, S.; Chen, X. Deep supervised hashing for fast image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2064–2072. [Google Scholar]
  36. Xia, R.; Pan, Y.; Lai, H.; Liu, C.; Yan, S. Supervised hashing for image retrieval via image representation learning. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014. [Google Scholar]
  37. Wang, D.; Cui, P.; Ou, M.; Zhu, W. Deep multimodal hashing with orthogonal regularization. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
  38. Zhu, H.; Long, M.; Wang, J.; Cao, Y. Deep hashing network for efficient similarity retrieval. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  39. Li, Q.; Sun, Z.; He, R.; Tan, T. Deep supervised discrete hashing. In Proceedings of the Advances in Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 2482–2491. [Google Scholar]
  40. Li, Y.; Zhang, Y.; Huang, X.; Zhu, H.; Ma, J. Large-scale remote sensing image retrieval by deep hashing neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 56, 950–965. [Google Scholar] [CrossRef]
  41. Makhzani, A.; Shlens, J.; Jaitly, N.; Goodfellow, I.; Frey, B. Adversarial autoencoders. arXiv 2015, arXiv:1511.05644. [Google Scholar]
  42. Datcu, M.; Daschiel, H.; Pelizzari, A.; Quartulli, M.; Galoppo, A.; Colapicchioni, A.; Pastori, M.; Seidel, K.; Marchetti, P.G.; d’Elia, S. Information mining in remote sensing image archives: System concepts. IEEE Trans. Geosci. Remote Sens. 2003, 41, 2923–2936. [Google Scholar] [CrossRef]
  43. Yang, Y.; Newsam, S. Geographic image retrieval using local invariant features. IEEE Trans. Geosci. Remote Sens. 2013, 51, 818–832. [Google Scholar] [CrossRef]
  44. Xu, S.; Fang, T.; Li, D.; Wang, S. Object classification of aerial images with bag-of-visual words. IEEE Geosci. Remote Sens. Lett. 2009, 7, 366–370. [Google Scholar]
  45. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  46. Tang, X.; Zhang, X.; Liu, F.; Jiao, L. Unsupervised deep feature learning for remote sensing image retrieval. Remote Sens. 2018, 10, 1243. [Google Scholar] [CrossRef]
  47. Jiao, L.; Tang, X.; Hou, B.; Wang, S. SAR images retrieval based on semantic classification and region-based similarity measure for earth observation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3876–3891. [Google Scholar] [CrossRef]
  48. Tang, X.; Jiao, L.; Emery, W.J. SAR image content retrieval based on fuzzy similarity and relevance feedback. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1824–1842. [Google Scholar] [CrossRef]
  49. Li, Y.; Zhang, Y.; Tao, C.; Zhu, H. Content-based high-resolution remote sensing image retrieval via unsupervised feature learning and collaborative affinity metric fusion. Remote Sens. 2016, 8, 709. [Google Scholar] [CrossRef]
  50. Ferecatu, M.; Boujemaa, N. Interactive remote-sensing image retrieval using active relevance feedback. IEEE Trans. Geosci. Remote Sens. 2007, 45, 818–826. [Google Scholar] [CrossRef]
  51. Demir, B.; Bruzzone, L. A novel active learning method in relevance feedback for content-based remote sensing image retrieval. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2323–2334. [Google Scholar] [CrossRef]
  52. Tang, X.; Jiao, L. Fusion similarity-based reranking for SAR image retrieval. IEEE Geosci. Remote Sens. Lett. 2016, 14, 242–246. [Google Scholar] [CrossRef]
  53. Tang, X.; Jiao, L.; Emery, W.J.; Liu, F.; Zhang, D. Two-stage reranking for remote sensing image retrieval. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5798–5817. [Google Scholar] [CrossRef]
  54. He, J.; Liu, W.; Chang, S.F. Scalable similarity search with optimized kernel hashing. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 25–28 July 2010; pp. 1129–1138. [Google Scholar]
  55. Heo, J.P.; Lee, Y.; He, J.; Chang, S.F.; Yoon, S.E. Spherical hashing. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2957–2964. [Google Scholar]
  56. Heo, J.P.; Lee, Y.; He, J.; Chang, S.F.; Yoon, S.E. Spherical hashing: Binary code embedding with hyperspheres. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 2304–2316. [Google Scholar] [CrossRef] [PubMed]
  57. Shen, F.; Shen, C.; Liu, W.; Tao Shen, H. Supervised discrete hashing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 37–45. [Google Scholar]
  58. Do, T.T.; Doan, A.D.; Nguyen, D.T.; Cheung, N.M. Binary hashing with semidefinite relaxation and augmented lagrangian. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 802–817. [Google Scholar]
  59. Liu, W.; Wang, J.; Ji, R.; Jiang, Y.G.; Chang, S.F. Supervised hashing with kernels. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2074–2081. [Google Scholar]
  60. Salakhutdinov, R.; Hinton, G. Semantic hashing. Int. J. Approx. Reason. 2009, 50, 969–978. [Google Scholar] [CrossRef] [Green Version]
  61. Do, T.T.; Doan, A.D.; Cheung, N.M. Learning to hash with binary deep neural network. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 219–234. [Google Scholar]
  62. Jiang, Q.Y.; Li, W.J. Asymmetric deep supervised hashing. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LO, USA, 2–7 February 2018. [Google Scholar]
  63. Li, Y.; Zhang, Y.; Huang, X.; Ma, J. Learning source-invariant deep hashing convolutional neural networks for cross-source remote sensing image retrieval. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6521–6536. [Google Scholar] [CrossRef]
  64. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 2014, Montreal, QC, Canada, 8–13 December 2014; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 2672–2680. [Google Scholar]
  65. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  66. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  67. De Boer, P.T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A tutorial on the cross-entropy method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  68. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  69. Ghasedi Dizaji, K.; Zheng, F.; Sadoughi, N.; Yang, Y.; Deng, C.; Huang, H. Unsupervised deep generative adversarial hashing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3664–3673. [Google Scholar]
  70. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
  71. Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A benchmark data set for performance evaluation of aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef]
  72. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  73. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  74. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  75. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
  76. Zhou, W.; Newsam, S.; Li, C.; Shao, Z. Learning low dimensional convolutional neural networks for high-resolution remote sensing image retrieval. Remote Sens. 2017, 9, 489. [Google Scholar] [CrossRef]
  77. Wu, C.; Zhu, J.; Cai, D.; Chen, C.; Bu, J. Semi-supervised nonlinear hashing using bootstrap sequential projection learning. IEEE Trans. Knowl. Data Eng. 2012, 25, 1380–1393. [Google Scholar] [CrossRef]
  78. Zhang, J.; Peng, Y. SSDH: Semi-supervised deep hashing for large scale image retrieval. IEEE Trans. Circuits Syst. Video Technol. 2017, 29, 212–225. [Google Scholar] [CrossRef]
  79. Cao, Y.; Long, M.; Wang, J.; Zhu, H.; Wen, Q. Deep quantization network for efficient image retrieval. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  80. Maaten, L.V.D.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Framework of basic remote sensing image retrieval.
Figure 1. Framework of basic remote sensing image retrieval.
Remotesensing 11 02055 g001
Figure 2. Framework of basic adversarial auto-encoder model.
Figure 2. Framework of basic adversarial auto-encoder model.
Remotesensing 11 02055 g002
Figure 3. Framework of the proposed semi-supervised deep adversarial hashing network.
Figure 3. Framework of the proposed semi-supervised deep adversarial hashing network.
Remotesensing 11 02055 g003
Figure 4. Comparison between the normal fully connection block and the the residual block: (a) normal fully connection block; and (b) residual block.
Figure 4. Comparison between the normal fully connection block and the the residual block: (a) normal fully connection block; and (b) residual block.
Remotesensing 11 02055 g004
Figure 5. Examples of different scenes in UCM dataset. The names of scenes can be found in Table 1.
Figure 5. Examples of different scenes in UCM dataset. The names of scenes can be found in Table 1.
Remotesensing 11 02055 g005
Figure 6. Examples of different scenes in AID dataset. The names of scenes can be found in Table 2.
Figure 6. Examples of different scenes in AID dataset. The names of scenes can be found in Table 2.
Remotesensing 11 02055 g006
Figure 7. Examples of different scenes in NWPU dataset. The names of scenes can be found in Table 3.
Figure 7. Examples of different scenes in NWPU dataset. The names of scenes can be found in Table 3.
Remotesensing 11 02055 g007
Figure 8. Retrieval precision–recall curve of different hashing methods counted on UCM archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Figure 8. Retrieval precision–recall curve of different hashing methods counted on UCM archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Remotesensing 11 02055 g008
Figure 9. Mean average precision counted on different number of retrieval images using diverse hashing methods for UCM archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Figure 9. Mean average precision counted on different number of retrieval images using diverse hashing methods for UCM archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Remotesensing 11 02055 g009
Figure 10. Retrieval precision–recall curve of different hashing methods counted on AID archive: (a) The length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Figure 10. Retrieval precision–recall curve of different hashing methods counted on AID archive: (a) The length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Remotesensing 11 02055 g010
Figure 11. Mean average precision counted on different number of retrieval images using diverse hashing methods for AID archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Figure 11. Mean average precision counted on different number of retrieval images using diverse hashing methods for AID archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Remotesensing 11 02055 g011
Figure 12. Retrieval precision–recall curve of different hashing methods counted on NPWU archive: (a)The length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Figure 12. Retrieval precision–recall curve of different hashing methods counted on NPWU archive: (a)The length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Remotesensing 11 02055 g012
Figure 13. Mean average precision counted on different number of retrieval images using diverse hashing methods for NPWU archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Figure 13. Mean average precision counted on different number of retrieval images using diverse hashing methods for NPWU archive: (a) the length of hash codes is 32 bits; (b) the length of hash codes is 64 bits; (c) the length of hash codes is 128 bits; (d) the length of hash codes is 256 bits; and (e) the length of hash codes is 512 bits.
Remotesensing 11 02055 g013
Figure 14. Two-dimensional scatterplots of different hash codes (128 bits) obtained by t-SNE over UCM with 21 semantic scenes. The relationships between number ID and scenes can be found in Table 1: (a) SDAH (Our method); (b) SSDH; (c) BTNSPLH; (d) KSH; (e) DQH; (f) DHN; and (g) DSH.
Figure 14. Two-dimensional scatterplots of different hash codes (128 bits) obtained by t-SNE over UCM with 21 semantic scenes. The relationships between number ID and scenes can be found in Table 1: (a) SDAH (Our method); (b) SSDH; (c) BTNSPLH; (d) KSH; (e) DQH; (f) DHN; and (g) DSH.
Remotesensing 11 02055 g014
Figure 15. Two-dimensional scatterplots of different hash codes (128 bits) obtained by t-SNE over AID with 30 semantic scenes. The relationships between number ID and scenes can be found in Table 2: (a) SDAH (Our method); (b) SSDH; (c) BTNSPLH; (d) KSH; (e) DQH; (f) DHN; and (g) DSH.
Figure 15. Two-dimensional scatterplots of different hash codes (128 bits) obtained by t-SNE over AID with 30 semantic scenes. The relationships between number ID and scenes can be found in Table 2: (a) SDAH (Our method); (b) SSDH; (c) BTNSPLH; (d) KSH; (e) DQH; (f) DHN; and (g) DSH.
Remotesensing 11 02055 g015
Figure 16. Two-dimensional scatterplots of different hash codes (128 bits) obtained by t-SNE over NWPU with 45 semantic scenes. The relationships between number ID and scenes can be found in Table 3: (a) SDAH (Our method); (b) SSDH; (c) BTNSPLH; (d) KSH; (e) DQH; (f) DHN; and (g) DSH.
Figure 16. Two-dimensional scatterplots of different hash codes (128 bits) obtained by t-SNE over NWPU with 45 semantic scenes. The relationships between number ID and scenes can be found in Table 3: (a) SDAH (Our method); (b) SSDH; (c) BTNSPLH; (d) KSH; (e) DQH; (f) DHN; and (g) DSH.
Remotesensing 11 02055 g016
Figure 17. Retrieval examples of “Storage Tanks” within UCM dataset based on the 128 bits hash codes learned by the different hash learning methods. The first images in each row are the queries. The remaining images in each row are the Top 10 retrieval results. The incorrect results are tagged in red, and the number of correct results among Top 50 retrieval images is provided for reference.
Figure 17. Retrieval examples of “Storage Tanks” within UCM dataset based on the 128 bits hash codes learned by the different hash learning methods. The first images in each row are the queries. The remaining images in each row are the Top 10 retrieval results. The incorrect results are tagged in red, and the number of correct results among Top 50 retrieval images is provided for reference.
Remotesensing 11 02055 g017
Figure 18. Retrieval examples of “Square” within AID dataset based on the 128 bits hash codes learned by the different hash learning methods. The first images in each row are the queries. The remaining images in each row are the Top 10 retrieval results. The incorrect results are tagged in red, and the number of correct results among Top 50 retrieval images is provided for reference.
Figure 18. Retrieval examples of “Square” within AID dataset based on the 128 bits hash codes learned by the different hash learning methods. The first images in each row are the queries. The remaining images in each row are the Top 10 retrieval results. The incorrect results are tagged in red, and the number of correct results among Top 50 retrieval images is provided for reference.
Remotesensing 11 02055 g018
Figure 19. Retrieval examples of “Forest” within NWPU dataset based on the 128 bits hash codes learned by the different hash learning methods. The first images in each row are the queries. The remaining images in each row are the Top 10 retrieval results. The incorrect results are tagged in red, and the number of correct results among Top 50 retrieval images is provided for reference.
Figure 19. Retrieval examples of “Forest” within NWPU dataset based on the 128 bits hash codes learned by the different hash learning methods. The first images in each row are the queries. The remaining images in each row are the Top 10 retrieval results. The incorrect results are tagged in red, and the number of correct results among Top 50 retrieval images is provided for reference.
Remotesensing 11 02055 g019
Figure 20. Influence of different parameters on our hash learning model. The results are counted on the UCM dataset using the 128 bits hash codes: (a) λ c ; (b) λ s ; (c) λ q ; (d) m; (e) t; and (f) T.
Figure 20. Influence of different parameters on our hash learning model. The results are counted on the UCM dataset using the 128 bits hash codes: (a) λ c ; (b) λ s ; (c) λ q ; (d) m; (e) t; and (f) T.
Remotesensing 11 02055 g020
Figure 21. Influence of different parameters on our hash learning model. The results are counted on the AID dataset using the 128 bits hash codes: (a) λ c ; (b) λ s ; (c) λ q ; (d) m; (e) t; and (f) T.
Figure 21. Influence of different parameters on our hash learning model. The results are counted on the AID dataset using the 128 bits hash codes: (a) λ c ; (b) λ s ; (c) λ q ; (d) m; (e) t; and (f) T.
Remotesensing 11 02055 g021
Figure 22. Influence of different parameters on our hash learning model. The results are counted on the NWPU dataset using the 128 bits hash codes: (a) λ c ; (b) λ s ; (c) λ q ; (d) m; (e) t; and (f) T.
Figure 22. Influence of different parameters on our hash learning model. The results are counted on the NWPU dataset using the 128 bits hash codes: (a) λ c ; (b) λ s ; (c) λ q ; (d) m; (e) t; and (f) T.
Remotesensing 11 02055 g022
Table 1. Names and volume of different scenes in UCM dataset.
Table 1. Names and volume of different scenes in UCM dataset.
Scene
Number
SceneVolumeScene
Number
SceneVolume
1Agricultural10012Intersection100
2Airplane10013Medium Residential100
3Baseball Diamond10014Mobile Home Park100
4Beach10015Overpass100
5Buildings10016Parking Lot100
6Chaparral10017River100
7Dense Residential10018Runway100
8Forest10019Sparse Residential100
9Freeway10020Storage Tanks100
10Golf Course10021Tennis Court100
11Harbor100
Table 2. Names and volume of different scenes in AID dataset.
Table 2. Names and volume of different scenes in AID dataset.
Scene
Number
SceneVolumeScene
Number
SceneVolume
1Airport36016Mountain340
2Bare Land31017Park350
3Baseball Field22018Parking390
4Beach40019Playground370
5Bridge36020Pond420
6Center26021Port380
7Church24022Railway Station260
8Commercial35023Resort290
9Dense Residential41024River410
10Desert30025School300
11Farmland37026Sparse residential300
12Forest25027Square330
13Industrial39028Stadium290
14Meadow28029Storage tanks360
15Medium Residential29030Viaduct420
Table 3. Names and volume of different scenes in NWPU dataset.
Table 3. Names and volume of different scenes in NWPU dataset.
Scene
Number
SceneVolumeScene
Number
SceneVolume
1Airplane70024Medium Residential700
2Airport70025Mobile Home Park700
3Baseball Diamond70026Mountain700
4Basketball Court70027Overpass700
5Beach70028Palace700
6Bridge70029Parking Lot700
7Chaparral70030Railway700
8Church70031Railway Station700
9Circular Farmland70032Rectangular Farmland700
10Cloud70033River700
11Commercial Area70034Roundabout700
12Dense Residential70035Runway700
13Desert70036Sea Ice700
14Forest70037Ship700
15Freeway70038Snow Berg700
16Golf Course70039Sparse Residential700
17Ground Track Field70040Stadium700
18Harbor70041Storage Tank700
19Industrial Area70042Tennis Court700
20Intersection70043Terrace700
21Island70044Thermal Power Station700
22Lake70045Wetland700
23Meadow700
Table 4. Structure of our semi-supervised deep adversarial hashing.
Table 4. Structure of our semi-supervised deep adversarial hashing.
Residual Auto-EncoderEncoder (Generator) d d / 2 d / 2 d / 2 d / 4 d / 4 d / 4
Latentc, K
Decoder d / 4 d / 4 d / 4 d / 2 d / 2 d / 2 d
Category Discriminator c d / 2 d / 4
Uniform Discriminator K d / 2 d / 4
Table 5. Optimal parameters for different visual features and different image archives.
Table 5. Optimal parameters for different visual features and different image archives.
BOWAlexFC7VGG16FC7AlexFineFC7VGG16FineFC7
UCM λ c 0.500.200.900.600.40
λ s 0.500.750.700.950.80
λ q 0.650.050.050.200.10
m2.601.401.002.604.00
AID λ c 1.000.450.500.300.35
λ s 0.850.950.850.800.70
λ q 0.010.010.010.010.01
m1.602.401.402.402.20
NWPU λ c 0.900.950.800.250.90
λ s 0.100.550.450.600.10
λ q 0.050.010.010.010.05
m3.003.002.003.003.00
Table 6. Retrieval mean average precision counted on three archives based on different visual features using Top 50 retrieval results.
Table 6. Retrieval mean average precision counted on three archives based on different visual features using Top 50 retrieval results.
BOWAlexFC7VGG16FC7AlexFineFC7VGG16FineFC7
UCMBaseline0.33380.48150.46130.65510.8027
Hash codes bits320.43620.70640.64140.81640.8994
640.44290.73140.70010.82140.9173
1280.45030.73190.73160.82860.9225
2560.47340.74710.73690.83190.9277
5120.48010.75720.75030.84280.9269
AIDBaseline0.27830.46830.42400.82760.9516
Hash codes bits320.38500.59750.62340.92300.9623
640.49430.72280.66250.93790.9673
1280.50180.73460.70640.93940.9754
2560.51140.75190.72840.94480.9755
5120.51590.76410.73190.94660.9779
NWPUBaseline0.22840.41280.41230.69810.9044
Hash codes bits320.34690.60200.61810.87150.9324
640.36090.65570.62340.88160.9572
1280.40480.67170.65230.88560.9614
2560.45240.67270.67080.89160.9688
5120.46760.68230.67950.89880.9710
Table 7. Retrieval Mean average precision of different hash learning methods across 21 semantic categories in UCM archive counted by Top 50 results. The retrieval results are obtained by the hash codes with 128 bits. The values in bold mean that highest mean average precision of different methods.
Table 7. Retrieval Mean average precision of different hash learning methods across 21 semantic categories in UCM archive counted by Top 50 results. The retrieval results are obtained by the hash codes with 128 bits. The values in bold mean that highest mean average precision of different methods.
SDAHSSDHBTNSPLHKSHDQNDHNDSH
Agriculture0.94870.94160.93490.93040.93860.93760.7457
Airplane0.94380.83730.81910.85100.92450.93400.4622
Baseball Diamond0.89310.93130.93780.81970.90500.85970.6699
Beach0.94970.99520.99550.96070.96470.93030.9383
Buildings0.67720.53530.79380.31250.26990.30370.3152
Chaparral0.97230.96540.94510.96850.96110.95620.7813
Dense Residential0.47960.34360.18340.25330.15220.16450.2382
Forest0.94870.99020.97000.99730.98030.94420.8299
Freeway0.96420.81150.66500.82530.75180.53100.4530
Golf Course0.76730.63570.76340.62260.64450.53730.4514
Harbor0.94520.92330.70230.97470.87250.75590.8240
Intersection0.78130.76960.77870.59060.46920.35670.5202
Medium-density Residential0.81820.57400.58770.72100.41760.38050.3175
Mobile Home Park0.72950.71570.65300.53040.54770.40640.4911
Overpass0.80700.80170.73140.60440.53130.39830.6388
Parking Lot0.94360.98090.49280.98240.97300.82940.7406
River0.81020.79060.72100.62560.52850.45580.4457
Runway0.91880.90990.86650.91460.82820.66440.8025
Sparse Residential0.88420.80720.96410.72470.62820.60980.4173
Storage Tanks0.57430.56430.24030.50590.53870.47350.2197
Tennis Courts0.64340.46960.78790.63730.55880.51840.1638
Average0.82860.77590.73970.73110.68510.61660.5460
Table 8. Retrieval Mean average precision of different hash learning methods across 30 semantic categories in AID archive counted by Top 50 results. The retrieval results are obtained by the hash codes with 128 bits. The values in bold mean that highest mean average precision of different methods.
Table 8. Retrieval Mean average precision of different hash learning methods across 30 semantic categories in AID archive counted by Top 50 results. The retrieval results are obtained by the hash codes with 128 bits. The values in bold mean that highest mean average precision of different methods.
SDAHSSDHBTNSPLHKSHDQNDHNDSH
Airport0.96610.58310.84480.94030.35970.45880.6298
Bare Land0.98730.79020.67310.98420.36240.53520.8339
Baseball Field0.97710.83070.89100.98040.57750.46450.8126
Beach0.95620.76870.96800.96530.83120.72240.8459
Bridge0.96490.77650.93390.96320.81860.71680.8117
Center0.90980.65770.73600.53050.24140.30020.7136
Church0.91260.55720.70940.81530.20170.25190.6206
Commercial0.93790.47500.75270.92710.30410.26710.6330
Dense Residential0.98270.69080.92880.97840.79000.67190.7718
Desert0.96510.81940.53600.93440.46610.52300.8857
Farmland0.94810.69250.94020.95200.85350.78930.6895
Forest0.97170.91830.89950.95660.88760.86310.9263
Industrial0.93530.57450.79460.91290.48370.38540.6467
Meadow0.97070.91880.80490.96910.82130.82950.9186
Medium Residential0.92810.67190.77800.86680.46770.44040.6896
Mountain0.96860.77840.96750.96000.89290.76060.8622
Park0.88840.57170.33210.66580.19910.26670.6503
Parking0.98940.91160.94360.97630.99460.94590.9513
Playground0.95020.73300.93090.97040.74960.64400.7991
Pond0.92540.68670.88090.96930.78670.59930.7784
Port0.96480.74000.80180.96420.59540.51810.7700
Railway Station0.90870.57890.81380.81200.41460.38520.6367
Resort0.84930.47890.57100.42570.11780.10690.5182
River0.96460.63120.88800.96280.64480.33070.6979
School0.79140.36560.62370.63860.10420.11270.3917
Sparse residential0.98280.85750.97240.97980.92270.86080.8533
Square0.85820.36660.47500.35650.19370.12850.5231
Stadium0.95410.89220.93820.97800.87400.75850.9034
Storage tanks0.87670.76680.76320.87410.81340.59350.8577
Viaduct0.99580.80230.93900.98460.97930.93700.8550
Average0.93940.69620.80110.87320.59170.53890.7493
Table 9. Retrieval Mean average precision of different hash learning methods across 45 semantic categories in NMPU archive counted by Top 50 results. The retrieval results are obtained by the hash codes with 128 bits. The values in bold mean that highest mean average precision of different methods.
Table 9. Retrieval Mean average precision of different hash learning methods across 45 semantic categories in NMPU archive counted by Top 50 results. The retrieval results are obtained by the hash codes with 128 bits. The values in bold mean that highest mean average precision of different methods.
SDAHSSDHBTNSPLHKSHDQNDHNDSH
Airplane0.92640.80640.92580.88010.62490.78010.7700
Airport0.86610.56730.39850.64750.23100.49760.2910
Baseball Diamond0.92590.70890.94700.87530.27540.46150.5786
Basketball Court0.85080.53910.93120.55630.12510.34890.4095
Beach0.93810.78830.87970.85490.43000.65060.5875
Bridge0.89050.79990.95250.87080.64610.74020.6834
Chaparral0.97540.96890.93840.96670.93910.94470.7546
Church0.77780.48590.69020.38490.07060.21380.2426
Circular Farmland0.95160.87730.91250.92410.90910.92700.7249
Cloud0.97230.92320.97280.97110.95190.93470.9370
Commercial Area0.78570.49030.81250.49870.19280.41740.3659
Dense Residential0.84340.63570.87550.79720.39670.46700.3909
Desert0.93830.88910.89390.90010.83940.84250.8296
Forest0.94860.86680.57510.93320.85690.82890.6403
Freeway0.87780.58660.79600.51190.32700.34710.4567
Golf Course0.87350.74550.91450.82550.60630.72050.6424
Ground Track Field0.91700.69860.90460.90940.28550.49430.6375
Harbor0.95340.82310.88510.89700.76800.80960.6062
Industrial Area0.88110.65430.57090.76200.55370.60510.4530
Intersection0.88430.72790.84140.81600.31740.66760.6722
Island0.93350.89250.87580.92270.75590.78860.8098
Lake0.92910.80560.81370.83940.71510.74470.7410
Meadow0.93760.86900.46550.87560.78020.74950.7957
Medium Residential0.78690.53560.66550.54480.24600.34780.2664
Mobile Home Park0.86700.72400.90510.85990.29150.54110.5180
Mountain0.90210.70570.89920.83810.64950.66740.2659
Overpass0.94030.74290.90220.82050.43580.61850.5801
Palace0.65880.45390.53260.26170.06590.10530.2487
Parking Lot0.94660.79140.89860.89850.56950.70660.7377
Railway0.89160.61060.70790.77090.41580.44290.5240
Railway Station0.84180.64990.28730.60350.19680.39250.4480
Rectangular Farmland0.87500.65600.78380.81940.69610.71550.3041
River0.83360.58900.79240.56170.30150.47670.3939
Roundabout0.83430.73460.92120.88790.63440.76190.6350
Runway0.84650.72360.78720.70540.37760.52100.6405
Sea Ice0.98120.94670.97950.96060.90850.91460.9311
Ship0.86790.59990.91940.81600.27720.60680.4182
Snowberg0.95530.89500.93690.94470.90710.88670.7867
Sparse Residential0.92850.68380.38090.86310.38580.66060.4903
Stadium0.88540.74990.87820.83020.62310.65260.7206
Storage Tank0.91840.78410.82260.78640.66350.77390.6866
Tennis Court0.79180.44020.89910.66530.09590.19870.2357
Terrace0.85420.66990.72170.80510.37270.59070.3062
Thermal Power Station0.85130.61470.83340.72510.29710.55580.5132
Wetland0.81160.65740.58180.51260.24840.40640.3991
Average0.88550.71350.79580.77560.49460.61170.5571
Table 10. Retrieval cost of similarity calculation and matching based on original features and hash codes with different length (unit: millisecond).
Table 10. Retrieval cost of similarity calculation and matching based on original features and hash codes with different length (unit: millisecond).
Target Image Archive Size1000500010,00015,00020,00025,00031,500
AlexFineFC731.79152.71306.00454.86563.74765.99949.25
Hash codes bits320.601.362.813.143.745.466.25
640.812.474.236.587.9911.0313.08
1281.284.769.6813.9718.0923.4329.56
2562.4110.4021.4532.7843.4054.3668.28
5124.7622.1645.7064.1679.11106.32132.72

Share and Cite

MDPI and ACS Style

Tang, X.; Liu, C.; Ma, J.; Zhang, X.; Liu, F.; Jiao, L. Large-Scale Remote Sensing Image Retrieval Based on Semi-Supervised Adversarial Hashing. Remote Sens. 2019, 11, 2055. https://doi.org/10.3390/rs11172055

AMA Style

Tang X, Liu C, Ma J, Zhang X, Liu F, Jiao L. Large-Scale Remote Sensing Image Retrieval Based on Semi-Supervised Adversarial Hashing. Remote Sensing. 2019; 11(17):2055. https://doi.org/10.3390/rs11172055

Chicago/Turabian Style

Tang, Xu, Chao Liu, Jingjing Ma, Xiangrong Zhang, Fang Liu, and Licheng Jiao. 2019. "Large-Scale Remote Sensing Image Retrieval Based on Semi-Supervised Adversarial Hashing" Remote Sensing 11, no. 17: 2055. https://doi.org/10.3390/rs11172055

APA Style

Tang, X., Liu, C., Ma, J., Zhang, X., Liu, F., & Jiao, L. (2019). Large-Scale Remote Sensing Image Retrieval Based on Semi-Supervised Adversarial Hashing. Remote Sensing, 11(17), 2055. https://doi.org/10.3390/rs11172055

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop