Next Article in Journal
Avoiding Void Holes and Collisions with Reliable and Interference-Aware Routing in Underwater WSNs
Next Article in Special Issue
A Concurrent and Hierarchy Target Learning Architecture for Classification in SAR Application
Previous Article in Journal
Fecal Volatile Organic Compounds in Preterm Infants Are Influenced by Enteral Feeding Composition
Previous Article in Special Issue
Target Recognition of SAR Images via Matching Attributed Scattering Centers with Binary Target Region
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification

Naval Aviation University, Yantai 264001, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(9), 3039; https://doi.org/10.3390/s18093039
Submission received: 10 August 2018 / Revised: 31 August 2018 / Accepted: 7 September 2018 / Published: 11 September 2018
(This article belongs to the Special Issue Automatic Target Recognition of High Resolution SAR/ISAR Images)

Abstract

:
With the continuous development of the convolutional neural network (CNN) concept and other deep learning technologies, target recognition in Synthetic Aperture Radar (SAR) images has entered a new stage. At present, shallow CNNs with simple structure are mostly applied in SAR image target recognition, even though their feature extraction ability is limited to a large extent. What’s more, research on improving SAR image target recognition efficiency and imbalanced data processing is relatively scarce. Thus, a lightweight CNN model for target recognition in SAR image is designed in this paper. First, based on visual attention mechanism, the channel attention by-pass and spatial attention by-pass are introduced to the network to enhance the feature extraction ability. Then, the depthwise separable convolution is used to replace the standard convolution to reduce the computation cost and heighten the recognition efficiency. Finally, a new weighted distance measure loss function is introduced to weaken the adverse effect of data imbalance on the recognition accuracy of minority class. A series of recognition experiments based on two open data sets of MSTAR and OpenSARShip are implemented. Experimental results show that compared with four advanced networks recently proposed, our network can greatly diminish the model size and iteration time while guaranteeing the recognition accuracy, and it can effectively alleviate the adverse effects of data imbalance on recognition results.

1. Introduction

Synthetic aperture radar (SAR) is an active ground observation system that can be installed on aircraft, satellites, spaceships and other flight platforms. Compared with the optical and infrared observation methods, SAR can overcome the adverse effects of weather and perform dynamic observations of ground and ocean targets, so it has bright application prospects in the field of remote sensing. Compared with natural images, SAR images reflect the backscattering intensity of electromagnetic information, so specialist systems are needed to interpret them, but searching for targets of interest in the massive SAR images by humans is time-consuming and extremely difficult, which justifies the urgent need for SAR automatic target recognition (SAR-ATR) algorithms [1]. In the era of big data, there are tons of SAR image data waiting to be processed every day. Therefore, SAR-ATR requires not only high recognition accuracy, but also efficient data processing flows.
The traditional SAR image target recognition methods are mainly composed of independent steps such as preprocessing, feature extraction, recognition and classification. The feature extraction process usually needs scale invariant feature transform (SIFT) [2], histogram of oriented gradient (HOG) [3] and other algorithms to extract good distinguishing features to better complete the classification task. However, both the accuracy and efficiency of SAR image recognition are seriously restricted due to the complicated process and hand-designed features [4,5].
In 2012, the deep CNN [6] proposed by Krizhevsky et al. achieved the error rate considerably lower than the previous state of the art results in the ImageNet Large Scale Visual Recognition Challenge (ILSVRC), and CNN became of great interest to the academic community. Since then, many CNN models such as VGGNet [7], GoogLeNet [8], ResNet [9], DenseNet [10] and SENet [11], have been proposed and constantly challenged the computer’s limit of image cognitive ability. In the last ILSVRC competition in 2017, the Top-5 error rate of the Image Classification task reached 2.251%, exceeding the human recognition level.
The exciting progress of CNN in the field of computer vision (CV) has encouraged people to think about how to apply CNN to target recognition in SAR images, and many scholars have made intensive studies of this topic. Some papers used CNN to accomplish the SAR image target classification experiments on the Moving and Stationary Target Acquisition and Recognition (MSTAR) [12] public data set. The accuracy has gradually increased from 84.7% [1] to more than 99% [13,14], which is higher than that of SVM [15], Cond Gauss [16], AdaBoost [17] and so on. Ding et al. [4] investigated the capability of a CNN combined with three types of data augmentation operations in SAR target recognition. Issues such as translation of target, randomness of speckle noise in different observations, and lack of pose images in training data are intensive studied. Huang et al. [18] studied the influence of different optimization methods on the recognition results of SAR images in CNN model. Huang et al. [19] discussed the problem of SAR image recognition under limited labeled data. Bentes et al. [20] compared four CNN models [4,21,22,23] used in SAR-ATR in recent years, and put forward a multiple resolution input CNN model (CNN-MR). In order to improve the learning ability of the network, the SAR images are processed to different resolution slices in CNN-MR. The performance of the CNN-MR in the experiments indicating that the informative features make a significant contribution to obtain higher accuracy in CNN models, but such data preprocessing will bring extra work. As summarized in [24], while deep learning has become the main tool for tasks like detection in CV on RGB imagery, however, it has not yet had the same impact on remote sensing. As far as we know, many CNN models [1,4,14,22] designed for SAR image recognition are shallow networks, and only a few frontier technologies are utilized in the field of CV. We consider that the following three aspects can be further studied in the task of SAR image recognition with CNN:
  • Many shallow CNN models just consist of several convolution layers, pooling layers and an output layer. The interdependencies between channels and spaces of feature maps are often overlooked. How to improve the expressive ability of CNN and extract informative features through network designing is a valuable research direction.
  • Because the source of SAR image acquisition is greatly limited, some data sets are highly imbalanced. When the traditional machine learning classification method is applied to the imbalanced dataset, the classifier is biased to minority classes in order to improve the overall accuracy, and the classification performance is seriously affected. As far as we know, the problem of data imbalance in SAR image target recognition has not been paid enough attention in the current research yet.
  • The huge amount of parameters is an obstacle when CNN is applied in practice. In SAR image recognition, attention should also be paid to reducing network parameters and computation consumption while ensuring accuracy.
The human visual system (HVS) can automatically locate the salient regions in visual images. Inspired by the HVS mechanism, several attention models are proposed to better understand how the regions of interest (ROIs) are selected in images [25]. The visual attention mechanism has been wildly applied in many prediction tasks such as natural language processing (NLP) [26], image/video caption [27,28], image classification [11,29] etc. In SAR image recognition, Karine et al. [25] combined the SIFT method with a saliency attention model and built a new feature named multiple salient keypoints descriptors (MSKD). MSKD is not used on the whole SAR image, but only the target area. The recognition experiments for both ISAR and SAR images show that MSKD can achieve a significant advantage over SIFT, which indicates that the application of the visual attention mechanism in SAR image recognition is feasible.
SENet [11] is a CNN model based on visual attention mechanism. It uses a gating mechanism to model channel-wise relationships and enhances the representation power of modules throughout the networks [30]. The authors of SENet developed a series of SE blocks that integrate with ResNet [9], ResNext [31] and Inception-ResNet [32], respectively. Experimental results on the ImageNet dataset show that the introduction of SEblock can effectively reduce the error rate. In ILSVRC 2017, SENet won the first place in image classification competition, indicating its effectiveness.
Depthwise separable convolution [33] is a kind of model compression technique that reduces the number of parameters and amount of computation used in convolutional operations while increasing representational efficiency [34]. It consists of a depthwise (DW) convolution, i.e., a spatial convolution performed independently over every channel of an input, followed by a pointwise (PW) convolution, i.e., a regular convolution with 1 × 1 kernel, projecting the channels computed by the DW convolution onto a new channel space. Depthwise separable convolution have been previously shown in Xception [33] to allow for image classification models that outperform similar networks with the same number of parameters, by making more efficient use of the parameters available for representation learning. Many state of the art CNN models such as MobileNets [35], ResNext [31], ShuffelNet [36], SqueezeNet [37] etc. also adopt depthwise separable convolutions to reduce model parameters and accelerate their calculations.
Data imbalance exists widely in practical applications, such as detecting sea surface oil pollution through satellite radar images [38], monitoring illegal trade in credit cards [39], and classifying medical data [40], etc. The general methods of dealing with imbalanced classification problems can be divided into two categories. The first one is data level methods including over-sampling and under-sampling [41,42,43]. The core idea of over-sampling is to randomly copy or expand the data of minority classes, but it easily leads to over fitting problems and deteriorates the generalization ability of the model. The under-sampling method balances the number of each class by removing part of the samples in the majority class, but it often losses some important data, which cause large offset or distortion in the decision boundary. The second is the algorithm level methods represented by the cost sensitive learning [44]. This method generally does not change the original distribution of the training data, but it gives different misclassification costs for different classes, i.e., the misclassification cost of a minority classes is higher than that of majority classes. The cost matrix in cost-sensitive learning is difficult to obtain directly from the data set and misclassification costs are often unknown [45,46]. Buda et al. [47] investigated the impact of class imbalance on the classification performance of CNNs and compared some frequently used methods. Experimental results indicate that over-sampling is almost universally effective in most situations where data imbalance occurs.
Inspired by SENet [11] and the extensive application of depthwise separable convolution, we consider applying them to SAR image recognition tasks. Based on the visual attention mechanism, we first designed a channel-wise and spatial attention block as the basic unit to construct our CNN model. Then, depthwise separable convolution wis utilized to replace the standard convolution in order to decrease network parameters and model size. We also use a new loss function named weighted distance measure (WDM) loss to reduce the influence of data imbalance on the accuracy. The main contributions of our work are:
  • Propose a lightweight CNN model based on visual attention mechanism for SAR image classification. The utilization of channel-wise and spatial attention mechanism can boost the representational power of network. Experiment on MSTAR [12] dataset indicate that compare with CNN model without visual attention mechanism (e.g., ResNet [9], Network in literature [23] and A-ConvNet [22]), our network achieves higher recognition accuracy. Meanwhile, the model parameters and calculation consumption are significantly reduced by using depthwise separable convolution.
  • A new WDM loss function is proposed to solve the data imbalance problem in the data set, and a comparative analysis is done of different ways to deal with the data imbalance problem. Experimental results of MSTAR [12] and OpenSARShip [48] indicate the new loss function has a good adaptability for the imbalanced data set.
The rest of this paper is organized as follows: Section 2 illustrates the key technologies used to build our lightweight CNN, including channel-wise and spatial attention, depthwise separable convolution and its implementation, and WDM loss function. Furthermore, the technical details of network construction and network topology are also given. Section 3 conducts a series of comparative experiments based on two open datasets, i.e., MSTAR [12] and OpenSARShip [48]. The performance of the proposed network is demonstrated, and how to choose the hyper-parameters is discussed. Section 4 summarizes our work and puts forward the future research.

2. Lightweight CNN Based on Visual Attention Mechanism

2.1. Channel-Wise and Spatial Attention

Convolution layers are the basic structure for CNNs. It learns filters that capturing local spatial features along all input channels, and generates feature maps of jointly encoding space and channel information. Squeeze and excitation (SE) block in [11] can be considered as a kind of channel-wise attention mechanism. It squeezes features along the spatial domain and reweights features along the channels. The structure of SE block is shown in the upper part of Figure 1. In SAR image target recognition, regions of interest are generally concentrated in a small area. Meanwhile, spatial information usually contains important features for accurate recognition, so it should also be used rationally. Inspired by SE block, we carry out similar operations on spatial, and introduce channel attention and spatial attention mechanisms on two parallel branches. Finally, we add the results from the two channels as the output. We call the above operation as channel-wise and spatial attention (CSA) mechanism, and the convolution unit is named CSA block, the structure of it is shown in Figure 1.
Suppose that the feature maps entering into CSA block is M H × W × C , where H , W and C are the spatial height, width and channel depth respectively. In channel attention by-pass, M is represented as M = [ m 1 , m 2 , , m C ] , m i H × W represents the feature maps on each channel. Spatial squeeze is performed by global average pooling (GAP), a statistic Z C is generated by shrinking M through spatial dimensions H × W , where the c -th element of Z is calculated by:
Z c = F s q ( m c ) = 1 W × H i = 1 W j = 1 H m c ( i , j )
After that, channel excitation is completed through a gating mechanism with sigmoid activation, vector Z is transformed to:
S = F e x ( z , W ) = σ ( g ( z , W ) ) = σ ( W 2 δ ( W 1 z ) )
In Equation (2), δ refers to the ReLU [49] function and σ represent sigmoid function, W 1 C r × C and W 2 C × C r . The utilization of two fully-connected (FC) layers aims at limiting model complexity and aiding generalization, it is composed of a dimensionality reduction layer with parameters W 1 with reduction ratio r (we set it to be 8, and the parameter choice is discussed in Section 3.5), a ReLU function, and then a dimensionality-increasing layer with parameters W 2 . The final output of the block is obtained by rescaling the transformation output M with the activations:
x ^ c = F s c a l e ( m c , S c ) = S c m c
The output of channel attention by-pass is X ^ c a = [ x ^ 1 , x ^ 2 , , x ^ c ] H × W × C ( x ^ c H × W ) , which represents the fusion features between channels.
In spatial attention by-pass, the input feature map is represented as M = [ m 1 , 1 , m 1 , 2 , , m i , j , , m H , W ] , m i , j 1 × 1 × C with i { 1 , 2 , , H } and j { 1 , 2 , , W } represents the spatial features that contain all the channel information. Channel squeeze is performed by a 1 × 1 convolution kernel K 1 × 1 × C × 1 , generating a projection tensor U H × W , i.e., U = K * M . Each U i , j of U represents the linearly combination for all C channels in a spatial location ( i , j ) . Similar to channel attention by-pass, we use the sigmoid function as nonlinear activation to complete spatial excitation. The output of spatial attention by-pass can be illustrated as:
X ^ s a = [ x ^ 1 , 1 , , x ^ i , j , , x ^ H , W ]
where, X ^ s a H × W × C and x ^ i , j = σ ( U i , j ) m i , j .
Finally, we add the results of two by-passes (channel attention by-pass and spatial attention by-pass) to get the output of CSA block, i.e., X ^ c s a = X ^ c a + X ^ s a . For the input feature map M , CSA block carries the future recalibrated through the channel and spatial, and it can enhance the expression ability of networks.

2.2. Depthwise Separable Convolution

In standard convolution, the channel of every kernel is the same as that of the current feature map C i n , and every channel is convoluted at the same time. The distribution of convolution kernel in standard convolution is shown in Figure 2a. Kernel size is N c o n v × N c o n v , and the number is C c o n v .
Depthwise separable convolution [33] uses DW convolution and 1 × 1 PW convolution to decompose convolution in channel level. DW refers to a convolution kernel that no longer carry out convolutions in all channels of the input image, but one input channel, i.e., one convolution kernel corresponds to one channel. After that, the PW convolution aggregates the multichannel output of the DW convolution layer to get the weight of the global response. The distribution of convolution kernel in depthwise separable convolution is shown in Figure 2b.
Through Figure 2a,b, we can make a brief analysis of the computation consumption of two convolution methods. The size of the input image is N i n × N i n , with C i n channels, the size of C c o n v kernels is N c o n v × N c o n v × C i n . In order to unify the output and input feature map in size, we assume the stride of convolution is 1, so the size of output features is C c o n v × N i n × N i n . Ignoring the addition of features aggregation, the calculation amount required is N i n × N i n × N c o n v × N c o n v × C i n × C c o n v , the first two items are the size of the input image, and the other four are the space dimensions of the convolution kernel. When deep separable convolution is used, the calculation consumption of DW convolution is N c o n v × N c o n v × C i n × N i n × N i n and the calculation consumption of PW convolution is 1 × 1 × C c o n v × C i n × N i n × N i n . So we can get the ratio of calculation consumption of two convolutions is as follows:
N i n × N i n × N c o n v × N c o n v × C i n × C c o n v N c o n v × N c o n v × C i n × N i n × N i n + C c o n v × C i n × N i n × N i n = 1 C conv + 1 N c o n v 2
It can be seen from the above formula that the calculation consumption of deep separable convolution can be effectively reduced compared with the standard convolution, and the ratio of calculation consumption is only related to the number and size of the convolution kernel.

2.3. Weighted Distance Measure Loss Function

Imbalanced data have a great influence on the classification results, mainly because majority class data have more influence on classifiers than minority classes, so the classification boundaries are biased toward the majority classes.
The common loss function in the field of machine learning, such as 0–1 loss function, log loss function and cross entropy loss function, have the same misclassification cost for all samples, and fail to be used directly in the problem of imbalance data classification. Therefore, new loss functions need to be designed for imbalanced data. On the other hand, the classification problem is a core problem in the research of pattern recognition, and a basic criterion in pattern recognition is to keep the inter class distance as large as possible and the intra class distance as small as possible.
Through the above analysis, we can conclude that the loss function used for imbalanced data classification in CNN should meet the following requirements:
  • It should strengthen the influence of minority samples on training process, and avoid the submergence of minority samples by majority samples.
  • It should be well compatible with the CNN training process and can be calculated in batches.
  • It should enhance the inter class distance and reduce the intra class distance.
Contrastive loss [50] is used to solve the face recognition problem with long tailed distribution (which mean the number of categories is very large and not known during training, and the number of training samples for a single category is very small, and it can be regarded as a form of data imbalance.) data. This method requires a pair of samples as input, learning a similarity measure based on the input data, and then using the similarity measure to determine whether the two samples belong to one class and achieve the recognition results. The core idea of contrastive loss is put a small distance between similar samples, and large distance for dissimilar samples [51]. In generally, the purpose of the SAR image classification is not to judge whether the two slices belong to one class, but to identify what category the image belongs to. So contrastive loss function cannot be used directly. Even so, the thought of the contrastive loss function is of great reference. We combine the idea of contrastive loss and cost sensitive learning to design a weighted distance measure (WDM) loss function used for the problem of imbalanced data classification in CNN. The target of WDM loss function lies in two aspects, the first one is maximize the inter class distance and minimize the intra class distance, and the second one is make the samples of minority classes obtain a large compensation weight.
The WDM loss function can be expressed as the following form.
L = α L 1 + β L 2
In Equation (6), L 1 represents intra class loss and L 2 represents inter class loss, α and β are loss weights of intra class and inter class respectively. α is set to 10 5 and β is set to 10 4 .
We use w indicates the compensation weight, which is used to control the wrong cost of different classes. Supposing that there are N samples in M class totally, and the number of each class are arranged from large to small as N 1 , N 2 , , N m ( m = 1 , 2 , , M ). Then, compensation weight w can be expressed as w = [ w 1 , w 2 , , w m ] = [ N m , N m 1 , , N 1 ] / N , which ensuring the minority classes can obtain a large compensation weight. L 1 can be further expressed as:
L 1 = i I L 1 i w i = i I k w i j = 1 k 1 D j
I represents the total classes of samples in a training batch, and j = 1 k 1 D j is defined as intra class distance measure. D j is the j -th longest Euclidean distance in one class. Suppose x 1 and x 2 are the two samples with the farthest distance in this class, x 3 and x 4 are the two samples with second-farthest distance, then there is D 1 = x 1 x 2 2 2 , D 2 = x 3 x 4 2 2 . k is a hyper-parameter ( k is not a sensitive parameter, it can be set to 1 or 2, experience shows k = 2 is a better choice.), showing the punishment strength of loss function to the intra class distance. The greater value of k means the greater the intensity of the punishment. Through Equation (7), we can see that the essence of intra class loss is the harmonic mean of the first k maximum distance measure.
L 2 is expressed as:
L 2 = max ( m D c , 0 ) = max ( m x A x B 2 2 , 0 )
In Equation (8), supposing that the inter class distance between the class A and B is the shortest. D c is defined as inter class distance measure, representing the shortest inter class. x A and x B denote the arithmetic mean of samples in class A and B after the last layer of CNN, which represents the center of the class characteristics. m is the threshold of loss function to punish the inter class distance. The smaller inter class distance will cause greater loss. We set m to 2 × 10 4 and the results sensitive to it is discussed in Section 3.5.
In general, in the WDM loss function, we introduce the intra class distance measure j = 1 k 1 D j and the inter class distance measure D c to punish the problem that the intra class distance is too large and the inter class distance is too small.
It should be explained that the contrastive loss function is based on a pair of samples, the optimization process is also aimed at a pair of samples and is a local optimization. The WDM loss function is based on a training batch, and the optimization process is also a global optimization for all kinds of samples.

2.4. Network Construction

2.4.1. The Implementation of Depthwise Separable Convolution and CSA Block

When we build the network, we learn from the basic structure of ResNet [9]. When ResNet works, the core unit of it, i.e., the residual block first uses 1 × 1 convolution to compress the dimension of the input feature maps. Therefore, the subsequent 3 × 3 convolution will be completed on a lower data dimension. Finally, the data dimension will be restored by 1 × 1 convolution.
The structure of the residual block is shown in Figure 3a. In the whole process, data is compressed firstly and then expanded, so this structure is also called the bottleneck block. The data processing process in the bottleneck structure is shown in Table 1, t represents expansion factor and generally takes 0.25 in residual structure.
As introduced in Section 2.2, DW convolution uses a convolution kernel with one channel (as shown in Figure 2b), feature extraction capability has decreased compared with standard convolution. If depthwise separable convolution is directly used to replace the 3 × 3 standard convolution in the bottleneck structure, DW convolution will face the data of compressed dimension, which is more unfavorable for DW convolution to extract features. Therefore, refer to literature [52], we first enhance the dimension of data by a PW unit before using DW, that is, set expansion factor t to an integer bigger than 1 (we take t = 6, and the choice of it is discussed in Section 3.5) to make DW convolution reach a higher dimension of data. After that, a PW convolution is used to compress the data dimension. This structure is called inverted residual block, as shown in Figure 3b. In addition, related studies [52] also show that using non-linear layers in bottlenecks indeed hurts the performance by several percent, so in the inverted residual block, we remove the ReLU layer after the last 1 × 1 convolution to better retain the features.
Finally, the CSA block mentioned in Section 2.1 is added to the inverted residual structure to complete the fusion of the channel and the spatial features. The structure of the inverted residual block with channel-wise and spatial attention (IR-CSA) is shown in Figure 3c. We use IR-CSA structure as the basic convolution block to form the main structure of the CNN we propose. It is similar to ResNet [9] and many other networks, the main structure of the network is constructed by continuously stacking the basic convolution units.

2.4.2. Network Topology

The main steps used in designing our network are summarized as below:
  • We use depthwise separable convolution instead of the 3 × 3 standard convolution in network to reduce the computational cost, and use the inverted residual block to improve the feature extraction ability of depthwise separable convolution.
  • The CSA block mentioned in Section 2.1 is introduced into the inverted residual structure to improve feature learning and fusion capabilities.
  • WDM loss function is applied to reduce the impact of imbalance data.
  • For the SAR image slice with input size 128 × 128, the larger size of convolution kernels are adopted to cope with the possible noise. We design the convolution kernel size in the first convolution layer to be 7 × 7, the performance of convolution kernels of different sizes under noise interference will be illustrated in Section 3.3.
The structure of lightweight network presented in this paper and ResNet50 [9] are shown in Table 2. Our network contains 12 IR-CSA blocks, and each IR-CSA block has 4 convolution layers and one CSA block. Similar to ResNet50, our network is also a 50-layer deep network, but its computing consumption is obviously less than it.
Only the main structure of the network is given in the Table 2. Other operations, such as batch normalization (BN), ReLU, etc. are not embodied in the table. The reduction of the size of the feature maps is achieved by setting the convolution step of 2.

3. Experiments and Discussion

3.1. Experimental Data Sets

3.1.1. MSTAR

One of the datasets used in our work is part of MSTAR program [10], which is jointly sponsored by the U.S. Defense Advanced Research Projects Agency (DARPA) and Air Force Research Laboratory (AFRL). Hundreds of thousands of SAR images were collected containing ground targets, including different target types, aspect angles, depression angles, serial number, and articulation. SAR images in the dataset are gathered by the X-band SAR sensors in spotlight mode [25], with the resolution of 0.3 m × 0.3 m and 0 ~ 360 ° azimuth coverage. Due to the lack of data, our dataset contains tanks: T62, T72; armored vehicles: BRDM2, BTR60; rocket launcher: 2S1; air defense unit: ZSU234; military trucks: ZIL131; bulldozer: D7; false target: SLICY nine types of targets, as shown in Figure 4. The lack of BMP2 and BTR70 also belong to the armored vehicles (the same as BRDM2 and BTR60), so influence on the cause can be ignored. Referring to the experiments in literature [22], 2770 images under 17 ° pitch angle were taken as training samples, and 2387 images were taken as testing samples under 15 ° pitch angle.
Table 3 gives a list of training and testing data for 9 types of targets. From the table, we can see that the number of samples is relatively balanced, without significant difference.

3.1.2. QpenSARShip

The OpenSARShip [48] is a new dataset built by Key Laboratory of Intelligent Sensing and Recognition, Shanghai Jiao Tong University, China. It contains more than ten thousands ship chips covering 17 AIS types from 41 Sentinel-1 SAR images with C-band [48]. These 41 Sentinel-1 SAR images are collected from five typical scenes because of their intense marine traffic: Shanghai Port (China), Shenzhen Port (China), Tianjin Port (China), Yokohama Port (Japan), and Singapore Port (Singapore). OpenSARShip provides two available products of the interferometric wide swath mode (IW): the single look complex (SLC) with 2.7 m × 22 m to 3.5 m × 22 resolution, and ground range detected (GRD) with 20 m × 20 m resolution [48].
We classify the data set according to different polarizations and imaging modes. The distribution of the samples under GRD and SLC mode are shown in Figure 5. Each mode contains the same number of VH and VV polarization images, e.g., in the 4738 cargo slices of the GRD mode, there are 2369 images of VH and VV polarization, respectively. The data set includes cargo, tankers, tugs and other eleven types of ships.
It can be seen from Figure 5 that the class imbalance is quite serious in this data set. The cargo class accounts for more than 60% of the total in both modes. This imbalance may have a great impact on the recognition results. We divide the data into a training set and testing set in the proportion of 7:3, and eliminate the minority samples that are not enough to build the training and testing set. The data we used in experiments is shown in Table 4.
In [48], the authors completed a series of SAR image recognition experiments under VH and VV polarization, but the mode of the SAR images (GRD or SLC) was not clarified. In order to study the effects of different polarizations and modes on the recognition results we conduct a series of prior classification experiments under our network with different polarizations and imaging modes (i.e., GRD mode with VH polarization, GRD mode with VV polarization and SLC mode with VH polarization) SAR images.
Compared with the overall accuracy, we can clearly understand the recognition result of each class through the confusion matrix, and avoid the influence of the high recognition accuracy of majority classes on the overall recognition accuracy. Therefore, we use the confusion matrix as the evaluation index in this place and the subsequent experiments in Section 3.4. The results of prior classification experiments are shown in Table 5, Table 6 and Table 7.
From Table 5, Table 6 and Table 7, we can see the total recognition accuracy in three groups are all about 78%, and the P of the majority class (cargo) is significantly higher than that of the minority classes. Experimental result indicates that polarizations and imaging modes have no significant effect on the recognition results, but data imbalance has an obvious influence on it, so we only utilize SAR images under GRD mode with VH polarization for the subsequent experiments in Section 3.4.

3.2. Experimental Environment and Configuration

Most CNN models (including our network) require input images of the same size. Meanwhile, the size of SAR chips in the OpenSARShip dataset is mainly concentrated in 100 × 100 to 150 × 150. Refer to the universal practice described in [20,22], we resize the SAR images to 128 × 128 by CenterCrop function in torchvision transforms toolkit of Pytorch (one of the most popular deep learning frameworks). If the image size is smaller than 128 × 128, it was cropped, if otherwise, it was expanded. The exceeded parts are filled with pixel dots with a gray-value of 0. The targets of OpenSARShip datasets are in the center of the images, so we do not change the distribution of the targets in the images by the center crop or expansion.
Xavier [53] is a widely used initialization method in CNNs. Its basic design principle is to make the information flow better in the network. The variance of activation value and gradient of each layer should be kept as constant as possible. Refer to many state of the art CNNs (e.g., ResNet [9], DenseNet [10], MobileNets [35] etc.), we also adopt Xavier as the initialization method. We train the network by using mini-batch SGD, with an initial learning rate of 0.01 and a reducing factor of 0.1 after 30 epoches. The momentum parameter is set to be 0.9 and the weight decay parameter 0.0001. The number of iterations in training is 50, and the batch size is set to 30.
Experiments are carried out in the 64-bit Ubuntu 14.04 system. The software is mainly based on deep learning architecture of Pytorch and python development environment Pycharm. The hardware is based on an Intel (R) Core (TM) i7-6770K @ 4.00GHz CPU and two NVIDIA GTX1080 GPUs, with CUDA8.0 accelerating calculation.

3.3. Classification Experiment on MSTAR

In order to test the performance of our network in SAR image recognition, we conducted a classification experiment based on the MSTAR dataset, and selected four CNN models with good performance in SARA-ATR or CV field, namely Network-1, proposed by Wilmanski et al [23]. Network-2, A-ConvNets proposed by Chen et al. in literature [22], ResNet18 [9], and SE-ResNet50 [11].
Figure 6 shows the training accuracy and loss curves of 5 models. It can be seen that due to the small amount of data in the MSTAR data set, the 5 CNN models can basically converge after 10 epoches, and all the networks can finally get close to 100% recognition accuracy on the training set. Our lightweight CNN and SE-ResNet50 have similar performance on the training set, and both of them converge faster than other models.
We define the accuracy of recognition P as the ratio of the number of samples correctly recognized to the total number of samples in the testing set, and use P as an indicator to evaluate the classification results. Table 8 shows the recognition accuracy on testing set, model size and total iteration times (total time spent on 50 training epoches) of five CNNs.
Our network obtained the highest recognition accuracy of 99.54%, compared with Network-1, Network-2 and ResNet50. The recognition precision of our lightweight network and SE-ResNet50 is higher, which justify the fact that the introduction of visual attention mechanism can significantly enhance the ability of feature learning of CNN models. While achieving a slightly higher recognition accuracy than SE-ResNet50, our lightweight network has an obvious advantage in terms of iteration time and model size. The model size is about 1/5 of SE-ResNet50 and the iteration time is about 1/4 of it. According to the information in Table 8, our lightweight network has achieved better results in recognition accuracy and recognition efficiency.
Table 9 shows the confusion matrix for the classification results of our lightweight network. As can be seen from the confusion matrix, each class has obtained an ideal accuracy, with a minimum recognition accuracy of 98.9% (BRDM-2) and maximum of 100% (T72, ZSU-131).
An important characteristic of SAR images is often accompanied by the effects of noise. In order to test the anti-noise ability of our network, referring to the experimental methods in literature [22,54], we add different intensities noise obeying gamma distribution [55] in the SAR image by controlling the proportion of noise pixels in the whole image pixels. First, we design a noise function to generate random noise that obeying gamma distribution G a ( α , β ) , where α = 1 , β   =   0.1 . Then, randomly select a certain proportion of pixels in the test images and replace their values with independent and identically distributed samples generated by noise function. Finally, under different noise intensity, the proposed lightweight network is used to make a contrast experiment by changing the size of the convolution kernel in the first convolution layer. Examples of images with different intensities of noise are shown in Figure 7, and the experimental results are shown in Figure 8 and Table 10.
It can be seen from Figure 8 and Table 10 that with the increase of the noise intensity, the recognition accuracy of the 5 × 5 convolution kernel and the 7 × 7 convolution kernel decreases obviously. The recognition accuracy of 7 × 7 convolution kernel at any noise intensity is higher than that of 5 × 5 convolution kernel, which shows that the use of 7 × 7 convolution kernel has a better adaptability to noise. It is worth noting that when the noise intensity increases from 10% to 15%, the reduction of the recognition accuracy of the 5 × 5 convolution kernel is obviously greater than the 7 × 7 convolution kernel. So we can infer that the feature extraction ability of the small convolution kernel will be greatly affected under the high intensity noise condition. Based on the above experimental results, we choose the 7 × 7 convolution kernel when designing the first convolution layer of the network.

3.4. Classification Experiment on OpenSARShip

There is a serious data imbalance problem in the OpenSARShip data set. The study in [47] shows that random over-sampling and under-sampling are two good methods solving data imbalance problem in CNNs. In order to compare the processing capabilities for imbalanced data of random over-sampling, under-sampling and the WDM loss function mentioned in Section 2.3, we design five groups of ablation experiments based on the proposed lightweight network. The experimental conditions are shown in Table 11.
The over-sampling in Table 11 refers to random copying of minority classes, and eventually the number of samples in minority classes is the same as that of the majority classes. In the GRD mode, we randomly copy minority samples and finally make the number of training samples for each class to be 1600. Under-sampling randomly removes samples of majority classes to balance the number of samples in minority classes. However, because the number of samples of minority classes in OpenSARShip is too small (as shown in Table 4, there are only 45 training samples and 18 test samples in the tug class under GRD mode), the exclusive utilization of under-sampling will cause the number of samples too small to constitute effective training and testing set. Therefore, we take a compromise in group 3, under-sampling is used of the majority classes while over-sampling is used of the minority classes, and finally the number of samples in every class reached 500. The first three groups adopt the cross entropy loss function, the difference is whether the data is preprocessed. Group 4 adopts the WDM loss function we proposed, and the data is not preprocessed. Group 5 can be seen as a combination of group 3 and group 4, the WDM loss function is used on the basis of data processing.
The results of prior classification experiments in Section 3.1.2 show that the recognition accuracy is not sensitive to different polarizations and imaging modes SAR image. So only the experimental results in the GRD mode with VH polarization are given here, as shown in Table 12, Table 13, Table 14, Table 15 and Table 16. The classification results of the five groups are summarized in Figure 9.
It can be seen from Table 12 and Figure 9 that group 1 obtains 78% of the overall recognition accuracy as the experimental benchmark, but the recognition rate of the four types of samples show a big difference. The recognition accuracy of majority class cargo reaches 90%, but the recognition accuracy of a minority class tug is only 17%, and the recognition accuracy of minority classes tanker and others is also at a lower level. In the experiment results of group 2 and group 3, the recognition accuracy of the minority classes is improved because of the use of over-sampling or the combination of over-sampling and under-sampling, but the recognition rate of tug is still very low (in the experiment of group 3, through data processing, the recognition accuracy can only be raised from 17% to 22%). The experiment results of group 4 shows that by using the WDM loss function, the recognition accuracy of the minority classes has greatly improved, with the recognition accuracy gap between cargo class and other minority classes obviously narrowed. The overall recognition accuracy reaches 83%, indicating that the WDM loss function effectively improves the adverse effects of imbalanced data on the recognition results. In group 5, we combine the data preprocessing method with the WDM loss function, and the recognition accuracy is slightly higher than that of group 4. It shows that the combination of the data level method and the algorithm level method will be a good way to solve the problem of data imbalance. In general, despite that the total classification accuracy from group 1 to group 5 slightly differs, the recognition accuracy of minority classes has greatly improved. It shows that the changes in the recognition accuracy of the minority classes are difficult to affect the overall recognition accuracy, only using the total recognition accuracy cannot accurately evaluate the recognition results.

3.5. Hyper-Parameters Experiment

In this section, we explain how to select the key hyper-parameters in the network. In Section 2.1, reduction ratio r is a variable parameter in channel attention bypass, and it represents the degree of compression of features on the channel. In order to get a suitable parameter, we use different r values to carry out recognition experiments on MSTAR dataset. The network used is the lightweight CNN proposed in this paper. In each comparison experiment, except for the value of r , the other conditions are the same. The results of the experiment are shown in Table 17.
The comparison in Table 17 reveals that with the increase of r , both accuracy and model size show a nonlinear downward trend. When r is increased from 8 to 16 and 32, the magnitude of the decrease in accuracy is significantly increased, so r is not as big as possible, the larger r can effectively compress the model size, but the over compression may also lead to the loss of information, and the decrease of the recognition accuracy. We found that when r = 8, a good tradeoff between accuracy and complexity is achieved, so we use this value for all experiments.
In Section 2.3, m is a hyper-parameter of the inter class loss L2, which is a limitation of distance between classes. In order to research its influence of the recognition accuracy, we conduct a series of comparative experiments under the same experimental environment of group 4 in Section 3.4. Experimental result is shown in Table 18.
We can conclude that the recognition accuracy is insensitive to m, when m is set to 2 × 10 4 , recognition accuracy of each class and entirety is more ideal. m represents a limitation of interclass distance and the bigger m brings the greater penalty. So we could also find that when m > 2 × 10 4 , the recognition accuracy is better than m < 2 × 10 4 .
In Section 2.4.1, we use the expansion factor t to control the data dimension, t is a coefficient. When t is less than 1, the data dimension is compressed, conversely, the data dimension is expanded. In residual structure, t is generally less than 1, while in the inverted residual structure, t is an integer greater than 1.
We set different t values for comparison experiments on the MSTAR dataset. The network used is the lightweight CNN proposed in this paper. In each comparison experiment, except for the value of t, the other conditions are the same. The results of the experiment are shown in the Table 19.
From Table 19, we can see that with the increase of t, the accuracy and the model size are increasing. When t is set to be 2 and 4, although it has a smaller model size, the accuracy rate is lower. When t is assigned to be 10, compared to 6, the accuracy rate is only a little higher, but model size has increased a lot, so we think 6 is the optimal value.

4. Conclusions and Future Work

This paper first designed a lightweight CNN based on visual attention and depthwise separable convolution for SAR image target classification. Then a new WDM loss function is proposed to solve the problem of data imbalance in data sets. Finally, a series of recognition experiments based on two open datasets of MSTAR and OpenSARShip are implemented. The experiment results on MSTAR show that compared with CNN model without visual attention mechanism (e.g., ResNet [9], Network in literature [23] and A-ConvNet [22]), our network achieves higher recognition accuracy, which indicate that the introduction of visual attention mechanism enhances the representation ability of CNN. Meanwhile, the model size and iteration time of our network is greatly reduced by the utilization of depthwise separable convolution. The ablation experiments on the OpenSARShip dataset compare the ability of several methods to handle imbalanced data. Experimental results indicate that the combination of resampling method and the WDM loss function can better weaken the impact of data imbalance on the recognition results. Nevertheless, there are still some limitations and shortcomings in our work.
  • The method we adopted in the paper belongs to supervised learning in machine learning field. The deep network needs a large number of data to train the parameters adequately, which restricts its application to a certain extent.
  • Our network needs the same size images as input, if the size of the input images is quite different, the recognition result will be affected. This problem can be solved by introducing space pyramid pooling (SPP) [56], which will be our future research direction.
  • The experimental results in Section 3.4 show that our network is somewhat sensitive to noise, and there is still much room for improvement in this aspect.
Last but not the least, weak supervised or unsupervised machine learning algorithm is an important development direction in the field of artificial intelligence. This kind of algorithm reduces the dependence on the training data to a certain extent, and makes the recognition process more intelligent. It is a worthwhile direction to introduce this algorithm into the field of SAR image and we think it will effectively enhance the intelligence and generalization ability of the recognition algorithm.

Author Contributions

J.S. and J.L. conceived and designed the CNN model; S.P. constructed the outline for the manuscript; J.S., J.L. and S.P. performed the experiments and analyzed the data; C.Q. contributed materials and computing resources; J.S. wrote the paper.

Funding

This research was funded in part by the Key Program of National Natural Science Foundation of China under Grant No. 61531020, and in part by the General Program of National Natural Science Foundation of China under Grant No. 61571454.

Acknowledgments

The authors would like to thank the anonymous reviewers for the constructive suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, S.; Wang, H. SAR target recognition based on deep learning. In Proceedings of the International Conference on Data Science and Advanced Analytics, Shanghai, China, 30 October–1 November 2014; IEEE: New York, NY, USA, 2015; pp. 541–547. [Google Scholar]
  2. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  3. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  4. Ding, J.; Chen, B.; Liu, H.; Huang, M. Convolutional Neural Network with Data Augmentation for SAR Target Recognition. IEEE Geosci. Remote Sens. Lett. 2016, 13, 364–368. [Google Scholar] [CrossRef]
  5. El-Darymli, K.; Gill, E.W.; Mcguire, P.; Power, D.; Moloney, C. Automatic Target Recognition in Synthetic Aperture Radar Imagery: A State-of-the-Art Review. IEEE Access 2017, 4, 6014–6058. [Google Scholar] [CrossRef]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012; Curran Associates Inc.: New York, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  7. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
  8. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; IEEE: Los Alamitos, CA, USA, 2015; pp. 1–9. [Google Scholar] [Green Version]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: Los Alamitos, CA, USA, 2016; pp. 770–778. [Google Scholar] [Green Version]
  10. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. arXiv, 2016; arXiv:1608.06993. [Google Scholar]
  11. Hu, J.; Shen, L.; Sun, G. Squeeze and Excitation Networks. arXiv, 2017; arXiv:1709.01507. [Google Scholar]
  12. Wissinger, J.; Ristroph, R.; Diemunsch, J.R.; Severson, W.E.; Fruedenthal, E. MSTAR’s extensible search engine and model-based inferencing toolkit. Int. Soc. Opt. Photonics 1999, 3721, 554–570. [Google Scholar]
  13. Wang, H.; Chen, S.; Xu, F.; Jin, Y.Q. Application of deep-learning algorithms to MSTAR data. In Proceedings of the Geoscience and Remote Sensing Symposium, Milan, Italy, 26–31 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 3743–3745. [Google Scholar]
  14. Profeta, A.; Rodriguez, A.; Clouse, H.S. Convolutional neural networks for synthetic aperture radar classification. Proc. SPIE 2016, 9843, 98430M. [Google Scholar] [CrossRef]
  15. Zhao, Q.; Principe, J.C. Support vector machines for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 643–654. [Google Scholar] [CrossRef]
  16. O’Sullivan, J.A.; DeVore, M.D.; Kedia, V.; Miller, M.I. SAR ATR performance using a conditionally Gaussian model. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 91–108. [Google Scholar] [CrossRef]
  17. Sun, Y.J.; Liu, Z.P.; Todorovic, S.; Li, J. Adaptive boosting for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2007, 43, 112–125. [Google Scholar] [CrossRef]
  18. Huang, P.; Qiu, W. A robust decision fusion strategy for SAR target recognition. Remote Sens. Lett. 2018, 9, 507–514. [Google Scholar] [CrossRef]
  19. Huang, Z.; Pan, Z.; Lei, B. Transfer Learning with Deep Convolutional Neural Network for SAR Target Classification with Limited Labeled Data. Remote Sens. 2017, 9, 907. [Google Scholar] [CrossRef]
  20. Bentes, C.; Velotto, D.; Tings, B. Ship Classification in TerraSAR-X Images with Convolutional Neural Networks. IEEE J. Ocean. Eng. 2017, 99, 1–9. [Google Scholar] [CrossRef]
  21. Bentes, C.; Velotto, D.; Lehner, S. Target classification in oceanographic SAR images with deep neural networks: Architecture and initial results. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 1–4. [Google Scholar]
  22. Chen, S.; Wang, H.; Xu, F.; Jin, Y.-Q. Target classification using the deep convolutional networks for SAR images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4806–4817. [Google Scholar] [CrossRef]
  23. Wilmanski, M.; Kreucher, C.; Lauer, J. Modern approaches indeep learning for SAR ATR. Proc. SPIE 2016, 9843, 98430N. [Google Scholar] [CrossRef]
  24. Ball, J.E.; Anderson, D.T.; Chan, C.S. Special Section Guest Editorial: Feature and Deep Learning in Remote Sensing Applications. J. Appl. Remote Sens. 2018, 11, 1–20. [Google Scholar] [CrossRef]
  25. Karine, A.; Toumi, A.; Khenchaf, A.; Hassouni, M.E. Radar Target Recognition Using Salient Keypoint Descriptors and Multitask Sparse Representation. Remote Sens. 2018, 10, 843. [Google Scholar] [CrossRef]
  26. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. arXiv, 2014; arXiv:1409.0473. [Google Scholar]
  27. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; Bengio, Y. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 2048–2057. [Google Scholar]
  28. Yao, L.; Torabi, A.; Cho, K.; Ballas, N.; Pal, C.; Larochelle, H.; Courville, A. Describing videos by exploiting temporal structure. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  29. Mnih, V.; Heess, N.; Graves, A. Recurrent models of visual attention. arXiv, 2014; arXiv:1406.6247. [Google Scholar]
  30. Shen, Z.; Shi, H.; Feris, R.; Cao, L.; Yan, S.; Liu, D.; Wang, X.; Xue, X.; Huang, T.S. Learning Object Detectors from Scratch with Gated Recurrent Feature Pyramids. arXiv, 2017; arXiv:1712.00886. [Google Scholar]
  31. Xie, S.; Girshick, R.B.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  32. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inceptionv4, inception-resnet and the impact of residual connections on learning. arXiv, 2016; arXiv:1602.07261. [Google Scholar]
  33. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 1800–1807. [Google Scholar] [Green Version]
  34. Kaiser, L.; Gomez, A.N.; Chollet, F. Depthwise Separable Convolutions for Neural Machine Translation. arXiv, 2017; arXiv:1706.03059. [Google Scholar]
  35. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv, 2017; arXiv:1704.04861. [Google Scholar]
  36. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. arXiv, 2017; arXiv:1707.01083. [Google Scholar]
  37. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv, 2016; arXiv:1602.07360. [Google Scholar]
  38. Kubat, M.; Holte, R.C.; Matwin, S. Machine Learning for the Detection of Oil Spills in Satellite Radar Images. Mach. Learn. 1998, 30, 195–215. [Google Scholar] [CrossRef] [Green Version]
  39. He, H.; Garcia, E.A. Learning from Imbalanced Data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [Green Version]
  40. Chang, R.F.; Wu, W.J.; Moon, W.K.; Chou, Y.H.; Chen, D.R. Support vector machines for diagnosis of breast tumors on US images. Acad. Radiol. 2003, 10, 189–197. [Google Scholar] [CrossRef]
  41. Guo, H.; Li, Y.; Shang, J.; Mingyun, G.; Yuanyue, H.; Bing, G. Learning from class-imbalanced data: Review of methods and applications. Expert Syst. Appl. 2016, 73, 220–239. [Google Scholar]
  42. Levi, G.; Hassner, T. Age and gender classification using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 34–42. [Google Scholar]
  43. Janowczyk, A.; Madabhushi, A. Deep learning for digital pathology image analysis: A comprehensive tutorial with selected use cases. J. Pathol. Inform. 2016, 7, 29–41. [Google Scholar] [CrossRef] [PubMed]
  44. Khan, S.H.; Bennamoun, M.; Sohel, F.; Togneri, R. Cost sensitive learning of deep feature representations from imbalanced data. arXiv, 2015; arXiv:1508.03422. [Google Scholar]
  45. Min, F.; Zhu, W. A competition strategy to cost-sensitive decision trees. In Proceedings of the International Conference on Rough Sets and Knowledge Technology, Chengdu, China, 17–20 August 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 359–368. [Google Scholar]
  46. Weiss, G.M.; Mccarthy, K.; Zabar, B. Cost-Sensitive Learning vs. Sampling: Which is Best for Handling Unbalanced Classes with Unequal Error Costs? In Proceedings of the International Conference on Data Mining, Las Vegas, NV, USA, 25–28 June 2007; pp. 35–41. [Google Scholar]
  47. Buda, M.; Maki, A.; Mazurowski, M.A. A systematic study of the class imbalance problem in convolutional neural networks. arXiv, 2017; arXiv:1710.05381. [Google Scholar]
  48. Huang, L.; Liu, B.; Li, B.; Guo, W.; Yu, W.; Zhang, Z.; Yu, W. OpenSARShip: A Dataset Dedicated to Sentinel-1 Ship Interpretation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 99, 1–14. [Google Scholar] [CrossRef]
  49. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  50. Hadsell, R.; Chopra, S.; Lecun, Y. Dimensionality Reduction by Learning an Invariant Mapping. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; IEEE Computer Society: Piscataway, NJ, USA, 2006; pp. 1735–1742. [Google Scholar]
  51. Hoffer, E.; Ailon, N. Deep Metric Learning Using Triplet Network. arXiv, 2014; arXiv:1412.6622v3. [Google Scholar]
  52. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. arXiv, 2018; arXiv:1801.04381. [Google Scholar]
  53. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  54. Dong, G.; Wang, N.; Kuang, G. Sparse Representation of Monogenic Signal: With Application to Target Recognition in SAR Images. IEEE Signal Process. Lett. 2014, 21, 952–956. [Google Scholar]
  55. Karine, A.; Toumi, A.; Khenchaf, A.; El Hassouni, M. Target Recognition in Radar Images Using Weighted Statistical Dictionary-Based Sparse Representation. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2403–2407. [Google Scholar] [CrossRef]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The structure of CSA block.
Figure 1. The structure of CSA block.
Sensors 18 03039 g001
Figure 2. The distribution of convolution kernel (a) in standard convolution (b) in depthwise separable convolution.
Figure 2. The distribution of convolution kernel (a) in standard convolution (b) in depthwise separable convolution.
Sensors 18 03039 g002
Figure 3. The structure of different basic blocks in CNNs (a) residual block (b) inverted residual block (c) inverted residual block with channel-wise and spatial attention (IR-CSA).
Figure 3. The structure of different basic blocks in CNNs (a) residual block (b) inverted residual block (c) inverted residual block with channel-wise and spatial attention (IR-CSA).
Sensors 18 03039 g003
Figure 4. Examples of MSTAR dataset.
Figure 4. Examples of MSTAR dataset.
Sensors 18 03039 g004
Figure 5. Statistical results of data set.
Figure 5. Statistical results of data set.
Sensors 18 03039 g005
Figure 6. Training curves of 5 CNN models (a) accuracy curves (b) loss curves.
Figure 6. Training curves of 5 CNN models (a) accuracy curves (b) loss curves.
Sensors 18 03039 g006
Figure 7. SAR images with different intensities of noise.
Figure 7. SAR images with different intensities of noise.
Sensors 18 03039 g007
Figure 8. Recognition accuracy of different noise intensities.
Figure 8. Recognition accuracy of different noise intensities.
Sensors 18 03039 g008
Figure 9. Result statistics of ablation experiments.
Figure 9. Result statistics of ablation experiments.
Sensors 18 03039 g009
Table 1. Data processing in the bottleneck structure.
Table 1. Data processing in the bottleneck structure.
InputOperatorOutput
h × w × k 1 × 1   c o n v h × w × ( t k )
h × w × ( t k ) 3 × 3   c o n v h s × w s × ( t k )
h s × w s × ( t k ) 1 × 1   c o n v h s × w s × k
Table 2. The main structure of ResNet50 and our network.
Table 2. The main structure of ResNet50 and our network.
Output SizeResNet50Our NetworkOutput Size
64 2 × 64 c o n v , 7 × 7 , 64 , s t r i d e 2 c o n v , 7 × 7 , 16 , s t r i d e 2 64 2 × 16
32 2 × 256 max p o o l , 3 × 3 , s t r i d e 2 32 2 × 24
[ c o n v , 1 × 1 × 64 , 64 c o n v , 3 × 3 × 64 , 64 c o n v , 1 × 1 × 64 , 256 ] × 3 [ c o n v , 1 × 1 × 16 , 96 c o n v , 3 × 3 × 1 , 16 c o n v , 1 × 1 × 16 , 96 c o n v , 1 × 1 × 96 , 24 C S A b l o c k ] × 2
16 2 × 512 [ c o n v , 1 × 1 × 256 , 128 c o n v , 3 × 3 × 128 , 128 c o n v , 1 × 1 × 128 , 512 ] × 4 [ c o n v , 1 × 1 × 24 , 144 c o n v , 3 × 3 × 1 , 24 c o n v , 1 × 1 × 24 , 144 c o n v , 1 × 1 × 144 , 32 C S A b l o c k ] × 3 16 2 × 32
8 2 × 1024 [ c o n v , 1 × 1 × 512 , 256 c o n v , 3 × 3 × 256 , 256 c o n v , 1 × 1 × 256 , 1024 ] × 6 [ c o n v , 1 × 1 × 32 , 192 c o n v , 3 × 3 × 1 , 32 c o n v , 1 × 1 × 32 , 192 c o n v , 1 × 1 × 192 , 96 C S A b l o c k ] × 4 8 2 × 96
4 2 × 2048 [ c o n v , 1 × 1 × 1024 , 512 c o n v , 3 × 3 × 512 , 512 c o n v , 1 × 1 × 512 , 2048 ] × 3 [ c o n v , 1 × 1 × 96 , 576 c o n v , 3 × 3 × 1 , 96 c o n v , 1 × 1 × 96 , 576 c o n v , 1 × 1 × 576 , 320 C S A b l o c k ] × 3 4 2 × 320
1 × 1 × N GAP, Fully connection, Cross entropy loss functionGAP, Fully connection, WDM loss function 1 × 1 × N
Table 3. Number of samples in training and testing set.
Table 3. Number of samples in training and testing set.
T62T72BRDM-2BTR-602S1ZSU-234ZIL-131D7SLICYTotal
Training2994232982562992992992992982770
Testing2732752741952742742742742742387
Table 4. List of Experimental Data (VH/VV polarization).
Table 4. List of Experimental Data (VH/VV polarization).
CargoTankerTugDredgingOther
trainingtestingtrainingtestingtrainingtestingtrainingtestingtrainingtesting
GRD1659710242103451813866
SLC122552534514619717618978
Table 5. Experimental results of GRD mode with VH polarization.
Table 5. Experimental results of GRD mode with VH polarization.
CargoTankerTugOtherP
Cargo638313830.90
Tanker41369170.35
Tug120330.17
Other34102200.30
Total 0.78
Table 6. Experimental results of GRD mode with VV polarization.
Table 6. Experimental results of GRD mode with VV polarization.
CargoTankerTugOtherP
Cargo629314550.89
Tanker38399170.38
Tug111330.17
Other33101220.33
Total 0.77
Table 7. Experimental results of SLC mode with VH polarization.
Table 7. Experimental results of SLC mode with VH polarization.
CargoTankerTugDredgingOtherP
Cargo4783386100.91
Tanker328779110.60
Tug212110.29
Dredging211200.33
Other29697270.35
Total 0.78
Table 8. Experimental results of recognition accuracy, model size and iteration time.
Table 8. Experimental results of recognition accuracy, model size and iteration time.
NetworksNetwork-1Network-2ResNet50SE-ResNet 50Our Network
P (%)95.4198.0598.2399.4199.54
Model size (Mb)19.221.698.5112.624.2
Iteration time (s)44255712591576403
Table 9. Confusion matrix for the experimental results of lightweight network.
Table 9. Confusion matrix for the experimental results of lightweight network.
T62T72BRDM-2BTR-602S1ZSU-234ZIL-131D7SLICY P
T622720010000099.64
T7202750000000100
BRDM-20227100001098.90
BTR-600001930110098.97
2S10000273010099.64
ZSU-23400000274000100
ZIL-1310100002730099.64
D70000010273099.64
SLICY0200000027299.27
Total 99.54
Table 10. Recognition accuracy of 5 × 5 and 7 × 7 kernel size.
Table 10. Recognition accuracy of 5 × 5 and 7 × 7 kernel size.
Noise Intensity Kernel Size1%5%10%15%
5 × 50.93330.89210.81330.6928
7 × 70.97140.93260.88350.8019
Table 11. Setting of experimental conditions on OpenSARShip dataset.
Table 11. Setting of experimental conditions on OpenSARShip dataset.
Over-SamplingUnder-SamplingCross Entropy LossWDM Loss
Group 1 (baseline)×××
Group 2××
Group 3×
Group 4×××
Group 5×
Table 12. Experimental results of group 1 (baseline, same as Table 5).
Table 12. Experimental results of group 1 (baseline, same as Table 5).
CargoTankerTugOtherP
Cargo638313830.90
Tanker41369170.35
Tug120330.17
Other34102200.30
Total 0.78
Table 13. Experimental results of group 2.
Table 13. Experimental results of group 2.
CargoTankerTugOtherP
Cargo635314130.89
Tanker38399170.38
Tug120330.17
Other3372240.36
Total 0.78
Table 14. Experimental results of group 3.
Table 14. Experimental results of group 3.
CargoTankerTugOtherP
Cargo632344130.89
Tanker30517150.50
Tug110430.22
Other3062280.42
Total 0.80
Table 15. Experimental results of group 4.
Table 15. Experimental results of group 4.
CargoTankerTugOtherP
Cargo630354140.89
Tanker2269570.67
Tug501030.56
Other1962390.59
Total 0.83
Table 16. Experimental results of group 5.
Table 16. Experimental results of group 5.
CargoTankerTugOtherP
Cargo632363930.89
Tanker2072560.70
Tug501030.56
Other1852410.62
Total 0.84
Table 17. Experimental results of different value of r .
Table 17. Experimental results of different value of r .
Reduction Ratio rAccuracy (%)Model Size (Mb)
499.5831.6
899.5424.2
1699.1621.9
3298.0818.5
Table 18. Recognition accuracy ( P ) of samples under different values of m.
Table 18. Recognition accuracy ( P ) of samples under different values of m.
CargoTankerTugOtherTotal
m = 1.5 × 10 4 0.870.650.500.560.81
m = 1.8 × 10 4 0.880.650.500.580.82
m = 2 × 10 4 0.890.670.560.590.83
m = 2.5 × 10 4 0.890.660.560.590.83
m = 3 × 10 4 0.880.650.500.580.82
Table 19. Experimental results of different value of t.
Table 19. Experimental results of different value of t.
Expansion Factor tAccuracy (%)Model Size (Mb)
295.6217.6
496.3721.5
699.5324.2
1099.5631.1

Share and Cite

MDPI and ACS Style

Shao, J.; Qu, C.; Li, J.; Peng, S. A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification. Sensors 2018, 18, 3039. https://doi.org/10.3390/s18093039

AMA Style

Shao J, Qu C, Li J, Peng S. A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification. Sensors. 2018; 18(9):3039. https://doi.org/10.3390/s18093039

Chicago/Turabian Style

Shao, Jiaqi, Changwen Qu, Jianwei Li, and Shujuan Peng. 2018. "A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification" Sensors 18, no. 9: 3039. https://doi.org/10.3390/s18093039

APA Style

Shao, J., Qu, C., Li, J., & Peng, S. (2018). A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification. Sensors, 18(9), 3039. https://doi.org/10.3390/s18093039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop