Next Article in Journal
Hyperspectral Image Classification Based on Spectral Multiscale Convolutional Neural Network
Previous Article in Journal
Global Mapping of Soil Water Characteristics Parameters— Fusing Curated Data with Machine Learning and Environmental Covariates
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Focal Loss Function Based on Transfer Learning for Few-Shot Radar Signal Intra-Pulse Modulation Classification

School of Electronic Engineering, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(8), 1950; https://doi.org/10.3390/rs14081950
Submission received: 8 March 2022 / Revised: 12 April 2022 / Accepted: 15 April 2022 / Published: 18 April 2022

Abstract

:
To solve the difficulty associated with radar signal classification in the case of few-shot signals, we propose an adaptive focus loss algorithm based on transfer learning. Firstly, we trained a one-dimensional convolutional neural network (CNN) with radar signals of three intra-pulse modulation types in the source domain, which were effortlessly obtained and had sufficient samples. Then, we transferred the knowledge obtained by the convolutional layer to nine types of few-shot complex intra-pulse modulation classification tasks in the target domain. We propose an adaptive focal loss function based on the focal loss function, which can estimate the parameters based on the ratio of hard samples to easy samples in the data set. Compared with other existing algorithms, our proposed algorithm makes good use of transfer learning to transfer the acquired prior knowledge to new domains, allowing the CNN model to converge quickly and achieve good recognition performance in case of insufficient samples. The improvement based on the focal loss function allows the model to focus on the hard samples while estimating the focusing parameter adaptively instead of tediously repeating experiments. The experimental results show that the proposed algorithm had the best recognition rate at different sample sizes with an average recognition rate improvement of 4.8%, and the average recognition rate was better than 90% for different signal-to-noise ratios (SNRs). In addition, upon comparing the training processes of different models, the proposed method could converge with the least number of generations and the shortest time under the same experimental conditions.

1. Introduction

Intra-pulse modulation classification of radar signals is an essential area within the field of electronic countermeasures (ECM), which determine the system, usage, and type of enemy radar by analyzing the data received from radar reconnaissance systems [1]. Generally, the classification of intra-pulse modulation can be divided into feature-based and data-based classification [2]. Feature-based classification needs to extract features from the radar signal and carefully design classifiers based on these features [3], while data-based classification can fully retain the data of the radar signal, and is the focus of this paper.
Deep learning is a new research direction in the field of machine learning and is a collective term for a class of pattern analysis methods [4]. With the rapid development of deep learning [4], many researchers now use deep learning to study the classification of intra-pulse modulation of radar signals. Compared to traditional feature extraction, deep learning can automatically extract depth features from radar signals. In [5], Z Huang proposed a deep convolutional neural network-based approach that used the amplitude information of single-polarization SAR images as input to automatically extract the hierarchical spatial characteristics. These features may be more abstract, but they are representative and suitable for classification. Deep learning models can learn all the representation layers together at the same time. By learning the common features, once the model modifies an internal feature, all other features that depend on it will automatically make the corresponding adaptation. To improve the classification accuracy of radar signals, various methods have been proposed. Many researchers convert one-dimensional (1D) radar signals into two-dimensional (2D) time-frequency images for classification. For example, short-time Fourier transformation and CNNs were used to identify six different intra-pulse modulations signals, and the overall classification success rate was over 90% [6]. Ma, XR et al. [3] designed a combination of the short-time Ramanujan Fourier transform and pseudo-Zernike moments invariant feature-based method to recognize different modulation schemes under different parameter variation conditions. To improve classification performance under a low signal-to-noise ratio (SNR), an improved convolutional de-noising automatic encoder [7] was proposed. When the SNR was −9 dB, the encoder could classify 12 kinds of modulated signals, and the classification accuracy was over 95%. Gao, LP et al. [8] proposed an image fusion algorithm using non-multi-scale decomposition to fuse images of a single signal with different time-frequency methods. Liu et al. [9] proposed an algorithm for radar emitter signal recognition transforming raw radio signals into time-frequency image using the Choi–Williams distribution function. Moreover, various studies show that using one-dimensional radar signals for identification makes sense, which is also the focus of this paper. Sun, J et al. [10] designed a novel encoding method to generate high-dimension sequences of equal length as new features in cases of inconsistent features between samples, and proposed a unidimensional convolutional neural network to classify the encoded high-dimension radar signals. Li, X et al. [11] proposed an attention-based approach for radar emitter classification using recurrent neural networks to classify the radar signals. To improve the classification accuracy of radar signals with different SNRs of −14~20 dB, a novel network was proposed, which combines a shallow convolutional neural network (CNN), a long-term memory network (LSTM), and a deep neural network (DNN). Wu, B et al. [12] proposed a novel 1D CNN with an attention mechanism to extract more discriminative features and recognize radar emitter signals. Although methods such as deep CNN can omit feature engineering and automatically extract and learn features from the data [5], at the same time, the number of samples required restricts the development of deep learning in the field of radar signal classification. Most researchers only focus on signal classification under different SNRs. These methods fail to overcome the obstacle of training the depth network with limited radar signals, which is a few-shot recognition problem [13].
To solve these issues, we propose a novel deep network based on transfer learning. Transfer learning is good at applying knowledge or patterns learned in one domain or task to a different but related domain or problem [14]. In general, transfer learning can be classified into three categories: instance-based, feature-based, and shared parameter-based [15]. Instance-based transfer learning studies how to select instances from the source domain that are useful for training in the target domain. For example, an effective assignment of weight to labeled data instances in the source domain can make the distribution of instances in the source and target domains close, so that a reliable learning model with high classification accuracy can be built in the target domain. Dai WY et al. [16] proposed the TrAdaBoost algorithm to improve the classification effect by adjusting the weights of misclassified samples in the source and target domains. Feature-based approaches extract and identify representative features shared between the source and target domains and then use these features to transfer knowledge [17]. Shared parameter-based methods investigate how to find common parameters or prior distributions between the spatial models of source and target data [18]. For few-shot Synthetic Aperture Radar (SAR) image classification, shared parameter-based methods are good at migrating labeled data or learned knowledge structures from related domains with sufficient samples [18,19,20,21,22]. Huang, Z et al. [18] designed an assembled CNN architecture consisting of a classification pathway and a reconstruction pathway, together with an additional feedback bypass. A novel method, deep memory convolution neural networks, for alleviating the problem of overfitting caused by insufficient SAR image samples was proposed in [19]. Rostami, M et al. [22] proposed a novel deep neural network for classifying SAR images that eliminates the need for a huge labeled dataset.
Different from the classification of 2D SAR images, we recognize the received 1D radar signals instead of converting them into images via time-frequency analysis. First, we trained a 1D deep CNN using three large numbers of simple radar signals with labels as the source dataset, and the source task was to classify the three radar signals as correctly as possible. We could easily obtain an optimal depth CNN, because the number of samples was simple and sufficient, and then discard the classifier and higher convolution layers, leaving only the structure and parameters of the lower layer to transfer to the target domain. The lower convolutional layers of the CNN (those layers closer to the input) extracted more general features, while some higher layers of classifier and convolutional layers were applied to specific features, with the higher layers containing more feature semantics and the lower layers containing less feature semantics but more location information. Meanwhile, we proposed an adaptive focus loss function based on focal loss [23], which can adjust the parameter according to the ratio of hard samples to easy samples in the data set. The experimental results demonstrate that, compared with existing algorithms, the proposed algorithm had significantly improved classification accuracy and convergence speed while using less training data.
The remainder of the paper proceeds as follows. Section 2 is concerned with the methodology used for this study. Detailed experimental procedures and discussion of the results are given in Section 3. Section 4 is the conclusion of this paper.

2. Methods

2.1. Related Work

2.1.1. Convolutional Neural Network

Deep learning arose rapidly in the first decade of the 21st century as computing power increased. The Convolutional Neural Network (CNN), the exemplar of deep learning, was first established by Y. L. Cun [24], who designed the famous LeNet-5 to classify handwritten numbers, drawing on artificial neurons and visual perception mechanisms. CNNs share many similarities with ordinary neural networks, in that both of them mimic the structure of human nerves and consist of neurons with learnable weights and bias constants [24]. However, CNNs are more widely used because they avoid complex pre-processing of data and can directly input raw data relying on convolution layers to extract feature maps [25]. In the following years, CNNs evolved based on their classical structure. In 2012 Geoffrey and his student Alex designed the Alex network [26], introducing a nonlinear activation function based on LeNet (ReLU) and a method to prevent overfitting (Dropout, Data augmentation). In 2014, K. S et al. [27] proposed VGG-Net, which contains more layers and uses the same size convolutional filter. The Inception structure of GoogLeNet [28] allows the entire network structure to be expanded in both width and depth. ResNet [29] proposes a residual learning framework that reduces the training burden on the network.

2.1.2. Radar Signal Intra-Pulse Module Classification

Radar emitter signal identification, which aims to obtain information concerning radar systems by analyzing the emitter signals, is an important aspect of electronic warfare and has been extensively studied by numerous researchers. All the CNNs mentioned in Section 2.1.1 have achieved good results in many fields, so it is reasonable to use CNNs to learn temporal correlations and deep features from radar signals for classification. Wei, SJ et al. [30] used sequences in the time, frequency, and autocorrelation domains of the original signal as inputs to a shallow CNN, after which the deep features extracted by the CNN were used as the input to an LSTM network, and finally, a DNN, as the classification network, would directly output the modulation type of the signal. This could achieve high accuracies for four common kinds of measured radar signals. In [9], Z. Liu created a deep CNN using the input of time-frequency spectrums of radar signal intra-pulse modulation to substitute manually constructed features that are time-demanding and neglect delicate characteristics. In [31], Y. Pan et al. used the Hilbert–Huang transform to obtain a wealth of information on the nonlinear and non-stationary properties of radar signals and built a deep residual network to avoid the degradation problem.

2.1.3. Transfer Learning

Transfer learning is a machine learning technique that can transfer knowledge learned in the source domain to the target domain for enhancing the learning of the target task. Transfer learning typically includes the following elements: a source domain D S , a target domain D Τ , a source learning task T S , and a target learning task T Τ . Based on the differences between the source/target domains and tasks, S.J. Pan et al. classified transfer learning into inductive transfer learning, unsupervised transfer learning, and transductive transfer learning [15]. The above three types of transfer learning can be further grouped into four cases: instance transfer learning, feature-representation transfer learning, parameter transfer learning, and relational-knowledge transfer learning [15]. Transfer learning has been successfully adopted in many fields, such as image and video quality, visual categorization, and machinery fault diagnosis. Ling, S et al. [32] surveyed state-of-the-art transfer learning algorithms in visual categorization applications such as object recognition, image classification, and human action recognition. Varga, D et al. [33] pre-trained different types of CNNs based on fusing the decisions of multiple image quality scores that can better characterize authentic image distortion and effectively estimate perceived image quality. In [34], an ImageNet database pre-trained CNN with global average pooling layers was proposed in order to transfer the learned knowledge so that the module can be easily generalized to any input image size and pre-trained CNNs. In [35], Li, C et al. reviews the research progress on deep transfer learning for machinery fault diagnosis in recent years. In the field of radar target classification, Huang, ZL et al. [18] used inductive transfer learning to transfer reconstructed knowledge from convolutional self-encoders to the SAR target classification task. They innovatively used a large number of unlabeled SAR scene images to train the convolutional self-encoder to reconstruct the features well and transferred only the encoder results during the target classification task. Qing, W et al. [36] designed a two-channel CNN combined with bi-directional LSTM architecture to improve the classification performance of the waveforms of cognitive passive radar. They used transductive transfer learning to initialize the target domain classifier with source domain parameters. In this paper, since the source and target domains have different but related distributions, we used parameter transfer learning, in which the source and target tasks share some of the network parameters.

2.1.4. The Focal Loss Function

The imbalance of different sample categories in target detection is a critical issue impacting accuracy. Online hard example mining (OHEM) [37], a typical algorithm to deal with class imbalance, increases the weight of misclassified samples but ignores the easy-to-classify samples. Focal loss is proposed in [23] to solve the category imbalance problem, and is obtained by modifying the standard cross-entropy loss. Compared to OHEM, the focal loss function allows the model to focus more on difficult samples by reducing the weights of easy-to-classify samples during the training process [23]. Currently, focal loss is commonly used in different fields. In the area of text detection in computer vision, X. Tian et al. [38] designed a focal text detection network that uses focal loss to train the network well with an uneven number of samples; it could obtain better performance when the number of samples was insufficient. In the field of medical image processing, where sample imbalance is a serious problem, [39] proposed a network framework using residual neural networks (Res-Net) [29] combined with focal loss for determining left ventricle segmentation from cardiac MRI images. For the problem of ship detection in high-resolution SAR images, ref. [40] designed a RetinaNet-Plus method based on the RetinaNet network, which uses focal loss in the training process to resolve class imbalance and reduce the loss weights of easy-to-classify samples.

2.2. The Proposed Methods

2.2.1. Transfer Learning-Based Convolutional Neural Network

Based on parameter transfer learning, we constructed the network frameworks for the source and target tasks separately, and the source domain used three simple intra-pulse modulation types of radar signals, which were effortlessly obtained and had sufficient samples. The goal was to classify these three radar signals as accurately as possible. In the target domain, we trained with a small number of complex intra-pulse modulation type signals, nine in total, and initialized the convolutional layers of the target domain network using the parameters learned in the source domain instead of random initialization. In the following, we describe the details of the method.
A sequential structure is a convolutional neural network in which the output of each layer is superimposed sequentially as the input of the next layer. Because of its simple structure, it has been widely used as a classical structure. VGG is a typical sequential structured network model, first proposed by K. Simonyan et al. [27]. They added convolutional layers to AlexNet one by one to study the effect of network depth on the recognition effect. Experiments showed that the deeper the network, the better the recognition effect; when the network structure increased to 16 and 19 layers, the effect improved significantly, and therefore these algorithms were called VGG-16 and VGG-19. Therefore, to keep the training simple and the model effortlessly transferred and tuned, we used a sequential structure and simplified the VGG network to build the source and target networks. To ensure that the layer parameters could transfer properly between the source and target domains, we construct a sequential structure with part of the same convolutional layers in the source and target domains. We designed a 1D CNN for the input 1D radar signal intra-pulse modulated sequence; the receptive field of the 1D convolution kernel was continuously translated over the data sequence to observe significant features due to the translation-invariant property of convolution. Each convolution filter of the convolution layer acted iteratively throughout the receptive field to convolve the input sequence, and the convolution result formed a feature map of the input sequence containing the local features of the radar signals. Each convolution filter shared the same parameters, including the same weight matrix and bias term, which were transferred to the target domain after the training of the source domain. The convolved 1D feature map was fed to the pooling layer, and maximum pooling was used for sampling, which was a nonlinear down-sampling method [41]. After acquiring radar signal sequence features by convolution, directly using all the extracted feature data to train the classifier for classification usually expends great computational effort, so the maximum pooling sampling method can be used to down-sample the convolutional features. The convolution and pooling process is shown in Figure 1.
By training the source domain network, we continuously optimized the feature extraction capability of the convolutional layers for 1D radar signals. Thus, these pre-trained convolutional layers could extract data features well in the face of complex target tasks with few-shot samples and inconsistent sample distribution.
To better assemble and train the network and highlight the effect of transfer learning, we designed a single-input, single-output convolutional network structure based on the above structure, as shown in Figure 2.
As shown in Figure 2, the source network consisted of five convolutional layers, each followed by batch normalization to prevent gradient disappearance and speed up training. To transfer the parameters properly between networks, we kept the structure of the lower convolutional layers of the source the same as that of the target network, with the same number and size of convolutional kernels. Table 1 shows the number of feature maps and the size of the convolutional kernels for each convolutional layer. We used multiple 1 × 3 (or 1 × 5) convolutional kernels instead of large-sized kernels to minimize the number of parameters and amount of computational effort while ensuring that the perceptual field of view was not altered [42]. As the number of layers deepened, we increased the number of convolutional kernels to extract more deep features and then down-sampled the feature maps via maximum pooling to prevent overfitting. This design gave the network an inverted triangular shape, i.e., the closer to the input layer, the smaller the number of parameters, and the closer to the output layer, the larger the number of parameters. Such an inverted triangular structure prevents the neural network from losing gradients too quickly during backpropagation [27]. In terms of activation function selection, the Rectifier Linear Unit (ReLU) is widely used because it can solve the gradient disappearance problem [43], but its sparsity tends to lead to dying ReLU, so we used Leaky ReLUs, which assign a non-zero slope to all negative values after each layer of convolution, to solve this problem [44,45].
In ref. [46], Y, W. et al. proved that different convolutional layers extract different-level feature information, and therefore that during the transfer learning process the appropriate convolutional layer parameters should be selected for transfer, instead of all of them. The lower convolutional layers of CNN (the layers closer to the input) extract more general features, i.e., the lower layers contain few feature semantics but have more location information, while some of the higher classifier and convolutional layers apply to specific features which extract more feature semantics, and the semantic features learned in the last few layers are quite different for different datasets. Therefore, the higher convolutional layers are generally related to task objectives and classification, and the lower convolutional layers are more suitable as feature extractors to extract general features for transfer learning. As a result, in the transfer process, we only kept the first four convolutional layers of the source network and discarded the other layers, which contained more semantic information. These optimized lower convolutional layers were more general and could effectively extract structural and detailed features in the radar signal, even in the face of new data.

2.2.2. Adaptive Focus Loss Function (AFL)

To further improve the classification performance of radar signals under few-shot learning, we replaced the original cross-entropy loss with adaptive focal loss, which could automatically adjust its application range depending on the number of samples determined via focal loss.
The focal loss function is widely used and has achieved good results in the field of target detection. The authors proposed focal loss (FL) based on cross-entropy (CE) loss:
C E ( p , y ) = { log ( p ) i f y = 1 log ( 1 p ) o t h e r w i s e .
where p [ 0 , 1 ] is the model’s estimated probability and y { 1 , 1 } indicates the ground-truth class. Facing the problem of sample imbalance, the authors added a factor to the CE loss that assigns different weights to the samples:
F L ( p t ) = ( 1 p t ) γ log ( p t )
p t = { p i f y = 1 1 p o t h e r w i s e .
The focusing parameter γ 0 . When γ = 0 , the focal loss is the traditional cross-entropy loss, and when γ increases, the modulation coefficient also increases. γ smoothly adjusts the proportion of loss accounted for by samples of different difficulties. If a hard sample is misclassified, the p t value is small:
p t 0
( 1 p t ) γ 1
and focal loss has not changed significantly compared to the original loss. By contrast, when the easily classified samples are correctly classified, p t 0 , and the contribution to the total loss is small. Based on this principle, focal loss solves most of the sample imbalance problems very well, but in some scenarios, its effect is not ideal.
Focal loss solves the sample imbalance problem by adjusting the parameter γ to give more weight to the hard samples with poor classification. However, the value of γ is not as large as possible. In ref. [23], the best recognition was achieved when γ = 2 , and the performance decreased when γ > 2 . For other models, the optimal value of γ needs to be determined through a large number of experiments. We can simplify the determination of the optimal parameters by estimating the range of γ . Therefore, we propose an adaptive focus loss function that estimates the value of γ based on the ratio of hard- to easy-to-classify samples. According to the research [47], there is a huge difference in quantity between easy and hard samples during the training process. First, we trained a base classifier using CE, then we predicted the training set and counted the numbers of easy and hard samples, denoted N e and N h , respectively (For radar signal intra-pulse modulation classification, we considered p t 0.1 to be a hard sample and p t 0.9 to be an easy sample. For different models and problems, one can change the judgment threshold of the hard and easy samples). According to the focal loss function, the loss gap is:
l o s s h a r d l o s s e a s y = ( 1 0.1 ) γ ( 1 0.9 ) γ = 9 γ
We defined the difficulty of the training set as the ratio of the number of easy samples to the number of hard ones:
r = N e N h
The focusing parameter γ was used to adjust the contribution of easy and difficult samples to the overall loss to balance their large quantitative differences. Therefore, the loss gap should not be less than the ratio of the number of easy samples to the number of hard ones, and the value of γ should increase as r increases, i.e., the more simple samples, the greater the focus on hard samples. Then we derived the estimate of γ as γ ^ , which should satisfy the following:
9 γ ^   r = N e N h
γ ^   log 9 N e N h
In summary, for the multi-classification problem of radar signal intra-pulse modulation, we propose the Adaptive Focus Loss function (AFL) as follows:
A F L = ( 1 p p r e d i c t i o n × y g r o u n d t r u t h ) γ ^ log ( p p r e d i c t i o n )
In this paper, we obtained γ ^ = log 9 N e N h , and p p r e d i c t i o n was a 1 × 9 vector, which was our model’s estimated probability for the nine radar signals classified. y g r o u n d t r u t h was the vector of the labels after one-hot encoding.

3. Experiments and Results

In this section, we simulated several radar signal datasets with different sample sizes to simulate different small sample cases, which were used to train and test the proposed method and other baseline methods. In all experiments, we used a computer equipped with an Intel 10900K CPU, 64GB of RAM, and a RTX 3070 GPU.

3.1. Dataset and Parameters Setting

Generally, the typical radar signal is dominated by high-power radio frequency (RF) pulses with a carrier band range from 3 MHz to 100 GHz. The radar receiver in our simulation used a local oscillator to mix with the high-frequency radar signal to reduce the frequency of the received signal, and then output a lower-frequency signal through the intermediate frequency (IF) amplifier. Specifically, to ensure that the frequency of the received signal was reduced to replicate a radar system operating in a real environment, a mixer was simulated. The mixer multiplied the RF signal by the local oscillator signal to obtain two output frequencies, the summation and the subtraction of the radio frequency f R F and the local oscillator frequency f L O , which can be expressed as:
f R F + f L O
f R F f L O
Through using a low-pass filter, the summed frequency f R F + f L O could be well suppressed, so that we could obtain the subtracted frequency f R F f L O , which is the IF signal. In this paper, we simulated the low frequency radar signals from the receiver output and used them to train and test our proposed method. For the source domain dataset, we selected three modulation types of simple and widely obtainable radar signals: single-carrier frequency (SCF) signals, linear frequency modulation (LFM) signals, and sinusoidal frequency modulation (SFM) signals. For the target domain dataset, we used nine different kinds of radar signals with complex modulation types comprising binary phase-shift keying (BPSK) signals, binary frequency-shift keying (BFSK) signals, quadrature frequency-shift keying (QFSK) signals, Frank phase-coded (Frank) signals, even quadratic frequency modulation (EQFM) signals, dual-frequency modulation (DLFM) signals, multiple linear frequency modulation (MLFM) signals, and two kinds of composite modulation (LFM–BPSK, BPSK–BFSK) signals. The sampling frequency was 1 GHz, the pulse width of all radar signals varied from 1 μ s to 10 μ s , and other signal parameters are shown in Table 2.
To simulate the real electromagnetic environment, we added additive Gaussian white noise (AWGN) to all signals. The model of the radar signal intercepted by the receiver is given by:
x ( t ) = s ( t ) + n ( t )
n ( t ) is white Gaussian noise, and s(t) is a radar signal. The SNR is defined as:
S N R = 10 log 10 P s P n
where P s represents the effective power of the signal and P n is the effective power of the noise. In Figure 3, taking the LFM signal as an example, we simulated the time-domain waveforms of the same signal at −5 dB, 0 dB, 5 dB, and noiseless, respectively.
Different research fields require a different number of samples in each field. To better investigate the relationship between the number of training samples and the classification effect of the model in radar intra-pulse modulation classification, we introduced a learning curve [48] to plot the classification accuracy versus the number of the training set. The learning curve equation is as follows:
y = 100 + b 1 x b 2
where y is the classification accuracy, x is the training dataset, and b 1 and b 2 correspond to the learning rate and decay rate, respectively. Figure 4 shows the learning curve of classification accuracy versus number of samples for the nine types of intra-pulse modulated radar signals in the target domain.
According to the learning curve, the classification accuracy curve reaches smoothness and the model converges when the total number of training samples is 2000, so in this paper, we defined radar signal intra-pulse modulation classification with less than 2000 training samples as small sample learning.
To validate our proposed method on cases of different sample sizes, we randomly generated samples from each type of signal in the target domain with numbers increasing from 50 to 140 at increments of 10, constituting 10 training sets with different sample sizes. For the three types of radar signals in the source domain, the number of samples for each signal was 5000. The number of training sets for each type of radar signal is shown in Table 3.
Each dataset was generated at the same SNR condition ranging from −5 dB to 5 dB, with a 1 dB interval. An additional set of noise-free signals was generated as a control group to verify the effect of noise on the model performance. For all the above training sets, we produced validation and test sets corresponding to the ratio of 4:1:1. For example, when the target domain training set had 450 samples for each SNR (−5~5 dB and a noise-free dataset for a total of 12 SNRs), the validation set and test set contained 112 samples. Figure 5 and Figure 6 respectively show the waveforms of the source-domain and target-domain intra-pulse modulated radar signals over time when the SNR was 0 dB.

3.2. Experiments on 1D-TLAFLCNN

3.2.1. Experiments on the Source Domain Network

In this section, the source domain network in Section 2 was trained with only three intra-pulse modulation radar signals. The source task was to classify these three radar signals as accurately as possible to optimize the feature extraction capability of the convolutional layers for 1D radar signals. In this stage of training the source domain network, we added three fully connected layers after the convolutional layer, which contained a hidden layer of 256 neurons and a “Leaky ReLU” activation function. The cross-entropy loss function and Adam optimizer were used with a 0.001 learning rate, the batch size was 64, and the network weights were saved for migration when the validation set had the highest accuracy. The result of the classification task is shown in Table 4.
In the subsequent transfer process, we only kept the first four convolutional layers of the source network and discarded the other layers, which contained more semantic information. These optimized lower convolutional layers were more general and could extract structural and detailed features in the radar signal well, even in the face of new data.

3.2.2. Experiments on the Target Domain Network

In this section, we transferred the learned weights to the target network and trained the proposed 1D-TLAFLCNN using the different datasets in the target domain generated in Section 3.1, through the following procedure:
  • Initialize the corresponding convolutional layers of the target domain network with the weights learned from the first four convolutional layers of the source domain network, and freeze these weights;
  • Randomly initialize the parameters of the fully connected layers using a Gaussian distribution;
  • Train the classification layers using the target domain dataset;
  • Fine-tune the entire network by unfreezing all convolutional layers and setting a low learning rate (set to 0.0001) to retrain the entire network in order to incrementally fit the pre-trained features to the new data.
When the SNR is 0 dB and the number of training samples is 450 (50 for each signal), the average accuracy value during the training process is shown in Figure 7
Figure 7 shows that the accuracy of the model stabilized using the validation dataset as the epoch increased, and the accuracy became essentially constant when the epoch reached 90, which indicates that the model converged. Subsequently, we repeated these experiments on training datasets with different SNRs and different sample sizes and tested the classification performance of the proposed model using the highest accuracy weights from the validation dataset. The experiment was repeated 20 times for each case and the average accuracy was taken as the final classification accuracy. The classification accuracies for the nine intra-pulse modulation signals, based on 1D-TLAFLCNN for different cases, are given in Table 5.
As shown in Table 5, we used the training sets of the target domain mentioned in Section 3.1. Each type of signal increased from 50 to 140 with a step of 10, constituting ten training sets with different sample sizes. It can be concluded that the proposed algorithm had good performance in the case of different numbers of small samples, and the average classification accuracy improved as the number of samples increased. Moreover, the proposed method, both in noiseless and noisy environments, had good performance and the classification accuracy steadily improved with increasing SNR. When the SNR was greater than or equal to −1dB, the classification accuracy of the different data sets was over 90%.

3.3. Comparisons with Other Baseline Methods

To show the effectiveness of transfer learning, focal loss, and our proposed adaptive focus loss function (AFL), we constructed five models based on whether or not to add these improvements, while ensuring that they had the same convolutional layer, convolutional kernel size, fully connected layers, batch normalization layer, etc. We used the 1D-CNN as a blank control group, denoted the 1D-CNN with only the focal loss function added (we took the default optimal value as γ = 2 [23]) as 1D-FLCNN ( γ = 2 ), the 1D-CNN with only transfer learning added as 1D-TLCNN, the 1D-CNN with the focal loss function added based on transfer learning as 1D-TLFLCNN ( γ = 2 ), and the transfer learning-based AFL proposed in this paper as 1D-TLAFLCNN, as shown in Table 6.
During the training process, we found that the three methods using transfer learning could have higher classification accuracy on the validation set at the beginning of the iteration, instead of learning from scratch. Transfer learning allows the model to gain some prior knowledge when facing new samples and converge at a faster speed. The accuracy of the five models during the training process at a SNR of 0 dB and 450 training samples (50 for each signal) is shown in Figure 8.
In addition, we compared the number of iterations used to reach convergence and the total time required for the five models in Table 7 and Figure 9.
In addition, we used some representative algorithms as a baseline, including CNN-Qu [7], CNN-Wu [12] and CNN-Wei [30]. All these methods are proposed for intra-pulse modulation classification of radar signals, and have been proved to have good accuracy advantages.
In Table 8 and Figure 10, we compare the classification accuracy of the method proposed in this paper (1D-TLAFLCNN) with other baseline methods at different sample sizes in the 0 dB case and calculate their average accuracy (AA).
In addition, we compared the classification accuracy of all methods on the test set for different SNRs. Figure 11 shows the variation in average accuracy with different SNRs when the sample size was 900 (100 for each signal).
The experimental results demonstrate that 1D-TLAFLCNN performed best under various SNR conditions, with different magnitudes of improvement compared to other algorithms. The classification effectiveness under different training sets in Table 8 shows that the addition of transfer learning and AFL resulted in a greater improvement than using only transfer learning or FL separately, in which transfer learning could reduce the number of convergence generations required for training and converge faster; AFL could estimate γ adaptively based on FL and improved the accuracy of classification.

4. Discussion

4.1. AFL Compared with Different Values of the Focusing Parameter Based on FL

To investigate the effect of different values of the focusing parameter γ in FL on the classification effect and thus prove the effectiveness of our proposed method, we compared the average accuracy of all SNRs with several different few-shot sample sizes, as shown in Figure 12.
As shown in the figure, for the same classification task, the value of the focusing parameter affected the classification accuracy, as it represents how much attention the model pays to the hard samples. A large value of the focusing parameter (e.g., γ = 5 ) tended to over-focus the model on some hard samples and bias the trained model toward these “outliers”, which is often fatal when the training sample is insufficient. In addition, even if the focus parameter took the same value, the classification accuracy kept fluctuating with the sample size. This is because the number of hard and easy samples changed with the sample size, which means that the classification task also changed, and the same value of the focusing parameter could not adjust to all classification tasks. Typically, for new classification tasks, we must conduct quantitative testing to find the appropriate value of the focus parameter γ , and our method is proposed to solve this problem. As we can see in the figure, our proposed method could estimate the range of γ by calculating the proportion of hard and easy samples in the dataset, which was a good improvement compared to other integer values of γ .

4.2. Effect of Different Noise Environments on Experimental Results

The classification performance of the model in noisy environments is one measure of model stability. To this end, we explored the effect of SNRs on the model in both noisy and noiseless environments. We compared the model classification performance of different models for a total of 12 scenarios from −5 dB to 5 dB along with pure signals without noise pollution, as shown in Table 9 and Figure 13. The accuracy in the chart is the average classification accuracy after repeating the experiment for ten different sample sets.
It can be seen that our proposed method had the highest average accuracy, 97.86%, in a noise-free environment. The classification performance of all models for noise-free signals was greatly improved compared to noisy environments, with an average accuracy improvement of 2–5%, since the models are more likely to extract features in pure signals, which are often masked in noisy environments. However, even in the −5dB case, the average accuracy of our proposed algorithm was over 85%. Meanwhile, the model recognition capability steadily improved as the SNR increased, which indicates that our algorithm has certain anti-noise performance.

4.3. Effect of Different Sample Sizes on Experimental Results

As shown in Table 8 and Figure 10, the proposed method had the best accuracy on different few-shot sample datasets, and the average accuracy improved by 8% compared with the traditional CNN algorithm. The comparison between 1D-TLAFLCNN and 1D-TLFLCNN shows that the estimate of γ calculated using AFL had a slightly better classification performance than the default value of taking γ = 2 . Overall, the classification results were better when using the transfer learning approach. The method using FL alone (default value γ = 2 ) yielded improved results in most cases, but the classification accuracy decreased in some cases, which may be because FL focuses excessively on outliers in the sample and makes the model misclassify. In comparison with other baseline methods, the proposed method had the best classification accuracy for all few-shot sample sizes. The method using TL alone had a flatter accuracy curve. It increased steadily as the sample size increased. Transfer learning enables the model to have some prior knowledge, which makes it have good feature extraction ability and effective performance when facing new few-shot problems.
However, as the number of samples increased, our method also showed certain shortcomings: the improvement effect was not obvious in some sample sets, with the accuracy improvement being less than 0.2%. We analyze that as the sample size increased, the difficulty of model training decreased, and the ratio of hard to easy samples tended to become balanced, which decreased the role of transfer learning and AFL.

4.4. Improvement in Training Model Time Consumption

Besides classification accuracy, fast classification is also necessary for radar reconnaissance systems, which need real-time classification. Shorter training times and faster convergence mean that our models could make predictions faster in real-world applications, helping to speed up our analysis of enemy radar systems.
In real radar reconnaissance systems, real-time training and classification of few-shot unknown signals for a new scene is an important aspect, which requires the model to have fast convergence capability. In X-band radar, the airborne radar pulse repetition frequency is f Hz, which emits n pulses at one wave position; therefore, the radar time-on-target collected in one scan is roughly n f seconds. Then, we can calculate the duration of the next received radar echo accordingly, which requires the model to finish converging when the signal is received again. Therefore, considering this practical application, the model convergence epoch and the time spent can be one of the judgments for the performance of the method.
By comparing the number of iterations and the total training time required to reach convergence for the five models in Table 7 and Figure 9, the proposed method converged at 90 generations and 258 s, which reduced the time cost by at least 10%, but this improvement is still far from enough to meet the time requirements for real-time training and testing in practical applications. In future work, we will continue to improve the convergence speed of the model and reduce the training time.

5. Conclusions

To solve the problems associated with the difficulty of training deep cellular neural networks and the insufficient training data available for radar signals, which leads to the low classification accuracy of intra-pulse modulation, a 1D-TLAFLCNN method is proposed for the classification of few-shot intra-pulse modulation radar signals. We used a transfer learning method to transfer the knowledge learned from a large number of simple intra-pulse modulated radar signals in the source domain to a complex modulation classification task in the target domain, and estimated the factor γ adaptively based on FL, which ensured that the model could focus more on the hard samples.
The experimental results show that the proposed method could reduce the number of generations to model convergence compared with other baseline methods, and the model converged within 90 generations with the shortest time of 258s. By comparing the experiments with different values of focusing parameters, AFL had a maximum accuracy improvement of approximately 1.5% based on FL and could reduce repeated experiments by estimating the range of parameters. In addition, the proposed method displayed noise immunity, and the average accuracy of our proposed method was over 85% in the −5 dB case. From further experiments exploring its classification performance on different few-shot datasets, we found that the application of transfer learning could well help the model gain rich prior knowledge and feature extraction ability in few-shot cases, which improved by 8% compared with the traditional CNN algorithm. However, we are also aware of the shortcomings of our method. First, the focusing parameter derived from AFL was a fixed value, but as training proceeded the proportion of hard and easy samples changed, and a fixed focusing parameter could not adapt to the new sample proportion, which made the focal loss function useless. Second, as the number of samples increased, the improvement effect of transfer learning was not obvious. In future work, we hope to develop a method to realize dynamic adjustment of the focusing parameters following changes in the ratio of hard and easy samples.

Author Contributions

Conceptualization, Z.J. and P.L.; methodology, Z.J. and P.L.; software, Z.J.; validation, Z.J.; formal analysis, P.L. and Z.J.; investigation, Z.J., S.Y. and Y.C.; resources, P.L. and B.W.; data curation, Z.J.; writing—original draft preparation, Z.J.; writing—review and editing, Z.J.; supervision, P.L. and B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

The author would like to show their gratitude to the editors and the reviewers for their insightful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, S. Research on recognition algorithm for intra pulse modulation of radar signals. In Proceedings of the 2018 IEEE 3rd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 12–14 October 2018; pp. 1092–1096. [Google Scholar] [CrossRef]
  2. Jin, Q.; Wang, H.Y.; Ma, F.F. An overview of radar emitter classification and identification methods. Telecommun. Eng. 2019, 59, 360–368. [Google Scholar]
  3. Ma, X.R.; Liu, D.; Shan, Y.L. Intra-pulse modulation recognition using short-time ramanujan Fourier transform spectrogram. EURASIP J. Adv. Signal Process. 2017, 1, 42. [Google Scholar] [CrossRef]
  4. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Huang, Z.L.; Datcu, M.; Pan, Z.X.; Lei, B. Deep SAR-Net: Learning objects from signals. ISPRS J. Photogramm. Remote Sens. 2020, 161, 179–193. [Google Scholar] [CrossRef]
  6. Wang, X.B.; Huang, G.M.; Zhou, Z.W.; Tian, W.; Yao, J.L.; Gao, J. Radar emitter recognition based on the energy cumulant of short time Fourier transform and reinforced deep belief network. Sensors 2018, 18, 3103. [Google Scholar] [CrossRef] [Green Version]
  7. Qu, Z.Y.; Wang, W.Y.; Hou, C.B.; Hou, C.F. Radar signal intra-pulse modulation recognition based on convolutional denoising autoencoder and deep convolutional neural network. IEEE Access 2019, 7, 112339–112347. [Google Scholar] [CrossRef]
  8. Gao, L.P.; Zhang, X.L.; Gao, J.P.; You, S.X. Fusion image based radar signal feature extraction and modulation recognition. IEEE Access 2019, 7, 13135–13148. [Google Scholar] [CrossRef]
  9. Liu, Z.; Shi, Y.; Zeng, Y.; Gong, Y. Radar emitter signal detection with convolutional neural network. In Proceedings of the 2019 IEEE 11th International Conference on Advanced Infocomm Technology (ICAIT), Jinan, China, 18–20 October 2019; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2019; pp. 48–51. [Google Scholar]
  10. Sun, J.; Xu, G.L.; Ren, W.J.; Yan, Z.Y. Radar emitter classification based on unidimensional convolutional neural network. IET Radar Sonar Navig. 2018, 12, 862–867. [Google Scholar] [CrossRef]
  11. Li, X.Q.; Liu, Z.M.; Huang, Z.T.; Liu, W.S. Radar emitter classification with attention-based multi-RNNs. IEEE Commun. Lett. 2020, 24, 2000–2004. [Google Scholar] [CrossRef]
  12. Wu, B.; Yuan, S.B.; Li, P.; Jing, Z.H.; Huang, S.; Zhao, Y.D. Radar emitter signal recognition based on one-dimensional convolutional neural network with attention mechanism. Sensors 2020, 20, 6350. [Google Scholar] [CrossRef]
  13. Li, F.Z.; Liu, Y.; Wu, P.X.; Dong, F.; Cai, Q.; Wang, Z. A survey on recent advances in meta-learning. Chin. J. Comput. 2021, 44, 422–446. [Google Scholar] [CrossRef]
  14. Li, Y.; Ding, Z.; Zhang, C.; Wang, Y.; Chen, J. SAR ship detection based on resnet and transfer learning. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 1188–1191. [Google Scholar] [CrossRef]
  15. Pan, S.J.L.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  16. Dai, W.; Yang, Q.; Xue, G.; Yu, Y. Boosting for transfer learning. Machine Learning. In Proceedings of the Twenty-Fourth International Conference (ICML 2007), Corvallis, OR, USA, 20–24 June 2007. [Google Scholar]
  17. Oquab, M.; Bottou, L.; Laptev, I.; Sivic, J. Learning and transferring midlevel image representations using convolutional neural networks. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1717–1724. [Google Scholar] [CrossRef] [Green Version]
  18. Huang, Z.L.; Pan, Z.X.; Lei, B. Transfer learning with deep convolutional neural network for SAR target classification with limited labeled data. Remote Sens. 2017, 9, 907. [Google Scholar] [CrossRef] [Green Version]
  19. Shang, R.H.; Wang, J.M.; Jiao, L.C.; Stolkin, R.; Hou, B.; Li, Y.Y. SAR targets classification based on deep memory convolution neural networks and transfer parameters. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2834–2846. [Google Scholar] [CrossRef]
  20. Zhang, W.; Zhu, Y.F.; Fu, Q. Deep transfer learning based on generative adversarial networks for SAR target recognition with label limitation. In Proceedings of the 2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Chongqing, China, 11–13 December 2019; pp. 1–5. [Google Scholar] [CrossRef]
  21. Huang, Z.L.; Dumitru, C.O.; Pan, Z.X.; Lei, B.; Datcu, M. Classification of large-scale high-resolution SAR images with deep transfer learning. IEEE Geosci. Remote Sens. Lett. 2021, 18, 107–111. [Google Scholar] [CrossRef] [Green Version]
  22. Rostami, M.; Kolouri, S.; Eaton, E.; Kim, K. Deep transfer learning for few-shot SAR image classification. Remote Sens. 2019, 11, 1374. [Google Scholar] [CrossRef] [Green Version]
  23. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Cun, Y.L.; Boser, B.; Denker, J.; Henderson, D.; Howard, R.; Habbard, W.; Jackel, L. Handwritten digit recognition with a back-propagation network. Adv. Neural Inf. Process. Syst. 1990, 2, 396–404. [Google Scholar]
  25. Cun, L.Y.; Kavukcuoglu, K.; Farabet, C. Convolutional networks and applications in vision. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems(ISCAS), Paris, France, 30 May–2 June 2010; pp. 253–256. [Google Scholar]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
  27. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. Comput. Sci. 2014, 1409, 1556. [Google Scholar]
  28. Szegedy, C.; Liu, W.; Jia, Y.Q.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  29. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  30. Wei, S.J.; Qu, Q.Z.; Su, H.; Wang, M.; Shi, J.; Hao, X.J. Intra-pulse modulation radar signal recognition based on CLDN network. IET Radar Sonar Navig. 2020, 14, 803–810. [Google Scholar] [CrossRef]
  31. Pan, Y.W.; Yang, S.H.; Peng, H.; Li, T.Y.; Wang, W.Y. Specific emitter identification based on deep residual networks. IEEE Access 2019, 7, 54425–54434. [Google Scholar] [CrossRef]
  32. Shao, L.; Zhu, F.; Li, X.L. Transfer learning for visual categorization: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 1019–1034. [Google Scholar] [CrossRef] [PubMed]
  33. Varga, D. No-reference video quality assessment using multi-pooled, saliency weighted deep features and decision fusion. Sensors 2022, 22, 2209. [Google Scholar] [CrossRef] [PubMed]
  34. Varga, D. Multi-pooled inception features for no-reference image quality assessment. Appl. Sci. 2020, 10, 2186. [Google Scholar] [CrossRef] [Green Version]
  35. Li, C.; Zhang, S.H.; Qin, Y.; Estupinan, E. A systematic review of deep transfer learning for machinery fault diagnosis. Neurocomputing 2020, 407, 121–135. [Google Scholar] [CrossRef]
  36. Wang, Q.; Du, P.F.; Yang, J.Y.; Wang, G.H.; Lei, J.J.; Hou, C.P. Transferred deep learning based waveform recognition for cognitive passive radar. Signal Process. 2019, 155, 259–267. [Google Scholar] [CrossRef]
  37. Shrivastava, A.; Gupta, A.; Girshick, R. Training region-based object detectors with online hard example mining. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 761–769. [Google Scholar] [CrossRef] [Green Version]
  38. Tian, X.W.; Wu, D.; Wang, R.; Cao, X.C. Focal text: An accurate text detection with focal loss. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 2984–2988. [Google Scholar] [CrossRef]
  39. Chen, M.Q.; Fang, L.; Liu, H.F. FR-NET: Focal loss constrained deep residual networks for segmentation of cardiac MRI. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 764–767. [Google Scholar] [CrossRef]
  40. Su, H.; Wei, S.J.; Wang, M.K.; Zhou, L.M.; Shi, J.; Zhang, X.L. Ship detection based on retinaNet-Plus for high-resolution SAR imagery. In Proceedings of the 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 26–29 November 2019; pp. 1–5. [Google Scholar] [CrossRef]
  41. Nagi, J.; Ducatelle, F.; Caro, G.A.D.; Cireşan, D.; Meier, U.; Giusti, A.; Nagi, F.; Schmidhuber, J.; Gambardella, L.M. Max-pooling convolutional neural networks for vision-based hand gesture recognition. In Proceedings of the 2011 IEEE International Conference on Signal and Image Processing Applications (ICSIPA), Kuala Lumpur, Malaysia, 16–18 November 2011; pp. 342–347. [Google Scholar] [CrossRef] [Green Version]
  42. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar] [CrossRef] [Green Version]
  43. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the 14th International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  44. Xu, B.; Wang, N.Y.; Chen, T.Q.; Li, M. Empirical evaluation of rectified activations in convolutional network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  45. Xu, J.; Li, Z.S.; Du, B.W.; Zhang, M.M.; Liu, J. Reluplex made more practical: Leaky ReLU. In Proceedings of the 2020 IEEE Symposium on Computers and Communications (ISCC), Rennes, France, 7–10 July 2020; pp. 1–7. [Google Scholar] [CrossRef]
  46. Yu, W.; Yang, K.Y.; Yao, H.X.; Sun, X.S.; Xu, P.F. Exploiting the complementary strengths of multi-layer CNN features for image retrieval. Neurocomputing 2017, 237, 235–241. [Google Scholar] [CrossRef]
  47. Li, B.; Liu, Y.; Wang, X. Gradient Harmonized Single-Stage Detector. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8577–8584. [Google Scholar] [CrossRef]
  48. Figueroa, R.L.; Zeng-Treitler, Q.; Kandula, S.; Ngo, L.H. Predicting sample size required for classification performance. BMC Med. Inform. Decis. Mak. 2012, 12, 8. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Convolution and pooling process.
Figure 1. Convolution and pooling process.
Remotesensing 14 01950 g001
Figure 2. The architecture of the one-dimensional convolutional neural network with adaptive focal loss function based on transfer learning.
Figure 2. The architecture of the one-dimensional convolutional neural network with adaptive focal loss function based on transfer learning.
Remotesensing 14 01950 g002
Figure 3. Differentiation between LFM signals based on the noise level. (a) −5 dB; (b) 0 dB; (c) 5 dB; (d) noiseless.
Figure 3. Differentiation between LFM signals based on the noise level. (a) −5 dB; (b) 0 dB; (c) 5 dB; (d) noiseless.
Remotesensing 14 01950 g003
Figure 4. The learning curve of classification accuracy versus number of samples.
Figure 4. The learning curve of classification accuracy versus number of samples.
Remotesensing 14 01950 g004
Figure 5. Different types of intra-pulse modulated radar signals in the source domain. (a) SCF; (b) LFM; (c) SFM.
Figure 5. Different types of intra-pulse modulated radar signals in the source domain. (a) SCF; (b) LFM; (c) SFM.
Remotesensing 14 01950 g005
Figure 6. Different types of intra-pulse modulated radar signals in the target domain. (a) BPSK; (b) BFSK; (c) QFSK; (d) FRANK; (e) EQFM; (f) DLFM; (g) MLFM; (h) LFM–BFSK; (i) BPSK–BFSK.
Figure 6. Different types of intra-pulse modulated radar signals in the target domain. (a) BPSK; (b) BFSK; (c) QFSK; (d) FRANK; (e) EQFM; (f) DLFM; (g) MLFM; (h) LFM–BFSK; (i) BPSK–BFSK.
Remotesensing 14 01950 g006
Figure 7. The average accuracy value during the training process.
Figure 7. The average accuracy value during the training process.
Remotesensing 14 01950 g007
Figure 8. The accuracy of the five models during the training process at a SNR of 0 dB and a number of training samples of 450 (50 for each signal).
Figure 8. The accuracy of the five models during the training process at a SNR of 0 dB and a number of training samples of 450 (50 for each signal).
Remotesensing 14 01950 g008
Figure 9. The number of iterations and time usage for training the model in different methods.
Figure 9. The number of iterations and time usage for training the model in different methods.
Remotesensing 14 01950 g009
Figure 10. The classification accuracy of different methods with different training sets at 0 dB.
Figure 10. The classification accuracy of different methods with different training sets at 0 dB.
Remotesensing 14 01950 g010
Figure 11. The classification accuracy of different methods with different SNRs when the sample size was 900 (100 for each signal).
Figure 11. The classification accuracy of different methods with different SNRs when the sample size was 900 (100 for each signal).
Remotesensing 14 01950 g011
Figure 12. The average classification accuracy of all SNRs for different focusing parameters with different sample sizes.
Figure 12. The average classification accuracy of all SNRs for different focusing parameters with different sample sizes.
Remotesensing 14 01950 g012
Figure 13. Classification accuracies with different noise environments.
Figure 13. Classification accuracies with different noise environments.
Remotesensing 14 01950 g013
Table 1. The design scheme for the source and target networks.
Table 1. The design scheme for the source and target networks.
LayerKernel SizeChannelMax Pooling
Source Network
Target Network
Conv11 × 5321 × 2
Conv21 × 5641 × 2
Conv31 × 51281 × 4
Conv41 × 32561 × 4
Source NetworkConv51 × 32561 × 4
Table 2. Parameters of the twelve different intra-pulse modulations of the radar emitter signals.
Table 2. Parameters of the twelve different intra-pulse modulations of the radar emitter signals.
TypeCarrier FrequencyParameter
Source
Domain
SCF50~500 MHzNone
LFM100~400 MHzBandwidth: 20 MHz~150 MHz
SFM100~400 MHzBandwidth: 20 MHz~100 MHz
Target
Domain
BPSK100~300 MHz5,7,11,13-bit Barker code
BFSK100~400 MHz5,7,11,13-bit Barker code
100~400 MHz
QFSK100~300 MHz16-bit Frank code
100~300 MHz
100~300 MHz
100~300 MHz
FRANK100~400 MHzPhase number: 4–6
EQFM100~300 MHzBandwidth: 5 MHz to 100 MHz
DLFM100~300 MHzBandwidth: 10 MHz to 150 MHz
MLFM100~300 MHzBandwidth: 30 MHz to 100 MHz
Bandwidth: 30 MHz to 100 MHz
Segment: 20–80%
100~300 MHz
LFM-BPSK100~300 MHzBandwidth: 5 MHz to 150 MHz
5,7,11,13-bit Barker code
BPSK-BFSK100~400 MHz5,7,11,13-bit Barker code
100~400 MHz
Table 3. The number of training sets for each type of radar signal.
Table 3. The number of training sets for each type of radar signal.
Signal NumberType
Source Domain5000 per signalSCF
LFM
SFM
Target DomainEach type of signal increase from 50 to 140 with a step of 10, constituting 10 training sets with different sample sizes respectively.BPSK
BFSK
QFSK
FRANK
EQFM
DLFM
MLFM
LFM–BPSK
BPSK–BFSK
Table 4. The classification accuracy for the source domain at different SNRs.
Table 4. The classification accuracy for the source domain at different SNRs.
SNR/dB−5−4−3−2−1012345
SCF0.98970.99230.99800.99900.99690.99801.00001.00001.00001.00001.0000
LFM0.98400.98840.99170.99150.99901.00000.99901.00001.00001.00001.0000
SMF0.98120.98420.98630.98880.99900.99690.99700.99831.00001.00001.0000
Average0.98500.98830.99200.99310.99830.99830.99870.99941.00001.00001.0000
Table 5. The classification accuracies of target domain signals, based on 1D-TLAFLCNN for different cases.
Table 5. The classification accuracies of target domain signals, based on 1D-TLAFLCNN for different cases.
Size−5 dB−4 dB−3 dB−2 dB−1 dB0 dB1 dB2 dB3 dB4 dB5 dBNoiseless
SNR
500.83530.84440.87810.89960.90310.91160.91480.92270.93470.93530.93650.9646
600.84000.85490.88410.90590.92080.92190.91710.92460.93480.93720.94400.9735
700.84480.86160.89660.91720.92640.92870.93770.93910.94410.95060.95110.9746
800.84630.86370.90670.92110.92730.92960.94330.94440.94540.95160.95120.9754
900.84800.86610.91680.92380.93810.93330.94520.94610.94560.95290.95390.9778
1000.84960.87770.92010.92650.94160.94050.94870.94640.94680.95490.95880.9798
1100.85570.88040.92160.93260.94230.94290.95110.94660.94800.95930.95940.9811
1200.85810.89190.92660.93970.94270.94610.95250.95570.95890.96390.96480.9845
1300.87010.91640.92780.93970.94640.95240.95620.95660.96420.96910.97670.9863
1400.87770.92060.93410.94110.95140.96490.96740.97090.96980.97860.98540.9889
Table 6. Differences between the proposed method and other baseline methods.
Table 6. Differences between the proposed method and other baseline methods.
Focal LossTransfer LearningAdaptive Focus Loss
1D-CNNnonono
1D-FLCNN ( γ = 2 )yesnono
1D-TLCNNnoyesno
1D-TLFLCNN ( γ = 2 )yesyesno
1D-TLAFLCNNnoyesyes
“yes” means that the model uses this improvement, “no” means it is not used.
Table 7. The number of iterations and time usage for training the model via different methods.
Table 7. The number of iterations and time usage for training the model via different methods.
Number of EpochsTotal Time (RTX3070)
1D-TLAFLCNN90258 s
1D-TLFLCNN ( γ = 2 )100284 s
1D-TLCNN120355 s
1D-FLCNN ( γ = 2 )140416 s
1D-CNN160483 s
Table 8. The classification accuracy of different methods with different training sets at 0dB.
Table 8. The classification accuracy of different methods with different training sets at 0dB.
Datasets5060708090100110120130140AA
1D-TLAFLCNN0.91160.92190.92870.92960.93330.94050.94290.94610.95240.96490.9372
1D-TLFLCNN ( γ = 2 )0.90070.91770.91850.92780.92610.93780.94070.93700.93840.95870.9303
1D-TLCNN0.89380.91110.92230.92780.93170.93890.93950.94070.94540.96190.9313
1D-FLCNN ( γ = 2 )0.90270.89630.91140.92220.92120.92880.91530.93150.93520.95240.9217
1D-CNN0.80740.81890.83510.84130.84560.84730.85040.85960.87040.86440.8441
CNN-Qu0.82860.82930.83890.83730.84110.84230.84390.85010.85770.86330.8433
CNN-Wu0.87510.87650.88010.88340.88620.88970.89350.89910.90330.90550.8892
CNN-Wei0.81770.82090.82370.82770.83070.83480.84550.83970.85080.85590.8347
Table 9. Classification accuracies with different noise environments.
Table 9. Classification accuracies with different noise environments.
SNR−5 dB−4 dB−3 dB−2 dB−1 dB0 dB1 dB2 dB3 dB4 dB5 dBNoiseless
1D-TLAFLCNN0.85250.87770.91120.92470.93400.93720.94340.94530.94920.95530.95810.9786
1D-TLFLCNN ( γ = 2 )0.84860.87280.90630.92180.93060.93230.93650.94010.94440.94990.95570.9702
1D-TLCNN0.84260.86190.88430.90490.91290.92350.92770.93020.93260.94050.94060.9605
1D-FLCNN ( γ = 2 )0.83930.86520.89450.90390.91330.91760.92960.93440.93560.94030.94270.9642
1D-CNN0.81170.82670.85240.87810.89470.91260.92050.92500.92680.93160.93380.9502
CNN-Qu0.74360.76330.81610.83910.84980.85160.86360.86930.87020.87250.87460.9011
CNN-Wu0.76240.79130.82180.83080.84740.85100.85880.86970.87110.87820.88680.9377
CNN-Wei0.80950.83900.85110.87440.89620.89760.91590.92320.93220.93520.93610.9599
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jing, Z.; Li, P.; Wu, B.; Yuan, S.; Chen, Y. An Adaptive Focal Loss Function Based on Transfer Learning for Few-Shot Radar Signal Intra-Pulse Modulation Classification. Remote Sens. 2022, 14, 1950. https://doi.org/10.3390/rs14081950

AMA Style

Jing Z, Li P, Wu B, Yuan S, Chen Y. An Adaptive Focal Loss Function Based on Transfer Learning for Few-Shot Radar Signal Intra-Pulse Modulation Classification. Remote Sensing. 2022; 14(8):1950. https://doi.org/10.3390/rs14081950

Chicago/Turabian Style

Jing, Zehuan, Peng Li, Bin Wu, Shibo Yuan, and Yingchao Chen. 2022. "An Adaptive Focal Loss Function Based on Transfer Learning for Few-Shot Radar Signal Intra-Pulse Modulation Classification" Remote Sensing 14, no. 8: 1950. https://doi.org/10.3390/rs14081950

APA Style

Jing, Z., Li, P., Wu, B., Yuan, S., & Chen, Y. (2022). An Adaptive Focal Loss Function Based on Transfer Learning for Few-Shot Radar Signal Intra-Pulse Modulation Classification. Remote Sensing, 14(8), 1950. https://doi.org/10.3390/rs14081950

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop