Next Article in Journal
Internal Damage Detection of Composite Structures Using Passive RFID Tag Antenna Deformation Method: Basic Research
Previous Article in Journal
Software-Defined NB-IoT Uplink Framework—The Design, Implementation and Use Cases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Specific Radar Recognition Based on Characteristics of Emitted Radio Waveforms Using Convolutional Neural Networks

Institute of Radioelectronics, Faculty of Electronics, Military University of Technology, 00-908 Warsaw, Poland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(24), 8237; https://doi.org/10.3390/s21248237
Submission received: 22 October 2021 / Revised: 29 November 2021 / Accepted: 3 December 2021 / Published: 9 December 2021
(This article belongs to the Topic Artificial Intelligence in Sensors)

Abstract

:
With the increasing complexity of the electromagnetic environment and continuous development of radar technology we can expect a large number of modern radars using agile waveforms to appear on the battlefield in the near future. Effectively identifying these radar signals in electronic warfare systems only by relying on traditional recognition models poses a serious challenge. In response to the above problem, this paper proposes a recognition method of emitted radar signals with agile waveforms based on the convolutional neural network (CNN). These signals are measured in the electronic recognition receivers and processed into digital data, after which they undergo recognition. The implementation of this system is presented in a simulation environment with the help of a signal generator that has the ability to make changes in signal signatures earlier recognized and written in the emitter database. This article contains a description of the software’s components, learning subsystem and signal generator. The problem of teaching neural networks with the use of the graphics processing units and the way of choosing the learning coefficients are also outlined. The correctness of the CNN operation was tested using a simulation environment that verified the operation’s effectiveness in a noisy environment and in conditions where many radar signals that interfere with each other are present. The effectiveness results of the applied solutions and the possibilities of developing the method of learning and processing algorithms are presented by means of tables and appropriate figures. The experimental results demonstrate that the proposed method can effectively solve the problem of recognizing raw radar signals with agile time waveforms, and achieve correct probability of recognition at the level of 92–99%.

1. Introduction

In radio-electronic reconnaissance systems we receive and then measure the basic time, frequency and spatial parameters (related to the scanning of the antenna) in order to recognize their emission sources (in our case, radars), and we do not visualize the spatial situation using signals. The radio-electronic reconnaissance systems extract the basic characteristic parameters from measured radar signals. Based on these parameters, we can obtain information such as the system, application, type and platform of the radar, and further deduce the battlefield situation, threat level, activity rule, tactical intention, etc., and provide important intelligence support for our own decision-making system. The modern electromagnetic environment is considered complex due to a multitude of signals originating from a number of different radars (emitters), and in the case of signals coming from the same radar their parameters (features) are measured with low accuracy. In many cases the radar may change one or several signal parameters in order to perform its task more efficiently [1,2]. Since each radar has limited parameter ranges (e.g., transmits within a limited frequency band) and often identifiable characteristics, it is assumed that radar signals with similar characteristics originate from the same device [3,4].
Artificial neural networks (ANNs) currently achieve high efficiency in the field of signal and image processing, especially in the field of different objects recognition [5,6,7,8,9,10] and extending situational awareness supporting, for example, SAR imaging [11,12,13]. Interest in the implementation of artificial intelligence methods in the field of broadly understood classification is constantly growing [14,15,16,17,18,19,20,21,22,23,24,25,26,27,28], which is caused by the continuous improvement of learning algorithms and the increase in the computing power of computers [29,30,31,32,33,34]. The conventional approach to digital signals processing with the use of convolutional artificial neural networks (CNNs) is the signal transition from the time domain to the time-frequency domain, i.e., to the signal image [35,36,37,38], which takes into account a typical structure of the convolutional neural network (CNN) [25,26,39], which is much easier to process than the raw signal.
Opponents of artificial intelligence methods, especially deep learning methods, will point out here that the learning process may take too long in relation to the use of manually programmed filters/processing algorithms. However, in the case of deep neural networks, it is possible to use already learned filters from convolutional layers [40,41,42,43,44]. The only problem that remains then is teaching a three-layer neural network, which is usually learned quite quickly in relation to convolutional layers [35,42,45,46,47]. Another cause for optimism is that modern graphics cards make it possible to reduce the training time of ANN classifiers from several weeks to hours or even minutes (here the leaders in this field are NVIDIA and AMD) [30,31,48]. The ANN processing methods discussed in this article will operate on digital signals, and will show the possibility of optimizing the methods of classifying these radar signals depending on the receiver that produces a given signal, as well as the acceleration of the processing itself. The nature of ANN structures and their operation can be used for data processing, and for signals at higher frequencies directly or indirectly (after rescale or transforming high-frequency signals to the lower baseband frequency) in which the telecommunications or radiolocation devices operate [8,39,49].
In our work we chose deep learning (DL) with convolutional neural networks for signal processing because we wanted to create an agile and adjustable (to radio background) radar signal recognition system [50,51,52]. DL for signal processing can be superior in comparison to the standard models of neural networks with full connections (i.e., as used in [53]) or processing with hand-crafted features (e.g., eigenvector analysis methods [54], random forest (RF), support vector machines (SVM), k-nearest neighbors [55]), because they do not require neither the use of preliminary signal processing, nor additional feature extraction, nor the standardization of measured vectors. The neural network acquires all the features of extraction during the learning process. However, as it is stated in other publications [55,56], to achieve better results in comparison to algorithms with hand-crafted extraction features we need enough data patterns to learn the deep model otherwise we do not achieve better performance but can lead the deep model to overfitting. A better approach to improve system performance can be achieved by supporting the DL method with hand-crafted pre-processing so we can speed up the learning process and outperformance application of each method separately [55]. In our work such preliminary processing can be called the use of the FIR filter before the entry of the classifier. Without the FIR filter, the convolutional neural network would still be able to classify the raw signals, except that we would probably have to add one layer or two to the network in order to teach it how to filter the desired signals from the frequency domain.
In paper [54] eigenvector analysis is used, which is quite efficient, however, as the number of analyzed devices increases, the eigenvector matrix grows, and this can gradually lead to deteriorating system performance. In addition, the methods based on the analysis of eigenvectors degrade some information when converting the classifying vectors to the matrix of eigenvectors. The degradation of some information may lead to deterioration of the classification under disturbance conditions or when geometrical changes of the input samples are inserted.
When using artificial neural networks (especially convolutional networks), adding further radar classifying vectors, it is not required to enlarge the structure of the neural network, but we can consider use of learning transfer, which means disconnecting the last layer of the neural network and connecting another one with the number of required classes and training this layer with the use of the pre-trained first layers [56]. Another argument regarding the use of convolutional neural networks is that there is no need to lose information in the pre-processing signals in the form of reducing the dimensions or averaging the available processing channels. The network, during the learning process through the use of layers such as convolutional and max pooling, decides on the basis of the training set which signal features are unimportant and what will be relevant in the process of proper classification [56].
It also should be noted that in the methods shown in other works [53,54] we process only post-detection data without introducing any significant disturbances. In the work presented here we process only the raw signals without complicated pre-processing, which is mostly carried out by layers of the neural network. Processing signals of the size 355 × 10 3 samples with the use of the method of eigenvector analysis would force us to create huge matrices, the calculation of which in real time may sometimes prove impossible.
In the following parts of this article, Section 2 contains a description of an electromagnetic environment, the measurement method and the processing of radar signal parameters, as well as the method of constructing the training dataset applied in the proposed model for recognizing emitter sources. Section 3 presents a detailed structure of CNN designed for radar signals recognition based on the measured parameters, i.e., pulse repetition interval, pulse duration, radio frequency and antenna scan period. The learning method designed for a CNN structure, the software used and the mathematical formulas are described in Section 4. The simulation environment used for testing the efficiency of radar signal recognition is described in Section 5. Analysis of experiment results, shown in the form of tables and figures, is presented in Section 6. Section 7 contains a summary of the obtained simulation results and the conclusion.

2. Description of the Database

2.1. Description of Radar Signal Parameters

The electronic support measures (ESM) systems measure the basic parameters of incoming radar signals (frequency, amplitude, bearing and elevation angles, pulse width, time of arrival and sometimes polarization). The data collected are sorted into groups considered to be from a single emitter and subsequently used to compute the time-dependent parameters (pulse repetition frequency, antenna rotation period, etc.) [1,2,3]. Finally, the system matches the “signal signatures”, composed from the average parameters from each group, with the characteristics of known emitters stored in the emitter database (EDB). This action enables the system to identify and classify the incoming radar signals which may have a high degree of inherent uncertainty arising from the methods of data gathering and processing [2,3]. In the electromagnetic environment a great deal of information collected by the receiver is processed in real time (Figure 1).
The basic measured radar signals in mobile systems for identification of radar signals MUR-20 or in the ELINT system of recognition of onboard RF emitters, produced by PIT-RADWAR, are the following:
-
Automatic detection direction that finds and monitors the emission sources with a frequency ranging from 500 MHz to 18 GHz;
-
Signal parameters measured: frequency, pulse width, amplitude, direction of arrival, pulse repetition frequency, antenna rotation period;
-
Deinterleaving;
-
Acousto-optical channel of spectrum analyzer 500 MHz and channel of compression spectrum analyzer 40 MHz;
-
Radio frequency measurement with 1 MHz accuracy;
-
Instantaneous time parameters measurement with 25 ns accuracy.
The signal structure generated by a single radar varies across time and depends on parameters such as pulse repetition interval (PRI), pulse duration (PD), radio frequency (RF) and antenna scan period (SP). The intervals of individual radar signal parameters may overlap, therefore the given signals may be more or less similar to each other at certain times. The recognition of measured radar signal parameters is also based on the analysis of their temporal structure, which will be reconstructed using the simulation environment, because unlike the PRI, PD, RF and SP parameters, we do not have signal temporal structures assigned to particular classes of radar signals.
The radar signals characteristics presented in this section belong to 18 different types of radars (classes). Table 1 presents the confidence intervals of radar signal parameters calculated based on their measurement data.
Figure 2 and Figure 3 show the time and frequency dependencies between individual classes of signal parameters. The intervals of individual classes have been marked with different colors. The number of classes is successively marked on the right side from 0 to 17. The interval width for each class means the time or frequency interval in which the signals will subsequently be generated for each class.

2.2. Constructing a Set of Data for Training a Neural Network

As noted in the previous section, based on the description and parameters intervals assigned to the individual radars, these classes in certain ranges can overlap each other, which can significantly affect the effectiveness of proper signal source recognition. To reduce the risk of a false recognition, the final decision concerning the recognition of a given signal source is announced only after receiving all the parameters described above. For this purpose, three training sets were created.
(a)
The first training dataset consists of time waveforms (TW) of the signals with variable PD, RF and intra-pulse modulation. An example of the time waveforms of a signal simulated with the use of a simulation environment (which is described in more detail in Section 5), based on the parameters presented in Table 1, is depicted in Figure 4.
(b)
The second training dataset consists of variable PRI waveforms which change depending on the applied inter-pulse modulation. Below, in Figure 5, these changes of PRI are shown.
(c)
The third training dataset consists of variable PD waveforms changing from pulse to pulse.

2.3. The Similarity between the Classes of Signals

From the waveforms of changes of individual parameters presented in Section 2.1 (Figure 2 and Figure 3), the overlapping of individual classes of signals can be noted, taking into account time and frequency dependencies. Table 2 below depicts the classes with overlapping ranges of changes in time (PRI, PD, SP) and frequency (RF).

3. Proposed Model

The CNN model applied in this paper is a multi-layer (deep) structure containing inter-area convolutional connections in the first layers, batch normalization layers [57] applied during the learning process to accelerate and stabilize it, a sub-sampling layer (max-pooling) [58] and dropout layers reducing the probability of overfitting [59]. Figure 6 below depicts the structure of the CNN used for training the signals of changeable intra-pulse modulation TW (RF and PD), PRI and PD signals with inter-pulse modulation.
The first CNN layers in this structure, in this case the convolutional layers 1D, are designed to extract the features from the signals tested [60]. Traditional convolutional 2D layers operate on images and use the 2D filters to extract features of input signal, where convolutional 1D layers have 1D filters and operate directly on the 1D signal without transformations (such as transforming signal 1D to a spectrogram). Extraction of selected signal features is carried out by the means of a convolution operation of the input signal with feature maps (filters) obtained during the learning process. Figure 7 shows examples of signal filters (visualization of weight maps) obtained during the learning process. Due to the CNN structure used together with the one-dimensional convolution layers, the illustration of these filters was presented in the form of sampled time waveforms.
In the designed structure feature extraction has been performed three times. After each extraction, the maps of neuron responses obtained at the output of the convolution layer were down-sampled with the help of the max pooling layer [58] in order to facilitate the features’ extraction in the successive convolution layers. The last layers—the dense (full connected) layers [4,61]—are the layers deciding on the basis of the extracted features which class of signals we deal with. Table 3 presents a detailed description of the structure of each layer in the CNN structure described above.
The second CNN structure uses three structures similar to the model described above. The structures are connected by concatenate layer [63] and work together simultaneously processing PRI, PD parameters and TW to determine the class of input signals. The last layer of this structure is the dense type and makes final recognition of the signal (Figure 8 and Table 4). The model described here is depicted in more detail in Figure 8.
The third CNN model is intended to recognize raw samples of vector signals. The first processing layers are comparable to structures described above, although input of the model is quite large with 335,544 samples and the last layers are transpose convolution layers [64]. Figure 9 and Table 5 presented below describe the model in more detail.
Table 6 presents the memory requirement parameters and performance while processing of single input vector data for the described above CNN architecture.

4. CNN Learning

The learning of CNN structures presented in this article was carried out using the TensorFlow [4] library, which enables the acceleration of the learning process using general-purpose computing in graphics processing units (GP-GPU) processors (graphics cards), the use of modern learning algorithms and constructing structures of artificial neural networks.
The learning process was carried out using Adam Optimizer, a modification of the stochastic gradient descent (SGD) algorithm [65], which allows for the efficient solving of the optimization problems for multidimensional objective functions. Due to its efficiency, this algorithm is implemented by default in ANN training libraries such as TensorFlow or Keras [66]. The Adam Optimizer gradient descent algorithm is an adaptive learning algorithm, which is an extension and combination of two methods, i.e., the AdaGrad and RMSProp methods [65]. Basic Equations (1–6) of the Adam Optimizer algorithm are presented below, according to which the successive values of changes of weight vector in the ANN are calculated in the following way:
g ( i ) = E ( W ( i ) ) ,
m ( i ) = β 1 ( i ) m ( i 1 ) + ( 1 β 1 ( i ) ) g ( i ) ,
v ( i ) = β 2 ( i ) v ( i 1 ) + ( 1 β 2 ( i ) ) g ( i ) ,
m ^ ( i ) = m ( i ) 1 β 1 ( i ) ,
v ^ ( i ) = m ( i ) 1 β 2 ( i ) ,
W ( i ) = W ( i 1 ) α m ^ ( i ) v ^ ( i ) + ϵ
where i is a number of the current epoch in the learning process, g ( i ) is the gradient of the objective function, m ( i ) is the first-order moment of the estimation of changes in the value of the weight vector, v ( i ) is the second-order moment of the estimate of changes in the value of the weight vector, m ^ ( i ) denotes the normalization of the moment m ( i ) , v ^ ( i ) denotes the normalization of the moment v ( i ) , β 1 ( i ) is the moment decay factor m(i), β 2 ( i ) is the moment decay factor v ( i ) , W ( i ) denotes the weight vector of CNN, α is the learning coefficient or the step of changes in the weight vector updating, and ϵ is the small value ensuring stability of calculations.
In the first step the Adam Optimizer program calculates the gradient of the objective function (multivariate error function) g ( i ) in Equation (1). Then, respectively, the first m ( i ) and second-order moments v ( i ) are calculated. These are the values of the estimation of the weight vector changes W ( i ) in Equations (2) and (3). Before the actual updating of the weights vector, Equation (6), the correction (normalization) of the moments v ( i ) and m ( i ) is performed. During each update, along with the learning progress, the effects of the moments of the first and second order of the estimates are minimized on the basis of the β 1 and β 2 coefficients.
Due to the fact that the presentation of the learning patterns is carried out using an indirect method between online (updating weights after presenting one learning pattern) or total batch (updating weights after presenting the entire training dataset), i.e., the learning dataset is divided into packages (mini-batches) with the number of patterns N. An additional element accelerating the achievement of the global minimum of the objective function of the problem under consideration (signal recognition) was the use of training data normalization (and the transmitted signals between CNN layers) using the batch normalization method [57].
A batch normalization algorithm normalizes the input training patterns and successive signal vectors propagated between successive ANN layers [57]. The normalization introduced by the discussed algorithm reduces or even eliminates possible oscillations of the ANN error minimization process and possibility to become stuck in the local minimum, resulting from the fact that small changes in the weight vector in the layers preceding subsequent layers may cause large changes in the weight vectors in subsequent layers (the deeper has an ANN structure), the so-called exploding gradient [67] or too-small changes to the weight vector value that will cause the weight vector gradient to fade away on subsequent layers, the so-called vanishing gradient [67].
In addition to the abovementioned stabilizing properties of the learning process, the batch normalization algorithm normalizes (providing the mean value and variance of the training pack: 0 and 1, respectively) in such a way that the training dataset that the range of values taken by ANN is approximately not variable, which is another aspect that reduces the probability of the occurrence of oscillations of objective function between successive epochs during the learning process.
The normalized input vectors for the successive ANN layers are calculated in the following way:
μ B = 1 m   i = 1 m x i ,
σ B 2 = 1 m   i = 1 m ( x i μ B ) 2 ,
x i ^ = x i μ B σ B 2 + ϵ ,
y i = γ x ^ + β = B N γ ,   β ( x i ) ,
where B is the number of the current training packet (mini-batch [57] number), m is the number of elements in the packet, μ B   is the average value of the training packet, σ B 2 is the variance of the teaching packet, x i ^ is the normalized input training vector, and y i   is the normalized and scaled training vector. Algorithm 1 presents the exemplary of batch normalization procedure for an input vector X.
Algorithm 1 Batch normalization procedure
Input: X[N] ,
Input vector
    N,
Size of X
     γ , β ,
Parameters to be learned
     ϵ
Very small value to stabilize calculation
Output: Y [ N ]
Normalized vector X
1: initialize: i = 0 , μ B = 0 ,     σ B 2 = 0 ,   s = 0 , x = 0
2: for  i = 0  in  N 1  do
Calculate mean of X
3:   μ B = μ B + X [ i ]
4: end for
5: μ B = μ B / N
6: for  i = 0  in  N 1  do
Calculate variance of X
7:   s = X [ i ] μ B
8:   σ B 2 = σ B 2 + s s
9: end for
10: σ B 2 = σ B 2 / N
11: for  i = 0  in  N 1  do
Normalize X vector as Y vector
12:   x = X [ i ] μ B σ B 2 + ϵ
13:   Y [ i ] = γ x + β
14: end for
In the first step, the normalizing algorithm calculates the mean value μ B and the variance σ B 2 of the presented training set (Equations (7) and (8)); then, based on these values, it calculates the normalized and scaled input vectors y i of the training set (Equations (9) and (10)), [57]. The operation described above is used not only for the input layer, but for each successive output layer in relation to the next input layer (Figure 10) of the entire CNN structure.

5. Simulation Environment

In order to test the designed structures of neural networks, a simulator was created using C++ [68,69] language, with the FFTW [70] library for signal processing and the OpenCL [71] library to speed up the process of generating signals. The simulator enables one to generate many digital signals, introducing interferences and noises into them. The operation of the simulator is based on the generation of signals on a one-time base (vector of signal samples of length N), which consists of N samples, and all available signal classes with variable parameters PD, PRI, RF and SP. These signals parameters vary from generation to generation of successive signals in the given class. The simulator works in a quasi-real mode and allows one to view the currently generated signals, the modulus of their spectrum in the frequency domain, and the vector of samples of the signal space (Figure 11).
In the first step, the simulator reads the signal parameters for the given radar devices from the configuration file. Then it should check whether the given hardware configuration allows one to process signals with the maximum sampling frequency obtained from the configuration file. If the system‘s maximum sampling rate is less for random access memory (RAM) than the highest-frequency signal of all signal classes, then the signals (their waveform form) are scaled to the system’s acceptable sampling rate. For faster learning and processing purposes, the scaling rate was chosen manually. After calculating these scaling factors (or setting them manually), the simulator creates the filters with a finite impulse response (FIR) [72,73] for each radar signal class. Filter coefficients for individual classes were calculated based on the Hann‘s time window [74] using the formulas presented in Equations (11)–(13).
c [ i ] = { sinc ( x [ i ] ) d l a   x > 0 1 d l a   x 0 ,     s i n c ( x ) = sin ( x π ) x π ,    
x   ϵ { l B + s · 0 , l B + s · 1 , , l B + s · ( N 1 ) } ,
h [ i ] = c [ i ] · ( 1 c o s ( 2 π i N ) ) 2 ,
where N is the number of FIR filter coefficients, x is the vector of discrete argument values for which the FIR filter coefficients are calculated, s is the discrete shift value between the sample x [ i ] and x [ i + 1 ] , c is the vector of coefficients calculated according to the s i n c ( x ) relationship, h is the vector of coefficients using the relationship for the Hann window,   i is the successive index of the FIR filter coefficient and i ( 0 ,   N 1 ) , l B means the cut-off value of the FIR filter corresponding to the upper and lower cut-off frequency. The block diagram of simulation and recognition radar signals is presented in Figure 12.
After the FIR filters are created, the main simulation loop follows. The operation of the simulation loop consists of the following four steps described below.
The first step is the digital generation of signals on the one vector of the output signals S of real numbers with the number of elements N = F m a x (number of samples). A single-class signal is added to the S vector at intervals depending on the PRI. Initially, the signal vector contains approximately K = 1 P R I [ s ] of the waveforms of a signal for the given class, where the exact number depends on the drawing of individual values of PRI and PD in the successive signal generations for the given class. The values of PRI, PD and RF at each successive signal generation are randomly selected in accordance with the uniform distribution or PRI, with PD inter-pulse modulation in intervals characterizing the allowable range of signal parameter changes for a given class of signals. Below, Algorithms 1–3 in the form of a pseudocode are presented, describing the process of adding subsequent signals to the output vector of signals S .
Algorithm 2 Add signals to vector space (Random PRI, PD, RF Modulation)
Input: L   = 18
number of signal classes
    N
number of samples for output vector of signals
    Cs [ L ]
L-Length vector of signals structures parameters
    ud ( minValue ,   maxValue )
function to create uniform range distribution for random engine
    rg
random engine generator
Output: S [ N ]
N-Length output vector of signals
1: initialize: i = 0
          sw  
vector of i-th time waveform signal
2: for  i = 0  in (L − 1) do
3:   cl = Cs [ i ]
get parameters of i-th class of signal
4:   priRange = ud ( cl . pri _ min ,   cl . pri _ max )
create range distribution for PRI parameters
5:   pdRange =   ud ( cl . pd _ min ,   cl . pd _ max )
create range distribution for PD parameters
6:   rfRange   =   ud ( cl . rf _ min ,   cl . rf _ max )
create range distribution for RF parameters
7:   shift = 0
8:   msl   =   cl . pd _ max   N
calculate max length of current class signal
9:  while ( ( shift + msl ) < N )  do
10:    cPRI =   priRange ( rg )
get random value from PRI range of cl-signal
11:    cPD =   pdRange ( rg )
get random value from PD range of cl-signal
12:    cRF =   rfRange ( rg )
get random value from RF range of cl-signal
13:    cWL = currentPD N
get length of time waveform (TW) signal
14:    sw = generate ( cPD ,   cRF )
generate signal with PD and RF
15:    AS 2 VS ( S ,   sw ,   cWL )
Add signal to vector space, Algorithm 3
16:    shift = shift + cPRI N
shift vector signal about shift value at output vector signal
17:  end while
18: end for
Algorithm 3 Add signal to vector space (random PRI modulation)
Input: S [ N ]
N-Length output vector of signals
    signalWaveform [ currentWaveformLength ]
number of samples for output vector of signals
    currentWaveformLength
length of generated signal
   shift
shift at output signal vector space
Output: S [ N ]
N-Length output vector of signals
1: initialize: i = 0
2: for  i = 0  in (currentWaveformLength − 1) do
3:   S [ i + shift ] = S [ i + shift ] + signalWaveform [ i ]
Adding signal to S vector space
11: end for
The second step is to introduce the interference and noise to the vector of the output signals. The noise introduced is additive. It is worth mentioning that the signals from different classes added to the same sample vector disturb each other by interfering with each other.
The third simulation step is the filtering step of the entire resulting vector from the output signals S , each of the 18 filters, and each of them is assigned to a given class signal. The block diagram of the output vector filter for each class of signals is shown in Algorithm 4.
Algorithm 4 Filter all signals
Input:  S [ N ]
N-Length output vector of signals
    L   = 18
number of signal classes
    filtersFIR [ L ]
L-Length vector of FIR filters
1: initialize: i = 0
       filteredSignalVector [ N ]
N-Length output vector of i-th signal after filtration
2: for  i = 0  in (L − 1) do
3:   firFilter   = filtersFIR [ i ]
get i-th FIR filter
4:   outputSignalVector   =   firFilter . filter ( S )
try to filter i-th signal from S[N]
5:   DNN _ isClassSignal ( outputSignalVector )
use DNN model to classify filtered signal
11: end for
As can be deducted from the above-presented algorithms in the form of pseudocode, an output vector is given as an argument for each of the 18th FIR filters. The filtering function of the firFilter instance of FIR class returns the signal vector filtered against the i-th class, which is then sent to the CNN input, and an attempt is made to detect the given signal class. DNN_isClassSignal is a function which analyzes the occurrence of a given signal class and returns the probability value in the situation where the recognized signal is located in the given input vector.

6. Experiment Results

The calculation results of the effectiveness of radar signal recognition based on the post-detection PRI, PD determination and the sampled time form of the signal are presented in Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12.
Where the disruption level means the amplitude level of the interfering signals.
The characteristics of changes in the probability of the correct recognition for 18 signal classes are presented in Figure 13, Figure 14 and Figure 15 on the basis of a separate analysis of the PRI, PD parameters and TW vectors and depending on the changes in the number of iterations of the training algorithm and the size of the training dataset.
The characteristic in Figure 13 indicates that teaching the CNN to recognize the correct class on the basis of PD parameter alone is unlikely: about 10% of the achieved effectiveness with a training dataset of 40–70 elements. This is due to large overlapping of the signal operation intervals for the PD parameter.
In the case of CNN analysis for the PRI parameter (Figure 14) in relation to the PD analysis (Figure 13), the results achieved are much better: approximately 70% of the effectiveness of the training package with the size of 20 elements and 60 epochs in the learning process. With larger packages at the level above 70, the quality of learning and CNN performance showed a descending trend.
The analysis of the results of the CNN recognition of TW samples (Figure 15) in terms of effectiveness was similar to the effectiveness of CNN in the PRI samples analysis. However, here the downward trend along with the growth of the learning dataset is much more noticeable. The optimal learning point turned out to be a package of 20 learning patterns and 140 epochs of the learning process. The combination of three CNN structures with the concatenation operation and an additional classifier with dense connections (Figure 7) which analyzes the combined tensor allowed one to achieve an efficiency of radar signal classes recognition at the level of 90–99% (Figure 16).
An important parameter that influences the qualitative operation of the structure (Figure 7, Table 4) is the size of the input vector accepted by the CNN. In this case, the size of CNN inputs, which analyzed the PRI and PD parameters, changed at the same time. The characteristics clearly indicate that the possibility of accepting a larger vector of samples (longer observation time) improved the efficiency of signal classification. The size of the package also turned out to be important. In this case, it should be at the level of about 350 elements for the CNN to achieve high effectiveness in the learning process.
The last examined structure (Table 5, Figure 8) contains efficiency of classification when the raw unprocessed sampled signal is considered. The characteristics of CNN operation in the case of disturbances are presented in Figure 17. The results of CNN operation without interference were at the level of 99% probability of correct recognition, therefore the tables and graphs with the operation of the CNN without interference were omitted.
The blue curve in Figure 17 shows the effectiveness of the CNN in the case of recognizing radar signals when there are a certain number of radio sources at different distances around the reconnaissance station. The noise levels (horizontal line) illustrate the signal amplitude level of interfering sources. The interfering signals, in this case, were signals for the 18 classes of radar signals examined earlier. To simulate the disturbance effect, the generated signals for the 18 classes to be recognized were superimposed on the remaining signals with a given amplitude (0.1–0.9). The red curve shows the disturbance noise.
A CNN working on the raw signals showed a high resistance to the noise of interference. However, it started to cope worse in the case of interference with other signals where the probability of correct recognition tended to decline along with the increase in the amplitude of the disturbing signals. To overcome this problem, the observation time (the number of samples received by the CNN at the input) was tested and increased, which resulted in an effectiveness equal to 92.2% (Table 11).

7. Conclusions

On the basis of the obtained results of the effectiveness measurements, it is possible to note the high effectiveness of the CNN in the process of recognizing digital signals when they are analyzed post-detection based on the PRI, and PD parameters and the form of the time waveform (RF and PD), or when we analyze the raw signal (not processed sampled signal). Simultaneous analysis of the set of three signal parameters was possible thanks to the concatenation operation of the three CNN network models (PD, PRI, TW), and the analysis of the response tensor obtained by the ANN with the dense full-connected architecture.
The achieved probabilities of correct signal recognition were high, ranging from 90–99%. However, in order to achieve such network efficiency in the case of post-detection analysis, it is required that the CNNs analyze more than two parameters of the radar signal. Otherwise, if each signal parameter is analyzed separately, then the radars’ signals cannot be properly classified. This is due to the overlapping of the ranges of the operating parameters of the classes of individual signals (Table 2), and was confirmed by the results of the signal recognition efficiency by the CNN (Table 7, Table 8 and Table 9 and Figure 13, Figure 14 and Figure 15), where the analysis of the parameters of each separately allowed one to achieve the maximum CNN values of the appropriate probability diagnosis at the level of 70–72% for PRI and TW, and a maximum of 11–16% for the PD parameter.
The CNN designed has the ability to classify signals on the basis of the analysis of raw data, also in the presence of interference at a 92–99% probability of correct recognition (Table 11 and Table 12). In the case of working with interference, the effectiveness of our CNN largely depends on the capabilities and sensitivity of the receiver, i.e., the ability to process signals such as a sampling frequency. Parameters such as package size and number of iterations (epochs) of the training algorithm were important for convergence by a particular CNN, and depend on both the architecture of the given CNN and the training dataset (Table 7, Table 8, Table 9 and Table 10) and (Figure 13, Figure 14, Figure 15 and Figure 16).
Increasing the size of the CNN input vector, which is basically the observation time or sampling rate, significantly improved the performance of the combined CNN (Table 4, Figure 7) and the CNN analyzing the raw signal (Table 5, Figure 8), especially in the event of disturbances (Table 11).
The results of the operation presented here were achieved in a relatively short time; about 2 h for a single learning process for post-detection analyzing networks, and about 24 h for a network working on the raw signals thanks to the scaling of high-frequency signals to lower frequencies. In the case of direct operation at high frequencies, changing the size of the CNN inputs and selecting an appropriate number of convolutional filters will be required. The learning time available and having the appropriate hardware architecture to carry out such a learning process should also be taken into account.

Author Contributions

Conceptualization, J.M. and D.P.; methodology, D.P.; software, D.P.; validation, J.M.; formal analysis, J.M. and D.P.; investigation, D.P. and D.P.; resources, J.M.; data curation, J.M.; writing—original draft preparation, J.M. and D.P.; writing—review and editing, J.M. and D.P.; visualization, D.P.; supervision, J.M.; project administration, J.M. and D.P.; funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financed/co-financed by the Military University of Technology under Research Project UGB 22-856.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Adamy, D.L. EW 102. A Second Course in Electronic Warfare; Horizon House Publications: London, UK, 2004. [Google Scholar]
  2. Vakin, S.A.; Shustov, L.N.; Dunwell, R.H. Fundamentals of Electronic Warfare; Artech House: Boston, MA, USA, 2001. [Google Scholar]
  3. Willey, R.G. ELINT: The Interception and Analysis of Radar Signals; Horizon House Publications: London, UK, 2006. [Google Scholar]
  4. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, G.; et al. TensorFlow. A system for large-scale machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; USENIX Association: Berkeley, CA, USA, 2016; pp. 265–283. [Google Scholar]
  5. Anitha, U.; Malarkkan, S. Underwater object identification and recognition with sonar images using soft computing techniques. Indian J. Geo Mar. Sci. 2018, 47, 665–673. [Google Scholar]
  6. Ding, J.; Liu, Z.C. A study of radar emitter recognition based on characteristic parameter matching method. Mod. Radar 2011, 33, 29–33. [Google Scholar]
  7. Kirichenko, L.; Radivilova, T.; Bulakh, V. Machine learning in classification time series with fractal properties. Data 2019, 4, 5. [Google Scholar] [CrossRef] [Green Version]
  8. Carter, C.A.P.; Masse, N. Neural Networks for Classification of Radar Signals; Defence Research Establishment: Ottawa, ON, Canada, 1993; Technical Note 93–33. [Google Scholar]
  9. Matuszewski, J.; Pietrow, D. Recognition of electromagnetic sources with the use of deep neural networks. In Proceedings of the XII Conference on Reconnaissance and Electronic Warfare Systems, Ołtarzew, Poland, 19–21 November 2018; Volume 11055, pp. 100–114. [Google Scholar] [CrossRef]
  10. Zhou, Z.; Huang, G.; Chen, H.; Gao, J. Automatic radar waveform recognition based on deep convolutional denoising auto-encoders. Circuits Syst. Signal Process. 2018, 37, 4034–4048. [Google Scholar] [CrossRef]
  11. Kechagias-Stamatis, O.; Aouf, N.; Belloni, C. SAR automatic target recognition based on convolutional neural networks. In Proceedings of the International Conference on Radar Systems (Radar 2017), Belfast, Ireland, 23–26 October 2017; pp. 1–4. [Google Scholar] [CrossRef]
  12. Zhang, T.; Zhang, X.; Ke, X.; Liu, C.; Xu, X.; Zhan, X.; Wang, C.; Ahmad, I.; Zhou, Y.; Pan, D.; et al. HOG-ShipCLSNet: A novel deep learning network with hog feature fusion for SAR ship classification. IEEE Trans. Geosci. Remote Sens. 2021, 1–22. [Google Scholar] [CrossRef]
  13. Zhang, T.; Zhang, X. A polarization fusion network with geometric feature embedding for SAR ship classification. Pattern Recognit. 2021, 123, 108365. [Google Scholar] [CrossRef]
  14. Nishizaki, H.; Makino, K. Signal classification using deep learning. In Proceedings of the 2019 IEEE International Conference on Sensors and Nanotechnology, Penang, Malaysia, 24–25 July 2019; pp. 1–4. [Google Scholar] [CrossRef]
  15. Zhang, G.; Rong, H.; Jin, W.; Hu, L. Radar emitter signal recognition based on resemblance coefficient features. In Rough Sets & Current Trends in Computing, International Conference; Rsctc: Uppsala, Sweden, 2014. [Google Scholar]
  16. Wu, X.; Shi, Y.; Meng, W.; Ma, X.; Fang, N. Specific emitter identification for satellite communication using probabilistic neural networks. Int. J. Satell. Commun. Netw. 2019, 37, 283–291. [Google Scholar] [CrossRef]
  17. Chen, H.; Zhang, F.; Tang, B.; Yin, Q.; Sun, X. Slim and efficient neural network design for resource-constrained SAR target recognition. Remote Sens. 2018, 10, 1618. [Google Scholar] [CrossRef] [Green Version]
  18. Shieh, C.S.; Lin, C.T. A vector neural network for emitter identification. IEEE Trans. Antennas Propag. 2002, 50, 1120–1127. [Google Scholar] [CrossRef] [Green Version]
  19. Pietkiewicz, T.; Sikorska-Łukasiewicz, K. Comparison of two classifiers based on neural networks and the DTW method of comparing time series to recognize maritime objects upon FLIR images. In Proceedings of the XII Conference on Reconnaissance and Electronic Warfare Systems, Ołtarzew, Poland, 19–21 November 2018; Volumn 110550V. [Google Scholar] [CrossRef]
  20. Tachibana, K.; Otsuka, K. Wind prediction performance of complex neural network with ReLU activation function. In Proceedings of the 57th Annual Conference of the Society of Instrument and Control Engineers of Japan (SICE), Nara, Japan, 11–14 September 2018; pp. 1029–1034. [Google Scholar] [CrossRef]
  21. Yun, L.; Jing-Chao, L. Radar signal recognition algorithms based on neural network and grey relation theory. In Proceedings of the 2011 Cross Strait Quad-Regional Radio Science and Wireless Technology Conference, Harbin, China, 26–30 July 2011; Volume 2, pp. 1482–1485. [Google Scholar] [CrossRef]
  22. Zhang, M.; Diao, M.; Gao, L.; Liu, L. Neural networks for radar waveform recognition. Symmetry 2017, 9, 75. [Google Scholar] [CrossRef] [Green Version]
  23. Mussab Elamien Abd Elmaggeed, A.J.A. Neural network algorithm for radar signal recognition. J. Eng. Res. Appl. 2015, 5, 123–125. [Google Scholar]
  24. Petrov, N.; Jordanov, I.; Roe, J. Radar emitter signals recognition and classification with feedforward networks. In Proceedings of the 17th International Conference in Knowledge Based and Intelligent Information and Engineering Systems—KES2013, Kitakyushu, Japan, 9–11 September 2013; Volume 22, pp. 1192–1200. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, J.; Zheng, T.; Lei, P.; Bai, X. A hierarchical convolution neural network (CNN)-based ship target detection method in spaceborne SAR imagery. Remote Sens. 2019, 11, 620. [Google Scholar] [CrossRef] [Green Version]
  26. Chi, Z.; Li, Y.; Chen, C. Deep convolutional neural network combined with concatenated spectrogram for environmental sound classification. In Proceedings of the 2019 IEEE 7th International Conference on Computer Science and Network Technology (ICCSNT), Dalian, China, 19–20 October 2019; pp. 251–254. [Google Scholar] [CrossRef]
  27. Kong, M.; Zhang, J.; Liu, W.; Zhang, G. Radar emitter identification based on deep convolutional neural network. In Proceedings of the 2018 International Conference on Control, Automation and Information Sciences (ICCAIS), Hangzhou, China, 24–27 October 2018; pp. 309–314. [Google Scholar] [CrossRef]
  28. Zhang, T.; Zhang, X. Squeeze-and-excitation Laplacian pyramid network with dual-polarization feature fusion for ship classification in sar images. IEEE Geosci. Remote Sens. Lett. 2021, 751–755. [Google Scholar] [CrossRef]
  29. Cruz-López, J.A.; Boyer, V.; El-Baz, D. Training many neural networks in parallel via back-propagation. In Proceedings of the 2017 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), Lake Buena Vista, FL, USA, 29 May–2 June 2017; pp. 501–509. [Google Scholar] [CrossRef] [Green Version]
  30. Aida-Zade, K.; Mustafayev, E.; Rustamov, S. Comparison of deep learning in neural networks on CPU and GPU-based frameworks. In Proceedings of the 2017 IEEE 11th International Conference on Application of Information and Communication Technologies (AICT), Moscow, Russia, 20–22 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
  31. Furukawa, M.; Itsubo, T.; Matsutani, H. An in-network parameter aggregation using DPDK for multi-GPU deep learning. In Proceedings of the 2020 Eighth International Symposium on Computing and Networking (CANDAR), Naha, Japan, 24–27 November 2020; pp. 108–114. [Google Scholar] [CrossRef]
  32. Hoskote, Y.; Vangal, S.; Dighe, S.; Borkar, N.; Borkar, S. Teraflops prototype processor with 80 cores. In Proceedings of the 2007 IEEE Hot Chips 19 Symposium (HCS), Stanford, CA, USA, 19–21 August 2007; pp. 1–15. [Google Scholar] [CrossRef]
  33. Clark, D. Breaking the teraFLOPS barrier. Computer 1997, 30, 12–14. [Google Scholar] [CrossRef]
  34. Dhabe, P.; Vyas, P.; Ganeriwal, D.; Pathak, A. Pattern classification using updated fuzzy hyper-line segment neural network and it’s GPU parallel implementation for large datasets using CUDA. In Proceedings of the 2016 International Conference on Computing, Analytics and Security Trends (CAST), Pune, India, 19–21 December 2016; pp. 24–29. [Google Scholar] [CrossRef]
  35. Dumin, O.; Prishchenko, O.; Pochanin, G.; Plakhtii, V.; Shyrokorad, D. Subsurface object identification by artificial neural networks and impulse radiolocation. In Proceedings of the 2018 IEEE Second International Conference on Data Stream Mining Processing (DSMP), Lviv, Ukraine, 21–25 August 2018; pp. 434–437. [Google Scholar] [CrossRef]
  36. Jurado-Lucena, A.; Montiel-Sanchez, I.; Escot Bocanegra, D.; Fernandez-Recio, R.; Poyatos Martinez, D. Class identification of aircrafts by means of artificial neural networks trained with simulated radar signatures. Prog. Electromagn. Res. 2011, 21, 243–255. [Google Scholar] [CrossRef]
  37. Li, P.; Chen, M.; Hu, F.; Xu, Y. A spectrogram-based voiceprint recognition using deep neural network. In Proceedings of the 27th Chinese Control and Decision Conference (2015 CCDC), Qingdao, China, 23–25 May 2015; pp. 2923–2927. [Google Scholar] [CrossRef]
  38. ŞEN, S.Y.; ÖZKURT, N. ECG arrhythmia classification by using convolutional neural network and spectrogram. In Proceedings of the 2019 Innovations in Intelligent Systems and Applications Conference (ASYU), Izmir, Turkey, 31 October–2 November 2019; pp. 1–6. [Google Scholar] [CrossRef]
  39. Kılıç, A.; Babaoğlu, S.; Babalik, A.; Arslan, A. Through-wall radar classification of human posture using convolutional neural networks. Int. J. Antennas Propag. 2019, 2019, 1–10. [Google Scholar] [CrossRef]
  40. Yonel, B.; Mason, E.; Yazıcı, B. Deep learning for passive synthetic aperture radar. IEEE J. Sel. Top. Signal Process. 2018, 12, 90–103. [Google Scholar] [CrossRef] [Green Version]
  41. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Matuszewski, J.; Pietrow, D. Artificial neural networks in the filtration of radiolocation information. In Proceedings of the 2020 IEEE 15th International Conference on Advanced Trends in Radioelectronics, Telecommunications and Computer Engineering (TCSET), Lviv-Slavske, Ukraine, 25–29 February 2020; pp. 680–685. [Google Scholar] [CrossRef]
  43. Wang, L.; Tang, J.; Liao, Q. A study on radar target detection based on deep neural networks. IEEE Sens. Lett. 2019, 3, 1–4. [Google Scholar] [CrossRef]
  44. Nalini, M.K.; Radhika, K.R. Comparative analysis of deep network models through transfer learning. In Proceedings of the 2020 Fourth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), Palladam, India, 7–9 October 2020; pp. 1007–1012. [Google Scholar] [CrossRef]
  45. Yang, X.; Denis, L.; Tupin, F.; Yang, W. SAR image despeckling using pre-trained convolutional neural network models. In Proceedings of the 2019 Joint Urban Remote Sensing Event (JURSE), Vannes, France, 22–24 May 2019; pp. 1–4. [Google Scholar] [CrossRef]
  46. Lee, D.Y.; Jeong, J.H.; Shim, K.H.; Lee, S.W. Decoding movement imagination and execution from eeg signals using bci-transfer learning method based on relation network. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1354–1358. [Google Scholar] [CrossRef] [Green Version]
  47. Zhang, T.; Chen, S.; Wei, S.; Chen, J. A data-efficient training model for signal integrity analysis based on transfer learning. In Proceedings of the 2019 IEEE Asia Pacific Conference on Circuits and Systems (APCCAS), Bangkok, Thailand, 11–14 November 2019; pp. 186–189. [Google Scholar] [CrossRef]
  48. Darzikolaei, M.A.; Ebrahimzade, A.; Gholami, E. Classification of radar clutters with artificial neural network. In Proceedings of the 2015 2nd International Conference on Knowledge-Based Engineering and Innovation (KBEI), Tehran, Iran, 5–6 November 2015; pp. 577–581. [Google Scholar] [CrossRef]
  49. Matuszewski, J.; Pietrow, D. Deep learning neural networks with using pseudo-random 3D patterns generator. In Proceedings of the 2019 IEEE International Scientific-Practical Conference Problems of Infocommunications, Science and Technology (PIC S&T), Kiev, Ukraine, 8–11 October 2019; pp. 262–267. [Google Scholar] [CrossRef]
  50. Du, X.; Liao, K.; Shen, X. Secondary radar signal processing based on deep residual separable neural network. In Proceedings of the 2020 IEEE International Conference on Power, Intelligent Computing and Systems (ICPICS), Shenyang, China, 28–30 July 2020; pp. 12–16. [Google Scholar] [CrossRef]
  51. Hannun, A.; Case, C.; Casper, J.; Catanzaro, B.; Diamos, G.; Elsen, E.; Prenger, R.; Satheesh, S.; Sengupta, S.; Coates, A.; et al. Deep Speech: Scaling up end-to-end speech recognition. arXiv 2014, arXiv:1412.5567. [Google Scholar]
  52. Hwang, M.-J.; Soong, F.; Song, E.; Wang, X.; Kang, H.; Kang, H.G. LP-wavenet: Linear prediction-based wavenet speech synthesis. In Proceedings of the 2020 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Auckland, New Zealand, 7–10 December 2020; pp. 810–814. [Google Scholar]
  53. Matuszewski, J.; Sikorska-Łukasiewicz, K. Neural network application for emitter identification. In Proceedings of the 2017 18th International Radar Symposium (IRS), Prague, Czech Republic, 28–30 June 2017; pp. 1–8. [Google Scholar] [CrossRef]
  54. Matuszewski, J. Identification of radar signals using discriminant vectors. In Proceedings of the MIKON 2008–17th International Conference on Microwaves, Radar and Wireless Communications, Wroclaw, Poland, 19–21 May 2008; pp. 433–436. [Google Scholar]
  55. Lin, W.; Hasenstab, K.; Moura Cunha, G.; Schwartzman, A. Comparison of handcrafted features and convolutional neural networks for liver MR image adequacy assessment. Sci. Rep. 2020, 10, 20336. [Google Scholar] [CrossRef] [PubMed]
  56. O’Mahony, N.; Campbell, S.; Carvalho, A.; Harapanahalli, S.; Hernandez, G.V.; Krpalkova, L.; Riordan, D.; Walsh, J. Deep Learning vs. Traditional Computer Vision; ImaR Technology Gateway, Institute of Technology Tralee: Trale, Ireland, 2019. [Google Scholar]
  57. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167v3. [Google Scholar]
  58. Giusti, A.; Cireşan, D.C.; Masci, J.; Gambardella, L.M.; Schmidhuber, J. Fast image scanning with deep max-pooling convolutional neural networks. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 4034–4038. [Google Scholar] [CrossRef] [Green Version]
  59. Özgür, A.; Nar, F. Effect of Dropout layer on Classical Regression Problems. In Proceedings of the 2020 28th Signal Processing and Communications Applications Conference (SIU), Gaziantep, Turkey, 5–7 October 2020; pp. 1–4. [Google Scholar] [CrossRef]
  60. Asgher, U.; Muhammad, H.; Hamza, M.M.; Ahmad, R.; Butt, S.I.; Jamil, M. Robust hybrid normalized convolution and forward error correction in image reconstruction. In Proceedings of the 2014 10th International Conference on Innovations in Information Technology (IIT), Abu Dhabi, United Arab Emirates, 9–11 November 2014; pp. 54–59. [Google Scholar] [CrossRef]
  61. Zhang, L.; Borggreve, D.; Vanselow, F.; Brederlow, R. Quantization considerations of dense layers in convolutional neural Networks for resistive crossbar implementation. In Proceedings of the 2020 9th International Conference on Modern Circuits and Systems Technologies (MOCAST), Bremen, Germany, 7–9 September 2020; pp. 1–6. [Google Scholar] [CrossRef]
  62. Marcu, D.C.; Grava, C. The impact of activation functions on training and performance of a deep neural network. In Proceedings of the 16th International Conference on Engineering of Modern Electric Systems (EMES), Oradea, Romania, 10–11 June 2021; pp. 1–4. [Google Scholar] [CrossRef]
  63. Kim, J.-C.; Chung, K. Hybrid multi-modal deep learning using collaborative concat layer in health bigdata. IEEE Access 2020, 8, 192469–192480. [Google Scholar] [CrossRef]
  64. Hong, S.; Song, S. Kick: Shift-N-Overlap cascades of transposed convolutional layer for better autoencoding reconstruction on remote sensing imagery. IEEE Access 2020, 8, 107244–107259. [Google Scholar] [CrossRef]
  65. Zhang, Z. Improved adam optimizer for deep neural networks. In Proceedings of the 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS), Banff, AB, Canada, 4–6 June 2018; pp. 1–2. [Google Scholar] [CrossRef]
  66. Tseng, C.; Lee, S. Design of digital differentiator using supervised learning on keras framework. In Proceedings of the 2019 IEEE 8th Global Conference on Consumer Electronics (GCCE), Osaka, Japan, 15–18 October 2019; pp. 162–163. [Google Scholar] [CrossRef]
  67. Al-Abri, S.; Lin, T.X.; Tao, M.; Zhang, F. A derivative-free optimization method with application to functions with exploding and vanishing gradients. IEEE Control Syst. Lett. 2021, 5, 587–592. [Google Scholar] [CrossRef]
  68. Masters, T. Practical Neural Network Recipes in C++; Academic Press: London, UK, 1993; ISBN 0124790402, 9780124790407. [Google Scholar]
  69. Stroustrup, B. The C++ Programming Language, 4th ed.; Addison-Wesley Professional: Boston, MA, USA, 2013. [Google Scholar]
  70. Bantle, M.; Bayerl, P.; Funken, S.; Thoma, M. Radar signal processing with openCL on integrated graphic processors. In Proceedings of the 2018 19th International Radar Symposium (IRS), Bonn, Germany, 20–22 June 2018; pp. 1–8. [Google Scholar] [CrossRef]
  71. Frigo, M.; Johnson, S.G. The design and implementation of FFTW3. Proc. IEEE 2005, 93, 216–231. [Google Scholar] [CrossRef] [Green Version]
  72. Visweswar, C.; Mitra, S.K. New linearly tapered window and its application to FIR filter design. In Proceedings of the 2005 European Conference on Circuit Theory and Design, Cork, Ireland, 29 August–1 September 2005; Volume 3, pp. III/445–III/448. [Google Scholar] [CrossRef]
  73. Ichige, K.; Iwaki, M.; Ishii, R. Accurate estimation of minimum filter length for optimum FIR digital filters. IEEE Trans. Circuits Syst. II Analog. Digit. Signal Process. 2000, 47, 1008–1016. [Google Scholar] [CrossRef]
  74. Shimauchi, S.; Ohmuro, H. Accurate adaptive filtering in square-root Hann windowed short-time Fourier transform domain. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 1305–1309. [Google Scholar] [CrossRef]
Figure 1. General structure of ESM system for measurement, recording, analysis and radar signals recognition [53].
Figure 1. General structure of ESM system for measurement, recording, analysis and radar signals recognition [53].
Sensors 21 08237 g001
Figure 2. The confidence intervals of 18 classes of radars: (a) for parameter PRI; (b) for parameter PD.
Figure 2. The confidence intervals of 18 classes of radars: (a) for parameter PRI; (b) for parameter PD.
Sensors 21 08237 g002
Figure 3. The confidence intervals of 18 classes of radars: (a) for parameter RF; (b) for parameter SP.
Figure 3. The confidence intervals of 18 classes of radars: (a) for parameter RF; (b) for parameter SP.
Sensors 21 08237 g003
Figure 4. Examples of generated waveforms for the parameters PD and RF for 18 classes.
Figure 4. Examples of generated waveforms for the parameters PD and RF for 18 classes.
Sensors 21 08237 g004
Figure 5. Illustration of training signals with inter-pulse modulation of PRI.
Figure 5. Illustration of training signals with inter-pulse modulation of PRI.
Sensors 21 08237 g005
Figure 6. Structure of the CNN used for radar signal recognition, where OT is the output tensor, SM is the size map or number output of the dense layer, and NM is the number of maps.
Figure 6. Structure of the CNN used for radar signal recognition, where OT is the output tensor, SM is the size map or number output of the dense layer, and NM is the number of maps.
Sensors 21 08237 g006
Figure 7. Example of obtained weight maps applied in the designed CNN.
Figure 7. Example of obtained weight maps applied in the designed CNN.
Sensors 21 08237 g007
Figure 8. CNN structure for radar signals recognition based on the simultaneous analysis of PD, PRI and TW (RF + PD) parameters.
Figure 8. CNN structure for radar signals recognition based on the simultaneous analysis of PD, PRI and TW (RF + PD) parameters.
Sensors 21 08237 g008
Figure 9. Structure of the CNN for radar signals recognition based on the analysis of the raw signal (raw samples of radar signal parameters).
Figure 9. Structure of the CNN for radar signals recognition based on the analysis of the raw signal (raw samples of radar signal parameters).
Sensors 21 08237 g009
Figure 10. Scheme of normalizing the input vectors for successive layers using the batch normalization algorithm.
Figure 10. Scheme of normalizing the input vectors for successive layers using the batch normalization algorithm.
Sensors 21 08237 g010
Figure 11. Example of the generation of signals parameters.
Figure 11. Example of the generation of signals parameters.
Sensors 21 08237 g011
Figure 12. Diagram of the simulation and recognition of radar signals.
Figure 12. Diagram of the simulation and recognition of radar signals.
Sensors 21 08237 g012
Figure 13. Effectiveness of radar signals recognition on the basis of the PD parameter in the function of the changes in the size of the training dataset and the number of iterations in the training algorithm.
Figure 13. Effectiveness of radar signals recognition on the basis of the PD parameter in the function of the changes in the size of the training dataset and the number of iterations in the training algorithm.
Sensors 21 08237 g013
Figure 14. Effectiveness of radar signals recognition on the basis of the PRI parameter in the function of the changes in the size of the training dataset and the number of iterations in the training algorithm.
Figure 14. Effectiveness of radar signals recognition on the basis of the PRI parameter in the function of the changes in the size of the training dataset and the number of iterations in the training algorithm.
Sensors 21 08237 g014
Figure 15. Effectiveness of recognizing radar signals on the basis of the time waveform shape modulation during the changes of the size of the training dataset and the number of iterations in the training algorithm.
Figure 15. Effectiveness of recognizing radar signals on the basis of the time waveform shape modulation during the changes of the size of the training dataset and the number of iterations in the training algorithm.
Sensors 21 08237 g015
Figure 16. Effectiveness of recognizing radar signals on the basis of simultaneous analysis of three parameters (PD, PRI and TW) depending on the size of the training set and the number of iterations in the training algorithm.
Figure 16. Effectiveness of recognizing radar signals on the basis of simultaneous analysis of three parameters (PD, PRI and TW) depending on the size of the training set and the number of iterations in the training algorithm.
Sensors 21 08237 g016
Figure 17. Effectiveness of signal recognition depending on the level of signal disruption.
Figure 17. Effectiveness of signal recognition depending on the level of signal disruption.
Sensors 21 08237 g017
Table 1. The confidence intervals of radar signal parameters for 18 classes examined.
Table 1. The confidence intervals of radar signal parameters for 18 classes examined.
Class NumberPRI
[ m s ]
PD
[ μ s ]
RF
[ GHz ]
SP
[s]
00.877–0.8780.929–1.7252.800–2.8323.97–4.00
11.229–1.2303.958–4.4921.255–1.3682.85–2.87
21.223–1.2232.512–2.8631.221–1.3395.80–5.96
31.223–1.2233.277–3.2771.228–1.3302.86–2.88
41.248–1.2502.512–3.0151.215–1.3512.86–2.87
51.247–1.2503.216–3.5711.248–1.3032.85–2.88
61.250–1.2523.727–3.8851.308–1.3652.87–2.88
71.751–1.7520.431–0.7053.144–3.1622.81–2.92
81.251–1.2522.018–2.3792.816–2.8425.04–5.08
90.768–0.7681.308–3.3842.832–2.8543.96–3.99
101.738–1.7392.811–3.5141.203–1.2546.02–6.07
111.775–1.7783.482–3.4821.220–1.2409.72–9.76
121.856–1.8581.727–4.5923.040–3.0926.03–6.09
131.905–1.9050.888–1.4662.219–2.2359.85–9.91
142.150–2.1504.898–5.5701.100–1.3895.41–5.53
152.225–2.2285.280–5.5291.180–1.2055.44–5.47
162.224–2.2264.138–4.9171.633–1.6505.43–5.59
172.375–2.3755.440–5.5481.171–1.1905.42–5.76
Table 2. Overlapping of individual signals in PRI, PD, RF and SP parameters.
Table 2. Overlapping of individual signals in PRI, PD, RF and SP parameters.
Number of Signal ClassNumber of Overlapping Signals in PRINumber of Overlapping Signals in PDNumber of Overlapping
Signals in the RF
Number of Overlapping Signals in the SP
0-9, 1389
1-12, 162, 3, 4, 5, 6, 143, 4, 5, 6, 7
234, 9, 10, 121, 3, 4, 5, 6, 10, 11, 14-
325, 9, 10, 121, 2, 4, 5, 6, 10, 11, 141, 4, 5, 6, 7
452, 9, 10, 121, 2, 3, 5, 6, 10, 11, 141, 3, 5, 6, 7
543, 9, 10, 11, 121, 2, 3, 4, 10, 141, 3, 4, 6, 7
68121, 2, 3, 4, 141, 3, 4, 5, 7
7---1, 3, 4, 5, 6
869, 120, 9-
9-0, 2, 3, 4, 5, 8, 10, 12, 1380
10-2, 3, 4, 5, 9, 11, 122, 3, 4, 5, 11, 14, 1512
11-5, 10, 122, 3, 4, 10, 14-
12-1, 2, 3, 4, 5, 6, 8, 9, 10, 11, 16-10
13-0, 9--
14-15, 16, 171, 2, 3, 4, 5, 6, 10, 11, 15, 1715, 16, 17
151614, 1710, 14, 1714, 16, 17
16151, 12, 14-14, 15, 17
17-14, 1514, 1514, 15, 16
Table 3. Detailed structure of designed CNN for separate signals recognition depending on the examined parameters (PRI, PD, TW).
Table 3. Detailed structure of designed CNN for separate signals recognition depending on the examined parameters (PRI, PD, TW).
Layer NumberLayer TypeLayer DimensionActivation Function
0B--
1R 1--
2C_1D 2[KS 3: 5, NK 4: 5, SS 5: 1]ReLU [20,62]
3B 6--
4MP_1D 7[SS: 2]-
5C_1D[KS: 3, NK: 9, SS: 1]ReLU
6B--
7MP_1D[SS: 2]-
8C_1D[KS: 2, NK: 6, SS: 1]ReLU
9B--
10MP_1D[SS: 2]-
11F--
12D 8-ReLU
13B--
14Dropout--
15D-Softmax [20]
16B--
1 Reshape layer [3], 2 Convolution 1D layer, 3 Kernel size, 4 Number of kernels, 5 Size stride, 6 Batch Normalization layer, 7 Max Pooling 1D layer, 8 Dense layer.
Table 4. Detailed structure of CNN designed for radar signal recognition with simultaneous input of three signal parameters (PRI, PD, TW).
Table 4. Detailed structure of CNN designed for radar signal recognition with simultaneous input of three signal parameters (PRI, PD, TW).
Input Vectors:
PD Samples
(Post-Processing)
PRI Samples
(Post-Processing)
TW Samples
(Raw Acquired Signal Samples)
Structure CNN for PD parameter 9Structure CNN for PRI parameterStructure CNN for TW parameter
Associated Outputs in CNN for PD, PRI and TW Parameters
Layer numberLayer typeDimensions of layerActivation function
0D-ReLU
1B--
2D-Softmax
3B--
9 For parameters PD, PRI, TW, the structure described in Table 3 was used.
Table 5. Detailed structure of designed CNN for recognizing sampled signals.
Table 5. Detailed structure of designed CNN for recognizing sampled signals.
Layer NumberLayer TypeDimensions of LayerActivation Function
0R--
1C_1D[KS: 5, NK: 30, SS: 1]ReLU
2B--
3MP_1D[SS: 2]-
4C_1D[KS: 4, NK: 30, SS: 2]ReLU
5B--
6MP_1D[SS: 2]-
7C_1D[KS: 3, NK: 30, SS: 3]ReLU
8B--
9MP_1D[SS: 2]-
10C_1D[KS: 4, NK: 30, SS: 2]ReLU
11B--
12C_1D[KS: 5, NK: 30, SS: 1]ReLU
13B--
14MP_1D[SS: 2]-
15F--
16D-Softmax
Table 6. The designed network parameters.
Table 6. The designed network parameters.
Processing
Network
Input
Tensor Size
Number of
Layers
Number of WeightsSize on DiskProcessing Time for a Single Tensor [s]GPU Processor
TW or PD
or PRI
(1, 128)17224863 kB0.015Geforce 1060 GTX 6GB
TW + PRI + PD(3, 128)3 × 17 (In parallel) + 3 (Output) = 542248 × 3 + 1332 = 8076(63 × 3 + 18) kB = 207 kB0.046
Table 7. Effectiveness of signal classification using CNN based on the PD vectors.
Table 7. Effectiveness of signal classification using CNN based on the PD vectors.
Input Size of ANN:512
Number of Samples 10:190Number of Tests 11:1900 η 12: 0.0001
Epochs 13 [B:S:E] 14Batch Size [B:S:E]Prediction 15 [Min–Max]
6010:10:900.056–0.111
7010:10:900.055–0.111
8010:10:900.056–0.056
9010:10:900.056–0.111
10010:10:900.056–0.056
11010:10:900.056–0.056
110:10:240900.056–0.167
10 number of learning patterns, 11 number of testing patterns, 12 learning factor, 13 number of executed learning iterations, 14 [B:S:E] is the changes of parameters (B is the begin value, S is the step value, E is the end value), 15 effectiveness of the correct recognition of the presented signals from one of the 18 classes.
Table 8. Effectiveness of signal classification using CNN based on the PRI vectors.
Table 8. Effectiveness of signal classification using CNN based on the PRI vectors.
Input Size of ANN:512
Number of Samples:190Number of Tests:1900 η : 0.0001
Number of EpochsBatch Size [B:S:E]Prediction [Min–Max]
6010:10:900.000–0.333
7010:10:900.056–0.500
8010:10:900.056–0.333
9010:10:900.056–0.556
10020:10:900.056–0.611
11020:10:900.056–0.778
12020:10:900.056–0.500
13020:10:900.111–0.722
14020:10:900.056–0.667
15020:10:900.111–0.556
Table 9. Effectiveness of signal classification using CNN based on the TW vectors.
Table 9. Effectiveness of signal classification using CNN based on the TW vectors.
Input Size of ANN:93
Number of Samples:190Number of Tests:1900 η : 0.0001
Number of Epochs
[B:S:E]
Batch Size
[B:S:E]
Prediction
[Min–Max]
6010:10:900.294–0.584
7010:10:900.334–0.538
8010:10:900.343–0.607
9010:10:900.392–0.589
10010:10:900.363–0.633
11010:10:900.447–0.664
12010:10:900.440–0.672
13010:10:900.474–0.667
14010:10:900.467–0.717
15010:10:900.463–0.652
Table 10. Effectiveness of signal recognition using the CNN on the basis of simultaneous analysis of 2 parameters PRI, PD and TW vectors.
Table 10. Effectiveness of signal recognition using the CNN on the basis of simultaneous analysis of 2 parameters PRI, PD and TW vectors.
Input Size of ANN (TW):93
Number of Samples:190Number of Tests:1900 η : 0.0001
Number of Epochs
[B:S:E]
Input Size (PRI, PD)
[B:S:E]
Prediction
[Min–Max]
128128:32:2880.774–0.889
128320:32:4800.776–0.944
128512:32:6720.808–0.944
128704:32:8640.670–0.889
128:32:160896:32:9280.778–0.889
160:32:224960:32:9920.832–0.923
256128:32:2880.722–0.944
256320:32:4800.722–1.000
Table 11. Efficiency of signal recognition using the CNN based on a sampled signal in the presence of interference from other signals.
Table 11. Efficiency of signal recognition using the CNN based on a sampled signal in the presence of interference from other signals.
Input Size of ANN: 3.3554 10 5
Number of Samples80Number of Tests80Batch Size:40 η : 5 10 5
Number of EpochsPredictionDisruption Level
2560.9910.1
0.9690.2
0.9430.3
0.8030.4
0.6440.5
0.6310.6
0.3280.7
0.4540.8
0.4460.9
Input Size of ANN: 8.3886 10 5
4990.9220.9
Table 12. Efficiency of signal recognition using the CNN based on a sampled signal in the presence of pseudorandom noise.
Table 12. Efficiency of signal recognition using the CNN based on a sampled signal in the presence of pseudorandom noise.
Input Size of ANN: 3.3554 10 5
Number of Epochs:256Number of Samples:80Number of Tests80Batch Size40 η : 5 10 5
PredictionDisruption Level
0.9990.1
1.0000.2
0.9970.3
0.9970.4
0.9990.5
0.9840.6
0.9560.7
0.6600.8
0.7530.9
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Matuszewski, J.; Pietrow, D. Specific Radar Recognition Based on Characteristics of Emitted Radio Waveforms Using Convolutional Neural Networks. Sensors 2021, 21, 8237. https://doi.org/10.3390/s21248237

AMA Style

Matuszewski J, Pietrow D. Specific Radar Recognition Based on Characteristics of Emitted Radio Waveforms Using Convolutional Neural Networks. Sensors. 2021; 21(24):8237. https://doi.org/10.3390/s21248237

Chicago/Turabian Style

Matuszewski, Jan, and Dymitr Pietrow. 2021. "Specific Radar Recognition Based on Characteristics of Emitted Radio Waveforms Using Convolutional Neural Networks" Sensors 21, no. 24: 8237. https://doi.org/10.3390/s21248237

APA Style

Matuszewski, J., & Pietrow, D. (2021). Specific Radar Recognition Based on Characteristics of Emitted Radio Waveforms Using Convolutional Neural Networks. Sensors, 21(24), 8237. https://doi.org/10.3390/s21248237

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop