1. Introduction
In recent years, with the rapid development of electronic countermeasure technology, jamming means have become complex and diverse, which puts forward higher requirements for the reliability of communication. Owing to its excellent performance, frequency-hopping communication has become widely utilized and is regarded as a secure method in military applications for hostile environments [
1]. However, the emergence of targeted interference has highlighted the limitations of traditional frequency-hopping techniques. To enhance the anti-interference ability of wireless communication systems [
2], this paper studies the anti-interference strategy based on Game Theory in frequency-hopping communication to deal with the interference attack in frequency-hopping communication and puts forward new ideas to solve the interference countermeasure problem, which is of great significance to improve the anti-interference ability of frequency-hopping communication systems.
As an important research topic in the field of digital signal processing, modulation recognition of communication signals has shown great potential in military and civil fields. In the military field, modulation recognition provides an important technical means for obtaining enemy intelligence in electromagnetic countermeasures and selecting the best jamming and suppression method. Accurately identifying the modulation mode of frequency-hopping signals can provide strong support for military information warfare by, for example, judging the attributes of enemy and our own targets and jamming enemy signals [
3]. Generally, after successfully intercepting enemy communication signals, it is undoubtedly a crucial task in communication countermeasure technology to determine the number levels and extract the feature level of the obtained mixed modulation signals and use the extracted features for further modulation recognition.
Traditional modulation recognition methods usually rely on manually designed features and complex signal processing algorithms, including maximum likelihood estimation based on hypothesis testing [
4] and feature extraction based on pattern recognition [
5]. These methods tend to perform poorly in the face of complex and variable frequency-hopping signals. In recent years, modulation recognition technology based on deep learning (DL) has attracted the close attention of researchers. Compared with traditional methods, modulation signal recognition based on DL does not need to rely on prior knowledge and can automatically extract features from data and classify them, so it not only has high classification accuracy but also stronger generalization ability in the face of large-scale data training. Mohamed A and others used a convolutional filter to use the basic convolutional neural network Alex Net and a residual neural network for compatibility with a constellation diagram, which significantly improved the accuracy of signal modulation classification [
6]. Lihong Guang et al. removed noise from a two-dimensional time–frequency map of a frequency-hopping signal by adaptive Wiener filtering and accurately extracted the time–frequency map of each hop signal by using the algorithm in image processing, which achieved the accurate recognition of the modulation mode of the frequency-hopping signal and achieved good results at −4 db [
7]. At present, DNNs are widely used in automatic modulation recognition (AMR) to complete signal detection and demodulation [
8], which greatly improves the accuracy of modulation recognition. In communication countermeasures, a jammer can accurately identify the modulation mode being used in a target communication system and decode the frequency-hopping signal by training a DNN to more effectively interfere with and destroy the enemy’s communication link.
Although deep learning modulation recognition technology has brought great convenience to people, its anti-interference performance has been questioned since 2013. In 2013, Szegedy et al. [
9] found adversarial examples that can attack the neural network model—examples that can make the machine learning model misjudge or misclassify by perturbing the normal examples slightly and imperceptibly. The study indicates that deep neural network (DNN) models are typically characterized by their high complexity and sensitivity, which enable them to detect minute variations within the input space. Exploiting this characteristic, it is possible to enhance resistance to attacks by introducing precisely calibrated minor perturbations to the original samples. This method constructs adversarial examples that can provoke incorrect classifications by the model, thereby demonstrating a critical vulnerability in its predictive accuracy. Goodfellow et al. [
10] proposed the fast gradient sign method (FGSM) in 2014. They added adversarial noise to the linear model and observed that when processing high-dimensional data input, the linear model was more vulnerable to the interference of adversarial examples, which overturned the theoretical explanation that the existence of adversarial examples was because the model was highly nonlinear. Kurakin et al. [
11] introduced the iterative fast gradient sign method (I-FGSM), building on prior work. This approach incrementally introduces perturbations through multiple iterations and reprojects the currently generated adversarial samples back into a predefined constraint set. Classification outcomes indicate that most of these adversarial examples are misclassified, thereby demonstrating the efficacy of adversarial attacks on neural network classifiers in practical scenarios. Dong et al. [
12] proposed a momentum iterative fast gradient sign method (MI-FGSM) to enhance resistance against sample attacks. This method integrates momentum into the gradient and gets rid of the bad local maximum in the iteration process to generate more mobile adversarial examples. Mardy et al. [
13] proposed projected gradient descent (PGD), which is different from the clipping operation of I-FGSM. It limits the size of a disturbance by projecting the results of each iteration to the
field of pure input.
At present, research on adversarial examples is mainly focused on image and audio. In the field of communication signals, communicators can add adversarial examples with specific disturbances to modulated signals. These adversarial examples can attack the modem of a communication system so that the DNN model of the reconnaissance party cannot correctly demodulate the signal or cause wrong decoding results, which significantly improves the ability of the communicators to resist smart interference. Therefore, this paper proposes a frequency-hopping modulation signal adversarial example attack method based on adaptive whitening and feature gradient smoothing to reduce the recognition rate of the modulation signal in the DNN model. The main contributions of this paper can be summarized as follows:
Experiments show that the conventional method of generating countermeasure samples has shortcomings when attacking the frequency-hopping modulation recognition model, and, according to the particularity of the frequency-hopping signal and the rich space–time characteristics of the hidden layer of the model, a countermeasure sample generation method AWFGS-MIFGSM suitable for the field of frequency-hopping signal modulation recognition is proposed.
The method initially considers that frequency-hopping signals are non-stationary signals whose frequencies change non-linearly over time. This typical time-varying characteristic results in a relatively concentrated energy distribution within a short time frame. To address this, the acquired frequency-hopping signals undergo an adaptive whitening process. This treatment enables a more uniform distribution of energy across frequencies, eliminates correlations between signals, and simplifies the generation of adversarial samples.
This method uses the high-dimensional spatial features of the hidden layer of the target model to calculate the gradient to launch the attack, which ensures that the amount of characteristic information of the spectrum signal sample is rich enough. Considering that single-point gradient information might be unreliable due to loss function surface oscillations, the characteristic gradient is smoothed using surrounding sample data to identify the optimal direction for countering disturbances and improving adversarial sample transfer.
Section 2 of this paper introduces the basic principle of adversarial samples and adversarial attack based on DNN modulation recognition. In
Section 3, the system model and the generation method of countermeasure samples based on adaptive whitening and feature gradient smoothing are described and analyzed. In
Section 4, the experimental setup is explained, and a series of experiments are described from the perspective of white box attack and black box attack, and the experimental results are analyzed. Finally, we discuss and conclude this work in
Section 5.
2. Related Literature Review
2.1. Adversarial Example Attack
Adversarial examples refer to the special samples formed by artificially adding subtle disturbances that are difficult to detect by the naked eye or that are visible to the naked eye after processing but that do not affect the overall system in the original data set. These disturbances are not random disturbances in the learning process but artificially constructed disturbances that can deceive the neural network model, as shown in Formula (1):
where
represents the added disturbance,
represents the neural network classifier,
represents the original image, and
represents the specified class. Since the minimum value of
is not easy to calculate, the loss function is introduced to change Formula (1) to Formula (2):
where
is the loss function, which is realized by calculating the cross entropy.
Adversarial samples possess strong camouflage capability, exploiting model vulnerabilities to launch targeted attacks that mislead the model into categorizing these samples into incorrect categories with high confidence. The impact of an adversarial example on modulation recognition is illustrated in
Figure 1. By introducing counter disturbance, the signal originally identified as a sine wave with 97.85% confidence is misclassified as a square wave with 99.92% confidence. This demonstrates that despite the incorrect classification results, the waveforms of the two signals are nearly identical.
Below are descriptions of the four most commonly used methods to generate adversarial examples.
2.1.1. FGSM
FGSM is an efficient and fast adversarial example generation method proposed by Goodfellow [
10] that is committed to generating adversarial examples close to original images. The generation formula is shown in Formula (3):
where
is the gradient of loss function
to input
,
is the parameter controlling the size of the disturbance, and
is the target category of the attack, that is, a single gradient iteration is performed in the direction of reducing the loss function corresponding to model category
. When the intention is to launch a no-target attack, the above formula is simply updated as follows:
where
is the correct category corresponding to the input sample
. The biggest feature of FGSM is its efficient running speed, so it is often widely used in scenarios that need to generate many adversarial examples, such as confrontation training. However, its disadvantage is that the overall performance of the generated adversarial samples is somewhat poor.
2.1.2. I-FGSM
I-FGSM [
11] can be regarded as a multiple-iteration version of FGSM. The original FGSM only adds a single-step disturbance along the direction of gradient increase, while I-FGSM makes a multi-step small disturbance along the direction of gradient increase through iteration and cuts the iteration results after each iteration update to ensure that they are kept within the valid interval (for example, it is usually the [0, 1] or [0, 255] interval for image data). Compared with FGSM, I-FGSM can construct more accurate disturbances, but the amount of calculation is increased. This method can be expressed as follows:
where the subscript
denotes the number of iteration rounds,
.
2.1.3. MI-FGSM
MI-FGSM [
12] attack incorporates momentum into the I-FGSM attack by introducing a small number of gradients generated by the current step while retaining some gradients from the previous step to stabilize the update direction and avoid falling into the local extremum. The improvement of this method is the accumulation of the velocity vector in the gradient direction by using momentum. The formula is as follows:
First, is input to classifier to obtain gradient ; then, the velocity vector is accumulated in the gradient direction through Formula (6) to update , and is updated by applying the symbol gradient in Formula (7), finally generating disturbance F. Compared with FGSM and I-FGSM, MI-FGSM gives higher mobility of adversarial examples.
2.1.4. PGD
Compared with the one-step confrontation of FGSM, PGD [
13] adopts the strategy of small-step and multi-step. PGD initializes with uniform random noise to project the gradient and clips the disturbance to a specified range after each iteration. The attack process is shown in Formula (8):
where
is the projection operation.
2.2. Modulation Recognition Adversarial Example Attack Based on a DNN
Modulation recognition can be regarded as a classification problem involving N modulation modes. The signal received by the communication receiver can be expressed as , where is the signal modulated by the transmitter according to a specific modulation scheme, transmits the impulse response of the wireless channel, is the frequency offset, is the phase offset, and σ indicates additive white Gaussian noise (AWGN). The purpose of any modulation classifier is to identify the modulation type of the signal given the received signal .
Modulation recognition can be categorized into classical and DL-based methods, depending on the use of deep learning algorithms. DL-based modulation recognition automates feature extraction and classification by feeding preprocessed signals directly into the network, significantly reducing the time needed to manually analyze communication signal characteristics. This advantage makes the method better adapted to future situations following the development of wireless communication where the amount of information may increase significantly, and it has higher recognition accuracy. The process is shown in
Figure 2.
DNNs are central to DL-based modulation recognition technology. They process signal characterization results, analyzing preprocessed and extracted signal data to infer and output the modulation mode. O’Shea et al. [
14] achieved the recognition and classification of three analog modulation signals and eight digital modulation signals based on a DNN model for the first time, and the accuracy rate reached 80%, proving the feasibility of applying DNNs to radio data recognition under the condition of a low signal-to-noise ratio. Ali et al. [
15] employed IQ samples, constellations, and high-order cumulants to train sparse self-coding for modulation recognition, confirming the DNN’s effectiveness in AWGN and flat fading channels via simulations. Xie et al. [
16] used high-order cumulants to extract different features of each signal type to train a DNN for modulation recognition. When the signal-to-noise ratio was −5 dB and −2dB, the overall recognition accuracy of the algorithm exceeded 99%. At present, research on the modulation recognition of communication signals mainly focuses on fixed-frequency signals, and there is a big gap in research on the modulation recognition of frequency-hopping signals at home and abroad. For frequency-hopping modulation signal recognition, reference [
17] introduced an algorithm that extracted instantaneous features and high-order cumulants from spread spectrum and conventional signals, enhancing recognition accuracy and reliability. Reference [
18] developed a method using time–frequency energy spectrum texture features for modulation recognition, employing a support vector machine classifier for training and classification.
Although DNNs have many advantages in the field of signal modulation recognition, there are also some problems and challenges, such as the large amount of data demands and lack of model generalization ability; the deep learning model is also more sensitive to targeted adversary attacks. Small and intentional disturbances may lead to classification errors in the model, which seriously affect the reliability and security of signal recognition.
Research on countermeasure samples for modulation recognition started late. In recent years, the academic community has gradually turned its attention to research on countermeasure sample attack methods based on modulation classification. In 2018, Sadeghi [
19] and others took the lead in research on countering sample attacks against the modulation recognition model of communication signals based on DL. The research results show that the modulation recognition model based on a DNN automatic encoder is vulnerable to interference. The paper further expounds on how attackers can effectively counterattacks. In 2020, Zhao [
20] and others studied and tested counterattack in the process of signal recognition, successfully reduced the recognition accuracy of the model through experiments, and verified the generalization ability of the model. In 2021, Lin et al. [
21] analyzed the effects of various gradient-based counterattack methods on modulation recognition; the experimental results showed that when the disturbance intensity was set to 0.001, the prediction accuracy could be reduced by 50%.
At present, the primary goal of counterattacks in modulation recognition is to improve attack performance, but research in the field of communication is still in its infancy, lacking the theoretical interpretation of counter samples. Most of the existing explanations are limited to a hypothetical interpretation and do not fully analyze the characteristics of the communication signal. Furthermore, current methods inadequately address the characteristics and gradient reliability of modulation signals, leading to issues like poor counterattack performance and limited black box adaptability. Improving the processing of the modulation signal’s characteristic gradient can significantly enhance both the effectiveness of attacks and the model’s security.
3. Anti Attack Method Based on Adaptive Whitening and Feature Gradient Smoothing
3.1. System Model
In the wireless communication environment, both the transmitter and receiver of frequency-hopping signals use the same communication protocol. During the communication process, the sender first modulates the frequency-hopping signal onto a carrier using a particular method to create a frequency-hopping modulation signal, which is then transmitted over the channel. The receiver needs to use the same modulation method as the sender to demodulate and reconstruct the received modulated signal and finally complete the communication process. Considering the existence of the reconnaissance party in the communication process, this party intercepts the communication signal and uses the intelligent DNN model to identify the modulation type of the signal, aiming to capture the content of the frequency-hopping signal. The system model is shown in
Figure 3.
To avoid this situation, the communication party needs to add adversarial examples to the communication signal on the premise of ensuring that its own communication is not affected as much as possible. This is done to flexibly attack the reconnaissance party deploying the DNN model and interfere with and mislead the identification results of the DNN model of the reconnaissance party so that the reconnaissance party cannot correctly identify the modulation type or demodulate and recover the intercepted signal, achieving the purpose of anti-reconnaissance. In this paper, an anti-attack method based on adaptive whitening and feature gradient smoothing (AWFGS) is proposed. Initially, the obtained frequency-hopping signal is adaptively whitened to enhance the useful features of the signal and facilitate subsequent feature extraction. Subsequently, the hidden layer feature extracted by the DNN model is used as the attack object, which significantly improves the attack accuracy and produces more refined adversarial examples, and the generated countermeasure samples have higher mobility.
3.2. Adaptive Whitening
Blind source separation refers to the process of recovering the source signal only by using the observed signal according to the statistical characteristics of the signal without any prior knowledge of the source signal and transmission channel. It has important applications in wireless communication and voice signal and digital image processing [
22]. As a necessary preprocessing step of blind source separation, whitening can identify the mixing matrix and directly realize the blind separation of non-stationary signals.
Currently, the whitening algorithm can be divided into a batch algorithm and an adaptive algorithm. The batch processing algorithm has good robustness, but it cannot meet the requirements of the system for real-time signal processing. The adaptive whitening algorithm, which is less complex, supports the online processing of mixed signals with effective real-time performance and has therefore been widely adopted and researched [
23]. Therefore, when whitening the original signal, the whitening algorithm with an adaptive form [
22] is often used. Since signal processing often involves processing signals with different characteristics and statistical properties, and these signals may have different distributions in time and frequency domains, adaptive whitening can better process different types and properties of signal data by adjusting the characteristics and statistical properties of the signal to preprocess the data, thus enhancing the overall effectiveness and quality of signal processing. Moreover, in feature extraction and pattern recognition, adaptive whitening can enhance the useful features in the signal, which is helpful for subsequent pattern recognition, classification, or prediction. Its structure is shown in
Figure 4 [
24].
is the whitening matrix with full rank, and the output whitening vector
meets the following characteristics:
where
is the observation signal,
is the identity matrix,
is the autocorrelation matrix of signal
, and
and
are the eigenvector matrix and eigenvalue matrix of
, respectively.
The adaptive whitening algorithm has excellent tracking performance and conditions for real-time signal processing. Its estimation of the whitening matrix
can be obtained by minimizing the cost function of Equation (11):
where
represents the determinant operation on matrix
. On the derivation of the instantaneous estimation of
over
, there are the following:
Based on Equation (12), the updated formula of whitening matrix
in the adaptive algorithm can be obtained as follows:
where
is the whitening signal and
is the step size parameter. In order to ensure convergence, its value should meet
, where
and
respectively represent the maximum eigenvalues of matrices
and
.
Different from the waveform of constant-frequency continuous signals, the waveform of frequency-hopping signals shows significant discontinuity, which leads to the inaccurate extraction of frequency-hopping signal features directly using the original signal and then affects the subsequent signal processing. However, gradient features are generally represented by high-dimensional data with high correlations and much redundant information, which not only increases the difficulty of data processing and model training but also reduces the amount of information on features, resulting in some gradient features being affected by abrupt points in the signal when representing modulated signals, making gradient calculation unstable. To solve the above problems, an adaptive whitening algorithm is introduced to minimize the interference between frequencies, effectively remove the correlation between data, improve the independence of sample features, and facilitate the accurate feature extraction of subsequent models. Additionally, the reduction in correlation reduces the dependence of the model on specific features, so the adversarial examples remain effective between different models, that is, there is a higher attack success rate between different models.
3.3. Feature Gradient Smoothing
Reference [
25] pointed out that the local non-smoothness of the loss surface impairs the transferability of generated adversary samples. To solve this problem, this study used the local average gradient instead of the original gradient to generate countermeasure samples, as shown in
Figure 5.
Source model a was used to generate countermeasure samples to attack target model B. and respectively represent the gradient of a corresponding point on the loss function surface of the two models. It can be seen that the loss function curve of model a showed an obvious oscillation phenomenon, which made the direction difference between and larger, which meant that the countermeasure samples generated on could not effectively attack model B, and the migration of countermeasure samples was low. If the gradient smoothing process was applied to model a, the local average gradient was obtained to replace the original -generated countermeasure samples to attack model B. Since the directions of and were closer, the migration of countermeasure samples could be higher, and the attack performance for model B was stronger, that is, .
In this study, we approximated the mathematical expectation of the gradient in the neighborhood by sampling n times in the neighborhood of the sample
:
where
is the average value of characteristics in the
neighborhood and
is the upper boundary of the
neighborhood, set as
, where
is the super parameter.
At present, although modulated signals based on gradient have destructiveness against attacks, they also have a series of limitations and challenges. Compared with high-dimensional data such as pictures, the amount of information in the spectrum signal sample is smaller, and the high-dimensional vector of the middle layer of the deep learning model can magnify the key features of the input sample. If the middle layer features extracted by the DNN model are used as the attack object, and the average gradient of its neighborhood is used to replace its single-point gradient, the surface oscillation of the loss function can be effectively smoothed, the accuracy of the attack can be improved, and a more refined modulated signal can be generated against the sample. In addition, for the same type of modulated signal samples, after different DNN models are trained, the output characteristics of the intermediate layer usually show some similarity, and the characteristics of the samples are transferable. Therefore, the disturbance generated by the counterattacks based on the characteristics of the middle layer should have better mobility.
3.4. Description of Attack Methods
Algorithm 1 introduces the process of generating countermeasure samples based on adaptive whitening and feature gradient smoothing. Firstly, the signal samples are adaptively whitened before the original signal input model, so that the sample features extracted after the input model are more effective, and the gradient can be calculated by using the rich space–time features in the hidden layer of the DNN model. Then, n samples are taken within a certain domain of the current data point
, n
samples are input into the intercepted hidden layer model
,
is calculated according to Formula (14), and then the mathematical expectation
of the gradient in the neighborhood of the data point is used to replace the gradient value of the point for subsequent iterations to reduce unstable factors, avoiding the algorithm falling into local extreme points and effectively smoothing the oscillation of the loss function surface. Then, a new loss function
is constructed by Formula (15), and the characteristic gradient is calculated and the attenuation factor
is updated. Finally,
is continuously updated to obtain the required countermeasure sample
. The complete block diagram of the algorithm is shown in
Figure 6.
After obtaining the original signal sample feature
after whitening, the average feature information is obtained. Different loss functions can be designed by using different
(
) norms to constrain the features, as shown in Formula (15).
To verify the effectiveness of the experimental method, AWFGS is introduced into MI-FGSM to obtain the momentum iteration fast gradient sign method AWFGS-MIFGSM, which is based on adaptive whitening and feature gradient smoothing. The pseudo-code of the algorithm is shown in Algorithm 1.
Algorithm 1 AWFGS-MIFGSM adversarial example attacks |
Input: Raw modulated signal sample , Truncate hidden layer model , New loss function , Norm constraint , Momentum decay factor , Disturbance size , Sampling times , Iterations , Attenuation factor , Neighborhood range size . |
Output: Optimize adversarial example . |
- 1:
Iteration step , neighborhood boundary
|
- 2:
,
|
- 3:
For t = 0 to T − 1 do
|
- 4:
is obtained by adaptive whitening of
|
- 5:
Take N samples randomly for neighborhood of
|
- 6:
Input N samples into the hidden layer model and obtain according to Formula (13)
|
- 7:
Calculate new loss function
|
- 8:
Calculate characteristic gradient, update
|
|
- 9:
Update
|
- 10:
End for
|
- 11:
Obtain optimized adversarial example
|
3.5. Analysis of Attack Methods
In machine learning, input data typically consist of various measurements, and there is a significant correlation between adjacent sampling points. If unprocessed data are fed into the network, this creates excessive redundancy and lowers the network’s training efficiency. A whitening operation before feature extraction can decrease data correlation and streamline the feature extraction process. Subsequently, the gradient calculated from these processed features is used to attack the DNN model. This approach enhances the mobility of the generated adversarial examples, increasing their attack success rate across different models. The reduction in correlation diminishes the model’s reliance on specific features, thereby increasing the likelihood that an adversarial example will be effective across various models.
Most of the attacks based on label gradients are methods that attackers try to maximize the gradient of the loss function with respect to the input data so that the model can produce a false classification of the adversarial example. In this process, the optimization goal is to maximize the classification loss. Adjusting the input data thus generates classification errors in the adversarial examples. The proposed algorithm does not use the classification loss as the optimization goal but uses extensive high-dimensional feature data in the DNN hidden layer to design adversarial examples, which not only makes the obtained sample signal features richer but also produces finer disturbances.
At present, most of the methods that have been used to combat sample attacks use the single-point data gradient value on the optimized path. Because the surface oscillation of the loss function leads to the unreliability of the single-point gradient information, the method proposed in this paper helps the model make full use of the data point neighborhood gradient information by whitening and neighborhood sampling, making the gradient direction on the loss function of the source model and the attack model closer so that the disturbance generated by this has better mobility and the success rate of black box attacks is higher.