Next Article in Journal
Cross-Domain Sentiment Analysis Based on Feature Projection and Multi-Source Attention in IoT
Previous Article in Journal
Dispersion Turning Attenuation Microfiber for Flowrate Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Modulation Classification Based on CNN-Transformer Graph Neural Network

1
Key Laboratory of Electronics and Information Technology for Space Systems, National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(16), 7281; https://doi.org/10.3390/s23167281
Submission received: 26 July 2023 / Revised: 16 August 2023 / Accepted: 18 August 2023 / Published: 20 August 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In recent years, neural network algorithms have demonstrated tremendous potential for modulation classification. Deep learning methods typically take raw signals or convert signals into time–frequency images as inputs to convolutional neural networks (CNNs) or recurrent neural networks (RNNs). However, with the advancement of graph neural networks (GNNs), a new approach has been introduced involving transforming time series data into graph structures. In this study, we propose a CNN-transformer graph neural network (CTGNet) for modulation classification, to uncover complex representations in signal data. First, we apply sliding window processing to the original signals, obtaining signal subsequences and reorganizing them into a signal subsequence matrix. Subsequently, we employ CTGNet, which adaptively maps the preprocessed signal matrices into graph structures, and utilize a graph neural network based on GraphSAGE and DMoNPool for classification. Extensive experiments demonstrated that our method outperformed advanced deep learning techniques, achieving the highest recognition accuracy. This underscores CTGNet’s significant advantage in capturing key features in signal data and providing an effective solution for modulation classification tasks.

1. Introduction

Wireless communication has become indispensable in various national undertakings and the daily routines of individuals. In light of the rapid advancements in wireless communication technology, achieving efficient data transmission in complex wireless environments has become imperative. Various modulation schemes are commonly employed to modulate transmitted signals, and the diversity and intricacy of modulation methods continue to expand. As an intermediate step between signal reception and demodulation, automatic modulation classification (AMC) is a critical technology in military and civilian communications [1,2]. It finds application in electronic warfare, intelligence operations, surveillance, threat analysis, and spectrum monitoring, facilitating the identification of modulation schemes and the subsequent decoding of received signals. Therefore, a reliable and effective AMC scheme is of paramount importance.
After extensive research and development, researchers categorized AMC methods into three main groups: likelihood-based (LB) methods, distribution-test-based (DT) methods, and feature-based (FB) methods. The LB [3,4,5,6] methods involve comparing observed signal samples with various modulation hypotheses, to estimate their similarity, leading to the derivation of a likelihood function based on the selected signal model. Classification decisions are then made by assessing the resemblance to different modulation hypotheses. However, this method often necessitates prior knowledge of channel parameters and other assumptions, resulting in increased computational complexity of the likelihood function.
The DT methods adhere to the symbol mapping mechanism framework [7], tailored for each specific supported modulation scheme. When the theoretical distributions of diverse modulation patterns are accessible, the goal is to identify the distribution that best aligns with the signal to be classified. The classification procedure involves determining the optimal goodness of fit (GoF) [8] among the hypothesized signal distributions. However, this method exhibits subpar recognition performance under low signal-to-noise ratio (SNR) conditions.
FB methods are carried out in two sequential steps: feature extraction and classifier design. These methods commonly employ various types of features, such as instantaneous features (e.g., instantaneous frequency, phase, and magnitude [9]), statistical characteristics (e.g., higher-order cumulants and higher-order moments [10]), and wavelet transform features [11]. Essentially, the FB approach segregates signals of different modulation modes within a high-dimensional feature space by designing an appropriate classifier. The choice of classifiers includes support vector machines (SVM) [12], decision trees [13], hidden Markov models [14], and artificial neural networks [15]. However, noise is often present in practical, complex, and dynamic communication environments, and manual feature extraction may need help to select the most discriminative features. Moreover, the classifiers designed are often not optimally tailored to the selected artificial features. These challenges can result in a diminished classification performance when facing real-world conditions.
The aforementioned LB and DT methods necessitate prior knowledge of specific channels and noises, and practical application scenarios often involve numerous uncertain factors, leading to a relatively low modulation recognition accuracy. Machine learning-based approaches extract meaningful attributes from raw data to describe and represent the data. These methods have shown remarkable performance in computer vision [16], natural language processing [17], recommendation systems [18], object detection [19], anomaly detection [20,21,22], and other domains. In recent years, the field of deep learning has witnessed remarkable progress, giving rise to a diverse range of network models, such as convolutional neural network (CNN) [23], recurrent neural network (RNN) [24], graph neural network (GNN) [25], and transformer network [26] models. These models have demonstrated exceptional achievements in the aforementioned applications. Consequently, modulation recognition methods based on deep learning have garnered increasing attention. In contrast to manual feature extraction, deep-learning-based methods seamlessly integrate feature extraction with classifier design, enabling automatic feature extraction and classification through an end-to-end model. For example, O’Shea et al. [27] utilized a CNN architecture comprising two convolutional layers and two fully-connected layers to extract distinctive features from raw IQ signals. Moreover, they extended their work in [28] by incorporating a 1D CNN to enhance the residual network’s performance compared to the VGG network. Remarkably, when SNR exceeds 5 dB, this approach demonstrates a favorable classification effect. Another notable approach, proposed by Hong et al. [29], introduced a modulation classification method based on RNNs utilizing gated recurrent units (GRUs). This method surpassed the recognition performance of certain CNN models. Furthermore, Sreeraj et al. [8] achieved high recognition accuracy by converting the IQ signal into magnitude and phase representations, and feeding them into an LSTM (long short-term memory) network.
Due to the successful application of GNNs, there has been a growing interest in converting time series into graph structures and utilizing GNNs for classification. Transforming signal sequences into graph structures allows modeling direct relationships between sample points. Compared to directly extracting raw IQ signal features using the aforementioned deep learning methods, this approach can extract potential discriminative features that might be challenging to obtain from the original IQ signals. Furthermore, the transformer network [26], initially introduced in natural language processing, has demonstrated significant advantages in handling sequential data. Therefore, by combining the transformer’s ability to capture dynamic relationships between long sequences and the remarkable benefits of GNNs in processing graph-structured data, in this paper, we propose a novel learnable architecture for modulation classification. This architecture combines a CNN, transformer, and GNN, enabling flexible construction of graph structures between different sample points. First, we introduce the CNN-transformer network (CTN), which adaptively maps raw IQ signals into graph structures, rather than using fixed mapping rules. Fixed mapping rules make constructing an optimal graph structure for signals challenging. In contrast, our proposed mapping method provides greater flexibility, allowing the learned graph structure to be adjusted for optimization, resulting in an improved modulation classification performance. This mapping reveals inherent complex relationships between different signal segments and models their topological connections, enabling a more accurate understanding of the intrinsic features of modulation schemes and enhancing the model’s ability to recognize unlabeled modulation types. Second, we construct a GNN architecture based on GraphSAGE [30] and DMoNPool [31]. This architecture leverages the graph structure obtained from CTN to perform the classification task. The CTN and GNN form our modulation classification framework, called a CNN-transformer graph neural network (CTGNet). The main contributions of this paper are as follows:
  • We preprocess the IQ signal using a sliding window and reorganize the subsequences into a signal matrix; we propose the CTN, which can dynamically map the subsequence matrix into a graph structure, thereby mining temporal and spatial relationships of signal data, to better understand modulation schemes;
  • We integrate CTN and GNN for end-to-end training and optimization, which can automatically learn the most reasonable graph representation of signals, and the network structure can be flexibly extended for application to different classification tasks;
  • We validated our method using the publicly available datasets RML2016.10a and RML2016.10b. Comparisons with CNN, RNN, and other GNN-based models demonstrated that our CTGNet achieved state-of-the-art performance.

2. Related Works

2.1. Transformer in Classification

The transformer network, introduced by Vaswani et al. [26], is a neural network architecture based on the attention mechanism, initially designed for sequence modeling in natural language processing tasks. Unlike traditional RNNs and CNNs, the transformer’s attention mechanism can simultaneously consider information from different positions in the sequence, making it highly effective for handling long-range dependencies and capturing global context. To extend the applicability of the transformer to image classification tasks, the vision transformer (ViT) [32] approach was proposed. ViT divides the image into a set of fixed-size image patches and treats each patch as a sequence input to the transformer. This enables ViT to capture global relationships within images, free from the constraints of receptive fields. Building on this, the transformer-in-transformer (TNT) [33] method incorporates a small transformer module within each image patch to capture local features. This enables interaction between local features and fine-grained relationship modeling across the entire image. Another advancement is the swin transformer [34], which adopts a hierarchical transformer architecture and performs self-attention modeling within each image patch. By interacting with image patches at different levels, a swin transformer can effectively capture multi-scale features in images, enhancing its modeling capabilities. Integrating various attention mechanisms, hierarchical structures, and feature interaction approaches in these models extends and improves the transformer architecture, leading to remarkable performance in image classification tasks. The successful application of these transformer-based models showcases their broad adaptability and powerful representation learning capabilities.

2.2. GNN in Classification

GNNs are a specialized category of neural network models specifically designed to handle graph-structured data. GNNs exhibit exceptional capabilities in learning and inferring from nodes and edges within a graph, enabling them to effectively capture intricate relationships and local patterns. A prominent example of a GNN are the graph convolutional networks (GCNs) proposed by Kipf et al. [35], which adapt convolutional neural network concepts to graph data. By iteratively aggregating information from neighboring nodes, GCNs update node representations, facilitating information propagation and feature extraction on graphs. In the quest for scalability, Hamilton et al. [30] introduced GraphSAGE, incorporating sampling and aggregation strategies to handle large-scale graph data, thus enhancing the model’s scalability without compromising its effectiveness. Another noteworthy advancement is the graph attention networks (GAT) introduced by Velickovic et al. [36]. GAT leverages attention mechanisms to learn node weights, allowing for more accurate capturing of node relationships and enhancing the model’s expressive power. To address the issue of graph dimensionality, Ying et al. [37] proposed DIFFPOOL, a differentiable pooling method that employs a hierarchical clustering strategy to partition nodes into subgraphs. This strategy effectively reduces the dimensionality of the entire graph, while preserving crucial structural information. Furthermore, deep modularity networks (DMoN), proposed by Anton et al. [31], are graph pooling methods that utilize a modularity measure. DMoNPool selects pooling nodes by optimizing modularity scores, enabling efficient learning and inference on graph data by capturing complex relationships effectively. Overall, these models demonstrate remarkable capabilities in handling graph-structured data through their effective relationship capture, incorporation of attention mechanisms, and employing pooling strategies for dimensionality reduction.

2.3. Deep-Learning-Based AMC

In recent years, deep learning methods have been extensively applied in the field of modulation classification, resulting in significant advancements. These methods effectively integrate feature engineering with classifier training, improving performance. Zhang et al. [38] first estimated the phase offset of the raw input signal, transformed the original signal based on this estimation, and then utilized the transformed signal for modulation classification. Inspired by hybrid models’ powerful feature extraction capabilities and the complementary information present in separate IQ channels, Xu et al. [39] proposed a novel multi-channel deep learning model named MCLDNN. This model extracts features from individual and combined IQ signals of the received data, considering the spatial and temporal aspects, and enhancing the classification performance. Yashashvi et al. [40] introduced a method to correct distorted signals by removing random frequency and phase offsets and then applied a CNN for feature extraction on the corrected signal. Liu et al. [41] designed a seven-layer residual network [42] for identifying ten types of modulation, showcasing the potential of deep learning in modulation classification. Tan et al. [43] employed a GRU to exploit temporal information for signal identification in exploring the temporal correlation feature. However, the direct use of a GRU network demonstrated comparable performance to that of a basic CNN in the work by O’Shea et al. [27]. Hu et al. [44] introduced an attention mechanism-based classifier framework, incorporating an RNN model with an attention mechanism to enhance classification performance. Zhang et al. [45] constructed a dual-stream network structure based on CNN and RNN, effectively extracting spatial and temporal-related features from the received signal. Furthermore, Sreeraj et al. [8] achieved a high recognition accuracy by transforming IQ signals into amplitude and phase representations and inputting them into an LSTM network, showcasing the benefits of using LSTM in modulation classification tasks. In another study detailed in [46], a novel approach was introduced to combine I/Q and A/P information. This method introduced a unique step attention fusion network (SAFN) module that amalgamates the diverse step outputs generated by the bidirectional gated recurrent unit (BiGRU) layer, each with distinct weights. Ghasemzadeh et al. [47] proposed an automatic modulation classification architecture and designed a stacking quasi-recurrent neural network (SQRNN) to simulate recurrent layer operations, and aggregate and extract features in time steps, thereby achieving high classification accuracy. In [48], the authors introduced a learning framework using an LSTM denoising auto-encoder to automatically extract reliable features from noisy radio signals for identifying modulation type. Zhang et al. [49] used fully-connected layers to build an autoencoder to enhance the information interaction between in-phase and quadrature channels, and the model learned the extracted interaction features and IQ data together. In [50], a hybrid model combining RNN and CNN, namely recurrent and convolutional neural network (R&CNN) was proposed. The model takes advantage of recurrent layers for time series data, as well as the spatial learning ability of convolutional layers. To solve the vulnerability of time–frequency representation to noise, Xu et al. [51] proposed a novel approach called deep TFT classification network (DTFTCNet), which integrates a time–frequency transform (TFT) and CNN within an end-to-end deep learning framework for radar signal modulation recognition. A modulation classification method based on deep feature fusion was proposed in [52], which utilizes ResNeXt for semantic feature extraction, GRU for time series representation, and combines ResNeXt with GRU output using discriminant correlation analysis (DCA) feature fusion models to improve AMC performance. Che et al. [53] introduced a spatial-temporal hybrid feature extraction network (STHFEN), which employs a dual feature extraction network to transform the signal into the spatial and temporal feature space and utilizes a hybrid inference classifier to combine the classification results. Zhang et al. [54] proposed high-order convolutional attention networks (HoCANs), which leverage a novel high-order attention mechanism to enhance feature correlations in radio signals. A complex-valued depthwise separable convolutional neural network (CDSCNN) was proposed in [55] for modulation classification. CDSCNN uses complex-valued operation units to achieve complex-valued feature learning. Zheng et al. [56] introduced a multi-scale radio transformer (Ms-RaT) with dual-channel representation to fuse frequency, amplitude, and phase information. A residual-attention convolutional network (RanNet) was proposed in [57], which employs advanced processing blocks, attention connections, and skip connections to enhance the intrinsic features of combined waveform data. Chang et al. [58] introduced a hierarchical classification head based convolutional gated deep neural network (HCGDNN) for modulation classification, leveraging the complementary information from different layers’ outputs in the deep learning model.
In addition to directly extracting features from raw time series, i.e., IQ signals, various methods have explored transforming IQ signals into alternative representations, such as images or matrices, to extract more discriminative features. Wang and Oates [59] utilized Gramian angular fields (GAF) and Markov transition fields (MTF) to convert time series into matrices, effectively capturing the underlying dynamics of the signals. Meanwhile, Peng et al. [60] transformed the original IQ signals into constellation diagrams and proposed an enhanced method for constellation diagram classification, leveraging AlexNet [61] and GoogLeNet [23] models for effective classification. Wang et al. [62] employed two CNN models for modulation classification. The first model identified seven modulation styles using raw IQ signals as input, while the second model, which we call ConsCNN, processed the input QAM constellation diagrams to recognize 16QAM and 64QAM modulations, demonstrating the significance of different signal representations in the classification task. Lin et al. [63] proposed a joint learning framework for robust modulation recognition in the presence of noise. The framework integrates three modules, dual-channel spectrum fusion, signal enhancement (SE), and signal classification (SC) into a single architecture, employing a multistage attention mechanism to enhance recognition-related features. In [64], a CNN-based time–frequency attention mechanism was proposed, focusing on learning meaningful frequency and temporal information for modulation recognition. Chen et al. [65] introduced the signal-to-matrix (S2M) method, where the I and Q channels underwent sliding window processing and were transformed into matrices. These matrices were then concatenated and fed into a CNN for feature extraction, showcasing the potential of matrix representations in feature learning. Statistical features have also been explored in modulation recognition. Lee et al. [66] designed a fully connected network using statistical features (such as skewness, kurtosis, and other high-order moments) as inputs, providing valuable insights into the significance of statistical characteristics in the classification process. On the other hand, Huang et al. [67] combined DenseNet [68] and LSTM to extract useful features from the cyclic correlation entropy vector (CCV) of the signal, highlighting the ability of deep learning models to capture intricate relationships within signals. Graph-based approaches have also gained traction. Liu et al. [69] treated each signal sample as a node in a graph and employed a CNN to learn node embeddings and adjacency matrices. Subsequently, graph convolutional networks were utilized for classification, demonstrating the power of graph representations in modeling complex relationships within the signals. Xuan et al. [70] used a graph neural network for modulation recognition, constructing an adjacency matrix through one-dimensional convolutions with different scales, to extract information between different sample points. Compared with the traditional methods of mapping time series to graphs (e.g., VG [71], HVG [72], LPVG [73]), their adaptive visibility graph algorithm (AVG) offered an adaptive way to convert the signal into a graph structure. However, this approach has limitations, as it relies on the size of the one-dimensional convolution kernel, overlooks the importance of global information, and cannot effectively capture long-distance dependencies within signals.

3. The Proposed Method

In this section, we comprehensively introduce CTGNet, describe the preprocessing of IQ signals, and utilize the CTN to generate graph representations of these signals. Furthermore, we also elucidate the architecture of GNN, which was specially designed for the feature extraction and classification of graphs. Before delving into the intricacies of our approach, it is worth highlighting two key considerations behind our proposed method. First, RNNs face the problem of vanishing and exploding gradients when dealing with long-term sequences. Furthermore, CNNs, while effective in capturing local information, may ignore critical global dependencies in the data. To circumvent these limitations, we leverage transformer networks, known for their remarkable ability to model long-range dependencies. By adopting a multi-head self-attention mechanism, a transformer network can directly identify the dependency between any two positions in the sequence. This property has proven invaluable in capturing the inherently multilevel information and multimodal features of signal sequences. Transformer networks thus become a compelling choice for modeling the underlying connections between different signal sequences, and these relationships are effectively represented as interactions between nodes in a graph structure. Furthermore, by building GNNs, we can better understand the topological relationships within the signal graph structure and the complex interaction patterns between nodes. This enhanced understanding facilitates the analysis of different modulation schemes. Considering all these factors, our CTGNet overcomes the limitations of traditional network models and fully exploits the potential of rich temporal information in time series data.

3.1. IQ Data Preprocessing

The raw IQ data represents samples of the signals’ in-phase (I) and quadrature (Q) components. However, a transformer network requires an input in the form of embeddings. To achieve this, we employ sliding window processing to partition the long IQ sequence into shorter subsequences, each with a fixed length. Additionally, we specify the overlap length between adjacent subsequences. After partitioning the signal, we reorganize the subsequences into a signal sequence matrix and apply one-dimensional convolutional encoding. This encoding process transforms the subsequences into embeddings, capturing their essential features, which serve as input to the transformer network. The specific preprocessing steps are illustrated in Figure 1. A more detailed description of the steps is provided below.
The I signal and Q signal can be represented as I = [ i 1 , i 2 , , i n ] and Q = [ q 1 , q 2 , , q n ] , where I and Q are one-dimensional vectors of length n. We partition both the I signal and Q signal into shorter subsequences of length w. The subsequences are generated with a sliding window of step size s:
I 1 = [ i 1 , i 2 , , i w ]
I 2 = [ i 1 + s , i 2 + s , , i w + s ]
I m = [ i 1 + ( m 1 ) · s , i 2 + ( m 1 ) · s , , i w + ( m 1 ) · s ]
Q 1 = [ q 1 , q 2 , , q w ]
Q 2 = [ q 1 + s , q 2 + s , , q w + s ]
Q m = [ q 1 + ( m 1 ) · s , q 2 + ( m 1 ) · s , , q w + ( m 1 ) · s ]
where I 1 , I 2 , , I m are the subsequences partitioned by the I signal, and Q 1 , Q 2 , , Q m are the subsequences partitioned by the Q signal. The number of overlapping elements between two adjacent subsequences is w s . Therefore, for the I and Q signals of length n, the number of subsequences partitioned by the sliding window is m:
m = n w s + 1
After performing this partitioning operation, to maintain the time characteristics between adjacent subsequences and further explore their internal relations, the subsequences are rearranged to form an m × w matrix. For the I signal, the resulting matrix X I is represented as
X I = [ I 1 ; I 2 ; ; I m ] = i 1 i 2 i w i 1 + s i 2 + s i w + s i 1 + ( m 1 ) · s i 2 + ( m 1 ) · s i w + ( m 1 ) · s
and for the Q signal, the resulting matrix X Q is represented as
X Q = [ Q 1 ; Q 2 ; ; Q m ] = q 1 q 2 q w q 1 + s q 2 + s q w + s q 1 + ( m 1 ) · s q 2 + ( m 1 ) · s q w + ( m 1 ) · s
The previous steps described the preprocessing of the IQ data, including using sliding window operations to obtain subsequences and rearranging them into a subsequence matrix. In the following subsection, we introduce the process of mapping this matrix onto a graph structure.

3.2. Mapping the Processed Data to Graph

After obtaining the subsequence matrices X I and X Q for the IQ data, in this section, we propose the CTN used to map them onto corresponding graphs, namely G I = { V I , ε I } and G Q = { V Q , ε Q } . The overall architecture of the CTN is illustrated in Figure 2.
Since the inputs of the transformer network are embeddings, after obtaining the matrices X I and X Q , we encode them using one-dimensional convolution to obtain the corresponding embeddings E I and E Q . Here, we assume that the dimension of the embeddings after one-dimensional convolution is d. For the input subsequence matrix of m × w , the output dimension after encoding is m × d . Next, in order to enable the model to capture the relative position and order information between subsequences and enhance the model’s ability to capture complex relationships between subsequences, we obtain E ˜ I and E ˜ Q by adding a learnable position encoding matrix P I and P Q , both with dimensions of m × d ,
E ˜ I = E I + P I
E ˜ Q = E Q + P Q
and in this way, E ˜ I and E ˜ Q will serve as the input to the transformer encoder.
The transformer encoder consists of multiple identical encoder layers containing two sub-layers: a multi-head self-attention network and feed-forward neural network. It also includes residual connections and layer normalization, which help address the vanishing gradient problem when applying deep models. The feed-forward neural network is a simple fully connected layer structure. The structure of the multi-head attention is shown in Figure 3. In the self-attention mechanism, each element in the input sequence interacts with other elements and computes attention weights between each element and the others. This allows the model to focus on the relationships between different positions in the input sequence. Multi-head attention divides the attention computation into multiple heads, where the attention is computed independently in each head. This enables different heads to focus on different aspects of the input sequence, providing more comprehensive information. Therefore, by learning the multi-head self-attention of the input sequence matrix, we obtain the adjacency matrix for the graph structure. According to [26], the specific computation process is as follows: The same computation process is applied to both the I signal and Q signal. Next, we will use the I signal as an example to explain this process.
To perform a linear transformation on the input subsequence embeddings E ˜ I and based on the number of attention heads n h , we multiply E ˜ I by weight matrix W i Q , W i K , and W i V , respectively. This operation yields the query matrix Q i , key matrix K i , and value matrix V i for each attention head:
Q i = E ˜ I W i Q
K i = E ˜ I W i K
V i = E ˜ I W i V
where i ranges from 1 to n h . W i Q , W i K , and W i V are weight matrices of dimensions d × d k , d × d k , and d × d v , respectively.
Then, by calculating the scaled dot-product between the query matrix Q i and the key matrix K i of each head, the correlation between them can be obtained. Through softmax function, the attention att i of each head can be obtained, and the attention can be weighted and summed over the value matrix V i to produce the output head i :
att i = softmax Q i K i T d k
head i = att i V i = softmax Q i K i T d k V i
and scaling makes the range of attention more appropriate, helping to train a stable model, and softmax makes all attention sum to 1.
In this way, each head i focuses on different representations of information, and different heads can focus on the dependencies between different locations. After that, concatenate each head i and multiply them with the matrix W O of size n h d v × d to obtain the matrix MultiHead :
MultiHead = head 1 , head 2 , , head n h W O
and this multi-head attention mechanism facilitates the model’s comprehension of the input data from diverse perspectives, thereby enhancing its capacity to capture a more extensive range of information.
Finally, after adding the MultiHead to the input E ˜ I and passing through LayerNorm, the result is sent to the feed-forward neural network, so that the output of the transformer encoder is obtained. Here, each element in att i represents the relative importance and correlation between each signal subsequence and other subsequences in the input subsequence matrix. By stacking the encoder layers as described above, we take the average value of the multi-head attention att i from the last encoder layer as the adjacency matrix A of the graph,
A = i = 1 n h att i n h
In addition, the input to the graph neural network includes the adjacency matrix and the node feature representation. For the node j in graph G I and G Q , we concatenate the corresponding subsequences as the feature representation of the node, denoted as
F j = [ i 1 + ( j 1 ) · s , i 2 + ( j 1 ) · s , , i w + ( j 1 ) · s , q 1 + ( j 1 ) · s , q 2 + ( j 1 ) · s , , q w + ( j 1 ) · s ]
We obtained the graphs of the I signal and the Q signal through the above steps. In the next subsection, we will use GNN for classification.

3.3. Classification of Graphs Using GNN

After mapping the I signal and Q signal onto graphs using the proposed CTN, we constructed a graph neural network for classification. The architecture of the graph neural network consists of GraphSAGE layers [30] followed by DMoNPool layer [31]. GraphSAGE is responsible for learning node representations in graph data. It achieves this by employing neighborhood sampling and feature aggregation. During this process, a set of neighboring nodes is sampled for each node, and their features are aggregated to generate a new representation for the central node. This operation is repeated for multiple iterations to progressively refine the node representations, capturing local and higher-order neighborhood information. On the other hand, DMoNPool handles the pooling of learned node representations to generate graph-level representations. By combining GraphSAGE and DMoNPool in CTGNet, the model can simultaneously leverage local node features and global graph structure features. After performing multi-layer node feature extraction and graph pooling, we obtained the feature vectors T I and T Q for the I signal and Q signal, respectively:
T I = [ T i 1 , T i 2 , ]
T Q = [ T q 1 , T q 2 , ]
and then concatenated T I and T Q to form:
T I Q = [ T i 1 , T i 2 , , T q 1 , T q 2 , ]
so the concatenated feature vector T I Q is then fed into a final fully connected layer for classification.
The above is our proposed CTGNet, as shown in Figure 4, and the information of each layer is shown in Table 1. It is able to learn graph-structured representations of signals in an end-to-end manner. By using CTN to map the I signal and Q signal to graphs G I and G Q , we construct a graph neural network based on a GraphSAGE layer and DMoNPool layer. CTGNet employs a holistic training approach for integrating graph construction and classification processes, ensuring that the transformed graph structure is optimal for modulation classification. Furthermore, representing signals as graphs can learn topological relations and extract global information among different sample points.

4. Dataset and Experiments

4.1. Dataset and Parameter Setting

In this paper, we evaluated our proposed method using the RML2016.10a [27] and RML2016.10b [74] datasets. The RML2016.10a dataset considers the effects of real-world electromagnetic environments and consists of 11 common modulation schemes. Each modulation scheme generated 20,000 samples at 20 different SNR levels, resulting in a total of 220,000 samples. The RML2016.10b dataset includes 10 modulation schemes and comprises a total of 1.2 million samples. Each sample in both datasets is organized into 2 × 128. The specific parameters of the datasets are presented in Table 2.
For the experiments, we split the RML2016.10a dataset into 80% for training and 20% for testing. As for the RML2016.10b dataset, we used 60% of the data for training and 40% for testing. During the model training process, we employed the cross-entropy loss function and utilized an Adam optimizer with a learning rate of 0.001. The batch size was set to 128. All our experiments were executed on a Nvidia GeForce GTX 1080 GPU, and all models were implemented using the PyTorch deep learning framework.

4.2. Baseline Methods

We compared our proposed method with nine different deep network models in terms of modulation classification performance. These models include AvgNet [70], MCLDNN [39], Resnet1d, VGG [28], CNN2d [27], GRU [29], LSTM [8], GAF [59], and ConsCNN [62]. Among them, GRU and LSTM are based on RNN, while Resnet1d, CNN2d and VGG are based on CNN. MCLDNN combines RNN and CNN architectures. Regarding AvgNet, it adopts one-dimensional convolution of different scales to construct the adjacency matrix of the signal, thus realizing classification based on GNN. GAF converts the signal into images and uses a resnet network to extract image features. ConsCNN processes the constellation diagram of the signal to realize the identification of modulation types.

4.3. Evaluation Metrics

In this paper, we used accuracy, F1 score, and recall to measure the performance of the different models. The accuracy was formalized as the proportion of correct predictions for the entire test set, as shown in Equation (23):
a c c u r a c y = i = 1 C ( T P i + T N i ) i = 1 C ( T P i + T N i + F P i + F N i )
where i = 1 , , C . C is the number of sample categories included in the dataset and the notations of T P i , F P i , F N i , T N i are described as the number of testing samples from i-th class in the conditions as below:
  • True Positive (TP): Truly positive, predicted to be positive
  • False Positive (FP): Truly negative, predicted to be positive
  • False Negative (FN): Truly positive, predicted to be negative
  • True Negative (TN): Truly negative, predicted to be negative
F1 Score is an indicator used to measure model performance in classification problems, defined as the harmonic mean of precision and recall. Therefore, the precision and recall of the i-th class prediction are shown in Equations (24) and (25):
p r e c i s o n i = T P i T P i + F P i
r e c a l l i = T P i T P i + F N i
and the total F1 score can be calculated as the mean of the F1 score in each class in Equation (26):
F 1 = 1 C i = 1 C 2 × p r e c i s i o n i × r e c a l l i p r e c i s i o n i + r e c a l l i

4.4. Results and Discussion

4.4.1. Ablation Study

In this section, an ablation study was performed to observe the effect of data preprocessing on the performance of CTGNet. The recognition performance of the model is presented in Table 3 for two scenarios: without data preprocessing and with data preprocessing. The results indicated that the model equipped with data preprocessing achieved a higher recognition accuracy than the model without preprocessing and exhibited a reduced training time. Through sliding window preprocessing, the original lengthy signal sequence is partitioned into shorter subsequences, which are organized into a subsequence matrix. The advantage of this approach lies in CTGNet’s ability to learn correlations among distinct subsequences, while considering the extraction of intrinsic features within each subsequence. This effectively captures both local and global features. Thus, data preprocessing not only enhances the model’s capability to understand intricate inter-subsequence relationships, but also facilitates the extraction of rich intra-subsequence features. Simultaneously, the reduced number of nodes in the learned graph significantly streamlines the overall processing time. The experimentation above confirmed the superiority of our proposed data preprocessing method.

4.4.2. Experiments on Different Sliding Window Sizes and Step Sizes

In this section, different sizes of sliding windows and strides were used in the data preprocessing stage, resulting in various sizes of samples and different degrees of overlap between adjacent sequences. Considering that each sample in both datasets contained 128 data points, we used sliding windows of size 8 and 16, with strides of 4, 6, and 8, and 8, 12, and 16, respectively. Employing these different window sizes and strides has an impact on the subsequent mapping to graph structures and classification using graph neural networks. Figure 5 displays the experimental results on both datasets under different settings.
From Figure 5, it can be observed that a higher recognition accuracy was achieved on the 2016.10a and 2016.10b datasets when using a window size of 16 with a stride of 8, as well as with a window size of 8 and a stride of 4. This indicates that under these parameter combinations, the model was able to construct graph structures more effectively and capture discriminative features for classification.

4.4.3. Comparisons with Other Baseline Methods

We compared our proposed model, which utilized a window size of 16 and a stride of 8, with various baseline methods. The experimental results of the two datasets are shown in Table 4 and Table 5, showing the accuracy, F1 score, and recall. In addition, Figure 6 shows the recognition accuracy of different methods as the signal-to-noise ratio changes. It is evident from the experimental results that our model achieved the highest performance on both datasets.
We also show the confusion matrices obtained using different recognition methods on the two datasets, as shown in Figure 7 and Figure 8. For the RML2016.10a dataset, it can be observed that all models tended to misclassify the other modulation types as AM-SSB. Additionally, except for CTGNet, AvgNet, MCLDNN, and ConsCNN, the other methods were more prone to misclassify QAM16 and QAM64. Furthermore, except for LSTM and GAF, the other methods tended to misclassify WBFM as AM-DSB, while LSTM more frequently misclassified AM-DSB as WBFM, and GAF misclassified AM-DSB as AM-SSB. In the case of CNN2d, it tended to misclassify QAM16, QAM64, and QPSK as 8PSK. For the RML2016.10b dataset, all methods tended to misclassify WBFM as AM-DSB. CNN2d was more inclined to misclassify other modulation types as 8PSK. In comparison to the other methods, CTGNet, AvgNet, and MCLDNN showed a better ability to differentiate between QAM16 and QAM64.

4.4.4. Discussion

As can be seen from the experiments in Section 4.4.1 and Section 4.4.2, the partition of subsequences of varying lengths and the amount of overlap between adjacent subsequences had an impact on the final recognition accuracy. Appropriately overlapping adjacent subsequences can bring several advantages when selecting window sizes and strides. First, overlapping adjacent subsequences implies the existence of overlapping regions, which leads to these regions receiving more attention when constructing the adjacency matrix. This facilitates increased information interaction between features within the sequence, aiding in extracting richer feature representations. Second, this can enhance contextual information. By overlapping adjacent subsequences, the model can consider the contextual information before and after the current sequence. This provides a more comprehensive contextual view, enabling the model to better understand the temporal relationships and semantic information within the sequence.
Compared with LSTM, GRU, and different CNN network models, our CTGNet transforms the signal into a graph structure and offers several advantages. By doing so, the model can capture the direct topological relationships between the sample points, enabling the model to effectively capture complex dependencies and intricate patterns. This is particularly beneficial for time series data, where long-range dependencies and non-linear relationships are prevalent. AvgNet, which constructs the graph structure using one-dimensional convolutions, is constrained by the convolutional kernel size, resulting in the oversight of long-range dependencies between sample points. In contrast, our proposed model not only captures local relationships, but also considers global contextual information. This comprehensive approach enhances the construction of topological relationships within time series data. Furthermore, our end-to-end model dynamically constructs the graph topology of the signal, enabling it to achieve optimal recognition accuracy. By leveraging the graph structure, our model could effectively integrate contextual information, capture complex dependencies, and adaptively represent the data, leading to superior performance compared to the baseline methods. In summary, our proposed CTGNet had the best performance among all the baseline methods, with a reasonable model size and training time. The model benefits from its ability to capture complex dependencies, incorporate contextual information, and dynamically construct a graph topology. These features collectively empowered our model to achieve exceptional recognition accuracy on both datasets.
At the same time, our proposed CTGNet has the potential limitation of a large computational complexity. In addition, when we partitioned the IQ signal, we used fixed lengths and steps. In the future, we will continue to study and devote ourselves to designing lightweight models to reduce computational complexity and design data preprocessing methods with variable parameters, such as a sliding window length and steps to improve model performance.

5. Conclusions

In this study, we introduced CTGNet, a novel approach that adaptively transforms signals into a graph structure for effective modulation classification. The proposed method comprises several key steps. First, we preprocess the raw IQ signal using a sliding window of fixed size and stride, obtaining subsequence samples. Next, we proposed a CTN to construct a graph structure based on these extracted subsequences. Finally, a graph neural network is utilized for the classification task. This end-to-end model demonstrated the capability to learn an optimal graph topology specifically tailored for signal classification. To validate the effectiveness of CTGNet, we conducted extensive experiments on the RML2016.10a and RML2016.10b datasets. Our results showcased the superiority of our proposed model in accurately classifying modulations. Additionally, we investigated the impact of different partition lengths of subsequences and the degree of overlap between adjacent subsequences on the recognition accuracy. The method proposed in this paper provides a reference value for the field of automatic modulation classification; that is, to explore different representation methods of signals and mine the potential features contained in them. In the future, different representation methods of signals, such as amplitude/phase, time–frequency representation, and graphs, can be fused to provide more abundant information, to promote the improvement of recognition performance.

Author Contributions

Conceptualization, D.W. and Y.Z.; methodology, D.W. and Y.Z.; software, D.W.; validation, D.W. and Y.Z.; formal analysis, D.W. and Y.H.; investigation, D.W., M.L. and X.Z.; resources, D.W.; data curation D.W., M.L. and X.Z.; writing—original draft preparation, D.W., M.L. and X.Z.; writing—review and editing, D.W., M.L., X.Z., Y.H. and Y.Z.; visualization, D.W., M.L. and X.Z.; supervision, Y.Z. and Y.H.; project administration, Y.Z. and Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, Z.; Nandi, A.K. Automatic Modulation Classification: Principles, Algorithms and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  2. Zhou, R.; Liu, F.; Gravelle, C.W. Deep learning for modulation recognition: A survey with a demonstration. IEEE Access 2020, 8, 67366–67376. [Google Scholar] [CrossRef]
  3. Xu, J.L.; Su, W.; Zhou, M. Software-defined radio equipped with rapid modulation recognition. IEEE Trans. Veh. Technol. 2010, 59, 1659–1667. [Google Scholar] [CrossRef]
  4. Panagiotou, P.; Anastasopoulos, A.; Polydoros, A. Likelihood ratio tests for modulation classification. In Proceedings of the MILCOM 2000 Proceedings, 21st Century Military Communications, Architectures and Technologies for Information Superiority (Cat. No. 00CH37155), Los Angeles, CA, USA, 22–25 October 2000; Volume 2, pp. 670–674. [Google Scholar]
  5. Hameed, F.; Dobre, O.A.; Popescu, D.C. On the likelihood-based approach to modulation classification. IEEE Trans. Wirel. Commun. 2009, 8, 5884–5892. [Google Scholar] [CrossRef]
  6. Dulek, B. Online hybrid likelihood based modulation classification using multiple sensors. IEEE Trans. Wirel. Commun. 2017, 16, 4984–5000. [Google Scholar] [CrossRef]
  7. Amuru, S.; da Silva, C.R. A blind preprocessor for modulation classification applications in frequency-selective non-Gaussian channels. IEEE Trans. Commun. 2014, 63, 156–169. [Google Scholar]
  8. Rajendran, S.; Meert, W.; Giustiniano, D.; Lenders, V.; Pollin, S. Deep learning models for wireless signal classification with distributed low-cost spectrum sensors. IEEE Trans. Cogn. Commun. Netw. 2018, 4, 433–445. [Google Scholar] [CrossRef]
  9. Moser, E.; Moran, M.K.; Hillen, E.; Li, D.; Wu, Z. Automatic modulation classification via instantaneous features. In Proceedings of the 2015 National Aerospace and Electronics Conference (NAECON), Dayton, OH, USA, 15–19 June 2015; pp. 218–223. [Google Scholar]
  10. Chang, D.C.; Shih, P.K. Cumulants-based modulation classification technique in multipath fading channels. IET Commun. 2015, 9, 828–835. [Google Scholar] [CrossRef]
  11. Hassan, K.; Dayoub, I.; Hamouda, W.; Berbineau, M. Automatic modulation recognition using wavelet transform and neural network. In Proceedings of the 2009 9th International Conference on Intelligent Transport Systems Telecommunications, (ITST), Lille, France, 20–22 October 2009; pp. 234–238. [Google Scholar]
  12. Han, L.; Gao, F.; Li, Z.; Dobre, O.A. Low complexity automatic modulation classification based on order-statistics. IEEE Trans. Wirel. Commun. 2016, 16, 400–411. [Google Scholar] [CrossRef]
  13. Swami, A.; Sadler, B.M. Hierarchical digital modulation classification using cumulants. IEEE Trans. Commun. 2000, 48, 416–429. [Google Scholar] [CrossRef]
  14. Ramkumar, B. Automatic modulation classification for cognitive radios using cyclic feature detection. IEEE Circuits Syst. Mag. 2009, 9, 27–45. [Google Scholar] [CrossRef]
  15. Popoola, J.J.; Olst, R. Automatic recognition of analog modulated signals using artificial neural networks. Comput. Technol. Appl. 2011, 2, 29–35. [Google Scholar]
  16. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  17. Wu, Y.; Schuster, M.; Chen, Z.; Le, Q.V.; Norouzi, M.; Macherey, W.; Krikun, M.; Cao, Y.; Gao, Q.; Macherey, K.; et al. Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv 2016, arXiv:1609.08144. [Google Scholar]
  18. Gao, C.; Wang, X.; He, X.; Li, Y. Graph neural networks for recommender system. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, Tempe, AZ, USA, 21–25 February 2022; pp. 1623–1625. [Google Scholar]
  19. Qian, X.; Lin, S.; Cheng, G.; Yao, X.; Ren, H.; Wang, W. Object detection in remote sensing images based on improved bounding box regression and multi-level features fusion. Remote Sens. 2020, 12, 143. [Google Scholar] [CrossRef]
  20. Lin, S.; Zhang, M.; Cheng, X.; Zhou, K.; Zhao, S.; Wang, H. Hyperspectral anomaly detection via sparse representation and collaborative representation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 946–961. [Google Scholar] [CrossRef]
  21. Lin, S.; Zhang, M.; Cheng, X.; Wang, L.; Xu, M.; Wang, H. Hyperspectral anomaly detection via dual dictionaries construction guided by two-stage complementary decision. Remote Sens. 2022, 14, 1784. [Google Scholar] [CrossRef]
  22. Lin, S.; Zhang, M.; Cheng, X.; Zhou, K.; Zhao, S.; Wang, H. Dual Collaborative Constraints Regularized Low-Rank and Sparse Representation via Robust Dictionaries Construction for Hyperspectral Anomaly Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 2009–2024. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  24. Graves, A.; Mohamed, A.r.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  25. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Networks Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  27. O’Shea, T.J.; Corgan, J.; Clancy, T.C. Convolutional radio modulation recognition networks. In Proceedings of the Engineering Applications of Neural Networks: 17th International Conference, EANN 2016, Aberdeen, UK, 2–5 September 2016; Proceedings 17. Springer: Berlin/Heidelberg, Germany, 2016; pp. 213–226. [Google Scholar]
  28. O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-air deep learning based radio signal classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef]
  29. Hong, D.; Zhang, Z.; Xu, X. Automatic modulation classification using recurrent neural networks. In Proceedings of the 2017 3rd IEEE International Conference on Computer and Communications (ICCC), Chengdu, China, 13–16 December 2017; pp. 695–700. [Google Scholar]
  30. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 2017, 30, 1024–1034. [Google Scholar]
  31. Müller, E. Graph clustering with graph neural networks. J. Mach. Learn. Res. 2023, 24, 1–21. [Google Scholar]
  32. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  33. Han, K.; Xiao, A.; Wu, E.; Guo, J.; Xu, C.; Wang, Y. Transformer in transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 15908–15919. [Google Scholar]
  34. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  35. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  36. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  37. Ying, Z.; You, J.; Morris, C.; Ren, X.; Hamilton, W.; Leskovec, J. Hierarchical graph representation learning with differentiable pooling. Adv. Neural Inf. Process. Syst. 2018, 31, 4805–4815. [Google Scholar]
  38. Zhang, F.; Luo, C.; Xu, J.; Luo, Y. An efficient deep learning model for automatic modulation recognition based on parameter estimation and transformation. IEEE Commun. Lett. 2021, 25, 3287–3290. [Google Scholar] [CrossRef]
  39. Xu, J.; Luo, C.; Parr, G.; Luo, Y. A spatiotemporal multi-channel learning framework for automatic modulation recognition. IEEE Wirel. Commun. Lett. 2020, 9, 1629–1632. [Google Scholar] [CrossRef]
  40. Yashashwi, K.; Sethi, A.; Chaporkar, P. A learnable distortion correction module for modulation recognition. IEEE Wirel. Commun. Lett. 2018, 8, 77–80. [Google Scholar] [CrossRef]
  41. Liu, X.; Yang, D.; El Gamal, A. Deep neural network architectures for modulation classification. In Proceedings of the 2017 51st Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 29 October–1 November 2017; pp. 915–919. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. Jiyuan, T.; Limin, Z.; Zhaogen, Z.; Wenlong, Y. Multi-modulation recognition using convolution gated recurrent unit networks. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2019; Volume 1284, p. 012052. [Google Scholar]
  44. Hu, S.; Pei, Y.; Liang, P.P.; Liang, Y.C. Deep neural network for robust modulation classification under uncertain noise conditions. IEEE Trans. Veh. Technol. 2019, 69, 564–577. [Google Scholar] [CrossRef]
  45. Zhang, Z.; Luo, H.; Wang, C.; Gan, C.; Xiang, Y. Automatic modulation classification using CNN-LSTM based dual-stream structure. IEEE Trans. Veh. Technol. 2020, 69, 13521–13531. [Google Scholar] [CrossRef]
  46. Chang, S.; Huang, S.; Zhang, R.; Feng, Z.; Liu, L. Multitask-learning-based deep neural network for automatic modulation classification. IEEE Internet Things J. 2021, 9, 2192–2206. [Google Scholar] [CrossRef]
  47. Ghasemzadeh, P.; Hempel, M.; Sharif, H. GS-QRNN: A high-efficiency automatic modulation classifier for cognitive radio IoT. IEEE Internet Things J. 2022, 9, 9467–9477. [Google Scholar] [CrossRef]
  48. Ke, Z.; Vikalo, H. Real-time radio technology and modulation classification via an LSTM auto-encoder. IEEE Trans. Wirel. Commun. 2021, 21, 370–382. [Google Scholar] [CrossRef]
  49. Zhang, F.; Luo, C.; Xu, J.; Luo, Y. An Autoencoder-based I/Q Channel Interaction Enhancement Method for Automatic Modulation Recognition. IEEE Trans. Veh. Technol. 2023, 21, 5977–5988. [Google Scholar] [CrossRef]
  50. Zhang, W.; Yang, X.; Leng, C.; Wang, J.; Mao, S. Modulation recognition of underwater acoustic signals using deep hybrid neural networks. IEEE Trans. Wirel. Commun. 2022, 21, 5977–5988. [Google Scholar] [CrossRef]
  51. Xu, S.; Liu, L.; Zhao, Z. DTFTCNet: Radar Modulation Recognition with Deep Time-Frequency Transformation. IEEE Trans. Cogn. Commun. Netw. 2023. early access. [Google Scholar] [CrossRef]
  52. Li, L.; Zhu, Y.; Zhu, Z. Automatic Modulation Classification Using ResNeXt-GRU with Deep Feature Fusion. IEEE Trans. Instrum. Meas. 2023, 72, 2519710. [Google Scholar] [CrossRef]
  53. Che, J.; Wang, L.; Bai, X.; Liu, C.; Zhou, F. Spatial-Temporal Hybrid Feature Extraction Network for Few-shot Automatic Modulation Classification. IEEE Trans. Veh. Technol. 2022, 71, 13387–13392. [Google Scholar] [CrossRef]
  54. Zhang, D.; Lu, Y.; Li, Y.; Ding, W.; Zhang, B. High-order convolutional attention networks for automatic modulation classification in communication. IEEE Trans. Wirel. Commun. 2022, 22, 4600–4610. [Google Scholar] [CrossRef]
  55. Xiao, C.; Yang, S.; Feng, Z. Complex-valued Depth-wise Separable Convolutional Neural Network for Automatic Modulation Classification. IEEE Trans. Instrum. Meas. 2023, 72, 2522310. [Google Scholar] [CrossRef]
  56. Zheng, Q.; Zhao, P.; Wang, H.; Elhanashi, A.; Saponara, S. Fine-grained modulation classification using multi-scale radio transformer with dual-channel representation. IEEE Commun. Lett. 2022, 26, 1298–1302. [Google Scholar] [CrossRef]
  57. Huynh-The, T.; Pham, Q.V.; Nguyen, T.V.; Nguyen, T.T.; da Costa, D.B.; Kim, D.S. RanNet: Learning residual-attention structure in CNNs for automatic modulation classification. IEEE Wirel. Commun. Lett. 2022, 11, 1243–1247. [Google Scholar] [CrossRef]
  58. Chang, S.; Zhang, R.; Ji, K.; Huang, S.; Feng, Z. A hierarchical classification head based convolutional gated deep neural network for automatic modulation classification. IEEE Trans. Wirel. Commun. 2022, 21, 8713–8728. [Google Scholar] [CrossRef]
  59. Wang, Z.; Oates, T. Encoding time series as images for visual inspection and classification using tiled convolutional neural networks. In Proceedings of the Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; AAAI: Menlo Park, CA, USA, 2015; Volume 1. [Google Scholar]
  60. Peng, S.; Jiang, H.; Wang, H.; Alwageed, H.; Zhou, Y.; Sebdani, M.M.; Yao, Y.D. Modulation classification based on signal constellation diagrams and deep learning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 718–727. [Google Scholar] [CrossRef]
  61. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1106–1114. [Google Scholar] [CrossRef]
  62. Wang, Y.; Liu, M.; Yang, J.; Gui, G. Data-driven deep learning for automatic modulation recognition in cognitive radios. IEEE Trans. Veh. Technol. 2019, 68, 4074–4077. [Google Scholar] [CrossRef]
  63. Lin, S.; Zeng, Y.; Gong, Y. Modulation Recognition Using Signal Enhancement and Multistage Attention Mechanism. IEEE Trans. Wirel. Commun. 2022, 21, 9921–9935. [Google Scholar] [CrossRef]
  64. Lin, S.; Zeng, Y.; Gong, Y. Learning of time-frequency attention mechanism for automatic modulation recognition. IEEE Wirel. Commun. Lett. 2022, 11, 707–711. [Google Scholar] [CrossRef]
  65. Chen, Z.; Cui, H.; Xiang, J.; Qiu, K.; Huang, L.; Zheng, S.; Chen, S.; Xuan, Q.; Yang, X. SigNet: A novel deep learning framework for radio signal classification. IEEE Trans. Cogn. Commun. Netw. 2021, 8, 529–541. [Google Scholar] [CrossRef]
  66. Lee, J.; Kim, B.; Kim, J.; Yoon, D.; Choi, J.W. Deep neural network-based blind modulation classification for fading channels. In Proceedings of the 2017 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 18–20 October 2017; pp. 551–554. [Google Scholar]
  67. Huang, S.; Lin, C.; Xu, W.; Gao, Y.; Feng, Z.; Zhu, F. Identification of active attacks in Internet of Things: Joint model-and data-driven automatic modulation classification approach. IEEE Internet Things J. 2020, 8, 2051–2065. [Google Scholar] [CrossRef]
  68. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  69. Liu, Y.; Liu, Y.; Yang, C. Modulation recognition with graph convolutional network. IEEE Wirel. Commun. Lett. 2020, 9, 624–627. [Google Scholar] [CrossRef]
  70. Xuan, Q.; Zhou, J.; Qiu, K.; Chen, Z.; Xu, D.; Zheng, S.; Yang, X. AvgNet: Adaptive visibility graph neural network and its application in modulation classification. IEEE Trans. Netw. Sci. Eng. 2022, 9, 1516–1526. [Google Scholar] [CrossRef]
  71. Lacasa, L.; Luque, B.; Ballesteros, F.; Luque, J.; Nuno, J.C. From time series to complex networks: The visibility graph. Proc. Natl. Acad. Sci. USA 2008, 105, 4972–4975. [Google Scholar] [CrossRef] [PubMed]
  72. Luque, B.; Lacasa, L.; Ballesteros, F.; Luque, J. Horizontal visibility graphs: Exact results for random time series. Phys. Rev. E 2009, 80, 046103. [Google Scholar] [CrossRef] [PubMed]
  73. Zhou, T.T.; Jin, N.D.; Gao, Z.K.; Luo, Y.B. Limited Penetrable Visibility Graph for Establishing Complex Network from Time Series; Acta Physica Sinica: Beijing, China, 2012. [Google Scholar]
  74. O’shea, T.J.; West, N. Radio machine learning dataset generation with gnu radio. In Proceedings of the GNU Radio Conference, Boulder, CO, USA, 12–16 September 2016; Volume 1. [Google Scholar]
Figure 1. Data preprocessing flow of the I signal and Q signal. w is the length of the sliding window, and s is the step size of the sliding window.
Figure 1. Data preprocessing flow of the I signal and Q signal. w is the length of the sliding window, and s is the step size of the sliding window.
Sensors 23 07281 g001
Figure 2. The architecture of CTN. This illustration demonstrates the architecture, using the I signal as an example. G I is the graph structure of the I signal obtained through CTN. Correspondingly, the graph structure of the Q signal is G Q .
Figure 2. The architecture of CTN. This illustration demonstrates the architecture, using the I signal as an example. G I is the graph structure of the I signal obtained through CTN. Correspondingly, the graph structure of the Q signal is G Q .
Sensors 23 07281 g002
Figure 3. The architecture of the multi-head attention. V , K , and Q are the value matrix, key matrix, and query matrix obtained through linear transformation, respectively, and n h is the number of heads of multi-head attention.
Figure 3. The architecture of the multi-head attention. V , K , and Q are the value matrix, key matrix, and query matrix obtained through linear transformation, respectively, and n h is the number of heads of multi-head attention.
Sensors 23 07281 g003
Figure 4. The architecture of the CTGNet. A I and A Q are the adjacency matrices of the graph structure of the I signal and Q signal, respectively. F j is the node feature. T I and T Q are the average feature vectors.
Figure 4. The architecture of the CTGNet. A I and A Q are the adjacency matrices of the graph structure of the I signal and Q signal, respectively. F j is the node feature. T I and T Q are the average feature vectors.
Sensors 23 07281 g004
Figure 5. Recognition accuracy for the two datasets with different sliding window sizes and step sizes. Here, the sliding window sizes were 8 and 16, and the corresponding step sizes were 4, 6, and 8, and 8, 12, and 16, respectively.
Figure 5. Recognition accuracy for the two datasets with different sliding window sizes and step sizes. Here, the sliding window sizes were 8 and 16, and the corresponding step sizes were 4, 6, and 8, and 8, 12, and 16, respectively.
Sensors 23 07281 g005
Figure 6. Recognition accuracy of ten methods with a change in SNR, where SNR ranged from −20 dB to 18 dB. (a) RML2016.10a; (b) RML2016.10b.
Figure 6. Recognition accuracy of ten methods with a change in SNR, where SNR ranged from −20 dB to 18 dB. (a) RML2016.10a; (b) RML2016.10b.
Sensors 23 07281 g006
Figure 7. Confusion matrices of the different models on RML2016.10a. Each row of the confusion matrix corresponds to the ground truth class, while each column corresponds to the class predicted by the respective models. (a) AvgNet; (b) MCLDNN; (c) Resnet1d; (d) VGG; (e) CNN2d; (f) GRU; (g) LSTM; (h) GAF; (i) ConsCNN; (j) CTGNet.
Figure 7. Confusion matrices of the different models on RML2016.10a. Each row of the confusion matrix corresponds to the ground truth class, while each column corresponds to the class predicted by the respective models. (a) AvgNet; (b) MCLDNN; (c) Resnet1d; (d) VGG; (e) CNN2d; (f) GRU; (g) LSTM; (h) GAF; (i) ConsCNN; (j) CTGNet.
Sensors 23 07281 g007aSensors 23 07281 g007b
Figure 8. Confusion matrices of the different models on RML2016.10b. Each row of the confusion matrix corresponds to the ground truth class, while each column corresponds to the class predicted by the respective models. (a) AvgNet; (b) MCLDNN; (c) Resnet1d; (d) VGG; (e) CNN2d; (f) GRU; (g) LSTM; (h) GAF; (i) ConsCNN; (j) CTGNet.
Figure 8. Confusion matrices of the different models on RML2016.10b. Each row of the confusion matrix corresponds to the ground truth class, while each column corresponds to the class predicted by the respective models. (a) AvgNet; (b) MCLDNN; (c) Resnet1d; (d) VGG; (e) CNN2d; (f) GRU; (g) LSTM; (h) GAF; (i) ConsCNN; (j) CTGNet.
Sensors 23 07281 g008aSensors 23 07281 g008b
Table 1. The structure of CTGNet. The length of the sliding window in data preprocessing is 16, and the step size is 8.
Table 1. The structure of CTGNet. The length of the sliding window in data preprocessing is 16, and the step size is 8.
OperationOutput Dimensions
Input128 × 1
Data Preprocessing15 × 16
Conv1D (size 5)15 × 96
Position Encoding15 × 96
TransformerEncoderLayer (num_layers 3, n h 8, d k 96, d v 96)15 × 96
Multi-Head Attention + Averageadjacency matrix: 15 × 15
Concatenate subsequences of I and Qnode feature: 15 × 32
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
Concatenate the output of above 6 layers15 × 576
DMoNPoolnode feature: 15 × 576, adjacency matrix: 15 × 15
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
Concatenate the output of above 6 layers15 × 576
DMoNPoolnode feature: 8 × 576, adjacency matrix: 8 × 8
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
GraphSAGE + BN + Relu15 × 96
Concatenate the output of above 6 layers15 × 576
Average576
FC + Relu288
FC + Relu96
Concatenate feature vectors T I and T Q 192
FC + Relu96
FC + Softmaxmodulation modes
Table 2. The parameters of the two datasets used in the experiments. The SNR of both datasets ranges from −20 to 18 dB, with RML2016.10a containing 11 modulation modes and RML2016.10b containing 10 modulation modes.
Table 2. The parameters of the two datasets used in the experiments. The SNR of both datasets ranges from −20 to 18 dB, with RML2016.10a containing 11 modulation modes and RML2016.10b containing 10 modulation modes.
DatasetSNRModulation Types
RML2016.10a−20 dB:2 dB:18 dBBPSK, QPSK, 8PSK, 16QAM, 64QAM, GFSK, CPFSK, PAM4, WBFM, AM-SSB, AM-DSB
RML2016.10b−20 dB:2 dB:18 dBBPSK, QPSK, 8PSK, 16QAM, 64QAM, GFSK, CPFSK, PAM4, WBFM, AM-DSB
Table 3. Recognition performance on two datasets with/without data preprocessing.
Table 3. Recognition performance on two datasets with/without data preprocessing.
DatasetModelAccuracyF1 ScoreRecallModel Size (MB)Training Time (s)
RML2016.10aw/processing0.62070.64390.62074.620.0440
wo/processing0.61680.63780.61689.550.2446
RML2016.10bw/processing0.64330.64630.64334.620.0444
wo/processing0.64180.63930.64189.550.2436
Table 4. Recognition accuracy, F1 score, and recall of ten methods on the RML2016.10a dataset.
Table 4. Recognition accuracy, F1 score, and recall of ten methods on the RML2016.10a dataset.
ModelAccuracyF1 ScoreRecallModel Size (MB)Training Time (s)
AvgNet0.61120.63090.61123.630.0897
MCLDNN0.61070.63700.61071.940.0219
Resnet1d0.58910.60370.58910.910.0072
VGG0.57320.58670.573247.320.0730
CNN2d0.52860.53440.528610.910.0082
GRU0.57200.58700.57202.520.0085
LSTM0.56140.57730.56140.970.0111
GAF0.53090.55710.530963.150.1547
ConsCNN0.54370.54490.543769.220.3006
CTGNet0.62070.64390.62074.620.0440
Table 5. Recognition accuracy, F1 score, and recall of ten methods on the RML2016.10b dataset.
Table 5. Recognition accuracy, F1 score, and recall of ten methods on the RML2016.10b dataset.
ModelAccuracyF1 ScoreRecallModel Size (MB)Training Time (s)
AvgNet0.63630.64060.63633.630.0858
MCLDNN0.63940.64120.63941.940.0219
Resnet1d0.63150.62980.63150.910.0078
VGG0.63650.63500.636547.320.0709
CNN2d0.55640.56380.556410.910.0080
GRU0.63070.62980.63072.520.0088
LSTM0.63290.63320.63290.940.0114
GAF0.55980.57660.559863.150.1685
ConsCNN0.56290.56410.562969.220.2998
CTGNet0.64330.64630.64334.620.0444
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, D.; Lin, M.; Zhang, X.; Huang, Y.; Zhu, Y. Automatic Modulation Classification Based on CNN-Transformer Graph Neural Network. Sensors 2023, 23, 7281. https://doi.org/10.3390/s23167281

AMA Style

Wang D, Lin M, Zhang X, Huang Y, Zhu Y. Automatic Modulation Classification Based on CNN-Transformer Graph Neural Network. Sensors. 2023; 23(16):7281. https://doi.org/10.3390/s23167281

Chicago/Turabian Style

Wang, Dong, Meiyan Lin, Xiaoxu Zhang, Yonghui Huang, and Yan Zhu. 2023. "Automatic Modulation Classification Based on CNN-Transformer Graph Neural Network" Sensors 23, no. 16: 7281. https://doi.org/10.3390/s23167281

APA Style

Wang, D., Lin, M., Zhang, X., Huang, Y., & Zhu, Y. (2023). Automatic Modulation Classification Based on CNN-Transformer Graph Neural Network. Sensors, 23(16), 7281. https://doi.org/10.3390/s23167281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop