Next Article in Journal
Image Processing Application for Pluripotent Stem Cell Colony Migration Quantification
Previous Article in Journal
Review of the Natural Time Analysis Method and Its Applications
Previous Article in Special Issue
Closed-Form Performance Analysis of the Inverse Power Lomax Fading Channel Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Deep Learning Framework for OFDM with Index Modulation Under Uncertain Channel Conditions †

by
Md Abdul Aziz
1,2,
Md Habibur Rahman
1,2,
Rana Tabassum
1,2,
Mohammad Abrar Shakil Sejan
3,
Myung-Sun Baek
3 and
Hyoung-Kyu Song
1,2,*
1
Department of Information and Communication Engineering, Sejong University, Seoul 05006, Republic of Korea
2
Department of Convergence Engineering for Intelligent Drone, Sejong University, Seoul 05006, Republic of Korea
3
Department of Electrical Engineering, Sejong University, Seoul 05006, Republic of Korea
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in ICTC 2024: The 15th International Conference on ICT Convergence, Jeju Island, Republic of Korea, 16–18 October 2024.
Mathematics 2024, 12(22), 3583; https://doi.org/10.3390/math12223583
Submission received: 30 October 2024 / Revised: 11 November 2024 / Accepted: 13 November 2024 / Published: 15 November 2024

Abstract

:
Index modulation (IM) is considered a promising approach for fifth-generation wireless systems due to its spectral efficiency and reduced complexity compared to conventional modulation techniques. However, IM faces difficulties in environments with unpredictable channel conditions, particularly in accurately detecting index values and dynamically adjusting index assignments. Deep learning (DL) offers a potential solution by improving detection performance and resilience through the learning of intricate patterns in varying channel conditions. In this paper, we introduce a robust detection method based on a hybrid DL (HDL) model designed specifically for orthogonal frequency-division multiplexing with IM (OFDM-IM) in challenging channel environments. Our proposed HDL detector leverages a one-dimensional convolutional neural network (1D-CNN) for feature extraction, followed by a bidirectional long short-term memory (Bi-LSTM) network to capture temporal dependencies. Before feeding data into the network, the channel matrix and received signals are preprocessed using domain-specific knowledge. We evaluate the bit error rate (BER) performance of the proposed model using different optimizers and equalizers, then compare it with other models. Moreover, we evaluate the throughput and spectral efficiency across varying SNR levels. Simulation results demonstrate that the proposed hybrid detector surpasses traditional and other DL-based detectors in terms of performance, underscoring its effectiveness for OFDM-IM under uncertain channel conditions.

1. Introduction

Because of its dependability and efficiency, many standards organizations have acknowledged and embraced multi-carrier (MC) transmission [1] for wireline and wireless systems [2,3]. Key advantages of this method include its enhanced capacity to reduce inter-symbol interference (ISI) and its resistance to single-frequency interference. MC is a popular choice for emerging applications such as wireless local area networks (WLANs) [4] and power line communications (PLC) [5] thanks to its capacity to withstand a range of channel impairments, including impulse noise and frequency selectivity. The fundamental idea behind the MC protocol is to use complex exponentials as information carriers to break the spectrum down into a number of orthogonal narrowband subchannels [6]. The two most frequently used MC techniques are discrete multi-tone (DMT) [7] for wireline systems and orthogonal frequency-division multiplexing (OFDM) [8], which is mostly utilized in wireless applications.
Recently, numerous advanced multi-carrier (MC) techniques based on OFDM have been explored to provide enhanced energy and spectrum [9,10]. Among these, OFDM employing index modulation (OFDM-IM) stands out as a promising approach. OFDM-IM is a viable alternative for next-generation wireless communication that transmits data more efficiently than conventional OFDM. Index modulation (IM) encodes information not only in data symbols but also in the indices of active elements, enabling additional data transmission without requiring extra bandwidth [11]. This technique enhances data throughput as well as resilience against frequency-selective fading and interference by leveraging diverse subcarriers and indices. Moreover, the selective activation of subcarriers supports energy-efficient operation, making it suitable for energy-conscious applications [12]. The ability of OFDM-IM to mitigate both inter-symbol and inter-carrier interference while balancing spectral efficiency, energy efficiency, and system complexity makes it a versatile choice for modern wireless networks [13].
Several works have been conducted on OFDM-IM [14,15]. In [16], a noncoherent OFDM-IM system based on maximum likelihood (ML) was proposed where the information bits were transmitted solely through the indices of active subcarriers. Low-complexity near-optimal detectors for OFDM-IM were introduced in [11] to address the computational challenges of ML detection, which is intensive due to the presence of index bits. The log-likelihood ratio (LLR) detector was introduced in [17]. Although it achieves near-ML performance, knowledge of the incoming signal’s noise power spectral density is required, which represents a drawback. A low-complexity OFDM-IM detector that encodes every potential subcarrier activation pattern (SAP) was presented in [18]. This approach overlooks the possibility of incorrect SAP detection by using every potential SAP to transmit data. In [19], a greedy detector (GD) based on energy detection was proposed to estimate OFDM-IM performance with significantly lower complexity, although it did not achieve near-optimal performance.
Reliable data transmission in OFDM-IM relies on accurate detection, as it ensures proper decoding of both index and data symbols, an enhanced bit error rate (BER), resilience to channel impairments, and preservation of communication quality [20]. Deep learning (DL) excels in handling high-dimensional complex data, adapting to varying conditions, and performing end-to-end optimization, making it ideal for efficient signal detection and data processing in communication systems [21,22,23]. Thus, DL may serve as a promising approach for OFDM-IM systems. Various DL-based strategies have been explored to simplify OFDM-IM implementations while achieving near-optimal performance [24,25,26]. A deep neural network (DNN)-based OFDM-IM system called DeepIM was proposed in [27]. The authors investigated the performance of the proposed model by utilizing both rectified linear units (ReLU) and hyperbolic tangent (Tanh) activation functions. A low-complexity dual-mode OFDM-IM system was proposed in [28], which employed a DL model known as DeepDM constructed through a combined DNN and convolutional neural network (CNN) to identify carrier bits and index bits independently. The authors of [29] proposed an index-bit detection neural network for OFDM-IM that incorporated fully connected layers and an inception module to effectively manage different modulation settings. In [30], the authors presented a DL-based index bit estimator (IE-DNN) for active LED index detection in visible-light communication systems. By employing fully connected or convolutional layers, IE-DNN significantly reduces the index error rate (IER) and BER with acceptable complexity. In [31], the authors presented a CNN-based detection framework for OFDM-IM systems. This framework uses convolutional layers to effectively identify indices by converting incoming symbols to polar coordinates, creating 2D matrices. In [32], the authors proposed a multiple-input–multiple-output OFDM-IM system using IMNet, a non-iterative detector. The IMNet design combines two subnets with the conventional least squares method to recover the transmitted signal. In [33], a transformer-based detector called TransIM was proposed. The structure of the transformer adds a great deal of complexity to TransIM. In an earlier work [34], we presented a long short-term memory (LSTM)-based detector for OFDM-IM systems. Although we mostly concentrated on perfect channel conditions in that study, we also assessed BER performance under uncertain channel state information (CSI) settings. Interestingly, under imperfect CSI conditions, our previous experiment outperformed the ML, GD, and DeepIM detectors at higher SNR levels while showing comparable performance at lower SNR levels.
CNN-based models can effectively learn and extract features from received signals [35]. Their ability to capture spatial hierarchies enhances the detection of active subcarriers, enabling more accurate index identification in OFDM-IM systems. The translation invariance of CNNs ensures consistent performance across varying signal positions, while their parameter efficiency reduces the risk of overfitting [31]. Furthermore, by incorporating dependencies from both past and future contexts, bidirectional LSTM (Bi-LSTM) models enhance sequential data processing. They address the vanishing gradient issue, enabling effective learning of long-range relationships, and provide parallel processing, which increases training efficiency [36,37]. In this paper, we propose a hybrid DL (HDL)-based OFDM-IM system designed to operate under uncertain CSI conditions. The proposed HDL detector integrates a one-dimensional CNN (1D-CNN) with a Bi-LSTM model. The 1D-CNN model extracts spatial information from the received signal, focusing on patterns associated with the activated subcarriers by analyzing the received symbols in the OFDM-IM system. This step enhances the model’s capacity to differentiate between various indices in the input data. The Bi-LSTM then processes these features, capturing temporal dependencies in the signal by considering both past and future information. This combined approach leverages spatial and temporal insights, significantly improving the OFDM-IM system’s BER performance as well as its adaptability to signal variations, resulting in more accurate index detection. We evaluate the BER performance of the proposed model across different signal-to-noise ratios (SNRs) for the OFDM-IM system under imperfect CSI conditions. Additionally, we assess the BER for various optimizers and equalizers. We determine the throughput and spectral efficiency (SE) to further illustrate the resilience of the suggested HDL-based OFDM-IM system. Our simulation results demonstrate that the proposed model outperforms traditional detectors as well as other DL-based detectors. The key contributions of this paper are summarized as follows:
  • We introduce an HDL-based OFDM-IM system that operates under uncertain CSI conditions. This HDL-based detector integrates a 1D-CNN model with a Bi-LSTM model; the 1D-CNN model captures spatial information from received signal, while the Bi-LSTM model processes these features. This combined approach improves detection performance even in the presence of CSI uncertainty.
  • The channel data and received signal are preprocessed using OFDM-IM domain knowledge before being fed into the HDL model. This preprocessing step enhances the HDL-based detector’s accuracy in detecting active subcarrier indices. Additionally, we apply several different equalizers to the received signal and evaluate their comparative performance.
  • We assess the BER performance of the proposed HDL-based detector at various SNR levels and explore its performance with different optimizers. Additionally, we calculate the SE and throughput of the model to further demonstrate its effectiveness. The results indicate that the proposed HDL detector significantly enhances detection performance in OFDM-IM systems.
The remainder of this paper is organized as follows: the system modeling is introduced in Section 2; a thorough discussion of the suggested model, including details on offline training and online testing protocols, is provided in Section 3; the simulation results are presented in Section 4 and the complexity in Section 5; finally, the conclusions are presented in Section 6.

2. System Model

OFDM-IM is a sophisticated modulation technique that extends the traditional OFDM architecture. Unlike conventional OFDM, which utilizes all subcarriers to carry data, OFDM-IM conveys additional information by using both data symbols and the indices of active subcarriers. By selectively activating a subset of subcarriers, OFDM-IM enhances spectral efficiency and improves the system’s resilience against interference and fading, leading to better BER performance. The signal transmission process of an OFDM-IM system is illustrated in Figure 1.
There are N subcarriers in each of the sections S that make up the entire transmitted signal bandwidth. Therefore, the total number of transmitted subcarriers in this OFDM-IM system is N T , where N T = S N . Each section’s signal processing system is independent from the others and operates separately. Of the N subcarriers, only A are active in each section, while the remaining ( N A ) subcarriers are deactivated (zero-padded). A total of m bits is transmitted per section in each transmission. These m bits consist of m 1 bits sent by the modulated symbols and m 2 bits sent by the active subcarriers’ indices. In this case, m 1 bits is represented as m 1 = A l o g 2 M and m 2 bits is represented as m 2 = l o g 2 C ( N , A ) ; hence, the total bits m can be expressed as follows:
m = m 1 + m 2 = A l o g 2 M + l o g 2 C ( N , A )
where M is the M - a r y modulation size. The m 1 bits are mapped to a set of A active subcarrier indices using combinatorial methods. As a result, the transmit vector x = [ x 1 , , x N ] is formed by assigning A nonzero data symbols to the A active subcarriers based on the incoming m bits. The mapping from bits to symbols is denoted as x = f O F D M I M ( D i ) , where D i represents the input sequence of m bits for each group.
The received signal in the frequency domain is represented as follows:
y = h x + w
where h = [ h 1 , , h N ] denotes the Rayleigh fading channel and h i CN ( 0 , 1 ) . The term w i CN ( 0 , σ ) represents the additive white Gaussian noise (AWGN).
For the imperfect CSI settings, we consider a system in which the receiver experiences channel estimation errors. Let h ^ ( α ) denote the estimated channel, while the actual channel is modeled as follows:
h ( α ) = h ^ ( α ) + e ( α )
where the channel estimation error e ( α ) follows a complex Gaussian distribution e ( α ) CN ( 0 , ϵ 2 ) , where ϵ 2 represents the channel estimation error variance. The estimated channel h ^ ( α ) is modeled as h ^ ( α ) CN ( 0 , 1 ϵ 2 ) . The CSI error variance ϵ 2 is inversely related to the average SNR by γ ¯ = E a 2 σ 2 , where E a 2 is the average energy of a transmitted M - a r y symbol and σ 2 is the noise variance.
In this experiment, ϵ is computed as ϵ = 1 1 + SNR , with SNR = 10 S N R d B 10 . This relationship adjusts the error variance based on the average SNR, which can more accurately reflect the real-world conditions of imperfect CSI under Rayleigh fading.

3. Architecture of the Proposed HDL Detector

Figure 2 illustrates the structure of the proposed HDL model. Similar to existing OFDM-IM detection techniques, the receiver is assumed to have knowledge of the channel information. As a result, both the channel h and received signal y are considered as inputs to the HDL model. To enhance detection performance, y and h undergo preprocessing using domain knowledge from OFDM-IM before being fed into the HDL detector. To obtain the equalized received signal vector, the widely used zero-forcing (ZF) equalizer is applied [38]. This is represented as follows:
y ¯ = y h 1
where the received signal y is multiplied element-wise by the inverse of the channel h . Along with this equalizer, we also utilized the minimum mean square error (MMSE) and decision feedback equalizer (DFE) to compare their performance. By minimizing the mean square error (MSE) between the transmitted and received signals, the MMSE equalizer optimizes the tradeoff between inter-symbol interference (ISI) suppression and noise enhancement [39]. The equalization process of the MMSE equalizer is expressed as follows:
y ¯ = y · h | h | 2 + σ 2
where y ¯ represents the received signal equalized by MMSE, | h | 2 represents the squared magnitude of the channel coefficient, and σ 2 is the noise variance.
The DFE equalizer achieves equalization by using the received signal along with feedback from previously detected symbols to reduce the ISI [40]. The equalization process of the DFE equalizer can be represented as follows:
y ¯ = y d · y ^
where y ¯ is the equalized output signal, d represents the feedback filter applied to the previously detected symbols, and y ^ denotes the decision (or detected) output from the previous symbol decisions. To generate the HDL’s input, the energy of the received signal y e = | y 1 | 2 , | y 2 | 2 , , | y N | 2 is calculated and then added to y ¯ . The 3 N -dimensional input vector I = [ y ¯ R , y ¯ I , y e ] is created by concatenating the real and imaginary parts of y ¯ , respectively denoted as y R and y I , with the received energy vector y e , as illustrated in Figure 2.
Our proposed HDL detector employs an organized methodology consisting of several essential elements: a 1D-CNN layer, a batch normalization layer, a Bi-LSTM layer, a one-dimensional max-pooling layer, and a dense layer with a sigmoid activation function in the output layer. The processed data I are fed into the 1D-CNN layer. A total of F = 64 filters of various 3 × 3 sizes are used in the convolutional layer to extract features [41]. The function of the 1D-CNN model with the data I can be explained as follows:
U = f ( I 1 w i + b i )
where the weight and bias vectors of the i-th layer are indicated by w i and b i , respectively. The batch normalization (BN) layer follows next. The BN technique is used to normalize the input of each layer in order to minimize internal covariate shifts and stabilize the learning process. This enables neural networks to achieve faster training, greater stability, and improved performance [42]. The Bi-LSTM layer takes its input from the output of the BN layer, denoted as B. The proposed Bi-LSTM layer is configured with Q = 128 hidden units.
Two LSTM layers (forward and backward layers) make up a Bi-LSTM hidden layer. Figure 3 shows that each LSTM cell contains three gates, namely, the input, forget, and output gates [43]. Additionally, the LSTM has two states, the hidden state h t 1 and cell state C t 1 . The cell state functions as memory to retain information extrapolated from previous inputs, while the hidden state is used to compute the output. Here, t represents the time instant and B t is the current input. The gates function updates the cell state, enabling the Bi-LSTM cell to add or remove data at each time step. The forget gate and input gate control the amount of cell state that needs to be updated and reset, respectively [44].
The functions of the forget, input, output, and candidate gates in the Bi-LSTM model are expressed as follows:
F o t = σ ψ ( W F o B t + T F o h t 1 + b F o ) ,
I n t = σ ψ ( W I n B t + T I n h t 1 + b I n ) ,
O u t = σ ψ ( W O u B t + T O u h t 1 + b O u ) ,
C a t = σ t a n h ( W C a B t + T C a h t 1 + b C a ) ,
where B t denotes the input vector of the Bi-LSTM at iteration t, the sigmoid activation function is represented by σ ψ , the terms F o t , I n t , O u t , and C a t correspond to the respective activation vectors of the forget, input, output, and candidate gates at iteration t, the vector h t 1 represents the hidden state from the previous iteration t 1 , and the weight matrices W F o , W I n , W O u , and W C a are associated with the forget, input, output, and candidate gates, respectively, with regard to the input B t . Similarly, the weight matrices T F o , T I n , T O u , and T C a are tied to the previous hidden state h t 1 . Finally, the bias vectors for the forget, input, output, and candidate gates are represented by b F o , b I n , b O u , and b C a , respectively. The Bi-LSTM model updates the previous cell state c t 1 to the current cell state according to the following equation:
U p t = + ( I n t C a t ) .
The hidden state function at any time step t can be represented as follows:
h t = O u t σ t a n h ( U p t ) .
The operation of the hidden state output in the Bi-LSTM layer is expressed as follows:
H t = σ f ( W H T T t + W H T T t + b v )
where T t and T t respectively represent the forward and backward sequence of the Bi-LSTM network.
The max-pooling layer now uses the Bi-LSTM layer’s output as its input. By choosing the maximum value within predetermined window sizes, the max-pooling layer shortens the input sequences, reducing dimensionality while preserving important temporal features for later layers [45]. The fundamental function of the max-pooling layer can be illustrated as follows:
O ( i , j ) = max q , r K H t ( i + q , j + r )
where O ( i , j ) is the output obtained after max-pooling, ( q , r ) denotes the indices within the pooling window, K is the pooling window, and H t ( i , j ) is the input feature map.
The output layer, which consists of a dense layer with a sigmoid activation function, receives the input O from the max-pooling layer. The weighted sum of the inputs is converted by the sigmoid activation function into a value between 0 and 1 indicating the output neuron’s probability or activity level [46]. The output D o of the HDL detector through the sigmoid activation function can be expressed as follows:
D o = f S i g m o i d ( W s × O + b s )
where W s is the weight vector of the output layer and b s is its bias.

Offline Training and Online Testing Procedure

The DL model must be trained offline using simulation data before applying the proposed HDL detector. Specifically, a set of transmitted vectors is generated by randomly producing different sequences of m bits, denoted as D i , such that x = f OFDM-IM D i . These vectors are subsequently transmitted to the receiver, where they are impacted by AWGN noise and the Rayleigh fading channel. According to their established statistical models, the noise vectors and the channel are also created at random, and vary from one bit-sequence to another in this case. The input dataset I , for which the labels are corresponding bit sequences D i , is obtained by preprocessing the received signal and channel vectors, i.e., y and h , as explained in the preceding section. The signal detection process of the proposed HDL-based decoder in the OFDM-IM system is comprehensively outlined in Algorithm 1, featuring each stage from data generation to HDL model design, training, and testing. This step-by-step breakdown offers a clear understanding of the entire workflow.
Algorithm 1 OFDM-IM symbol detection with HDL model
  1:
Initialize Parameters:
  2:
Set subcarriers N, active subcarriers A, modulation order M, SNR SNR dB , batch size, training epochs, and learning rate.
  3:
Calculate derived parameters: bits per symbol m, QAM symbols, power allocation, and noise level.
  4:
Define Subcarrier Patterns:
  5:
Generate index patterns based on N and A (e.g., index set idx).
  6:
Generate OFDM-IM Signal:
  7:
Define OFDM_IM_received(bits, SNRdb) for training and testing data generation:
 • 
Convert bits into subcarrier indices and QAM symbols.
 • 
Apply channel and noise effects.
 • 
Normalize and return the received signals.
  8:
Design Neural Network (HDL model):
  9:
Construct HDL_model(x):
 • 
Use Conv1D for feature extraction.
 • 
Apply BatchNormalization.
 • 
Add bidirectional LSTM layers.
 • 
Add GlobalMaxPool1D.
 • 
Use a dense output layer with sigmoid activation for bit prediction.
10:
Training Phase:
11:
for each epoch from 1 to training epochs do
12:
   Initialize average cost, avg _ cost = 0
13:
   for each batch from 1 to total batches do
14:
     Initialize empty batches batch_x and batch_y
15:
     for each sample from 1 to batch size do
16:
        Generate random bits
17:
        Compute received signal using OFDM_IM_received
18:
        Append signal to batch_x and bits to batch_y
19:
     end for
20:
     Run optimizer, calculate training cost, and update avg _ cost
21:
   end for
22:
   Append avg _ cost to training loss list
23:
end for
24:
Testing Phase:
25:
for each SNR in range of SNR values do
26:
   Set test number, test_number, based on SNR
27:
   Initialize empty test batches batch_x and batch_y
28:
   for each test sample from 1 to test_number do
29:
     Generate random bits
30:
     Compute received signal using OFDM_IM_received
31:
     Append signal to batch_x and bits to batch_y
32:
   end for
33:
   Calculate BER using model predictions and store result
34:
end for
35:
Display Results:
36:
Plot BER vs. SNR
The HDL model is trained using the gathered data in order to reduce the BER, or more specifically the discrepancy between D i and its prediction D o . Therefore, we use the mean-squared error (MSE) loss function for the training process. The MSE effectively penalizes larger errors more heavily than smaller ones due to squaring of the differences, making it highly sensitive to outliers. Mathematically, it is expressed as follows [47]:
L ( D i , D o ; θ ) = 1 n i = 1 n ( D i D o ) 2
where D i represents the true values, D o represents the predicted values, and n is the number of samples. Adaptive moment estimation (Adam) is a DL optimization technique that combines the strengths of both RMSprop and momentum. By adjusting the learning rate for each parameter, Adam efficiently handles sparse gradients and noisy data, making it well suited for complex models and large datasets [48]. Adam is readily implementable in a number of commercial DL platforms, including Keras 2.10.0 and TensorFlow 2.10.0. Additionally, we evaluate the performance of the RMSProp, Adagrad, and Gradient Descent optimizers for comparison.
Figure 4 and Figure 5 illustrate the training loss of the proposed model for the different equalizers and modulation schemes, respectively. We evaluate the training loss across each epoch to monitor the model’s performance over time. As observed in Figure 4, the proposed model demonstrates strong training loss performance across all equalizers. While the DFE equalizer exhibits slightly higher loss, both the ZF and MMSE equalizers achieve very low and nearly identical training loss levels. Figure 5 illustrates the training loss for different modulation schemes under various data setups, specifically for ( 4 , 1 , 4 ) , ( 8 , 2 , 8 ) , and ( 8 , 4 , 16 ) with the ZF equalizer. Notably, this training loss is estimated at 15 dB training SNR with a batch size of 1000 and learning rate of 0.005. As can be observed, the training loss tends to increase with higher modulation orders and data setup.
As the model’s performance depends significantly on the SNR level chosen for training, denoted by γ ¯ train ; thus, careful selection is crucial in order to optimize HDL’s effectiveness. Specifically, the model trained at this SNR level must perform reliably across other relevant SNR values, making it essential to select an optimal γ ¯ train . For instance, if γ ¯ train is set too low, the model may fail to generalize adequately, as it will not account for the noise impact during training. The simulation results presented in the next section detail the process of selecting an appropriate training SNR for each experimental setup.
We trained our proposed HDL model over 1000 epochs across all experimental settings, with each epoch comprising 20 batches and each batch containing 1000 data samples (often referred to as the batch size). This setup resulted in a total of 2 × 10 4 batches, yielding approximately 2 × 10 7 unique training samples. The samples for each batch were generated randomly during the training process. The HDL model was trained for various configurations of ( N , A , M ) . Specifically, we trained the proposed model with modulation orders of QPSK, 8QAM, and 16QAM. All training configurations for the proposed model are detailed in Table 1. Our proposed model was trained using a batch size of 1000 over 1000 epochs.
The proposed system can operate in real time to estimate data bits across different channel fading scenarios without the need for further modifications to θ after it has been trained offline. Upon receiving the signal and channel characteristics, the HDL model efficiently generates an estimated bit and performs comparably to the ML detector in the presence of channel estimation errors. We applied 100,000 samples to test the proposed model.
Figure 6 presents the confusion matrix of the proposed model under the ( 4 , 1 , 4 ) data setup and with a training SNR of 15 dB, providing a bit-wise performance evaluation on the test data. The confusion matrix shows the model’s accuracy in predicting “0” and “1” bits at the specified SNR level. Test samples with true bit labels were passed through the trained model to obtain predicted values, which were thresholded at 0.5 to produce binary predictions. The confusion matrix indicates that the model correctly classified 197,769 “0” bits and 197,822 “1” bits, with 2214 “0” bits misclassified as “1” bits and 2195 “1” bits misclassified as “0” bits. These results highlight the model’s strong bit-wise accuracy with a low misclassification rate.

4. Results

The BER performance of the proposed HDL detector was evaluated with respect to SNR under imperfect CSI conditions. A model’s learning rate balances speed and stability, determining how quickly it learns; an optimal rate accelerates convergence, prevents instability, and ensures efficient and accurate outputs [49]. Our proposed model demonstrates efficient performance across different learning rates, as shown in Figure 7a. The performance of the HDL-based detector at various learning rates was calculated with a training SNR of 15 dB and a ( 4 , 1 , 4 ) data setup. At lower SNR levels, it performs similarly for learning rates of 0.01, 0.005, and 0.002; at higher SNR levels it shows a slight advantage, with a learning rate of 0.005 compared to the others. Consequently, all experimental setups were designed with a learning rate of 0.005.
The BER performance of the proposed HDL-based model for different batch sizes is shown in Figure 7b. The results demonstrate that the model exhibits robust performance across all batch sizes. At lower SNR levels, the performance remains nearly consistent for batch sizes of 500, 1000, and 2000. However, at higher SNR levels, the model achieves slightly superior performance with a batch size of 1000.
The training SNR is critical for DL models in communication systems, as it directly impacts the model’s ability to generalize across various channel conditions. The performance of the proposed detector across various training SNR levels is illustrated in Figure 8. Notably, the model demonstrates the most stable performance compared to the other levels at a training SNR of 15 dB. Increasing or decreasing the training SNR away from this optimal point results in diminished performance. For training SNR levels of 0 dB and 10 dB, the HDL model performs well at lower SNRs but exhibits slightly weaker performance at higher SNRs. In contrast, at training SNRs of 20 dB and 25 dB the proposed model demonstrates good performance at higher SNR levels while showing comparatively weaker results at lower SNRs. Thus, the experimental results indicate that the proposed HDL-based detector achieves superior performance at a training SNR of 15 dB. Consequently, all experimental setups were designed with a training SNR of 15 dB.
In all experimental settings, the results were calculated using the ZF equalizer and Adam optimizer, except in Figure 9 and Figure 10, where BER performance is compared across different equalizers and optimizers. Additionally, all experiments in this setup utilized ( N , A , M ) = ( 4 , 1 , 4 ) except in Figure 11, which presents a comparison of model performance across different modulation orders and data setups.
Equalizers play a crucial role in communication systems by mitigating channel distortions such as ISI and multipath fading. Equalizers ensure reliable data transmission in noisy environments by improving BER and overall system performance by providing improved signal quality and dependability [50]. The BER performance of the proposed model with various equalizers is depicted in Figure 9. The results reveal that the proposed detector achieves impressive BER performance with both the ZF and MMSE equalizers, while its performance with the DFE equalizer is notably less effective. Although the model delivers nearly equivalent performance with the ZF and MMSE equalizers, it demonstrates slightly enhanced performance when employing the ZF equalizer, highlighting its efficiency in this configuration.
The BER performance of the proposed model with different optimizers is illustrated in Figure 10. Specifically, we examined the performance of the HDL model using the Adam, RMSProp, Adagrad, and Gradient Descent optimizers. The figure clearly indicates that the proposed model demonstrates acceptable performance across all optimizers. However, its performance with the Gradient Descent optimizer is somewhat less favorable compared to the others. Notably, while the HDL detector exhibits nearly comparable performance with both Adam and RMSProp, the Adam optimizer achieves a slightly superior performance in comparison to RMSProp.
Figure 11 illustrates the BER performance of the HDL detector in the OFDM-IM system with different modulation orders under imperfect channel conditions. We evaluated the performance with the QPSK, 8QAM, and 16QAM modulation schemes, where the configurations were set as follows: ( N , A , M ) = ( 4 , 1 , 4 ) for QPSK, ( 8 , 2 , 8 ) for 8QAM, and ( 8 , 4 , 16 ) for 16QAM. We observed that increasing the modulation order and data setup leads to a reduction in performance; specifically, the model achieves 5 dB and 9 dB superior performance for the ( 4 , 1 , 4 ) configuration compared to the ( 8 , 2 , 8 ) and ( 8 , 4 , 16 ) configurations, respectively.
Figure 12 compares the BER performance of the proposed HDL-based detector against ML, GD, DeepIM [27], and LSTM-IM [34] detectors. This evaluation used a 15 dB training SNR and a 0.005 learning rate along with the ZF equalizer and Adam optimizer. The results clearly demonstrate that the HDL detector consistently outperforms GD, ML, DeepIM, and LSTM-IM models with the ( 4 , 1 , 4 ) data setup. Specifically, the proposed detector shows approximately 2.5 dB improved performance over GD and achieves nearly 0.9 dB better BER performance compared to both the ML and DeepIM detectors under imperfect CSI conditions. Additionally, we compared the performance of the proposed model with the ( 4 , 3 , 4 ) data setting against a CNN-based model [31]. Table 2 presents the BER performance of all comparable models at an SNR of 20 dB, with the results illustrated in Figure 12. The simulation results confirm that the HDL model outperforms the CNN model at higher data rates.
To further demonstrate the robustness of the proposed HDL-based OFDM-IM system, we calculated the throughput and spectral efficiency (SE), as shown in Figure 13. These metrics were evaluated for the (4,1,4), (4,2,4), and (4,3,4) data configuration, with each setup representing a distinct combination of active indices. The results indicate that both throughput and SE increase with higher active indices. The model demonstrated the highest performance with (4,3,4) setup. These results confirm that our proposed HDL model effectively achieves high throughput and SE.

5. Computational Complexity

The computational complexity of a DL model can be estimated by calculating the number of floating point operations (FLOPs) required for each layer [51,52,53]. In the given network, the computational cost of the 1D-CNN layer with ReLU activation function is O ( 2 × input channels × kernel size × N 2 × F ) = O ( 576 N ) . The subsequent BN layer normalizes the feature maps, with a complexity of O ( 2 × N 2 × F ) FLOPs. Moving on to the Bi-LSTM layer, the FLOP count is provided by O ( 8 × N × Q × ( 64 + Q ) ) . The global max-pooling performs a simple pooling operation with a computational cost of O ( N 2 × 2 Q ) . Finally, the output dense layer with sigmoid activation function incurs a FLOP count of O ( ( 2 × Q × m ) + m ) . Summing these operations provides the total FLOPs for the HDL model as O ( 197376 N + 1028 ) , providing a complete picture of the computational demands. This measure is crucial for model optimization and deployment, particularly in environments with limited resources.

6. Conclusions

In this study, we investigate the BER performance of our hybrid DL (HDL) model combining a 1D-CNN and Bi-LSTM under imperfect channel conditions. By preprocessing the channel vector and received signal based on domain-specific knowledge, the HDL model effectively recovers transmitted data bits. Our evaluation included a thorough analysis of various batch sizes, learning rates, optimizers, and equalizers, finding that the HDL model achieves optimal results with the Adam optimizer and ZF equalizer. Although our proposed model demonstrates effective performance at positive training SNR values, its performance is comparatively lower at negative training SNR values, which led us to exclude such values during training. Our comparative results reveal that the HDL model outperforms GD, ML, DeepIM, CNN, and LSTM-IM detectors, demonstrating its robustness compared to both traditional and conventional models. Additionally, we explored the performance of our HDL detector across different modulation orders and data configurations, confirming that the best performance is achieved with QPSK modulation and the ( 4 , 1 , 4 ) data setup. These promising results highlight the advantages of HDL, suggesting that it can provide valuable contributions towards further integration of DL within advanced IM systems. Furthermore, to substantiate the resilience of the proposed HDL-based OFDM-IM system, we evaluated it in terms of both throughput and SE. Our findings should inspire continued research on ways to leverage DL techniques in next-generation communication systems.

Author Contributions

Conceptualization, M.A.A. and M.H.R.; methodology, M.A.A., R.T. and M.A.S.S.; software, M.A.A., M.H.R. and M.-S.B.; validation, M.A.A. and M.A.S.S.; formal analysis, M.A.A., M.H.R., R.T. and M.-S.B.; investigation, M.A.A., M.H.R. and M.A.S.S.; resources, H.-K.S.; data curation, M.A.A. and R.T.; writing—original draft preparation, M.A.A.; writing—review and editing, M.A.A. and H.-K.S.; visualization, M.A.A. and M.-S.B.; supervision, H.-K.S.; project administration, H.-K.S.; funding acquisition, H.-K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by an Institute of Information and Communications Technology Planning and Evaluation (IITP) grant funded by the Korean Government (Ministry of Science and ICT (MSIT), South Korea); in part under the Metaverse Support Program to Nurture the Best Talents, Grant IITP-2024-RS-2023-00254529; in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education under Grant 2020R1A6A1A03038540; and in part by the MSIT (Ministry of Science and ICT, Korea) under the ITRC (Information Technology Research Center) support program (IITP-2024-RS-2024-00437494) supervised by the IITP (Institute for Information and Communications Technology Planning and Evaluation).

Data Availability Statement

The data will be made available by the authors upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AWGNAdditive White Gaussian Noise
BERBit Error Rate
Bi-LSTMBidirectional Long Short-Term Memory
BNBatch Normalization
CNNConvolutional Neural Network
CSIChannel State Information
DFEDecision Feedback Equalizer
DNNDeep Neural Network
DLDeep Learning
GDGreedy Detector
HDLHybrid Deep Learning
ISIInter-Symbol Interference
MCMulti-Carrier
MMSEMinimum Mean Square Error
MLMaximum Likelihood
OFDM-IMOrthogonal Frequency-Division Multiplexing with Index Modulation
16QAM16-Quadrature Amplitude Modulation
QPSKQuadrature Phase-Shift Keying
ReLURectified Linear Units
SNRSignal-to-Noise Ratio
SESpectral Efficiency
TanhHyperbolic Tangent
ZFZero-Forcing

References

  1. Bingham, J.A.C. Multicarrier Modulation for Data Transmission: An Idea Whose Time Has Come. IEEE Commun. Mag. 1990, 28, 5–14. [Google Scholar] [CrossRef]
  2. Taura, K.; Tsujishita, M.; Takeda, M.; Kato, H.; Ishida, M.; Ishida, Y. A digital audio broadcasting (DAB) receiver. IEEE Trans. Consum. Electron. 1996, 42, 322–327. [Google Scholar] [CrossRef]
  3. Telecommunication Standardization Sector of ITU. Asymmetric Digital Subscriber Line (ADSL) Transceivers; Telecommunication Standardization Sector of ITU: Geneva, Switzerland, 1999. [Google Scholar]
  4. Nee, R.v.; Prasad, R. OFDM for Wireless Multimedia Communications; Artech House, Inc.: London, UK, 2000. [Google Scholar]
  5. Biglieri, E. Coding and modulation for a horrible channel. IEEE Commun. Mag. 2003, 41, 92–98. [Google Scholar] [CrossRef]
  6. Garg, V.K. Fourth generation systems and new wireless technologies. In Wireless Communications & Networking; Elsevier: Amsterdam, The Netherlands, 2007; pp. 1–22. [Google Scholar]
  7. Cioffi, J.M. A multicarrier primer. ANSI T1E1 1991, 4, 91–157. [Google Scholar]
  8. Terry, J.; Heiskala, J. OFDM Wireless LANs: A Theoretical and Practical Guide; Sams Publishing: Carmel, IN, USA, 2002. [Google Scholar]
  9. Bury, A.; Egle, J.; Lindner, J. Diversity comparison of spreading transforms for multicarrier spread spectrum transmission. IEEE Trans. Commun. 2003, 51, 774–781. [Google Scholar] [CrossRef]
  10. Mao, T.; Wang, Z.; Wang, Q.; Chen, S.; Hanzo, L. Dual-mode index modulation aided OFDM. IEEE Access 2016, 5, 50–60. [Google Scholar] [CrossRef]
  11. Başar, E.; Aygölü, Ü.; Panayırcı, E.; Poor, H.V. Orthogonal frequency division multiplexing with index modulation. IEEE Trans. Signal Process. 2013, 61, 5536–5549. [Google Scholar] [CrossRef]
  12. Wen, M.; Cheng, X.; Ma, M.; Jiao, B.; Poor, H.V. On the achievable rate of OFDM with index modulation. IEEE Trans. Signal Process. 2015, 64, 1919–1932. [Google Scholar] [CrossRef]
  13. Mao, T.; Wang, Q.; Wang, Z.; Chen, S. Novel index modulation techniques: A survey. IEEE Commun. Surv. Tutor. 2018, 21, 315–348. [Google Scholar] [CrossRef]
  14. Abu-Alhiga, R.; Haas, H. Subcarrier-index modulation OFDM. In Proceedings of the 2009 IEEE 20th International Symposium on Personal, Indoor and Mobile Radio Communications, Tokyo, Japan, 13–16 September 2009; pp. 177–181. [Google Scholar]
  15. Fan, R.; Yu, Y.J.; Guan, Y.L. Generalization of orthogonal frequency division multiplexing with index modulation. IEEE Trans. Wirel. Commun. 2015, 14, 5350–5359. [Google Scholar] [CrossRef]
  16. Choi, J. Noncoherent OFDM-IM and its performance analysis. IEEE Trans. Wirel. Commun. 2017, 17, 352–360. [Google Scholar] [CrossRef]
  17. Başar, E.; Aygölü, Ü.; Panaylrcl, E. Orthogonal frequency division multiplexing with index modulation in the presence of high mobility. In Proceedings of the 2013 IEEE First International Black Sea Conference on Communications and Networking (BlackSeaCom), Batumi, Georgia, 3–5 July 2013; pp. 147–151. [Google Scholar]
  18. Siddiq, A.I. Low complexity OFDM-IM detector by encoding all possible subcarrier activation patterns. IEEE Commun. Lett. 2016, 20, 446–449. [Google Scholar] [CrossRef]
  19. Crawford, J.; Ko, Y. Low complexity greedy detection method with generalized multicarrier index keying OFDM. In Proceedings of the 2015 IEEE 26th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), Hong Kong, China, 30 August–2 September 2015; pp. 688–693. [Google Scholar]
  20. Basar, E. Index modulation techniques for 5G wireless networks. IEEE Commun. Mag. 2016, 54, 168–175. [Google Scholar] [CrossRef]
  21. Zappone, A.; Di Renzo, M.; Debbah, M. Wireless networks design in the era of deep learning: Model-based, AI-based, or both? IEEE Trans. Commun. 2019, 67, 7331–7376. [Google Scholar] [CrossRef]
  22. Erpek, T.; O’Shea, T.J.; Sagduyu, Y.E.; Shi, Y.; Clancy, T.C. Deep learning for wireless communications. In Development and Analysis of Deep Learning Architectures; Springer: Berlin/Heidelberg, Germany, 2020; pp. 223–266. [Google Scholar]
  23. Sejan, M.A.S.; Rahman, M.H.; Aziz, M.A.; Kim, D.S.; You, Y.H.; Song, H.K. A comprehensive survey on MIMO visible light communication: Current research, machine learning and future trends. Sensors 2023, 23, 739. [Google Scholar] [CrossRef]
  24. Dhanke, S.S.; Sharma, S.; Kumar, A.; Mandloi, M. Index Modulation Multiple Access via Deep Learning based Detection. In Proceedings of the 2020 IEEE 3rd 5G World Forum (5GWF), Bangalore, India, 10–12 September 2020; pp. 312–317. [Google Scholar]
  25. Esmaiel, H.; Leftah, H.A.; Junejo, N.U.R.; Sun, H. Deep Learning-Based Index Modulation for Underground Communications. IEEE Open J. Commun. Soc. 2023, 4, 2122–2132. [Google Scholar] [CrossRef]
  26. Liu, H.; Zhang, Y.; Zhang, X.; El-Hajjar, M.; Yang, L.L. Deep learning assisted adaptive index modulation for mmWave communications with channel estimation. IEEE Trans. Veh. Technol. 2022, 71, 9186–9201. [Google Scholar] [CrossRef]
  27. Van Luong, T.; Ko, Y.; Vien, N.A.; Nguyen, D.H.; Matthaiou, M. Deep learning-based detector for OFDM-IM. IEEE Wirel. Commun. Lett. 2019, 8, 1159–1162. [Google Scholar] [CrossRef]
  28. Kim, J.; Ro, H.; Park, H. Deep learning-based detector for dual mode OFDM with index modulation. IEEE Wirel. Commun. Lett. 2021, 10, 1562–1566. [Google Scholar] [CrossRef]
  29. Kim, J.; Park, H. Deep Learning-Based Detector With Modulation Parameter-Independent Structure for OFDM With Index Modulation. IEEE Access 2023, 11, 130358–130367. [Google Scholar] [CrossRef]
  30. Le-Tran, M.; Kim, S. Deep Learning-Assisted Index estimator for generalized LED index modulation OFDM in visible light communication. Photonics 2021, 8, 168. [Google Scholar] [CrossRef]
  31. Wang, T.; Yang, F.; Song, J.; Han, Z. Deep convolutional neural network-based detector for index modulation. IEEE Wirel. Commun. Lett. 2020, 9, 1705–1709. [Google Scholar] [CrossRef]
  32. Liu, J.; Lu, H. IMNet: A learning based detector for index modulation aided MIMO-OFDM systems. In Proceedings of the 2020 IEEE Wireless Communications and Networking Conference (WCNC), Seoul, Republic of Korea, 25–28 May 2020; pp. 1–6. [Google Scholar]
  33. Zhang, D.; Wang, S.; Niu, K.; Dai, J.; Wang, S.; Yuan, Y. Transformer-based detector for OFDM with index modulation. IEEE Commun. Lett. 2022, 26, 1313–1317. [Google Scholar] [CrossRef]
  34. Aziz, M.A.; Rahman, M.H.; Sejan, M.A.S.; Tabassum, R.; Hwang, D.d.; Song, H.K. Deep Recurrent Neural Network Based Detector for OFDM with Index Modulation. IEEE Access 2024, 12, 89538–89547. [Google Scholar] [CrossRef]
  35. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  36. Ratnam, D.V.; Rao, K.N. Bi-LSTM based deep learning method for 5G signal detection and channel estimation. AIMS Electron. Electr. Eng. 2021, 5, 334–341. [Google Scholar] [CrossRef]
  37. Aziz, M.A.; Rahman, M.H.; Sejan, M.A.S.; Tabassum, R.; Hwang, D.d.; Song, H.K. Bidirectional Deep Learning Decoder for Polar Codes in Flat Fading Channels. IEEE Access 2024, 12, 149580–149592. [Google Scholar] [CrossRef]
  38. Jiang, Y.; Varanasi, M.K.; Li, J. Performance analysis of ZF and MMSE equalizers for MIMO systems: An in-depth study of the high SNR regime. IEEE Trans. Inf. Theory 2011, 57, 2008–2026. [Google Scholar] [CrossRef]
  39. Hanoon, S.S.; Khazal, H.F.; Jamel, T.M. A Deep Learning Approach to Evaluating SISO-OFDM Channel Equalization. Wasit J. Eng. Sci. 2024, 12, 110–124. [Google Scholar] [CrossRef]
  40. Belfiore, C.A.; Park, J.H. Decision feedback equalization. Proc. IEEE 1979, 67, 1143–1156. [Google Scholar] [CrossRef]
  41. Li, H.; Li, Y.; Zhou, S.; Wang, J. Wireless channel feature extraction via GMM and CNN in the tomographic channel model. J. Commun. Inf. Netw. 2017, 2, 41–51. [Google Scholar] [CrossRef]
  42. Bjorck, N.; Gomes, C.P.; Selman, B.; Weinberger, K.Q. Understanding batch normalization. arXiv 2018, arXiv:1806.02375. [Google Scholar]
  43. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef]
  44. Aziz, M.A.; Rahman, M.H.; Sejan, M.A.S.; Baik, J.I.; Kim, D.S.; Song, H.K. Spectral Efficiency Improvement Using Bi-Deep Learning Model for IRS-Assisted MU-MISO Communication System. Sensors 2023, 23, 7793. [Google Scholar] [CrossRef]
  45. Yu, D.; Wang, H.; Chen, P.; Wei, Z. Mixed pooling for convolutional neural networks. In Rough Sets and Knowledge Technology, Proceedings of the 9th International Conference, RSKT 2014, Shanghai, China, 24–26 October 2014, Proceedings 9; Springer: Berlin/Heidelber, Germany, 2014; pp. 364–375. [Google Scholar]
  46. Narayan, S. The generalized sigmoid activation function: Competitive supervised learning. Inf. Sci. 1997, 99, 69–82. [Google Scholar] [CrossRef]
  47. Hodson, T.O.; Over, T.M.; Foks, S.S. Mean squared error, deconstructed. J. Adv. Model. Earth Syst. 2021, 13, e2021MS002681. [Google Scholar] [CrossRef]
  48. Zhang, Z. Improved adam optimizer for deep neural networks. In Proceedings of the 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS), Banff, AB, Canada, 4–6 June 2018; pp. 1–2. [Google Scholar]
  49. You, K.; Long, M.; Wang, J.; Jordan, M.I. How does learning rate decay help modern neural networks? arXiv 2019, arXiv:1908.01878. [Google Scholar]
  50. Ahmed, S.; Khan, Y.; Wahab, A. A review on training and blind equalization algorithms for wireless communications. Wirel. Pers. Commun. 2019, 108, 1759–1783. [Google Scholar] [CrossRef]
  51. Brownlee, J. A gentle introduction to pooling layers for convolutional neural networks. Mach. Learn. Mastery 2019, 22, 1–16. [Google Scholar]
  52. Zhang, M.; Wang, W.; Liu, X.; Gao, J.; He, Y. Navigating with graph representations for fast and scalable decoding of neural language models. arXiv 2018, arXiv:1806.04189. [Google Scholar]
  53. Chaudhary, N.; Misra, S.; Kalamkar, D.; Heinecke, A.; Georganas, E.; Ziv, B.; Adelman, M.; Kaul, B. Efficient and generic 1d dilated convolution layer for deep learning. arXiv 2021, arXiv:2104.08002. [Google Scholar]
Figure 1. Generalized data transmission process for an OFDM-IM system.
Figure 1. Generalized data transmission process for an OFDM-IM system.
Mathematics 12 03583 g001
Figure 2. Structure of the proposed HDL detector for OFDM-IM systems.
Figure 2. Structure of the proposed HDL detector for OFDM-IM systems.
Mathematics 12 03583 g002
Figure 3. The internal configuration of an LSTM cell.
Figure 3. The internal configuration of an LSTM cell.
Mathematics 12 03583 g003
Figure 4. Training loss of the proposed HDL model for different equalizers with data setup ( N , A , M ) = ( 4 , 1 , 4 ) : (a) training loss for the ZF equalizer, (b) training loss for the MMSE equalizer, and (c) training loss for the DFE equalizer.
Figure 4. Training loss of the proposed HDL model for different equalizers with data setup ( N , A , M ) = ( 4 , 1 , 4 ) : (a) training loss for the ZF equalizer, (b) training loss for the MMSE equalizer, and (c) training loss for the DFE equalizer.
Mathematics 12 03583 g004
Figure 5. Training loss of the proposed HDL model for different modulation orders and data combinations with the ZF equalizer: (a) training loss for the ( N , A , M ) = ( 4 , 1 , 4 ) setup, (b) training loss for the ( N , A , M ) = ( 8 , 2 , 8 ) setup, and (c) training loss for the ( N , A , M ) = ( 8 , 4 , 16 ) setup.
Figure 5. Training loss of the proposed HDL model for different modulation orders and data combinations with the ZF equalizer: (a) training loss for the ( N , A , M ) = ( 4 , 1 , 4 ) setup, (b) training loss for the ( N , A , M ) = ( 8 , 2 , 8 ) setup, and (c) training loss for the ( N , A , M ) = ( 8 , 4 , 16 ) setup.
Mathematics 12 03583 g005
Figure 6. The confusion matrix of the proposed HDL-based model.
Figure 6. The confusion matrix of the proposed HDL-based model.
Mathematics 12 03583 g006
Figure 7. Performance of the HDL-based detector with (a) different learning rates and (b) different batch sizes in for the ( 4 , 1 , 4 ) data combination.
Figure 7. Performance of the HDL-based detector with (a) different learning rates and (b) different batch sizes in for the ( 4 , 1 , 4 ) data combination.
Mathematics 12 03583 g007
Figure 8. Performance of the HDL-based detector at various training SNRs for the ( 4 , 1 , 4 ) data configuration.
Figure 8. Performance of the HDL-based detector at various training SNRs for the ( 4 , 1 , 4 ) data configuration.
Mathematics 12 03583 g008
Figure 9. Performance of the proposed HDL-based detector with various equalizers for the ( 4 , 1 , 4 ) data configuration.
Figure 9. Performance of the proposed HDL-based detector with various equalizers for the ( 4 , 1 , 4 ) data configuration.
Mathematics 12 03583 g009
Figure 10. BER performance of the proposed HDL-based detector utilizing different optimizers for the ( 4 , 1 , 4 ) data configuration.
Figure 10. BER performance of the proposed HDL-based detector utilizing different optimizers for the ( 4 , 1 , 4 ) data configuration.
Mathematics 12 03583 g010
Figure 11. BER performance of the proposed HDL-based detector for various modulation orders and data setup.
Figure 11. BER performance of the proposed HDL-based detector for various modulation orders and data setup.
Mathematics 12 03583 g011
Figure 12. BER performance comparison of the proposed HDL-based detector with other detectors under imperfect CSI conditions for the ( 4 , 1 , 4 ) data combinations.
Figure 12. BER performance comparison of the proposed HDL-based detector with other detectors under imperfect CSI conditions for the ( 4 , 1 , 4 ) data combinations.
Mathematics 12 03583 g012
Figure 13. Throughput and SE of the proposed HDL-based OFDM-IM system: (a) throughput performance and (b) SE performance.
Figure 13. Throughput and SE of the proposed HDL-based OFDM-IM system: (a) throughput performance and (b) SE performance.
Mathematics 12 03583 g013
Table 1. Simulation parameters for the proposed system.
Table 1. Simulation parameters for the proposed system.
ParametersValue
Training samples10,000,000
Testing samples100,000
Modulation schemeQPSK, 8QAM, 16QAM
Noise typeAWGN
Fading channelRayleigh fading channel
Number of hidden layers128
Learning rate0.01, 0.005, 0.002
Number of epoch1000
Batch size1000
Training SNR (dB)0, 5, 10, 15, 20, 25
N4, 8
A1, 2, 3, 4
OptimizerAdam, RMSProp, Adagrad, Gradient Descent
Table 2. Comparative BER performance at 20 dB SNR for all models.
Table 2. Comparative BER performance at 20 dB SNR for all models.
SNRModel NameBERModel NameBER
20 dBHDL0.0033074783DeeIM0.00387089132682052
LSTM-IM0.00355340813397947ML0.00392567922218776
HDL0.0074599785 (A = 3)CNN0.00854427859235968 (A = 3)
GD0.00585573211245060
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aziz, M.A.; Rahman, M.H.; Tabassum, R.; Sejan, M.A.S.; Baek, M.-S.; Song, H.-K. A Hybrid Deep Learning Framework for OFDM with Index Modulation Under Uncertain Channel Conditions. Mathematics 2024, 12, 3583. https://doi.org/10.3390/math12223583

AMA Style

Aziz MA, Rahman MH, Tabassum R, Sejan MAS, Baek M-S, Song H-K. A Hybrid Deep Learning Framework for OFDM with Index Modulation Under Uncertain Channel Conditions. Mathematics. 2024; 12(22):3583. https://doi.org/10.3390/math12223583

Chicago/Turabian Style

Aziz, Md Abdul, Md Habibur Rahman, Rana Tabassum, Mohammad Abrar Shakil Sejan, Myung-Sun Baek, and Hyoung-Kyu Song. 2024. "A Hybrid Deep Learning Framework for OFDM with Index Modulation Under Uncertain Channel Conditions" Mathematics 12, no. 22: 3583. https://doi.org/10.3390/math12223583

APA Style

Aziz, M. A., Rahman, M. H., Tabassum, R., Sejan, M. A. S., Baek, M. -S., & Song, H. -K. (2024). A Hybrid Deep Learning Framework for OFDM with Index Modulation Under Uncertain Channel Conditions. Mathematics, 12(22), 3583. https://doi.org/10.3390/math12223583

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop