3.1. Resource Grouping Pattern
The channel estimation based on the K-means algorithm identifies the centroids of each cluster, effectively counteracting signal distortion. However, this channel estimation method must adequately account for significant channel variability along the time and frequency domains to perform clustering effectively. For instance, if the channel varies rapidly in the time domain due to high Doppler effects, the groups must be densely configured along the time domain. Similarly, if the channel exhibits rapid variation in the frequency domain due to significant delay spread, the groups must be densely configured with single or multiple subcarriers along the frequency domain. Such clustering-based K-means algorithm channel estimation cannot be optimally performed for all UEs. In this paper, to address this limitation, we propose a method that generates resource group patterns within the resource grid for received signals to enable the K-means-based blind channel estimation to operate effectively by identifying the characteristics of the dynamically changing channel. The resource group pattern consists of consecutive OFDM symbols in the time domain and consecutive subcarriers in the frequency domain. The number of consecutive OFDM symbols, , can range from {1, 2, …, 14}. If is not a divisor of 14, which is the total number of OFDM symbols in a single slot, the remaining OFDM symbols that do not satisfy this condition are grouped into the maximum possible number of consecutive OFDM symbols.
In this paper, such groups are defined as remainder groups. Additionally, the number of consecutive subcarriers,
, can take any value among the divisors of 120 within 10 RBs. The divisors of 120,
are expressed in Equation (6) and are listed as {1, 2, 3, 4, 5, 6, 8, 10, 12, 15, 20, 24, 30, 40, 60, 120}, for a total of 16 values:
In other words, the resource grouping patterns for the received signal can consider 14 cases in the time domain and 16 cases in the frequency domain, resulting in a total of 224 possible patterns. For instance, in a pattern where and within 1 slot in the time domain and 10RBs in the frequency domain, one resource group consists of 7 consecutive OFDM symbols in the time domain and 60 consecutive subcarriers in the frequency domain.
As shown in
Figure 2, the clustering pattern consists of a total of four groups, and K-means clustering-based channel estimation is performed for each group within the pattern. In
Figure 2, within the resource grid corresponding to one slot in the time domain and 10 RB in the frequency domain, each group is distinguished by different colors, and K-means clustering-based channel estimation is executed for each group. Channel estimation is performed for a single resource group, as shown by the red dashed lines. The estimated channel values are represented as
, and each resource group is estimated to have the same channel.
The generated resource groups must identify QPSK symbols through K-means-based channel estimation, requiring at least four REs within each group. In this paper, this requirement is defined as the minimum RE count condition. Patterns that do not satisfy this condition are excluded. For example, in the pattern where and , a single group consists of two consecutive OFDM symbols in the time domain and one subcarrier in the frequency domain, resulting in a total of 840 groups. Since each group in this pattern contains only two REs, channel estimation using the K-means algorithm cannot be performed. Therefore, this pattern is excluded for failing to meet the minimum RE count condition.
Another case is the pattern where and , in which a single group consists of one OFDM symbol in the time domain and three subcarriers in the frequency domain, resulting in a total of 560 groups for the entire pattern. Since the REs within each group of this pattern are four or fewer, it is excluded. Additionally, the minimum RE count condition is equally applied to the remainder groups. For example, in the grouping pattern where and , a resource group consists of six consecutive OFDM symbols in the time domain and a single subcarrier in the frequency domain. In this case, the 13th and 14th OFDM symbols in the time domain combined with a single subcarrier in the frequency domain form the remainder group, which contains only two REs and, thus, fails to meet the minimum RE count condition. Consequently, this resource grouping pattern is also excluded. Therefore, considering and , 212 patterns can be configured out of the 224 resource group patterns, excluding those that do not meet the minimum resource count condition.
3.2. Data Set Generation
This paper evaluates channel estimation performance by supporting optimal resource groups for all UEs. To this end, among all configured resource grouping patterns, the pattern that achieves the highest channel estimation performance in arbitrary channel environments is derived based on Equation (5) and utilized as the correct label for the CNN. The channel environment considers factors such as the UE’s velocity and delay spread. For instance, when the UE’s velocity is 30 km/h, the channel environment is relatively stable, making a pattern composed of resource groups with many OFDM symbols effective. Conversely, when the velocity is 150 km/h, a pattern composed of resource groups with a smaller number of OFDM symbols will be more effective.
The delay spread is considered from 50 ns to 300 ns in increments of 50 ns, and velocity is considered from 30 km/h to 150 km/h in increments of 30 km/h, resulting in 30 channel conditions. The SNR is assumed to be perfectly calculated. The parameters used for dataset generation, including these, are summarized in
Table 1 [
16].
The K-means algorithm is employed for channel estimation without relying on pilot signals. In this study, an exhaustive search is performed to identify the optimal pattern among the 212 proposed resource grouping patterns for each of the 30 considered channel conditions using the K-means algorithm. The MSE for each pattern is computed using Equation (5), and the pattern with the best performance is selected as the optimal pattern for the corresponding channel condition. The identified optimal patterns are used to generate correct labels for each step of the two-step training process, which is detailed in
Section 3.3. Subsequently, a CNN is employed to effectively group the data according to various channel conditions, including delay spread, velocity, and SNR. The CNN is trained to learn the optimal grouping method for each channel condition and outputs the corresponding grouping pattern for each received signal, thereby enabling more accurate channel estimation. K-means algorithm-based channel estimation requires a sufficient number of data samples to achieve accurate channel estimation, and the minimum number of required data samples for channel estimation may vary with the SNR level. When performing K-means algorithm-based channel estimation under low SNR conditions, a relatively larger number of resource grouping patterns from the 212 proposed in this study have proven to be effective. In contrast, at higher SNR levels, patterns with a relatively smaller number of resources are utilized for channel estimation. The proposed 212 patterns can be effective depending not only on the SNR but also on the delay spread, which affects channel variations in the frequency domain, and the velocity, which impacts the time domain.
Therefore, in this study, the optimal pattern obtained through an exhaustive search of the proposed 212 patterns is analyzed to examine the influence of delay spread, velocity, and SNR on channel estimation performance, as illustrated in
Figure 3,
Figure 4,
Figure 5 and
Figure 6. The 212 patterns proposed in this paper are grouped by each channel parameter and presented in
Figure 3,
Figure 4,
Figure 5 and
Figure 6 to illustrate the distribution of the optimal patterns.
Figure 3 shows the distribution of the optimal number of OFDM symbols as
, which is the parameter determining the pattern in the time domain under different velocity conditions. The x-axis in each graph, labeled “Number of OFDM symbols per group”, represents the optimal
. In
Figure 3, as the velocity of the UE increases, channel state variations in the time domain become more frequent and rapid, resulting in smaller values of
appearing more frequently at higher velocities. This indicates that as velocity increases, the number of OFDM symbols per group decreases, necessitating channel estimation at smaller group sizes. Consequently, for relatively low velocities, such as 30 km/h, patterns with larger values of
per group are determined to be optimal. In contrast, at higher velocities, such as 150 km/h, smaller group sizes are more suitable, leading to patterns with fewer OFDM symbols per group being identified as optimal.
Figure 4 illustrates the distribution of the optimal number of subcarriers,
, which is the parameter that determines the resource grouping pattern in the frequency domain under different delay spread conditions. This is represented on the
x-axis of each graph.
In
Figure 4, as the delay spread increases, channel variations in the frequency domain occur more rapidly and irregularly, requiring smaller values of
per group for channel estimation in the frequency domain. For a low delay spread, such as 50 ns, forming groups is frequently determined as optimal patterns. In contrast, for a high delay spread, such as 300 ns, the group size becomes smaller to respond more finely to channel variations, and smaller values of
are more widely distributed. This indicates that creating groups with smaller units is more effective under conditions of high delay spread, emphasizing the necessity of finer resource grouping in the frequency domain as the delay spread increases.
Figure 3 and
Figure 4 demonstrate that resource grouping into smaller units is effective under relatively high levels of velocity and delay spread. However, a substantial proportion of cases were identified where larger units of
and
were optimal, which shows that the
with a value of 120 was the most frequently identified optimal value. Because the channel environments considered in this study include both low and high SNR levels, channel estimation based on the K-means algorithm requires an appropriate amount of data to achieve sufficient channel estimation performance based on the SNR level. This requirement increases as the SNR decreases, requiring more data. To meet this requirement,
, which offers more data points, is more effective than
in terms of data quantity. Consequently, when performing K-means-based channel estimation under high SNR conditions, smaller units of
and
are determined while satisfying the relatively small data requirements.
The optimal data requirements for channel estimation under varying SNR levels are shown in
Figure 5 and
Figure 6.
Figure 5 illustrates the optimal number of data counts per group at SNR levels of −6, −4, −2, and 0 dB, while
Figure 6 shows the corresponding values at SNR levels of 2, 4, 6, 8, and 10 dB. In low-SNR conditions, relatively large data quantities, such as 960, 1080, 1200, and 1680, were detected to ensure channel estimation performance. In contrast, under high-SNR conditions, smaller data quantities were more broadly distributed. As shown in
Figure 3,
Figure 4,
Figure 5 and
Figure 6, the analysis of optimal patterns demonstrates that appropriate resource grouping in the time and frequency domains is essential for K-means-based channel estimation methods, considering channel factors such as UE velocity, delay spread, and SNR. Furthermore, it was confirmed that the amount of data required to ensure the performance of K-means-based channel estimation must also be considered.
The input data for the CNN is generated based on simulations. The two-step CNN training process is conducted using the final optimal patterns derived from exhaustive search, characterized by the frequency domain parameter and the time domain parameter . Thus, each step involves distinct input data and correct labels. For instance, during training in the frequency domain, the optimal patterns are classified based on the number of subcarriers within the patterns. The grouped patterns are then defined with as the correct label. The input data is generated based on the channel conditions associated with the patterns sharing the same correct label. Single or multiple patterns with the defined correct label have associated channel conditions, such as delay spread, velocity, and SNR. These channel conditions are used to generate a tapped-delay line (TDL)-D channel. The channel generation process is repeated times, and for each iteration, the channel conditions are randomly selected from the patterns with the same , following a discrete uniform distribution. The transmitted signal passes through the generated TDL-D channel, and noise is added to produce , as expressed in Equation (1).
The received signals are processed as input data for the CNN by separating the real and imaginary components and adjusting the size to . The values are normalized between −1 and 1. In this study, is set to 100,000. The generated dataset of 100,000 samples is divided into training, validation, and test datasets in proportions of 0.8, 0.1, and 0.1, respectively. The training dataset serves as input data for the CNN.
3.3. Optimal Resource Grouping Pattern Selection
To reduce the complexity of selecting the optimal resource grouping pattern and to efficiently extract features, a two-step learning process is performed by dividing the proposed 212 resource grouping patterns into frequency and time domains. In the first step, learning is focused on selecting the optimal parameter , which is used to determine the resource grouping pattern in the frequency domain among the proposed resource grouping patterns. In the second step, the learning is focused on the parameter , which determines the resource grouping pattern in the time domain. After completing the learning process, the trained CNN combines the optimal values of and to output the optimal resource grouping pattern within the given channel condition.
In the first step, since the network focuses solely on the frequency domain, learning is performed based exclusively on the number of subcarriers. First, the 212 resource grouping patterns are classified into groups that have the same number of subcarriers, . Then, for the resource grouping patterns that use the same number of subcarriers, the corresponding subcarrier count, , is defined as the new correct label for CNN training. The dataset is normalized to values between −1 and 1 under the channel conditions corresponding to this correct value, with a size of , and the network is constructed accordingly. After completing the first-step of training, the CNN outputs the optimal . In the second step, since the network focuses solely on the time domain, learning focuses exclusively on the number of OFDM symbols in the proposed resource grouping patterns. First, the proposed patterns are grouped according to the same number of OFDM symbols, . Then, for the patterns with the same number of OFDM symbols, the corresponding number, , is defined as the new correct label for the second step of CNN training. For the channel conditions corresponding to this correct label, the dataset is normalized to values between −1 and 1, with a size of and the network is constructed accordingly. Upon completing the second stage of training, the CNN outputs the optimal , which is combined with the optimal from the first stage to determine the optimal resource grouping pattern for the given channel environment.
3.4. Proposed CNN Structure
The proposed CNN architecture in this study is illustrated in
Figure 7. As shown in
Figure 7, the CNN consists of an input layer and 6 convolutional (Conv) layers. The input layer has dimensions of
, where 1680 represents the total number of resources in the resource grid considered in this study, 2 corresponds to the real and imaginary parts of the received signal, and
denotes the total number of data samples. The network includes batch normalization (BN), rectified linear unit (ReLU) activation functions, and max pooling (MaxPool) layers, which are sequentially arranged in the structure.
The filter size for the convolutional layers is set to
, and the number of filters gradually increases as the layers deepen, specifically to 16, 24, 32, 48, 64, and 96 filters. This incremental increase in the number of filters is designed to enable the CNN to extract progressively complex features, aligning with prior studies [
17,
18,
19], which have shown that increasing the number of filters enhances data representation capabilities and learning accuracy in large-scale MIMO and mmWave systems. Furthermore, prior examples in OFDM systems demonstrate the effectiveness of this design in enabling the CNN to learn temporal and frequency-domain features effectively [
20,
21,
22]. The increase in the number of filters in the CNN layer design is a critical criterion, ensuring efficient data processing in bandwidth- and noise-intensive channel environments. This approach maximizes the feature extraction capability, particularly in complex channel conditions [
23,
24]. The filter size is optimized for effectively learning long time-series data, and previous studies have reported that such designs contribute to improving the SNR [
25,
26].
Batch normalization and ReLU activation functions are employed as key components to maintain training stability and prevent overfitting. Following each convolutional layer, a MaxPool layer with a size of
is used to reduce the data dimensions, thereby decreasing the computational load while retaining essential features [
27,
28]. This multi-layer structure is designed to enable the CNN to efficiently learn spatial and temporal features.