Next Article in Journal
An Ensemble CNOP Method Based on a Pre-Screening Mechanism for Targeted Observations in the South China Sea
Next Article in Special Issue
Anti-Eavesdropping by Exploiting the Space–Time Coupling in UANs
Previous Article in Journal
Pile Driving and the Setup Effect and Underlying Mechanism for Different Pile Types in Calcareous Sand Foundations
Previous Article in Special Issue
Large-Scale Long-Term Prediction of Ship AIS Tracks via Linear Networks with a Look-Back Window Decomposition Scheme of Time Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Underwater Acoustic Channel Estimation Method Based on Bias-Free Convolutional Neural Network

1
State Key Laboratory of Acoustics, Institute of Acoustics, Chinese Academy of Sciences, Beijing 100190, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Unité Mixte de Recherche, Institut de Chimie Moléculaire, Centre National de la Recherche Scientifique 6302, Université Bourgogne Franche-Comté, 21078 Dijon, France
4
Laboratoire d’Etude de l’Apprentissage et du Développement, Unité Mixte de Recherche, Centre National de la Recherche Scientifique 5022, Université Bourgogne Franche-Comté, 21078 Dijon, France
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(1), 134; https://doi.org/10.3390/jmse12010134
Submission received: 23 November 2023 / Revised: 10 December 2023 / Accepted: 30 December 2023 / Published: 9 January 2024
(This article belongs to the Special Issue Underwater Acoustic Communication and Network)

Abstract

:
In recent years, the study of deep learning techniques for underwater acoustic channel estimation has gained widespread attention. However, existing neural network channel estimation methods often overfit to training dataset noise levels, leading to diminished performance when confronted with new noise levels. In this research, a “bias-free” denoising convolutional neural network (DnCNN) method is proposed for robust underwater acoustic channel estimation. The paper offers a theoretical justification for bias removal and customizes the fundamental DnCNN framework to give a specialized design for channel estimation, referred to as the bias-free complex DnCNN (BF-CDN). It uses least squares channel estimation results as input and employs a CNN model to learn channel characteristics and noise distribution. The proposed method effectively utilizes the temporal correlation inherent in underwater acoustic channels to further enhance estimation performance and robustness. This method adapts to varying noise levels in underwater environments. Experimental results show the robustness of the method under different noise conditions, indicating its potential to improve the accuracy and reliability of channel estimation.

1. Introduction

Underwater acoustic (UWA) communication systems have become a critical component to meet the rising demand for marine exploration and commercial activities. These systems facilitate a diverse set of applications, including oceanographic research, offshore oil and gas exploration, and the Internet of Underwater Things [1,2,3]. The UWA wave is often regarded as one of the most demanding communication mediums as it is characterized by limited bandwidth, significant multipath spread, fast fading, and intricate oceanic noise [4,5].
Orthogonal frequency division multiplexing (OFDM) for UWA communication has gained substantial attention due to its high spectral efficiency and resistance to long multipath spread in UWA channels [6]. The basic idea behind OFDM is to split the channel bandwidth into evenly spaced subchannels in the frequency domain. Given the challenges posed by frequency-selective fading and substantial noise fluctuations in certain UWA environments, accurate channel estimation is crucial for a UWA OFDM communication system. Traditional channel estimation methods, such as the least squares (LS) estimation algorithm, suffer from poor performance in low signal-to-noise ratio (SNR) conditions, as the estimated mean square error (MSE) is inversely proportional to the SNR. In comparison to LS estimation, minimum mean square error (MMSE) estimation utilizing second-order statistics of the channel can achieve higher estimation accuracy. However, it requires prior knowledge of noise and channel statistics, and its complexity is much higher than that of LS estimation.
In recent years, machine learning, especially deep learning (DL), has rapidly progressed and been widely applied in various fields, offering new solutions to the challenges in UWA communication [7,8,9]. Zhang et al. [10] developed a UWA OFDM communication receiver system based on a five-layer fully connected deep neural network (DNN) for UWA channel estimation and equalization. The model demonstrated advantages over traditional algorithms, especially in scenarios with limited pilot subcarriers in OFDM communication. In a subsequent work, Zhang et al. [11] introduced a UWA OFDM communication receiver system utilizing a combination of a convolutional neural network (CNN) for feature extraction and multilayer perceptron (MLP) for data symbol recovery. This model outperforms traditional methods and fully connected DNN-based UWA OFDM frameworks. Liu et al. [12] proposed a CNN-based UWA OFDM receiver system that effectively integrates channel estimation and equalization. The design features an encoder and decoder structure with convolutional layers for feature extraction and signal reconstruction, which reduces network complexity. Qiao et al. [13] introduced CsiPreNet, a learning model comprising a one-dimensional CNN and long short-term memory (LSTM) network. This model captures the temporal and spectral characteristics of UWA channel state information (CSI) and outperforms existing recursive least square (RLS) predictors. Ouyang et al. [14] modified a super-resolution neural network to address the channel estimation problem, resulting in the channel super-resolution network (CSRNet). Simulation results showed superior performance compared to LS. Liu et al. [15] introduced a method for UWA channel estimation based on a denoising sparsity-aware DNN (DeSA-DNN). Their approach uses DNN to simulate the iterative process of classical sparse reconstruction algorithms, leveraging the sparsity of UWA channels. It incorporates an effective denoising module using CNN to mitigate the impact of interference on channel estimation.
In summary, CNN-based methods offer simplicity and flexibility in adapting to the characteristics of UWA channels. However, existing CNN-based channel estimation approaches often focus on single data-block estimations and lack consideration of the temporal correlation within the channel. Moreover, these methods are typically trained within specific SNR ranges, while UWA channels often exhibit significant SNR fluctuations. This can potentially lead to overfitting within the training SNR range and a decrease in performance outside of it.
In this paper, a robust underwater acoustic channel estimation method based on a bias-free CNN is introduced. The main contributions of this research can be summarized as follows:
  • We incorporate the “bias-free” concept [16] into denoising convolutional neural network (DnCNN) enhances the stability of the model performance and aims to overcome overfitting the training SNR conditions. And through theoretical justification and framework customization, we develop a specialized neural network for channel estimation known as bias-free complex DnCNN (BF-CDN).
  • Utilizing the temporal correlation of the channel over a certain time period, the input to the model consists of the coarse channel estimation results of data blocks received within a certain time segment. This results in further improvement and robustness in estimation performance.
  • Simulations and real sea experimental data results confirm the robustness of the method under different noise conditions and highlight its potential to improve the accuracy and reliability of channel estimation.
The rest of this paper is structured as follows: Section 2 provides a brief overview of the UWA-OFDM system model. Section 3 introduces the proposed BF-CDN model for channel estimation and covers the problem transformation, theoretical explanation, and model architecture design. Section 4 presents simulation and experimental results and provides an analysis of these findings. Finally, in Section 5, we conclude the paper.

2. UWA-OFDM System Model

Assume that h n represents the UWA channel
h ( n ) = i = 1 N A i n δ τ τ i n .
The UWA channel is assumed to be approximated by N dominant discrete paths, where each path is associated with a complex gain A i n and time delay τ i n at the i-th discrete sample time. The formation of multipath in UWA channels is a result of various factors, including reflection on the water’s surface, seafloor, and object surfaces as well as refraction in water. These factors collectively contribute to the time-varying nature of UWA channels.
We consider a cyclic prefix (CP) OFDM baseband system in this context. The block diagram of the system is shown in Figure 1. After passing through the channel, the signal obtained by the receiver can be expressed as
y ( n ) = x ( n ) h ( n ) + w ( n ) ,
where ⊗ represents the circular convolution, and x ( n ) and w ( n ) denote the transmitted signal and additive noise, respectively. After removing the CP, the received signal in the frequency domain can be obtained by DFT transformation:
Y = d i a g ( X ) Fh + W ,
where d i a g ( X ) is the diagonal matrix of the transmitted symbols, F is the corresponding Fourier transform matrix, h is the time domain channel, and Y and W denote the frequency domains of y ( n ) and w ( n ) , respectively.

3. Methods

3.1. Problem Transformation

In traditional DL-based UWA channel estimation methods, several approaches, such as [10,11,17], employ model inputs consisting of received symbols and transmitted pilots. This type of input often requires keeping the pilot symbols fixed during the training phase, leading to strong coupling between the model parameters and the transmitted symbols, which is not conducive to transferability. Therefore, in our approach, we utilize the coarse channel estimation results as the model input. To reduce computational complexity, we employ the LS estimator.
In traditional UWA-OFDM systems, pilots are extracted and used for channel estimation. Assume that the pilots are equispaced; the LS solution in time domain can be expressed as [18]
h ^ L S n = 1 K p F p H d i a g X p H Y p ,
where K p is the number of pilots, Y p is a K p × 1 vector that consists of the received pilot symbols, d i a g ( X p ) is the K p × K p diagonal matrix with the known pilot positions along its diagonal, and F p represents the K p × N DFT matrix. The MSE of the LS estimator is calculated as [19,20]
D L S = E h ^ L S n h n 2 2 = E w n 2 x p n 2 = 1 σ x 2 σ x 2 σ w 2 σ w 2 ,
where x p n represents the transmitted pilot signal. This formula demonstrates that the MSE of the LS estimation is inversely proportional to the SNR, where σ w 2 and σ x 2 represent the variances of the pilots and the noise, respectively. Assuming the LS channel estimation results as the model input, with the goal to approximate the true UWA channel, the model output can be represented as
h ^ output = F ( h ^ LS n ; θ ) ,
where θ { θ l } l = 1 L { w l , b l } l = 1 L represents the training parameters in the neural network, L represents the number of network layers, w l and b l denote the weight items and bias items of the l-th layer, respectively, and F ( · ) represents the forward computation of the model.
If the MSE is assumed as the training loss function, then
L ( Θ ) = 1 M tr m = 1 M tr h n m F ( h ^ LS n m ; θ ) 2 2 1 SNR ,
where M tr is the batch size, and ∝ is defined as “is proportional to”. Based on the analysis of the aforementioned issues, the LS channel estimation problem shares similarities with image denoising. Specifically, it involves removing noise impact on estimation results. Leveraging this insight, this paper applies DL denoising techniques to address the channel estimation problem.
Among DL-based image denoising methods, DnCNN [21] is one of the most representative approaches. Building upon this foundation, this paper adopts and modifies the fundamental framework of DnCNN to design a channel estimation neural network.

3.2. Theoretical Analysis of DnCNN Estimator

UWA channels often exhibit significant SNR variations, and traditional neural network channel estimation methods tend to overfit within specific SNR ranges, leading to performance degradation outside the training SNR. To address this challenge, we apply the bias-free concept, which was initially introduced and validated for its robustness to SNR variations in image denoising [16], to the field of UWA channel estimation. Furthermore, we provide an explanation of why bias-free is relevant to channel estimation.
Under the criterion of minimizing MSE, MMSE estimation is optimal, and it can be expressed as [22]
h ^ M M S E n = r hh r hh + σ w 2 σ x 2 I d 1 h ^ L S n ,
where r hh = E [ h ( n ) h ( n ) H ] refers to the autocorrelation matrix of the channel vector h ( n ) in the time domain, and I d denotes the identity matrix. The MSE of the MMSE estimator can be expressed as [23]
D M M S E = E h ^ M M S E ( n ) h ( n ) 2 2 = t r r hh ( σ w 2 σ x 2 r hh + I d ) 1 D L S ,
where t r represents the trace of a matrix.
In Equation (6), the function F computed by a DnCNN can be expressed as
h ^ D n C N N ( n ) = F ( h ^ LS ( n ) ) = w L · R ( w L 1 · · R ( w 1 · h ^ LS ( n ) + b 1 ) + + b L 1 ) + b L = A h h ^ LS ( n ) + b h ,
where w L to w 1 are the weight items of the neural network, R represents the activation function rectified linear unit (ReLU), and b L to b 1 are the bias terms. The matrix A h serves as the equivalent of the overall weight matrix of the network, and vector b h plays an equivalent role to the overall bias of the network. As the error vector of the optimal estimator (in terms of MSE) [ h ^ M M S E ( n ) h ( n ) ] is orthogonal to any possible estimator [24], we can express the MSE of the DnCNN estimator as [20]
D D n C N N = E h ^ D n C N N ( n ) h ( n ) 2 2 = E { h ^ D n C N N ( n ) h ^ M M S E ( n ) + h ^ M M S E ( n ) h ( n ) 2 2 } = E { h ^ D n C N N ( n ) h M M S E ( n ) 2 2 } + E { h ^ M M S E ( n ) h ( n ) 2 2 } + 2 E { ( h ^ D n C N N ( n ) h M M S E ( n ) ) H ( h ^ M M S E ( n ) h ( n ) ) } = E { h ^ D n C N N ( n ) h ^ M M S E ( n ) 2 2 } + D M M S E .
Substituting Equations (8) and (10) into Equation (11) results in
D D n C N N = E { h ^ D n C N N ( n ) h ^ M M S E ( n ) 2 2 } + D M M S E = E { A h h ^ LS n + b h r hh r hh + σ w 2 σ x 2 I d 1 h ^ L S ( n ) 2 2 } + D M M S E .
Next, let us represent r hh ( r hh + σ w 2 σ x 2 I d ) 1 as A M M S E ; we can restate D D n C N N as
D D n C N N = E { A h h ^ LS ( n ) + b h A M M S E h ^ L S ( n ) 2 2 } + D M M S E .
It can be observed that A M M S E represents a linear transformation. A h serves as the weight matrix for the DnCNN estimator and aims to approximate A M M S E to minimize the MSE. Notably, the presence of b h does not contribute to minimizing the MSE. Therefore, it can be omitted to prevent overfitting.
Therefore, the performance comparison among the aforementioned methods can be outlined as follows:
D L S D D n C N N D M M S E .

3.3. Proposed BF-CDN Architecture

Beyond the bias removal requirement discussed earlier, channel estimation differs from image denoising in two significant ways. (1) Complex vs. real values: While image denoising deals with real-value inputs in three-channel RGB matrices, channel estimation involves complex values. (2) One-dimensional vs. two-dimensional: Images are two-dimensional matrices and are typically processed using two-dimensional convolutions. In channel estimation, the results are represented as one-dimensional vectors.
Due to these differences, this paper employs one-dimensional complex convolution layers (ComplexConv1d) to effectively leverage complex values in UWA channel estimation. The complex convolution layer can be represented as [25]
R x O U T I x O U T = A B B A R x I N I x I N ,
where x I N and x O U T represent the input and output of the network layer, respectively, R ( · ) and I ( · ) are responsible for extracting the real and imaginary parts, respectively, and A and B are the real and imaginary parts, respectively, of a complex kernel w = A + i B . Furthermore, a complex activation function, ComplexReLU, is utilized in this context; it is defined as [26]
ComplexReLU x O U T = ReLU R x O U T + i ReLU I x O U T .
Based on the optimizations of and modifications to the basic DnCNN as described above, we obtain a BF-CDN architecture suitable for UWA-OFDM channel estimation.
As shown in Figure 2, for a given depth of BF-CDN, there are three types of layers: (i) For the first layer, a ComplexConv1d and CompelxReLU are employed. The input feature map has a dimension corresponding to the length of the LS estimation results. (ii) For the hidden layers, batch normalization (BN) is added after the ComplexConv1d, and the number of features in the hidden layers remains unchanged; these are followed by a ComplexReLU. (iii) For the last layer, a ComplexConv1d is used to restore the number of features to the size of the channel dimension.
It is worth noting that the bias terms are removed in each convolutional layers to mitigate overfitting; this architecture can be represented as follows:
h ^ B F C D N ( n ) = F ( h ^ LS ( n ) ) = w L · R ( w L 1 · · R ( w 1 · h ^ LS ( n ) ) ) = A h h ^ LS ( n ) .
Meanwhile, the bias terms are also removed from the BN used during training. On the other hand, UWA channels are typical time-varying channels, and several studies have shown their strong temporal correlation at time scales within the channel coherence time [27,28]. Therefore, the temporal correlation of the channel can be utilized to improve the performance of channel estimation. Thus, we utilize LS channel estimation results within the channel coherent time as inputs. LS channel estimation results from distinct data blocks are employed as different channel inputs for the model. This approach allows the model to leverage temporal correlations to distinguish between the channel and additive noise more effectively, thereby better extracting the characteristics of the channel.
As described in Equation (7), the model employs MSE as the loss function and utilizes the Adam optimizer.

4. Results and Discussion

4.1. Simulations

To ensure the reliability of the simulations, the channel impulse responses (CIRs) from the real sea UWA communication experiment are used. These CIRs are subjected to manual denoising and then employed as the training and testing datasets for the model. Details about the experiment are presented in Section 4.2.
A total of 700 CIRs were used for the simulations, with 535 CIRs allocated for the training dataset and 135 CIRs for the testing dataset. Other simulation parameters are provided in Table 1. Additional system parameters are provided in Table 2. The model is implemented using the PyTorch framework, while Python serves as the simulation platform. The flow graph of the model is presented in Table 3.

4.1.1. Robustness under Various Noise Levels

To evaluate the robustness of the proposed method under various noise levels for UWA channel estimation, we first compare the performance between DnCNN and BF-CDN.
In Figure 3, we depict the average error levels on the test set for both methods across different training SNR ranges. It is observed that within their respective training SNR ranges, DnCNN performs well. However, when subjected to new noise levels, it experiences a substantial decline in performance, revealing evident overfitting issues within the model-training noise level. In contrast, BF-CDN maintains relatively stable performance outside the training range, even when faced with different noise conditions.
For a more intuitive explanation, a comparison of channel estimation results is provided for scenarios outside the training SNR range. Channel estimation results of three samples are shown by using DnCNN and BF-CDN models, with both models trained within an SNR range of 10 to 15 dB. In Figure 4, with a testing SNR of 15 dB, it is evident that within the training SNR range, both DnCNN and BF-CDN exhibit excellent performance, accurately estimating the channel conditions. Conversely, in Figure 5, with a testing SNR of −10 dB, DnCNN exhibits poor performance. By comparison, BF-CDN demonstrates a significant improvement in performance, highlighting the enhanced robustness of the BF-CDN model across different noise conditions.

4.1.2. Gains from Temporal Correlation

Traditional DL-based UWA channel estimation typically adopts estimating individual data blocks to capture static channel characteristics while often overlooking the common temporal correlation gains inherent in UWA channels. For the time-varying channel of several consecutive OFDM blocks within the channel coherence time, the positions of non-zero delays are similar, and the corresponding gains exhibit temporal correlation. Building upon the orthogonal matching pursuit (OMP) algorithm for UWA channel estimation [28], Zhou et al. [29] introduced the simultaneous orthogonal matching pursuit (SOMP) algorithm, which takes into account the temporal correlation between multiple channels. Experimental results have demonstrated promising performance with this method. In this context, we leverage this characteristic to improve the performance of DL-based UWA channel estimation. We provide a comparative analysis with the performance of the SOMP algorithm in Section 4.1.3 and Section 4.2.
In this context, based on the BF-CDN framework, a performance comparison is conducted between using LS channel estimation of individual data blocks and using LS channel estimation of 10 data blocks as model inputs (joint estimation, i.e., BF-CDN). Figure 6 illustrates the channel estimation performance of both methods within different training SNR ranges of the test set. It can be observed that, compared to BF-CDN-individual estimation, BF-CDN exhibits smaller estimation errors and more stable performance under various SNR conditions. Furthermore, Figure 7 illustrates examples at an SNR of 15 dB. It is evident that under conditions of high SNR, both BF-CDN-individual and BF-CDN achieve accurate channel estimation. However, as depicted in Figure 8, when the SNR decreases to 10 dB, the estimation performance of BF-CDN-individual significantly underperforms, indicating a weakened ability to suppress noise. In contrast, BF-CDN demonstrates more precise estimation under the same conditions. This demonstrates that by employing joint estimation with multiple channels over the channel coherence time, the method can capture more channel information and suppress noise effectively, leading to improved channel estimation accuracy.

4.1.3. UWA Channel Estimation Performance

In order to assess the channel estimation performance of the proposed method, several commonly used channel estimation methods were compared in simulations. The compared methods include the LS method, the OMP method, the SOMP method, and the condition of known CSI. The LS method serves as the foundational approach for channel estimation, while OMP has been widely adopted as an advanced method in recent years and demonstrates superior performance according to various studies. SOMP, building upon OMP, similarly leverages temporal correlation. Hence, these methods are selected as comparative benchmarks.
The real-time performance of the methods is summarized in Table 4. Simulations were conducted on a computer equipped with an Intel Core i7 CPU operating at 2.50 GHz and with 16 GB of memory. In terms of computational complexity, the DL-based algorithm undergoes a two-phase process involving offline training and online deployment. Following offline training, the model parameters remain constant. During the subsequent online deployment phase, channel estimation is executed through forward propagation using the pre-trained DL model.
Figure 9 and Figure 10 depict the average MSE and bit error rate (BER) performance of the channel test set. It is evident that compared to the LS, OMP, and SOMP algorithms, the proposed method, BF-CDN, significantly reduces channel estimation errors. At an SNR of 15 dB, the MSE performance improvement is 13.8 dB, 7.9 dB, and 7.3 dB, respectively. Furthermore, it consistently maintains optimal BER performance within the SNR of [−15 dB, 15 dB]. Correspondingly, the spectral efficiency under various SNRs is presented in Figure 11. It is evident that the proposed method closely approximates the spectral efficiency of the known CSI, demonstrating its efficacy at channel estimation.

4.2. Processing of Real Experimental Data

In this section, we utilized real-world sea trial data to validate the proposed method’s performance. The experiments were conducted in the western Pacific Ocean with fixed positions for both the transmitter and receiver, which were spaced approximately 55 to 60 km apart (with additional distance due to sensors and boats floating). The hydrophone was positioned at a depth of approximately 1040 m underwater. The experimental setup is depicted in Figure 12, while the sound speed profile near the hydrophone is illustrated in Figure 13. Additional experimental parameters are detailed in Table 5.
A comparison of channel estimation results for 20 consecutive received OFDM frames is presented in Figure 14, and the corresponding received SNRs are shown in Figure 15. It can be observed that in comparison to LS, OMP, and SOMP, the proposed method consistently maintains superior BER performance under different SNR conditions, demonstrating its excellent robustness for noise fluctuations.

5. Conclusions

In this paper, a robust UWA channel estimation method based on the bias-free CNN is introduced. Initially, the LS channel estimation results are employed as a channel affected by noise, and a denoising neural network is applied to achieve accurate channel estimation. Subsequently, the bias-free concept is introduced, and its necessity is theoretically explained and validated through simulations. Then, modifications are made to the model to adapt it for channel estimation. Compared to the DnCNN network, the proposed method exhibits superior robustness for noise fluctuations: even those not encountered during training. Furthermore, simulations confirm that joint estimation further improves the performance and robustness compared to individual estimation by leveraging the temporal correlation of the channel.
Finally, a performance comparison is conducted between the proposed method and classical methods such as LS, OMP, and SOMP. The simulation results show that the method outperforms these classical methods across different SNR levels. At an SNR of 15 dB, the MSE performance improvement is 13.8 dB, 7.9 dB, and 7.3 dB, respectively. Real sea trail data processing further validates the superior performance of the proposed method. In future work, exploration will be conducted on the impact of implementing the bias-free technique within various CNN architectures on channel estimation. The objective is to design and refine models for enhanced performance.

Author Contributions

D.W.: methodology, software, investigation, and writing—original draft preparation; Y.Z.: formal analysis, validation, funding acquisition, and writing—review and editing; L.W.: investigation, project administration, resources, and supervision; Y.T.: data curation, formal analysis, project administration, and writing—review and editing; H.W.: data curation, funding acquisition, investigation, and project administration; J.W.: data curation, resources, investigation, and project administration; F.M.: conceptualization, formal analysis, supervision, and writing—review and editing; F.Y.: conceptualization, formal analysis, supervision, and writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the China Scholarship Council, Chinese Academy of Sciences, CAS Specific Research Assistant Funding Program, and the National Natural Science Foundation of China (grant No. 62301551).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the Institute of Acoustics of the Chinese Academy of Sciences. Restrictions apply to the availability of these data, which were used under license for this study. Data are available from the authors upon reasonable request and with the permission of the Institute of Acoustics of the Chinese Academy of Sciences.

Acknowledgments

The authors would like to thank the editors and reviewers for their comments on the manuscript of this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UWAUnderwater acoustic
OFDMOrthogonal frequency division multiplexing
LSLeast squares
SNRSignal-to-noise ratio
MSEMean square error
MMSEMinimum mean square error
DLDeep learning
DNNDeep neural network
CNNConvolutional neural network
MLPMultilayer perceptron
LSTMLong short-term memory
CSIChannel state information
RLSRecursive least square
CSRNetChannel super-resolution network
DeSA-DNNDenoising sparsity-aware DNN
DnCNNDenoising convolutional neural network
BF-CDNBias-free complex denoising convolutional neural network
CPCyclic prefix
ReLURectified linear unit
CompelxReLUCompelx rectified linear unit
ComplexConv1dOne-dimensional complex convolution layers
BNBatch normalization
CIRsChannel impulse responses
OMPOrthogonal matching pursuit
SOMPSimultaneous orthogonal matching pursuit
BERBit error rate

References

  1. Stojanovic, M.; Beaujean, P.P.J. Acoustic communication. In Springer Handbook of Ocean Engineering; Springer: Cham, Switzerland, 2016; pp. 359–386. [Google Scholar]
  2. Howe, B.M.; Miksis-Olds, J.; Rehm, E.; Sagen, H.; Worcester, P.F.; Haralabus, G. Observing the oceans acoustically. Front. Mar. Sci. 2019, 6, 426. [Google Scholar] [CrossRef]
  3. Mohsan, S.A.H.; Li, Y.; Sadiq, M.; Liang, J.; Khan, M.A. Recent advances, future trends, applications and challenges of internet of underwater things (iout): A comprehensive review. J. Mar. Sci. Eng. 2023, 11, 124. [Google Scholar] [CrossRef]
  4. Stojanovic, M.; Preisig, J. Underwater acoustic communication channels: Propagation models and statistical characterization. IEEE Commun. Mag. 2009, 47, 84–89. [Google Scholar] [CrossRef]
  5. Huang, J.g.; Wang, H.; He, C.b.; Zhang, Q.f.; Jing, L.y. Underwater acoustic communication and the general performance evaluation criteria. Front. Inf. Technol. Electron. Eng. 2018, 19, 951–971. [Google Scholar] [CrossRef]
  6. Li, B.; Huang, J.; Zhou, S.; Ball, K.; Stojanovic, M.; Freitag, L.; Willett, P. MIMO-OFDM for high-rate underwater acoustic communications. IEEE J. Ocean. Eng. 2009, 34, 634–644. [Google Scholar]
  7. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  8. Qin, Z.; Ye, H.; Li, G.Y.; Juang, B.H.F. Deep learning in physical layer communications. IEEE Wirel. Commun. 2019, 26, 93–99. [Google Scholar] [CrossRef]
  9. Lu, H.; Jiang, M.; Cheng, J. Deep learning aided robust joint channel classification, channel estimation, and signal detection for underwater optical communication. IEEE Trans. Commun. 2020, 69, 2290–2303. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Li, J.; Zakharov, Y.; Li, X.; Li, J. Deep learning based underwater acoustic OFDM communications. Appl. Acoust. 2019, 154, 53–58. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Li, C.; Wang, H.; Wang, J.; Yang, F.; Meriaudeau, F. Deep learning aided OFDM receiver for underwater acoustic communications. Appl. Acoust. 2022, 187, 108515. [Google Scholar] [CrossRef]
  12. Liu, J.; Ji, F.; Zhao, H.; Li, J.; Wen, M. CNN-based underwater acoustic OFDM communications over doubly-selective channels. In Proceedings of the 2021 IEEE 94th Vehicular Technology Conference (VTC2021-Fall), Norman, OK, USA, 27–30 September 2021; pp. 1–6. [Google Scholar]
  13. Liu, L.; Cai, L.; Ma, L.; Qiao, G. Channel state information prediction for adaptive underwater acoustic downlink OFDMA system: Deep neural networks based approach. IEEE Trans. Veh. Technol. 2021, 70, 9063–9076. [Google Scholar] [CrossRef]
  14. Ouyang, D.; Li, Y.; Wang, Z. Channel estimation for underwater acoustic OFDM communications: An image super-resolution approach. In Proceedings of the ICC 2021-IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar]
  15. Liu, S.; Mou, Y.; Wang, X.; Su, D.; Cheng, L. Denoising enabled channel estimation for underwater acoustic communications: A sparsity-aware model-driven learning approach. Intell. Converg. Netw. 2023, 4, 1–14. [Google Scholar] [CrossRef]
  16. Mohan, S.; Kadkhodaie, Z.; Simoncelli, E.P.; Fernandez-Granda, C. Robust and interpretable blind image denoising via bias-free convolutional neural networks. arXiv 2019, arXiv:1906.05478. [Google Scholar]
  17. Zhang, Y.; Wang, H.; Tai, Y.; Li, C.; Meriaudeau, F. A machine learning label-free method for underwater acoustic OFDM channel estimations. In Proceedings of the 15th International Conference on Underwater Networks & Systems, Shenzhen, China, 22–24 November 2021; pp. 1–5. [Google Scholar]
  18. Li, B.; Zhou, S.; Stojanovic, M.; Freitag, L.; Willett, P. Multicarrier Communication over Underwater Acoustic Channels with Nonuniform Doppler Shifts. IEEE J. Ocean. Eng. 2008, 33, 198–209. [Google Scholar]
  19. Minn, H.; Bhargava, V.K. An investigation into time-domain approach for OFDM channel estimation. IEEE Trans. Broadcast. 2000, 46, 240–248. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Wang, H.; Li, C.; Chen, X.; Meriaudeau, F. On the performance of deep neural network aided channel estimation for underwater acoustic OFDM communications. Ocean Eng. 2022, 259, 111518. [Google Scholar] [CrossRef]
  21. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  22. Kay, S.M. Fundamentals of Statistical Signal Processing: Estimation Theory; Prentice-Hall, Inc.: Hoboken, NJ, USA, 1993. [Google Scholar]
  23. Gizzini, A.K.; Chafii, M.; Nimr, A.; Fettweis, G. Enhancing least square channel estimation using deep learning. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–5. [Google Scholar]
  24. Moon, T.K.; Stirling, W.C. Mathematical Methods and Algorithms for Signal Processing; Prentice-Hall, Inc.: Hoboken, NJ, USA, 2000. [Google Scholar]
  25. Hirose, A. Complex-Valued Neural Networks; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 400. [Google Scholar]
  26. Guberman, N. On complex valued convolutional neural networks. arXiv 2016, arXiv:1602.09046. [Google Scholar]
  27. Huang, S.; Yang, T.; Huang, C.F. Multipath correlations in underwater acoustic communication channels. J. Acoust. Soc. Am. 2013, 133, 2180–2190. [Google Scholar] [CrossRef]
  28. Qiao, G.; Song, Q.; Ma, L.; Liu, S.; Sun, Z.; Gan, S. Sparse Bayesian learning for channel estimation in time-varying underwater acoustic OFDM communication. IEEE Access 2018, 6, 56675–56684. [Google Scholar] [CrossRef]
  29. Zhou, Y.h.; Tong, F.; Zhang, G.q. Distributed compressed sensing estimation of underwater acoustic OFDM channel. Appl. Acoust. 2017, 117, 160–166. [Google Scholar] [CrossRef]
Figure 1. The considered CNN-based UWA-OFDM system architecture.
Figure 1. The considered CNN-based UWA-OFDM system architecture.
Jmse 12 00134 g001
Figure 2. The architecture of BF-CDN.
Figure 2. The architecture of BF-CDN.
Jmse 12 00134 g002
Figure 3. The residual norms of the estimated channels and the true channels within different training SNR ranges. The training SNR ranges are highlighted in the blue area: (a) SNR 10 , 15 dB , (b) SNR 0 , 5 dB , and (c) SNR 10 , 15 dB .
Figure 3. The residual norms of the estimated channels and the true channels within different training SNR ranges. The training SNR ranges are highlighted in the blue area: (a) SNR 10 , 15 dB , (b) SNR 0 , 5 dB , and (c) SNR 10 , 15 dB .
Jmse 12 00134 g003
Figure 4. Examples of channel estimation results for DnCNN and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 15 dB .
Figure 4. Examples of channel estimation results for DnCNN and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 15 dB .
Jmse 12 00134 g004
Figure 5. Examples of channel estimation results for DnCNN and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 10 dB .
Figure 5. Examples of channel estimation results for DnCNN and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 10 dB .
Jmse 12 00134 g005
Figure 6. The residual norms of the estimated channels and the true channels within different training SNR ranges. The training SNR ranges are highlighted in the blue area: (a) SNR 10 , 15 dB , (b) SNR 0 , 5 dB , and (c) SNR 10 , 15 dB .
Figure 6. The residual norms of the estimated channels and the true channels within different training SNR ranges. The training SNR ranges are highlighted in the blue area: (a) SNR 10 , 15 dB , (b) SNR 0 , 5 dB , and (c) SNR 10 , 15 dB .
Jmse 12 00134 g006
Figure 7. Examples of channel estimation results for BF-CDN-individual estimation and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 15 dB .
Figure 7. Examples of channel estimation results for BF-CDN-individual estimation and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 15 dB .
Jmse 12 00134 g007
Figure 8. Examples of channel estimation results for BF-CDN-individual estimation and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 10 dB .
Figure 8. Examples of channel estimation results for BF-CDN-individual estimation and BF-CDN, with both trained at noise level SNR 10 , 15 dB and tested at SNR = 10 dB .
Jmse 12 00134 g008
Figure 9. The MSE performance on the test set.
Figure 9. The MSE performance on the test set.
Jmse 12 00134 g009
Figure 10. The BER performance on the test set.
Figure 10. The BER performance on the test set.
Jmse 12 00134 g010
Figure 11. The spectral efficiency on the test set.
Figure 11. The spectral efficiency on the test set.
Jmse 12 00134 g011
Figure 12. The environment of the seal trail.
Figure 12. The environment of the seal trail.
Jmse 12 00134 g012
Figure 13. Sound speed profile.
Figure 13. Sound speed profile.
Jmse 12 00134 g013
Figure 14. BER performances on sea trial data.
Figure 14. BER performances on sea trial data.
Jmse 12 00134 g014
Figure 15. The SNR for each received frame in sea trial.
Figure 15. The SNR for each received frame in sea trial.
Jmse 12 00134 g015
Table 1. Parameters of UWA communication simulation.
Table 1. Parameters of UWA communication simulation.
ParameterValue
UWA modulation schemeOFDM with 4QAM
Bandwidth100 Hz
Center frequency300 Hz
Number of subcarriers256
Number of pilots64
Number of data subcarriers192
Length of cyclic prefix0.44 s
Number of blocks in a frame10
Table 2. Other system parameters.
Table 2. Other system parameters.
ParameterValue
OptimizerAdam
Learning rate 1 × 10 4
Batch number20
Epoch number500
Table 3. Flow graph of the proposed model.
Table 3. Flow graph of the proposed model.
Layer * Input LayerOperationOutput Shape
Input---(20, 160, 10)
Conv1ComplexConv1d layer (256, 7, 1, 3)InputComplexReLU(20, 160, 256)
Conv2ComplexConv1d layer (256, 7, 1, 3)Conv1BN + ComplexReLU(20, 160, 256)
Conv3ComplexConv1d layer (256, 7, 1, 3)Conv2BN + ComplexReLU(20, 160, 256)
Conv4ComplexConv1d layer (256, 7, 1, 3)Conv3BN + ComplexReLU(20, 160, 256)
OutputComplexConv1d layer (10, 7, 1, 3)Conv4-(20, 160, 10)
* The convolutional layer (C, K, stride, padding). C: channel number, K: kernel size, stride: stride of the convolution, padding: padding added to both sides of the input.
Table 4. Real-time performances of different methods.
Table 4. Real-time performances of different methods.
AlgorithmRuntime (ms)
LS3.7
OMP304.0
SOMP170.6
DnCNN16.3
BF-CDN17.4
Table 5. Parameters for UWA communication in real trial.
Table 5. Parameters for UWA communication in real trial.
ParameterValue
UWA modulation schemeOFDM with 4QAM
Bandwidth100 Hz
Center frequency300 Hz
Number of subcarriers256
Number of pilots64
Number of null subcarriers9
Number of data subcarriers183
Length of cyclic prefix0.44 s
Number of blocks in a frame10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, D.; Zhang, Y.; Wu, L.; Tai, Y.; Wang, H.; Wang, J.; Meriaudeau, F.; Yang, F. Robust Underwater Acoustic Channel Estimation Method Based on Bias-Free Convolutional Neural Network. J. Mar. Sci. Eng. 2024, 12, 134. https://doi.org/10.3390/jmse12010134

AMA Style

Wang D, Zhang Y, Wu L, Tai Y, Wang H, Wang J, Meriaudeau F, Yang F. Robust Underwater Acoustic Channel Estimation Method Based on Bias-Free Convolutional Neural Network. Journal of Marine Science and Engineering. 2024; 12(1):134. https://doi.org/10.3390/jmse12010134

Chicago/Turabian Style

Wang, Diya, Yonglin Zhang, Lixin Wu, Yupeng Tai, Haibin Wang, Jun Wang, Fabrice Meriaudeau, and Fan Yang. 2024. "Robust Underwater Acoustic Channel Estimation Method Based on Bias-Free Convolutional Neural Network" Journal of Marine Science and Engineering 12, no. 1: 134. https://doi.org/10.3390/jmse12010134

APA Style

Wang, D., Zhang, Y., Wu, L., Tai, Y., Wang, H., Wang, J., Meriaudeau, F., & Yang, F. (2024). Robust Underwater Acoustic Channel Estimation Method Based on Bias-Free Convolutional Neural Network. Journal of Marine Science and Engineering, 12(1), 134. https://doi.org/10.3390/jmse12010134

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop