1. Introduction
Aircraft wake turbulence is a strong airflow formed from the pressure difference between the upper and lower surfaces of the wings during high-speed flight, rotating rapidly around the wingtips and extending backward and downward [
1]. Wake turbulence, a byproduct of aircraft lift, is a critical factor affecting flight safety. When an aircraft encounters the wake of a preceding aircraft, the wake imparts a rolling moment to the following aircraft’s wings, potentially leading to loss of control [
2,
3,
4]. Therefore, wake incidents can have severe consequences.
Given the significance of aircraft wake for civil aviation safety, wake detection has become a key research area in civil aviation [
5,
6,
7,
8]. Extensive research on detection methods has been conducted domestically and internationally, including Doppler radar, microwave radar, and acoustic radar detection. Jun Li et al. [
9]. detected the wake of a transport aircraft 1.4 km away using a newly developed X-band Doppler weather radar (peak power 50 kW) at a base in Nanjing. Doppler radar is suitable for long-range detection, covering a wide area but with relatively low resolution and high equipment costs, requiring improvements in observation accuracy. As research progresses, LiDAR has gradually become a tool for observing aircraft wake, using laser beams to detect the motion and intensity of wake vortices. LiDAR receivers capture scattered laser signals, and by analyzing the time delay and frequency shift of these signals, three-dimensional images and velocity distributions of wake are obtained, providing high-resolution detection imaging for precise wake morphology and changes. Michel D T [
10] proposed an airborne LiDAR sensor for tracking wake vortices produced by aircraft in formation flight, demonstrating that airborne LiDAR can detect wake vortices generated by preceding aircraft. Liu, X [
11] applied LiDAR during aircraft departure stages, analyzing LiDAR detection methods and studying the effects of crosswinds on wake vortex evolution.
Although LiDAR can detect aircraft wake, extracting information in real-time is key to reducing wake intervals and enhancing operational safety. Köpp [
12] successfully detected distant aircraft wake using a 2 µm LiDAR. In 2011, Fibertek [
13] simulated the wake of a Boeing 747 and conducted on-site detection and validation using a developed 1.5 µm coherent Doppler wind LiDAR. In 2014, Mitsubishi Corporation [
14] discovered clear-air turbulence using an airborne coherent Doppler wind LiDAR.
In recent years, with the rapid development of deep learning, convolutional neural networks have demonstrated capabilities surpassing human performance in various computer vision tasks [
15]. Convolutional neural networks have also become a new method for LiDAR target recognition, offering robustness and strong non-linear representation abilities [
16].
Pan, W [
17] proposed a new method to recognize aircraft wake vortices by modifying the VGG16 network, providing a binary classification of uncertain wake vortex behavior patterns. Nana Chu [
18] proposed a two-stage probabilistic deep learning framework for wake vortex recognition and duration assessment. The first stage involves vortex core localization using a convolutional neural network (CNN), and the second stage predicts the vortex intensity within the region of interest (ROI) based on initial core localization results. Zhang, X [
19] proposed an improved CNN method for WV localization and grading based on PCDL data, avoiding the effects of unstable environmental wind fields on WV localization and classification results.
With the development of convolutional neural networks, the Inception-ResNet-v2 network [
20], with its depth and breadth, has achieved significant success in large-scale image classification tasks. However, the standard Inception-ResNet-v2 network still lacks in handling fine-grained features in complex environments. To address this, we proposed the IRSN-WAKE network, enhancing feature representation capability and robustness by integrating the SE (Squeeze-and-Excitation) module [
21] and a noise suppression module. The SE module adaptively reweights feature channels, improving the network’s focus on key features and addressing insufficient channel interaction. Concurrently, we designed and introduced a noise suppression module, using residual connections and element-wise operations to effectively suppress noise, enhancing feature extraction robustness and accuracy. With these modules, we built a new deep-learning framework based on the Inception-ResNet-v2 network and combined it with LiDAR data to achieve precise aircraft wake recognition. This technology is anticipated to be applied to rapid and efficient wake recognition to ensure flight safety and increase airport capacity.
This paper’s primary contributions include:
Proposing an improved Inception-ResNet-v2 network integrated with SE and noise suppression modules, effectively enhancing the network’s feature representation capability and robustness;
Designing a wake recognition framework based on LiDAR data, introducing deep learning models to achieve precise extraction and recognition of wake features;
Validating the superior performance of the proposed method in wake recognition tasks through experiments, demonstrating its potential for aircraft wake monitoring applications, and providing references for ensuring flight safety.
The structure of this paper is as follows:
Section 2 discusses the principles of aircraft wake formation, the principles of Doppler effect measurement, and the methods for collecting Doppler LiDAR data;
Section 3 details the IRSN-WAKE network architecture;
Section 4 introduces the wake recognition method based on LiDAR data;
Section 5 concludes the research and proposes future research directions.
3. Convolutional Neural Networks and Their Implementation in Detection
In recent years, deep Convolutional Neural Networks (CNNs) have achieved breakthroughs in various computer vision tasks such as image classification, object detection, sentiment recognition, and scene segmentation. Doppler LiDAR is widely used in wind field measurements [
11], and deep learning provides an effective method for detecting aircraft wake vortices. However, due to limitations in storage space and computational power, traditional neural network models still face significant challenges in storage and computation on embedded devices. Currently, research focusing on lightweight CNNs has gained attention, with the Inception and ResNet architectures being two prominent models. The former is noted for its computational efficiency and superior performance, while the latter effectively addresses the vanishing gradient problem in deep networks through residual connections. Applying lightweight networks to LiDAR data recognition has emerged as a novel approach. However, LiDAR data is single-channel and varies significantly from images with very low resolution.
To improve the accuracy and real-time performance of aircraft wake vortex detection, we propose the Inception-ResNet-Wake model. This model introduces a noise suppression module in the convolutional process to effectively suppress noise, preserve critical features, and combine the SE module with the Inception-ResNet-v2 network. The SE module readjusts the weights of feature channels adaptively, enhancing the model’s representational power to capture richer feature information across different scales, thereby improving detection accuracy. The network architecture of the model is illustrated in
Figure 5, and the pseudo-code is shown in Algorithm 1.
Algorithm 1 IRSN-WAKE Network |
1: procedure IRSN-WAKE_NETWORK(I) ← NoiseSuppression(I) ) do
6: end for do ← INCEPTION_RESNET _B_SE 9: end for do ← INCEPTION_RESNET_C_SE 12: end for ← Final_Classification_Block
15: end procedure |
3.1. Noise Suppression Module
In Algorithm 1, IRSN-WAKE introduces a noise suppression module that effectively suppresses noise and preserves key features during the convolution process. The input to the noise suppression module first undergoes initial convolution processing through a BasicConv2 module, producing the convolved output feature map . The BasicConv2 module comprises a convolutional layer, batch normalization layer, and activation function. The convolutional layer extracts local spatial features, the batch normalization layer normalizes feature values to accelerate training and improve stability, and the normalized feature map undergoes non-linear transformation via the activation function, enhancing the model’s expressive power. Subsequently, is combined with the input image through residual connection to form a new feature map . undergoes convolution processing again via the BasicConv2 module to obtain . Finally, undergoes element-wise multiplication with , and is connected back to the input image through residual connection to form the ultimate noise-suppressed feature map . This effectively enhances feature representation capability while suppressing noise influence.
3.2. Inception-ResNet-v2 Network
The size of images in pixels is 48 × 56 × 3. As shown in
Figure 6, the Inception-ResNet-v2 network combines the Inception architecture with residual connections, offering efficient computation and demonstrating high performance in image recognition tasks. The network consists of several types of modules, each designed to handle features at different scales. The Stem module serves as the entry point of the network and is responsible for initial feature extraction through three 3 × 3 convolutions and pooling layers. The Inception-ResNet-A module combines various convolution operations and residual connections to capture fine-grained features, including 1 × 1 convolutions, 3 × 3 convolutions, and 7 × 1 and 1 × 7 convolutions, which are combined through the Inception feature concatenation layer. The Reduction-A module reduces spatial dimensions while increasing the depth of the feature maps, facilitating the processing of more abstract features in deeper layers. Its core operation involves 3 × 3 convolutions and max pooling to decrease spatial dimensions. The Inception-ResNet-B module is similar to the Inception-ResNet-A module but is configured differently to capture mid-level features, including more 7 × 1 and 1 × 7 convolutions, focusing on capturing more complex patterns. The Reduction-B module further reduces spatial dimensions, preparing the network for the final stage, shrinking the spatial dimensions to 5 × 6 while increasing the depth to 1152 through 3 × 3 convolutions and max pooling. The Inception-ResNet-C module acts as the network’s final convolutional operation, capturing high-level features and refining the feature maps through a series of 1 × 1 and 3 × 3 convolutions to produce the final output.
3.3. The IRSN-WAKE Network
The IRSN-WAKE network integrates a denoising module on the basis of the Inception-ResNet-v2 architecture and incorporates the SE (Squeeze-and-Excitation) module.
The SE module consists of two main steps: Squeeze and Excitation. The Squeeze step addresses the issue of channel dependencies by compressing global spatial information into a channel descriptor. It achieves this by globally averaging the spatial dimensions of feature maps, compressing a W × H × C feature map into a 1 × 1 × C feature vector Z that encapsulates global information. Each channel’s feature maps are compressed into a single numerical value, ensuring that the generated channel-wise statistics Z contain contextual information, thereby alleviating channel dependencies. This involves applying global average pooling to each channel’s feature map, generating channel-level statistical data, as shown in Equation (5).
where
is the global average pooling using the channel,
is the value of channel
at position
,
and
are the height and width of the feature map, respectively.
To leverage the information pooled during the compression operation, SE uses the Excitation operation to comprehensively capture channel dependencies. A gating mechanism composed of two fully connected layers is employed; the first layer reduces the number of channels from
to
to decrease the computational load, followed by a
non-linear activation layer. The second fully connected layer restores the number of channels back to
, and then a Sigmoid activation yields the weights
. The final dimension of
is
, which is used to characterize the weights of
feature maps in the feature map
.
refers to the reduction ratio. The gating mechanism, including the fully connected layers, captures dependencies between channels and recalibrates the feature map, as seen in the formula.
Here, represents the activation function, represents the Sigmoid activation function, and are the weight matrices of the fully connected layers, and is the channel statistics generated by the Squeeze operation.
A reweighting operation is then performed. The attention weights previously obtained are applied to each channel’s features. Each feature map in feature map
is multiplied by its corresponding weight to produce the final output
of the SE module.
where
is the weight for channel
,
is channel
of the input feature map and
is the reweighted feature map.
To enhance the feature representation ability of the Inception-ResNet-v2 network, allowing it to adaptively weight different feature channels, the SE module can be inserted after each Inception module (Inception-ResNet-A, B, C modules). Take the Inception-ResNet-A module as an example, as shown in
Figure 7.
First, the feature map
is generated by the Inception-ResNet-A, which is then passed to the SE module for processing. The output
from the Inception-ResNet-A module is:
The integration of the SE module ensures that the network can utilize both fine-grained and abstract features, focusing on the most informative features, thereby enhancing the model’s robustness and accuracy.
4. Experimental Setup and Results
4.1. Experimental Platform
The experiments were conducted on the Windows 11 operating system using the Python programming language with the PyTorch 2.4.0 training framework. The GPU used was NVIDIA GeForce RTX 4090 (NVIDIA, Santa Clara, CA, USA), the processor was 13th Gen Intel® Core™ i7-13700K 24 (Intel, Santa Clara, CA, USA), and the memory was 32GB.
4.2. Data Processing and Network Parameter Settings
In the initial stages of the experiment, we collected wake vortex data from different aircraft models, including A380 and A320, using a Doppler lidar at Chengdu Shuangliu International Airport. Due to the varying wind speeds in the wake vortex background, the collected radar data was normalized to remove scale effects before inputting them into the network. Normalizing the data accelerates gradient descent in convolutional neural networks, enhancing the model’s convergence speed.
Figure 8 depicts the average wind speed variation over time in the background wind field.
In total, 3500 data samples were collected for this experiment. After randomizing the data, 60% was allocated for the training set, 20% for the validation set, and the remaining 20% for the test set.
This experiment used the Adam adaptive gradient descent algorithm. This method is simple to implement, computationally efficient, and consumes less memory [
19]. It is widely applied in the fields of computer vision and natural language processing. The initial learning rate for the optimizer in this experiment was set to 0.001; momentum was set to 0.9; weight decay was 1 × 10
−4. Additionally, the convolutional neural network training model utilized the cross-entropy loss function. Cross-entropy measures the distance between the actual output p and the expected output probability q. A smaller cross-entropy value indicates that the model fits better with the training data. Mathematically, as shown in Equation (9):
4.3. Ablation Experiments
To verify the effectiveness of each module, we conducted ablation studies, as shown in
Table 3. Inception-ResNet refers to the original Inception-ResNet-v2 model, Inception-ResNet-SE is the model with the addition of the SE module to the base Inception-ResNet-v2, Inception-ResNet-ND involves the addition of a noise suppression module to the base Inception-ResNet-v2, and ISRN-wake is our constructed model which incorporates both SE and noise suppression modules to the Inception-ResNet-v2 framework. The experimental results demonstrate that the addition of the SE and noise suppression modules significantly improves the ISRN-wake model across various evaluation metrics, including Accuracy, Precision, Recall, and F1-Score., validating our architectural enhancements.
4.4. Results and Analysis
In this experiment, preprocessed training data was fed into IRSN-WAKE for training, which stopped after 200 iterations.
Figure 9 shows the iteration-wise change in loss values during the training process.
To further demonstrate the effectiveness of the constructed target network, we compared common target recognition networks: VGG16, SVM, KNN, and RF. We used accuracy, precision, recall, and F1 scores as evaluation metrics. Accuracy refers to the proportion of correctly classified samples to the total number of samples, as shown in Equation (10); precision refers to the proportion of actual positive samples among all samples predicted to be positive, as shown in Equation (11); recall refers to the proportion of correctly predicted positive samples among all actual positive samples, as shown in Equation (12); and the F1 score is the harmonic mean of precision and recall, used to comprehensively evaluate the model’s performance, as shown in Equation (13).
where
TP represents the number of true positive predictions made by the model.
TN represents the number of true negative predictions made by the model.
FP represents the number of false positive predictions made by the model (also known as Type I errors). FN represents the number of false negative predictions made by the model (also known as Type II errors).
The results are shown in
Table 4, where the ISRN-wake model exhibits outstanding performance across all evaluation metrics, achieving the highest accuracy of 0.986, precision of 0.967, and F1 score of 0.960. This indicates the effectiveness of the constructed ISRN-wake network in recognizing aircraft wake vortices and minimizing false alarms.
To further evaluate the performance of the models in binary classification tasks, we calculated the confusion matrix for each model and conducted an in-depth analysis of their classification capabilities across different categories. As shown in
Figure 10, the results indicated that deep learning models, such as VGG16 and ISRN-wake, performed exceptionally well in these tasks, especially in situations where the ratio of positive to negative samples was imbalanced. Although traditional machine learning models like SVM, KNN, and RF also showed good performance, they exhibited some limitations in balancing precision and recall.
The ISRN-wake model, with a TP of 1665, TN of 424, FP of 57, and FN of 80, demonstrated extremely high precision and recall in detecting positive samples. Compared to the other models, it had the lowest miss rate and false alarm rate. Moreover, the ISRN-wake model achieved a detection speed of 180 fps while utilizing only 0.4 M of resources, proving its real-time capability and accuracy in handling complex tasks such as wake vortex detection. In contrast, although VGG16 is also a deep learning model, its detection speed was only 50fps, with resource usage as high as 120 M.
5. Conclusions
This study proposes the IRSN-WAKE network, which achieves efficient identification of aircraft wake vortices by integrating SE (Squeeze-and-Excitation) and denoising modules. The SE module enhances the sensitivity of the network to critical features by adaptively weighting feature channels, while the denoising module effectively suppresses noise, thereby improving feature extraction robustness and accuracy. Experimental results demonstrate that the framework accurately identifies aircraft wake features in complex airflow environments, showing high recognition accuracy and robustness. Through experimental validation, we found:
IRSN-WAKE network has enhanced feature extraction capabilities: The SE module significantly improves the network’s sensitivity to key features, while the denoising module effectively reduces noise interference.
Importance of combining Doppler LiDAR data with deep learning in capturing aircraft wake vortices: Integrating deep learning techniques with sensor data enables efficient wake recognition in complex environments, offering new solutions for aircraft wake monitoring.
However, for aircraft wake recognition tasks, we only used Doppler LiDAR data from a single airport in this study. To address these limitations, future research will consider integrating heterogeneous data sources such as weather radar data and ADS-B data into aircraft wake recognition tasks to improve recognition accuracy. Additionally, collecting aircraft wake data from various types of civil airports and training deep learning models accordingly will enhance model generalizability, thereby providing technical support for safe and efficient flight operations.