Next Article in Journal
Aerodynamic Optimization Method for Propeller Airfoil Based on DBO-BP and NSWOA
Previous Article in Journal
Planning and Evaluation of Water-Dropping Strategy for Fixed-Wing Fire Extinguisher Based on Multi-Resolution Modeling
Previous Article in Special Issue
Integrated Technologies for Anti-Deicing Functions and Structures of Aircraft: Current Status and Development Trends
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Temperature Distribution on an Aircraft Hot-Air Anti-Icing Surface by ROM and Neural Networks

1
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 610054, China
2
State Key Laboratory of Aerodynamics, Mianyang 621000, China
3
Low Speed Aerodynamics Institute, China Aerodynamics Research and Development Center, Mianyang 621000, China
4
School of Mechanical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Aerospace 2024, 11(11), 930; https://doi.org/10.3390/aerospace11110930
Submission received: 16 October 2024 / Revised: 2 November 2024 / Accepted: 6 November 2024 / Published: 11 November 2024
(This article belongs to the Special Issue Deicing and Anti-Icing of Aircraft (Volume IV))

Abstract

:
To address the inefficiencies and time-consuming nature of traditional hot-air anti-icing system designs, reduced-order models (ROMs) and machine learning techniques are introduced to predict anti-icing surface temperature distributions. Two models, AlexNet combined with Proper Orthogonal Decomposition (POD-AlexNet) and multi-CNNs with GRU (MCG), are proposed by comparing several classic neural networks. Design variables of the hot-air anti-icing cavity are used as inputs of the two models, and the corresponding surface temperature distribution data serve as outputs, and then the performance of these models is evaluated on the test set. The POD-AlexNet model achieves a mean prediction accuracy of over 95%, while the MCG model reaches 96.97%. Furthermore, the proposed model demonstrates a prediction time of no more than 5.5 ms for individual temperature samples. The proposed models not only provide faster predictions of anti-icing surface temperature distributions than traditional numerical simulation methods but also ensure acceptable accuracy, which supports the design of aircraft hot-air anti-icing systems based on optimization methods such as genetic algorithms.

1. Introduction

When an aircraft flies under icing conditions, ice may accumulate on its surface and pose a severe threat to flight safety and even lead to accidents [1]. Effective anti-icing measures enhance the ability of aircraft to handle icing conditions. Hot-air anti-icing systems, widely used in large civil airliners [2], are particularly suitable for areas requiring extensive anti-icing, such as wings and engine inlets. The hot-air anti-icing system uses compressor bleed air to prevent icing by conducting heat through the skin to the outer surface [3]. The piccolo tube structure within the hot-air anti-icing cavity mainly determines the temperature distribution on the surface and the performance of the hot-air anti-icing system [4]. Therefore, a rapid and accurate prediction method of anti-icing surface temperature related to the piccolo tube structure is crucial for designing and optimizing the hot-air anti-icing system.
With advancements in icing wind tunnel experiments, researchers have investigated the effects of the internal and external aspects of the hot-air anti-icing system on surface temperature distribution. The anti-icing characteristics of engine inlet guide vanes and other components were studied in the YBF-02 icing wind tunnel, and the effects of liquid water content and hot-air mass flow rate on the performance of the anti-icing system were analyzed [5,6]. The wing hot-air anti-icing system was studied in the China Aerodynamics Research and Development Center Icing Wind Tunnel (CARDC IWT), and the effects of icing meteorological conditions, hot-air mass flow rate, and hot-air temperature on the double-skin heat transfer enhancement hot-air anti-icing system were investigated [7]. Experiments in icing wind tunnels provide controllable environments and accurate data, but the preparation cycle is long and costly.
Advanced numerical methods have become crucial for the design and verification of anti-icing systems due to the development of thermodynamic models and numerical simulations. A series of numerical calculations using the SST k ω turbulence model were performed [8], comparing three injection methods—impingement jets, offset jets, and swirl jets—and demonstrating that the swirling effect is able to enhance internal heat transfer. The influence of the velocity, height, temperature, liquid water content (LWC), and median volume diameter (MVD) on the anti-icing thermal load was investigated [9]. Numerical simulation for predicting the surface temperature of the anti-icing cavity skin involves calculating flow fields and anti-icing performance, which is computationally intensive and time-consuming. These limitations indicate that existing numerical simulation methods cannot meet the requirements for the design of aircraft hot-air anti-icing systems based on optimization methods such as genetic algorithms.
Compared to numerical simulations, there are few studies using non-numerical methods to predict the surface temperature of hot-air anti-icing systems, primarily involving ROM-based and machine learning-based approaches. ROMs integrate data reduction techniques such as POD with surrogate models like neural networks to predict low-dimensional approximations of the original data. The back propagation neural network improved with the particle swarm optimization algorithm (PSO-BP) [10] effectively maps the flight condition parameters to characteristic coefficients of surface temperature, achieving an average absolute error of 3.87 K. Additionally, a meta-model was employed based on POD and a general regression neural network to evaluate the performance of electrothermal anti-icing systems [11]. These approaches demonstrate that ROMs effectively capture the main characteristics of surface temperature, thereby improving prediction efficiency while maintaining approximation accuracy. An artificial neural network (ANN) has been employed to predict wing temperatures based on experimental and computational fluid dynamics (CFD) data [12], but this method predicts temperatures at only three lines within the anti-icing region. K-nearest neighbor (KNN) and a local linear weighted regression algorithm [13] were also employed to predict the temperature trend of the electrothermal anti-icing surface, but this method converts predicted temperature change rates into actual temperatures rather than predicting surface temperature directly. Overall, existing non-numerical simulation methods primarily focus on temperature predictions for electrothermal anti-icing systems or target only specific cross-sections within the anti-icing region, with insufficient research on predicting the complex three-dimensional surface temperatures for hot-air anti-icing systems. However, the ROMs introduce additional errors during data reduction and reconstruction, and neural network methods are currently seldom applied directly to regression prediction of high-dimensional temperature data.
This study focuses on enhancing the approximation accuracy of a ROM and developing a high-dimensional neural network model to predict the surface temperature distribution across the entire hot-air anti-icing region. Comparative experiments are conducted to validate the efficacy of the proposed methods, demonstrating their superiority over traditional approaches. The Section 2 of this paper describes the methods for obtaining and processing samples, while the Section 3 explains the basic principles and network architectures of the proposed models. The Section 4 presents the experimental results and analyses.

2. Sample Data Acquisition and Processing

This paper focuses on a turbofan engine nacelle inlet and gathers surface temperature data for neural network experiments through three-dimensional anti-icing numerical simulations. Relevant data processing operations are then performed to facilitate network training.

2.1. Data Collection

Figure 1a shows the three-dimensional model of the turbofan engine nacelle inlet. Numerical simulations are conducted on the localized hot-air anti-icing system at the 12 o’clock position of the inlet (Figure 1b) to obtain the corresponding surface temperature distribution data.
The jet holes on the piccolo tube are arranged in a diamond shape, i.e., one of the two neighboring columns has only one jet hole, and the other column has two. The design parameters are selected based on the internal structural characteristics of the inlet hot-air anti-icing system (Figure 2): (1) z-direction jet hole spacing between two neighboring columns L; (2) x-coordinate of the piccolo tube center x pic ; (3) y-coordinate of the piccolo tube center y pic ; (4) the outflow direction angle of jet holes in the middle row θ 0 ; (5) the relative angle between the clockwise (positive y-axis direction) jet holes and the middle row of jet holes θ 1 ; (6) the relative angle between the counterclockwise (negative y-axis direction) jet holes and the middle row of jet holes θ 2 .
The design variable space is six-dimensional, with parameter ranges detailed, and the baseline design presented in Table 1. A total of 5000 samples of piccolo tube design variables are generated using Latin Hypercube Sampling (LHS) [14] to ensure a near-random distribution across the unit hypercube.
On this basis, the three-dimensional anti-icing numerical simulation method [15] is used for calculating the surface temperature of the hot-air anti-icing system with the sampled structural design variables. This method includes airflow field calculation, water droplet collection efficiency calculation, and three-dimensional surface icing thermodynamics calculation. Comparing the numerical simulation results of this method with those from the traditional tight-coupled method and experimental results demonstrates the reliability of this approach [1,15]. The skin solid domain used for calculating the temperature distribution data employs a structured grid consisting of 198 grids in the chord direction and 121 grids in the span direction, totaling 23,958 grids (Figure 3). The design conditions are shown in Table 2.

2.2. Training Data Processing

Each sample consists of two parts: anti-icing design variables and surface temperature distribution data. All 5000 samples are divided into three datasets in a ratio of 8:1:1. A total of 4000 samples are used as the training set for training network models, with 500 samples allocated as the validation set for adjusting the model hyperparameters, and 500 samples set aside as the test set for evaluating model performance. Assessing the performance of models using the test set that was not involved in training ensures their reliability in real-world environments.
Temperature distribution data are normalized using the Z-score method to facilitate network learning, as shown in the following equation:
B i = A i μ i δ i i = 1 , 2 , , 23958
Here, B i represents the standardized temperature at the i-th point, A i denotes the original temperature at the i-th point, and μ i and δ i are the mean and standard deviation, respectively, of the temperature at the i-th point. The mean of the temperature dataset is 0, and the standard deviation is 1 after standardization. Normalization enables data with different features to be compared on the same scale, thereby enhancing the performance and stability of networks.

3. Rapid Prediction Method of Temperature Distribution

3.1. Basic Principles

3.1.1. Proper Orthogonal Decomposition

Proper Orthogonal Decomposition (POD), also known as Principal Component Analysis (PCA), is a mathematical method used for data dimensionality reduction and pattern recognition, allowing the extraction of main feature patterns from multi-dimensional datasets. The fundamental idea of POD is to decompose complex datasets into orthogonal basis functions that capture the main features of the data. These basis functions project the original features into a new orthogonal feature space, approximating the original data with fewer parameters. This process achieves data dimensionality reduction, eliminates redundant information between features, and retains the maximum variance, thus maintaining the key features of the data.
The steps of the POD are as follows:
  • Data collection: Organize the temperature distribution matrix X so that each column represents a temperature distribution sample and each row corresponds to a grid point.
  • Calculate the covariance matrix: The covariance matrix C of the matrix X can be expressed using the following equation:
    C = 1 m X X T
    Here, X is an n × m matrix, where n represents the number of temperature points on the surface and m represents the number of samples; X T denotes the transpose of matrix X, resulting in the covariance matrix C being of dimension n × n .
  • Eigenvalue decomposition: The eigenvalue decomposition of the covariance matrix C yields the eigenvalues λ i and the corresponding eigenvectors ϕ i :
    C ϕ i = λ i ϕ i
  • Select main eigenvalues and eigenvectors: The top k eigenvalues λ i and their corresponding eigenvectors ϕ i are selected based on the magnitude of the eigenvalues. These eigenvectors will serve as POD basis functions.
  • Construct the basis mode matrix: Construct the POD basis mode matrix Φ using the selected k eigenvectors ϕ i :
    Φ = [ ϕ 1 , ϕ 2 , , ϕ k ]
  • Data dimensionality reduction: Project the original data X onto the POD basis mode matrix Φ to obtain a low-dimensional representation Y:
    Y = Φ T X
    Here, Y is referred to as the matrix of fitting coefficients.
POD is commonly used in fluid mechanics to analyze complex fluid dynamics problems, such as eddy currents and turbulence [16,17]. By reducing the dimensionality of high-dimensional temperature distribution data, POD overcomes the limitations of traditional numerical calculations and improves prediction efficiency.

3.1.2. Convolutional Neural Networks

Convolutional neural networks (CNNs) [18] are deep learning models primarily designed to process data with a grid structure, such as images and videos. CNNs can automatically learn the spatial features of images through their layered architectures and local connections.
The basic structure of CNNs includes convolutional layers, pooling layers, and fully connected layers [19]. The convolutional layer is the core of CNNs, extracting local features from the input data through convolution operations. The convolution operation in this layer uses a small convolution kernel (or filter) to slide over the input data, calculating the dot product between the kernel and the input data to generate a feature map. By sharing convolution kernel parameters, the convolutional layer significantly reduces the number of model parameters, enhancing computational efficiency. The pooling layer is used to reduce the size of feature maps, decrease computational complexity, and prevent overfitting. Pooling operations include max pooling and average pooling [20]. The fully connected layer is usually located at the end of CNNs, mapping the flattened feature map to output for classification or regression tasks.
Classic CNNs include LeNet [20], AlexNet [21], UNet [22], and ResNet [23], among others. The architecture of AlexNet used in this paper is illustrated in Figure 4, where the layers in the feature block follow the classic AlexNet layer configurations. In the experiments described in Section 4, a six-dimensional input vector is reshaped into a six-channel image of size 128 × 128 before being fed into AlexNet.

3.1.3. Recurrent Neural Networks

Recurrent neural networks (RNNs) [24] are a type of neural network designed to process sequence data, making them particularly suitable for time series prediction and natural language processing tasks. The basic unit of RNNs consists of an input layer, hidden layers, and an output layer. Unlike feedforward neural networks, the hidden layers of RNNs have recurrent connections. At each time step, the hidden layers receive both the input of the current time step and the hidden state of the previous time step. This recurrent structure enables RNNs to capture dynamic changes in sequence data. At each time step, the hidden layers of RNNs update their state as shown in the following equation:
h t = f ( W x h x t + W h h h t 1 )
Here, h t is the hidden state at time t; h t 1 is the hidden state at time t 1 ; W x h and W h h are learnable weight matrices; x t is the input at the current time step; f is typically a nonlinear activation function.
To address the gradient vanishing and exploding problems in traditional RNNs, long short-term memory (LSTM) [25] and gated recurrent unit (GRU) [26] introduce gating mechanisms to control the flow of information. GRU is a simplified version of LSTM, using only two gates (update gate and reset gate), reducing the model complexity while still capturing long-term dependencies effectively. The classic structure of a GRU network is shown in Figure 5.
The reset gate r t controls how much past information influences the current candidate hidden state. Its calculation is given by Equation (7). The current candidate hidden state h ˜ t is computed as shown in Equation (8). The update gate z t manages how much information from the previous hidden state is retained and how much from the new candidate hidden state is incorporated into the current hidden state. Its calculation is provided in Equation (9), and the calculation for the current hidden state h t is shown in Equation (10).
r t = σ ( W r · [ h t 1 , x t ] )
h ˜ t = tanh ( W h · [ r t h t 1 , x t ] )
z t = σ ( W z · [ h t 1 , x t ] )
h t = ( 1 z t ) h t 1 + z t h ˜ t
Here, W r is the weight matrix of the reset gate, x t represents the input vector at time t, and σ denotes the sigmoid activation function, whose output ranges between 0 and 1. W h is the weight matrix for the candidate hidden state, ∘ denotes element-wise multiplication, and tanh represents the hyperbolic tangent activation function. W z is the weight matrix for the update gate. The functional expressions for the sigmoid and tanh activation functions are provided in the Equations (11) and (12), respectively.
σ ( x ) = 1 1 + e x
tanh ( x ) = e x e x e x + e x

3.2. Temperature Distribution Prediction Method Based on ROM

This section combines the POD and neural network models to predict the three-dimensional temperature distribution of the hot-air anti-icing system. First, POD is applied to the temperature distribution to obtain the basis mode matrix and fitting coefficients. Then, key design variables serve as inputs for training the neural network, with the predicted fitting coefficients as outputs. Finally, data reconstruction combines the basis mode matrix with these predicted fitting coefficients to yield the final temperature distribution prediction results. The technical workflow is shown in Figure 6.
  • Data acquisition and preprocessing stage: First, the LHS method is used to sample, generating 5000 design variable samples of the piccolo tube. Then, the three-dimensional anti-icing numerical simulation method calculates the surface temperature of the hot-air anti-icing cavity in the turbofan engine inlet, resulting in the formation of the temperature distribution matrix. Subsequently, the POD method, described in Section 3.1.1, reduces data dimensionality on the temperature distribution matrix, obtaining the basis mode matrix and the matrix of fitting coefficients. During this process, the top 128 eigenvectors, chosen based on the magnitude of their eigenvalues, are compiled into the basis mode matrix. The truncated number, 128, is determined based on the cumulative energy ratio of the modes, as shown in Figure 7.
  • Network model training and predicting stage: As illustrated in Figure 8, during the training of the neural network model, the design variables serve as inputs, while the decomposed fitting coefficients are the target values. For predictions, the trained neural network uses the provided design variables to predict the fitting coefficients.
  • Data post-processing stage: The predicted fitting coefficients from the neural network, combined with the basis mode matrix obtained from the POD decomposition, are used for data reconstruction, resulting in the final predicted temperature distribution.

3.3. Temperature Distribution Prediction Method Based on High-Dimensional Data

In this section, a neural network model predicts the anti-icing surface temperature distribution using structural design variables as inputs, with the surface temperature distributions as the outputs. The neural network architecture MCG proposed in this section primarily consists of two-dimensional convolution layers (Conv), two-dimensional transposed convolution layers (ConvTranspose), GRU layers, LeakyReLU activation layers, fully connected layers (Linear), and other structures. Figure 9 illustrates the network structure of MCG.
MCG includes two different paths to process input data:
  • Path 1: The initial part of this path expands a 1 × 6 input vector into a 32 × 32 matrix through a fully connected layer and a reshaping operation, facilitating subsequent processing. Subsequently, a CNN block composed of multiple convolution layers is applied to capture features and patterns in the expanded vector. Following the CNN block, a TCNN block, made up of several transposed convolution layers, aims to enlarge the dimensions of the feature map while retaining feature correlations. Finally, a fully connected layer processes the output to promote high-level abstraction and feature aggregation.
  • Path 2: This path employs three layers of GRU to extract relevant features from the input sequence, effectively capturing patterns and correlations within it. It learns the mapping relationship between the anti-icing design variables and the temperature distribution at the z = 0 m position.
  • Integration and output: The output of Path 1 is reshaped into a 198 × 121 format. Then, the central column is replaced with the output of Path 2 to obtain the final result.
All activation functions in the model use LeakyReLU, which is defined by the following equation:
f ( x ) = x , x > 0 α x , x 0
The LeakyReLU function, with its non-zero output in the negative region, helps prevent neuron death that can occur with ReLU, enabling the network to learn a broader range of features. Additionally, compared to other more complex activation functions, LeakyReLU is computationally simpler, thus providing higher efficiency during training.

4. Experiments and Results Analysis

This paper employs the PyTorch framework to build prediction network models. The computer configuration used for model development and training includes an Intel (R) Core (TM) i9-10900 CPU running at 2.80 GHz × 20, 32 GB of RAM, and an NVIDIA GeForce RTX 3080 Ti graphics card. Both network training and prediction are performed using the GPU.

4.1. Loss Function

SmoothL1Loss is chosen as the loss function during the training process, and its functional expression is as follows:
SmoothL1Loss ( x , y ) = 0.5 ( x y ) 2 , | x y | < 1 | x y | 0.5 , | x y | 1
When the difference between the predicted value x and the true value y is less than 1, SmoothL1Loss behaves as L2 loss (squared error), making it more sensitive to small errors. When the difference is greater than or equal to 1, the loss behaves as L1 loss (absolute error), thereby reducing the impact of larger errors.

4.2. Performance Metrics

In the experiments, the root mean square error (RMSE), mean relative error (MRE), and mean absolute error (MAE) are chosen as performance metrics to compare the performances on the test set of various prediction methods. The equations for each metric are as follows:
RMSE = 1 m i = 1 m j = 1 l ( T i , j T ^ i , j ) 2
MRE = 1 m i = 1 m j = 1 l T i , j T ^ i , j T i , j
MAE = 1 m i = 1 m j = 1 l T i , j T ^ i , j
Here, m is the total number of samples, l is the total number of grids, and T i , j and T ^ i , j represent the target and predicted values of i-th sample at the j-th grid, respectively. In order to better evaluate the established method, the mean prediction accuracy (MPA) [27] is proposed:
MPA = 1 m i = 1 m N s , i l × 100 %
Here, an absolute error less than s serves as the standard for accuracy. If this standard is met, the temperature prediction for the grid point is deemed accurate. Let N s , i denote the number of grids with accurate predictions for the i-th sample. With s = 5 , N s , i / l represents the prediction accuracy (PA) of the i-th sample. The PA for all test samples is averaged to calculate the MPA on the test set.

4.3. Temperature Distribution Prediction Based on ROM

4.3.1. Comparative Test Results Analysis of ROMs

This section compares multiple classic neural networks, including AlexNet, UNet, ResNet, LSTM, GRU, multi-layer perceptron (MLP) [28], and variational autoencoder (VAE) [29], in predicting POD fitting coefficients and evaluates their performance on the test set. Figure 10 illustrates the variation in loss values during the ROM training process.
All models have reached convergence, indicating their training losses have stabilized after a certain number of epochs. However, the extent to which the loss has decreased varies across different models. The ROMs based on GRU, LSTM, and VAE, respectively, show little improvement in their training losses, suggesting that these models may not have effectively learned the underlying data patterns as effectively as others. The POD-MLP reaches a moderate level after convergence. In contrast, the ROMs based on ResNet, UNet, and AlexNet, respectively, show the most significant reduction in training losses, with rapid and effective convergence. Notably, POD-AlexNet stands out by reducing the loss more rapidly than other networks and consistently maintains the lowest loss value throughout the training process, which further highlights its learning efficiency compared to other models.
Table 3 presents the test performance evaluation results of different ROMs, where t t e s t denotes the average prediction time for a single sample. The comparative results indicate that convolution-based models outperform sequence-based models in predicting the fitting coefficients, which are essential for reconstructing the temperature distribution. This disparity arises from the inherent characteristics of these models and the nature of the data involved.
LSTM and GRU are well-suited for sequence data, excelling at capturing temporal dependencies [30]. However, the primary challenge in this study is not modeling temporal sequences but accurately capturing spatial patterns in the fitting coefficients. LSTM and GRU struggle to model these spatial relationships effectively, which results in suboptimal performance.
POD-MLP and POD-VAE exhibit weaker performances. MLP, as a fully connected model, struggles to efficiently capture spatial dependencies inherent in the data [31], leading to slower and less accurate predictions. On the other hand, VAE, which aims to learn a probabilistic latent representation, often suffers from information loss and tends to produce overly smooth outputs [32,33]. This limitation hinders its ability to capture fine spatial details in the temperature distribution.
In contrast, convolution-based models like AlexNet are inherently better suited for spatial data. The POD-AlexNet method developed in this paper, despite its relatively simple architecture, demonstrates significant improvements across all performance metrics. Furthermore, the average prediction time is just 1 ms, considerably lower than that of traditional numerical simulation methods.

4.3.2. Prediction Results and Analysis

Figure 11 compares the predicted results of POD-AlexNet with the target results. The three contour plots on the left illustrate the predicted results, target results, and absolute errors of the surface temperature distribution. In these plots, the vertical coordinate s = 0 m represents the stagnation point at the leading edge of the inlet. Positive values represent the upper surface of the inlet, while negative values correspond to the lower surface. The line graph on the right illustrates the predicted and target temperature results at z = 0 m (the 12 o’clock position, as shown in Figure 1). It can be observed that the predicted results align closely with the target results, and the predicted values, along with their change patterns at various positions on the anti-icing surface, are notably accurate.

4.4. Temperature Distribution Prediction Based on High-Dimensional Model

4.4.1. Comparative Test Results Analysis of High-Dimensional Model

This section conducts comparative experiments on several classic neural networks, including LeNet, AlexNet, UNet, VAE, LSTM, and GRU, as well as our proposed model, MCG. Their effectiveness and superiority are evaluated by comparing their performance on a test set.
Figure 12 illustrates the variation in loss values during the high-dimensional model training process. In this figure, some of the first epochs have been ignored for the sake of a graphical representation of data. All models’ training losses have converged, but there are some variation in their final loss values. The VAE exhibited the largest loss after convergence, followed by UNet with the second-largest loss, showing slightly better performance than VAE, but still not ideal compared to the other models. AlexNet, LeNet, LSTM, and GRU all converged to similar loss values, occupying the middle ground, which suggest that the performance of these models is comparable. The MCG model demonstrated the greatest reduction in loss and achieved the best convergence, maintaining superior performance throughout the training process.
Table 4 presents the performance evaluation results for various networks. Comparative tests reveal that sequence-based models, such as LSTM and GRU, outperform convolution-based models like LeNet, AlexNet, and UNet when directly predicting high-dimensional temperature distribution data. Among all models, VAE exhibits the lowest accuracy. These performance differences arise from the distinct characteristics of the models and the nature of the high-dimensional data.
Convolution-based models like LeNet, AlexNet, and UNet are highly effective in tasks dominated by local spatial features, such as image classification or segmentation, as shown in numerous studies [19,21,22]. However, predicting high-dimensional temperature distributions in this study requires capturing complex dependencies across the entire spatial field, which may limit the effectiveness of convolutional models that prioritize local features extraction.
Conversely, LSTMs and GRUs capture long-term dependencies and patterns across data effectively [34]. Their gating mechanisms dynamically adjust the importance of different data points, which make them particularly suited for handling complex, high-dimensional data [32,33]. This likely explains their superior performance over convolutional models, which focus on local feature extraction and may overlook broader, non-local dependencies.
VAEs encode data into a latent space and then decode it, making them more prone to information loss, particularly in high-dimensional tasks [35,36]. In addition, VAEs aim to learn the data’s probability distribution rather than directly optimize prediction accuracy, often resulting in overly smooth outputs. These characteristics make VAEs less effective at capturing fine details in temperature distributions, potentially explaining their weaker performance in experimental results.
Based on the above experimental and analytical results, this study proposes the MCG model, which combines convolutional networks and GRUs in parallel, leveraging the advantages of both. The convolutional networks excel at extracting spatial features, while the GRU captures long-term dependencies in high-dimensional temperature distribution data. The results indicate that this combination significantly enhances prediction performance. The MCG model achieves notable gains in RMSE, MRE, MAE, and MPA on the test set. Moreover, its average prediction time per sample is about 5.5 ms—much faster than traditional numerical simulations that typically take hours or even days, marking a substantial improvement in computational efficiency.

4.4.2. Prediction Results and Analysis

Figure 13 compares the predicted results of MCG with the target results. The three contour figures on the left display the predicted results, target results, and absolute errors of the surface temperature distribution. The vertical coordinate s = 0 m in the figure represents the stagnation point at the leading edge of the inlet. Positive values represent the upper surface of the inlet, while negative values represent the lower surface. The line graph on the right shows the predicted and target temperature results at z = 0 m (the 12 o’clock position as shown in Figure 1). The predicted results align well with the target results. The predicted values and change patterns at various positions on the anti-icing cavity surface are accurately predicted.

4.5. Comparative Analysis of POD-Alexnet and MCG

This section compares two proposed methods for predicting the surface temperature distribution through experiments. The results indicate that both methods achieve high prediction accuracy but emphasize different aspects.
Figure 14 compares the predicted absolute errors of the two methods, POD-AlexNet and MCG, using the same samples. As indicated by the comparative figures, the error distribution of the POD-AlexNet is relatively even, with no particular local absolute errors standing out significantly. In contrast, the prediction of the MCG exhibits regions with high-error points. However, aside from these regions, MCG’s overall error distribution is lower than that of POD-AlexNet. Further analysis suggests that the errors in the POD-AlexNet primarily result from truncating the basis modes and inaccuracies in the neural network’s prediction of fitting coefficients. Since each basis mode captures specific temperature distribution features across the entire three-dimensional surface, especially the first few modes, errors in the fitting coefficients propagate across the surface, leading to a relatively uniform distribution of prediction errors at each grid point.

5. Conclusions

Two models for predicting the surface temperature distribution of aircraft hot-air anti-icing systems are developed in this paper. Specifically, a prediction model is based on POD and AlexNet, termed POD-AlexNet, and another model is based on high-dimensional data, termed MCG. Additionally, extensive comparative experiments are conducted to validate the effectiveness of these models, leading to the following conclusions:
(1)
The POD-AlexNet model enables rapid predictions of the POD fitting coefficients and obtains anti-icing temperature distributions by reconstructing the POD basis modes based on these fitting coefficients. By selecting the appropriate neural network model, AlexNet, the RMSE of test samples is less than 2, the MRE is less than 0.5%, the MAE is less than 1.5, and the MPA is higher than 95%. In addition, the time cost for predicting each sample is about 1 ms, achieving fast and efficient prediction.
(2)
The MCG model enables the rapid and direct predictions of anti-icing surface temperature distributions. It achieves an RMSE of 1.75 on the test set, an MRE of 3.23‰, an MAE of 1.02, and an MPA of 96.97%. Additionally, the average single-sample prediction time is about 5.5 ms, which is a significant improvement compared to the traditional numerical simulation method, which takes hours or even days.
(3)
The error distribution of POD-AlexNet is relatively uniform; in contrast, the MCG exhibits localized high-error points. However, the overall error distribution of the MCG is significantly lower than that of the POD-AlexNet. In general, the POD-AlexNet and MCG models can provide faster predictions of anti-icing surface temperature distributions than traditional numerical simulation methods with acceptable error, which supports the design of aircraft hot-air anti-icing systems based on optimization methods such as genetic algorithms.

Author Contributions

Conceptualization, Q.Y. and X.Y.; methodology, Z.C. and J.G.; software, Z.C.; validation, Z.C. and Q.Y.; data curation, W.D.; writing—original draft preparation, Z.C.; visualization, Z.C. and Q.Y.; supervision, X.Y. and J.G.; funding acquisition, X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Science and Technology Major Project No. J2019-III-0010-0054 and National Natural Science Foundation of China No. 12172372.

Data Availability Statement

The data are unavailable due to privacy.

Acknowledgments

The authors would like to sincerely thank the relevant organizations and institutions for their support of this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Nomenclature

Dthe diameter of jet holes
Hflight altitude, km
Ljet hole spacing, mm
lthe total number of grids
mthe total number of samples
m ˙ single hot-air mass flow rate for a single jet hole, g / s
N s , i the number of grids with accurate predictions for the i-th sample
P tot total bleed air pressure, MPa
T tot total bleed air temperature, K
T temperature at far-field, K
T i , j the target value of i-th sample at j-th grid
T ^ i , j the predicted value of i-th sample at j-th grid
t t e s t the average prediction time for a single sample, ms
x pic x-coordinate of the piccolo tube center, mm
y pic y-coordinate of the piccolo tube center, mm
θ 0 the outflow direction angle of jet holes in the middle row, °
θ 1 relative angle between the positive y-axis direction jet hole and the middle row, °
θ 2 relative angle between the negative y-axis direction jet hole and the middle row, °
Θ the value of θ 0 in the baseline design
Abbreviations
AoAangle of attack, °
LWCliquid water content, g / m 3
MaMach number
MAEmean absolute error, K
MPAmean prediction accuracy
MREmean relative error
MVDmedian volumetric diameter, mm
RMSEroot mean square error

References

  1. Yang, Q.; Guo, X.; Zheng, H.; Dong, W. Single- and multi-objective optimization of an aircraft hot-air anti-icing system based on Reduced Order Method. Appl. Therm. Eng. 2023, 219, 119543. [Google Scholar] [CrossRef]
  2. Filburn, T. Anti-ice and Deice Systems for Wings, Nacelles, and Instruments. In Commercial Aviation in the Jet Era and the Systems That Make It Possible; Springer International Publishing: Cham, Switzerland, 2020; pp. 99–109. [Google Scholar] [CrossRef]
  3. Hannat, R.; Weiss, J.; Garnier, F.; Morency, F. Application of the Dual Kriging Method for the Design of Hot-Air-Based Aircraft Wing Anti-Icing System. Eng. Appl. Comput. Fluid Mech. 2014, 8, 530–548. [Google Scholar] [CrossRef]
  4. Hoffmann Domingos, R.; da Cunha Brandão Reis, B.; Martins da Silva, D.; Malatesta, V. Numerical Simulation of Hot-Air Piccolo Tubes for Icing Protection Systems. In Handbook of Numerical Simulation of In-Flight Icing; Habashi, W.G., Ed.; Springer International Publishing: Cham, Switzerland, 2020; pp. 1–30. [Google Scholar] [CrossRef]
  5. Dong, W.; Zhu, J.; Zheng, M.; Chen, Y. Thermal Analysis and Testing of Nonrotating Cone with Hot-Air Anti-Icing System. J. Propuls. Power 2015, 31, 1–8. [Google Scholar] [CrossRef]
  6. Dong, W.; Zheng, M.; Zhu, J.; Lei, G.; Zhao, Q. Experimental Investigation on Anti-Icing Performance of an Engine Inlet Strut. J. Propuls. Power 2017, 33, 379–386. [Google Scholar] [CrossRef]
  7. Guo, Z.; Zheng, M.; Yang, Q.; Guo, X.; Dong, W. Effects of flow parameters on thermal performance of an inner-liner anti-icing system with jets impingement heat transfer. Chin. J. Aeronaut. 2021, 34, 119–132. [Google Scholar] [CrossRef]
  8. Liu, Y.; Yi, X. Investigations on Hot Air Anti-Icing Characteristics with Internal Jet-Induced Swirling Flow. Aerospace 2024, 11, 270. [Google Scholar] [CrossRef]
  9. Ni, Z.; Liu, S.; Zhang, J.; Wang, M.; Wang, Z. Influnce of environment parameters on anti-icing heat load for aircraft. J. Aerosp. Power 2021, 36, 8–14. [Google Scholar] [CrossRef]
  10. Mao, H.; Lin, X.; Li, Z.; Shen, X.; Zhao, W. Anti-Icing System Performance Prediction Using POD and PSO-BP Neural Networks. Aerospace 2024, 11, 430. [Google Scholar] [CrossRef]
  11. Jung, S.; Raj, L.P.; Rahimi, A.; Jeong, H.; Myong, R.S. Performance evaluation of electrothermal anti-icing systems for a rotorcraft engine air intake using a meta model. Aerosp. Sci. Technol. 2020, 106, 106174. [Google Scholar] [CrossRef]
  12. Abdelghany, E.S.; Farghaly, M.B.; Almalki, M.M.; Sarhan, H.H.; Essa, M.E.S.M. Machine Learning and IoT Trends for Intelligent Prediction of Aircraft Wing Anti-Icing System Temperature. Aerospace 2023, 10, 676. [Google Scholar] [CrossRef]
  13. Ran, L.; Xiong, J.; Zhao, Z.; Zuo, C.; Yi, X. Prediction of Surface Temperature Change Trend of Electric Heating Anti-icing and De-icing Based on Machine Learning. Equip. Environ. Eng. 2021, 18, 29–35. [Google Scholar]
  14. McKay, M.D.; Beckman, R.J.; Conover, W.J. Comparison of Three Methods for Selecting Values of Input Variables in the Analysis of Output from a Computer Code. Technometrics 1979, 21, 239–245. [Google Scholar] [CrossRef]
  15. Yang, Q.; Zheng, H.; Guo, X.; Dong, W. Experimental validation and tightly coupled numerical simulation of hot air anti-icing system based on an extended mass and heat transfer model. Int. J. Heat Mass Transf. 2023, 217, 124645. [Google Scholar] [CrossRef]
  16. Ruan, d.R.; Kevin, J.; Barbara, H. Aerodynamic design of an electronics pod to maximise its carriage envelope on a fast-jet aircraft. Aircr. Eng. Aerosp. Technol. 2024, 96, 10–18. [Google Scholar] [CrossRef]
  17. Gutierrez-Castillo, P.; Thomases, B. Proper Orthogonal Decomposition (POD) of the flow dynamics for a viscoelastic fluid in a four-roll mill geometry at the Stokes limit. J. Non-Newton. Fluid Mech. 2019, 264, 48–61. [Google Scholar] [CrossRef]
  18. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  19. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  20. Boureau, Y.L.; Ponce, J.; LeCun, Y. A theoretical analysis of feature pooling in visual recognition. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 111–118. [Google Scholar]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  24. Rumelhart, D.E.; Hintont, G.E.; Williams, R.J. Learning by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  25. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  26. Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  27. Zhu, Y.; Yang, J.; Zhong, S.; Zhu, W.; Li, Z.; Wei, T.; Li, Y.; Gu, T. Research on Temperature Forecast Correction by Dynamic Weight Integration Based on Multi-neural Networks. J. Trop. Meteorol. 2024, 40, 156. [Google Scholar] [CrossRef]
  28. Lippmann, R.P. An introduction to computing with neural nets. ACM SIGARCH Comput. Archit. News 1988, 16, 7–25. [Google Scholar] [CrossRef]
  29. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2022, arXiv:1312.6114. [Google Scholar]
  30. Velarde, G.; Brañez, P.; Bueno, A.; Heredia, R.; Lopez-Ledezma, M. An Open Source and Reproducible Implementation of LSTM and GRU Networks for Time Series Forecasting. Eng. Proc. 2022, 18, 30. [Google Scholar] [CrossRef]
  31. Yu, T.; Li, X.; Cai, Y.; Sun, M.; Li, P. S2-MLP: Spatial-Shift MLP Architecture for Vision. In Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 3615–3624. [Google Scholar] [CrossRef]
  32. Zafar, N.; Haq, I.U.; Chughtai, J.u.R.; Shafiq, O. Applying Hybrid Lstm-Gru Model Based on Heterogeneous Data Sources for Traffic Speed Prediction in Urban Areas. Sensors 2022, 22, 3348. [Google Scholar] [CrossRef]
  33. Mateus, B.C.; Mendes, M.; Farinha, J.T.; Assis, R.; Cardoso, A.M. Comparing LSTM and GRU Models to Predict the Condition of a Pulp Paper Press. Energies 2021, 14, 6958. [Google Scholar] [CrossRef]
  34. Mienye, I.D.; Swart, T.G.; Obaido, G. Recurrent Neural Networks: A Comprehensive Review of Architectures, Variants, and Applications. Information 2024, 15, 517. [Google Scholar] [CrossRef]
  35. Zhang, Y.; Carvalho, D.; Valentino, M.; Pratt-Hartmann, I.; Freitas, A. Improving Semantic Control in Discrete Latent Spaces with Transformer Quantized Variational Autoencoders; Association for Computational Linguistics: St. Julian’s, Malta, 2024; pp. 1434–1450. [Google Scholar]
  36. Rocha, M.B.; Krohling, R.A. VAE-GNA: A variational autoencoder with Gaussian neurons in the latent space and attention mechanisms. Knowl. Inf. Syst. 2024, 66, 6415–6437. [Google Scholar] [CrossRef]
Figure 1. Turbofan engine nacelle model and hot-air anti-icing system. (a) Turbofan engine nacelle model; (b) Hot-air anti-icing system at the 12 o’clock position.
Figure 1. Turbofan engine nacelle model and hot-air anti-icing system. (a) Turbofan engine nacelle model; (b) Hot-air anti-icing system at the 12 o’clock position.
Aerospace 11 00930 g001
Figure 2. Design variables of inlet hot-air anti-icing system.
Figure 2. Design variables of inlet hot-air anti-icing system.
Aerospace 11 00930 g002
Figure 3. Structured mesh of inlet hot-air anti-icing system skin.
Figure 3. Structured mesh of inlet hot-air anti-icing system skin.
Aerospace 11 00930 g003
Figure 4. Structure of AlexNet suggested by Krizhevsky [21].
Figure 4. Structure of AlexNet suggested by Krizhevsky [21].
Aerospace 11 00930 g004
Figure 5. Structure of GRU suggested by [26].
Figure 5. Structure of GRU suggested by [26].
Aerospace 11 00930 g005
Figure 6. The technical workflow diagram.
Figure 6. The technical workflow diagram.
Aerospace 11 00930 g006
Figure 7. The curve of energy ratio relative to the number of POD modes.
Figure 7. The curve of energy ratio relative to the number of POD modes.
Aerospace 11 00930 g007
Figure 8. The data flow diagram.
Figure 8. The data flow diagram.
Aerospace 11 00930 g008
Figure 9. The network structure of MCG.
Figure 9. The network structure of MCG.
Aerospace 11 00930 g009
Figure 10. The variation in loss values during the ROM training process.
Figure 10. The variation in loss values during the ROM training process.
Aerospace 11 00930 g010
Figure 11. Comparison between target results and predicted results of POD-AlexNets. (a) Comparison of testing sample #50; (b) Comparison of testing sample #210; (c) Comparison of testing sample #225; (d) Comparison of testing sample #428.
Figure 11. Comparison between target results and predicted results of POD-AlexNets. (a) Comparison of testing sample #50; (b) Comparison of testing sample #210; (c) Comparison of testing sample #225; (d) Comparison of testing sample #428.
Aerospace 11 00930 g011
Figure 12. The variation in loss values during the high-dimensional model training process.
Figure 12. The variation in loss values during the high-dimensional model training process.
Aerospace 11 00930 g012
Figure 13. Comparison between target results and predicted results of MCG. (a) Comparison of testing sample #50; (b) Comparison of testing sample #210; (c) Comparison of testing sample #225; (d) Comparison of testing sample #428.
Figure 13. Comparison between target results and predicted results of MCG. (a) Comparison of testing sample #50; (b) Comparison of testing sample #210; (c) Comparison of testing sample #225; (d) Comparison of testing sample #428.
Aerospace 11 00930 g013
Figure 14. Absolute error of POD-AlexNet and MCG. (a) Comparison of testing sample #50; (b) Comparison of testing sample #210; (c) Comparison of testing sample #225; (d) Comparison of testing sample #428.
Figure 14. Absolute error of POD-AlexNet and MCG. (a) Comparison of testing sample #50; (b) Comparison of testing sample #210; (c) Comparison of testing sample #225; (d) Comparison of testing sample #428.
Aerospace 11 00930 g014
Table 1. The range of design variables for inlet hot-air anti-icing system.
Table 1. The range of design variables for inlet hot-air anti-icing system.
Design
Variables
L/D x pic /D y pic /D θ 0 / Θ θ 1 / Θ θ 2 / Θ
Baseline14.60010.860.73
Range[10, 20][−10, 7.5][−10, 7.5][0.54, 1.08][0.54, 1.08][0.54, 1.08]
Table 2. Design conditions for inlet hot-air anti-icing system.
Table 2. Design conditions for inlet hot-air anti-icing system.
HAoAMa T MVDLWC P tot T tot m ˙ single
6 km0.427263.55 K20 μm0.43 g/ m 3 0.25 MPa555 K1.33 g/s
Table 3. Performance evaluation results of methods based on ROM.
Table 3. Performance evaluation results of methods based on ROM.
NetworksRMSEMREMAEMPA t test
POD-MLP3.270.80%2.4488.25%4.0 ms
POD-LSTEM11.192.46%7.7457.28%0.7 ms
POD-GRU8.982.07%6.4558.81%0.4 ms
POD-VAE4.160.93%2.8783.57%0.4 ms
POD-UNet4.411.15%3.4378.48%1.2 ms
POD-ResNet2.810.69%2.1190.86%0.5 ms
POD-AlexNet1.990.47%1.4595.83%1.0 ms
Note: Bold shows the performance evaluation results of the model proposed in this study.
Table 4. Performance evaluation results of methods based on high-dimensional model.
Table 4. Performance evaluation results of methods based on high-dimensional model.
NetworksRMSEMREMAEMPA t test
LeNet2.855.26‰1.6691.76%7.5 ms
AlexNet2.804.98‰1.5992.23%8.5 ms
UNet2.845.26‰1.6691.82%1.3 ms
VAE3.626.40‰2.0588.46%4.6 ms
LSTM-3L2.564.68‰1.4893.18%5.4 ms
GRU-3L2.474.56‰1.4593.64%5.5 ms
MCG1.753.23‰1.0296.97%5.5 ms
Note: Bold shows the performance evaluation results of the model proposed in this study.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chu, Z.; Geng, J.; Yang, Q.; Yi, X.; Dong, W. Prediction of Temperature Distribution on an Aircraft Hot-Air Anti-Icing Surface by ROM and Neural Networks. Aerospace 2024, 11, 930. https://doi.org/10.3390/aerospace11110930

AMA Style

Chu Z, Geng J, Yang Q, Yi X, Dong W. Prediction of Temperature Distribution on an Aircraft Hot-Air Anti-Icing Surface by ROM and Neural Networks. Aerospace. 2024; 11(11):930. https://doi.org/10.3390/aerospace11110930

Chicago/Turabian Style

Chu, Ziying, Ji Geng, Qian Yang, Xian Yi, and Wei Dong. 2024. "Prediction of Temperature Distribution on an Aircraft Hot-Air Anti-Icing Surface by ROM and Neural Networks" Aerospace 11, no. 11: 930. https://doi.org/10.3390/aerospace11110930

APA Style

Chu, Z., Geng, J., Yang, Q., Yi, X., & Dong, W. (2024). Prediction of Temperature Distribution on an Aircraft Hot-Air Anti-Icing Surface by ROM and Neural Networks. Aerospace, 11(11), 930. https://doi.org/10.3390/aerospace11110930

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop