4.2. DNN Model Construction and Analysis (Divergence, Convergence, and Stability Boundaries)
Using the linear aeroelastic equations of motion for three-dimensional airfoils from
Section 2, we established a database with a total of 707,472,000 samples. Training on the entirety of this data would overwhelm the hardware, leading to memory issues and excessively long training times. Please see
Figure 12 for the message of memory error.
To manage the data volume, we categorized the data into three different flight outcomes: 0 for divergence, 1 for stability boundary, and 2 for convergence. The sampled data were divided into a 70% training set, a 25% validation set, and the remaining 5% as a test set. Our study utilized a significantly large dataset, consisting of 707,472,000 data points. Given the substantial size of the dataset, even a smaller percentage can represent a large absolute number of data points. In our case, 5% of the dataset equates to approximately 35,373,600 data points. This is a sufficiently large sample to ensure the testing set is representative of the overall data distribution and provides a reliable measure of model performance. The choice of 5% was made with the consideration that it provides an adequate representation of the data’s variability, covering various flight conditions, panel configurations, and other factors relevant to aeroelastic behavior. This ensures that the test set encompasses a broad spectrum of scenarios that the model may encounter. Training and evaluating deep learning models, particularly those with complex architectures like LSTM-NN, on such a large dataset demands significant computational resources and time. By using 5% of the data for testing, we balanced the need for a representative test set with the practical considerations of computational efficiency and resource utilization. A smaller test set size allowed for quicker evaluation cycles, enabling more frequent and timely feedback on model performance. This facilitated iterative improvements to the model and more efficient use of computational resources without compromising the robustness of the evaluation. The primary focus of our study was to investigate the efficacy of deep learning models in predicting panel flutter, rather than the exhaustive evaluation of every possible test set size. By using a 5% test set, we ensured that the core research objectives were met while maintaining computational efficiency and timely feedback for model improvements.
We constructed the DNN with an input layer, a hidden layer, and an output layer. A ReLU function was used as the activation function in the hidden layer. Since our training goal involved three classifications, we used Softmax as the activation function in the output layer. The cross-entropy loss function was chosen for loss calculation. Initially, we set up 10 neurons, with the batch size and epochs adjusted and fixed at 15,000 and 1000, respectively. Although the overall model accuracy reached 88.29%, the accuracy for predicting stable boundaries was only 30.56%. The one-layer DNN training model architecture is shown in
Figure 13, 1-layer DNN model accuracy is shown in
Figure 14, and 1-layer DNN model loss is shown in
Figure 15. The prediction results are shown in
Figure 16.
The situation arises because there are relatively fewer data points in scenario 1, resulting in this outcome. After various attempts, we decided to adjust the weights using class weights to address the problem of data imbalance. The goal was to ensure the model could better predict the minority class and highlight the importance of different classes, thereby improving model performance and prediction accuracy. However, improper setting of proportions can lead to overfitting and a decrease in accuracy. Ultimately, the weights we decided to use are shown in
Figure 17.
After adjusting the weights, there was a noticeable improvement in the training data for scenario 1, as shown in
Figure 18.
Due to the prediction accuracies for scenarios 0 and 2 already exceeding 90%, we will directly compare the prediction accuracy of scenario 1 across different hidden layers. We experimented with different numbers of hidden layers, using ReLU as the activation function after each hidden layer, and Softmax as the activation function for the output layer. The accuracy results are shown in
Figure 19.
Based on the displayed prediction accuracy for scenario 1 in
Figure 19, it is evident that the optimal number of hidden layers is six. After determining the number of hidden layers, we proceeded to experiment with different numbers of neurons to identify the best deep neural network model, as shown in
Figure 20.
From
Figure 20, it can be inferred that selecting appropriate neurons and hidden layers is crucial as it significantly affects the training outcomes. When using 90 neurons per layer, the training prediction achieved an accuracy of 98.89%. Therefore, we chose a DNN model with six hidden layers and 90 neurons each.
Figure 21 depicts the optimal DNN model architecture for this study. The training accuracy, loss, and prediction results are shown in
Figure 22,
Figure 23 and
Figure 24, respectively.
4.3. LSTM Model Establishment and Analysis (Divergence, Convergence, and Stable Boundary)
The primary reason for choosing LSTM networks is their ability to effectively handle temporal dependencies and sequential data. Panel flutter phenomena in aerospace structures exhibit significant temporal correlations due to the dynamic nature of the aeroelastic interactions over time. LSTMs are specifically designed to capture and model these long-term dependencies and sequential patterns, which are crucial for accurate predictions in such dynamic systems. With regard to GPR and shallow ANNs, while GPR and shallow ANNs are powerful tools for regression tasks, they are generally not optimized for handling sequential data with complex temporal dependencies. GPR assumes a fixed covariance structure and may struggle with large datasets due to computational constraints. Shallow ANNs, on the other hand, may not capture the intricate temporal relationships as effectively as LSTMs. The LSTM networks are highly effective at modeling complex non-linear relationships within the data, which are inherent in the aeroelastic behaviors of panel flutter. The recurrent nature of LSTMs allows them to maintain and update memory of previous inputs, enabling them to learn and represent the intricate dynamics of the system. The GPR, while flexible, may not scale well with the large dataset used in our study (707,472,000 data points) and may encounter difficulties in modeling very complex non-linear relationships. Shallow ANNs can model non-linearities to some extent but might require significantly more neurons and layers to achieve the same level of performance as LSTMs, which could lead to overfitting and other challenges. The ability of LSTMs to leverage temporal information resulted in more robust and accurate predictions of panel flutter under various flight conditions. Although LSTMs typically require longer training times due to their recurrent structure, the trade-off is justified by their superior performance and generalization capabilities. Our study included measures to balance training time and model accuracy, ensuring efficient training processes. The LSTM networks have been successfully applied to various time-series prediction tasks in engineering and other fields, demonstrating their effectiveness in modeling dynamic systems. By applying LSTMs to predict panel flutter, we leverage a proven technique for capturing the temporal evolution of aeroelastic responses, aligning with the specific requirements of aerospace engineering applications.
Following the methodology used in
Section 4.2, the data were split into 70% for the training set, 25% for the validation set, and the remaining 5% for the test set. For the LSTM architecture, we adopted a simple configuration consisting of one input layer, one hidden layer, and one output layer. We set 10 neurons per layer and a batch size of 15,000. The model underwent training for 1000 epochs with class weights applied similarly. In each hidden layer, ReLU was employed as the activation function, while Softmax was used in the output layer. The model achieved an accuracy of 96.07%, with a prediction accuracy of 86.11% specifically for stable boundaries. These results are illustrated in
Figure 25,
Figure 26,
Figure 27 and
Figure 28.
We use the method from
Section 4.2, keeping the number of neurons fixed at 10, and test various hidden layer configurations to identify the optimal training model. The results are shown in
Figure 29.
Based on the assessment in
Figure 29, we ultimately decided to use three LSTM hidden layers. Keeping these three layers fixed, we then adjusted the number of neurons to find the optimal model.
From
Figure 30, it can be seen that having 120 neurons in the three LSTM hidden layers yields the highest accuracy. Therefore, we chose to use 120 neurons for our long short-term memory deep learning model.
Figure 31 shows the architecture of the optimal long short-term memory model and the training accuracy, loss, and prediction results are shown in
Figure 32,
Figure 33 and
Figure 34.
4.4. LSTM-NN Model Development and Analysis (Divergence, Convergence, and Stability Boundary)
While the DNN and LSTM models both achieved over 98% accuracy, we aimed to see if even higher accuracy could be achieved. Therefore, we experimented with adding neural network hidden layers to the LSTM model. Using the same LSTM approach as mentioned in
Section 4.3, we first established the input layer and divided the data into 70% training set, 25% validation set, and the remaining 5% test set. Following
Section 4.2, we kept the class weight, number of neurons at 10, epoch, and batch size the same as in the previous sections.
Table 2 shows the accuracy, model loss, and stable boundary prediction accuracy for different hidden layer configurations. The model with three LSTM layers and one NN hidden layer achieved the highest accuracy and lowest loss.
After selecting the model architecture, we adjusted the number of neurons. As in the previous Sections, since the accuracy for scenarios 0 and 2 reached 99%, we compared the prediction accuracy for the stability boundary.
When the neuron count was set to 80, the prediction accuracy for the stability boundary reached 99.44%, as shown in
Figure 35. Therefore, the final model architecture set the number of neurons in the hidden layer to 80.
Figure 36 shows the LSTM-NN model architecture for this study, with training accuracy, loss, and prediction results shown in
Figure 37,
Figure 38 and
Figure 39.
In this study, we explored various deep learning architectures, specifically Deep Neural Networks (DNN), Long Short-Term Memory (LSTM), and Long Short-Term Memory Neural Networks (LSTM-NN). Each architecture was tested with different configurations of hidden layers and neurons to understand their impact on performance. We investigated shallow networks (fewer hidden layers) and deep networks (more hidden layers) to determine the optimal depth for each model (
Figure 19 and
Figure 29). Shallow networks are typically faster to train and less prone to overfitting, while deep networks can capture more complex patterns but may require more regularization. Our findings indicated that increasing the number of hidden layers generally improved the models’ ability to learn complex patterns from the data. However, beyond a certain point, additional layers led to diminishing returns in accuracy and increased the risk of overfitting. For example, DNN and LSTM models showed significant improvements up to 4–5 hidden layers but adding more layers resulted in marginal gains or even reduced performance due to overfitting. We also varied the number of neurons (
Figure 20 and
Figure 30) in each hidden layer to study their impact on performance. A higher number of neurons allows the model to learn more detailed representations of the input data but can also increase the risk of overfitting. Our present studies revealed that an optimal range of neurons provided the best balance between model complexity and performance. Models with too few neurons struggled to capture the underlying patterns, while those with too many neurons tended to overfit the training data. For DNNs, we observed that 3–5 hidden layers with 100–200 neurons per layer achieved the best performance. These configurations provided sufficient capacity to learn the complex aeroelastic behaviors without overfitting. LSTM models benefited from having between two and four LSTM layers with 50–150 neurons per layer. This architecture effectively captured temporal dependencies and complex non-linear relationships in the data, making it ideal for predicting flutter Mach numbers. The LSTM-NN models, which combine LSTM and feedforward layers, showed optimal performance with two to three LSTM layers followed by one to two fully connected layers, each with 100–150 neurons. This hybrid architecture leveraged the strengths of both LSTM and DNN models, providing high accuracy in both classification and regression tasks. We compared the performance of different architectures using key metrics such as Mean Squared Error (MSE), R-square values, and training/validation accuracy. These comparisons were conducted across multiple datasets and configurations to ensure robustness. The empirical results showed that while deeper models and those with more neurons generally provided better performance, there was a clear threshold beyond which additional complexity did not yield significant gains. For instance, LSTM-NN models consistently outperformed pure DNN and LSTM models in classification tasks, while LSTM models excelled in regression tasks with fewer layers and neurons. The choice of model architecture should consider the specific requirements of the task at hand. For tasks involving sequential data and temporal dependencies, LSTM and LSTM-NN models are more suitable. For tasks requiring high classification accuracy, hybrid architectures like LSTM-NN provide the best results.
4.6. Analysis of Flutter Mach Number Predictions Using Different Models
Using the number of hidden layers and neurons determined in
Section 4.2, we trained the data for the flutter speed of small panels separately. We used the same model architecture as before to train and compare the performance of different algorithms. The data were split into 70% training set, 25% validation set, and 5% test set. The input features were the aspect ratio of the small panel, the decay rate (
), longitudinal load (
), frequency (
), and total damping (
gT). The label was the Mach number at which panel flutter occurs.
Table 4 compares the Mean Squared Error (MSE) and R-square values of DNN, LSTM, and LSTM-NN algorithms.
From
Table 4, we can see that all three models have very low MSE values and R-square values very close to 1, indicating that the models have converged well and fit the data accurately. We then compared the prediction results, residual percentage, and relative error of the different algorithms for easier comparison. In
Figure 40, the horizontal axis represents the actual values and the vertical axis represents the predicted values. The dashed line represents the baseline where the actual value equals the predicted value. In
Figure 41, the horizontal axis represents the actual values and the vertical axis represents the residual percentage. In
Figure 42, the horizontal axis represents the actual values and the vertical axis represents the error percentage. From the comparison results in the figures, we can see that the LSTM performs better across all three figures and has lower MSE and higher R-square values. The average relative error is as low as 0.2182%.
Our dataset comprises 707,472,000 data points, capturing a wide range of flight conditions, panel configurations, and other variables. This extensive dataset helps ensure that the models are trained on diverse scenarios, reducing the risk of overfitting to any specific subset of data. Despite using 5% of the dataset for testing, this still amounts to approximately 35,373,600 data points, providing a substantial and representative test set to evaluate model performance. We divided our dataset into distinct training, validation, and test sets. This approach allows us to monitor the model’s performance on unseen data and adjust hyperparameters to reduce overfitting. In addition to the above techniques, we implemented a weighting strategy to adjust the importance of different data points during training (
Figure 17). Weighting helps in managing the influence of varying data points, ensuring that the model captures significant patterns without being overly influenced by outliers. This approach enhances the model’s ability to learn from diverse flight conditions and panel configurations, contributing to more balanced and generalized learning. We carefully selected the number of hidden layers and neurons to balance model complexity and performance (
Figure 19,
Figure 20,
Figure 29 and
Figure 30). This process involved experimenting with different architectures to find the optimal configuration that minimizes overfitting while maintaining high accuracy.
Solving the full-order model (Equation (19)) 707,472,000 times requires substantial computational effort, as it involves solving PDEs for each sample. This process is highly time-consuming, taking approximately 5.5 h on a CPU Intel i9-13900. Using CFD for grid points and fluid–solid coupling would significantly increase this time. In contrast, deep learning models are more efficient. Once trained, they can predict results for the entire dataset in less than 1 min, drastically reducing computational costs. To validate our deep neural network within aeroelasticity theory, we analyzed various flight conditions of small panels. After normalizing the input features,
Figure 43 shows the relationship between panel aspect ratio (
a/
b) and flutter occurrences. The horizontal axis represents the aspect ratio and the vertical axis shows the number of flutter occurrences. Longer panels (greater
a) better withstand deformation in the wind direction, absorbing energy from longitudinal forces and reducing flutter amplitude.
Figure 44 shows the relationship between the air density (altitude) and the number of occurrences of flutter. The horizontal axis represents the normalized air density (altitude) and the vertical axis represents the number of flutter occurrences. As the altitude increases, the air density decreases, which reduces the aeroelastic impact on the panel, thereby reducing the occurrence of flutter.
Figure 45 shows the relationship between the frequency (
) of the panel and the number of occurrences of flutter. The horizontal axis represents the frequency and the vertical axis represents the number of flutter occurrences. As the vibration frequency increases, most of the energy is absorbed by the vibration frequency, thus reducing the amplitude of flutter and the occurrence of flutter.
Figure 46 shows the relationship between the longitudinal compression (
) on the panel and the number of occurrences of flutter. The horizontal axis represents the longitudinal compression and the vertical axis represents the number of flutter occurrences. When the panel is subjected to a larger longitudinal force, the structure deforms more, making the panel more unstable and increasing the risk of flutter.
Figure 47 shows the relationship between the total damping (
gT) of the panel and the number of occurrences of flutter. The horizontal axis represents the total damping and the vertical axis represents the number of flutter occurrences. As damping increases, vibration energy is rapidly dissipated. When the amplitude of the panel decreases, the interaction between aerodynamic forces and the structure weakens, reducing the risk of flutter.
Figure 48 shows the decay rate (
) and the number of occurrences of flutter. The horizontal axis represents the decay rate and the vertical axis represents the number of flutter occurrences. A larger damping rate means that the vibration energy in the system dissipates faster. When the small plate is subjected to external disturbances, the vibration energy quickly decays, preventing the system from entering an unstable state.
Through these graphical explanations, we ensure that the data in this study align with physical principles, thereby confirming the reliability of the deep neural network models used in this research.
A description of the pros and cons of this method compared to modern numerical complex systems that allow very detailed coupled simulations is given as follows:
DL Models: Deep learning (DL) models offer significant computational efficiency and speed. Once trained, they can quickly predict panel flutter, making them suitable for real-time monitoring and design optimization applications.
Numerical Complex Systems: High-fidelity numerical simulations, such as CFD-FEM coupled systems, are computationally intensive and time-consuming, which can be a limitation for real-time applications;
DL Models: DL models are highly scalable and can be trained on large datasets, as demonstrated in our study with over 700,000,000 data points. This scalability ensures robustness and generalizability to various flight conditions and panel configurations.
Numerical Complex Systems: These systems, while highly accurate, require substantial computational resources, limiting their scalability and practicality for extensive parameter studies or real-time applications;
DL Models: Our study demonstrates that DL models, particularly LSTM-NN, achieve high predictive accuracy in capturing flutter phenomena. The results in
Figure 42,
Figure 43,
Figure 44,
Figure 45,
Figure 46 and
Figure 47 are consistent with the physical phenomena of flutter, proving the sufficiency of our theoretical model in capturing conditions that constitute flutter.
Numerical Complex Systems: These systems provide highly accurate and detailed simulations, which are essential for critical design and safety assessments.
DL Models: While DL models offer efficient predictions, they may lack the detailed physical insights that numerical simulations provide. This could potentially limit their explanatory power in understanding the underlying mechanisms of aeroelastic phenomena. However, it is important to note that the consistency of our results with known physical phenomena, as shown in
Figure 42,
Figure 43,
Figure 44,
Figure 45,
Figure 46 and
Figure 47, and the use of a typical panel plate aeroelastic equation (Reference [
30]) demonstrate that our theoretical model is sufficient as a database for machine learning. This indicates that the perceived lack of physical insight is not necessarily a limitation in this context.
Numerical Complex Systems: High-fidelity simulations offer detailed insights into physical phenomena, aiding in a deeper understanding of the underlying mechanics and interactions. These insights are crucial for developing comprehensive and accurate aeroelastic models.
While deep learning models provide significant advantages in terms of computational efficiency, scalability, and predictive accuracy, high-fidelity numerical simulations offer unmatched physical insights. Our study bridges this gap by demonstrating that the theoretical model used is sufficient to capture the conditions leading to flutter, providing a robust database for machine learning. This hybrid approach leverages the strengths of both methodologies, ensuring efficient and accurate predictions while maintaining physical relevance.
We used the constitutive equations of a simply supported panel due to their simplicity and well-understood behavior, making them ideal for initial model validation. This approach allowed us to focus on the effectiveness of deep learning models without the added complexity of wing flutter equations. The results in
Figure 42,
Figure 43,
Figure 44,
Figure 45,
Figure 46 and
Figure 47 are consistent with physical flutter phenomena and validate the theoretical model. Reference [
30] also supports the use of this typical aeroelastic equation, proving its adequacy for machine learning. Our study’s input features included the panel aspect ratio, Mach number, air density, and decay rate, which were chosen for their relevance to aeroelastic behavior. In our sensitivity analysis, the Mach number had the highest impact on model predictions, highlighting its crucial role in aeroelastic phenomena. The panel aspect ratio significantly affected model performance, altering structural dynamics and flutter characteristics. Air density showed a moderate impact, which was more pronounced in high-speed conditions. The decay rate had the least impact but influenced model fine-tuning.
The analysis underscores the importance of including the Mach number and panel aspect ratio, which are relevant to guiding feature selection and potentially reducing model complexity. The accurate prediction of panel flutter ensures the structural integrity of aircraft components, providing early warnings and reducing the need for extensive physical testing. Our models can explore a wider range of design parameters and flight conditions, enhancing design optimization and robustness. Applications extend to wind turbine blades, the automotive industry, and structures like bridges and tall buildings, predicting and mitigating aeroelastic instabilities to improve safety and performance.