Next Article in Journal
Hybrid Artificial Intelligence Strategies for Drone Navigation
Previous Article in Journal
Deep Learning in Finance: A Survey of Applications and Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Airfoil Shape Generation and Feature Extraction Using the Conditional VAE-WGAN-gp

by
Kazuo Yonekura
*,
Yuki Tomori
and
Katsuyuki Suzuki
Department of Systems Innovation, The University of Tokyo, Tokyo 113-8656, Japan
*
Author to whom correspondence should be addressed.
AI 2024, 5(4), 2092-2103; https://doi.org/10.3390/ai5040102
Submission received: 13 September 2024 / Revised: 17 October 2024 / Accepted: 18 October 2024 / Published: 28 October 2024

Abstract

:
A machine learning method was applied to solve an inverse airfoil design problem. A conditional VAE-WGAN-gp model, which couples the conditional variational autoencoder (VAE) and Wasserstein generative adversarial network with gradient penalty (WGAN-gp), is proposed for an airfoil generation method, and then, it is compared with the WGAN-gp and VAE models. The VAEGAN model couples the VAE and GAN models, which enables feature extraction in the GAN models. In airfoil generation tasks, to generate airfoil shapes that satisfy lift coefficient requirements, it is known that VAE outperforms WGAN-gp with respect to the accuracy of the reproduction of the lift coefficient, whereas GAN outperforms VAE with respect to the smoothness and variations of generated shapes. In this study, VAE-WGAN-gp demonstrated a good performance in all three aspects. Latent distribution was also studied to compare the feature extraction ability of the proposed method.

1. Introduction

In mechanical design, it is desirable to design shapes that indicate the required performance. This task is called an inverse design problem. Recently, deep generative models such as generative adversarial networks (GANs) and variational autoencoders (VAEs) have been used for inverse designs [1,2]. Both GANs and VAEs can generate desired shapes that satisfy these requirements [3]. In VAEs, designers can choose the shapes to generate by analyzing latent vectors [2], which is difficult in GAN. In this study, to enable latent engineering in GAN models, a variational autoencoder generative adversarial network (VAEGAN) model, a combination of VAE and GAN is used for inverse designs.
One approach to an inverse problem and optimization is to use a surrogate model [4,5]. The surrogate model is constructed using data, and it outputs the performance parameters from the input shapes in a short time. The authors in [6] used surrogate models for turbine cooling hole optimization problem. However, the number of design parameters was large, and it was difficult to use it for inverse problems. Hence, they conducted a principal component analysis to reduce dimensionality, but since PCA was necessary, the shape diversity was low and accuracy was low. Hence, when a certain shape indicates a certain performance requirement, designers can search from the surrogate models. The surrogate model is used for inverse problems and optimization in industrial applications such as turbine cooling holes [7]. Deep neural networks are also used as surrogate models to predict performance [8,9]. However, the shapes obtained from surrogate models are like those obtained from a training dataset, and completely different shapes cannot be obtained.
Deep generative models have recently been used to solve inverse problems. The GAN [10] is a widely used generative model. GAN-based inverse design is utilized in various industrial fields [11,12,13,14,15]. Apart from these industrial applications, the airfoil design task is an important benchmark problem for inverse design method [16,17,18]. The performance of the output shape sometimes differ from the required labels. The desired accuracy depends on each task. In [19], a flow machinery is designed using DNN as a regression model. In the literature, the error was less than 5 to 10%. In the GAN-based industry problem [15], the target accuracy is 5 to 10%. Those target accuracy is indeed not high. Hence, in real applications, the method is used to obtain good initial solutions for further numerical optimization. For example, in turbine blade optimization [20], it needs a good initial solution to start, but how to obtain such a solution is not trivial.
In order to improve accuracy, one approach is to couple physical models to NN models. The authors in [21] proposed physics-informed neural networks (PINNs) that use the residual of the physical equations in the loss functions. When PINNs are used, the physical model has to be implemented within the NN model to conduct backpropagation through the physical calculations. Therefore, general purpose software cannot be used with PINNs. However, in the inverse design, general purpose software is used in most cases. In the airfoil design benchmark problem, XFoil [22], the general purpose software for subsonic airfoils, is utilized, and hence, PINNs cannot be used. To overcome this limitation, refs. [23,24] proposed and utilized PG-GAN to the inverse design.
In airfoil designs, because the aim is to obtain shapes that satisfy specific aerodynamic requirements, it is important to obtain a smooth shape such that a numerical analysis, such as fluid analysis, can be conducted. For this purpose, a conditional GAN is used for the airfoil inverse design [25,26]. However, one of the issues with GAN models in airfoil generation is that the output shapes are not smooth, and fluid analysis is not applicable. Savitzky–Golay smoothing filters [27] are used in [26], but the filter changes the trailing edge. To overcome this issue, a conditional Wasserstein GAN with a gradient penalty (cWGAN-gp) [28,29] is employed, and it outputs a smoother airfoil than the ordinary GAN models as reported in [3]. The uncertainty of generated shapes is also evaluated [30]. However, GAN models generate data from random latent vectors, and the output shape is chosen randomly.
The VAE [31] is another generative model. It is used to generate three-dimensional models [32,33] and the airfoil [1]. The VAE is also used for airfoil generation tasks; ref. [2] reported that two different types of shapes are mixed when using VAE models, and such mixing is not possible in the GAN models. One of the differences between the GAN and VAE lies in the latent space. In the GAN model, the generator generates data from a random latent vector or noise vector, whereas in the VAE model, feature vectors are extracted from the training dataset, and data are embedded in the latent space. Hence, in the GAN model, the output shape is chosen randomly using the random noise vector.
VAEGAN [34] is a GAN-based neural network that uses a VAE model as its generator. The VAE part of the VAEGAN enables feature extraction from data, whereas the entire model is based on the GAN architecture. The VAEGAN was applied to the airfoil generation task in [25], but although the latent space of the VAEGAN is essentially different from that of the GAN, the latent space has not been discussed in airfoil generation tasks. In this study, the VAEGAN was coupled with cWGAN-gp to increase the smoothness of shapes, and it was then applied for the airfoil generation task. The latent space was compared with the cWGAN-gp model; from the comparison, the latent space is well-organized.
The remainder of this paper is organized as follows: Section 2 introduces the GAN, VAE, and VAEGAN models. The conditional VAEGAN model with the Wasserstein distance and gradient penalty for airfoil generation is proposed in Section 3. Numerical experiments are presented in Section 4, and conclusions are presented in Section 5. In order to follow the notation of existing literature, G and f w both represent the generator network, and D and g θ represent the discriminator network.

2. Conditional GAN, VAE, and VAEGAN Models

2.1. Conditional GAN and WGAN-gp

GAN [10] consists of a generator network G and discriminator network D as illustrated in Figure 1. The generator generates data that mimic the training data, which are also called real data, and the discriminator distinguishes real data from fake data. The generator and discriminator minimize and maximize the loss function, respectively.
min G max D L G A N ( G , D ) .
The loss function is defined as follows:
L G A N ( G , D ) = E x p r ( x ) log ( D ( x ) ) + E z p z ( z ) log 1 D ( G ( z ) ) ,
where x denotes real data, and p r ( x ) is the probability distribution of the real data. z denotes a random noise vector sampled from probability distribution p z ( z ) . G and D denote the generator and discriminator networks, respectively. G ( z ) is the output from the generator, which is referred to as fake data.
Although many successful applications of GAN have been reported [35], an instability in training GAN was also reported [36,37]. WGAN-gp [29] is a solution for improving the stability. WGAN-gp uses the earth-mover’s (EM) distance instead of the Jensen–Shannon (JS) divergence in GAN. EM distance W ( · , · ) is a distance between two probability distributions defined by
W ( p r , p g ) = inf γ Π ( p r , p g ) E x , x γ [ x x ] ,
where p r and p g are probability distributions of train data and generated data, respectively. Then, the loss function of WGAN-gp is
L W G A N g p ( f w , g θ ) = W ( p r , p g ) + λ E x p r [ x f w ( x ) 1 2 ] ,
where f w and g θ represent the generator and discriminator, respectively.

2.2. Conditional VAE

The VAE [31] is an encoder–decoder-type deep neural network. The VAE model is shown in Figure 2. The encoder embeds features into the latent space, and the decoder reconstructs the data from the latent space:
f e n c ( x ) = μ , σ 2 ,
z N ( μ , σ 2 ) ,
f d e c ( z ) = x ,
where μ and σ 2 are the mean and variance vectors of the latent space. The loss function of the VAE model is
L V A E = L llike + L prior
= E x x x 2 + D KL ( q ( z | x ) | | p ( z ) ) ,
where D K L ( q ( z | x ) | | p ( z ) ) is the Kullback–Leibler (KL) divergence between the distribution of samples in latent space q ( z | x ) and the prior p ( z ) . The standard normal distribution is used as prior p ( z ) .

2.3. Conditional VAEGAN

A VAEGAN couples the VAE and GAN models, and it consists of an encoder, decoder, and discriminator, as illustrated in Figure 3. The encoder extracts features from the data and embeds them into the latent space, whereas the decoder generates data from latent vectors. When generating new data in the VAEGAN, latent vectors are specified, and the decoder is processed from the latent vectors. The loss function of the VAEGAN model is
L VAEGAN = L llike Dis + L GAN + L prior .
L llike Dis denotes the reconstruction loss represented by the discriminator. L GAN and L prior are the GAN loss and prior loss defined in the GAN and VAE models, respectively.

3. CVAE -WGAN-gp Model for Airfoil Generation

3.1. Conditional VAE-WGAN-gp

VAEGAN is coupled with WGAN-gp to improve stability and is trained in a conditional manner. The loss function of VAE-WGAN-gp is expressed as
L VAE WGAN gp = L llike Dis + L WGAN gp + L prior .
The architecture of the model is the same as that in Figure 3. The training dataset was fed to the encoder with labels, and the data were embedded into the latent space. The prior of the latent space was employed as the standard normal distribution. The decoder reconstructs data from the latent vector and labels. The generated fake data, x , were fed into the discriminator with the label and training data. The discriminator distinguishes between real data and fake data. When generating new data, the latent vector is inputted into the decoder, and the decoder outputs new data.
The network architectures are shown in Figure 4. The encoder consisted of five layers with 512, 256, 128, and 64 nodes. The latent dimension was fixed to 4 because [1] reported that the sensitivity of the latent dimensionality is low for this airfoil design task. The decoder also consisted of five layers with 64, 128, 256, 512, and 496 nodes. The discriminator consisted of three layers with 512, 256, and 1 node. A leaky rectified linear unit (leaky ReLU) [38] was employed as the activation function. The dimension of the latent space was grid-searched, and 4 was chosen. The Adam optimizer was used, the learning rate was set as 0.0001 , and the model was trained for 50,000 epochs.

3.2. Airfoil Generation Framework

First, a training dataset was prepared. The dataset consisted of shape data and the lift coefficients C L , the corresponding performance index, is calculated. The data preparation procedure is explained in Section 4.1. The CVAE-WGAN-gp model was trained using this dataset. The lift coefficient C L was used as label to be input into the model. Once the training was completed, the decoder was used to generate new data. The decoder inputs latent vectors and labels and outputs new data. When generating the data, latent vectors were sampled using a standard normal distribution. The aerodynamic performance of the generated data was evaluated. In the numerical experiments, XFoil [22] was used to evaluate airfoils. The aim of the inverse design is to obtain shapes that satisfy the required labels. However, the loss function of the CVAE-WGAN-gp model does not consider the reproduction of aerodynamic performance. Therefore, the reproduction of the aerodynamic performance is not guaranteed by the model. To evaluate the model, the error in aerodynamic performance must be evaluated. Adversarial attacks on the machine learning models are an important topic [39]. However, in this study, we assume that the dataset is under full control by a designer, and we do not consider adversarial attacks.

4. Numerical Experiments

4.1. Dataset

The airfoil dataset was constructed using the NACA 4-digit airfoil dataset [40]. The 4-digit airfoil is defined by three parameters: maximum camber, position of maximum camber, and maximum thickness normalized by chord length. Totally, 10,000 airfoil shapes are generated by the definition. The lift coefficients C L are calculated using XFoil version 6.97 [22]. Among 10,000 shapes, the XFoil computation converges for only 3709 airfoils. Some airfoils indicate too high camber, which leads to unconvergence. The airfoil shape is represented by a set of points, where the number of points is 248 because the XFoil calculation requires more than 120 points. x and y coordinates of the points were assembled into one vector as x = x 1 , x 2 , , x 248 , y 1 , y 2 , , y 248 (Figure 5).
The histogram of the lift coefficients is shown in Figure 6. The histogram is close to the uniform distribution except that the frequency increases around C L ∼0.5 and decreases at C L ∼1.5. The coordinates x and the label C L are standardized before feeding into the neural networks.

4.2. Airfoil Generation

The computation is carried out on a machine manufactured by FUJITSU Client Computing Limited, Tokyo, Japan, equipped with Intel Core i7 CPU (2.8 GHz) with 32 GB memory. The wall clock time for training 50,000 epochs was 16,650 s, and less than 1 s for generating 100 shapes. The learning curve is shown in Figure 7. The proposed CVAE-WGAN-gp and conventional cWGAN-gp models were trained, and the generated shapes are shown in Figure 8. The red figures indicate that the XFoil calculations for the shapes did not converge, whereas the blue figures indicate otherwise.
Some of the generated shapes are similar to NACA airfoils. The examples of airfoils that are generated under the condition of C L = 0.5 are shown in Figure 9. The figure shows a reasonable solution with respect to the aeronautical point of view, except that some shapes indicate zig-zag curves. Such zig-zag curves lead to failure of flow analysis. Some shapes are weird from an aeronautical point of view; e.g., some figures in the bottom of Figure 8d have protrusions that prevent smooth flow.
To evaluate the accuracy of shape generation quantitatively, ref. [3] used following indices. In this study, we also use the same indices:
  • ϕ mean : Smoothness index.
  • MSE: Mean squared error of C L .
  • μ : Index of variety of generated shapes.
The smoothness index ϕ mean is the mean of the smoothness indices ϕ defined for each shape. ϕ is defined as
v k = x k + 1 x k , y k + 1 y k
ϕ = k = 1 N arccos v k v k + 1 v k v k + 1
If a shape is a complete circle, ϕ = 2 π , and if a shape contains zig-zag lines, ϕ indicates a larger value than 2 π . ϕ mean indicates the smoothness of the generated shapes, and a smaller value indicates better smoothness. ϕ mean of the training dataset is ϕ mean = 2.14 π . The number is slightly larger than 2 π because of the camber line of the airfoils; the centerline of many airfoils is an upward convex curve.
The MSE is a quantitative index of the lift coefficient error. The MSE is defined as the mean of C L r C L l 2 for all shapes whose XFoil calculation converges. This MSE is not included in the loss function because the reconstruction error in the loss function is the distance between the output shape and the training shape, whereas this MSE is the distance between the lift coefficients. μ is defined as the mean deviation of shapes from the mean shape, that is, g i g mean , where g i is a vector of a shape and g mean is the mean of g i .
The indices of the proposed CVAE-WGAN-gp model and other models are compared in Table 1. cWGAN-gp and CVAE-WGAN-gp achieved better scores than the others with respect to ϕ mean . cWGAN-gp and CVAE indicated better scores only for one or two indices, that is, cWGAN-gp was better in ϕ mean and μ , and CVAE was better in MSE. However, CVAE-WGAN-gp showed good scores for all the three indices.
It is reasonable that the MSE of CVAE-WGAN-gp and CVAE is almost similar because both models use the loss between the training data and generated data, that is, L llike and L llike Dis , which does not contain the loss of cWGAN-gp. Because loss L llike measures the direct distance between the training and generated data, the model can generate data similar to the training data, which leads to a lower MSE.
ϕ mean was smaller in the cWGAN-gp and CVAE-WGAN-gp models than in the CVAE model. A smaller ϕ mean was obtained owing to the Wasserstein distance, as shown in the difference between cGAN and cWGAN-gp. The discriminator measures the distance between the training data and generated data using the Wasserstein distance.
μ was significantly large in cWGAN-gp, and CVAE-WGAN-gp followed cWGAN-gp. In cWGAN-gp, some of the generated shapes were wired, as shown in Figure 8. The lift coefficients cannot be calculated for these wired shapes, as illustrated in red in Figure 8, which leads to a larger MSE value. CVAE-WGAN-gp indicated a larger μ than CVAE, whereas the MSEs of both models were almost equal.
It has been already reported that VAEGAN outperforms the GAN model [34], which is consistent with this study. CVAE-WGAN-gp exhibited good properties of both CVAE and cWGAN-gp, owing to its architecture and loss functions.

4.3. Latent Distribution

The latent distributions of cWGAN-gp and CVAE-WGAN-gp are shown in Figure 10. When discussing the latent space of GAN, Achour et al. [26] used the t-SNE [41] to reduce the latent dimension so that the latent space is visualized. We used the same t-SNE to visualize the latent space of CVAE-WGAN-gp. We note that t-SNE has some hyperparameters, and by changing the hyperparameters, the visualization result sometimes drastically changes. In this manuscript, the latent space with t-SNE is provided to convey the image of the difference in latent space. In the latent space of WGAN-gp, data with both large and small C L s are randomly distributed throughout the space. On the other hand, the latent space of CVAE-WGAN-gp has a clumped distribution of those with small C L s, and the small and large data are separated. This difference is due to the presence of the encoder. Let x and x be latent variables that satisfy x x ε } , c and c be the labels, and functions f and f VAEGAN be the generator of WGAN-gp and CVAE-WGAN-gp, respectively. Then, the generated shapes are written as f ( x , c ) and f ( x , c ) . In the case of WGAN-gp, if we focus on a point in the latent space, both data with large CL and data with small CL exist around that point, which are generally significantly different. It implies that f ( x , c ) f ( x , c ) is significantly large. In the case of CVAE-WGAN-gp, C L of the generated shapes from nearby two latent variables is close. It implies that f VAEGAN ( x , c ) f VAEGAN ( x , c ) is expected to be smaller than f ( x , c ) f ( x , c ) . Hence,
f ( x , c ) f ( x , c ) x x = f ( x , c ) f ( x , c ) x x f ( x , c ) f ( x , c ) f ( x , c ) f ( x , c ) 1 ,
holds, and consequently,
f ( x , c ) f ( x , c ) x x f ( x , c ) f ( x , c ) f ( x , c ) f ( x , c )
holds. The latent space distributions are quite different between WGAN-gp and CVAE-WGAN-gp. This difference arises from the encoder part of CVAE-WGAN-gp.
The generator or decoder generates new data from latent data. In cWGAN-gp, the generator generates data with both high C L and low C L from a similar latent variable. Because the label is also an input to the generator as well as the latent variable, the generator can output different shapes from the same latent variable. However, it is desirable if different data are located in different areas of the latent space. In the CVAE-WGAN-gp model, the decoder can generate different data from different latent variables, which results in a lower MSE than cWGAN-gp.

5. Conclusions

This study proposed CVAE-WGAN-gp, a model that generates an airfoil that satisfies required lift coefficients. The proposed model showed a better performance in terms of smoothness, reproduction of lift coefficients, and shape varieties, whereas CVAE and cWGAN-gp exhibited some but not all these three desired properties. The proposed model has properties of both the CVAE and cWGAN-gp models. The latent space was compared with both the proposed and cWGAN-gp models, and it was shown that the encoder of the CVAE-WGAN-gp model embeds data in the latent space, and the latent distributions are different from that of WGAN-gp.A better MSE is explained by the neatness of the latent space. The fact that different data are embedded in different latent areas helps the decoder to generate accurate data from the latent space. The hyper parameters have to be properly chosen to obtain better results.

Author Contributions

Conceptualization, K.Y.; methodology, K.Y.; software, Y.T.; validation, Y.T.; formal analysis, Y.T.; investigation, Y.T.; resources, K.Y.; data curation, Y.T.; writing—original draft preparation, K.Y.; writing—review and editing, K.Y.; visualization, Y.T. and K.Y.; supervision, K.Y. and K.S.; project administration, K.Y.; funding acquisition, K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by JSPS KAKENHI grant number JP21K14064 and JP23K13239.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available at https://github.com/utyone/airfoilVAEGAN (accessed on 12 September 2024).

Conflicts of Interest

K.Y. is a board member of MJOLNIR SPACEWORKS, Co., Ltd. The remaining authors declare that the research was concluded in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Yonekura, K.; Suzuki, K. Data-driven design exploration method using conditional variational autoencoder for airfoil design. Struct. Multidiscip. Optim. 2021, 64, 613–624. [Google Scholar] [CrossRef]
  2. Yonekura, K.; Wada, K.; Suzuki, K. Generating various airfoils with required lift coefficients by combining NACA and Joukowski airfoils using conditional variational autoencoders. Eng. Appl. Artif. Intell. 2022, 108, 104560. [Google Scholar] [CrossRef]
  3. Yonekura, K.; Miyamoto, N.; Suzuki, K. Inverse airfoil design method for generating varieties of smooth airfoils using conditional WGAN-gp. Struct. Multidiscip. Optim. 2022, 65, 173. [Google Scholar] [CrossRef]
  4. Braconnier, T.; Ferrier, M.; Jouhaud, J.C.; Montagnac, M.; Sagaut, P. Towards an adaptive POD/SVD surrogate model for aeronautic design. Comput. Fluids 2011, 40, 195–209. [Google Scholar] [CrossRef]
  5. Mack, Y.; Goel, T.; Shyy, W.; Haftka, R. Surrogate Model-Based Optimization Framework: A Case Study in Aerospace Design. In Evolutionary Computation in Dynamic and Uncertain Environments; Yang, S., Ong, Y.S., Jin, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 323–342. [Google Scholar]
  6. Yonekura, K.; Watanabe, O. A shape parameterization method using principal component analysis in application to shape optimization. J. Mech. Des. 2014, 136, 121401. [Google Scholar] [CrossRef]
  7. Nita, K.; Okita, Y.; Nakamata, C.; Kubo, S.; Yonekura, K.; Watanabe, O. Film cooling hole shape optimization using proper orthogonal decomposition. In Proceedings of the ASME Turbo Expo 2014: Turbine Technical Conference and Exposition, Düsseldorf, Germany, 16–20 June 2014. [Google Scholar] [CrossRef]
  8. Pfrommer, J.; Zimmerling, C.; Liu, J.; Kärger, L.; Henning, F.; Beyerer, J. Optimisation of manufacturing process parameters using deep neural networks as surrogate models. Procedia Cirp 2018, 72, 426–431. [Google Scholar] [CrossRef]
  9. Liang, L.; Liu, M.; Martin, C.; Sun, W. A deep learning approach to estimate stress distribution: A fast and accurate surrogate of finite-element analysis. J. R. Soc. Interface 2018, 15, 20170844. [Google Scholar] [CrossRef]
  10. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Cambridge, MA, USA, 8–13 December 2014; Volume 2, NIPS’14. pp. 2672–2680. [Google Scholar]
  11. Yonekura, K.; Omori, K.; Qi, X.; Suzuki, K. Designing ship hull forms using generative adversarial networks. arXiv 2023, arXiv:2311.05470. [Google Scholar]
  12. Li, H.; Zheng, Y.; Wu, X.; Cai, Q. 3D Model Generation and Reconstruction Using Conditional Generative Adversarial Network. Int. J. Comput. Intell. Syst. 2019, 12, 697–705. [Google Scholar] [CrossRef]
  13. Li, R.; Li, X.; Hui, K.H.; Fu, C.W. SP-GAN: Sphere-Guided 3D Shape Generation and Manipulation. ACM Trans. Graph. 2021, 40, 1–12. [Google Scholar] [CrossRef]
  14. Wu, J.; Zhang, C.; Xue, T.; Freeman, B.; Tenenbaum, J. Learning a Probabilistic Latent Space of Object Shapes via 3D Generative-Adversarial Modeling. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2016; Volume 29. [Google Scholar]
  15. Kato, N.; Suzuki, K.; Kondo, Y.; Suzuki, K.; Yonekura, K. Automotive motor rotor design synthesis using cWGAN-gp with distortion penalty. Res. Sq. 2024. [Google Scholar] [CrossRef]
  16. Chen, W.; Chiu, K.; Fuge, M. Airfoil Design Parameterization and Optimization using Bézier Generative Adversarial Networks. arXiv 2020, arXiv:2006.12496. [Google Scholar] [CrossRef]
  17. Sekar, V.; Zhang, M.; Shu, C.; Khoo, B.C. Inverse Design of Airfoil Using a Deep Convolutional Neural Network. AIAA J. 2019, 57, 993–1003. [Google Scholar] [CrossRef]
  18. Yilmaz, E.; German, B. Conditional Generative Adversarial Network Framework for Airfoil Inverse Design. In Proceedings of the AIAA AVIATION 2020 FORUM, Virtual. 15–19 June 2020. [Google Scholar]
  19. Saito, H.; Kanzaki, D.; Yonekura, K. Applications of machine learning in surge prediction for vehicle turbochargers. Mach. Learn. Appl. 2024, 16, 100560. [Google Scholar] [CrossRef]
  20. Yonekura, K.; Hattori, H.; Shikada, S.; Maruyama, K. Turbine blade optimization considering smoothness of the Mach number using deep reinforcement learning. Inf. Sci. 2023, 642, 119066. [Google Scholar] [CrossRef]
  21. Raissi, M.; Perdikaris, P.; Karniadakis, G. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  22. Drela, M. XFOIL: An Analysis and Design System for Low Reynolds Number Airfoils. In Proceedings of the Low Reynolds Number Aerodynamics, Notre Dame, IN, USA, 5–7 June 1989; Mueller, T.J., Ed.; Springer: Berlin/Heidelberg, Germany, 1989; Volume 54, Lecture Notes in Engineering. pp. 1–12. [Google Scholar]
  23. Wada, K.; Suzuki, K.; Yonekura, K. Physics-guided training of GAN to improve accuracy in airfoil design synthesis. Comput. Methods Appl. Mech. Eng. 2024, 421, 116746. [Google Scholar] [CrossRef]
  24. Yonekura, K. Physics-guided generative adversarial network to learn physical models. arXiv 2023, arXiv:2304.11488. [Google Scholar]
  25. Wang, J.; Li, R.; He, C.; Chen, H.; Cheng, R.; Zhai, C.; Zhang, M. An inverse design method for supercritical airfoil based on conditional generative models. Chin. J. Aeronaut. 2022, 35, 62–74. [Google Scholar] [CrossRef]
  26. Achour, G.; Sung, W.J.; Pinon-Fischer, O.J.; Mavris, D.N. Development of a Conditional Generative Adversarial Network for Airfoil Shape Optimization. In Proceedings of the AIAA Scitech 2020 Forum, Orlando, FL, USA, 6–10 January 2020; p. 2261. [Google Scholar]
  27. Press, W.H.; Teukolsky, S.A. Savitzky-Golay Smoothing Filters. Comput. Phys. 1990, 4, 669–672. [Google Scholar] [CrossRef]
  28. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  29. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of Wasserstein GANs. arXiv 2017, arXiv:1704.00028. [Google Scholar]
  30. Yonekura, K.; Aoki, R.; Suzuki, K. Quantification and reduction of uncertainty in aerodynamic performance of GAN-generated airfoil shapes using MC dropouts. Theor. Appl. Mech. Lett. 2024, 100504. [Google Scholar] [CrossRef]
  31. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  32. Nash, C.; Williams, C.K.I. The shape variational autoencoder: A deep generative model of part-segmented 3D objects. Comput. Graph. Forum 2017, 36, 1–12. [Google Scholar] [CrossRef]
  33. Guan, Y.; Jahan, T.; van Kaick, O. Generalized Autoencoder for Volumetric Shape Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  34. Larsen, A.B.L.; Sønderby, S.K.; Larochelle, H.; Winther, O. Autoencoding beyond pixels using a learned similarity metric. In Proceedings of the 33rd International Conference on Machine Learning; New York, New York, USA, 20–22 June 2016, Balcan, M.F., Weinberger, K.Q., Eds.; Proceedings of Machine Learning Research: Cambridge, MA, USA; Volume 48, pp. 1558–1566.
  35. Gui, J.; Sun, Z.; Wen, Y.; Tao, D.; Ye, J. A Review on Generative Adversarial Networks: Algorithms, Theory, and Applications. IEEE Trans. Knowl. Data Eng. 2021, 35, 3313–3332. [Google Scholar] [CrossRef]
  36. Arjovsky, M.; Bottou, L. Towards Principled Methods for Training Generative Adversarial Networks. arXiv 2017, arXiv:1701.04862. [Google Scholar]
  37. Goodfellow, I. NIPS 2016 Tutorial: Generative Adversarial Networks. arXiv 2017, arXiv:1701.00160. [Google Scholar]
  38. Xu, B.; Wang, N.; Chen, T.; Li, M. Empirical Evaluation of Rectified Activations in Convolutional Network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  39. Kwon, H. Adversarial image perturbations with distortions weighted by color on deep neural networks. Multimed. Tools Appl. 2023, 82, 13779–13795. [Google Scholar] [CrossRef]
  40. Abbot, I.H.; von Doenhoff, A.E.; Stivers, L., Jr. Summary of Airfoil Data; Langley Memorial Aeronautical Laboratory: Langley, Hampton, VA, USA, 1945. [Google Scholar]
  41. Van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Conditional GAN.
Figure 1. Conditional GAN.
Ai 05 00102 g001
Figure 2. Conditional VAE.
Figure 2. Conditional VAE.
Ai 05 00102 g002
Figure 3. Conditional VAEGAN.
Figure 3. Conditional VAEGAN.
Ai 05 00102 g003
Figure 4. Network architectures of the encoder, decoder, and discriminator.
Figure 4. Network architectures of the encoder, decoder, and discriminator.
Ai 05 00102 g004
Figure 5. Shape discretization.
Figure 5. Shape discretization.
Ai 05 00102 g005
Figure 6. Histogram of C L .
Figure 6. Histogram of C L .
Ai 05 00102 g006
Figure 7. Learning curve.
Figure 7. Learning curve.
Ai 05 00102 g007
Figure 8. Generated shapes. Numbers on top of each shape represent re-calculated C L . Red figure implies C L calculation did not converge.
Figure 8. Generated shapes. Numbers on top of each shape represent re-calculated C L . Red figure implies C L calculation did not converge.
Ai 05 00102 g008
Figure 9. Generated shapes C L = 0.5 . Different color represents different shapes.
Figure 9. Generated shapes C L = 0.5 . Different color represents different shapes.
Ai 05 00102 g009
Figure 10. Latent distribution.
Figure 10. Latent distribution.
Ai 05 00102 g010
Table 1. Success rates and errors of generated shapes.
Table 1. Success rates and errors of generated shapes.
ϕ mean MSE ↓ μ
cGAN ( d = 3 ) 4.91 π 0.0470.152
cWGAN-gp ( d = 3 ) 3.46 π 0.0470.320
N -CVAE [1]3.95 π 0.0270.226
VAE-WGAN-gp ( d = 4 ) 3.50 π 0.0280.243
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yonekura, K.; Tomori, Y.; Suzuki, K. Airfoil Shape Generation and Feature Extraction Using the Conditional VAE-WGAN-gp. AI 2024, 5, 2092-2103. https://doi.org/10.3390/ai5040102

AMA Style

Yonekura K, Tomori Y, Suzuki K. Airfoil Shape Generation and Feature Extraction Using the Conditional VAE-WGAN-gp. AI. 2024; 5(4):2092-2103. https://doi.org/10.3390/ai5040102

Chicago/Turabian Style

Yonekura, Kazuo, Yuki Tomori, and Katsuyuki Suzuki. 2024. "Airfoil Shape Generation and Feature Extraction Using the Conditional VAE-WGAN-gp" AI 5, no. 4: 2092-2103. https://doi.org/10.3390/ai5040102

APA Style

Yonekura, K., Tomori, Y., & Suzuki, K. (2024). Airfoil Shape Generation and Feature Extraction Using the Conditional VAE-WGAN-gp. AI, 5(4), 2092-2103. https://doi.org/10.3390/ai5040102

Article Metrics

Back to TopTop