1. Introduction
Carbon-fiber-reinforced polymer (CFRP) has high strength-to-weight ratio, high stiffness-to-weight ratio, light-weight, corrosion resistance, fatigue resistance, designability and other properties, and is widely used in aerospace, energy buildings, sports equipment, land transportation, marine fields, etc. In the context of the depletion of mineral energy, the global natural gas trade is active, and the market is increasing. Accordingly, composite wound gas cylinders commonly used for hydrogen storage or energy storage are also being used more widely and have a better prospect [
1,
2].
The appearance of commonly used composite gas cylinders is shown in
Figure 1. However, improper storage, prepreg production and preparation of CFRP during material processing may lead to fiber fracture, matrix debonding, fiber migration and other damages. During component manufacturing, fiber delamination, kinking, blistering, pollution, holes and other damages are caused due to improper laying, curing, machining or assembly. In service engineering, fiber delamination, fracture, erosion, matrix cracking and other damages may be caused by collision, chemical corrosion, local overload, fatigue and harsh environments. Due to the accumulation of impact, load, service time, environment and other factors, fatigue damage is the most dangerous and the main form of mechanical structure failure in the service process, accounting for 50–90% of the total mechanical structure failure [
3]. At the initial stage, there will be slight fatigue damage in the structure, which is difficult to detect. The internal fatigue damage gradually expands, and when it expands to a certain extent, it will break, and when it is serious, it will cause huge economic losses and endanger personal safety [
4]. Because of this, composite gas cylinders also have high safety risks, and now many accidents caused by the health problems of gas cylinders have aroused everyone’s attention. Therefore, it is of great significance to monitor and evaluate the health status of CFRP in real time, and efficient detection and monitoring technology has become the focus of current research. Currently, many people are studying experiments on composite plates and cylinders. Generally, basic research can be carried out on plates before being extended to gas cylinders.
At present, major research institutions have developed many structural health monitoring (SHM) methods. Common SHM methods include the acoustic emission method, electromechanical impedance method, vibration method, strain method, Lamb wave method, etc. Among them, the acoustic emission method depends on the acoustic emission signal when the damage occurs and expands, which is suitable for online use; The electromechanical impedance method is low in cost and simple in structure, but it is difficult to detect minor damage with low accuracy and small monitoring range [
5,
6]. Although the vibration method can monitor large-area damage, it requires more sensors, is insensitive to small damage, and is easily affected by the external environment; The strain method is only applicable to the monitoring of large-scale damage to structures. Compared with other SHM methods, because Lamb wave propagates in the measured structure with small signal attenuation, long propagation distance, sensitivity to minor damage, large area monitoring, high accuracy, low cost and other advantages, the structural health monitoring method based on Lamb wave is widely used in the industrial field and favored by researchers at home and abroad [
4].
Piezoelectric (PZT) is the main sensing element used in Lamb wave experiments at present and is widely used in signal generation and the acquisition of Lamb waves. A piezoelectric ceramic chip is a dual-mode transducer made of zirconium, titanium and lead oxides after chemical reaction. The structure includes a piezoelectric ceramic sheet and metal vibration sheet, which have obvious piezoelectric characteristics and can be used for ultrasonic excitation and reception.
Our current research mainly focuses on the corresponding signal acquisition and subsequent identification processing for several types of common CFRP damage, such as holes, notches, fatigue cracks, etc. This study is based on the health monitoring principle of Lamb wave. A Lamb wave in composite laminate is an elastic wave generated and propagated between the upper and lower surface boundaries of the plate by particle movement, as shown in
Figure 2. When the wave packet encounters a defect, reflection and scattering occur. Then, by analyzing the difference in waveform signals between healthy samples and damaged samples, the defects in the medium can be easily detected. Compared with broadband excitation signal, narrowband excitation signal is more suitable for structural health monitoring technology [
7]. The active Lamb wave is excited by the driving element to the monitored structure. Lamb wave has been widely used in the detection of materials and plates due to its flexible excitation and detection methods, its interaction with plate defects and carrying a large amount of information required for detection. Additionally, it has become an important and effective means and method for evaluating the performance of plate-like structures. In 2011, Mirahmadi et al. [
8] effectively improved the time resolution and signal-to-noise ratio of Lamb wave S0 mode detection by using signal processing methods such as Wiener filtering. In 2021, Xiang Yanxun from East China University of Science and Technology proposed a crack localization method based on group velocity matching, which can well locate breathing cracks on the guided wave propagation path [
9].
Recently, intelligent manufacturing has been thriving, and various types of production data can be collected faster and more extensively than ever before. This provides new opportunities for data-driven fault diagnosis methods to fully utilize massive data of various types [
10], which has attracted increasing attention from researchers and engineers. In the rapidly developing world of technology, artificial neural networks are one of the most mature data-driven fault diagnosis methods. With the rapid development of machine learning, deep learning (DL) has become an effective method for structural health monitoring. DL can automatically learn abstract representations of raw data [
11], which can better extract features and improve analysis effectiveness. In certain scenarios, data types for fault diagnosis can be presented in two-dimensional formats (such as time-frequency spectra), and then combined with image processing methods for classification. Through the processing of these images, analysis of various time-series signals can be facilitated, providing better ideas and directions for structural health monitoring today [
12]. Ince et al. [
13] applied one-dimensional CNN to real-time motor fault diagnosis in their research. Abdeljaber et al. [
14] conducted experiments on real-time damage detection using one-dimensional CNN. Chong [
15] proposed a method for extracting induction motor fault features, in which he converted one-dimensional vibration signals into two-dimensional grayscale images. Liu et al. [
16] applied a CNN model to wave-based crack damage detection, training a deep CNN model based on Lamb waves data to avoid the calculation and extraction of damage features. Wu et al. [
17] used continuous wavelet transform to convert Lamb wave time series signals into two-dimensional images, and then used a CNN model to classify the images, achieving accurate localization of layered damage in composite structures. Through these methods, researchers can simplify the process of waveguide signal processing, and to a certain extent, simplify the analysis of waveguide’s multimode and dispersion problems, while also improving the ability to extract signal features. At the same time, these methods can be combined with image processing methods for signal classification, improving the classification effect of temporal signals. However, in practice, damaged and defective data is relatively rare and costly. There are also many challenges in the engineering application of this technology, including the lack of training data, the imbalance between damaged and undamaged data, and the impact on the construction and subsequent processing of the overall dataset. Therefore, in this study, we hope to find a method to expand the acquired damaged signal samples without affecting accuracy, and improve this situation through data augmentation, combined with innovative classifiers to optimize signal classification. In the signal dimension transformation method, commonly used methods have many limitations and complex parameters, and the effect of feature extraction for Lamb waves is generally not ideal. Therefore, this study attempts a new signal-to-image conversion method, which can extract Lamb wave features more specifically, simplify the processing process, and improve the effect of signal dimension transformation. In addition, this study uses attention mechanisms to capture the time and feature dependencies of time series, improving the quality of data generation and subsequent signal classification. This overall framework for deep learning and signal processing is the focus of this study
The rest of this article is organized as follows.
Section 2 is about the design of the experimental scheme and the introduction of relevant experimental equipment and materials.
Section 3 includes the selection and design of data enhancement algorithm and signal classification algorithm.
Section 4 presents the experimental test results of the proposed method and the comparison of common methods. Finally,
Section 5 concludes this article.
3. Methodology
In order to realize the identification and classification of different damages and defects, neural networks are considered for processing. Processing direction mainly faces the application of data-driven modeling within the concepts of sensor data fusion, feature extraction and pattern recognition for Lamb wave signals. Data-driven methods require a large volume of historic data to establish the fault modes of the systems without a priori known models or signal patterns. The ultrasonic Lamb wave signal is obtained during the experiment, and the neural network model is constructed and trained using the signal data obtained from the original sample. The neural network model automatically learns the mode and damage-sensitive characteristics of ultrasonic signals during training. It should be added that in practice, data in damaged cases are rare and very costly. There are also many challenges in the engineering application of this technology, including the lack of training data, the imbalance between damaged and undamaged data, and the impact on the construction and subsequent processing of the overall dataset. Therefore, it is necessary to find a method to expand the obtained damage signal samples without affecting the accuracy. The underlying neural networks are trained on a small subset of data and are then used to process any incoming new sample. Considering these premises and foundations, the target model is selected and parameterized. On the basis of parametric research, the structure and parameters of the model are optimized. The test signal is introduced into the optimized model to verify the damage detection capability and classify the defects and damages. The idea in this study is to design a composite framework for identifying and classifying complex damage signals based on data augmentation and signal classification. At the same time, a dimension transformation method is introduced to process time series signals, which converts one-dimensional time series signals into two-dimensional images, making it easier to combine data with the latest and most mature neural network systems. In the specific operation process, we continuously optimize and improve the model by adding attention mechanisms and designing other relevant parameters. The theoretical basis, method selection and design of this study are depicted in
Figure 6:
3.1. A Method of Signal Dimension Conversion
In the acquisition experiment of Lamb waves, the collected signals are timing signals, which need to be processed and analyzed later. In traditional fault diagnosis methods, the data pre-processing method is vital since most data-driven methods cannot handle the raw signals directly. One of the main functions of the data pre-processing method is to extract the features of the raw signals from a large volume of historic data. However, extracting the proper features is exhaustive work, and these features have great effects on the final results. In addition, current deep learning methods mainly deal with two-dimensional structural data. The depth learning methods used for image processing are relatively mature, and there are many effective methods. Therefore, if we can convert the experimental one-dimensional time series data into two-dimensional (image) data through specific methods, we can better adapt to the current popular and cutting-edge models and improve the training and classification effect. Now, the following methods are adopted to convert the time-domain raw signals to images.
As shown in
Figure 7, in this conversion method, the time-domain raw signals fulfill the pixels of the image by sequence. First, we need to obtain an
×-sized image. To fill the square image, we need to randomly intercept a segment signal with the length
from the raw signal. First, let
denote the value of the segment signal. While
denotes the pixel strength of the image, as shown in equation
[
18]. Later, we will analyze this formula. The value of the segmented signal at each position is defined by
, and the value corresponding to a specific position is determined by
and
. The image intensity at the corresponding position is obtained by performing the operation in the equation.
The function
round is the rounding function, and the whole pixel value has been normalized from 0 to 255. The number size refers to the pixel intensity of the grayscale image. The 2 × 2 filters are commonly used in this study, and the size of image features on each layer would be reduced by half. Therefore, the recommended value of
in this case is
. In this study, the selection of 64 × 64 and 16 × 16 depends on the amount of signal data. In combination with the sparsity of guided wave signal features and the large amount of data collected each time, in order to ensure the features of the data, 64 × 64 images are chosen as the research object. The advantage of this data processing method is that it can convert one-dimensional signals into two-dimensional signals through this method, and it provides a way to explore the 2D features of the raw signal [
15]. Moreover, this data pre-processing method can be calculated without any predefined parameters and can reduce the experiences as more as possible. These can simplify the training process and optimize the calculation process. In this study, the size of the image varies with the volume of the signal data, and the guided wave signal data targeted in this experiment has a large volume. Moreover, a larger image size can facilitate classification results. Therefore, in this study, the size of the image is 64 × 64. In the subsequent processing, we also apply the zero-padding method to the model to prevent dimension loss. The process of quantifying temporal signals into two-dimensional images may introduce noise, which can cause the loss of image clarity and details, thereby affecting image performance to some extent. Therefore, this experiment used filters for the generated images, and in this experiment, we chose the common Gaussian filter to smooth the image and remove noise. Through subsequent testing, this experiment minimized the impact of noise on image quality and subsequent processing as much as possible.
3.2. Data Enhancement Based on Diffusion Model
This study mainly aims to build a data-driven SHM system, which, most importantly, needs data support. Considering the limitations of the experimental conditions and the specific conditions of the experimental object, some experimental signals have poor detection effects, and some experiments have insufficient signal detection to build a sufficient number of data sets for training. It may have certain impact on the training of the experimental model, and will also have a certain interference with the analysis of the experiment. At the same time, the experimental data has uneven samples, so we consider expanding and enhancing some data to increase the number of samples. Data enhancement is also called data amplification, which means that limited data can generate the value equivalent to more data without substantially increasing data. These can expand the dataset, improve the robustness of the model, improve the generalization ability of the model and avoid sample imbalance. Now, we mainly consider testing several generation models in this study, including Generic Adversary Network (GAN), Variable Automatic Encoder (VAE) and diffusion models (DM).
3.2.1. GAN
GAN consists of a generator and discriminator. The generator is responsible for generating realistic data to “cheat” the discriminator, and the discriminator is responsible for judging whether a sample is real or “made”. The training of GAN involves the two models learning from each other and obtaining new data through confrontation. The generating network should constantly optimize the data generated by itself so that the discriminant network cannot judge, and the discriminant network should also optimize itself to make its judgment more accurate. The relationship between the two forms confrontation, so it is called a confrontation network [
19]. When selecting GAN, after comparison and analysis, the more advanced and practical DCGAN network was selected. DCGAN is a variant of GAN. DCGAN combines CNN with the original GAN, and both the generating network and the recognition network are applied to the deep convolution neural network. DCGAN improves the stability of basic GAN and the quality of generated results, and improves the ability and effect of feature extraction as a whole.
3.2.2. VAE
VAE wants to train a generation model. First, it assumes the distribution of a hidden variable z, and constructs a model from z to target data x, which can map the sampling probability distribution to the training set probability distribution. The hidden variable z is generated, which contains data information and noise, so that the learned target data is close to the probability distribution of the real data. VAE generally matches a Gaussian distribution for each sample, and the hidden variable Z is obtained by sampling from the Gaussian distribution. In addition to restoring input sample data, it can also be used to generate new data. VAE is also a mature model for data generation and expansion.
3.2.3. Diffusion Model
A diffusion model is a kind of generation model, different from Variable Automatic Encoder (VAE), Generic Adversary Network (GAN) and other generation networks. A diffusion model gradually applies noise to the image in the forward phase until the image is damaged and becomes a complete Gaussian noise, and then learns the process of restoring from Gaussian noise to the original image in the reverse phase. The purpose of a diffusion model is to decompose specific data distribution and reconstruct data step by step from random noise. Because the intermediate process is stochastic and unstable, the whole process is generated. In addition, reversible random steps ensure extensive fidelity of the model and sample quality.
Specifically, in the forward phase, the noise is gradually increased on the original image, and the new image obtained in each step is only related to the results in the previous step until the image becomes pure Gaussian noise. The diffusion process is a Markov process. The reverse phase is a process of continuous noise removal. This is to give Gaussian noise to the original image, and then gradually remove the noise until the original image is finally restored [
20]. The schematic diagram of the forward and reverse processes of the diffusion model is shown in
Figure 8. More detailed processes will be described later:
These are the following points about the forward process. We have a sequence of
variables, where
~
is our observed data and
are latent variables. we can think of diffusion models as a specific realization of a hierarchical VAE. What sets them apart is a unique inference model, which contains no learnable parameters and is constructed so that the final latent distribution
converges to a standard gaussian [
21]. This “forward process” model is defined as follows:
The variables
define a fixed variance schedule, chosen such that
The forward process transforms sample distribution into gaussian noise. The image is transformed according to Gaussian distribution after adding noise. A nice property of the forward process is that we can directly sample from any timestep. Next, parameter transformation and iteration are carried out to obtain a new expression of the formula:
One consequence of this is that we can draw random samples
~
as part of the training procedure. Another consequence is that we can make additional manipulations to the lower bound to reduce the variance. This proceeds by observing the following result (from Bayes rule):
Due to the Markov property of the forward process, we have:
. Therefore, by conditioning all terms on
we arrive at the following expression.
This is the main framework of the forward process. Now, let us look at the reverse process.
The real noise generated by each forward propagation part is recorded as a label. In addition to inference, the forward diffusion process also includes the “data set construction process” similar to that used in this mathematical model. When the model is in reverse diffusion, the Gaussian noise generated in the forward diffusion can be predicted and inferred step by step to restore the initial sample data. The corresponding reverse process distributions are defined as:
The variance
is a time-dependent constant (
). The
is a neural network with an input
. In order to share parameters between timesteps, the conditioning variable
is introduced in the form of positional embeddings [
22]. The forward process samples
can be written as a function of
and some noise
.
We assume
is a Gaussian distribution, and we can use a neural network to fit it. Combine the generated noise tags as training reference. The inverse process is also a Markov chain process.
The corresponding process and results can be obtained as =. Set variance to constant and continue to substitute the Gaussian probability density function for subsequent processing.
Finally, it can be summarized as such:
Then, calculate the likelihood function of the target data distribution, so that the network can be trained. The mean fitted is (), and this can be used for sampling.
The subsequent sampling process samples
from the standard normal distribution and calculates
iteratively. Known mean
and constant variance
,
to
can be calculated by parameter renormalization. This training objective also be viewed as a weighted combination of denoising score matching used for training score-based generative models [
23].
In fact, the idea of the Denoising Diffusion Probabilistic Model (DDPM) is very simple. In reverse diffusion,
is used to predict
images. Instead of optimizing the
to
conversion, it is to optimize how noise is added when
to
. At the same time, DDPM also adds time embedding [
24]. In this model, we will use U-Net. U-Net is a U-shaped network structure, which is widely used in the segmentation field. The structure of this network is symmetrical, filled with convolutional layers but without fully connected layers. This method combines high-resolution feature maps with low-resolution feature maps after upsampling, which makes the segmentation results more accurate. At the same time, in the upsampling stage, the model still has a large number of channels, and more detailed object information will be transmitted back through upsampling. From the shape, the upsampling stage and the downsampling stage are symmetrical. The function of time embedding is to tell U-Net (used to generate random noise and other work) the position of reverse diffusion. Another reason for adding time embedding is that U-Net shares parameters. Adding time embedding can generate different outputs based on different inputs. Through these methods, U-Net can form some rough contours in the process of reverse diffusion. As it progresses close to the original image, we hope to obtain object edge information and some high-frequency information, so as to make the output image more realistic [
25]. We utilize an attention mechanism to capture the temporal and feature dependencies of time series. To capture temporal and feature dependencies of multivariate time series, we utilize a two-dimensional attention mechanism in each residual layer instead of a convolution architecture [
26]. The attention mechanism allows the model to handle various lengths. For batch training, we apply zero padding to each sequence so that the lengths of the sequences are the same. This is also the key model used in this study.
3.3. Signal Recognition and Classification Based on DenseNet
In this section, we demonstrate the recognition and classification ability of the above comprehensive methods for Lamb wave signals of different defects in composite materials. Now we consider using neural networks for signal classification. When the network selected for signal classification is very deep, the effect of the model becomes worse (the higher the error rate), indicating that the deeper the network is, the better. When the network level increases to a certain number, the training accuracy and test accuracy are rapidly reduced, which shows that when the network becomes very deep, the deep network becomes more difficult to train. In the process of back propagation, the neural network needs to continuously propagate the gradient. When the number of network layers deepens, the gradient will gradually disappear in the process of propagation, which makes it impossible to effectively adjust the weight of the previous network layer. Therefore, deep residual network (ResNet) [
27] is considered. The residual network is easier to be optimized by adding shortcut connections to “skip connections”. Several layers of networks containing a shortcut connection are called a residual block, as shown in the following
Figure 9:
As shown in the figure above, x represents the input, represents the output of the residual block before the activation function of the second layer, and σ Represents the ReLU activation function. The output of the final residual block is.
Directly transfer the input x to the output as the initial result, and the output result is
. When
, then
, that is, the identity mapping mentioned above. Thus, ResNet is equivalent to changing the learning goal. It is no longer a complete output of learning, but the difference between the target value
and
, that is, the so-called residual
Stochastic depth shortens [
28] ResNet by randomly dropping layers during training to allow better information and gradient flow.
The basic idea of the densely connected constructive network (DenseNet) model is the same as ResNet, but it establishes a dense connection between all the preceding layers and the following layers, and its name comes from this. One advantage of ResNet is that gradients can flow directly from the following layer to the front layer through the identity function. However, the output of the identity function and H is cumulative, which may hinder the information flow in the network. To improve the problem of information flow between different layers, DenseNet directly connects all inputs to the output layer. Specifically, each layer will accept all the layers in front of it as its additional input. The connection method is through element-level addition. To preserve the feed-forward nature, each layer obtains additional inputs from all preceding layers and passes on its own feature-maps to all subsequent layers. In DenseNet, each layer is connected to the channel dimensions of all previous layers and is used as the input of the next layer. Another major feature of DenseNet is feature reuse through feature connections on the channel. These features enable DenseNet to achieve better performance than ResNet with fewer parameters and computing costs [
29]. Besides better parameter efficiency, one big advantage of DenseNets is their improved flow of information and gradients throughout the network, which makes them easy to train. Each layer has direct access to the gradients from the loss function and the original input signal, leading to an implicit deep supervision.
The performance differences between the main transmission forms of these two types of networks are shown below. The nonlinear transformation equation of ResNet is as follows:
The nonlinear transformation equation of DenseNet is as follows:
The dense connection mode of this model has many advantages. It increases the transmission of gradient, reuses features, and even reduces overfitting on small sample data. At the same time, DenseNet has fewer parameters than ResNet, less computation and higher accuracy. The Dense block is a basic module of DenseNet, and DenseNet is divided into several Dense blocks. In each node of the Dense block, the input is concatenated feature maps, and the size of the feature maps within each Dense block is the same. A Transition module is used to perform down-sampling transition connections between each Dense block. The nonlinear combination function in the Dense block refers to the combination of BN + ReLU + Conv. The simple model structure of the Dense block is shown in
Figure 10.
Figure 11 shows a complete DenseNet structure, which includes 3 Dense Blocks and 2 Transition layers. The Transition layers connect each Dense Block and consist of convolution and pooling to downsample and compress the model. After completing these steps, the data will enter the prediction or classification module for processing. Each layer of DenseNet is designed to be very narrow to reduce redundancy. Instead of drawing representational power from extremely deep or wide architectures, DenseNet exploits the potential of the network through feature reuse, yielding condensed models that are easy to train and highly parameter-efficient. Concatenating feature-maps learned by different layers increases variation in the input of subsequent layers and improves efficiency. The network improves the flow of information and gradient, making it easy to train, and the intensive connection has regularization benefits, reducing the overfitting problem of a small training set. Additionally, after verification, the model has a good perception effect on the defect signal in this experiment and a strong feature extraction ability.
4. Experimental Validation
In this section, we demonstrate the recognition and classification ability of the above comprehensive methods for Lamb wave signals of different defects in composite materials. We classify different experimental test situations, and group the existing data into corresponding training sets to train the best model.
Through the introduction and discussion in the previous chapter, it can be seen that DenseNet increases the transmission of gradients, reduces overfitting, and has low computational complexity and high accuracy. Therefore, DenseNet is used as the main model in the classification stage of the entire deep learning framework in this study. Moreover, this classifier has good perception effect on defect signals in this experiment and has strong feature extraction ability after verification. After research and optimization, the DenseNet model structure used in this experiment is shown in
Table 2.
In this study, a large number of simulation data about the defect are used to test the neural network model, and the recognition ability and feature extraction ability of different neural network models for this type of defect signal are compared and analyzed, which is convenient for model selection. After converting one-dimensional data into two-dimensional images and putting them into corresponding neural networks for training, the network designed in this experiment is compared with common classification network models. This study combines the algorithms used for data expansion and classification recognition. In the data set and experimental setup, we used five combinations to test. The first is this proposed framework, which combines data augmentation and classification algorithms (including the improved DDPM). To validate the effectiveness of this method, we compared four different algorithm combinations: “VAE + DenseNet”, “DCGAN + DenseNet”, “DDPM + ResNet”, and “DDPM + DenseNet” (including the original DDPM).
In order to test the quality and effect of data generated by different models, several common comparison indicators are selected in combination with common methods. First, the accuracy rate, that is, the ratio of the number of correctly predicted results to the total number of samples. The higher the accuracy, the better the quality of the generated data. Then there is the recall rate, which is aimed at the original sample. It means to correctly predict the probability of positive samples in all positive samples, that is, to correctly predict the ratio of the number of positive samples to all positive samples. The higher the recall rate, the more accurate results will be retrieved. Finally, Fréchet Inception Distance (FID) index is selected. FID represents the diversity and quality of generated objects. The smaller the FID, the better the diversity and quality. In the later tests, in order to highlight the overall trend and distribution characteristics, we use bold to mark the best results in different test schemes.
Next, different data grouping forms are tested to see which grouping situation has the best effect on data feature learning and recognition, and the data quality is the highest. We will adjust the number of groups and proportion relationship of three types of data, namely include intact data, hole defect data and crack defect data. The best proportion of data sets with the best learning ability for these three situations will be found by testing in different situations. The design and allocation of the experiment are shown in
Table 3, the best values in each test result are marked in bold:
Through the table, we can observe that the fourth group should be selected for experiment and training. In this grouping case, the lowest FID value and the highest accuracy rate can be obtained, and the higher recall value can also be maintained, indicating that the learning and training effect of input data is the best in this case.
After determining the classification method of data groups, use the classification method with the best learning effect mentioned above, and use the data collected from the experiment to make a training set. First, the generation model is used to expand the unbalanced data of some samples, and then the expanded data is mixed with the original data to make a new dataset. Then put, the new data set obtained after learning and training into the previously selected classification network for testing, and verify the overall experimental design through the accuracy and effect of classification. Relevant details of the training process are supplemented as follows. For the dataset, we randomly divided the dataset into five parts and used one of them as test data for each run. We also randomly split the remaining data into train and validation data with a ratio of 6:1. For the training of all tasks, we normalize each feature to have zero mean and unit variance. We set the batch size as 16 and the number of epochs as 200. We used Adam optimizer with a learning rate of 0.001. The core values in the diffusion model are also given here. We set the number of the diffusion step T = 100, the minimum noise level β1 = 0.0001, and the maximum noise level βT = 0.5. For the several types of data enhancement and data classification methods used in this study, grouping and collocation tests are conducted, respectively. This is to test which data enhancement and data classification methods work best.
After determining the testing plan and training details, extensive training and testing were conducted for the five methods proposed earlier in this experiment. The framework combining data augmentation and classification algorithms (including the improved DDPM) proposed in this paper, as well as four different algorithm combinations: “VAE + DenseNet”, “DCGAN + DenseNet”, “DDPM + ResNet”, and “DDPM + DenseNet” (including the original DDPM), were tested as a whole. In addition to the accuracy index, the FID index was also retained as a reference in the final test results. After multiple rounds of training, the results are shown in
Table 4 and the best values in each test result are marked in bold:
It can be seen that the improved DDPM model is the best for learning and generating such defect signals. The accuracy obtained by putting the generated data back into the classification model is also the highest, and the performance is relatively stable. Hence, suggesting that this type of architecture is well suited to detect features and patterns in the raw time-domain signals that can unequivocally determine the presence of defect reflections. For the signal data of this experiment, DDPM can extract features well, and generate new signals that are close to the original signal properties, which plays a great role in defect signal processing and recognition. At the same time, we evaluated the effect of DenseNet in the classification task and compared it with ResNet architecture. The feature extraction ability and classification effect of Densenet in this study are also stronger than Resnet, and these two methods can be better complementary and improved. To sum up, the signal recognition neural network framework proposed in this experiment combined with improved DDPM and DenseNet shows good comprehensive effect in the above tests.
The experiment and test results show that the method designed in this study based on DenseNet classification and DDPM data enhancement can accurately detect and classify the damage signals of common defects in CFRP composite plates, and can achieve high accuracy. It is practical for the nondestructive testing of composite materials.
5. Conclusions
In this study, a method for identifying and classifying composite damage signals based on DenseNet classification and DDPM data enhancement is proposed. At the same time, a dimension transformation method is introduced to process defect signals. This method converts one-dimensional timing signals into two-dimensional images, which is convenient for combining the data with the latest and most mature neural network system. This method, by adding attention mechanism and other related parameter designs, has been successfully improved. After determining the test plan and training details, the five application methods proposed at the end of this experiment were fully trained and tested, and the advantages and disadvantages of the proposed method and the other combination methods were compared. The experiment and test results show that the method designed in this study based on DenseNet classification and DDPM data enhancement can accurately detect and classify the damage signals of common defects in CFRP composite plates, and can achieve high accuracy. In the final comprehensive test, the test results of the method proposed in this study for Lamb wave signal-datasets of different defects reached 88.10%. Compared with other common data expansion and signal classification methods, this scheme also has advantages, and is practical for the nondestructive testing of composite materials. The method proposed in this experiment mainly focuses on the study of micro-damage in composite structures, and it may not be able to detect defects smaller than 1 mm. The application object of this method is mainly carbon fiber composites, but it can also be applied to other composite materials, such as glass fiber, and has been tested, whereas other materials have not been tested yet. In the future, we may try to extend it to other materials and other types of damage, and further improve the accuracy. We believe that this scheme will continue to improve accuracy in the future and will be extended to more types and components of composite products; it may play a role in composite nondestructive testing and structural health monitoring.