Next Article in Journal
The XMUSPEECH System for Accented English Automatic Speech Recognition
Next Article in Special Issue
Anisotropy Corrected FMC/TFM Based Phased Array Ultrasonic Imaging in an Austenitic Buttering Layer
Previous Article in Journal
Cardioprotective Mechanisms of Interrupted Anesthetic Preconditioning with Sevoflurane in the Setting of Ischemia/Reperfusion Injury in Rats
Previous Article in Special Issue
Deep Learning Based Inversion of Locally Anisotropic Weld Properties from Ultrasonic Array Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

System Invariant Method for Ultrasonic Flaw Classification in Weldments Using Residual Neural Network

1
Department of Mechanical Engineering, College of Engineering, Sungkyunkwan University, Suwon 16419, Korea
2
Korea Institute of Nuclear Safety, Daejeon 34142, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(3), 1477; https://doi.org/10.3390/app12031477
Submission received: 8 December 2021 / Revised: 10 January 2022 / Accepted: 12 January 2022 / Published: 29 January 2022

Abstract

:
The industrial use of ultrasonic flaw classification using neural networks in weldments must overcome many challenges. A major constraint is the use of numerous systems, including a combination of transducers and equipment. This causes high complexity in the datasets used in the training of neural networks, which decreases performance. In this study, the performance of a neural network was enhanced using signal processing on an ultrasonic weldment flaw dataset to achieve system invariance. The dataset contained 5839 ultrasonic flaw signals collected by various types of transducers connected to KrautKramer USN60. Every signal in the dataset was from 45 FlawTech/Sonaspection weldment specimens with five types of flaw: crack, lack of fusion, slag inclusion, porosity, and incomplete penetration. The neural network used in this study is a residual neural network with 19 layers. The performance evaluation of the same network structure showed that the original database can achieve 62.17% ± 4.13% accuracy, and that the invariant database using the system invariant method can achieve 91.45% ± 1.77% accuracy. The results demonstrate that using a system invariant method for ultrasonic flaw classification in weldments can improve the performance of a neural network with a highly complex dataset.

1. Introduction

Ultrasonic testing is a non-destructive testing technique that is commonly used in many industries. Because most defects occur in weldments, ultrasonic inspections are usually focused on weldments, and thus detecting and classifying ultrasonic flaw signals is a long-standing task in the field of ultrasonic testing. Determining flaws requires numerous experiences, and it is difficult to perform non-destructive testing efficiently because there is the possibility of human misjudgment. To solve this problem, many studies on ultrasonic testing to determine the presence, type, shape, and size of flaws using artificial intelligence are being conducted.
Song et al. [1] used probabilistic neural networks on ultrasonic weldment defect signals. Masnata et al. [2] used a neural network and Fisher discriminant analysis to classify cracks and volumetric defects. Polikar et al. [3] applied a neural network to develop a frequency-invariant weldment defect classification system for pipes. Margrave et al. [4] used neural networks to detect flaws in steel pipes. Song et al. [5] developed an intelligent ultrasonic evaluation system (IUES) for real-time weldment flaw classification. Simas Filho et al. [6] developed a flexible decision support system based upon neural network and principal component analysis to evaluate fiber-metal laminates. Cruz et al. [7] also developed a decision support system that selects parameters for the neural network to detect flaws in steel-welded joints. Meng et al. [8]. used a combination of wavelet transform and a convolutional neural network to classify void and delamination defects in CERP specimens. With the exception of [4], all these methods are limited due to their complexity [9,10,11].
A large amount of data must be collected to ensure the performance of artificial intelligence (AI). It is highly difficult to collect a large amount of real inspection data in the ultrasonic testing research field. Many studies collect data from specimens to solve this problem; however, there is a limit on the amount of data that one research group can collect. This works well for systems with little variation, such as those that use an ultrasonic transducer from the same manufacturer. However, inspectors use various types of transducer and equipment for industrial ultrasonic testing. The limitation of dataset size can be a problem in this scenario for the following reasons. First, the AI may not sufficiently train signals from each system. Second, flaw signals from various systems may interfere with the training process, finding global minimum. To solve this problem, this study proposes a system invariant technique to reduce the complexity of ultrasonic weldment flaw databases and compares the performance of two different databases, one without a system invariant method and the other with a system invariant method, using the same neural network architecture.

2. Ultrasonic Weldments Flaw Databases

To properly train neural networks, a database of well-inspected ultrasonic weldments flaw signals was developed. The signals were collected from 45 FlawTech/Sonaspection weldment specimens, including five types of flaw: crack, lack of fusion, slag inclusion, porosity, and incomplete penetration. The signals were collected with 21 different ultrasonic transducers to make the system complex. This group of transducers included 2, 4, and 5 MHz central frequencies, 45, 60, and 70 degree inspection angles, and four different manufacturers (KrautKramer, GE, Olympus, and TKS). All the inspections were performed using KrautKramer and USN-60 flaw detector as the ultrasonic pulser-receiver instrument and a 50 MHz sampling frequency. The list of transducers is presented in Table 1.
To construct the database, 5839 signals were acquired for each flaw and, subsequently, 2048 points were extracted with the flaw signal (Figure 1). The number of each weldment flaw is presented in Table 2.
To check the performance of the neural networks for training, the database was separated. To balance the learning between each flaw, 40% of the crack signal and 10% of other flaw signals were randomly divided into the test signal. The compositions of the training and testing signals are described in Table 3.

3. Database Augmentation

Although 4386 signals were obtained, considering the complexity of the system, the amount of data was not sufficient. Therefore, the number of training data was increased through augmentation. In terms of data augmentation, noise addition and time shifting were effective in various studies [3,10].
Because ultrasonic testing is usually performed in harsh environments, the possibility of the testing results being corrupted by electric noise, presented as white Gaussian noise (WGN), exists. Additive WGN was used to initiate this. The mathematical details of WGN are described in [12,13]. The WGN is added with two different signal-to-noise ratios (SNR) levels i.e., SNR 20 and SNR 15 (Figure 2). The number of testing signals increased threefold from 4386 to 13,158, including the original signal.
Another augmentation technique was adjusted after developing the augmentation database with WGN. In terms of ultrasonic testing signals, time shifting is equivalent to changing the distance between the transducer and the flaw location [10]. Therefore, new signals were created that shifted 20 and 10 points backward, and 20 and 10 points forward (Figure 3). Because the system features a sampling frequency of 50 MHz, one point on the time axis is 0.02 μs. Because of time shifting, including the original signal, the testing signal number increased fivefold, from 13,158 to 65,790. This augmented dataset is referred to as the original database and is described in Table 4.

4. Artificial Neural Networks

Artificial neural networks (ANNs) are computational neural models inspired by a sequence of biological neurons. ANNs operate on the basis of the backpropagation algorithm proposed by Rumelhart et al. [14]. In the early stage of ANNs, a significantly simplified form of neural network, known as a multilayer perceptron (MLP), was developed, which led to fully connected neural networks (FCNNs) [15]. FCNNs are composed of input, hidden, and output layers. Each layer contains nodes with weights, and each node in the layer is connected to every node in the next layer. Deep neural networks are networks with more than two hidden layers.
The process of training a neural network involves adjusting the weight and connections between the layers. First, the data enter the input layer and pass through the hidden layers, while the dot product between the input data and weight matrix, whose size depends on the shape of the hidden layers, is computed. While computing, the activation function is used for tuning the training direction. After calculating the output, a loss function is computed, which gives the error between the true label and the predicted output. The error is then inverted using a backpropagation algorithm, and the weights are slightly updated to reduce this error. This process can be repeated until the error becomes sufficiently small or saturated [16,17,18,19].

4.1. Convolutional Neural Network

Convolutional neural networks (CNNs) were first proposed by Lecun et al. [20]. They later gained popularity in the computer vision community [21]. CNNs feature a convolutional layer that is connected to a specific spatial region in the previous layer and is based on the size of a defined filter, in contrast to FCNNs. This architecture allows the network to focus on lower-level features before assembling them into higher-level features. CNNs also include pooling layers, which help to reduce the computational load by downsampling the input data, but increase the complexity of the network [22].

4.2. Residual Neural Network

Residual neural networks (ResNets), first proposed by Kaiming He et al. [23], are based on CNN but include shortcuts between the input and output of the unit layer. If the input data are referred to as x, the output of ResNets is described by Equation (1):
F(x) + x
where F is the weight of computation. The calculational direction of ResNets minimizes the value of F(x) + x, which means F(x) becomes 0 since x is a fixed (input) value here. Conventional CNNs feature a weak spot in significantly deep networks, as their performance degrades after a certain depth. However, significantly deep networks are required to train with large amounts of data. ResNets solve this problem through their distinct network architecture as previously described, obtaining the best results by training 152 layers of ResNet using the CIFAR-10 dataset [24,25,26,27].

4.3. Residual Neural Network Architecture

The ResNet used in this study was designed in Keras, based on Tensorflow (Google open-source software for deep learning). The algorithm included an input layer, nineteen convolutional layers, and three stages of ResNet blocks. Each stage features the same architecture, but a different number of feature maps. The detailed architecture of the stage with feature map FM is described in Table 5. The final ResNet architecture used when these unit stages were combined is shown in Table 6.
The number of nodes in the input layer was maintained at the same level as the sampling point of the input signal, which was 2048. The first convolutional layer contained 64 kernels of size 1 × 32, which remained large because of their performance for ultrasonic and vibration signals. [10,11,28]. The pooling layer reduces the number of layers but does not affect performance [22]. After the data have passed through stage 3, there is one fully connected (dense) layer and an output layer. The number of nodes in the fully connected layer was chosen as the best performance over multiple trials. The number of nodes in the output layer was kept equal to the number of flaws for classification. The activation function used in the layers was rectified linear unit (Relu), described by Equation (2)
R e l u = x < 0   f x = 0 x 0   f x = x
because of its state-of-the-art performance in deep learning [29,30,31]. The loss function used is sparse categorical cross-entropy, described by Equation (3):
i = 1 M y i log p i
where M is the number of data, y is a vector of the true label, and p is the predicted label. Sparse categorical cross-entropy gives the result as integer form, and it is widely used for multiclass classification problems. The activation function applied to the output layer is softmax, described by Equation (4):
σ y i = e y i k = 1 K e y k
where y is the input vector to the output layer, and i indexes the output nodes, i = 1, 2, K. In this study, the number of classes is 5, so K is 5. Additionally, dropout normalization was used to avoid overfitting. The network is then trained to minimize cross-entropy loss.
ResNet is trained with optimizer adaptive moment estimation (Adam) for 500 epochs. Adam is an algorithm for the first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments. The hyper-parameters feature intuitive interpretations and typically require little tuning. The method is suitable for the classification of flaw signals in complex systems since it is appropriate for non-stationary objectives and problems with very noisy and/or sparse gradients [32].

4.4. ResNet Performance Evaluation

The performance of ResNet was then evaluated on the original database (Figure 4). The results show that highly complex datasets with a comparably low amount of data can be a significant penalty when training a neural network. In epoch 461, the training accuracy increased to 92.63%, whereas the testing accuracy reached 70.75% in epoch 241. Here, the meaning of testing accuracy is simply how much data were predicted to be the same as the true label. For a more detailed analysis of the testing accuracy, the average accuracy after epoch 100 was 62.17%, with a standard deviation of 4.13%. This means that training ResNet with the original database can result in a network performance of 62.17% ± 4.13%.
Prior research demonstrates that for simpler databases acquired with KrautKramer transducers, the testing accuracy exceeds 90% [10,11]. This low testing accuracy has two possible explanations. One is that the network is overfitted, and the other is that the complexity of the database is too high. However, since this network already includes many regularization techniques that are widely known to solve overfitting, it is difficult to regard it as a problem of overfitting alone. Or, at least, it is difficult to solve this problem with only the regularization technique. Therefore, reducing the complexity of the database down can be a possible solution for this problem.

5. System Invariant Method

Because the database was obtained from 21 different transducers, the neural network may be confused by other types of signals because of the unique characteristics of each transducer. There are several approaches to making ultrasonic signals invariant to system or frequency to improve the performance of the neural network training. Polikar et al. created a frequency invariant system to enhance the performance of a neural network for ultrasonic signals by normalizing the frequency [3]. Song et al. developed an intelligent system for ultrasonic flaw classification in weldments [5]. During the process, the system was normalized using a reference signal in the frequency domain.

5.1. Principle of System Invariant Method

Weldment flaws signal V f include the system, flaw scattering pattern, and specimen as variables. Among these, systems and specimens increase the complexity of the database, which can inhibit performance in classifying neural network flaws, so the database must be refined by extracting only the flaw scattering patterns. The schematic experiment setup is shown in Figure 5.
Because the ultrasound system is linear time-invariant (LTI), V f can be expressed in a time domain with convolution between system variables, as in Equation (5) [33]:
V f t = β t * C 1 t * T 12 t * C 2 t * A t * C 2 t * T 21 t * C 1 t
where β is system factor, C 1 is the diffraction correction in the wedge, T 12 is the transmission coefficient of the wedge to the specimen, C 2 is the diffraction correction in the specimen, A is the scattering pattern from the flaw, and T 21 is the transmission coefficient of the specimen to the wedge.
To eliminate system and specimen parameters, signals were acquired from a reference specimen using the same system used for the weldment signal acquisition. A reference signal was acquired using a circular reference block to obtain the same ultrasound path for every inspection angle for each transducer–system combination. Here, 21 reference signals by 21 different types of transducers were obtained using the same ultrasonic pulser-receiver device. The schematic experimental setup for the reference signal is shown in Figure 6.
Furthermore, the reference signal V R can be described in a time domain as a convolution between system factors, as in Equation (6):
V R t = β t * C 1 t * T 12 t * C 2 t * R t * C 2 t * T 21 t * C 1 t
where C 2 is diffraction correction in the reference specimen and R is the reflection coefficient from the round surface. In the frequency domain, the signals V f and V R can be combined by simple multiplication, not convolution [34,35]. Using fast Fourier transforms (FFT), V f and V R are described in the frequency domain as Equations (7) and (8):
V f ω = β ω C 1 ω T 12 ω C 2 ω A ω C 2 ω T 21 ω C 1 ω
V R ω = β ω C 1 ω T 12 ω C 2 ω R ω C 2 ω T 21 ω C 1 ω
Divide these two terms, flaw scattering pattern A can be extracted as Equation (9):
V f ω V R ω = C 2 ω C 2 ω 2 A ω R ω
The reflection R is affected only by the mechanical property of the two mediums and the incidence angle. Here, the round surface of the reference block fixes the incidence angle as normal when the transducer angle varies. Therefore, R i is constant, while i represents each transducer with a different angle. Furthermore, the reflection coefficient for steel to air is significantly close to −1 in this case. This means that Equation (9) is described as Equation (10):
V f ω V R ω C 2 ω C 2 ω 2 A ω
Diffraction correction C ω in far-field is represented as Equation (11):
C ω = i k a 2 2 z
where k is the wavenumber, a is the radius of the transducer and z is the flight distance. C 2 and C 2 feature the same frequency and radius as the transducer, except for distance. Thus, Equation (10) reduces to:
V f ω V R ω z z 2 A ω
where z is the distance between the transducer and the flaw of the weldment specimen, while z is the distance between the transducer and the round surface of the reference specimen. With constant a = z / z 2 , A ω can be represented as Equation (13):
a A ω V f ω V R ω
Applying inverse fast Fourier transform to A ω can restore the time domain signal of the flaw scattering pattern. When training a neural network, each signal becomes normalized for increasing performance; the constant a disappears. The flow diagram of this process is shown in Figure 7.

5.2. Applying System Invariant Method to Original Database

To reduce complexity, the system invariant method should be used to reconstruct the ultrasonic weldment signals in the original database. Examples of crack signals before and after using the system invariant method are shown in Figure 8. The system factor was removed from the original signals, leaving only the scattering pattern from flaws. The database that uses the system invariant method is referred to as the invariant database.

5.3. ResNet Performance Evaluation with Invariant Database

Because the original database and the invariant database featured the same number and composition, training the same architecture of ResNet with the original database and invariant database is suitable for verifying the effect of the system invariant method.
The performance of ResNet was then evaluated on the invariant database, as shown in Figure 9. The results show that the system invariant method can reduce database complexity and enhance network performance. The training accuracy increased to 92.15% in epoch 446, which is comparable to the original database, whereas the testing accuracy was higher than 95.04% in epoch 491. For a more detailed analysis of the testing accuracy, the average accuracy after epoch 100 was 91.45%, with a standard deviation is 1.77%. This means that training ResNet with the invariant database can obtain a network performance of 91.45% ± 1.77%.
A further reason why the invariant method can improve network performance is that the test curve behaves similarly to the training curve, with only a minor shift up and down. This implies that the system invariant method can confuse the network into believing that training and testing data with the same flaw type are similar signals with which to classify.

5.4. Performance Comparison

Table 7 and Figure 10 show a performance comparison. Comparing the training accuracies, the invariant database featured lower average accuracy, but a smaller standard deviation than the original database, making it difficult to determine which is better. It is more rational to consider this as an error that can occur naturally during the training process of a neural network, rather than a performance deviation.
Comparing the test accuracies, there was a significant improvement in accuracy for the invariant database. The average accuracy improved by 29.28 pp from 62.17% to 91.45%, making it possible to classify ultrasonic weldment flaw signals more precisely; furthermore, the standard deviation improved by 2.36 pp from 4.13% to 1.77%, making the network performance more stable. Furthermore, the average training and test accuracies of the invariant database were almost similar, suggesting that the system invariant technique can effectively secure the system invariant.

6. Summary

In this study, a system invariant method was developed and tested using a complex database obtained from 21 different transducers. The flaw signal database contained 65,790 training data and 1453 test data, referred to as the original database. By developing a system invariant method in the frequency domain and using it in the original database, an invariant database was constructed. Consequently, comparing performance using a residual neural network, the invariant database showed a significant improvement in accuracy compared to original database. The proposed system invariant method showed meaningful performance improvements against complex ultrasonic equipment. In future studies, practical research is needed in order to overcome difficulties that may arise when the system invariant technique is applied to the actual ultrasonic testing site.

Author Contributions

Conceptualization, H.-J.K. and S.-S.K.; methodology, H.-J.K. and S.-J.S.; software, J.P.; validation, H.-J.K., S.-S.K. and J.P.; formal analysis, J.P.; investigation, J.P.; resources, J.P.; data curation, J.P. and S.-E.L.; writing—original draft preparation, J.P.; writing—review and editing, J.P.; visualization, J.P.; supervision, H.-J.K. and S.-J.S.; project administration, S.-S.K.; funding acquisition, H.-J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Nuclear Safety Research Program of the Korea Foundation of Nuclear Safety (KoFONS), funded by the Nuclear Safety and Security Commission (No. 1805005).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

WGNWhite Gaussian noise
ANNArtificial neural network
FCNNFully connected neural network
CNNConvolutional neural network
ResNetResidual neural network

References

  1. Song, S.J.; Schmerr, L.W. Ultrasonic flaw classification in weldments using probabilistic neural networks. J. Nondestruct. Eval. 1992, 11, 69–77. [Google Scholar] [CrossRef]
  2. Masnata, A.; Sunseri, M. Neural network classification of flaws detected by ultrasonic means. NDT E Int. 1996, 29, 87–93. [Google Scholar] [CrossRef]
  3. Polikar, R.; Udpa, L.; Udpa, S.S.; Taylor, T. Frequency invariant classification of ultrasonic weld inspection signals. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 1998, 45, 614–625. [Google Scholar] [CrossRef] [PubMed]
  4. Margrave, F.W.; Rigas, K.; Bradley, D.A.; Barrowcliffe, P. The use of neural networks in ultrasonic flaw detection. Measurement 1999, 25, 143–154. [Google Scholar] [CrossRef]
  5. Song, S.J.; Kim, H.J.; Cho, H. Development of an intelligent system for ultrasonic flaw classification in weldments. Nucl. Eng. Des. 2002, 212, 307–320. [Google Scholar] [CrossRef] [Green Version]
  6. Simas Filho, E.F.; Silva, M.M.; Farias, P.C.; Albuquerque, M.C.; Silva, I.C.; Farias, C.T. Flexible decision support system for ultrasound evaluation of fiber–metal laminates implemented in a DSP. NDT E Int. 2016, 79, 38–45. [Google Scholar] [CrossRef]
  7. Cruz, F.C.; Simas Filho, E.F.; Albuquerque, M.C.; Silva, I.C.; Farias, C.T.; Gouvêa, L.L. Efficient feature selection for neural network based detection of flaws in steel welded joints using ultrasound testing. Ultrasonics 2017, 73, 1–8. [Google Scholar] [CrossRef]
  8. Meng, M.; Chua, Y.J.; Wouterson, E.; Ong, C.P.K. Ultrasonic signal classification and imaging system for composite materials via deep convolutional neural networks. Neurocomputing 2017, 257, 128–135. [Google Scholar] [CrossRef]
  9. Munir, N.; Kim, H.J.; Song, S.J.; Kang, S.S. Investigation of deep neural network with drop out for ultrasonic flaw classification in weldments. J. Mech. Sci. Technol. 2018, 32, 3073–3080. [Google Scholar] [CrossRef]
  10. Munir, N.; Kim, H.J.; Park, J.; Song, S.J.; Kang, S.S. Convolutional neural network for ultrasonic weldment flaw classification in noisy conditions. Ultrasonics 2019, 94, 74–81. [Google Scholar] [CrossRef]
  11. Munir, N.; Park, J.; Kim, H.J.; Song, S.J.; Kang, S.S. Performance enhancement of convolutional neural network for ultrasonic flaw classification by adopting autoencoder. NDT E Int. 2020, 111, 102218. [Google Scholar] [CrossRef]
  12. Rimoldi, B. Principles of Digital Communication: A Top-Down Approach; Cambridge University Press: Cambridge, UK, 2016. [Google Scholar]
  13. Gallager, R.G. Principles of Digital Communication; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  14. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation; California University San Diego La Jolla Inst for Cognitive Science (ICS-8506): San Diego, CA, USA, 1985. [Google Scholar]
  15. Rosenblatt, F. Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms; Spartan Books: Washington, DC, USA, 1961. [Google Scholar]
  16. Géron, A. Hands on Machine Learning with Scikit-Learn and Tensorflow: Concepts, Tools, and Techniques to Build Intelligent Systems; O’Reilly Media Inc.: Newton, MA, USA, 2017. [Google Scholar]
  17. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  18. Nielsen, M.A. Neural Networks and Deep Learning; Determination Press: San Francisco, CA, USA, 2015. [Google Scholar]
  19. Li, F.F.; Karpathy, A.; Johnson, J.; Yeung, S. CS231n: Convolutional Neural Networks for Visual Recognition, Stanford University: Stanford, CA, USA, 2016.
  20. Lecun, Y.; Bottou, L.; Benjio, Y.; Haffner, P. Gradient-based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  21. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  22. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for simplicity: The all convolutional net. arXiv, 2014; arXiv:1412.6806. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; Technical Report No. TR-2009; Computer Science Department, University of Toronto: Toronto, ON, Canada, 2009; Available online: https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf (accessed on 11 January 2022).
  25. Wu, S.; Zhong, S.; Liu, Y. Deep residual learning for image steganalysis. Multimed. Tool. Appl. 2018, 77, 10437–10453. [Google Scholar] [CrossRef]
  26. Shibata, N.; Tanito, M.; Mitsuhashi, K.; Fujino, Y.; Matsuura, M.; Murata, H.; Asaoka, R. Development of a deep residual learning algorithm to screen for glaucoma from fundus photography. Sci. Rep. 2018, 8, 14665. [Google Scholar] [CrossRef] [Green Version]
  27. Cheng, X.; Zhang, Y.; Chen, Y.; Wu, Y.; Yue, Y. Pest identification via deep residual learning in complex background. Comput. Electron. Agric. 2017, 141, 351–356. [Google Scholar] [CrossRef]
  28. Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
  29. Nair, V.; Hinton, G.E. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  30. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. Proc. icml 2013, 30, 3. [Google Scholar]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  32. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv, 2014; arXiv:1412.6980. [Google Scholar]
  33. Schmerr, L.W. Fundamentals of Ultrasonic Nondestructive Evaluation; Springer: New York, NY, USA, 2016; pp. 650–681. [Google Scholar]
  34. McGillem, C.D.; Cooper, G.R. Continuous and Discrete Signal and System Analysis, 2nd ed.; Holt, Rinehart & Winston: New York, NY, USA, 1984; p. 118. ISBN 0030617030. [Google Scholar]
  35. Weisstein, E.W. Convolution Theorem. From MathWorld—A Wolfram Web Resource. Available online: https://mathworld.wolfram.com/ConvolutionTheorem.html (accessed on 11 January 2022).
Figure 1. Flaw A scan signal example for database construction: (a) crack; (b) porosity.
Figure 1. Flaw A scan signal example for database construction: (a) crack; (b) porosity.
Applsci 12 01477 g001
Figure 2. Additive WGN: (a) non-; (b) SNR 15.
Figure 2. Additive WGN: (a) non-; (b) SNR 15.
Applsci 12 01477 g002
Figure 3. Time shifting of signals: (a) 10 points (0.2 μs) forward and backward shift; (b) 20 points (0.4 μs) forward and backward shift.
Figure 3. Time shifting of signals: (a) 10 points (0.2 μs) forward and backward shift; (b) 20 points (0.4 μs) forward and backward shift.
Applsci 12 01477 g003
Figure 4. Performance of ResNet trained with original database: (a) training accuracy curve; (b) testing accuracy curve.
Figure 4. Performance of ResNet trained with original database: (a) training accuracy curve; (b) testing accuracy curve.
Applsci 12 01477 g004
Figure 5. Schematics of the experimental setup to obtain a flaw signal in the weldment.
Figure 5. Schematics of the experimental setup to obtain a flaw signal in the weldment.
Applsci 12 01477 g005
Figure 6. Schematics of experiment setup to obtain reference signal.
Figure 6. Schematics of experiment setup to obtain reference signal.
Applsci 12 01477 g006
Figure 7. Flow diagram of the system invariant method.
Figure 7. Flow diagram of the system invariant method.
Applsci 12 01477 g007
Figure 8. Examples of cracks and porosity signals when applying system invariant method, from left to right: original signal, flaw scattering pattern in the frequency domain, and invariant signal; (a) crack signal with Olympus A430S (45 degrees, 2 MHz); (b) crack signal with KrautKramer WB70-4 (70 degrees, 4 MHz); (c) porosity signal with Olympus A430S (45 degrees, 2 MHz); (d) porosity signal with KrautKramer WB70-4 (70 degrees, 4 MHz).
Figure 8. Examples of cracks and porosity signals when applying system invariant method, from left to right: original signal, flaw scattering pattern in the frequency domain, and invariant signal; (a) crack signal with Olympus A430S (45 degrees, 2 MHz); (b) crack signal with KrautKramer WB70-4 (70 degrees, 4 MHz); (c) porosity signal with Olympus A430S (45 degrees, 2 MHz); (d) porosity signal with KrautKramer WB70-4 (70 degrees, 4 MHz).
Applsci 12 01477 g008aApplsci 12 01477 g008b
Figure 9. ResNet performance training with invariant database: (a) training accuracy curve; (b) testing accuracy curve; (c) comparison between accuracy and testing accuracy curves.
Figure 9. ResNet performance training with invariant database: (a) training accuracy curve; (b) testing accuracy curve; (c) comparison between accuracy and testing accuracy curves.
Applsci 12 01477 g009
Figure 10. ResNet testing performance comparison between original and invariant database.
Figure 10. ResNet testing performance comparison between original and invariant database.
Applsci 12 01477 g010
Table 1. List of transducers.
Table 1. List of transducers.
ManufacturerKrautKramerGEOlympusTKS
MWB 45-2WB 45-2WB 45-4MWB 45-4SWB 45-5A430S (45 ° , 2 MHz)4C14 X 14A45
MWB 60-2WB 60-2WB 60-4MWB 60-4SWB 60-5A430S (60 ° , 2 MHz)4C14 X 14A60
MWB 70-2WB 70-2WB 70-4MWB 70-4SWB 70-5A430S (70 ° , 2 MHz)4C14 X 14A70
Table 2. Ultrasonic weldment flaw database.
Table 2. Ultrasonic weldment flaw database.
Title 1FlawsNo. of Signals
1Crack2899
2Lack of Fusion1196
3Slag Inclusion634
4Porosity493
5Incomplete Penetration617
Total5839
Table 3. Ultrasonic weldment flaw database.
Table 3. Ultrasonic weldment flaw database.
Weldment Flaw Database
Training DatabaseTesting Database
Crack1740Crack1159
Lack of Fusion1076Lack of Fusion120
Slag Inclusion571Slag Inclusion63
Porosity444Porosity49
Incomplete Penetration555Incomplete Penetration62
Total4386Total1453
Table 4. Augmented database (original database).
Table 4. Augmented database (original database).
Original Database
Training DatabaseTesting Database
Crack26,100Crack1159
Lack of Fusion16,140Lack of Fusion120
Slag Inclusion8565Slag Inclusion63
Porosity6660Porosity49
Incomplete Penetration8325Incomplete Penetration62
Total65,790Total1453
Table 5. Detailed architecture of unit stage of ResNet.
Table 5. Detailed architecture of unit stage of ResNet.
Adopted ResNet Stage
Layer TypeKernel Size/StrideFeature MapsActivation FunctionDescription
1Input layer---x
2Conv 1 ( 1 × 3 ) / ( 1 × 1 )FMReluy = Conv 1 (x)
3Dropout0.5--y = Dropout(y)
4Conv 2 ( 1 × 3 ) / ( 1 × 1 )FMReluy = Conv 2 (y)
5Dropout0.5--y = Dropout (y)
6Shortcut---x = x
7Conv s ( 1 × 1 ) / ( 1 × 1 )FMRelux = Conv s (x)
8Add shortcut--Relux = x + y
9Conv 3 ( 1 × 3 ) / ( 1 × 1 )FMReluy = Conv 3 (x)
10Dropout0.5--y = Dropout (y)
11Conv 4 ( 1 × 3 ) / ( 1 × 1 )FMReluy = Conv 4 (y)
12Dropout0.5--y = Dropout (y)
13Shortcut---x = x
14Add shortcut--Relux = x + y
15Conv 5 ( 1 × 3 ) / ( 1 × 1 )FMReluy = Conv 5 (x)
16Dropout0.5--y = Dropout (y)
17Conv 6 ( 1 × 3 ) / ( 1 × 1 )FMReluy = Conv 6 (y)
18Dropout0.5--y = Dropout (y)
19Shortcut---x = x
20Add shortcut--Relux = x + y
Table 6. Architecture of ResNet.
Table 6. Architecture of ResNet.
Adopted ResNet Architecture
Layer TypeKernel Size/StrideFeature MapsOutput Size
1Input Layer--2048
2Conv 1 ( 1 × 32 ) / ( 1 × 8 )64 ( 256 × 64 )
3Dropout0.5- ( 256 × 64 )
4ActivationRelu- ( 256 × 64 )
5Max Pool ( 1 × 3 ) / ( 1 × 2 )- ( 128 × 64 )
6Stage 1-64 ( 128 × 64 )
7Max Pool ( 1 × 3 ) / ( 1 × 2 )- ( 64 × 64 )
8Stage 2-128 ( 64 × 128 )
9Max Pool ( 1 × 3 ) / ( 1 × 2 )- ( 32 × 128 )
10Stage 3-256 ( 32 × 256 )
11Dense Layer300-300
12Dropout0.5--
13Output Layer5-5
Table 7. ResNet performance comparison between the original and the invariant database.
Table 7. ResNet performance comparison between the original and the invariant database.
Original DatabaseInvariant Database
(Epoch > 100)TrainingTestingTrainingTesting
Average Accuracy (%)92.6362.1792.1591.45
Standard Deviation (%)0.644.130.431.77
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, J.; Lee, S.-E.; Kim, H.-J.; Song, S.-J.; Kang, S.-S. System Invariant Method for Ultrasonic Flaw Classification in Weldments Using Residual Neural Network. Appl. Sci. 2022, 12, 1477. https://doi.org/10.3390/app12031477

AMA Style

Park J, Lee S-E, Kim H-J, Song S-J, Kang S-S. System Invariant Method for Ultrasonic Flaw Classification in Weldments Using Residual Neural Network. Applied Sciences. 2022; 12(3):1477. https://doi.org/10.3390/app12031477

Chicago/Turabian Style

Park, Jinhyun, Seung-Eun Lee, Hak-Joon Kim, Sung-Jin Song, and Sung-Sik Kang. 2022. "System Invariant Method for Ultrasonic Flaw Classification in Weldments Using Residual Neural Network" Applied Sciences 12, no. 3: 1477. https://doi.org/10.3390/app12031477

APA Style

Park, J., Lee, S. -E., Kim, H. -J., Song, S. -J., & Kang, S. -S. (2022). System Invariant Method for Ultrasonic Flaw Classification in Weldments Using Residual Neural Network. Applied Sciences, 12(3), 1477. https://doi.org/10.3390/app12031477

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop