Next Article in Journal
Preliminary Study of Novel Bio-Crypto Key Generation Using Clustering-Based Binarization of ECG Features
Next Article in Special Issue
Context-Enhanced Network with Spatial-Aware Graph for Smartphone Screen Defect Detection
Previous Article in Journal
Multisensor Integrated Platform Based on MEMS Charge Variation Sensing Technology for Biopotential Acquisition
Previous Article in Special Issue
FR-PatchCore: An Industrial Anomaly Detection Method for Improving Generalization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advancing Ultrasonic Defect Detection in High-Speed Wheels via UT-YOLO

1
School of Physical Science and Technology, Southwest Jiaotong University, Chengdu 610036, China
2
Chengdu Lead Science & Technology Co., Ltd., Chengdu 610073, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(5), 1555; https://doi.org/10.3390/s24051555
Submission received: 31 January 2024 / Revised: 23 February 2024 / Accepted: 27 February 2024 / Published: 28 February 2024
(This article belongs to the Special Issue Sensing and Imaging for Defect Detection)

Abstract

:
In the context of defect detection in high-speed railway train wheels, particularly in ultrasonic-testing B-scan images characterized by their small size and complexity, the need for a robust solution is paramount. The proposed algorithm, UT-YOLO, was meticulously designed to address the specific challenges presented by these images. UT-YOLO enhances its learning capacity, accuracy in detecting small targets, and overall processing speed by adopting optimized convolutional layers, a special layer design, and an attention mechanism. This algorithm exhibits superior performance on high-speed railway wheel UT datasets, indicating its potential. Crucially, UT-YOLO meets real-time processing requirements, positioning it as a practical solution for the dynamic and high-speed environment of railway inspections. In experimental evaluations, UT-YOLO exhibited good performance in best recall, [email protected] and [email protected]:0.95 increased by 37%, 36%, and 43%, respectively; and its speed also met the needs of real-time performance. Moreover, an ultrasonic defect detection data set based on real wheels was created, and this research has been applied in actual scenarios and has helped to greatly improve manual detection efficiency.

1. Introduction

Wheels play a pivotal role as integral components of trains, directly influencing the safety and reliability of train operations [1]. With the rapid development and increasing speed of high-speed trains, corresponding safety risks are also escalating. The quality and structural condition of high-speed rail wheels have a direct impact on the safety and overall performance of an entire train. Operating at high speeds exposes the wheels to various external factors, including wear, impacts, and fatigue cracks, which will lead to deformation, damage, or even detachment, particularly under high-load and -frequency conditions. Wheel defects can lead to severe traffic accidents.
Non-destructive testing is a method that preserves the integrity of the object being examined [2], primarily being used for detecting defects in materials, components, and equipment. Traditional non-destructive testing (NDT) methods encompass ultrasonic testing (UT) [3], magnetic particle testing (MT) [4], and penetrant testing (PT) [5]. In the railway industry, owing to the unique environment, nearly all inspections rely on non-destructive testing. X-ray computed tomography has been employed to address the reconstruction of rolling contact fatigue cracks in rails [6], and various ultrasonic testing methods have been applied in axle defect detection [7,8]. Given the complex structures of wheels and the multitude of components, there are a variety of detection methods [9,10,11].
Traditional wheel defect detection methods predominantly hinge on visual inspection and ultrasonic testing. Visual inspection, however, is inefficient and prone to inaccuracies. Ultrasonic testing, relying on the principles of ultrasonic wave propagation in materials, involves emitting ultrasonic waves and capturing reflected signals. This approach enables the detection of various defects within wheels, such as cracks, inclusions, and porosities. Characterized by high sensitivity and resolution, this technology proves effective in identifying minute defects, offering early warnings of potential issues and contributing to the prevention of safety hazards arising from wheel failures.
The application of ultrasonic testing to train wheels goes beyond routine periodic inspections; it is also applicable in emergencies, such as post-collision or other external impact scenarios. Continuous monitoring of wheel health empowers operators to implement timely maintenance and repair measures, ensuring the safety and stability of trains during high-speed operations. However, due to the huge quantity of data, relying solely on manual detection is too inefficient. Traditional detection methods rely more on fixed methods. Matching and efficiency are difficult to guarantee. At the same time, subject to the limitations of expert knowledge, the detection process is relatively complicated. Therefore, a set of novel methods that can meet the above needs continues to be applied and promoted.
Deep learning technology represents a significant advancement in enhancing the efficiency and accuracy of computer vision as well as NDT. Its application extends to the automated analysis of various NDT data, including ultrasonic signals, magnetic particle images, and penetrant images, thereby substantially improving defect detection accuracy. Specifically, in ultrasonic wheel defect detection, deep learning technology offers several advantages. Firstly, it facilitates the automatic analysis of ultrasonic signals, thereby significantly enhancing detection efficiency. Secondly, it can extract intricate features from ultrasonic signals that may be challenging for human recognition, leading to a substantial improvement in detection accuracy. Additionally, deep learning technology contributes to cost reduction and minimizes manual intervention in the testing process. One notable strength of deep learning models is their ability to classify defects into different types, accurately locate defect positions, and measure defect sizes. These affordances not only streamline the analysis process but also provide a comprehensive understanding of the detected defects. By annotating ultrasound B-scan images, we can annotate a defect, which will give us a UT dataset. By learning from a large quantity of effective data, a model can efficiently detect data.
Despite these advantages, the application of deep learning technology in ultrasonic wheel defect detection still encounters certain challenges. One of the primary challenges is the creation of high-quality datasets. The performance of deep learning models heavily relies on the quality and diversity of the training data. Ensuring that the datasets used for training are representative of real-world scenarios is essential for achieving robust and reliable defect detection. Another challenge lies in the high complexity of deep learning models. While these models exhibit remarkable capabilities, their complexity can lead to challenges in interpretability and the need for significant computational resources. Striking a balance between model complexity and interpretability is crucial for practical implementation and widespread adoption. While deep learning technology has immense potential for revolutionizing ultrasonic wheel defect detection by significantly improving efficiency and accuracy, addressing challenges related to dataset creation and model complexity is essential for realizing its full benefits in real-world applications. To this end, we need to explain the categories of defects. In this work, the defect data are divided into four different forms according to the depth of scanning, namely, surface, near-surface, internal, and wheel plate cracks.
Given that natural images seldom exhibit exaggerated aspect ratios, standard deep-learning approaches may fall short when applied directly to ultrasonic images. Hence, this article explores specialized preprocessing techniques aimed at enhancing the compatibility between deep learning models and the intricate features inherent in ultrasonic wheel defect B-scan images. By addressing these specific challenges through dedicated preprocessing methods, the goal is to optimize the learning capacity of deep neural networks, fostering improved accuracy in detecting defects. This approach is crucial in bridging the gap between conventional image-processing methodologies and the distinctive characteristics of ultrasonic datasets, ultimately enhancing the efficacy of defect detection in high-speed railway train wheels. The design and training of deep learning models need to be tailored to accommodate these distinctive characteristics, ensuring optimal performance in detecting ultrasound defects. The main contributions of this paper are as follows:
  • An ultrasonic defect detection data set based on real wheels was created, and the defects were subdivided according to different categories;
  • An advanced UT-YOLO network was proposed to increase a network’s ability to perceive small defects and detection;
  • This research has been applied in actual scenarios and has helped to greatly improve manual detection efficiency.

2. Related Work

2.1. Traditional Defect Detection

Traditional defect detection methods are based on the characteristics of an object, such as color, size, shape, etc., and use image-processing methods and traditional machine learning methods for detection. Among them, the representative target detection methods include Scale-Invariant Feature Transform (SIFT) [12], Histogram of Oriented Gradients (HOG) [13], Oriented FAST and Rotated BRIEF (ORB) [14], and Harr [15]. It is necessary to use a pre-designed template to slide the window on the target image, and the steps are cumbersome. The extracted object features are all low-level, manually selected features, so they are not conducive to accurate detection and offer poor real-time detection. Moreover, these methods are mostly used for obvious features in natural images which have rich information. For these simple tasks, generally, Prewitt, Sobel, and Canny can be used to find solutions. However, in practice, especially in complicated wheel defect detection situations, noise is very serious, and much depends on the accuracy of the equipment and the skilled operators. All in all, due to the limitations of many factors, ultrasonic detection images are often not used in this type of detection method.
While conventional methods exhibit proficiency in straightforward industrial scenarios governed by clear rules, they encounter challenges when confronted with the intricacies and diversities inherent in real-world defects. Preceding the era of deep learning, object detection heavily relied on meticulously crafted features. Although this approach attained some degree of success, it grappled with notable limitations. For instance, the manual design of features often necessitated subjective decisions, resulting in inconsistencies and a lack of generalizability. The manual approach struggled to accommodate the vast diversity and complexity of real-world objects, and as the number of object categories and image variations expanded, the manual design process became increasingly cumbersome and impractical. Moreover, the broad spectrum of potential defects demanded multiple sets of pre-defined templates, constraining these methods’ capacity to identify novel and unexpected issues. Consequently, when confronted with intricate data, traditional methods frequently falter, necessitating laborious post-processing efforts.

2.2. Deep-Learning-Based Defect Detection

The rise of deep learning has transformed object detection, offering notable advantages over traditional methods. Convolutional neural networks (CNNs) excel at extracting high-level features from images, overcoming the limitations of manual feature crafting. Their seamless integration of feature extraction, selection, and classification enhances overall efficiency and effectiveness. Scholarly work consistently demonstrates the prowess of deep learning in defect detection, particularly in industrial settings.
The R-CNN network was proposed by Girshick et al. [16] in 2013. It is a landmark effort in the field of object detection. He et al. proposed SPP-Net [17], which no longer extracts features from all candidate areas and only extracts features of an entire image once, which greatly reduces the amount of calculation. Faster-RCNN [18] was the most representative model in the R-CNN series that was proposed by Ren et al. Because the Region Proposal Network (RPN) realized end-to-end training and testing of a network, detection accuracy was highly improved. However, Faster-RCNN is generally more complex, and its performance on small objects is not very impressive. He et al. proposed Mask RCNN [19], which has high accuracy in detection, but at the same time, instance segmentation will increase label price. The biggest problem for practical use is that the speed of the algorithm still cannot meet the industrial real-time detection requirements. In the industrial field, some researchers have adjusted the common detection frameworks to suit the detection of industrial data. Liu et al. [20] improved Faster-RCNN by utilizing cascade heads to detect metal defects. Cha et al. [21] changed the backbone to improve the detection time for bridge defects based on Faster-RCNN. He et al. [22] improved detection accuracy using multi-scale feature diffusion. The pursuit of greater detection speed has favored a large number of one-stage algorithms.
Redmon et al. [23] proposed the first one-stage detector, YOLO v1, which uses a multi-scale region centered on a grid to replace the RPN, greatly improving detection speed and meeting the needs of real-time detection but not in terms of accuracy. A novel object detection method called SSD [24] was proposed in 2016, combining the advantages of both Faster-RCNN and YOLO v1. It ensures high-precision detection while taking into account detection speed by using multi-scale regional features for regression, particularly via combining high-level and low-level feature maps. YOLO v2 [25] employs Darknet-19 as a backbone network of a model to extract features and reduce the computational complexity of the model. In addition, it further improves detection speed and accuracy by adding batch normalization, multi-scale training, and K-means dimension clustering after each convolutional layer. YOLO v3 [26] introduced several key advancements over its predecessors, leading to improved performance in both speed and accuracy. One of its most significant contributions is the Darknet-50 residual network, a newly designed architecture that utilizes residual connections for deeper feature extraction. This network was combined with the feature pyramid network (FPN) [27], which enabled multi-scale fusion prediction and detection on three different-scale feature maps. YOLO v4 [28] uses CSPDarkNet53 as a backbone network and optimizes to varying degrees using data processing, training methods, activation functions, loss functions, etc., to achieve the best detection effect at a given time. Li et al. [29] provided an end-to-end solution for the surface defect detection of steel strips based on improved YOLO. Zhang et al. [30] improved YOLO v3 by introducing a novel transfer learning method with fully pre-trained weights from a geometrically similar dataset to detect bridge surface damage.
However, all algorithms are based on the premise of sufficient data. However, in actual situations, especially in industrial environments, due to the particularity of application scenarios, sufficient and balanced labeled data sets are difficult to obtain, even if ordinary data augmentation cannot completely solve this problem. A novel automated defect synthesis network called Defect-GAN was designed by Zhang et al. [31] to generate realistic and diverse defect samples for training accurate and robust defect inspection networks. Valente et al. [32] used synthetic training data by simulating two types of print effects with image-processing and computer graphic techniques.
Building upon YOLO v5s as the baseline, this work introduces UT-YOLO, a novel ultrasonic testing defect detection method. UT-YOLO incorporates several key advancements over the baseline model, namely, enhanced backbone architectures with Swin-Transformer [33] and ResNet [34] variants, improved neck modules like BiFPN [35] for better feature aggregation, a dedicated multi-detection head for the accurate localization of small defects, a novel attention module for enhanced feature representation, and a practical pre-processing pipeline specifically tailored for real-world ultrasonic B-scan images, ensuring robust and effective defect detection in industrial applications.

3. Methodology

The basic architecture of YOLO v5 is shown in Figure 1. The whole operation process is as follows: Firstly, the images are readied for feature extracted by the backbone network, named CSPDarkNet53 [28]; then, the features in the neck section are separated into different sizes for feature fusion to obtain more feature information. After that, the different detection heads are designed, which allows the algorithm to fulfill the specific requirements of object detection.

3.1. Backbone

The architecture of the backbone is designed for feature extraction, but its efficiency should be given more attention. Due to this consideration, some novel backbone structures were applied in this task. The defects from ultrasonic testing B-scan images appear in different regions, so a useful feature extraction backbone should be designed to solve this problem. Regarding the wheels of railway trains, the defects are always small and noise-like, making them hard to identify easily. Figure 2 shows a typical residual connection, which is widely used in the design of backbone improvements. By introducing residual connections, the neural network is allowed to skip certain levels of learning, thereby alleviating the vanishing and exploding gradient problems and making training easier. This network also supports deeper networks. Traditional deep neural networks suffer from the problem wherein the training difficulty increases as the depth of the network increases. The residual structure allows building deeper networks by adding more layers without degradation issues.
Figure 3 shows the structure of the Swin-Transformer block. It allows a model to better capture global relationships among pixels in an image and avoids full connection to reduce calculation costs, affordances that are of vital importance in object detection, especially in ultrasonic defect testing. At the same time, by stacking multiple transformer blocks, the network is able to learn more complex and abstract feature representations.
Here, Z l 1 represents the output of layer l 1 or the input of layer l . The information passes from one layer to the next layer sequentially.
Firstly, the input is linearly transformed and normalized; then, it is passed through Weighted Multi-Head Self Attention (W-MSA), which focuses on a different area of the input rather than the global average. Secondly, the output will be added by the residual connection. The new output will repeat another Layer Normalization (LN) step and then connect with a Multi-Layer Perceptron (MLP) to fit the complicated input data. Finally, the residual connection will be used again to improve model stability. It is usually used as a repeated block, which helps alleviate vanishing or exploding gradient problems during training, making the network easier to train and optimize. In wheel defect feature extraction, after convolution and residual connection, the feature extraction module of Swin-Transformer can be superimposed multiple times to obtain more semantic information, which plays an important role in distinguishing background noise and defects.

3.2. Neck

BiFPN, a bidirectional feature propagation network, facilitates the bidirectional flow of information, allowing features to traverse from high-level layers to low-level layers, as well as enabling the propagation of low-level features to high-level layers. This dual-directional information exchange enhances the model’s ability to capture semantic information across various feature levels, thereby augmenting the model’s perceptual capabilities towards the target. In contrast, PANet adopts unidirectional propagation, limiting the flow of information to a single direction. The corresponding structure is shown in Figure 4.
Another noteworthy aspect is that BiFPN was meticulously designed to mitigate computational complexity without compromising model efficacy. It achieves this by judiciously fusing features at distinct levels, thereby sidestepping the computational overhead associated with full connections. In contrast, PANet exhibits a more intricate design that could entail a higher computational burden. Nevertheless, a novel enhancement of BiFPN made by incorporating the Simple, Parameter-Free Attention Module (SimAM) [36] was proposed. This addition enables the network to selectively concentrate on varying levels of features, thus refining its capacity for nuanced feature extraction. A diagram of SimAM is shown in Figure 5.

3.3. Head

Given that ultrasonic testing typically involves small targets, the existing detection heads may struggle to meet the required standards. Consequently, this study enhances the number of detection heads and simultaneously augments the depth of the detection heads within the target detection model. The objective is to enhance the model’s discernment of intricate scenes and nuanced features in pursuit of more effective performance. The new structure is shown in Figure 6.
The added detection head is shown in Figure 6. The structure in the dotted box in the lower half is consistent with that in Figure 1 but is rendered in gray. On this basis, in the upper block diagram of the network, a small target detection layer was added. Using color marking, the structure is used for the detection of small targets.

3.4. Loss Function

The original YOLOv5 loss function generally consists of three components, as represented by Equation (1). The first term quantifies localization loss, evaluating the accuracy of the model’s predictions for the position of the target bounding box. The second term corresponds to confidence loss, measuring how accurately the model predicts the presence of a target. The final term assesses the accuracy of the classification
L ori ( t p , t gt ) = k = 0 K [ α k w α box i = 0 S 2 j = 0 B I k i j o b j L CIoU + α obj i = 0 S 2 j = 0 B I k i j o b j L obj + α cls i = 0 S 2 j = 0 B I k i j o b j L cls ]
where K , S 2 , and B are the number of output features, the number of grids, and the number of anchors, respectively. α * is the weight for different sections; α box , α obj , and α cls are set as 0.05, 0.3, and 0.7, respectively. I k i j o b j represents whether the i - th cell, the j - th anchor box in the k - th feature map, is a positive sample. If it is positive, it is defined as 1; otherwise, it is set as 0. t p and t gt are the prediction vector and ground truth (GT), and α k w is the weight for different sizes of the output.
Equation (2) is the new loss function called L UT , which is designed for UT defect detection. The localization loss is replaced by L EIoU .
L UT ( t p , t gt ) = k = 0 K [ α k w α box i = 0 S 2 j = 0 B I k i j o b j L EIoU + α obj i = 0 S 2 j = 0 B I k i j o b j L obj + α cls i = 0 S 2 j = 0 B I k i j o b j L cls ]
Intersection over Union (IoU) is used to evaluate the quality of object detection models. A schematic diagram of IoU is shown in Figure 7.
Here, the value of IoU should be described as in Equation (3).
IoU = A B A B
Simply put, the overlapping region of A and B is C, which is denoted as A B . The area of A + B C is denoted as A B . The ratio of A B and A B is IoU, which is an indicator of prediction accuracy. In the original loss function, a complete IoU (CIoU) was used to solve the special cases with the same central point but different ratios of height to width. Equations (4)–(6) define the loss of CIoU
L CIoU = 1 CIoU
CIoU = IoU ( ρ 2 ( b , b g t ) c 2 + α ν )
ν = 4 π 2 ( a r c t a n w g t h g t a r c t a n w h ) 2
where α is weight and the ν is the parameter for measuring the ratio of height to width. A schematic diagram is shown in Figure 8: the green line (AB) is the parameter c , which signifies the diagonal distance of the minimum bounding box (the gray, dotted one) capable of encompassing both the predicted box and the GT box. Meanwhile, the parameter ρ is the Euclidean distance between two center points b and b g t , which are indicated by the red line.
At the same time, the definition of ν is shown in Equation (7):
ν = 4 π 2 ( a r c t a n w g t h g t a r c t a n w h ) 2
In Equation (2), the penalty term of EIoU consists of separating the influence factors of the aspect ratio based on the penalty term of CIoU to calculate the length and width of the target frame and anchor frame, respectively. So, L EIoU is changed according to Equation (8).
L EIoU = 1 IoU + ρ 2 ( b , b g t ) c 2 + ρ 2 ( w , w g t ) c w 2 + ρ 2 ( h , h g t ) c h 2
Therefore, L EIoU can provide a more accurate assessment when an object’s position is imprecise or slightly offset. However, to use L EIoU as the loss function, the dataset should be more closely scrutinized, especially regarding the balance between various types of data.
To strike a balance between real-time requirements and computational complexity, the original YOLOv5 algorithm was chosen as the baseline model. This work enhanced the baseline model by (1) utilizing dataset-preprocessing methods that incorporate the characteristics of ultrasonic B-scan data, enhancing the model’s compatibility with neural networks; (2) employing a combination of ResNet, Swin-Transformer, and BiFPN structures for feature extraction to capture richer feature information, thereby increasing the capability of feature extraction; (3) introducing SimAM and a dedicated small-object detection head to further optimize the detection accuracy for small objects; and (4) enhancing the accuracy of bounding box regression through the refinement of the loss function’s design.

4. Experiments and Results

Diverging from natural images, ultrasonic detection images primarily convey intensity information, rendering them particularly susceptible to noise interference. Moreover, as the focus of this study is on detecting defects in wheels, the acquired data hold exceptional value. The experiments were conducted alongside an exploration of datasets and various experiments.

4.1. Dataset

This work was evaluated with reference to a real dataset, which was collected with phased array UT (PAUT) and transmit–receive (TR) probes to attain B-scan images about four different defects according to different depths, surfaces near other surfaces, and internal and rim cracks in different service depots using the LU system [37]. The entire dataset of UT defects consisted of around 15,000 B-scan images, in which peeling, scratches, and cracks in different positions and depths were recorded. It is hard to find different real images with which to show the different types of defects, especially some inner defects, so a diagram of typical wheel defects is shown in Figure 9.
UT is a very useful solution for detecting inner defects, but analyzing surface and sub-surface areas is not its strength because of the large amount of initial wave interference at the contact surface, as well as the noise; last but not the least, due to the particularity of the wheel structure, the collection is generally completed in one cycle, so the aspect ratio of the collected images will be very large. Examples of the original collection data are shown in Figure 10. Image (a) below depicts a surface defect, and the periodic signal represents the wheel plate hole; there are six holes in a wheel.
The environment of the experimental data acquisition is shown in Figure 11. In the service depots, the wheel runs in the designated area, and the robotic arm will carry a phased array probe to collect the testing information. The collected signals will be stored in the industrial computer for preliminary analysis like different gains to give a short judgement.
To illustrate the practicability of the proposed method, this work briefly displays the typical real defects, including wheel rim cracks and some surface defects like scratches, with their B-scan images given below in Figure 12.
Another thing that should be taken into consideration is the variation in the depths of the scan with the different ratios of width to height. It was necessary to perform a data-preprocessing procedure. First, the images were cropped and reorganized into rectangles with an aspect ratio of 1 to the greatest extent possible, as in the algorithm for data preprocessing shown below (Algorithm 1).
Algorithm 1: Data Preprocessing
Input: Original Images.
Output: Processed Image after Data Preprocessing.
1:
// Obtain the width (w) and height (h) of the image
2:
w, h = getImageDimensions(originalImage)
3:
// Check if the width is greater than 3 times the height
4:
if w > 3 * h:
5:
// Crop the image into two halves and stack them vertically
6:
processedImage = cropAndStackImage(originalImage)
7:
else:
8:
// If width is not greater than 3 times height, no processing is done
9:
processedImage = originalImage
10:
end if
11:
// Return the processed image
12:
return processedImage
Then, the processed images needed to be labeled with different classes. Finally, the images were divided into three sets: a training set, a validation set, and a testing set with a ratio of 7:2:1, containing 10,395, 2970, and 1485 images, respectively.

4.2. Evaluated Index

In this work, two sets of metrics from academia and industry, respectively, are used to evaluate the model. Generally, the academic indexes that should be used are average precision (AP), recall, mAP, and frames per second (FPS); the industrial indexes always focus on the true alarm rate (TAR) and the false alarm rate (FAR) of detection.

4.2.1. Academic Indexes

A confusion matrix is often used to summarize the prediction effect of classification models, as shown in Table 1.
The calculations are shown in Equations (9)–(13):
p r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
A P = 0 1 p ( r ) d r
m A P = 1 n k = 1 n A P k
F P S = f r a m e N u m e l a p s e d T i m e = N T

4.2.2. Industrial Indicators

An example describing the true alarm rate (TAR) and false alarm rate (FAR) is provided here. Assuming there are T samples in total, and D is the number of defect samples, the rest of the samples, called N, are normal. Using an algorithm of detection, in the problem section, D − 1 out of D samples were found to be defects, and in the normal section, 2 out of N were identified as defects; an explanatory diagram is shown in Figure 13, and the calculation formulae should be like those in Equations (14) and (15).
T A R = D 1 D
F A R = 2 N
In this example, there is one more definition to provide, that is, the miss alarm rate (MAR), which in this case is 1 out of D, as represented in Equation (16).
M A R = 1 D 1 D = 1 D
In practice, there is a big gap between academia and industry. The requirements for statistical results are often stringent in the industrial setting, making the utilization of industrial indexes more challenging. In this work, academic metrics were adopted primarily to validate the feasibility of algorithmic improvements, serving as an assessment tool for evaluating the effectiveness of the algorithm.

4.3. Implementation Details

All the experiments were carried out on a computer running Windows 10. Two Nvidia GeForce RTX3090 GPUs were used, and the CPU version was Intel Core i9-1098XE. The deep learning framework was Pytorch 1.7.1, and the python version was 3.7.11. The image input was resized to 640 × 640, the batch size was 128, the optimizer was SGD, the initial learning rate was set to 0.001, the total number of iterations was set as 500 epochs, and the patience was set to 50, signifying that if there are 100 consecutive rounds without improvement, the system defaults to convergence.

4.4. Experiment Results

To evaluate the performance of the algorithm in question in terms of accuracy and speed in terms of both academic and industrial applications, UT-YOLO was compared with different algorithms. Table 2 presents the best recall, [email protected], and [email protected]:0.95 results for each algorithm, derived from the experimental observations of the proposed model with respect to the test set. Figure 14 and Figure 15 depict the [email protected] and [email protected]:0.95.
The results depicted in Figure 14 and Figure 15 show the effectiveness of UT-YOLO. Compared with the original baseline, the IoU0.5 indicator and the 0.5:0.95 indicator have been greatly improved, which is a very exciting result. In Figure 16, a comparative analysis of the detection results across different models is presented, highlighting the performance variations for a common object. Notably, the examination distinctly underscores the superior efficacy of the UT-YOLO model.
It can be observed in Figure 17 that the UT-YOLO model demonstrated precise localization of the defects, especially the small-sized defects. Furthermore, for defects near the wheel plate hole, the model exhibited commendable feature extraction capabilities, as evident in the outcomes presented in Figure 17. Despite the presence of some minor defects associated with background blur, the model displays robustness and can effectively detect defects, even in challenging conditions like strong background noise and electrical interference. The left side of each image is the original detection result, and the blue box on the right corresponds to the enlarged local features.
For real-world industrial inspection, additional sets of 50 defective samples and 50 non-defective samples were carefully chosen to assess the detection rate and false alarm rate. Among the 50 defective samples, consisting of a total of 6000 images, 210 were identified as defects by the UT inspectors. Using UT-YOLO, a total of 206 defect instances were successfully detected, resulting in a TAR of 98.10% (206 out of 210). Notably, the four undetected defects existed in three distinct wheelsets. In practical applications, the achieved TAR was 94% (47 out of 50), underscoring the heightened challenges faced in real-world scenarios.
Furthermore, a thorough examination of 6000 defect-free images from 50 wheelsets revealed 83 instances of false alarms, originating from 6 different wheelsets. According to statistical analysis based on the images, the calculated FAR was 1.38% (83 out of 6000). However, in actual application scenarios, the observed FAR was 12% (6 out of 50). These results emphasize the complexity and intricacies encountered in practical industrial applications, where achieving optimal detection indicators proves to be a more demanding task.
Last but not least, the efficiency of this method aligns with the demands of real-time collection and detection. In typical scenarios, a wheel completes one full rotation in about one minute. With a phased array probe utilizing 120 channels for data collection, the algorithm achieves a detection speed of approximately 2 s per wheel. Considering factors such as software response and data loading time, on-site application can accomplish the detection of a wheel and provide defect location information within 10 s, significantly diminishing the expenses associated with manual inspection.

5. Discussion

In this study, the application of the UT-YOLO method to ultrasonic inspection data of railway wheels was inspected, with a focus on detecting various types of defects. The evaluation of results underscores the superiority of this method, emphasizing its reliability both in field applications and on site. Among the models employed for object detection, the proposed UT-YOLO model exhibits significant advancements compared to several benchmark models. The outcomes indicate that UT-YOLO achieved a best recall of 0.94, an [email protected] of 0.89, and an [email protected]:0.9 of 0.64, surpassing baseline YOLOv5s by 37%, 36%, and 42%, respectively. Particularly noteworthy is UT-YOLO’s substantial superiority over other comparative algorithms in terms of [email protected]:0.9. Moreover, UT-YOLO demonstrated the fastest speed among the evaluated models. This is attributed to its unique features, such as an added small object detection layer, an attention mechanism module, data enhancement, and data preprocessing, specially designed for ultrasonic wheel b-scan image defect detection.
While the current study has effectively addressed challenges in ultrasonic wheel detection, especially in regard to speed and accuracy, achieving promising outcomes in terms of efficiency and practical applicability, it is crucial to acknowledge the diverse nature of real-world defects. Despite this success, certain limitations may arise in addressing specific prevalent issues. In particular, some false alarms caused by external factors such as electrical interference are still relatively serious, so the data collection process is very dependent on the stability of the equipment because the surface conditions of the wheels are not the same, which will cause the data to be easily invalidated, and a large amount of noise, especially electrical noise, is generated at the same time. However, to avoid excessive investment, more targeted distinctions will be made on the algorithm side to eliminate some special interferences. This will also be reflected in future work. Therefore, future endeavors will prioritize an in-depth exploration of actual wheel inspection equipment and the quality of inspection data. These tasks will involve a meticulous classification of various defects, aiming to enhance this methodology’s performance across a broader spectrum of real-life scenarios.

6. Conclusions

The study presents a noteworthy breakthrough in defect detection methods utilizing ultrasound B-scan images obtained through the application of UT-YOLO. The UT-YOLO model incorporates specific modifications that significantly enhance the accuracy of defect detection, particularly in the context of B-scan images of ultrasonic defects. The notable enhancements include the integration of small-object detection layers, residual structures for feature fusion, attention mechanisms, and specialized processing methods. Collectively, these adjustments culminate in a robust model poised to address the challenges inherent in wheel defect detection in real-world ultrasonic B-scan images. When compared with the baseline YOLOv5s model, UT-YOLO demonstrated marked performance improvements. In the wheel defect dataset test results, it achieves noteworthy enhancements of 37%, 36%, and 43% in best recall, [email protected], and [email protected]:0.95, respectively. Simultaneously, the speed of UT-YOLO reached an impressive 69FPS, indicating its capacity to operate effectively in real-time applications and making it well suited for deployment in practical scenarios.
Despite the UT-YOLO algorithm’s notable successes in UT defect detection, there are still some areas that may be further investigated and refined. To further lower the false alarm rates and missed alarm rates, future work will concentrate on distinguishing the noise and the noise-like defects as well as the whole automatic process in train wheel defect detection to achieve truly intelligent detection.

Author Contributions

Conceptualization, Q.Z.; methodology, Q.Z.; software, Q.Z. and K.T.; validation, Q.Z., K.T. and J.P.; formal analysis, Q.Z.; investigation, Q.Z. and K.T.; resources, A.W. and J.P.; data curation, Q.Z., K.T. and J.P.; writing—original draft preparation, Q.Z.; writing—review and editing, J.P.; visualization, Q.Z. and K.T.; supervision, J.P., J.L. and X.G.; project administration, J.P., J.L. and X.G.; funding acquisition, J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Nature Science Foundation of China (No. 61771409) and Science and Technology Project of Sichuan (No. 2021YJ0080).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We appreciate the hard work of every author, the valuable suggestions of every reviewer, and the patience and meticulousness of the academic editors.

Conflicts of Interest

Author Ai wang was employed by the company Chengdu Lead Science & Technology. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zerbst, U.; Mädler, K.; Hintze, H. Fracture mechanics in railway applications-An overview. Eng. Fract. Mech. 2005, 72, 163–194. [Google Scholar] [CrossRef]
  2. Dwivedi, S.K.; Vishwakarma, M.; Soni, A. Advances and researches on non destructive testing: A review. Mater. Today Proc. 2018, 5, 3690–3698. [Google Scholar] [CrossRef]
  3. Drinkwater, B.W.; Wilcox, P.D. Ultrasonic arrays for non-destructive evaluation: A review. NDT E Int. 2006, 39, 525–541. [Google Scholar] [CrossRef]
  4. Kasai, N.; Takada, A.; Fukuoka, K.; Aiyama, H.; Hashimoto, M. Quantitative investigation of a standard test shim for magnetic particle testing. NDT E Int. 2011, 44, 421–426. [Google Scholar] [CrossRef]
  5. Roshan, C.C.; Raghul, C.; Ram, H.V.; Suraj, K.; Solomon, J. Non-destructive testing by liquid penetrant testing and ultrasonic testing—A review. Int. J. Adv. Res. Ideas Innov. Technol. 2019, 5, 694–697. [Google Scholar]
  6. Naeimi, M.; Li, Z.; Qian, Z.; Zhou, Y.; Wu, J.; Petrov, R.H.; Sietsma, J.; Dollevoet, R. Reconstruction of the rolling contact fatigue cracks in rails using X-ray computed tomography. NDT E Int. 2017, 92, 199–212. [Google Scholar] [CrossRef]
  7. Kappes, W.; Hentschel, D.; Oelschlägel, T. Potential improvements of the presently applied in-service inspection of wheelset axles. Int. J. Fatigue 2016, 86, 64–76. [Google Scholar] [CrossRef]
  8. Mineo, C.; Cerniglia, D.; Pantano, A. Numerical study for a new methodology of flaws detection in train axles. Ultrasonics 2014, 54, 841–849. [Google Scholar] [CrossRef] [PubMed]
  9. Le, M.; Jun, J.; Kim, J.; Lee, J. Nondestructive testing of train wheels using differential-type integrated Hall sensor matrixes embedded in train rails. NDT E Int. 2013, 55, 28–35. [Google Scholar] [CrossRef]
  10. Montinaro, N.; Epasto, G.; Cerniglia, D.; Guglielmino, E. Laser ultrasonics inspection for defect evaluation on train wheel. NDT E Int. 2019, 107, 102145. [Google Scholar] [CrossRef]
  11. Cavuto, A.; Martarelli, M.; Pandarese, G.; Revel, G.M.; Tomasini, E.P. Train wheel diagnostics by laser ultrasonics. Measurement 2016, 80, 99–107. [Google Scholar] [CrossRef]
  12. Lindeberg, T. Scale invariant feature transform. Scholarpedia 2012, 7, 10491. [Google Scholar] [CrossRef]
  13. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 886–893. [Google Scholar]
  14. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  15. Rainer, L.; Maydt, J. An extended set of Haar-like features for rapid object detection. In Proceedings of the International Conference on Image Processing, Rochester, NY, USA, 22–25 September 2002; Volume 1. [Google Scholar]
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  19. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  20. Liu, Z.; Tang, R.; Duan, G.; Tan, J. TruingDet: Towards high-quality visual automatic defect inspection for mental surface. Opt. Lasers Eng. 2021, 138, 106423. [Google Scholar] [CrossRef]
  21. Cha, Y.J.; Choi, W.; Suh, G.; Mahmoudkhani, S.; Büyüköztürk, O. Autonomous structural visual inspection using region-based deep learning for detecting multiple damage types. Comput.-Aided Civ. Infrastruct. Eng. 2018, 33, 731–747. [Google Scholar] [CrossRef]
  22. He, Y.; Song, K.; Meng, Q.; Yan, Y. An end-to-end steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans. Instrum. Meas. 2019, 69, 1493–1504. [Google Scholar] [CrossRef]
  23. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  24. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  25. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  26. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  27. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  28. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  29. Li, J.; Su, Z.; Geng, J.; Yin, Y. Real-time detection of steel strip surface defects based on improved yolo detection network. IFAC-PapersOnLine 2018, 51, 76–81. [Google Scholar] [CrossRef]
  30. Zhang, C.; Chang, C.C.; Jamshidi, M. Concrete bridge surface damage detection using a single-stage detector. Comput.-Aided Civ. Infrastruct. Eng. 2020, 35, 389–409. [Google Scholar] [CrossRef]
  31. Zhang, G.; Cui, K.; Hung, T.-Y.; Lu, S. Defect-GAN: High-fidelity defect synthesis for automated defect inspection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 2524–2534. [Google Scholar]
  32. Valente, A.; Wada, C.; Neves, D.; Neves, D.; Pere, F.V.M.; Megeto, G.A.S.; Cascone, M.H.; Gomes, O.; Lin, Q. Print defect mapping with semantic segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass, CO, USA, 1–5 March 2020; pp. 3551–3559. [Google Scholar]
  33. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  35. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  36. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  37. Zhang, Y.; Tan, Y.; Peng, J.; Peng, C.; Yang, K.; Gao, X. LU automatic ultrasonic inspection system for in-service wheelset and its application. In Proceedings of the 2016 18th International Wheelset Congress (IWC), Chengdu, China, 7–10 November 2016; pp. 100–104. [Google Scholar]
Figure 1. The architecture of YOLO v5.
Figure 1. The architecture of YOLO v5.
Sensors 24 01555 g001
Figure 2. The structure of a residual connection.
Figure 2. The structure of a residual connection.
Sensors 24 01555 g002
Figure 3. The structure of Swin-Transformer.
Figure 3. The structure of Swin-Transformer.
Sensors 24 01555 g003
Figure 4. (a) Schematic diagram of PANet; (b) schematic diagram of BiFPN.
Figure 4. (a) Schematic diagram of PANet; (b) schematic diagram of BiFPN.
Sensors 24 01555 g004
Figure 5. The diagram of SimAM.
Figure 5. The diagram of SimAM.
Sensors 24 01555 g005
Figure 6. The new small-object detection head.
Figure 6. The new small-object detection head.
Sensors 24 01555 g006
Figure 7. The schematic diagram of IoU.
Figure 7. The schematic diagram of IoU.
Sensors 24 01555 g007
Figure 8. A schematic diagram of prediction box (blue) and GT (yellow).
Figure 8. A schematic diagram of prediction box (blue) and GT (yellow).
Sensors 24 01555 g008
Figure 9. Diagram of typical wheel defects.
Figure 9. Diagram of typical wheel defects.
Sensors 24 01555 g009
Figure 10. Examples of (a) original UT B-scan with a defect; (b) original UT B-scan with a defect and annotation; and (c) original UT B-scan data without defects.
Figure 10. Examples of (a) original UT B-scan with a defect; (b) original UT B-scan with a defect and annotation; and (c) original UT B-scan data without defects.
Sensors 24 01555 g010
Figure 11. Typical scenarios: (a) wheel detection in service depots; (b) wheel; (c) PA probes.
Figure 11. Typical scenarios: (a) wheel detection in service depots; (b) wheel; (c) PA probes.
Sensors 24 01555 g011
Figure 12. Typical defects: (a) wheel rim crack; (b) scratch; (c) ultrasonic b-scan defect image of rim crack; (d) surface and near-surface b-scan defect images.
Figure 12. Typical defects: (a) wheel rim crack; (b) scratch; (c) ultrasonic b-scan defect image of rim crack; (d) surface and near-surface b-scan defect images.
Sensors 24 01555 g012
Figure 13. Example explaining the industrial indicators.
Figure 13. Example explaining the industrial indicators.
Sensors 24 01555 g013
Figure 14. [email protected] for the ablation experiment.
Figure 14. [email protected] for the ablation experiment.
Sensors 24 01555 g014
Figure 15. [email protected]:0.95 for the ablation experiment.
Figure 15. [email protected]:0.95 for the ablation experiment.
Sensors 24 01555 g015
Figure 16. Comparison of detection results achieved using different models.
Figure 16. Comparison of detection results achieved using different models.
Sensors 24 01555 g016
Figure 17. Detection results and magnified details. (a) surface and near-surface crack; (b) inner defect; (c) the defect near the web plate hole region; (d) Surface and near-surface hole-like defects; (e) wheel rim crack.
Figure 17. Detection results and magnified details. (a) surface and near-surface crack; (b) inner defect; (c) the defect near the web plate hole region; (d) Surface and near-surface hole-like defects; (e) wheel rim crack.
Sensors 24 01555 g017
Table 1. Confusion matrix.
Table 1. Confusion matrix.
Actual PositiveActual Negative
Predicted PositiveTPFP
Predicted NegativeFNTN
Table 2. Results regarding detection for the UT-B scan dataset.
Table 2. Results regarding detection for the UT-B scan dataset.
ModelBest Recall[email protected][email protected]:0.95FPS
YOLOv5s0.570.530.2155
YOLOv5s + BiFPN0.630.580.2158
YOLOv5s + BiFPN + SimAM0.670.630.2557
YOLOv5s + Swin-Transformer0.850.820.6164
UT-YOLO (Ours)0.940.890.6469
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Q.; Peng, J.; Tian, K.; Wang, A.; Li, J.; Gao, X. Advancing Ultrasonic Defect Detection in High-Speed Wheels via UT-YOLO. Sensors 2024, 24, 1555. https://doi.org/10.3390/s24051555

AMA Style

Zhang Q, Peng J, Tian K, Wang A, Li J, Gao X. Advancing Ultrasonic Defect Detection in High-Speed Wheels via UT-YOLO. Sensors. 2024; 24(5):1555. https://doi.org/10.3390/s24051555

Chicago/Turabian Style

Zhang, Qian, Jianping Peng, Kang Tian, Ai Wang, Jinlong Li, and Xiaorong Gao. 2024. "Advancing Ultrasonic Defect Detection in High-Speed Wheels via UT-YOLO" Sensors 24, no. 5: 1555. https://doi.org/10.3390/s24051555

APA Style

Zhang, Q., Peng, J., Tian, K., Wang, A., Li, J., & Gao, X. (2024). Advancing Ultrasonic Defect Detection in High-Speed Wheels via UT-YOLO. Sensors, 24(5), 1555. https://doi.org/10.3390/s24051555

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop