Next Article in Journal
Negentropy-Based Sparsity-Promoting Reconstruction with Fast Iterative Solution from Noisy Measurements
Previous Article in Journal
Polarization-Sensitive and Wide Incidence Angle-Insensitive Fabry–Perot Optical Cavity Bounded by Two Metal Grating Layers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Method for Detecting the Randomness of Barkhausen Noise in a Material Fatigue Test Using Sensitivity and Uncertainty Analysis

1
School of Aeronautics and Astronautics, University of Electronic Science and Technology of China, Chengdu 611731, China
2
China Special Equipment Inspection and Research Institute, Beijing 100029, China
3
School of Mechanical Engineering, North University of China, Taiyuan 030051, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5383; https://doi.org/10.3390/s20185383
Submission received: 27 July 2020 / Revised: 4 September 2020 / Accepted: 17 September 2020 / Published: 20 September 2020
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
The magnetic Barkhausen noise (MBN) signal provides interesting clues about the evolution of microstructure of the magnetic material (internal stresses, level of degradation, etc.). This makes it widely used in non-destructive evaluation of ferromagnetic materials. Although researchers have made great effort to explore the intrinsic random characteristics and stable features of MBN signals, they have failed to provide a deterministic definition of the stochastic quality of the MBN signals. Because many features are not reproducible, there is no quantitative description for the stochastic nature of MBN, and no uniform standards to evaluate performance of features. We aim to make further study on the stochastic characteristics of MBN signal and transform it into the quantification of signal uncertainty and sensitivity, to solve the above problems for fatigue state prediction. In the case of parameter uncertainty in the prediction model, a prior approximation method was proposed. Thus, there are two distinct sources of uncertainty: feature(observation) uncertainty and model uncertainty were discussed. We define feature uncertainty from the perspective of a probability distribution using a confidence interval sensitivity analysis, and uniformly quantize and re-parameterize the feature matrix from the feature probability distribution space. We also incorporate informed priors into the estimation process by optimizing the Kullback–Leibler divergence between prior and posterior distribution, approximating the prior to the posterior. Thus, in an insufficient data situation, informed priors can improve prediction accuracy. Experiments prove that our proposed confidence interval sensitivity analysis to capture feature uncertainty has the potential to determine the instability in MBN signals quantitatively and reduce the dispersion of features, so that all features can produce positive additive effects. The false prediction rate can be reduced to almost 0. The proposed priors can not only measure model parameter uncertainties but also show superior performance similar to that of maximum likelihood estimation (MLE). The results also show that improvements in parameter uncertainties cannot be directly propagated to improve prediction uncertainties.

1. Introduction

The Barkhausen effect specifically refers to the phenomenon in which the magnetic flux of a magnetized material changes discontinuously due to magnetic domain inversion during magnetization [1]. As an important nondestructive testing (NDT) method, the magnetic Barkhausen noise (MBN) signal is sensitive to the changes in many material properties and has many applications in the fields of material stress and hardness detection [2,3,4,5], metal fatigue state analysis [6], metal microstructure transformation, and grain size measurement [7].
As the direct carrier for excitation and reception of MBN signal, the characteristics of sensor determine the quality of the MBN detection result. We adopted a self-developed MBN sensor with high spatial resolution. In detail, the MBN sensor is mainly composed of an excitation device and an MBN signal receiving device. The excitation part is a U-shaped yoke composed of a plurality of overlapping silicon steel sheets, and the excitation coil is composed of a multi-hit coil wound around the U-shaped yoke, which are used to excite an alternating magnetic field. The receiving device uses a magnetic core-wound coil receiver, which is generally placed between the two yokes of the U-shaped yoke while keeping perpendicular to the test piece for measurement. In order to improve the spatial resolution and measurement stability of the sensor, we add a layer shielding shell with high permeability and conductivity to the outside of the receiver, limiting the effective signal receiving area of the MBN sensor to a relatively small local area. In addition, we increased the number of turns for the pickup coil, so as to achieve the purpose of getting good sensitivity and frequency response characteristics in a lower frequency range.
Intensity changes in an MBN signal with material state transformation are usually reflected and described by the representative parameters calculated in time-frequency (TF) domain such as the amplitude, energy, root mean square (RMS), waveform full width at half maximum (FWHM), envelope, peak time, threshold, and power spectrum [3,4,5,8,9,10]. However, affected by the microscopic magnetic anisotropy of the material itself, measurement performance, and experimental magnetization parameters (such as magnetization intensity and frequency, excitation waveform), the MBN has an obvious stochastic nature and the application of more automatic signals processing procedures used for extraction, selection, and fusion of signal features containing critical and distinctive information about the material properties are urgently required. Figure 1 provides an intuitive understanding of the stochastic characteristic of the MBN signal, in which distinctive differences can be drawn from different wave packets of MBN signals. According to the relevant literatures, multiple signal processing methods have been applied to get more representative information of MBN signal. Magalas [11] introduced wavelet transform into MBN signal analysis; then, Miesawicz et al. [12] further expanded the study. Luo et al. [13] applied the auto-regressive modeling method to obtain a single-sided power spectrum (also called a PSD or an AR spectrum), which is a near-deterministic expression of the MBN signal. In [14,15,16], Hilbert–Huang Transform and Short-Time Fourier Transform (STFT) are used to analyze the joint TF representation and properties of MBN signal. In addition, wide new features are also proposed. Vashista et al. [17] proposed two parameters of “count” and “event” of MBN signal; on this basis, Hang [10] used a genetic algorithm (GA) for appropriate threshold selection. Su et al. [4] performed first- and second-order derivation on the AR spectrum and manually extracted some peaks from it as new features. Li et al. [5] adopted a modified slow feature analysis (m-SFA) to enrich the content of PSD by fusing the AR spectrum with different orders. There are many types of features, which can be roughly divided into energy feature, statistical feature, and shape feature, and great differences may manifest for different features used in different applications. When a variety of new features continue to be extracted, the problem of feature explosion arises, and various dimensionality reduction methods are needed to decrease the redundancy of features such as PCA [4,9], FCA [9], and forward-selection combined with cross-validation [18]. It is complex work.
Although researchers have proposed many excellent methods for MBN signal processing so that features can have good linearity and low dispersion, reducing the influence of the random characteristics of MBN signal on the detection or prediction results. There are three problems that must be considered in these methods. The first one is that many new features are not reproducible on cycling. In some cases, once the materials, experimental equipment and methods are changed, the new features may be useless and are not always superior to the conventional features, lacking in universality and versatility. The second, definite explanations and definitions about the laws of the stochastic nature of MBN signal haven’t been made. Thirdly, due to the particularity of different features, we cannot use uniform standards to analyze, define, and classify the performance of features, making the features analysis and selection complicated.
Summarizing the works of others, we aim to make further study on the internal quality of the random characteristics of MBN signal and transform it into the accurate quantification of signal uncertainty and sensitivity, to solve the three problems above. For data uncertainty measurement, the most commonly used are probabilistic methods based on data statistics such as Monte Carlo method and Bayesian analysis [19]. Among them, Bayesian analysis has a great advantage in making full use of historical information and continuously integrating new data to learn the accurate probability distribution of data when the sample is unbalanced and short. In Bayesian learning, clarifying the role of prior and likelihood information in Bayesian inference of large models is necessary. Likelihood reflects the amount of data to some degree and Prior represents how certain the researcher is about model parameters being estimated [20]. In [21,22], the discussion has also been given out. It can be concluded that, when accurate informed priors are specified, the amount of theory (via the prior) incorporated into the model estimation process increases. Specifically, with strong and accurate priors, less data would likely be needed to properly estimate parameter values. In contrast, more data would be needed in cases where no theory was incorporated. Helton et al. [23] carried out eigenvalue uncertainty analysis on nuclear data using a survey of sampling-based statistical methods and discussed the influence of data uncertainty caused by the data amount and the degree of dispersion on the uncertainty of model prediction. In [21], Sarah et al. used Bayesian statistics for mixture uncertainty modeling, illustrating the importance of a prior sensitivity analysis, and discussing how to interpret results that fluctuate with different prior settings.
Based on these analyses, in a complex system, two distinct sources of uncertainty can be considered: feature(observation) uncertainty and model uncertainty [24,25]. In this paper, we specify that the stochastic quality of MBN signal is characterized as features uncertainty. Model uncertainty arises from the inability to specify an exact value for a parameter that is assumed to have a constant value in the respective investigation, and accounting for uncertainty in model parameter estimations makes the model robust to alleviate overfitting without a need for regularization [26]. Specifically, Bayes by Backprop modeled on variational inference was proposed to measure the weight uncertainty in a neural network [27]. Likewise, we try to model the parameters uncertainty of a multivariable linear regression (MLR) model, in which the posterior distribution of parameters can be inferred by Bayes’ theorem without using complicated variational inference. Moreover, output uncertainty analysis can be used to study how the sensitivity propagates from the uncertainty changes of input features and model parameters to the model predictions, which is called the prediction uncertainty. In other words, uncertain outputs can be treated as functions of uncertainty analysis of inputs and model parameters. In general, we aim to quantify uncertainty in the input features of MBN signals, uncertainty in the parameters of the learning model, and uncertainty in the output predictions produced by the input and parameters characterized as probability distribution based on the modeling of sensitivity and uncertainty. This has the potential to improve modeling flexibility and accuracy. The benefit of using Bayesian learning in our study is that they offer a unified and consistent set of tools for uncertainty and sensitivity measurements for modeling, inference, prediction, and model selection [28,29]. The methods given in this article include the following:
First, feature uncertainty is modeled by reparameterization sampling. Specifically, we perform interval sensitivity analysis based on the confidence intervals of the feature distribution. A certain interval value is selected as the measurement index of signal uncertainty by applying reparameterization sampling of feature matrices. When the uncertainty is maintained, a more robust feature group is reconstructed by uniformly quantizing the features extracted from the same signal wave packet. Then, model uncertainty is measured by BLR. For BLR, prior distribution (prior) selection and training are of great importance. Incorporating informed priors into the estimation process can improve convergence and prediction accuracy. Due to a lack of prior knowledge, we propose an effective method in this paper to assign model parameters more reasonable priors by calculating the gradient descent of the Kullback–Leibler (KL) divergence between the prior distribution and posterior distribution. In addition, prediction uncertainty is characterized as the statistical mean and variance of prediction for each fatigue state and is used as the evaluation index of the results of interval sensitivity analysis and prior sensitivity analysis. Prediction uncertainty is applied to measure the sensitivity of observation uncertainty and model uncertainty.
The rest of this paper is arranged as follows: Section 2 introduces the experimental setup and measurements. The main algorithm derivation and modeling process are illustrated in Section 3. Section 4 presents the experiment results and related analysis. Finally, the research in this paper is summarized and concluded in Section 5.

2. Experimental Setup and Measurements

The experimental system mainly consists of a low-frequency fatigue test machine (Shimadzu (China) Co., Ltd. Beijing, China) the MBN measurement instrument and the fatigue test piece. The measurement system was self-built and included a signal generator Handyscope hs3 (TiePie engineering, Amazing Tech Co., Ltd, Shenzhen, China), power amplifier Newton LPA05A (Newtons4th Ltd, Beijing Miko-Xinye Electronics Technology Co., Ltd. Beijing, China), preamplifier Stanford SR560 (Stanford Research Systems), self-developed current collector and magnetic Barkhausen signal sensor, and computer. The testing specimens are manufactured from carburized 20R steel (American: AISI/SAE, 1020, Britain: BS, IC22), whose chemical composition is shown in Table 1, and the full-size and shape were designed according to National standard. In the process of specimen preparation, the surface is ground to Ra1.6 roughness, then annealed at a high temperature of 550 °C for 6 h, and then the surface oxide scale with diluted hydrochloric acid is removed. The maximum magnetic field strength in the test sample is 99 A/m, and the saturation induction intensity of the sample is 1.5–2T, and the magnetic field intensity must be greater than 10,000 A/m when it reaches saturation. Then, tensile tests were carried out on specimens before the fatigue test. The lower yield strengths of the two specimens were 297 MPa and 301 MPa, and the corresponding tensile strengths were 443 MPa and 448 MPa. Through trial and error, 345 MPa was selected as the maximum tensile stress for the fatigue test. The test stress ratio was 0, and the stress change was in the form of a sinusoid wave with a frequency of 15 Hz. The MBN signal and the magnetic characteristic signal excited at the center of the specimen were measured after a certain cycle period.
The sensor used in the experiment is a magnetic core wound coil type shielded sensor designed based on non-destructive testing technology, which aims to solve the problem of low spatial resolution of the current coil wound sensor. The coil is made of manganese-zinc ferrite, and the noise of the coil itself without the specimen is consistent with the noise level of the system without the sensor. In the process of measurement, we use a rubber band to fix the sensor probe on the surface of the material. The tension of the rubber band is very small and can be ignored. The yoke of sensor contacts the specimen, and the pick-up is located between the two yokes and not in contact with specimen, so it is not affected by head pressure. The configuration block diagram of the experimental system is shown in Figure 2; during the experiment, the software on the computer controls the signal generator to emit a sinusoidal signal at the specified frequency. The sinusoidal signal is amplified by the power amplifier and then sent to the excitation coil through the current collector so that the alternating magnetic field is generated in the tested component and the MBN signal is excited.
Moreover, considering that the received signals are susceptible to the excitation signal and external factors, the measured MBN signals were filtered by a bandpass filter to remove the low-frequency and high-frequency interference signals. The detailed excitation and acquisition parameter settings are listed in Table 2. In order to better observe and compare the random characteristics of the different wave packets, we set a long signal acquisition channel to ensure that a sufficient number of signal wave packets (150) with the same sampling points (35,000) could be split from the continuous signal. Meanwhile, a dataset with enough samples was created for prediction.
The whole life cycle of ferromagnetic materials consists of the loading time from loading free to break. Due to the dispersiveness of fatigue, in order to reduce the experimental error, the same fatigue experiment and MBN signal measurement were carried out on three specimens taken from the same base material. The fatigue life cycle of three specimens measured in the experiment were 42,712 times, 54,643 times and 47,762 times, respectively. To explore the MBN signal across the whole life cycle of the ferromagnetic materials, the MBN signal was collected every 1000 cycles until the materials broke. To avoid the influence of experimental measurement errors, we randomly selected MBN signals under 12 measuring moments that corresponded to certain fatigue loading times for analysis, and we regarded each loading time as a different fatigue state.

3. Modeling for the Uncertainty

Given a dataset, the choice of model and parametrization of the regression function are of the utmost importance. Our aim is to establish an accurate model using probability uncertainty and sensitivity analysis that does this reasonably well. This paper focuses on uncertainty modeling, including the uncertainty analysis of features, model parameters, and predictions. In any practical setting, we have access to only a finite, potentially large, amount of data for selecting the model class and the corresponding parameters. Given that this finite amount of training data does not cover all possible scenarios, we may want to describe the remaining parameter uncertainty to obtain a measure of confidence of the model’s prediction at test time. The smaller the training set is, the more important the uncertainty modeling.
Consistent modeling of uncertainty provides model predictions with confidence bounds. The overall implementation process of this paper is illustrated in Figure 3 and includes two main parts. The first part summarizes the modeling process of observation uncertainty, in which signal preprocessing and feature extraction are carried out and the uncertainties in the signal are characterized as multiple normal distributions of the features. The second part is the measurement of model parameter uncertainties for prior setting and posterior training.
Given a set of training example pairs D = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x N , y N ) } , we aim to find a function f that maps input x R K to corresponding function values y R . For this dataset, N and K denote the size of the feature matrices characterized as X = { x 1 , x 2 , , x N } R N × K . Each element in the sample space X refers to an example with K features. We index the examples and features as n = 1 , 2 , , N and k = 1 , 2 , , K , respectively. We assume that a linear model is used for prediction. The functional relationship between x and y is given by
y = x T θ + ε , ε ~ N ( 0 , σ 2 )
where θ R K are the model parameters (regression coefficients) that we are about to measure and ε is an independent random variable that describes the measurement noise, which can also be called zero-mean Gaussian-distributed noise.

3.1. Observation Uncertainty Analysis Based on Reparameterization Sampling

In this section, the method of modeling observation uncertainty using probability distributions is given. The stochastic nature of MBN signal in this paper can be understood as signal instability, which affects the robustness and characterization of features. That is, the same feature extracted from different wave packets has different values. Based on these considerations, we aim to reconstruct the features to varying reparameterization degrees (manifested as confidence intervals of a Gaussian distribution) from the probability perspective to infer and define the uncertainty in the MBN signals.
In observation uncertainty analysis of any multi-to-one response system using sampling methods, three steps are mainly considered: (1) Calculate the mean vector and covariance matrices to define the probability distribution for observation uncertainty characterization; (2) Reconstruct the sample space of observations by reparameterization sampling; and (3) Send the reconstructed sample space into the predictor for the response, in addition to analyzing the uncertainty in the response. The feature transformation based on reparameterization sampling is shown in Figure 4.
Feature reparameterization is conducted based on the substantial wave packets of the same fatigue states. To reconstruct the sample space, the distribution space and joint probability density function of each input parameter x R K must be determined first. Let a multivariate Gaussian distribution approximate the spatial probability distribution of the input parameters, which is parameterized by a mean vector μ x R K and a covariance matrix Σ x . We write X ~ N ( μ x , Σ x ) , and the joint probability density is given as
p ( x | μ x , Σ x ) = ( 2 π ) K 2 | Σ | 1 2 exp ( 1 2 ( x μ x ) T | Σ x | 1 ( x μ x ) )
The empirical mean vector μ ( μ = [ μ 1 , μ 2 , , μ K ] ) and covariance matrix are defined as
μ x = 1 N n = 1 N x n
Σ x = C o v [ x , x ] = [ cov [ x 1 , x 1 ] cov [ x 1 , x 2 ] cov [ x 1 , x K ] cov [ x 2 , x 1 ] cov [ x 2 , x 2 ] cov [ x 2 , x K ] cov [ x K , x 1 ] cov [ x K , x K ] ]
where cov [ x i , x j ] is the covariance between two univariate random feature variables. The covariance matrix is symmetric and positive semidefinite and tells us something about the spread of the data space. The elements on its diagonal represent the variance of the features.
For a general multivariate Gaussian distribution, that is, where the mean is nonzero and the covariance is not the identity matrix, characterized as N ( μ x , Σ x ) , we use the properties of linear transformation of a Gaussian random variable. Specifically, we define x s = f ( x ) , where f is a function of a kind of linear transformation. The characteristic of the MBN signal is positively correlated with its feature uncertainty. The greater the instability of MBN signal is, the greater the feature uncertainty. We know that different confidence intervals in the feature joint probability distribution refer to different value ranges of feature variables. The larger the confidence interval is, the greater the feature uncertainty. Considering the three confidence intervals (68.2%, 95.4% and 99.7%) shown in Figure 5, we would like to construct three different multiple sample spaces from a sampler that provides samples from the multivariate Gaussian distribution of the features.
According to the reparameterization method, we assume ζ N ( 0 , 1 ) and x s = k A ζ + μ , where is pointwise multiplication, Σ = A A T is a Gaussian distribution with mean vector μ and covariance matrix Σ , and k is the coefficient that is set to increase the diversity of the sampling transformation. One convenient choice for A is to use the Cholesky decomposition of the covariance matrix Σ = A A T . We should know that computing the Cholesky factorization of a matrix is symmetric and positive definite and that the covariance matrices calculated in our problem just meet this requirement. Three types of reconstructed sample space with different value ranges can be parameterized as
X s = { μ ± Σ 1 / 2 ζ , [ μ σ , μ + σ ] μ ± 2 Σ 1 / 2 ζ , [ μ 2 σ , μ + 2 σ ] μ ± 3 Σ 1 / 2 ζ , [ μ 3 σ , μ + 3 σ ]

3.2. Model Uncertainty Analysis of Bayesian Priors

We specify the conditional probability distribution of the output given the input for a particular parameter setting. The Gaussian likelihood is then defined as
p ( y | x , θ ) = N ( y | x T θ , σ 2 I )
Previously, we looked at linear regression models where we estimated the model parameters θ by means of maximum likelihood estimation (MLE: θ M L = arg max θ p ( Y | X , θ ) ) or maximum a posteriori estimation (MAP: θ M A P = arg max θ p ( θ | X , Y ) ). In practice, both MLE and MAP use a point estimate from which single specific parameter values are calculated and easily lead to overfitting. Bayesian linear regression allows us to reason about model parameters, that is, to place an a posteriori probability distribution over plausible parameter settings when making predictions. This means that we do not fit any specific parameters but introduce uncertainty in linear model parameters, which is achieved by approximating the corresponding distributions over all the parameters (prior and posterior distributions are considered). The probability relationship between the priors and the posterior probability distribution can be established by Bayesian theory:
p ( θ | X , Y ) = p ( Y | X , θ ) p ( θ ) p ( Y | X )
Although the prior and its corresponding posterior do not necessarily obey the same distribution in the actual situation, in order to facilitate calculations in the experiment, we only consider the general ideal situation. That is, we assume that the prior probability and posterior probability follow the same form of distribution with different parameters. We place a Gaussian prior distribution over the model parameters. Then, we consider the priors (prior distributions) and likelihood:
p ( θ ) = N ( θ | μ 0 , Σ 0 )
p ( Y | X , θ ) = N ( y | x T θ , σ 2 I )
According to Equation (7), we know that the posterior is proportional to the product of the prior and the likelihood, so if both the likelihood and prior distributions are Gaussian, then the posterior can also be modeled by a Gaussian distribution. We assume the posterior of model parameters
p ( θ | D ) = N ( θ | μ p , Σ p )
where μ p and Σ p can be parameterized over the parameters of priors and likelihood based on Bayesian theory (7). In some detail,
p ( θ | X , Y ) = N ( θ | Σ p ( Σ 0 1 μ 0 + σ 2 x T y ) , ( Σ 0 1 + σ 2 x T x ) 1 )
From the above formula, we know that Σ p is a function of Σ 0 , and μ p is a function of μ 0 and Σ 0 . The selection of priors tends to be subjectively biased. In general, enough sample data can largely eliminate this prior bias. Any priors will yield roughly the same results; this is also known as the prior robustness of Bayesian learning. However, when there is a lack of sample data, as in our experiment, the limitations of the sample information and prior bias will lead to a large number of blind searches, causing low efficiency and precision of the posterior [30,31]. The fitting degree of the posterior parameters to the data depends largely on prior selection. That is, the closer the prior is to the true distribution, the faster and more accurately the posterior distributions can be found. From another perspective, we ignore the restrictive effect of the data and make the priors and the posterior as consistent as possible. Then, the result will be more ideal. On this basis, we compute the parameters μ 0 and Σ 0 of the prior distributions on the regression coefficients θ to minimize the Kullback–Leibler divergence with the functional Bayesian posterior:
p * ( θ ) = arg min μ 0 , Σ 0 K L ( p ( θ ) | | p ( θ | D )
Let
L ( θ ) = K L ( p ( θ ) | | p ( θ | D )    = 1 2 ( t r ( Σ p 1 Σ 0 ) + ( μ 1 μ 0 ) T Σ p 1 ( μ 1 μ 0 ) k + log ( det Σ p det Σ 0 ) )
where k is the dimensionality of the distribution. Then, we respectively calculate the gradient of L ( θ ) with respect to the mean μ 0 and standard deviation σ 0 ( σ 0 is the root of elements on the diagonal of covariance matrix Σ 0 ). The prior parameters are updated as μ 0 = μ 0 α Δ μ 0 and σ 0 = σ 0 α Δ σ 0 . Once the priors are obtained, the posterior is updated based on Equation (11). Then, we iterate to solve the more ideal posterior probability distribution. The iterative process uses the posterior probability parameters calculated at each step to readjust (or replace) the prior, and then the corresponding posterior is calculated again. If the effect of the second posterior is better than that of the first through some measurement parameters or methods, then this process continues until the newly obtained posterior probability no longer changes (converges). Otherwise, the iteration is terminated. Finally, based on the above reasoning, we obtain the predictive distribution of y * at an arbitrary test input x * as
p ( y * | X , Y , x * ) = N ( y * | x * T μ p , x * T Σ p x * + σ 2 )

4. Experimental Results and Discussion

Based on a review of other studies on MBN signals, a commonly used preselected feature set (including energy, rms, mean, entropy, skewness, and the peak of the power spectral density) is used for our experiment, in which the three types of features are included. We use a sliding window with a length of a signal period to split continuous MBN signals and directly calculate the features of the single wave packet. The uncertainty in the MBN signals is modeled based on their multiple wave packets. The experiments in this paper were verified by 10-fold cross-validation (CV). To compare the prediction results of different methods, the prediction uncertainties are measured (one for mean and one for variance). The regression evaluation index root mean square error ( R M S E ) and r-square ( R 2 ) are calculated as well.

4.1. Interval Sensitivity Analysis Based on Disorder Sampling

As described in Section 3.1, we statistically analyze the feature matrices corresponding to different wave packets measured under the same fatigue state based on Equations (3) and (4). This is used to create their associated probability distributions. Then, a set of probability distributions T M × K is defined to characterize the uncertainty in features, in which the elements are represented in the form of Equation (2). M denotes the measuring moments m = 1 , 2 , , M . Then, we set up three different sensitivity intervals corresponding to the three predictive confidence intervals 68.2%, 95.4%, and 99.7% of the standard normal distribution to reconstruct the sample space.
Based on Equation (5), three sets of sample space consistent with the probability distribution T are reconstructed, where the value of Gaussian noise ζ is randomly sampled from the standard normal distribution to simulate the stochastic condition of the MBN signal. The exact uncertainty interval conditions are listed in Table 3. The sensitivity interval ± in Table 3 represents the original sample space, which is preserved to facilitate comparison with the reconstructed sample space. The lager the sensitivity interval is, the more scattered the feature distribution.
The interval sensitivity analysis of the predictions is performed by exploring the change rule of prediction uncertainty, as shown in Table 4. Table 4 lists the statistical information of the predicted mean E ( R i ) = 1 S j = 1 S R i , j i = 1 , 2 , , m and standard error V ( R i ) = 1 S 1 j = 1 S ( E ( R i ) R i , j ) 2 i = 1 , 2 , , m for the prediction results of each fatigue loading time under the conditions of using the original sample space and the reconstructed sample space as model inputs. R i represents the fatigue state response of predictions, and S is the total number of test data points for each fatigue state.
Figure 6 is a box plot corresponding to the statistical information in Table 4, where the black broken line describes the change in fatigue loading times corresponding to each fatigue state in the test dataset. The overall prediction accuracies, characterized by R M S E and R 2 , are listed in Table 5. According to Table 4, Figure 6, and Table 5, it can be observed that the larger the sensitivity interval that is set, the greater the prediction uncertainty is. At the same time, the prediction accuracy is significantly reduced, indicating that the observation uncertainty will directly affect the uncertainty in and accuracy of the predictions. In addition, by comparing the prediction results from the original sample space (sensitivity interval: ± ) and the reconstructed sample space with a sensitivity interval of 68.2% (sensitivity interval: ± σ ), we found that their prediction accuracy and uncertainty under different fatigue states are somewhat similar. To some extent, we have successfully characterized the uncertainty conditions of feature distributions in the original sample space. On the basis of adding random Gaussian noise, the uncertainty of the MBN signals can be highly restored within the [ μ σ , μ + σ ] confidence interval of the feature distributions.

4.2. Feature Reparameterization Based on Ordered Sampling

According to Section 4.1, the stochastic condition of MBN signal is defined by analyzing feature uncertainty using interval sensitivity analysis. However, our idea is not limited to this approach. Our purpose is to improve the robustness of features through feature reconstruction on the basis of retaining feature uncertainty. We know that the stochastic quality of the MBN signal is intuitively presented in the disordered form of signal wave packets measured under the same fatigue state. Therefore, we must perform the noise quantization on these feature differences, making the different features extracted from the same wave packet have a consistent law to follow based on the random sampling within a confidence interval centered on the mean. Here, quantization refers to applying the same Gaussian noise value ζ i N ( 0 , 1 ) , i = 1 , 2 , , N to the feature reconstruction in each example of the sample space. All reconstruction sensitivity intervals are selected as ± σ based on the conclusion of Section 4.1. Following this idea, a more robust feature sample space is reconstructed.
To better compare the quality of features before and after reconstruction, the mapping between inputs and outputs is established by using two model types: multiple linear regression (MLR) and multilayer perceptron (MLP). The prediction accuracies of different numbers of features are shown in Figure 7. We found that no matter how many features are selected, the prediction results of the reconstructed features are superior to those of the original features. For the original features, the prediction accuracy improved slightly when the number of features increased, indicating that the feature dimension has little effect on the improvement of prediction results. However, for the reconstructed features, as the number of features increased, the prediction results improved significantly, and the prediction accuracy value was closer to 1. According to the analysis, the positive addiction effect is produced between features, and we infer that, as long as there are enough features that can be selected, the prediction accuracy can be infinitely close to 1.
A prediction comparison using 12 features is presented in Table 6. Figure 8 shows detailed information about the prediction uncertainty of various fatigue states using MLR and MLP. It can be concluded that, no matter which model is chosen, the reconstructed features can greatly improve the prediction accuracy and reduce the output uncertainty compared with the original features. To some extent, the feature uncertainty analysis using the reparameterization sampling method with ordered random noise is beneficial for eliminating the influence of the stochastic characteristic of MBN signal. This provides clear distinctions between the reconstructed features associated with different fatigue states (labels in experiments) so that the models can more easily recognize different labels.

4.3. Developed Priors Setting for Model Uncertainty Measurement

Despite the popularity of the training methods of MLE and MAP, a lack of parameter uncertainty measures makes MLR prone to making overconfident predictions. In addition, regularization is always needed. Taking parameter uncertainties into account, the Bayesian linear regression (BLR) developed with MLR is applied, in which Bayesian estimation is used to estimate the probability distributions of regression coefficients. In BLR, priors (prior distributions) are of the highest significance and represent the knowledge accumulation of the researchers about the model parameters being estimated. Actually, the diagonal Gaussian distribution is typically used as the priors in the absence of prior knowledge, with the aim of relying on a good amount of data and multiple iterations to approximate a robust posterior. However, this greatly fails to make the prediction accuracy equivalent to that of MLE. A new method of prior setting is introduced in Section 3.2., and the effect will be verified and compared with that of MLE. Table 7 lists the posterior probability distributions trained by our method and the diagonal Gaussian distribution. The optimal parameters given by MLE are also listed.
Figure 9 shows the prediction uncertainty obtained when using the two types of priors. According to Table 6 and Figure 9, the prediction results of our proposed method for prior setting outperform the commonly used one. Meanwhile, the effect of our method is almost equivalent to that of MLE. This indicates that assigning a prior distribution reasonably is important for improving the efficiency and accuracy of model prediction and that our proposed method can enhance the precision of prior assignments without any informed prior information. Meanwhile, we found that, no matter which priors are used, the uncertainty of the obtained posterior parameters has little effect on the prediction results and can be ignored. This indicates that improvements in parameter uncertainties cannot be directly propagated to improve prediction uncertainty.

5. Conclusions

Understanding the stochastic characteristic of MBN signals is of great significance for improving the efficiency and accuracy of material property analysis. In this paper, a series of uncertainty and sensitivity analyses using Bayesian statistics is applied to model uncertainty in MBN signal features, model parameters, and predictions characterized as multivariate probability distributions. Feature uncertainty usually manifests as signal instability and is often difficult to suppress or evaluate. Model uncertainty is caused by unstable parameter tuning in the process of training. In addition, the uncertainty of the output is measured correspondingly as the input or model parameters change under different probability uncertainties. The main contributions of this paper include the reconstruction of the original feature space by reparameterization sampling. Through experimental analysis, we finally determined that the original feature space can be better restored by modeling within the probability confidence interval of [ σ , + σ ] and adding random noise that obeys the standard normal distribution. The probability confidence interval is thus used as the measurement index of the stochastic quality of MBN signal. On this basis of reconstruction, interval [ σ , + σ ] is maintained, and a more robust feature space can be obtained by adding the uniform quantization Gaussian noise to each example in the feature space, making the features extracted from the same wave packet have the same random distribution characteristics. The results proved that the performance of reconstructed features is much better than original features, and a good linear superposition effect is generated between the features. With the increase of feature dimensions, the prediction accuracy has been significantly improved. Moreover, to measure the uncertainties of the model parameters given that both the prior knowledge and data volume are limited, we proposed a more useful method for incorporating informative prior into training using the method of approximating Kullback–Leibler divergence between the prior distribution and posterior distribution. The criteria for selecting the optimal priors were chosen with the goal of maintaining a consistent prior and posterior. The results showed that our method is superior to the commonly used diagonal Gaussian prior and is comparable to MLE.
The analysis of MBN signal uncertainty proposed in this paper is not limited to the current experimental background. Rather, this analysis is expected to be further verified and expanded in other application scenarios, such as hardness and stress testing and material characteristic curve calibration.

Author Contributions

Conceptualization, Y.H. and X.L.; Data curation, Y.H. and Y.Z.; Formal analysis, Y.H.; Funding acquisition, X.L. and Y.Z.; Investigation, X.L., J.Z., and J.T.; Methodology, Y.H. and X.L.; Project administration, Y.Z. and X.C.; Resources, Y.Z., J.Z., and J.T.; Software, Y.H. and J.T.; Validation, X.L.; Writing—original draft, Y.H.; Writing—review and editing, X.L. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Key Research and Development Program of China, Grant No. 2017YFF0209702, the National Natural Science Foundation of China, Grant No. 51405066; the Fundamental Research Funds for the Central Universities, Grant No. ZYGX2019J083.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Manh, T.L.; Caleyo, F.; Hallen, J.M. Model for the correlation between magnetocrystalline energy and Barkhausen noise in ferromagnetic materials. J. Magn. Magn. Mater. 2018, 454, 155–164. [Google Scholar] [CrossRef]
  2. Sorsa, A.; Leiviska, K.; Santa-aho, S. An efficient procedure for identifying the prediction model between residual stress and barkhausen noise. J. Nondestruct. Eval. 2013, 32, 341–349. [Google Scholar] [CrossRef]
  3. Sorsa, A.; Leiviska, K.; Santa-aho, S. A data-based modelling scheme for estimating residual stress from Barkhausen noise measurements. Insight Non Destruct. Test. Cond. Monit. 2012, 54, 278–283. [Google Scholar] [CrossRef]
  4. Sun, G.G.; Liu, H.; He, C.F. A Novel prediction method for hardness using auto-regressive spectrum of barkhausen noise. J. Nondestruct. Eval. 2018, 37, 85. [Google Scholar] [CrossRef]
  5. Li, Z.B.; Sun, G.G.; He, C.F. Prediction of the hardness of X12m using Barkhausen noise and component analysis methods. J. Magn. Magn. Mater. 2019, 478, 59–67. [Google Scholar] [CrossRef]
  6. Lindgren, M.; Lepisto, T. Barkhausen noise evaluation of fatigue in high strengthsteel. Int. J. Mater. Prod. Technol. 2006, 26, 140–151. [Google Scholar] [CrossRef]
  7. Ranjan, R.; Jiles, D.; Rastogi, P.K. Magnetic properties of decaburized steels: An investigation of the effects of grain size and carbon content. IEEE Trans. Magn. 1987, 23, 1869–1876. [Google Scholar] [CrossRef]
  8. Ding, S.; Tian, G.Y.; Moorthy, V. New feature extraction for applied stress detection on ferromagnetic material using magnetic Barkhausen noise. Measurement 2015, 73, 515–519. [Google Scholar] [CrossRef]
  9. Ding, S.; Tian, G.Y.; Sutthaweekulb, R. Non-destructive hardness prediction for 18CrNiMo7-6 steel based on feature selection and fusion of Magnetic Barkhausen Noise. NDT E Int. 2019, 107, 102138. [Google Scholar] [CrossRef]
  10. Hang, C.; Liu, W.B.; Wang, P. A Method of Barkhausen Noise feature extraction based on an adaptive threshold. Appl. Sci. 2019, 9, 2964. [Google Scholar] [CrossRef] [Green Version]
  11. Magalas, L.B. Application of the wavelet transform in mechanical spectroscopy and in Barkhausen noise analysis. J. Alloy Compd. 2000, 310, 269–275. [Google Scholar] [CrossRef]
  12. Miesowicz, K.; Staszewski, W.J.; Korbiel, T. Analysis of Barkhausen noise using wavelet-based fractal signal processing for fatigue crack detection. Int. J. Fatigue 2016, 83, 109–116. [Google Scholar] [CrossRef]
  13. Luo, X.Y.; Wang, Y.L.; Wang, L. Non-destructive hardness measurement of hot-stamped high strength steel sheets based on magnetic barkhausen noise. Proc. Eng. 2014, 81, 1768–1773. [Google Scholar] [CrossRef] [Green Version]
  14. Yang, M.; Liu, W. Hilbert-Huang Transform Analysis and Feature Extraction of Barkhausen Signal. Electra. Meas. Tech. 2017, 8, 180–183. [Google Scholar]
  15. Michal, M.; Grzegorz, P. Use of time-dependent multispectral representation of Magnetic Barkhausen Noise signals for the needs of non-destructive evaluation of steel materials. Sensors 2019, 19, 1443. [Google Scholar]
  16. Michal, M.; Grzegorz, P. Time-Frequency Analysis of Barkhausen Noise for needs of anisotropy evaluation of grain-oriented steels. Sensors 2020, 20, 768. [Google Scholar]
  17. Vashista, M.; Paul, S. Novel processing of Barkhausen noise signal for assessment of residual stress in surface ground component exhibiting poor magnetic response. J. Magn. Magn. Mater. 2013, 323, 2579–2584. [Google Scholar] [CrossRef]
  18. Mao, K.Z. Orthogonal forward selection and backward elimination algorithms for feature subset selection. IEEE Trans. Syst. Man Cybern. Part B 2004, 34, 629–634. [Google Scholar] [CrossRef]
  19. Wang, R.Q.; Liu, F. Uncertainty quantification of low-cycle fatigue life model based on Bayesian theory. Acta Aeronaut. Astronaut. Sin. 2017, 9, 249–258. [Google Scholar]
  20. Muller, U.K. Measuring prior sensitivity and prior informativeness in large Bayesian models. J. Monet. Econ. 2012, 59, 581–597. [Google Scholar] [CrossRef]
  21. Depaoli, S.; Yang, Y.Z.; Felt, J. Using Bayesian Statistics to model uncertainty in mixture models: A sensitivity analysis of priors. Struct. Equ. Model. 2017, 24, 198–215. [Google Scholar] [CrossRef]
  22. Baldwin, S.A.; Fellingham, G.W. Bayesian methods for the analysis of small sample multilevel data with a complex variance structure. Psychol. Methods 2013, 18, 151–164. [Google Scholar] [CrossRef] [PubMed]
  23. Helton, J.C.; Johnson, J.D. Survey of sampling-based methods for uncertainty and sensitivity analysis. Reliab. Eng. Syst. Saf. 2006, 91, 1175–1209. [Google Scholar] [CrossRef] [Green Version]
  24. Armen, D.K.; Ove, D. Aleatory or epistemic? Does it matter? Struct. Saf. 2009, 31, 105–112. [Google Scholar]
  25. Cacuci, D.G.; Ionescu-Bujor, M. A Comparative review of sensitivity and uncertainty analysis of large-scale systems—II: Statistical methods. Nucl. Sci. Eng. 2004, 147, 204–217. [Google Scholar] [CrossRef]
  26. Shridhar, K.; Laumann, F.; Liwicki, M. A Comprehensive guide to Bayesian Convolutional Neural Network with Variational Inference. arXiv 2019, arXiv:1901.02731. [Google Scholar]
  27. Blundell, C.; Julien, C. Weight Uncertainty in Neural Networks. In Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015. [Google Scholar]
  28. Alex, G. Practical variational inference for neural network. Adv. Neur. Inf. Process. Syst. J. 2011, 391, 2348–2356. [Google Scholar]
  29. Muthen, B.; Asparouhov, T. Bayesian structural equation modeling: A more flexible representation of substantive theory. Psychol. Methods 2012, 17, 313–335. [Google Scholar] [CrossRef]
  30. Lin, S.M.; Wang, S.C.; Lu, Y.C. Bayesian method of computational learning mechanism and problem solving. J. Tsinghua Univ. Sci. Technol. 2000, 40, 61–64. [Google Scholar]
  31. Depaoli, S. Mixture class recovery in GMM under varying degrees of class separation: Frequentist versus Bayesian estimation. Psychol. Methods 2013, 18, 186–219. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Random characteristics of magnetic Barkhausen noise (MBN) signal.
Figure 1. Random characteristics of magnetic Barkhausen noise (MBN) signal.
Sensors 20 05383 g001
Figure 2. Experiment system.
Figure 2. Experiment system.
Sensors 20 05383 g002
Figure 3. Flowchart of uncertainty modeling.
Figure 3. Flowchart of uncertainty modeling.
Sensors 20 05383 g003
Figure 4. Feature reparameterization.
Figure 4. Feature reparameterization.
Sensors 20 05383 g004
Figure 5. Selection of confidence intervals for standard normal distribution.
Figure 5. Selection of confidence intervals for standard normal distribution.
Sensors 20 05383 g005
Figure 6. Comparison of prediction uncertainty using the four different sample spaces.
Figure 6. Comparison of prediction uncertainty using the four different sample spaces.
Sensors 20 05383 g006
Figure 7. The prediction for various multiple linear regression (MLR) models.
Figure 7. The prediction for various multiple linear regression (MLR) models.
Sensors 20 05383 g007
Figure 8. Prediction comparison of original and reconstructed feature sample spaces using MLR and MLP.
Figure 8. Prediction comparison of original and reconstructed feature sample spaces using MLR and MLP.
Sensors 20 05383 g008
Figure 9. Comparison of prediction uncertainty using different priors setting.
Figure 9. Comparison of prediction uncertainty using different priors setting.
Sensors 20 05383 g009
Table 1. The chemical composition of steel 20R (: %).
Table 1. The chemical composition of steel 20R (: %).
CSiMnPSNiCrCu
0.17~0.230.17~0.370.35~0.65≤0.035≤0.035≤0.03≤0.25≤0.25
Table 2. Parameters table of signal acquisition.
Table 2. Parameters table of signal acquisition.
ParametersValue
Excitation frequency9 Hz/16 Hz/35 Hz
Output voltage of signal generator0.5 V
Sampling frequency1 MHz
Barkhausen wave packet number150
Power magnification10
MBN signal magnification20
Bandpass filtering5 kHz–50 kHz
Table 3. The setting of uncertainty intervals used in the experiment.
Table 3. The setting of uncertainty intervals used in the experiment.
Serial Num1234
Sensitivity interval ± ± σ ± 2 σ ± 3 σ
Table 4. The uncertainty in the predictions.
Table 4. The uncertainty in the predictions.
Serial Num1234
0268.47 ± 0.52   ×   10 3 46.29 ± 1.42   ×   10 3 −202.78 ± 1.67   ×   10 3 −169.63 ± 2.64   ×   10 3
30003483.88 ± 0.67   ×   10 3 2774.34 ± 1.98   ×   10 3 1308.89 ± 3.14   ×   10 3 1177.93 ± 3.13   ×   10 3
50004680.88 ± 0.81   ×   10 3 4663.44 ± 1.43   ×   10 3 3891.78 ± 2.45   ×   10 3 3999.97 ± 2.73   ×   10 3
70007414.28 ± 1.33   ×   10 3 8032.62 ± 1.27   ×   10 3 8959.23 ± 2.37   ×   10 3 9244.01 ± 2.62   ×   10 3
10,0009197.50 ± 1.23   ×   10 3 10,749.56 ± 1.62   ×   10 3 11,264.93 ± 1.99   ×   10 3 11,997.91 ± 2.77   ×   10 3
13,00012,793.82 ± 1.81   ×   10 3 13,537.09 ± 1.55   ×   10 3 13,978.91 ± 2.24   ×   10 3 15,265.55 ± 2.46   ×   10 3
20,00021,699.65 ± 2.07   ×   10 3 21,135.74 ± 1.28   ×   10 3 22,116.94 ± 1.96   ×   10 3 22,448.67 ± 3.16   ×   10 3
24,00024,810.07 ± 2.21   ×   10 3 24,876.67 ± 1.54   ×   10 3 25,253.09 ± 2.42   ×   10 3 25,771.84 ± 2.46   ×   10 3
28,00029,094.94 ± 1.99   ×   10 3 29,142.15 ± 1.30   ×   10 3 28,612.62 ± 2.25   ×   10 3 28,719.35 ± 3.03   ×   10 3
32,00028,419.50 ± 2.38   ×   10 3 28,615.83 ± 1.21   ×   10 3 27,961.96 ± 1.90   ×   10 3 29,384.89 ± 2.32   ×   10 3
37,00038,271.63 ± 1.57   ×   10 3 36,976.86 ± 0.86   ×   10 3 36,156.01 ± 1.55   ×   10 3 34,802.78 ± 1.63   ×   10 3
42,00041,429.98 ± 1.60   ×   10 3 41,416.01 ± 0.78   ×   10 3 39,830.73 ± 1.25   ×   10 3 38,786.25 ± 1.59   ×   10 3
Table 5. Comparison of prediction accuracy.
Table 5. Comparison of prediction accuracy.
Serial Num1234
R M S E (Root Mean Square Error)2010.912012.712962.473378.40
R 2 (R-square)0.97660.97650.94760.9310
Table 6. Prediction accuracy of using different training data.
Table 6. Prediction accuracy of using different training data.
Method NumMethodFeaturesRMSER2
1Multiple Linear Regression, MLR
(Maximum Likelihood Estimation, MLE)
X / X s 2110.91/1126.330.9766/0.9934
2Multilayer Perceptron, MLP
(Back Propagation, BP)
X / X s 2040.65/1070.120.9771/0.9937
Table 7. Parameter posterior calculation using different training methods.
Table 7. Parameter posterior calculation using different training methods.
Priors
Condition
Diagonal
Gaussian
OursMLE
Posterior x 1 9.97 × 104 ± 5.63 × 10−32.217 × 103 ± 4.12 × 1042.684 × 103
x e n −8.88 × 102 ± 7.25 × 10−65.102 × 103 ± 9.37 × 10−74.932 × 103
x r m −6.22 × 105 ± 2.75−9.476 × 106 ± 2.00−8.997 × 106
x m a −7.15 × 103 ± 9.06 × 10−22.854 × 104 ± 4.67 × 10−42.794 × 104
x e t r −1.02 × 106 ± 4.281.941 × 106 ± 2.221.516 × 106
x s k 5.16 × 102 ± 3.95 × 10−4−6.507 × 103 ± 2.17 × 10−5−6.198 × 103
x p s d −4.08 × 104 ± 5.93 × 10−35.317 × 104 ± 4.43 × 10−44.858 × 104
Estimation R M S E 2795.272014.412110.91
R 2 0.95240.97650.9766

Share and Cite

MDPI and ACS Style

Hou, Y.; Li, X.; Zheng, Y.; Zhou, J.; Tan, J.; Chen, X. A Method for Detecting the Randomness of Barkhausen Noise in a Material Fatigue Test Using Sensitivity and Uncertainty Analysis. Sensors 2020, 20, 5383. https://doi.org/10.3390/s20185383

AMA Style

Hou Y, Li X, Zheng Y, Zhou J, Tan J, Chen X. A Method for Detecting the Randomness of Barkhausen Noise in a Material Fatigue Test Using Sensitivity and Uncertainty Analysis. Sensors. 2020; 20(18):5383. https://doi.org/10.3390/s20185383

Chicago/Turabian Style

Hou, Yuting, Xiang Li, Yang Zheng, Jinjie Zhou, Jidong Tan, and Xiaoping Chen. 2020. "A Method for Detecting the Randomness of Barkhausen Noise in a Material Fatigue Test Using Sensitivity and Uncertainty Analysis" Sensors 20, no. 18: 5383. https://doi.org/10.3390/s20185383

APA Style

Hou, Y., Li, X., Zheng, Y., Zhou, J., Tan, J., & Chen, X. (2020). A Method for Detecting the Randomness of Barkhausen Noise in a Material Fatigue Test Using Sensitivity and Uncertainty Analysis. Sensors, 20(18), 5383. https://doi.org/10.3390/s20185383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop