1. Introduction
Stopping the industrial equipment used on production lines is normally detrimental to companies. Downtime means a loss of production, plus costs due to stopping and restarting the equipment. This underscores the importance of predictive approaches to anticipate behaviours corresponding to potential failures that could trigger these stoppages.
In this context, predictive modelling emerges as a fundamental tool, allowing companies to anticipate equipment failures based on historical and real-time data. By employing advanced data analysis and machine learning techniques, organizations can identify hidden patterns in equipment’s operational data and develop accurate predictive models that alert operators about potential impending failures. This way it is possible to minimize unexpected equipment downtime, improve service quality for customers, and also reduce the additional costs caused by over-maintenance in preventive maintenance policies [
1]. There are a variety of approaches to this, each with its advantages and disadvantages [
2], but in all of them it is essential to follow the best practices for data preparation, model training, and validation to ensure the accuracy and reliability of their predictions. Some approaches stand out for their ability to capture temporal dependencies in sequential data, which is essential for predicting equipment failures over time. Outstanding state-of-the-art models include recurrent neural networks, namely Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRUs) [
3,
4].
Other models are effective in modelling linear and stationary relationships in time series data, such as the Autoregressive Integrated Moving Average (ARIMA) or the Seasonal Autoregressive Integrated Moving Average (SARIMA), although these models are normally more efficient in short-term prediction [
5,
6]. It is also common to use ensemble approaches, which combine multiple models, such as Random Forests (RFs) and Gradient Boosting Machines (GBMs) [
7,
8]. However, each model should be developed according to the specific needs and characteristics of a problem.
The prediction of future sensor values alone is not sufficient to identify trends automatically. Predictive models only forecast the future, but additional tools are needed to classify features and determine whether the patterns that are being anticipated are normal or abnormal. This classification is normally accomplished through classical machine learning models, such as neural networks, Support Vector Machines, and other classifiers [
9].
The present work aims to detect possible failures in an industrial wood chip pump. The methodology consisted of analysing the p-value produced by the Augmented Dickey–Fuller (ADF) test for stationarity. The ADF test establishes that a p-value above a certain level, typically 0.05, indicates a possibly non-stationary time series. In industrial equipment, vibration should be eliminated or contained as much as possible, so that the equipment’s vibration signature will have a profile similar to that recommended by the manufacturer. Over time, however, it normally increases at some point because screws, joints, and fragile parts in general start to become loose. When it reaches a certain value, maintenance interventions are required to fix the problem and reduce vibration again. Hence, the hypothesis is that a trend which may lead to an equipment failure might be detected by the observation of the p-value of a time series in near real time using a rolling window for the test. Specific patterns of increase or decrease or large amplitude variations might reveal a possibly approaching failure in the equipment. These patterns must, therefore, be analysed, aiming to prevent failures from happening and reduce downtime and maintenance costs. To the best of the authors’ knowledge, no previous predictive maintenance study has been reported using the ADF test on chip pump time series. However, this method could be a useful approach for all variables which are stationary in normal conditions, offering great potential for the predictive maintenance of chip pumps and possibly other industrial equipment.
Section 2 describes the state of the art.
Section 3 describes the materials and methods used.
Section 4 shows the results obtained.
Section 5 presents a balance and comparison of the results to the state of the art.
Section 6 draws some conclusions and suggests directions for future work.
2. Literature Review
Fault prediction in industrial equipment is an area of research and practice that encompasses a wide range of approaches and methodologies. Given the complexity of industrial systems and the diversity of factors that can lead to faults, there is no single approach that is universally applicable. Instead, engineers and researchers have explored a variety of techniques and strategies to anticipate, avoid, and mitigate equipment failures [
10].
Statistical models aim at analysing the behaviour of random variables based on recorded data. For predictive maintenance, statistical models are used to determine the current degradation and the expected remaining life of the equipment. This type of model is often implemented in multi-model approaches [
11].
Jie Zhao et al. [
12] presented a novel methodology that uses the Autoregressive Moving Average (ARMA). Instead of using the most common data transformation methods, they developed their own method by creating a transfer function and took the transformed sequence as the input data to build an ARMA model. They achieved a Mean Absolute Percentage Error (MAPE) 2.48%, which is lower than that of the ARMA model with traditional methods.
ARIMA is the most common forecast model used in time series, due to the adaptability of its linear patterns to all time series strategies, according to Chafak Tarmanini et al. [
13]. Based on daily real electricity load data for 709 individual households that were randomly chosen over an 18 month period, they compared the performance of two forecasting methods: ARIMA and an Artificial Neural Network (ANN). After tests, the ANN’s MAPE was 1.80% and ARIMA’s was approximately 2.61%.
In order to prove the good performance, when applied to problems related to anomaly detection and technological parameter forecasting, of the ARIMA algorithm, using sparse data obtained from an urban sewer system, Karthick Thiyagarajan et al. [
14] developed a model and obtained appealing results. Over a time span of 30 days, during which 24 days were used for training, they achieved a Mean Absolute Error (MAE) of approximately 0.0962.
Jing Xu and Yongbo Zhang [
15] developed a model, aiming to study quality predictions in power grid equipment based on key inspection parameters. By gathering live monitoring data from power grid equipment and creating an intelligent predictive alerting algorithm using time series and a trend analysis, they achieved a MAPE of 5.62% for the leakage current and 4.16% for the resistive current.
Luigi De Simone et al. [
16] proposed an LSTM methodology for the predictive maintenance of railway rolling stock equipment. As the dataset contained some imperfections, they developed an algorithm for filtering the spikes found and another one for standardizing them. They also used 10-fold cross-validation and Adaptive Moment Estimation (ADAM) and obtained the following results: a 0.0184 MAE for a 60 min predicted time window, including for low- and high-severity diagnostic events (L&H). For high-severity diagnostic events only (H), they achieved 0.0008 MAE.
Based on supervisory control and data acquisition (SCADA) data, Wisdom Udo and Yar Muhammad developed eight models using extreme gradient boosting (XGBoost) and LSTM to build the characteristic behaviour of critical wind turbine components [
17]. They tested them on two real case studies regarding six different wind turbines with the aim of predicting the gearbox bearing temperature and obtained a MAPE between 0.8% and 6.2%.
Hansika et al. [
18] propose that RNNs have been shown to be promising options for prediction tasks, often surpassing the current state-of-the-art statistical benchmarks in the field. One of the existing types of RNNs is a GRU. Mateus et al. [
19] have employed neural networks to develop two models aimed at predicting the behaviour of an industrial paper press. Their achievement lies in accurately forecasting the press’ behaviour over a 30-day period, achieving a MAPE of less than 10% when utilizing GRU. Mateus et al.also developed a model using LSTM to compare it with a GRU network. Using the same training and test data in both experiments, with changes such as different resample rates, different layer sizes, and different activation functions in the hidden layer, they concluded the existence of some instability in the LSTM model, with its results varying significantly. To examine the most efficient pre-processing approaches to predicting sensory data trends based on GRU neural networks, Mateus et al.developed a model that can anticipate future values with a MAPE of 1.2%. They proved that it is possible to forecast the future behaviour of industrial paper pulp presses up to 30 days in advance.
Martins et al. [
20] also developed a model using a GRU, successfully predicting, within a 7-day timeframe and with a MAPE of 2.79%, the three states classified by a Hidden Markov Model and the observable states obtained from clustering. The methodology also involved the use of a Principal Component Analysis (PCA) and K-Means algorithm during data clustering.
With a variety of models and ways to predict time series values, and, in turn, equipment failures, Peter T. Yamak et al. [
21] conducted a study on the performance of LSTM, GRU, and ARIMA models. They used data from Bitcoin transactions from a Crypto data download website. The data were normalized and forced into stationarity, and their trend and seasonality were removed. All models were tested over a time span of 500 days, yielding the following MAPE results: LSTM, 6.80%; GRU, 3.97%; and ARIMA, 2.76%, with this last being the best and taking the least amount of time.
Although the introduction of stationarity tests into predictive maintenance has not yet been the subject of a more in-depth study, there are already some articles on the topic. Phong B. Dao and Wieslaw J. Staszewski [
22] investigated the application of tests such as the ADF and Kwiatkowski–Phillips–Schmidt–Shin (KPSS) tests to detect damage in structures using Lamb wave data. This study emphasized the challenges imposed by environmental conditions, such as temperature variations, that affect wave responses. The KPSS test was found to be more sensitive in detecting damage, although it is more affected by temperature changes. Their study highlights the need to mitigate these effects for reliable detection. In this type of study, it is also very common to use techniques such as cointegration, which is capable of removing both natural and artificial common trends, and then use peak-to-peak and variance values to detect and classify damages more efficiently [
23].
More recently, Phong B. Dao et al. [
24] also introduced a new method for early fault detection through the analysis of time series stationarity across sliding windows of SCADA data with a high level of confidence. This method demonstrated some superiority over others in detecting unusual or abrupt changes and being able to monitor multiple parameters simultaneously, once again showcasing the simplicity, efficiency, and promise of such methods.
Table 1 shows a comparison of some of the relevant approaches to time series prediction and fault detection, including both classical statistical models and modern approaches based on deep neural networks. As the table shows, predictive models already achieve very reasonable accuracy, with errors less than 10 %. Nonetheless, these predictions still need additional analysis in order to determine a potential failure. No studies were founded based on the analysis of the
p-value to determine ongoing trends, as we aim to in the present study.
3. Materials and Methods
3.1. Methodology
The CRoss Industry Standard Process for Data Mining (CRISP-DM) methodology was followed during this research. The first steps are described in the present section, while the analysis of the results is described in
Section 4. The method’s deployment is left for future work.
The first step in the project was the study of the problem, which consisted of a visit to the company and a literature review. The second step, “Data understanding”, consisted of a statistical analysis of the variables involved, as well as visualization experiments. The third step, “Data preparation”, consisted of data selection and cleaning. After that, seasonal decomposition and a stationarity analysis were performed in the modeling phase. The analysis of the results consisted of the evaluation of the p-value over different time windows.
3.2. Experimental Setup
In the experimental part, the Python programming language was used in a multiplatform open-source integrated development environment (IDE) for scientific programming called Spyder. Additionally, in the development of this project, the following libraries were used: pandas, numpy, matplotlib, pickle, statsmodels, and seaborn. The specifications of the devices used in this study are an Intel(R) Core(TM) i5-10300H CPU @ 2.50GHz 2.50 GHz, with approximately 8.00 GB of RAM.
3.3. Dataset
The dataset used in this project contains sensor data from three industrial wood chip pumps, which are operating in a large Portuguese paper industry. The data were collected by a set of sensors installed in the pumps and from logging readings with 1 min sampling rate. The dataset contains data for the years 2017, 2018, and 2019. The number of variables monitored is 31, but for the simplicity of this analysis only three were used in this project: the vibration of chip pump 1 (VCP 1), vibration of chip pump 2 (VCP 2), and vibration of chip pump 3 (VCP 3). Vibration was chosen because it of its importance for condition monitoring. It is one of the most telling variables and also one of the most difficult to analyse because of the influence of noise. Increasing vibration in industrial equipment is a serious risk, so this is one variable where a stationarity analysis may be more important.
Some statistical results from different years and different chip pumps, taken from the provided dataset, are shown in
Table 2. As the table shows, there are some negative numbers, which obviously correspond to noise and outliers in the data. The table also shows that VCP1 has the lowest mean and the highest standard deviation in 2017, which means that it potentially had problems in this year. VCP2 has the highest mean of the three chip pumps. It also reaches its highest values in 2017 and 2018, which means that it may need maintenance interventions to improve its working condition and prevent potential failures.
Table 2 also shows there are some missing samples, otherwise the AF (absolute frequency) would be 525k for all these variables, considering the sampling period is 1 min. Some of the missing samples correspond to periods of planned maintenance, when the equipment and sensors are turned off. Other missing values, however, correspond to failures in the process of reading or recording the data. When this happens, there are still records in the database; however, they are incomplete and may pose challenges to data analysis.
Table 3 summarizes the number of missing values that may need attention during the data analysis process.
3.4. Data Preparation
In order to make the dataset less extensive, and to reduce noise at the same time, downsampling was performed, changing the period to 5 min instead of the original 1 min—i.e., the mean of each period of 5 min was taken from the original dataset. Thus, the dataset was reduced to one fifth of its original size; approximately 100k rows instead of approximately 500k. This will not affect the information that is needed for the present research, since any high-frequency components possibly lost are not necessary for the stationarity analysis. Maximum and minimum values may be affected, but again they are not relevant for this goal. Additionally, many maximum and minimum values may be outliers caused by reading errors or possible external interference in the sensors, or other extreme values which can be smoothed without affecting the result of this analysis.
Table 4 shows a summary of the statistical parameters of the dataset after downsampling.
After dataset reduction, it is clear that the average values remained largely unchanged, but the maximum values were affected. After further analysis, some periods were identified as candidates for subsequent analysis, given their small percentage of missing values and their temporal extent. The following ranges were selected:
Range 1 (R1): 5 November 2017–20 January 2018;
Range 2 (R2): 4 February 2018–27 October 2018;
Range 3 (R3): 4 November 2018–16 February 2019.
These periods are of interest because of their low number of missing data samples.
Table 5 shows a summary of the statistical parameters of these intervals. As the table shows, there were still some missing points in the second interval, so they were imputed using linear interpolation. Moreover, it is also possible to notice a difference between the values of absolute frequency observed in the second interval due to its longer time span.
As shown in the previous
Table 6, after applying linear interpolation to the second time interval, the values hardly changed.
3.5. ADF Test
As mentioned in
Section 1, the main objective of this study is to understand the variation of the
p-value through stationarity tests, namely using the ADF test. This test shows whether the time series can be considered stationary or not. It consists of testing the null hypothesis, which states that there is a unit root in the time series sample and therefore the series is not stationary. If the null hypothesis cannot be rejected, then it is not possible to reject the existence of a unit root and the series is therefore non-stationary [
26]. The lower the test’s
p-value, the stronger the rejection of the null hypothesis. Conversely, a higher
p-value indicates a greater likelihood of accepting non-stationarity.
The ADF test is applied to the time series model
where
is the time series,
is a constant,
is the coefficient of a time trend,
is the lag order of the autoregressive process,
t is time,
are coefficients, and
is the error.
is the coefficient more directly associated with stationarity. We perform a linear regression of
against
t and
; if
, then the process is a random walk, otherwise the series is non-stationary.
Based on the results obtained from the analyses of the previous time series, it was concluded that the time series corresponding to the interval from 2 April 2018 to 27 October 2018 would be the most suitable for the modelling phase, considering its temporal extent and statistical properties.
4. Results
4.1. Variable Analysis
Through individual analyses of the variables, an attempt was made to understand their behaviour over time.
Figure 1 shows the plots of the three variables chosen for the year 2018. Regarding variable VCP 1, it is possible to observe a growing trend over time, but as shown in
Figure 2, it also presents some discrepant values, which may happen for different reasons. As for variable VCP 2, it exhibits some irregularity and instability and shows seasonal behaviour due to the repetition of patterns throughout the time series, as revealed in
Figure 3. Notably, there are values that are nearly zero and values associated with the mean value. Variable VCP 3, between February and May, demonstrates dynamic behaviour with some randomness in its values. From there, similar to the analysis conducted on the behaviour of variable VCP 2, some seasonal patterns start to emerge. The charts give an idea of the long-term behaviour of these variables, but it is clear that the signals are very noisy and it is difficult to identify clear patterns just from this visual information.
4.2. Seasonal Decomposition
Individual decompositions of the data were performed, focusing on their trend, seasonality, and residual components. Previous analysis of the charts shows that the series do not vary too much over time, but even so they may contain some seasonal components, with additives being more probable than multiplicatives, considering that the average and standard deviation do not suffer large changes over time. Therefore, seasonal decomposition was performed using the additive model. In terms of the period parameter, different values were tried. The value 2016 (
) was used to produce the charts shown in
Figure 2, taking into account the number of intervals present weekly, in order to understand the weekly pattern. As the figure shows, the variable VCP 1 exhibits an absence of significant trends in the time series; there are no clear upward or downward movements. Additionally, it is worth noting the presence of reduced seasonality, as the seasonal component is fluctuating between the values of −0.02 and 0.03.
Figure 3 shows the seasonal decomposition of variable VCP 2. Again, despite the presence of numerous oscillations in the trend, crystal clear long-term growth and declining movements are not visible. The oscillations show there is more noise in this signal than in VCP 1, but the noise does not show clear patterns that could be identified at this point. Despite the noise and small amplitude of most mid-term trend variations, the chart seems to show that there is a growing trend until mid-April, followed by a sudden decline. And then, after this decline, there is a trend of slow growth until late September, which is again followed by a decline, this time with a period of instability. These faint variations over time may be important for further analysis. As for the seasonality, it exhibits similarities to the behaviour observed previously in variable VCP 1, so there are no clear seasonal components of significant amplitude.
Regarding variable VCP 3, its seasonal decomposition is shown in
Figure 4. The charts show the peculiar characteristics of this signal, which is different from VCP 1 and 2. Again, the seasonal component is negligible. The trend is variable in the first months and more stable after April. Until May, a quite pronounced but unstable trend is observed, which denotes problems in the equipment’s normal operation. At the end of April, something changed and the trend decreased suddenly, before slowly increasing in subsequent months.
4.3. Stationarity Analysis
The ADF test was employed in order to analyse stationarity of the signals based on the Akaike Information Criterion.
Table 7 shows the
p-values, for all of the period under analysis (R2), of the three pumps. As the table shows, all the ADF statistic test values are strongly negative, indicating strong evidence against the null hypothesis of non-stationarity. In other words, the time series are proven to be stationary in this period, based on the visual evidence seen when looking at
Figure 1,
Figure 2,
Figure 3 and
Figure 4 and also the information that comes from the factory: the pumps receive periodic maintenance interventions so that their vibrations are kept at safe levels over time.
Nonetheless, when the analysis is performed in smaller time windows, stationarity may not be guaranteed, according to the initial research question of the present research. The most important question, however, is what window sizes will work to show the relevant non-stationarity of the time series, if there is any. Considering thsi project’s objectives, we decided to experiment with two distinct sizes of rolling windows: weekly and daily. This segmentation aims to obtain information that could lead to a more holistic approach.
The weekly variation of the
p-value throughout the time series is illustrated in
Figure 5. The figure shows that the
p-value varies from approximately zero to values above 0.75 in VCP 2 and 0.6 in VCP 1 and VCP 3. This indicates that there are weeks when the vibration follows a non-stationary path, which is the pattern being sought after in the present research.
An example of the daily variation of the
p-value throughout the time series is illustrated in
Figure 6. As this figure shows, daily variations exhibit constant fluctuations, which make any analysis or monitoring very difficult to perform. The weekly signals produce more stable results, but even so there is certainly a lot of noise, and a more stable signal is required for more confidence in the results.
4.4. Impact of Outliers on the p-Value
Discrepant values are often called outliers. They are normal in industrial sensor readings and they have an impact on the data analysis process, as they affect data averages, standard deviation, and many other statistical characteristics of a series.
The relationship between the discrepant data and the p-value was also sought. We are particularly interested in outliers because, in the absence of labelled data about the failures’ true timestamps, outliers are possible indicators of failures. They may be sources of true or false positives.
In this study, a discrepant value was defined as one 200% higher than the mean recorded for each analysed variable. In our data, 5 days of discrepant values were identified.
Figure 7 shows the discrepant values, which are marked with bounding boxes. They correspond to the following dates:
Performing a more specific analysis of the
p-value values on the different dates identified, it is possible to observe some variations in the
p-value during or around the week that includes the date of the discrepant values. As an example,
Figure 8 shows the weekly and daily
p-values for VCP 3, where a spike in vibration was observed on 6 March. As the figure shows, there is an increase in the
p-value in the first week of March, followed by a steady decrease until the end of the month. The daily chart shows the highest
p-values were registered on days 12, 16, and 30—in other words, after the discrepant value was registered. These values are compared in
Figure 6, which shows the variation of the
p-value and the mean of the variable VCP 3 corresponding to the identified day and which highlights some peaks of the
p-value over four months, including March.
The sources of these spikes may be numerous, and during the present study it was not possible to explain them. Hence, the conclusion from this observation is that outliers have an influence on the p-value, but whether they are true or false positives, regarding fault diagnosis, is left for future work, where more data may be available.
4.5. Minimizing Noise
In order to minimize the impact of discrepant data on the variation of the
p-value, two noise filters were applied. They consisted of applying a rolling average, with sliding windows of 1, 3, 7, or 15 days, followed by the application of a LOWESS filter with
frac = 0.15. The impact of those filters can be observed in
Figure 9,
Figure 10 and
Figure 11.
Figure 9 and
Figure 10 show a portion of the VCP1 signal, smoothed with rolling averages of 1- and 3-day windows, respectively, and then a LOWESS filter with a fraction of 15%. Analysing
Figure 9 and
Figure 10, it is possible to observe some stability in the mean value of VCP1 throughout most of the time window. However, towards the end of the window, there is a sudden increase. Looking at the variation of the
p-value, it is possible to see that the same pattern occurs. Additionally, in
Figure 9, it is possible to observe an increase in the
p-value immediately after the mean of the variable rises. This shows one example of the pattern being sought, as the series becomes non-stationary because of an increasing rolling average due to an increased vibration amplitude.
In
Figure 10, it is also possible to see that before the final increase in the VCP1 mean, the
p-value reacts twice, rising above the threshold value of 0.05 twice in a short time span. This is the same signal as in
Figure 9. The figures show that the width of the rolling window can have a significant impact in the results and possible alarms generated, in cases where the
p-value is used to generate an alarm when it rises above a certain threshold.
This same
p-value behaviour can also be observed in
Figure 11, which shows the vibration of VCP3 with a rolling average of 15 days. In this case, before a negative peak in the mean value of the variable is observed, it is possible to see a significant increase in the
p-value followed by a sudden drop. The
p-value reacted to a spike in vibration that lasted several days, and at the end of May it also rose above 0.05, reacting to a consistent increase in vibration levels.
In summary, the pictures show the following:
The impact of short spikes on the p-value when no noise-filtering methods are used. The p-value responds to the spikes and can generate true or false positives, with more data being necessary to determine which is the case.
When applying noise-filtering methods such as a rolling average and LOWESS filter, the impact of short spikes is minimized. The size of the rolling window has a direct impact on the number of times the p-value crosses the significance threshold defined.
5. Discussion
The p-value of the ADF test is one indicator of the stationarity of a time series. In the case of industrial equipment, some variables, such as vibration, would be stationary in ideal conditions. However, what normally happens is that vibration increases with increased workload or between maintenance interventions. The present research shows that analysis of the p-value may be a promising method to detect trends, or even outliers, in the vibration signals of three industrial pumps. The ADF test is conducted using a rolling window, thus avoiding direct repetition of the test on the same data points, reducing the likelihood of false positive inflation. Balancing false positives and false negatives is always challenging and the cost analysis of further domain-specific errors must be conducted before this approach is used in the industry.
To the best of the authors’ knowledge, this method has never been tried before in wood chip pumps. Nonetheless, the analysis of the p-value could have several advantages compared to other methods of detecting and predicting possible faults:
One advantage is that it is known that a p-value above 0.05 indicates the non-stationarity of the series. This threshold is easier to establish than fixing limits for each variable. So this could be a kind of universal approach to detect the non-stationarity of a variable in a given time window.
Another advantage is that the ADF test is able to work with noisy data, simplifying the analysis process. The presence of noise in the data has some impact on the results but does not cause a catastrophic failure of the method.
On the other hand, the method also has some drawbacks, as demonstrated in the results of the experimental work. One important drawback is that it is necessary to determine the best time windows, both for calculating the
p-value and for noise filters, if they are applied, because different window sizes will produce different results and different numbers of possible alarms. Also, even though the
p-value method may be universal for different types of equipment, the noise and quality of the data still require careful analysis and specific cleaning or treatment methods. An analysis of possible cointegration issues may also be necessary for ruling out the influence of external factors [
27].
Table 8 shows the advantages and disadvantages of the proposed method, making its interpretation more straightforward.
6. Conclusions
Fault detection and prediction in industrial equipment is an important area that is subject to heavy research. To the best of the authors’ knowledge, the analysis of the p-value from an ADF test of a time series is a novel approach, which has never been applied to industrial chip pump vibration analyses. It offers immense potential for monitoring and determining trends in the state of an equipment. The hypothesis was further investigated through the analysis of the p-value of three time series, using rolling windows of different sizes to calculate the p-value and also to apply noise-filtering techniques to stabilize the results obtained.
The results showed a stable increase in their p-value, which was increasingly above the threshold of 0.05, when there were large variations in the vibration of the chip pumps. Equipment prognostic and fault prediction using the p-value is an appealing idea, because it can be a universal method that is simpler than defining limits for each variable, and the present study showed that vibration, although a stationary time series in the long term, is non-stationary during some periods.
Future work requires the analysis of additional data, namely to study in deeper detail the behaviour of the p-value in situations of actual faults, both for vibration and other variables.