Next Article in Journal
Gear Teeth Deflection Model for Spur Gears: Proposal of a 3D Nonlinear and Non-Hertzian Approach
Previous Article in Journal
PSIC-Net: Pixel-Wise Segmentation and Image-Wise Classification Network for Surface Defects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vibrodiagnostics Faults Classification for the Safety Enhancement of Industrial Machinery

1
Faculty of Mechanical Engineering, Brno University of Technology, 616 69 Brno, Czech Republic
2
TOSHULIN, a.s., Wolkerova 845, 768 24 Hulin, Czech Republic
*
Author to whom correspondence should be addressed.
Machines 2021, 9(10), 222; https://doi.org/10.3390/machines9100222
Submission received: 3 September 2021 / Revised: 22 September 2021 / Accepted: 23 September 2021 / Published: 30 September 2021

Abstract

:
The current digitization of industrial processes is leading to the development of smart machines and smart applications in the field of engineering technologies. The basis is an advanced sensor system that monitors selected characteristic values of the machine. The obtained data need to be further analysed, correctly interpreted, and visualized by the machine operator. Thus the machine operator can gain a sixth sense for keeping the machine and the production process in a suitable condition. This has a positive effect on reducing the stress load on the operator in the production of expensive components and in monitoring the safe condition of the machine. The key element here is the use of a suitable classification model for data evaluation of the monitored machine parameters. The article deals with the comparison of the success rate of classification models from the MATLAB Classification Learner App. Classification models will compare data from the frequency and time domain, the data source is the same. Both data samples are from real measurements on the CNC vertical machining center (CNC-Computer Numerical Control). Three basic states representing machine tool damage are recognized. The data are then processed and reduced for the use of the MATLAB Classification Learner app, which creates a model for recognizing faults. The article aims to compare the success rate of classification models when the data source is a dataset in time or frequency domain and combination.

1. Introduction

The current development of Industry 4.0 brings, in addition to the digitization of production, the collection of large amounts of data, which are suitable for the use of artificial intelligence methods. Particularly in the field of technical diagnostics, this opens possibilities for the early detection state of an object (machine) based on the measured data, which are obtained from different standard diagnostic methods (e.g., vibrodiagnostics, thermodiagnostics or electrodiagnostics) and it is possible to use special methods of technical diagnostics such as measuring the magnetic field of the machine [1]. It is therefore a multi-parameter diagnostic that can more accurately detect, identify, and localize the emerging fault, thanks to a combination of input data from different areas of diagnostics.
Vibrations are an important carrier of information about the condition of rotating equipment and this fact is commonly used in the field of vibrodiagnostics. This article focuses on the machining center and here it is possible to use the vibrodiagnostic system to detect other information such as machining quality or tool damage. Thus, unexpected states can be detected by a vibrodiagnostic signal, which may indicate a fault or a safety risk. However, these signals are unexpected and a simple evaluation mechanism cannot be designed, for example, with the detection of frequent failures of rotary machines such as misalignment, unbalance, or bearing damage. In these cases, it is appropriate to use self-learning artificial intelligence methods, i.e., they can learn to recognize a fault condition from historical data for a specific machine or specific machining operations.
In the development process of smart machine tools capable of operating within the framework of industry 4.0, it is necessary to implement risk analysis already during the machine design stage not only in terms of potential damage to the health of the machine operator (according to ISO 12100 [2]), but also in terms of quality management and monitoring of the capability of the manufacturing process on the machine tool (according to ISO 9001 [3]). We can apply the good technical practice as described in IEC 31010 [4] and then implement the necessary diagnostic tools into the machine design or modify this machine design so that the measurement uncertainty of the sensor system (according to ISO 3534-1 [5]) would not affect the interpretation of the measured data and the subsequent decision-making processes in a negative way. A typical example could be early detection of tool damage and subsequent bringing the machine to a safe state. Breaking and ejecting the carbide insert from a tool holder can cause damage to the transparent guard of the machine’s working area, and thus lead to a subsequent loss of its safety function, or it may cause irreversible damage to the machined surface. Therefore, by early detection of a machine fault, we can protect the health of the operator as well as reduce the financial losses caused by poor production or increased wear of the smart machine’s spindle.
This article will, therefore, compare classification methods that can detect a specific fault from real data, namely a missing insert on a machining tool, the absence of which could have occurred due to tool damage. A method will be used where predictors for machine learning will be evaluated in advance and their choice is therefore up to the expert. The correct choice of predictors has a major impact on the classification success rate. The authors of the work have already dealt with the issue of evaluating vibrodiagnostic signals in [6,7,8]. The idea of using artificial intelligence methods in diagnostics is described in other publications such as in [9,10,11,12], where artificial intelligence methods are used for the recognition of gearbox faults. Also, in the paper [13] authors focus on the classification using neural networks in the planetary gear. The article [14] uses artificial intelligence methods to recognize the shape of a rotary machine shaft’s orbit; it is a picture classification according to the shape of shaft orbit. The aim of this research is to determine a fast method (real-time evaluation) for the timely evaluation of a failure that would reduce the safety of the machine. Thus, the primary goal is not to diagnose the condition of the machine, such as a bearing or gearbox failure, and to predict residual life, but to detect failures related to machine safety and to warn the operator immediately.
This article has several goals and they are:
  • Verify that the missing insert on the tool is detectable on the vibrodiagnostic signal measured on the machine spindle.
  • Select and compare predictors from time and frequency domains.
  • Effect spindle speeds on the classification success rate.
  • Select and compare the most successful classification methods
  • Evaluate detection success rate of fault-free state
It is assumed that the process of data collection and evaluation takes place on an IPC (Industrial PC) computer near the machine (for example IPC SIMATIC PC-based, where one IPC will be prepared for each CNC machine), and the learning and model generated process will take place on a server containing commercial software such as MATLAB (server will be used for many CNC machines). Learning and exporting the model is an important operation and will take place under strict supervision. The created MATLAB model will be exported for the use of individual IPCs without MATLAB.
The Python environment also includes Machine Learning tools such as scikit [15] or Artap used in [16], however in this case the ambitions are to be used in a safety environment and the authors there rely on the professional Matlab environment. In the future, the above-mentioned toolboxes will be used and tested to evaluate the condition of the machine.
Python environment is used for signal processing, the MATLAB and Classification Learner App [17] are used for data classification.

2. Experimental Procedure

The data were obtained from the machine MCV 754 QUICK (Figure 1), it is a three-axis CNC (Computer Numerical Control) machining centre, more information can be obtained on the manufacturer’s website [18]. The obtained data are from the measuring instrument Microlog CMXA48 and the measurement parameters are in the Table 1.
The instrument Microlog uses a CMSS 2111 accelerometer sensor that has been attached to the spindle with a magnet. The position of the sensor on the machine is shown in Figure 2. Used tool H490 F90AX D040-4-16-12 has the possibility of changing the teeth (inserts), which are fixed with a screw, as shown in Figure 3. Three states/classes were chosen for the experiment, namely:
  • Class 0—Tool in the fault-free state (Figure 3-left)
  • Class 1—Tool without insert (carbide insert), but with screw-simulation of damage and breakage of an insert (Figure 3-middle)
  • Class 2—Tool without insert and screw (Figure 3-right)
Expected measurable changes are caused by the unbalance of the tool when this vibration is transmitted to the spindle. The failure of unbalance is characterized by an increase in vibrations in the region of the first harmonic frequency. The unbalance is described in the [19,20,21] and is divided into a static unbalance and a dynamic unbalance.
The first experiment verified that the missing insert was reflected in spindle vibrations at 700, 1050, and 1400 RPM. The results are shown in Figure 4, which shows the area around the first harmonic frequency. The aim of this experiment was to determine at which spindle speeds the manifestations are visible in the frequency spectrum and which speeds are suitable for data collection in a machine learning dataset. From the result, it is evident that at the speed 700 RPM the effect of the unbalances in the frequency domain is insignificant, and therefore undetectable. For this reason, speeds were selected to obtain a dataset from 1000 to 1500 RPM.

2.1. Dataset Obtain

For each state/class, a series of data was measured for different operating speeds that are within the recommended speed range for the tool. Specifically, the speed was from 1000 RPM to 1500 RPM after the step of 50 RPM in two repetitions and 20 s is maintained at each speed level. The total time is therefore 20 × 11 × 2 = approx. 440 s for one state (class). The speed setting was programmed and started for each state in the same way, Figure 5 shows the time record of the speed change over time. The aim is to obtain data at a wide range of operating speeds. The data were captured in a recorder mode, i.e., one long time record was obtained, which was later processed into individual samples, which formed a dataset for machine learning.

2.2. Predictors

Predictors are arbitrary parameters, and their values (or combinations thereof) determine the properties of the class being searched. If these parameters did not adequately reflect the properties of the class, the data classification would not be successful. An example of the use of a predictor is the known area for image data processing. The input is information about individual pixels and at full HD resolution ( 1920 × 1080 ) it is more than 2 × 10 6 pixels and each pixel contains information about three colours (RGB). This amount of data is disproportionately large and contains a lot of useless information, so there is a need to reduce data such as reducing the number of colours, edge detection, and more. The outputs of this reduction are predictors, which are the input for artificial intelligence, and based on their values (and their combinations) it is possible to classify individual classes. The properties of predictors should be:
  • Clearly describe the class properties (state)-improved achievement classification
  • Low computational complexity-speeds up the predictor calculation
  • Low number of predictors-speeds up the learning and evaluation process
The input data in vibrodiagnostics are time domain signals. In this particular case, the sampling frequency is 2560 Hz (sampling period is approximately 0.4 ms), which means 2560 values every second of recording for one axis. It is, therefore, necessary to reduce this data flow and evaluate suitable predictors from it, which will be the input for artificial intelligence. The aim is to select appropriate predictors and compare their impact on the success of data classification.

2.3. Data Processing and Selection of Predictors

In this article, a method is chosen that classifies data into individual classes based on predictors. The selection and calculation of predictors are performed before the classification of data and the correct choice of these predictors has a significant impact on the results of the classification and it is necessary to know the issue and have experience with similar experiments. The author of the article has already dealt with this issue in [7,8] and verified which predictors are applicable for a similar type of problem. This article will compare the classification successes for time and frequency domain predictors and their combinations. It was also necessary to choose the length of the time signal and divide the resulting samples. The times of 0.25 s and 0.5 s were chosen and again the advantages and disadvantages of these lengths were compared. The following predictors were used.
Predictors from the time domain:
  • Speed-Current RPM during one sample
  • RMS-Root Mean Square of acceleration amplitudes from one sample
  • STD-The standard deviation of the acceleration amplitudes from a single sample
  • PCA-Principal Component Analysis realized with scikit learn library R2 (with parameters n_components = 1 and explained_variance_ = 0) [22]
The result is a vector that contains 4 predictors for each sample (RPM, RMS, STD and PCA); the values in this vector are further labelled as “T predictors”.
Frequency domain predictors are based on FFT (Fast Fourier Transform) frequency analysis performed using the NumPy library “Standard FFTs” [23] and a range of 4 to 100 Hz is used. The minimum value of 4 Hz is determined by the minimum measuring range of the sensor used. The maximum value of 100 Hz is the maximum value of the occurrence of the first harmonic in all operating speeds of the tool and any tooth frequency. The maximum operating speed is 1432 RPM, the maximum tooth frequency is therefore (1432 × 4)/60 = 95.47 Hz, therefore the maximum frequency of 100 Hz was chosen.
  • Speed-Current RPM during one sample
  • Spectrum energy (eF)-calculated as A f ( i ) 2 for the 4–100 Hz range, where A f ( i ) are amplitudes of spectral lines.
  • Af(i)-the sequence of amplitudes of individual spectral lines for frequencies 4–100 Hz
The result is a vector whose size varies according to the length of the signals used, i.e., division into individual samples. For example, if the signal is 0.25 s long and has a sampling frequency of 2560 Hz, the situation is as follows:
N = f s · t = 2560 · 0.25 = 640
B i n ( Δ f ) = f s N = 2560 640 = 4
For range 4–100 Hz:
N B i n = f m a x f m i n B i n = 100 4 4 = 24
For a 0.25 s signal (t1), the output will be a vector containing 26 (24 + 1 + 1) predictors for each data sample. The disadvantage of such a short signal is the frequency resolution ( B i n ) of 4 Hz, where this value may be insufficient. Furthermore, signals with a length of 0.5 s will be compared. There is the following situation
N = f s · t = 2560 · 0.5 = 1280
B i n ( Δ f ) = f s N = 2560 1280 = 2
For range 4–100 Hz:
N B i n = f m a x f m i n B i n = 100 4 2 = 48
For a 0.5 s signal (t2), the output will be a vector containing 50 (48 + 1 + 1) predictors for each data sample. With this signal distribution, we get a frequency resolution Δ f ( B i n ) of 2 Hz, which is more suitable than the previous case, but we get only half of the samples for machine learning from the limited recording.
The resulting vector for frequency domain predictors contains the measured RPM value, the calculated Spectrum energy value (labeled as eF) and the sequence of amplitudes of individual spectral lines (Af1 to Af24 for t1 and Af1 to Af48 for t2), calculated using the FFT function, see Figure 6. The values in this vector are further labelled as “F predictors”.
Both variants will be compared with each other.
The detailed data processing procedure is as follows:
Data processing (outside MATLAB)
  • Combination of vibration channel and tacho channel data
  • Splitting the signal according to the selected time (t1 = 0.25 s or t2 = 0.5 s)
  • Checking whether the change is not greater than 10 RPM during the duration (t1 or t2) (elimination of transients)
  • Sorting the signal into individual classes
  • Predictor calculation, for each sample (T-time domain, F-Frequency domain or FT-all predictors)
  • Saving predictors to a file according to individual classes (preparation for MATLAB)
Processed in MATLAB:
  • Random mixing and merging of data of individual classes into one dataset
  • Evaluating in the environment “Classification learner app”
  • Generating and saving the most successful model
Verification of the generated model:
  • Randomly selected n samples from each class (selected n = 100 per class)
  • Accuracy and TPR (True positive rate) and FNR (False negative rate) evaluated
  • Results displayed and saved as the Confusion matrix

3. Classification Methods

Classification Learner App (CL) compared different classification methods to determine which methods are appropriate for the problem. This application allows you to compare different methods and then select and create a model containing the selected method (chosen were the most successful). The individual methods are described in more detail in [24]. Datasets were tested on all methods that CL allows and their success is in a Table 2, Table 3 and Table 4. The number of repetitions of the learning process was 10 for each model, and the interval estimate was calculated using “The t Confidence interval” [25] with alpha = 0.05. Furthermore, the hyperparametric optimization provided by Matlab was verified. Each winning model was subjected to hyperparametric optimization with the same dataset, and these results are shown in Table 5. Due to the time consuming, only one repetition was performed, the average improvement of the prediction accuracy with hyperparametric optimization is only 0.55%. Models without hyperparametric optimization were used for further validation.
Briefly, these are the following classification methods:

3.1. Classification Trees

Binary decision trees for multiclass learning [26]. This method is easy to interpret and fast for data prediction, has low memory requirements, but may not achieve sufficient prediction accuracy. Decision making takes place through a tree structure from the beginning (root) to the final class (leaves). The method was probably first published by J. Ross Quinlan in 1975. The tested variants are:
  • Fine tree
  • Medium tree
  • Coarse tree

3.2. Discriminant Analysis

Regularized linear and quadratic discriminant analysis [27]. This method is easy to interpret and fast for data prediction for large datasets. Using “Quadratic Discriminant” increases memory requirements. It is one of the methods of N-dimensional statistical analysis (multivariate statistical analysis), where, based on the decision rule, objects are divided into groups according to probability densities. The method was probably first published by Ronald Fisher in 1936. The tested variants are:
  • Linear discriminant
  • Quadratic discriminant

3.3. Naive Bayes

Naive Bayes model with Gaussian, multinomial, or kernel predictors [28]. This method is easy to interpret and achieves good results for multi-class classification. Prediction speed is slow (medium for Gaussian), memory usage is a medium (small for Gaussian). The method is based on Bayes’ theorem on conditional probabilities and class affiliation testing. Bayes’ Theorem was published by Thomas Bayes in 1736 and the classifier was probably first published in 1960. The tested variants are:
  • Gaussian Naive Bayes
  • Kernel Naive Bayes

3.4. Support Vector Machine Classification

Support vector machines (SVM) for binary or multiclass classification [29].
This method is difficult to interpret (not for the linear variant). If there are more than two classification classes, then the problem must be divided into several partial binary problems, therefore, it has high memory requirements and is therefore slow in multi-class classification. The principle of the method is to find a hyperplane that separates individual objects based on class affiliation. The method was probably first published by Vladimir N. Vapnik and Alexey Ya. Chervonenkis in 1963. The tested variants are:
  • Linear SVM
  • Quadratic SVM
  • Cubic SVM
  • Fine Gaussian SVM
  • Medium Gaussian SVM
  • Coarse Gaussian SVM

3.5. Nearest Neighbors

k nearest neighbours (KNN) classification using Kd-tree search [30]. This method is difficult to interpret and the success of the method decreases with the growth of classes. Prediction speed and memory requirements are media. The principle of the method consists of finding the nearest neighbours of objects in N-dimensional space with the properties of the class. The method was probably first published by Evelyn Fix and Joseph Hodges in 1951. The tested variants are:
  • Fine KNN
  • Medium KNN
  • Coarse KNN
  • Cosine KNN
  • Cubic KNN
  • Weighted KNN

3.6. Classification Ensembles

Boosting, random forest, bagging, random subspace, and ECOC (Error-Correcting Output Codes) ensembles for multiclass learning [31]. This method combines several simple methods into one powerful model. Interpretability is difficult, however, the prediction speed is from fast to medium (depending on the combination used) and memory usage is low (medium for Subspace KNN). The tested variants are:
  • Boosted Trees
  • Bagged Trees
  • Subspace Discriminant
  • Subspace KNN
  • RUSBoosted Trees

4. Comparison of Classification Results

The experiments were divided into several variants, according to the properties of the input data, to be able to compare results of the classification and determine which parameters/predictors and under what conditions they affect the success of the classification. The first step is to use a sample length of 0.25 s and 0.5 s (t1 and t2). The choice of this parameter affects the number of predictors-(frequency spectrum resolution) and also the number of samples for the dataset because the measured data have a limited length (for more details, see Section 2 in Figure 5). Another division will focus on determining the effect of spindle speed variability. Datasets labeled R1-R3 have a constant speed (1000, 1250, 1500 RPM) and experiments labelled R4 have a dataset with combinations of all speeds, so it is a random mixing of sample R1 to R3 regardless of speed. And the last experiments contain a combination of all predictors, ie. from the time and frequency domain (without T or F marking). A unique code name was chosen for each experiment so that it was possible to refer to a specific variant in the text. The overall division of the experiment is shown in Figure 7 and Figure 8.

5. Verification of Resulting Models

The following section compares the success of each method for all variants of experiments. The goal was to find out what methods are suitable for the given predictors, and how the success of the classification differs with the spindle speed, and how successful the time and frequency domain predictors are.
The accuracy of the models is compared and is determined as follows. In the case of selecting the most successful model, an evaluation method was used, which the Matlab Learner app calls “Accuracy” (in the generated code as “validationAccuracy”)
A c c u r a c y = A l l T r u e P r e d i c t i o n A l l P r e d i c t i o n · 100
where:
A c c u r a c y —success rate of classification (ideal value is 100%)
A l l T r u e P r e d i c t i o n —True prediction from all class
A l l P r e d i c t i o n —All (true and false) prediction from all class
To verify the model (Figures 10 and 11) the method used by Matlab was used, namely “True positive rate”-TPR, where each class was evaluated separately; the remaining fields on the lines in CM (Confusion matrix) are False negative rate-FNR for the class. The line amount is therefore 100%.
T P R = T r u e C l a s s P r e d i c t i o n A l l C l a s s P r e d i c t i o n · 100
where:
T P R —True positiv rate (ideal value is 100%)
T r u e C l a s s P r e d i c t i o n —True (correct) prediction per class
A l l C l a s s P r e d i c t i o n —All (true + false) prediction per class
F N R = T r u e C l a s s P r e d i c t i o n P r e d i c t e d C l a s s · 100
where:
F N R -False positiv rate (ideal value is 0%)
P r e d i c t e d C l a s s —Predicted class (incorrect) for True class
A l l C l a s s P r e d i c t i o n —All (true + false) prediction per class
In Table 2, Table 3 and Table 4 are the results of individual methods and it is clear that some methods are more successful for predictors from the time domain than predictors from the frequency domain. More interesting is the graphic display in Figure 9, where the success rate of the classification depends on the spindle speed. Descriptions in the graph mean “Classification success rate” is the Accuracy ( A l l T r u e P r e d i c t i o n / A l l P r e d i c t i o n ) for the created models at the number of a random selection of 100 samples per class and “Fault-free detection success rate” is the TPR (True positive rate) for class 0, i.e., faultless condition.
In the case of predictors from the frequency domain, the success rate increases with spindle speed, because the effect of the unbalances fault is more noticeable with increasing spindle speed, but in the case of the predictor from the time domain it the course is the opposite, see Figure 9.
The next step was to select the most successful models and subject them to further verification. From the confusing matrix in Figure 10 and Figure 11 is clear that the models had a problem recognizing classes 1 and 2, it is a variant without an insert with a screw and a variant without an insert and a screw, i.e., a defect was determined, but incorrectly classified. However, the classification of the fault-free state is more important, it determines the success with which the model classifies the fault-free state-class 0, this information is the most important and is graphically shown in Figure 9-bottom. The experiment, which combined all predictors and contained datasets from all speeds, achieved a success rate of the classification of the fault-free state (classification of class 0-marked “TF mix”) equally 97% for the dataset for lengths of 0.25 s (t1) and 98% for the dataset for lengths of 0.5 s (t2). Experiment with speed 1500 RPM and combined predictors achieved a success rate of 100% for both datasets for lengths of 0.25 s (t1) and 0.5 s (t2).

Time Consumption of Prediction

The time consumption of the prediction is highly dependent on the performance of the computing unit that will perform the prediction. In general, the time-consuming process is the learning of the model, but this operation is only performed once when the systems are deployed or when a new dataset is available for learning. Thus, the learning process can be efficiently scheduled and is not critical to the running of the system.
More important is the prediction time and the possible reaction of the system to a possible problem. The prediction time consists of two steps, namely data preprocessing, i.e., extraction of predictors, and then the prediction itself. These times were tested on a personal computer with Intel(R) Core(TM) i7-2600 CPU @ 3.40 GHz, 16 GB RAM and a similar configuration is possible for the current IPC. Time was measured for 100 repetitions each time to eliminate random variations in performance. Measured time includes only computation time without loading or saving data from recording media. Winning models for the most challenging variants containing all predictors (time and frequency domain) were used. For experiment T1R4, the winning model was “Bagged Trees” and the results are as follows. The data processing time for 100 repetitions is 1.13 s and the prediction time for 100 repetitions is 5.12 s, that is, the average total time is (1.13 + 5.12)/100 = 0.0625 s i.e., 62.5 ms. For T2R4 experiment, the winning model was “Cubic SVM” and the results are as follows. The data processing time for 100 repetitions is 2.04 s and the prediction time for 100 repetitions is 1.43 s, that is, the average total time is (2.04 + 1.43)/100 = 0.0347 s i.e., 34.7 ms. The measured times are only approximate and depend on many circumstances, also there is a possibility for optimizing the classifier selection according to its prediction time. For our purposes, the achieved times are sufficient.

6. Implementation of the Trained Model

The procedure for implementing a trained model is as follows. The model can be generated as a function of MATLAB, or as source code in C language, then it is easy to compile it for any platform. Its function (ability to classify) does not change over time because there is no learning process. However, provided the conditions have changed or a new dataset is available, it is possible to start the learning process again and train the model for the new data.
Figure 12 schematically shows the possible connection of the server with individual clients. Each IPC performs data collection and especially timely evaluation based on a trained model that was learned in the MATLAB environment on the server and then exported. The process is as follows:
(a)
Data acquisition and learning process
  • The server requests data from individual IPCs and creates a dataset for creating a classification model (Data is predictors-it evaluates IPCs, only predictors are transmitted).
  • If there is a sufficient amount of data (or conditions have changed, etc.), then MATLAB performs learning and creates a trained classification model for a specific IPC and exports it to the C language (this MATLAB allows).
  • The trained model is sent to a specific IPC.
Individual machines and their IPCs can be distributed anywhere in different places and there will be only one server with MATLAB, which will generate a model for a specific machine and send it to a specific IPC. If necessary, the MATLAB SW can be replaced by another system, however, the functionality will be maintained and the change will be made on only one server.
(b)
Evaluation process
  • IPC performs data collection and predictor calculation-this is a simple operation and can be performed in any programming language, for simplicity and availability Python was chosen, which performs statistical operations and frequency analysis.
  • Performs classification using MATLAB created and trained model, which was exported to C language (Python allows to run code in C language).
Such a topology and partitioning allows for easy changes on both the IPC and server-side, including changing the programming environment.
As for the specific deployment for monitoring the machine, it is important to note that nowadays it is not easy to implement elements of artificial intelligence in the safety segment, where it is necessary to certify the software and verify its functions. However, it is possible to deploy a similar system only as an informative element that would alert the operator, but not interfere with the operation of the machine. The use of artificial intelligence methods is particularly advantageous in the field of technical diagnostics for the evaluation of multiple criteria and operator alerts.

7. Results and Discussions

Figure 9 summarizes the results in graphical form and expresses the dependence of the predictors on the spindle speed. For predictors from the frequency domain, an increase in success with increasing speed is evident due to a more pronounced manifestation of an unbalance fault in the frequency spectrum (increasing the first harmonic). In contrast, time domain predictors achieve a higher success rate at lower speeds than frequency domain predictors. Thus, using all predictors (time and frequency) increases the success rate at lower speeds and does not decrease at higher speeds. The figure also focuses on the detection of a fault-free state, i.e., assuming that we modify the classes only to the fault-free state (class0) and any fault state (class 1 or 2), then it is obvious that the success of fault-free detection will increase because we do not include misclassification between class 1 and class 2, which are similar. This is important if we only want to detect a fault-free condition and in any other case, an operator warning is issued.
Important results are shown in the Table 2, Table 3 and Table 4, which compare the successes of individual classification methods. Furthermore, Figure 10 and Figure 11 compare the successes of the classification of individual classes and the most important is the comparison in Figure 9, which shows the success of the classification and the success of fault-free state detection, wherewith variant R4 (TF mix-the combination of all speeds), 97% for t1 and 98% for t2 success rate is achieved.

8. Conclusions

The article aimed to verify the possibilities of artificial intelligence for the classification of machine tool failure based on an experiment from a real machine. Early detection of the disorder is an important factor in protecting the health and eliminating damage. The use of artificial intelligence methods brings the possibility of creating universal algorithms, which, thanks to the ability to learn, adapt to a specific machine, and a specific phase of machining.
The described method approaches the possibility of using artificial intelligence for data classification, in this case for the classification of various machine states. A similar method can be used to classify any classes, where only a change of predictors that describe the problem sought is needed.
The choice of predictors used is an important factor which, together with the operating conditions, influences the success of the classification. Future experiments will focus on the possibility of detecting a fault condition of a tool during machining and will focus on obtaining a larger dataset with different machining conditions.

Author Contributions

Resources, Software, Writing—original draft, D.Z.; Writing—original draft, P.B.; Validation, Visualization, T.M.; Resources, J.T.; Writing—review and editing, R.H.; Validation, K.M.; Resources, V.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Education, Youth and Sports of the Czech Republic and The European Union-European Structural and Investments Funds in the frame of Operational Programme Research Development and Education-Project: Machine Tools and Precision Engineering (Project No. CZ.02.1.01/0.0/0.0/16_026/0008404).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Oravec, M.; Pacaiova, H.; Izarikova, G.; Hovanec, M. Magnetic Field Image-Source of Information for Action Causality. In Proceedings of the 2019 IEEE 17th World Symposium on Applied Machine Intelligence and Informatics (SAMI), Herl’any, Slovakia, 24–26 January 2019; pp. 101–105, ISBN 978-1-7281-0250-4. [Google Scholar]
  2. ISO 12100:2010. Safety of Machinery—General Principles for Design—Risk Assessment and Risk Reduction; ISO: Genova, Switzerland, 2010. [Google Scholar]
  3. ISO 9001:2008. Quality Management Systems; ISO: Genova, Switzerland, 2008. [Google Scholar]
  4. ISO 31010:2019. Risk Management—Risk Assessment Techniques; IEC: Genova, Switzerland, 2019. [Google Scholar]
  5. ISO 3534-1:2006. Statistics—Vocabulary and Symbols—Part 1: General Statistical Terms and Terms Used in Probability; ISO: Genova, Switzerland, 2006. [Google Scholar]
  6. Zuth, D.; Marada, T. Using artificial intelligence to determine the type of rotary machine fault. Mendel 2018, 24, 49–54. [Google Scholar] [CrossRef]
  7. Zuth, D.; Marada, T. Utilization of machine learning in vibrodiagnostics. Adv. Intell. Syst. Comput. 2018. [Google Scholar] [CrossRef]
  8. Zuth, D.; Marada, T. Comparison of Faults Classification in Vibrodiagnostics from Time and Frequency Domain Data. In Proceedings of the 2018 18th International Conference on Mechatronics-Mechatronika, Brno, Czech Republic, 5–7 December 2018; Volume 18, pp. 482–487, ISBN 978-802145544-3. [Google Scholar]
  9. Ympa, A. Learning Methods for Machine Vibration Analysis and Health Monitoring. Ph.D. Dissertation, Delft University of Technology, Delft, The Netherlands, 2001. [Google Scholar]
  10. Maluf, D.A.; Daneshmend, L. Application of machine learning for machine monitoring and diagnosis. Fla. Artif. Intell. Res. Symp. 1997, 10, 232–236. [Google Scholar]
  11. Li, C.; Sánchez, R.V.; Zurita, G.; Cerrada, M.; Cabrera, D. Fault Diagnosis for Rotating Machinery Using Vibration Measurement Deep Statistical Feature Learning. Sensors 2016, 16, 895. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Chen, Z.; Chen, X.; Li, C.; Sanchez, R.V.; Qin, H. Vibration-based gearbox fault diagnosis using deep neural networks. J. Vibroeng. 2017, 19, 2475–2496. [Google Scholar] [CrossRef]
  13. Popiołek, K. Diagnosing the technical condition of planetary gearbox using the artificial neural network based on analysis of non-stationary signals. Diagnostyka 2016, 17, 57–64. [Google Scholar]
  14. Haedong, J.; Park, S.; Woo, S.; Lee, S. Rotating Machinery Diagnostics Using Deep Learning on Orbit Plot Images. Procedia Manuf. 2016, 5, 1107–1118. [Google Scholar] [CrossRef] [Green Version]
  15. Scikit-Learn: Machine Learning in Python. Available online: https://scikit-learn.org/stable/index.html (accessed on 6 January 2018).
  16. Pánek, D.; Orosz, T.; Karban, P. Artap: Robust Design Optimization Framework for Engineering Applications. In Proceedings of the 2019 Third International Conference on Intelligent Computing in Data Sciences (ICDS), Marrakech, Morocco, 28–30 October 2019; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  17. Train Models to Classify Data Using Supervised Machine Learning. MathWorks-Makers of MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/classificationlearner-app.html (accessed on 20 April 2018).
  18. MCV 750 | KOVOSVIT MAS. Machine tools, CNC machines, CNC lathes | KOVOSVIT MAS [online]. Copyright© KOVOSVIT MAS 2016. Available online: https://www.kovosvit.com/mcv-750-p7.html (accessed on 20 April 2018).
  19. Broch, J.T. Mechanical Vibration and Shock Measurements, 2nd ed.; Bruel & Kjaer: Nellum, Denmark, 1980; ISBN 978-8787355346. [Google Scholar]
  20. Blata, J.; Juraszek, J. Metody Technické Diagnostiky, Teorie a Praxe. Metody Diagnostyki Technicznej, Teoria i Praktika; Repronis: Ostrava, Czech Republic, 2013; ISBN 978-80-248-2997-5. [Google Scholar]
  21. Zuth, D.; Vdoleček, F. Měření vibrací ve vibrodiagnostice. In Automa; FCC Public: Praha, Czech Republic, 2010; pp. 32–36. ISSN 1210-9592. [Google Scholar]
  22. Scikit-Learn 0.23.1 Documentation. Scikit-Learn: Machine learning in Python—Scikit-learn 0.16.1 Documentation. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.decomposition.PCA.html (accessed on 20 April 2018).
  23. Discrete Fourier Transform (numpy.fft)—NumPy v1.19 Manual. NumPy. Available online: https://numpy.org/doc/stable/reference/routines.fft.html (accessed on 20 April 2018).
  24. Classification, MathWorks-Makers of MATLAB and Simulink-MATLAB & Simulink. Available online: https://www.mathworks.com/help/stats/classification.html (accessed on 20 April 2018).
  25. Montgomery, D. Applied Statistics and Probability for Engineers; John Wiley: New York, NY, USA, 2003; ISBN 0-471-20454-4. [Google Scholar]
  26. Decision Trees, MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/decision-trees.html (accessed on 20 April 2018).
  27. Discriminant Analysis Classification, MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/discriminant-analysis.html (accessed on 20 April 2018).
  28. Naive Bayes Classification, MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/naive-bayes-classification.html (accessed on 20 April 2018).
  29. Support Vector Machines for Binary Classification, MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/support-vector-machines-for-binary-classification.html (accessed on 20 April 2018).
  30. Classification Using Nearest Neighbors, MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/classification-using-nearest-neighbors.html (accessed on 20 April 2018).
  31. Framework for Ensemble Learning, MATLAB and Simulink. Available online: https://www.mathworks.com/help/stats/framework-for-ensemble-learning.html (accessed on 20 April 2018).
Figure 1. CNC vertical machining center MCV 754.
Figure 1. CNC vertical machining center MCV 754.
Machines 09 00222 g001
Figure 2. Placement of the CMSS 2111 accelerometer to the machine MCV 754.
Figure 2. Placement of the CMSS 2111 accelerometer to the machine MCV 754.
Machines 09 00222 g002
Figure 3. The tool H490 F90AX D040-4-16-12 in fault-free state-Class 0 (left), without the insert Class 1 (middle) and without the insert and without the screw Class 2 (right).
Figure 3. The tool H490 F90AX D040-4-16-12 in fault-free state-Class 0 (left), without the insert Class 1 (middle) and without the insert and without the screw Class 2 (right).
Machines 09 00222 g003
Figure 4. The first test of the effect of the fault in frequency domain (C0−C2 are Class 0−Class 2, 1H—first harmonic frequency, max(1H)—maximum of the first harmonic frequency).
Figure 4. The first test of the effect of the fault in frequency domain (C0−C2 are Class 0−Class 2, 1H—first harmonic frequency, max(1H)—maximum of the first harmonic frequency).
Machines 09 00222 g004
Figure 5. The variation of the spindle speed over time during the dataset obtain.
Figure 5. The variation of the spindle speed over time during the dataset obtain.
Machines 09 00222 g005
Figure 6. Scheme of predictors calculation using IPC (“eF”—Spectrum energy, “Af1” to “Afn”—the sequence of amplitudes of individual spectral lines).
Figure 6. Scheme of predictors calculation using IPC (“eF”—Spectrum energy, “Af1” to “Afn”—the sequence of amplitudes of individual spectral lines).
Machines 09 00222 g006
Figure 7. The code names of experiments.
Figure 7. The code names of experiments.
Machines 09 00222 g007
Figure 8. The code names of experiments (the combined predictors).
Figure 8. The code names of experiments (the combined predictors).
Machines 09 00222 g008
Figure 9. The resulting successes rate of the classification and the success rate of the prediction of the fault-free state. The waveforms from the graph are T—Time domain predictors only, F—Frequency domain predictor only, TF—All predictors together, TF mix—All predictors together for all RPM (from 1000 to 1500 with step 50).
Figure 9. The resulting successes rate of the classification and the success rate of the prediction of the fault-free state. The waveforms from the graph are T—Time domain predictors only, F—Frequency domain predictor only, TF—All predictors together, TF mix—All predictors together for all RPM (from 1000 to 1500 with step 50).
Machines 09 00222 g009
Figure 10. Resulting confusion matrix for experiments with sample time t1 (0.25 s).
Figure 10. Resulting confusion matrix for experiments with sample time t1 (0.25 s).
Machines 09 00222 g010
Figure 11. Resulting confusion matrix for experiments with sample time t2 (0.5 s).
Figure 11. Resulting confusion matrix for experiments with sample time t2 (0.5 s).
Machines 09 00222 g011
Figure 12. IPC and server location topology.
Figure 12. IPC and server location topology.
Machines 09 00222 g012
Table 1. Measurement parameters of Microlog CMXA48.
Table 1. Measurement parameters of Microlog CMXA48.
Measurement Parameters:
Analysis type:Acceleration time domain
Units:g
Frequency range:1000 Hz
Y-Axis units:g
WAV (Waveform Audio File Format) File Parameters:
Bits per sample:16
Samples per second:2560
Average bytes per second:10,240
Format:PCM (Pulse Code Modulation)
Table 2. Resulting success rate of classification of individual methods for experiments with sample time t1 (0.25 s). The best results are marked in bold.
Table 2. Resulting success rate of classification of individual methods for experiments with sample time t1 (0.25 s). The best results are marked in bold.
t1R1Tt1R1Ft1R2Tt1R2Ft1R3Tt1R3Ft1R4Tt1R4F
Fine tree79.9 ± 0.5569.9 ± 1.6667.9 ± 1.1480.1 ± 1.2363.9 ± 0.9594.1 ± 0.3259.4 ± 0.6073.2 ± 0.37
Medium tree77.2 ± 0.8959.4 ± 1.2962.3 ± 0.8979.2 ± 1.2955.6 ± 1.1594.1 ± 0.3254.9 ± 0.4067.1 ± 0.24
Coarse tree76.8 ± 0.5447.2 ± 1.0857.6 ± 0.8972.4 ± 0.9751.6 ± 0.9894.5 ± 0.4652.9 ± 0.2958.1 ± 0.42
Linear discriminant74.1 ± 0.3060.1 ± 0.9556.8 ± 0.3074.1 ± 0.8051.8 ± 0.6295.0 ± 0.3051.3 ± 0.1864.3 ± 0.19
Quadratic discriminant74.2 ± 0.5269.1 ± 1.4159.9 ± 0.2081.4 ± 0.8644.7 ± 0.9395.6 ± 0.5852.3 ± 0.0769.5 ± 0.18
Gaussian Naive Bayes69.6 ± 0.6857.5 ± 1.1356.7 ± 0.3873.7 ± 0.6142.0 ± 0.6591.1 ± 0.6148.8 ± 0.1259.2 ± 0.37
Kernel Naive Bayes68.1 ± 0.5160.3 ± 0.8860.8 ± 0.6177.0 ± 0.8248.1 ± 0.8294.0 ± 0.3153.6 ± 0.1664.1 ± 0.19
Linear SVM73.0 ± 0.5060.3 ± 0.8057.1 ± 0.5675.4 ± 0.9351.9 ± 0.6294.9 ± 0.5152.1 ± 0.1365.1 ± 0.23
Quadratic SVM73.5 ± 0.3173.8 ± 1.0659.9 ± 0.4983.5 ± 0.5048.9 ± 0.6996.9 ± 0.2851.2 ± 0.8074.6 ± 0.26
Cubic SVM73.0 ± 0.7378.6 ± 1.1160.2 ± 1.0885.2 ± 0.5934.9 ± 1.1596.6 ± 0.3135.5 ± 0.8083.4 ± 0.35
Fine Gaussian SVM73.8 ± 0.9371.0 ± 1.2065.3 ± 1.1367.4 ± 1.0755.9 ± 0.8770.2 ± 1.8061.7 ± 0.1979.0 ± 0.65
Medium Gaussian SVM71.7 ± 0.4573.9 ± 0.6460.5 ± 0.6083.1 ± 0.7150.7 ± 0.9796.7 ± 0.4857.6 ± 0.2475.6 ± 0.27
Coarse Gaussian SVM73.0 ± 0.4457.2 ± 0.8457.2 ± 0.5471.4 ± 0.9649.8 ± 0.8793.5 ± 0.2352.8 ± 0.1365.4 ± 0.17
Fine KNN82.2 ± 0.6375.0 ± 1.2478.7 ± 1.1280.2 ± 0.9474.5 ± 0.6689.6 ± 0.5977.5 ± 0.2181.7 ± 0.29
Medium KNN72.1 ± 0.7253.5 ± 0.6959.7 ± 0.7864.8 ± 0.7951.6 ± 0.8681.4 ± 0.7460.9 ± 0.3267.7 ± 0.37
Coarse KNN68.4 ± 0.3044.9 ± 1.0056.2 ± 0.5965.5 ± 0.9247.8 ± 0.8375.2 ± 1.1159.3 ± 0.1864.5 ± 0.11
Cosine KNN68.2 ± 0.9855.7 ± 0.9457.4 ± 1.0663.0 ± 0.6151.4 ± 1.5781.2 ± 1.1359.4 ± 0.3066.6 ± 0.17
Cubic KNN72.1 ± 0.7355.4 ± 1.2659.2 ± 0.9464.0 ± 1.0751.3 ± 1.1078.1 ± 2.1260.9 ± 0.2366.8 ± 0.27
Weighted KNN82.8 ± 0.4076.1 ± 1.0180.6 ± 1.2481.8 ± 0.7075.3 ± 0.8489.4 ± 0.7778.5 ± 0.3182.6 ± 0.34
Boosted Trees79.0 ± 0.9572.9 ± 1.1265.2 ± 1.0584.3 ± 0.9061.5 ± 1.0435.4 ± 0.0057.0 ± 0.3971.5 ± 0.31
Bagged Trees84.1 ± 0.5876.9 ± 1.1378.8 ± 1.1977.0 ± 1.0075.7 ± 0.9496.1 ± 0.3978.4 ± 0.2784.4 ± 0.33
Subspace Discriminant73.4 ± 0.4259.6 ± 0.8355.6 ± 0.6173.0 ± 0.6452.3 ± 0.4495.3 ± 0.3251.1 ± 0.1163.2 ± 0.29
Subspace KNN82.1 ± 0.6772.6 ± 1.1876.3 ± 1.1674.8 ± 0.7174.7 ± 0.9884.3 ± 1.6475.5 ± 0.2577.4 ± 0.51
RUSBoosted Trees76.9 ± 0.5864.1 ± 1.3463.4 ± 0.9782.4 ± 1.0361.5 ± 1.5766.6 ± 9.8555.8 ± 0.4168.4 ± 0.28
Table 3. Resulting success rate of classification of individual methods for experiments with sample time t2 (0.5 s). The best results are marked in bold. Note: F-model failed (one or more of the classes have singular covariance matrices).
Table 3. Resulting success rate of classification of individual methods for experiments with sample time t2 (0.5 s). The best results are marked in bold. Note: F-model failed (one or more of the classes have singular covariance matrices).
t2R1Tt2R1Ft2R2Tt2R2Ft2R3Tt2R3Ft2R4Tt2R4F
Fine tree81.8 ± 1.7072.9 ± 1.7965.5 ± 1.5885.1 ± 1.3963.3 ± 2.2394.2 ± 0.6062.6 ± 0.4882.9 ± 0.81
Medium tree81.6 ± 1.6172.8 ± 1.6762.4 ± 1.8685.1 ± 1.3960.3 ± 2.2194.2 ± 0.6058.1 ± 0.5273.9 ± 0.74
Coarse tree75.7 ± 0.7855.2 ± 1.0258.0 ± 1.1279.4 ± 1.6059.5 ± 1.6994.1 ± 0.9553.9 ± 2.2560.9 ± 0.39
Linear discriminant76.9 ± 0.9070.6 ± 2.2855.1 ± 0.8384.4 ± 1.4659.0 ± 0.8697.7 ± 0.3154.5 ± 0.2472.0 ± 0.17
Quadratic discriminant72.3 ± 0.86F53.3 ± 0.75F50.5 ± 1.05F54.3 ± 0.3183.2 ± 0.28
Gaussian Naive Bayes68.2 ± 0.6767.2 ± 1.6852.6 ± 1.1774.2 ± 1.6041.0 ± 0.5985.4 ± 0.5051.8 ± 0.3164.5 ± 0.51
Kernel Naive Bayes69.3 ± 0.9977.8 ± 1.4650.7 ± 0.7784.6 ± 1.9248.0 ± 1.2298.4 ± 0.7255.0 ± 0.2774.4 ± 0.27
Linear SVM74.2 ± 0.7671.7 ± 0.7855.7 ± 1.3083.1 ± 1.3559.9 ± 0.7895.7 ± 0.4355.7 ± 0.2772.7 ± 0.28
Quadratic SVM77.4 ± 0.7284.0 ± 1.8755.4 ± 1.4089.5 ± 0.8461.6 ± 0.9998.0 ± 0.5056.6 ± 0.3187.4 ± 0.33
Cubic SVM76.8 ± 1.2584.9 ± 1.7654.0 ± 1.5488.2 ± 0.8354.8 ± 1.8597.9 ± 0.6739.3 ± 1.3589.6 ± 0.31
Fine Gaussian SVM78.6 ± 1.3873.9 ± 1.6564.9 ± 1.8175.4 ± 2.4264.0 ± 1.0772.9 ± 1.4763.9 ± 0.3471.7 ± 0.34
Medium Gaussian SVM76.7 ± 1.8584.9 ± 1.9556.2 ± 1.1787.8 ± 1.3062.1 ± 1.0797.9 ± 0.4960.9 ± 0.3084.7 ± 0.30
Coarse Gaussian SVM71.8 ± 1.0758.1 ± 0.6454.3 ± 0.9172.0 ± 1.2950.8 ± 1.1881.7 ± 0.5955.4 ± 0.2170.0 ± 0.22
Fine KNN86.2 ± 1.6182.0 ± 2.1378.4 ± 1.4483.2 ± 1.5274.7 ± 1.7791.7 ± 1.1978.3 ± 0.4685.1 ± 0.40
Medium KNN69.9 ± 1.2859.3 ± 1.8254.5 ± 1.6270.8 ± 1.3959.5 ± 2.0385.1 ± 1.1964.2 ± 0.2173.6 ± 0.45
Coarse KNN49.9 ± 1.0553.9 ± 1.0952.0 ± 1.2154.9 ± 1.3758.5 ± 0.9572.0 ± 0.6861.8 ± 0.2468.3 ± 0.20
Cosine KNN64.0 ± 0.9162.5 ± 1.5357.9 ± 1.0967.6 ± 1.2657.7 ± 1.7384.2 ± 1.1261.7 ± 0.3373.5 ± 0.61
Cubic KNN71.5 ± 1.2359.1 ± 1.3255.7 ± 1.1469.4 ± 1.3660.8 ± 1.4379.9 ± 0.7564.0 ± 0.3972.5 ± 0.32
Weighted KNN87.6 ± 1.6181.7 ± 1.6978.7 ± 1.4184.3 ± 0.8777.7 ± 1.6891.5 ± 1.4679.6 ± 0.4086.2 ± 0.43
Boosted Trees87.1 ± 1.1979.6 ± 1.2970.6 ± 1.8037.3 ± 0.0070.3 ± 2.4036.7 ± 0.0060.3 ± 0.3781.3 ± 0.38
Bagged Trees88.4 ± 1.1879.9 ± 1.1877.2 ± 1.3887.8 ± 1.2277.4 ± 1.3094.3 ± 1.2179.8 ± 0.3986.0 ± 0.41
Subspace Discriminant75.5 ± 0.8671.6 ± 1.5951.7 ± 0.8084.1 ± 1.3458.6 ± 1.1396.9 ± 0.4153.7 ± 0.2071.1 ± 0.12
Subspace KNN85.0 ± 1.5973.4 ± 1.4172.5 ± 2.2278.3 ± 2.0073.7 ± 1.9785.5 ± 1.9376.7 ± 0.3679.7 ± 0.49
RUSBoosted Trees86.2 ± 1.4477.7 ± 1.5367.1 ± 1.5647.9 ± 3.8662.6 ± 2.0876.0 ± 9.0459.4 ± 0.3477.6 ± 0.46
Table 4. Resulting success rate of classification of individual methods for experiments with combined predictors. The best results are marked in bold. Note: F-model failed (one or more of the classes have singular covariance matrices).
Table 4. Resulting success rate of classification of individual methods for experiments with combined predictors. The best results are marked in bold. Note: F-model failed (one or more of the classes have singular covariance matrices).
t1R1t1R2t1R3t1R4t2R1t2R2t2R3t2R4
Fine tree78.2 ± 0.9685.4 ± 0.6493.3 ± 0.4777.6 ± 0.4980.2 ± 1.6688.3 ± 1.7297.0 ± 0.4283.5 ± 0.66
Medium tree77.5 ± 1.0684.3 ± 0.8093.3 ± 0.4772.3 ± 0.3780.2 ± 1.6688.3 ± 1.7297.0 ± 0.4273.7 ± 0.44
Coarse tree76.0 ± 0.8174.5 ± 0.8692.2 ± 0.3658.8 ± 0.4478.1 ± 1.2878.8 ± 1.3697.0 ± 0.4262.1 ± 0.45
Linear discriminant76.2 ± 0.6878.3 ± 0.5996.1 ± 0.3270.3 ± 0.1381.0 ± 1.6681.4 ± 18.0198.1 ± 0.2676.4 ± 0.20
Quadratic discriminant81.5 ± 0.7983.1 ± 1.0193.7 ± 0.5277.6 ± 0.16FFF87.5 ± 0.25
Gaussian Naive Bayes71.4 ± 0.8571.2 ± 0.5981.6 ± 0.6561.0 ± 0.2962.3 ± 23.5280.0 ± 1.2793.7 ± 0.5967.7 ± 0.56
Kernel Naive Bayes78.4 ± 0.8976.7 ± 0.7792.7 ± 0.2870.5 ± 0.2682.5 ± 1.2685.8 ± 0.5899.0 ± 0.2778.7 ± 0.36
Linear SVM76.5 ± 0.5377.6 ± 0.3995.6 ± 0.4871.3 ± 0.2683.6 ± 1.2182.8 ± 0.8997.6 ± 0.5076.3 ± 0.33
Quadratic SVM85.2 ± 0.9185.0 ± 0.6495.6 ± 0.3881.6 ± 0.2186.6 ± 0.7088.2 ± 0.5298.4 ± 0.2489.5 ± 0.36
Cubic SVM85.8 ± 0.8585.6 ± 0.5596.3 ± 0.3388.1 ± 0.2786.2 ± 0.8288.1 ± 0.8998.6 ± 0.3990.9 ± 0.22
Fine Gaussian SVM70.8 ± 0.9470.4 ± 0.7772.1 ± 1.3681.5 ± 0.4069.4 ± 1.2875.7 ± 1.6066.8 ± 1.3374.9 ± 0.41
Medium Gaussian SVM85.5 ± 0.7884.4 ± 0.7795.2 ± 0.6582.5 ± 0.2087.9 ± 0.7886.6 ± 0.5598.1 ± 0.5088.4 ± 0.25
Coarse Gaussian SVM71.5 ± 0.4071.1 ± 0.4392.9 ± 0.4070.8 ± 0.1577.2 ± 0.7971.6 ± 0.7598.1 ± 0.3474.0 ± 0.31
Fine KNN78.2 ± 0.8380.8 ± 0.8490.1 ± 0.8884.5 ± 0.1483.1 ± 1.5785.4 ± 0.9593.7 ± 0.7287.1 ± 0.33
Medium KNN61.3 ± 0.7664.9 ± 1.0383.2 ± 0.7373.9 ± 0.1770.5 ± 1.4068.7 ± 1.8691.7 ± 0.6878.2 ± 0.38
Coarse KNN55.9 ± 0.7959.0 ± 0.9782.3 ± 0.4772.5 ± 0.2558.2 ± 0.6163.1 ± 1.0590.3 ± 0.7869.5 ± 0.69
Cosine KNN64.0 ± 1.1567.0 ± 0.8784.5 ± 0.9873.7 ± 0.2276.8 ± 1.4167.3 ± 2.0088.3 ± 0.7677.4 ± 0.32
Cubic KNN62.4 ± 0.6962.2 ± 0.8880.2 ± 0.7272.8 ± 0.2570.8 ± 1.3370.7 ± 1.7186.3 ± 0.6975.7 ± 0.42
Weighted KNN82.2 ± 0.9683.0 ± 1.0789.9 ± 0.6186.4 ± 0.2583.7 ± 1.2687.5 ± 1.3694.1 ± 0.6988.5 ± 0.33
Boosted Trees84.7 ± 0.8687.7 ± 0.5734.4 ± 0.0077.6 ± 0.2450.8 ± 8.9936.0 ± 0.0035.4 ± 0.0082.0 ± 0.37
Bagged Trees84.7 ± 1.0285.8 ± 0.5592.3 ± 0.5188.2 ± 0.3682.1 ± 1.3086.3 ± 1.8295.6 ± 0.8385.2 ± 0.51
Subspace Discriminant76.0 ± 0.4076.5 ± 0.4795.7 ± 0.6069.3 ± 0.1985.2 ± 1.2686.0 ± 0.9998.5 ± 0.2374.8 ± 0.31
Subspace KNN75.5 ± 1.3071.4 ± 1.1584.4 ± 0.9878.0 ± 0.3972.3 ± 1.4478.8 ± 1.0392.0 ± 0.6480.7 ± 0.63
RUSBoosted Trees81.8 ± 0.7485.1 ± 1.0364.9 ± 9.9973.3 ± 0.4681.6 ± 2.9369.7 ± 8.0960.5 ± 10.4275.7 ± 0.34
Table 5. Results of prediction accuracy after hyperparametric optimization, “Hyperparameter opt”-best model with hyperparameter optimization, “Difference”-Difference between result with and without hyperparameter optimization, “Avr. difference”-average difference from all experiments.
Table 5. Results of prediction accuracy after hyperparametric optimization, “Hyperparameter opt”-best model with hyperparameter optimization, “Difference”-Difference between result with and without hyperparameter optimization, “Avr. difference”-average difference from all experiments.
t1R1Tt1R1Ft1R2Tt1R2Ft1R3Tt1R3Ft1R4Tt1R4F
Best model:84.0%74.1%77.8%85.6%74.7%96.7%78.3%84.7%
Hyperparameter opt:85.3%77.6%81.5%85.8%75.2%94.6%80.9%86.2%
Difference:1.3%3.5%3.7%0.2%0.5%−2.1%2.6%1.5%
t2R1Tt2R1Ft2R2Tt2R2Ft2R3Tt2R3Ft2R4Tt2R4F
Best model:86.7%85.8%79.8%87.3%76.3%97.5%79.5%88.9%
Hyperparameter opt:83.1%84.9%81.1%86.4%80.8%97.5%79.5%88.7%
Difference:−3.6%−0.9%1.3%−0.9%4.5%0.0%0.0%−0.2%
t1R1t1R2t1R3t1R4t2R1t2R2t2R3t2R4
Best model:86.6%86.7%97.1%87.4%89.3%89.5%99.2%91.0%
Hyperparameter opt:87.3%87.1%96.7%89.1%88.4%89.5%99.2%91.3%
Difference:0.7%0.4%−0.4%1.7%−0.9%0.0%0.0%0.3%
Average difference: 0.55%.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zuth, D.; Blecha, P.; Marada, T.; Huzlik, R.; Tuma, J.; Maradova, K.; Frkal, V. Vibrodiagnostics Faults Classification for the Safety Enhancement of Industrial Machinery. Machines 2021, 9, 222. https://doi.org/10.3390/machines9100222

AMA Style

Zuth D, Blecha P, Marada T, Huzlik R, Tuma J, Maradova K, Frkal V. Vibrodiagnostics Faults Classification for the Safety Enhancement of Industrial Machinery. Machines. 2021; 9(10):222. https://doi.org/10.3390/machines9100222

Chicago/Turabian Style

Zuth, Daniel, Petr Blecha, Tomas Marada, Rostislav Huzlik, Jiri Tuma, Karla Maradova, and Vojtech Frkal. 2021. "Vibrodiagnostics Faults Classification for the Safety Enhancement of Industrial Machinery" Machines 9, no. 10: 222. https://doi.org/10.3390/machines9100222

APA Style

Zuth, D., Blecha, P., Marada, T., Huzlik, R., Tuma, J., Maradova, K., & Frkal, V. (2021). Vibrodiagnostics Faults Classification for the Safety Enhancement of Industrial Machinery. Machines, 9(10), 222. https://doi.org/10.3390/machines9100222

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop