Proactive Fault Diagnosis of a Radiator: A Combination of Gaussian Mixture Model and LSTM Autoencoder
Abstract
:1. Introduction
- Section 2—Analyzing the Effects of Vibration on Radiator Integrity: In this section, an in-depth analysis of frequency response signals obtained from a real operational facility is undertaken to simulate the impact of vibrations on the structural integrity of radiators. A randomized durability vibration bench test is conducted, employing state-of-the-art acceleration sensors for real-time signal capture. The rationale behind the selection of GMM and LSTM autoencoders for fault diagnosis is elaborated upon.
- Section 3—Feature Engineering and Data Refinement: This section is dedicated to the critical task of feature engineering. Here, the focus is on extracting time-domain statistical features from raw data and refining them through Principal Component Analysis (PCA). The insights gained from PCA guide the subsequent training of the GMM. Findings related to fault diagnosis for stages 2 and 4, along with an in-depth exploration of anomaly detection for the unlabeled stage 3, are presented.
- Section 4—Leveraging LSTM Autoencoders: In this section, attention turns to the practical implementation of LSTM autoencoders for fault diagnosis and anomaly detection. The advantages of this approach are highlighted, showcasing its superiority over traditional methods, especially when dealing with time-series data.
- Section 5—Evaluation and Technical Contributions: The final section provides a comprehensive evaluation of the proposed combination of GMM and LSTM autoencoders. Here, meticulous detailing of the technical contributions of the methodology to the field of machine learning for fault diagnosis is presented. Emphasis is placed on its capability for precise fault detection and anomaly prediction. The practical implications of this approach for real-world applications, such as enhancing equipment uptime, preventing critical failures, and minimizing maintenance costs, are explored.
2. Radiator Dataset and Research Objective
2.1. Random Durability Vibration Bench Test and Dataset
2.2. Methods and Evluation Metrics
3. GMM-Based Fault Diagnosis
3.1. Feature Engineering and PCA of Radiator Dataset
3.2. Fault Diagnosis and Anomaly Detection Result Using GMM
- (1)
- Training the Gaussian Mixture Model.
- (2)
- Establishing a threshold value based on the absolute log-likelihood distribution derived from stage 1.
- (3)
- In stages 2–4, identifying data points with absolute log-likelihood values exceeding the predefined threshold as faulty.
- (1)
- Feature Engineering: Applying a ten-second window combined with sliding window augmentation.
- (2)
- Gaussian Distributions: Opting for three components, a selection validated by both BIC and AIC analyses as well as expert insights in fault diagnosis.
4. LSTM Autoencoder-Based Fault Diagnosis
4.1. Rationale for LSTM Autoencoder Selection
4.2. Fault Diagnosis and Anomaly Detection Result Using LSTM Autoencoder
- (1)
- Feature engineering method: Employing a 10-s window alongside sliding augmentation.
- (2)
- LSTM AE structure: Configured as 100/1/100 architecture.
- (3)
- Dropout and L2 Regularization: Not implemented.
5. Discussion
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Tsui, K.L.; Chen, N.; Zhou, Q.; Hai, Y.; Wang, W. Prognostics and health management: A review on data driven approaches. Math. Probl. Eng. 2015, 2015, 793161. [Google Scholar] [CrossRef]
- Lee, J.; Wu, F.; Zhao, W.; Ghaffari, M.; Liao, L.; Siegel, D. Prognostics and health management design for rotary machinery systems—Reviews, methodology and applications. Mech. Syst. Signal Process. 2014, 42, 314–334. [Google Scholar] [CrossRef]
- Meng, H.; Li, Y.F. A review on prognostics and health management (PHM) methods of lithium-ion batteries. Renew. Sustain. Energy Rev. 2019, 116, 109405. [Google Scholar] [CrossRef]
- Ahmad, R.; Kamaruddin, S. An overview of time-based and condition-based maintenance in industrial application. Comput. Ind. Eng. 2012, 63, 135–149. [Google Scholar] [CrossRef]
- Xu, Z.; Saleh, J.H. Machine learning for reliability engineering and safety applications: Review of current status and future opportunities. Reliab. Eng. Syst. Saf. 2021, 211, 107530. [Google Scholar] [CrossRef]
- Lan, T.; Gao, Z.-W.; Yin, H.; Liu, Y. A Sensor-Fault-Estimation Method for Lithium-Ion Batteries in Electric Vehicles. Sensors 2023, 23, 7737. [Google Scholar] [CrossRef]
- Li, Y.; Liu, C.; Li, D.; Hua, J.; Wan, P. Documentation of Tool Wear Dataset of NUAA_Ideahouse. IEEE Dataport. 2021. Available online: https://ieee-dataport.org/open-access/tool-wear-dataset-nuaaideahouse (accessed on 6 January 2023).
- Sayyad, S.; Kumar, S.; Bongale, A.; Kotecha, K.; Abraham, A. Remaining Useful-Life Prediction of the Milling Cutting Tool Using Time–Frequency-Based Features and Deep Learning Models. Sensors 2023, 23, 5659. [Google Scholar] [CrossRef]
- Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
- Cao, S.; Hu, Z.; Luo, X.; Wang, H. Research on fault diagnosis technology of centrifugal pump blade crack based on PCA and GMM. Measurement 2021, 173, 108558. [Google Scholar] [CrossRef]
- Chen, Z.; Peng, Z.; Zou, X.; Sun, H. Deep learning based anomaly detection for muti-dimensional time series: A survey. In Proceedings of the China Cyber Security Annual Conference, Beijing, China, 20–21 July 2021; pp. 71–92. [Google Scholar]
- Zhao, H.; Sun, S.; Jin, B. Sequential fault diagnosis based on LSTM neural network. IEEE Access 2018, 6, 12929–12939. [Google Scholar] [CrossRef]
- Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef] [PubMed]
- Janiesch, C.; Zschech, P.; Heinrich, K. Machine learning and deep learning. Electron. Mark. 2021, 31, 685–695. [Google Scholar] [CrossRef]
- Kamat, P.; Sugandhi, R. Anomaly detection for predictive maintenance in industry 4.0—A survey. In Proceedings of the E3S Web of Conferences, Lisboa, Portugal, 19–21 October 2020; Volume 170, p. 02007. [Google Scholar]
- Chalapathy, R.; Chawla, S. Deep learning for anomaly detection: A survey. arXiv 2019, arXiv:1901.03407. [Google Scholar]
- Dara, S.; Tumma, P. Feature extraction by using deep learning: A survey. In Proceedings of the 2018 Second International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 29–31 March 2018; pp. 1795–1801. [Google Scholar]
- Bergman, T.L.; Lavine, A.S.; Incropera, F.P.; DeWitt, D.P. Introduction to Heat Transfer, 5th ed.; John Wiley & Sons: New York, NY, USA, 2005; pp. 632–633. [Google Scholar]
- Klyatis, L.M. Accelerated Reliability and Durability Testing Technology; John Wiley & Sons: New York, NY, USA, 2012; Volume 70. [Google Scholar]
- Elsayed, E.A. Overview of reliability testing. IEEE Trans. Reliab. 2012, 61, 282–291. [Google Scholar] [CrossRef]
- Kapur, K.C.; Pecht, M. Reliability Engineering; John Wiley & Sons: New York, NY, USA, 2014; Volume 86. [Google Scholar]
- O’Connor, P.; Kleyner, A. Practical Reliability Engineering; John Wiley & Sons: New York, NY, USA, 2012. [Google Scholar]
- Zio, E. Reliability engineering: Old problems and new challenges. Reliab. Eng. Syst. Saf. 2009, 94, 125–141. [Google Scholar] [CrossRef]
- IEC 60068-2-6:2007; Environmental Testing—Part 2–6: Tests—Test Fc: Vibration (Sinusoidal). IEC: London, UK, 2011.
- MIL-STP-810; Environmental Engineering Considerations and Laboratory Tests. U.S. Department of Defense: Washington, DC, USA, 1964.
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3146–3154. [Google Scholar]
- Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and Regression Trees; CRC Press: Boca Raton, FL, USA, 1984. [Google Scholar]
- Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
- Noble, W.S. What is a support vector machine? Nat. Biotechnol. 2006, 24, 1565–1567. [Google Scholar] [CrossRef]
- Dreiseitl, S.; Ohno-Machado, L. Logistic regression and artificial neural network classification models: A methodology review. J. Biomed. Inform. 2002, 35, 352–359. [Google Scholar] [CrossRef]
- Reynolds, D.A. Gaussian mixture models. Encycl. Biom. 2009, 741, 659–663. [Google Scholar]
- Rasmussen, C. The infinite Gaussian mixture model. Adv. Neural Inf. Process. Syst. 1999, 12, 554–560. [Google Scholar]
- Nguyen, H.D.; Tran, K.P.; Thomassey, S.; Hamad, M. Forecasting and Anomaly Detection approaches using LSTM and LSTM Autoencoder techniques with the applications in supply chain management. Int. J. Inf. Manag. 2021, 57, 102282. [Google Scholar] [CrossRef]
- Said Elsayed, M.; Le-Khac, N.A.; Dev, S.; Jurcut, A.D. Network anomaly detection using LSTM based autoencoder. In Proceedings of the 16th ACM Symposium on QoS and Security for Wireless and Mobile Networks, Alicante, Spain, 16–20 November 2020; pp. 37–45. [Google Scholar]
- Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B Stat. Methodol. 1999, 61, 611–622. [Google Scholar] [CrossRef]
- Holland, S.M. Principal Components Analysis (PCA); Department of Geology, University of Georgia: Athens, GA, USA, 2008; p. 2501. [Google Scholar]
- Xuan, G.; Zhang, W.; Chai, P. EM algorithms of Gaussian mixture model and hidden Markov model. In Proceedings of the 2001 International Conference on Image Processing (Cat. No. 01CH37205), Thessaloniki, Greece, 7–10 October 2001; Volume 1, pp. 145–148. [Google Scholar]
- Kuha, J. AIC and BIC: Comparisons of assumptions and performance. Sociol. Methods Res. 2004, 33, 188–229. [Google Scholar] [CrossRef]
- Aurélien, G. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, 2nd ed.; O’Reilly Media, Inc.: Newton, MA, USA, 2019. [Google Scholar]
- Graves, A.; Graves, A. Long short-term memory. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar]
- Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef]
- Pascanu, R.; Mikolov, T.; Bengio, Y. On the difficulty of training recurrent neural networks. In Proceedings of the 30th International Conference on Machine Learning (PMLR), Atlanta, GA, USA, 17–19 May 2013; pp. 1310–1318. [Google Scholar]
- Hochreiter, S. The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef]
- Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
- Bao, W.; Yue, J.; Rao, Y. A deep learning framework for financial time series using stacked autoencoders and long-short term memory. PLoS ONE 2017, 12, e0180944. [Google Scholar] [CrossRef]
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
- Cortes, C.; Mohri, M.; Rostamizadeh, A. L2 regularization for learning kernels. arXiv 2012, arXiv:1205.2653. [Google Scholar]
State | Label | Time | Observation Result |
---|---|---|---|
Stage 1 | Normal 1 | 0–100 min | No coolant leakage in 100 min |
Stage 2 | Normal 2 | 100–300 min | No coolant leakage in 300 min |
Stage 3 | Unknown | 300–400 min | Estimate a coolant leakage between 300 and 400 min |
Stage 4 | Abnormal | 400–808 min | Coolant leakage at 400 min |
Category | Description |
---|---|
Variable type | Numerical data |
Sampling rate | 12,800 Hz |
Measurement unit | m/s2 |
Data Acquisition Equipment | Siemens SCADAS recorder with SCM-V24-II module
|
Sensors | Isotron accelerometer: Model 65-10
: +/− 1.0%, 100.0 < frequency <= 10,000.0 Hz : +/− 2.1%, 10,000.0 < frequency <= 15,000.0 Hz |
Missing data | None |
Window Size | The Number of Gaussian Components | ||||
---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | |
1 s | 0.6286 | 0.7018 | 0.7094 | 0.7095 | 0.7076 |
2 s | 0.6182 | 0.7536 | 0.7649 | 0.7630 | 0.7609 |
3 s | 0.6117 | 0.7872 | 0.8016 | 0.7971 | 0.7979 |
4 s | 0.6070 | 0.8183 | 0.8306 | 0.8291 | 0.8293 |
5 s | 0.6125 | 0.8384 | 0.8483 | 0.8532 | 0.8517 |
6 s | 0.6141 | 0.8572 | 0.8674 | 0.8704 | 0.8717 |
7 s | 0.6153 | 0.8696 | 0.8828 | 0.8822 | 0.8823 |
8 s | 0.6240 | 0.8642 | 0.8833 | 0.8852 | 0.8871 |
9 s | 0.6275 | 0.8724 | 0.8915 | 0.8904 | 0.8899 |
10 s | 0.6226 | 0.8873 | 0.9000 | 0.8995 | 0.9009 |
10 s (sliding) | 0.6310 | 0.8873 | 0.8998 | 0.8985 | 0.8998 |
Variables | Window Size | Hyperparameters | AUC | ||
---|---|---|---|---|---|
LSTM AE Structure | Dropout Rate | L2 Regularization | |||
Window size | 1 s | 100/1/100 | None | None | 0.8579 |
2 s | 100/1/100 | None | None | 0.9151 | |
3 s | 100/1/100 | None | None | 0.9631 | |
4 s | 100/1/100 | None | None | 0.9639 | |
5 s | 100/1/100 | None | None | 0.9572 | |
6 s | 100/1/100 | None | None | 0.9656 | |
7 s | 100/1/100 | None | None | 0.9672 | |
8 s | 100/1/100 | None | None | 0.9647 | |
9 s | 100/1/100 | None | None | 0.9683 | |
10 s | 100/1/100 | None | None | 0.9728 | |
10 s (sliding) | 100/1/100 | None | None | 0.9893 | |
The number of nodes | 10 s (sliding) | 200/1/200 | None | None | 0.9902 |
10 s (sliding) | 300/1/300 | None | None | 0.9912 | |
Dropout rate | 10 s (sliding) | 100/1/100 | 0.3 | None | 0.9868 |
10 s (sliding) | 100/1/100 | 0.5 | None | 0.9845 | |
10 s (sliding) | 100/1/100 | 0.7 | None | 0.9782 | |
L2 Regularization | 10 s (sliding) | 100/1/100 | None | 0.1 | 0.9729 |
10 s (sliding) | 100/1/100 | None | 0.01 | 0.9853 | |
10 s (sliding) | 100/1/100 | None | 0.001 | 0.9702 | |
Dropout rate +L2 Regularization | 10 s (sliding) | 100/1/100 | 0.5 | 0.01 | 0.9751 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lee, J.-G.; Kim, D.-H.; Lee, J.H. Proactive Fault Diagnosis of a Radiator: A Combination of Gaussian Mixture Model and LSTM Autoencoder. Sensors 2023, 23, 8688. https://doi.org/10.3390/s23218688
Lee J-G, Kim D-H, Lee JH. Proactive Fault Diagnosis of a Radiator: A Combination of Gaussian Mixture Model and LSTM Autoencoder. Sensors. 2023; 23(21):8688. https://doi.org/10.3390/s23218688
Chicago/Turabian StyleLee, Jeong-Geun, Deok-Hwan Kim, and Jang Hyun Lee. 2023. "Proactive Fault Diagnosis of a Radiator: A Combination of Gaussian Mixture Model and LSTM Autoencoder" Sensors 23, no. 21: 8688. https://doi.org/10.3390/s23218688
APA StyleLee, J. -G., Kim, D. -H., & Lee, J. H. (2023). Proactive Fault Diagnosis of a Radiator: A Combination of Gaussian Mixture Model and LSTM Autoencoder. Sensors, 23(21), 8688. https://doi.org/10.3390/s23218688