Next Article in Journal
A New Method for the Exact Controllability of Linear Parabolic Equations
Next Article in Special Issue
Multiscale 1D-CNN for Damage Severity Classification and Localization Based on Lamb Wave in Laminated Composites
Previous Article in Journal
Mode Selection for Device to Device Communication in Dynamic Network: A Statistical and Deep Learning Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Generalized Autonomous Power Plant Fault Detection Model Using Deep Feature Extraction and Ensemble Machine Learning

by
Salman Khalid
1,
Muhammad Muzammil Azad
2 and
Heung Soo Kim
1,*
1
Department of Mechanical, Robotics and Energy Engineering, Dongguk University-Seoul, 30 Pildong-ro 1-gil, Jung-gu, Seoul 04620, Republic of Korea
2
Department of Mechanical Engineering, Dongguk University-Seoul, 30 Pildong-ro 1-gil, Jung-gu, Seoul 04620, Republic of Korea
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(3), 342; https://doi.org/10.3390/math13030342
Submission received: 9 December 2024 / Revised: 20 January 2025 / Accepted: 20 January 2025 / Published: 22 January 2025
(This article belongs to the Special Issue Artificial Intelligence for Fault Detection in Manufacturing)

Abstract

:
Ensuring operational reliability and efficiency in steam power plants requires advanced and generalized fault detection methodologies capable of addressing diverse fault scenarios in boiler and turbine systems. This study presents an autonomous fault detection framework that integrates deep feature extraction through Convolutional Autoencoders (CAEs) with the ensemble machine learning technique, Extreme Gradient Boosting (XGBoost). CAEs autonomously extract meaningful and nonlinear features from raw sensor data, eliminating the need for manual feature engineering. Principal Component Analysis (PCA) is employed for dimensionality reduction, enhancing computational efficiency while retaining critical fault-related information. The refined features are then classified using XGBoost, a robust ensemble learning algorithm, ensuring accurate fault detection. The proposed model is validated through real-world case studies on boiler waterwall tube leakage and motor-driven oil pump failure in steam turbines. Results demonstrate the framework’s ability to generalize across diverse fault types, detect anomalies at an early stage, and minimize operational downtime. This study highlights the transformative potential of combining deep feature extraction and ensemble machine learning for scalable, reliable, and efficient fault detection in power plant operations.

1. Introduction

Steam thermal power plants are essential to global energy production, providing a substantial portion of electricity to meet increasing industrial and domestic demands [1,2,3]. These plants operate by converting thermal energy into mechanical energy to drive steam turbines, which are connected to generators. Despite their critical role, the reliability and efficiency of steam power plants are frequently compromised by operational faults, particularly in boilers and steam turbines [4]. Boiler faults, which account for 52% of all power plant failures, are often caused by waterwall tube leakages due to thermal stress, material degradation, and corrosion [5,6]. Similarly, steam turbines are prone to mechanical faults such as rotor imbalances, gear faults, misalignments, and bearing defects [7,8,9]. A significant fault in turbine systems arises from failures in motor-driven oil pumps, which are critical for providing hydraulic lubrication and maintaining the control oil system. Inadequate oil supply, often caused by the malfunction or failure of the electric motor driving the pump, can lead to catastrophic bearing failures in the turbine’s rotating components [10]. Research indicates that approximately 40% of AC motor failures are caused by rolling bearing defects, emphasizing the importance of diagnosing bearing conditions proactively to prevent such failures [11]. Other prevalent faults in AC motors include winding failures, rotor and stator imbalances, broken rotor bars, and eccentricity-related problems [12].
As shown in Figure 1, the survey presents the fault distribution in steam power plants based on forced outages and severity, highlighting boiler and turbine faults as the major contributors [13]. These faults critically impact operational efficiency, lead to increased maintenance costs, and result in prolonged downtime, emphasizing the need for effective fault detection and diagnostic strategies. The significance of robust fault detection strategies has been highlighted in numerous studies. For instance, Babak et al. [14] reviewed the causes of boiler tube failures in steam power plants and stressed the importance of advanced monitoring systems to mitigate their impact on power generation. Similarly, Huang et al. [15] investigated vibration-based diagnostics for steam turbines, revealing that mechanical faults often go undetected until significant performance degradation occurs. This highlights the need for early detection techniques that can minimize operational disruptions.
Traditional fault detection methods, including manual inspections [16], rule-based monitoring [17], and basic signal analysis [18], have provided a foundation for addressing these challenges. Techniques such as vibration analysis and thermal imaging, while effective, remain reactive, labor-intensive, and limited in their ability to predict faults before significant disruptions occur [19,20]. To overcome these limitations, model-based methods have been introduced, utilizing mathematical models to represent normal system behavior and identify anomalies through deviations [21]. Techniques such as observer-based approaches, parameter estimation, and Kalman filters have demonstrated precision but are hindered by their reliance on accurate system modeling and sensitivity to parameter uncertainties [22,23,24]. Moreover, knowledge-based methods have utilized expert systems and fuzzy logic algorithms to detect anomalies based on predefined knowledge [25,26]. However, these approaches face challenges in adapting to evolving systems and handling unforeseen faults, limiting their scalability.
Recent advances in sensor technologies and data-driven methods have revolutionized fault detection in steam power plants [27]. Modern sensors generate high-fidelity data streams, facilitating the development of sophisticated analytical methods. For example, Min et al. [28] demonstrated the utility of piezoelectric sensors for real-time acoustic emission monitoring in steam boilers, achieving precise fault localization. Similarly, Ukil et al. [29] utilized distributed optical fiber sensors to monitor temperature variations in real time, enabling early detection of thermal anomalies. Data-driven methods have further enhanced fault detection by incorporating advanced statistical and machine learning techniques. Statistical techniques like PCA play a pivotal role in managing high-dimensional data. PCA facilitates dimensionality reduction while retaining essential features critical for anomaly detection. For instance, Jungwon et al. [30] applied PCA to detect plugged tubes in superheater banks during power plant startups, enabling timely decision-making and mitigating critical failures such as tube leakage. Miroslaw et al. [31] extended this approach with multiway PCA (MPCA) to model healthy system behavior in steam boilers, creating a confidence ellipsoid that allowed for early leak detection and improved maintenance efficiency. Ajami et al. [32] further demonstrated the efficacy of using PCA in turbine systems through a PCA-based inverse neural network control strategy for fault-tolerant control.
Machine learning techniques have significantly advanced fault detection by integrating robust preprocessing and classification algorithms. Khalid et al. [33] developed a sensor optimization framework for boiler waterwall tube leakage detection, combining correlation analysis with supervised learning for accurate fault classification. Jaswanth et al. [34] and Liang et al. [35] highlighted the robustness of XGBoost, a gradient boosting algorithm, in accurately classifying faults in boilers and turbines. Additionally, Zijun et al. [36] combined XGBoost with Dynamic Time Warping (DTW) for turbine health prognostics, achieving high reliability in detecting early signs of degradation. Similarly, Zhanhong et al. [37] explored a hybrid model combining XGBoost and genetic algorithms for fault detection in complex thermal systems, demonstrating superior adaptability to evolving operating conditions. Despite their success, many machine learning methods depend on manually engineered features, which may limit adaptability in dynamic industrial environments. Deep learning addresses the limitations of manual feature engineering by enabling autonomous feature extraction directly from raw data [38,39]. Hyeongmin et al. [40] proposed the Optimal Temporal Convolutional Auto-Encoder (Opt-TCAE) for boiler fault detection, capturing inter-sensor and temporal relationships to improve accuracy and reduce false alarms. Zhang et al. [41] combined Robust Long Short Term Memory (LSTM) Autoencoders with 1D CAEs to detect boiler leaks, effectively managing corrupted data and capturing dependency patterns. For turbines, Jinxing et al. [42] utilized CAEs to detect anomalies based on reconstruction errors, while Jose et al. [43] demonstrated the efficacy of CAEs in modeling normal operating conditions in gas turbines without labeled data. Despite their strengths, CAEs are not inherently designed for fault classification tasks [44]. While they excel in feature extraction, their effectiveness relies on the quality of the reconstructed latent representations, which can lead to overfitting or poor generalization when datasets are imbalanced or noisy.
The proposed framework directly addresses the challenges of overfitting and poor generalization through the following key contributions: CAEs are employed to autonomously extract meaningful and nonlinear features directly from raw sensor data, eliminating the reliance on manual feature engineering. Manual approaches often introduce biases that can lead to overfitting, whereas CAEs ensure a systematic and unbiased extraction of fault-relevant patterns. Additionally, PCA is integrated to refine these features by reducing dimensionality, focusing the model’s attention on critical fault-related information while discarding noise and redundancy. This step further mitigates overfitting risks and enhances the robustness of the feature set. Finally, the ensemble learning capabilities of XGBoost significantly enhance the generalization of the framework. By combining multiple decision trees, XGBoost effectively handles the challenges posed by noisy and imbalanced datasets, ensuring reliable and accurate fault detection across diverse scenarios. The proposed method is validated through real-world case studies, including boiler waterwall tube leakage and motor-driven oil pump failure in steam turbines, demonstrating its effectiveness and practicality for industrial fault detection applications.

2. Proposed Autonomous Fault Detection Methodology and Theoretical Foundations

2.1. Description of the Proposed Methodology

The proposed methodology, illustrated in Figure 2, begins with the collection of multi-sensor data from thermal power plants, including temperature data for boiler waterwall tube leakage and vibration data for turbine motor-driven oil pump faults. This dataset encompasses both healthy and faulty operating states, providing a robust foundation for analysis. During preprocessing, the raw data undergo normalization and segmentation to ensure consistency and prepare them for analysis. Autonomous feature extraction is performed using a CAE, which compresses high-dimensional data into a compact and informative feature space while preserving fault-relevant patterns. The extracted features are further refined through PCA for dimensionality reduction, optimizing computational efficiency while retaining critical information. The refined feature set is then classified using XGBoost, a robust ensemble learning algorithm, to accurately distinguish between healthy and faulty system states.

2.2. Theoretical Background of Applied Algorithms

(a)
PCA
PCA is a widely employed statistical technique for dimensionality reduction, feature extraction, and data visualization [45]. By transforming high-dimensional data into a lower-dimensional space, PCA preserves the most significant variance in the dataset, enabling efficient processing and analysis of complex systems. It achieves this by identifying principal components, which represent the orthogonal directions of maximum variance within the data. The PCA process begins with the calculation of the covariance matrix to capture the relationships between features in the dataset. For a mean-centered dataset X , the covariance matrix Σ is computed as [46]:
Σ = 1 n 1 X T X
Here, X is assumed to be mean-centered, ensuring that all features have zero mean. The covariance matrix Σ is symmetric and forms the basis for identifying directions of maximum variance in the data. PCA then derives the eigenvectors and eigenvalues of Σ , where the eigenvectors represent the principal components, and the eigenvalues quantify the variance explained by each component. The original data X are projected onto the new principal component space, resulting in a transformed dataset Y , as defined by:
Y = X V
where Y   R n × k represents the transformed data in the reduced k -dimensional space ( k < p ), and V   R n × k is the matrix of the top k eigenvectors corresponding to the largest eigenvalues of Σ . In steam power plants, PCA is extensively utilized for reducing the dimensionality of sensor data while retaining critical information [30,32]. This dimensionality reduction not only enhances computational efficiency but also filters out irrelevant variations, thereby improving signal quality. PCA’s robustness and ability to efficiently manage high-dimensional data make it an indispensable technique for fault detection and detection in complex industrial systems.
(b)
XGBoost
XGBoost is an advanced ensemble machine learning algorithm rooted in the gradient boosting framework [47]. It is widely acclaimed for its exceptional efficiency, scalability, and accuracy, particularly in structured data applications. XGBoost constructs an ensemble of decision trees by iteratively adding models that correct the errors of previous iterations, optimizing both predictive accuracy and generalization performance. The core of XGBoost lies in its objective function, which balances model predictive performance and complexity. The objective function is defined as [48]:
O b j e c t i v e = i = 1 n L y i , y ^ i + t = 1 T Ω f t
where L y i , y ^ i is the loss function measuring the difference between the true value y i and the predicted value y ^ i . Ω f t is the regularization term that penalizes model complexity. XGBoost employs gradient descent to minimize the objective function. At each iteration t , it fits a new decision tree f t ( x ) to the negative gradients (residuals) of the loss function from the previous iteration:
g i = L y i , y ^ i t 1 y ^ i t 1
The new predictions are then updated as:
y ^ i t = y ^ i t 1 + η f t ( x i )
where η is the learning rate, contributing each new tree to the overall prediction. In industrial applications, particularly in steam power plants, XGBoost has proven to be highly effective in fault detection and classification, owing to its ability to manage high-dimensional data and capture complex relationships [35,36]. Its capacity to identify subtle deviations from normal behavior makes it a powerful tool for early fault detection, enabling timely interventions and minimizing operational disruptions. When integrated with dimensionality reduction techniques such as PCA, XGBoost processes reduced feature sets with remarkable computational efficiency, maintaining high accuracy without overburdening resources.
(c)
SVM
SVM is a versatile algorithm for classification and regression tasks, capable of processing both linear and nonlinear data by utilizing kernel functions [49,50]. A linear kernel (LK), appropriate for linearly separable data, calculates the similarity between data points through the dot product of feature vectors:
K x , x = x · x
where x   a n d   x are feature vectors. For data requiring nonlinear decision boundaries, the polynomial kernel (PK) transforms data into higher-dimensional spaces, enabling the algorithm to capture complex relationships:
K x , x = x · x + c d
Here, c is a constant controlling the trade-off between high-order and low-order terms, and d is the degree of the polynomial. The radial basis function Kernel (RK) is well suited for highly nonlinear data. It maps data into infinite-dimensional space, capturing intricate patterns with the equation:
K x , x = e x p γ x x 2
In this equation, γ is a hyperparameter controlling the influence of individual data points, and x x 2 represents the squared Euclidean distance between two feature vectors. SVM’s adaptability through kernel functions makes it versatile for various classification challenges. By scaling features and employing dimensionality reduction techniques, SVM achieves high performance even in complex datasets. In fault detection scenarios, such as in boiler and turbine systems, SVM effectively identifies patterns indicative of faults, contributing to enhanced system reliability and minimized downtime [51]. Its robust decision-making capabilities are particularly advantageous in industrial environments where precise fault classification is critical.
(d)
Artificial neural networks (ANNs)
ANNs are computational models inspired by biological neural networks, designed to approximate complex nonlinear relationships in data [52]. They consist of layers of interconnected neurons, where each neuron applies a weighted sum of inputs followed by an activation function. ANNs learn by minimizing a loss function through backpropagation and adjusting weights and biases iteratively to improve accuracy. The output of a single-layer ANN can be expressed as:
y = f i = 1 n w i x i + b
where y is the output w i x i are weights, x i are inputs, b is the bias, and f is the activation function. In industrial applications, such as fault detection in power plants, ANNs are highly effective due to their adaptability and ability to process large-scale, high-dimensional data [53]. They excel in tasks requiring nonlinear decision boundaries, such as anomaly detection and classification. By utilizing their flexibility and scalability, ANNs provide robust solutions for identifying faults in complex systems, ensuring operational reliability and efficiency.
(e)
CAE
CAEs are a class of deep learning models specifically designed for unsupervised feature extraction and dimensionality reduction [54]. CAEs combine the principles of CNNs and traditional autoencoders, utilizing spatial hierarchies in data to extract meaningful features. Their architecture is particularly well suited for processing high-dimensional structured data, such as images, time series, and acoustic signals, where preserving spatial or temporal relationships is crucial. A CAE typically consists of two main components: an encoder and a decoder [55]. The encoder compresses the input data X into a lower-dimensional latent representation Z . It employs convolutional layers to capture local patterns and hierarchies within the data. The encoder’s function is mathematically represented as:
Z = f e n c o d e r ( X )
where Z represents the latent representation capturing essential features of the input data. The decoder reconstructs the input X from the latent representation Z , using transposed convolutional layers (deconvolutions) to restore the data to its original shape. The decoder’s function is given by:
X ^ = f e n c o d e r ( Z )
The objective of the CAE is to minimize the reconstruction error, which quantifies the difference between the original input X and reconstructed output X ^ . This reconstruction error is expressed as:
ζ = X X ^ 2
In industrial fault detection, CAEs are widely applied to analyze sensor data, acoustic emissions, and time-series signals [40,42]. By learning meaningful representations directly from raw data, CAEs are particularly effective for tasks such as anomaly detection, unsupervised learning, and feature extraction for classification. However, CAEs are not inherently optimized for fault classification, as their performance heavily depends on the quality of reconstructed latent representations. This limitation can lead to overfitting or poor generalization, particularly when datasets are noisy or imbalanced. To address the complexities of fault detection in boiler and turbine systems, this study presents an integrated framework combining CAEs, PCA, and XGBoost. The CAE architecture implemented in this study, as shown in Figure 3, comprises three encoding and three decoding layers, all using ReLU activation functions. This comprehensive approach achieves reliable and efficient fault detection, addressing critical challenges in maintaining the operational reliability of steam power plant systems.

3. Implementation of the Proposed Model on Real-World Steam Power Plant Data

This section presents two real-world case studies, steam turbine fault detection and boiler waterwall tube leakage detection, to validate the proposed autonomous fault detection model. These critical fault scenarios in steam power plants demonstrate the model’s ability to generalize across diverse fault types.

3.1. Data Acquisition

Accurate fault detection in power plants relies heavily on the acquisition of high-quality sensor data that reflects real-world operating conditions. For this study, critical data were obtained from steam power plant systems, focusing on two major fault scenarios: steam turbine motor-driven oil pump failure and boiler waterwall tube leakage.

3.1.1. Steam Turbine Motor-Driven Oil Pump Fault

For this case study, data were collected to analyze the performance and failures of the turbine’s motor-driven oil pump, a critical component of the steam turbine lubrication and control system. This oil pump ensures a continuous supply of hydraulic lubrication and control oil to the turbine’s rotating components, and its failure can result in insufficient oil supply, leading to severe bearing damage and potential turbine failure. To capture relevant fault patterns, three critical sensors were selected by power plant experts to monitor bearing vibration in the horizontal direction. The dataset consists of 8 days of healthy operational data and 8 days of faulty data, recorded at a sampling rate of 1 sample per minute to effectively monitor fault progression over time. As shown in Figure 4, the faulty data exhibit higher fluctuations, indicating significant anomalies in the system during fault conditions.

3.1.2. Boiler Waterwall Tube Leakage

For this case study, three critical sensors were selected by power plant experts to capture the most relevant steam temperature data for analysis. These sensors measured temperatures at the outlets of Superheater I (SH-I), Superheater II (SH-II), and Reheater I (RH-I). The dataset included 17.5 days of healthy operational data and 17.5 days of leakage data, collected at a sampling rate of one sample per second, resulting in a comprehensive dataset for fault detection and analysis. As shown in Figure 5, the healthy data are represented by the green curve, while the leakage data are depicted by the red curve. The leakage data exhibit noticeable fluctuations compared to the relatively stable patterns observed in the healthy data.

3.2. Data Preprocessing

To ensure consistency across the datasets and eliminate biases caused by varying scales, all sensor readings were normalized using Min-Max scaling [56]. This technique transforms the raw data into a standardized range of [0, 1], retaining the relative differences between the values and ensuring that no feature dominates the analysis due to its magnitude. Min–Max scaling was chosen for its simplicity and effectiveness in preserving the integrity of the data while standardizing it for efficient processing [57]. Normalization was performed for both healthy and faulty data collected from all sensors, enabling the models to better identify anomalies and patterns related to fault detection. This preprocessing step ensures that the fault classification process remains robust and reliable by enhancing the model’s ability to extract meaningful features. The normalized data for one sensor is depicted in Figure 6, which illustrates the transformed healthy and faulty states, demonstrating the effectiveness of this approach in preparing data for analysis.

3.3. Model Development and Evaluation

The model development and evaluation process, as illustrated in Figure 7, begins with the collection of raw sensor data from power plant systems, including both temperature and vibration signals to represent healthy and faulty states. The data undergo preprocessing, including cleaning and normalization, to ensure consistency and reliability. The pre-processed data are then randomly split into training (70%), validation (15%), and testing (15%) subsets. During the training phase, three distinct approaches were employed: the ANN, CAE-SVM and CAE-XGBoost models. PCA was utilized for dimensionality reduction, enhancing computational efficiency while retaining essential fault-related features. Once the models achieved convergence, the best-performing ones were saved and evaluated on the unseen testing dataset, ensuring a reliable assessment of their fault detection accuracy and generalization capabilities. To evaluate the proposed model’s robustness and accuracy, multiple performance metrics were employed. Classification accuracy served as the primary metric, indicating the proportion of correctly classified instances. To ensure a more comprehensive assessment, precision, recall, and F1-score were also analyzed. The equations for these metrics are provided in Reference [58]. Additionally, the confusion matrix provided detailed insights into the model’s strengths and areas for improvement by visualizing classification results.

4. Case Studies

This section presents the computational results obtained from applying the proposed fault detection framework to two case studies: steam turbine motor-driven oil pump fault and boiler waterwall tube leakage.

4.1. Turbine Fault Detection

The results of the turbine motor-driven oil pump fault detection, based on autonomous features extracted using the CAE, are presented in this section. A comparison is made for the traditional ANN-based deep learning model with the proposed hybrid models with CAE as a deep feature extractor integrated with machine learning classifiers for fault detection. The evaluation metrics in terms of training, validation, and testing accuracy are shown in Figure 8. The ANN model achieved a high training accuracy of 98.88%; however, its performance on the validation and testing datasets dropped to 93.10%. This indicates that while the ANN could learn patterns in the training data effectively, it struggled to generalize to unseen data, potentially due to overfitting or insufficient capacity to capture more complex fault-related features. The CAE-SVM models displayed varied performance depending on the kernel function utilized. The linear kernel (CAE-SVM-LK) demonstrated results similar to the ANN model, achieving 98.88% accuracy during training and maintaining 93.10% accuracy for validation and testing datasets. In comparison, the polynomial kernel (CAE-SVM-PK) exhibited slightly reduced training accuracy at 98.51%, though it achieved a testing accuracy of 95.55%, reflecting its ability to model more complex relationships. This consistency suggests that the polynomial kernel effectively captured the fault patterns in the turbine extracted by the CAE model without overfitting. The radial kernel (CAE-SVM-RK) performed well during training, achieving 99.25% accuracy, but validation and testing accuracy remained at 93.10%. Among the evaluated models, CAE-XGBoost delivered the most robust performance, achieving a perfect training accuracy of 100% and maintaining high validation and testing accuracies of 96.55%. The ensemble nature of XGBoost, combined with the ability of CAE to capture intricate relationships in the data, enabled it to outperform other models in detecting turbine faults. Table 1 presents the computational resource requirements for each model, measured in terms of training time, testing time, and model size. The hybrid models exhibit slightly higher computational times compared to the traditional ANN model; however, this increase is minimal and is well-justified by the significant reduction in overfitting and improvement in overall improvement in turbine fault detection performance. Moreover, the testing times for all hybrid models remain in the millisecond range; thus, the data acquisition rate of one sample per minute demonstrates the suitability of hybrid models for real-time monitoring. Although the hybrid models exhibit larger model sizes (approximately 1.6 MB), this remains negligible in the context of the modern big data era. These results underscore the effectiveness of combining CAE for autonomous feature extraction with a highly scalable and precise classifier like XGBoost.
The confusion matrices presented in Figure 9 illustrate the performance of the developed models on the unseen test dataset, emphasizing their capability to classify turbine faults accurately. The ANN model demonstrated highly unreliable fault classification, achieving 86.21% accuracy in identifying a healthy state and 100% accuracy in identifying a faulty case. Therefore, a 13.79% misclassification rate for healthy states limits the generalization ability of the ANN model. The CAE-SVM models, particularly using linear (LK) and polynomial (PK) kernels, improved the accuracy for healthy state to 93.10%, with a minor misclassification rate of 6.90%, showcasing their enhanced ability to model turbine fault features extracted by the CAE. The radial kernel (RK), however, showed identical performance to the ANN model with 86.21% accuracy for a healthy state. This highlights the kernel-specific limitations in capturing fault patterns through the SVM model. Moreover, the CAE-XGBoost achieved good performance and was consistent with CAE-SVM-LK and CAE-SVM-PK. The CAE-XGBoost model showed 93.10% accuracy for identifying a healthy state and maintained 100% accuracy for identifying the fault in the power plant turbine. This underscores the robustness and precision of the hybrid models by utilizing the CAE-extracted features for turbine fault classification.
Table 2 presents the detailed performance metrics of the ANN, CAE-SVM-based hybrid models, and the CAE-XGBoost model for turbine fault detection, evaluated in terms of accuracy, precision, recall, and F1-score. The ANN model achieved moderate accuracy for the healthy state (86.21%) and perfect accuracy for the faulty state (100%), but its precision and recall values indicate limited generalization. The ANN model also showcased large differences of 13~14% between the two health states in terms of all metrics, thus revealing an overall F1-score of 92.59% for the healthy state and 93.55% for the faulty state. The CAE-SVM models with linear and polynomial kernels demonstrated superior performance, achieving 93.10% accuracy for the healthy state and 100% for the faulty state, with consistently high F1-scores of 96.43% for the healthy state and 96.67% for the faulty state. The radial kernel again showed identical performance to ANN, while CAE-XGBoost showed identical performance to the CAE-SVM-LK and CAE-SVM-PK models. Thus, the detailed evaluation revealed that the hybrid models perform better compared to the traditional ANN model, and the kernel function in SVM plays a significant role in classifying the autonomous features extracted by the CAE model.
Table 3 provides a comparison of the proposed approach with existing popular methods for turbine fault detection, highlighting the superiority of the hybrid models. The CAE-SVM-LK and CAE-XGBoost models achieved the highest accuracy (96.55%), outperforming traditional methods such as SVM (88.10%), KNN (86.80%), and Naive Bayes (93.00%), as well as the standalone ANN (93.10%). This demonstrates the effectiveness of integrating autonomous feature extraction with advanced classifiers for enhanced fault detection performance.

4.2. Boiler Fault Detection

The ability of the three best models for turbine fault detection to generalize across different power plant components has been evaluated in this section. For this purpose, the same three hybrid models have been used for boiler waterwall tube leakage detection. Figure 10 illustrates the training, validation, and testing accuracies of these models for boiler leakage detection. The results highlight the generalization capability of the hybrid models when applied to a different fault detection scenario within the same power plant. Among the models, CAE-XGBoost exhibited the highest generalization ability, achieving 96.80% training accuracy, 94.93% validation accuracy, and 93.20% testing accuracy. Meanwhile, the CAE-SVM-LK showed relatively robust performance, maintaining 93.34%, 93.73%, and 92.00% for training, validation, and testing accuracy, respectively. Moreover, the CAE-SVM-PK model exhibited a slight decline in testing accuracy to 90.53%. These results emphasize the adaptability of the hybrid CAE-based framework, particularly the CAE-XGBoost model, which utilizes autonomous feature extraction and robust ensemble classification to handle diverse fault types.
The confusion matrices presented in Figure 11 illustrate the classification performance of the three hybrid models, CAE-SVM-LK, CAE-SVM-PK, and CAE-XGBoost, on the unseen test dataset for boiler leakage detection. Among these models, CAE-XGBoost demonstrated the most balanced performance, achieving 92.53% accuracy in correctly classifying the healthy state and 93.87% accuracy for the leakage state, with minimal misclassification rates of 7.47% and 6.13%, respectively. The CAE-SVM-LK model also performed reasonably well, with 92.00% classification accuracy for both healthy and faulty states, though it exhibited slightly higher misclassification rates of 8.00% compared to CAE-XGBoost. Moreover, the CAE-SVM-PK model showed reduced accuracy for the healthy state, classifying 85.87% of the instances correctly while achieving a higher 95.20% accuracy for the faulty state. However, it exhibited a misclassification rate of 14.13% for healthy conditions, indicating challenges in generalizing boiler leakage detection. Overall, these results validate the robustness of the CAE-XGBoost model in maintaining high classification accuracy across diverse fault types, demonstrating its superiority in utilizing CAE-extracted features for fault detection in critical components of the power plant.
Table 4 provides a comprehensive evaluation of the performance metrics for these three models applied to boiler fault detection. The CAE-SVM-LK model demonstrated balanced performance, achieving 92.00% accuracy, precision, recall, and F1-score for both healthy and leakage states, reflecting its consistent ability to classify healthy and leakage conditions with equal reliability. However, the CAE-SVM-PK model displayed a disparity in its performance metrics, achieving 85.87% accuracy for the healthy state, accompanied by a higher precision of 94.71%, but with a lower recall of 85.87%, indicating a tendency to overestimate healthy state identification. For the leakage state, the CAE-SVM-PK model achieved higher accuracy (95.20%) but at the cost of a slightly reduced precision (87.07%). In contrast, the CAE-XGBoost model demonstrated the most robust and balanced performance across all metrics, achieving 92.53% accuracy for the healthy state with precision and recall of 93.78% and 92.53%, respectively, resulting in an F1-score of 93.15%. Similarly, for the leakage state, it achieved 93.87% accuracy, with precision, recall, and F1-scores consistently above 92%. The robust generalization observed in the CAE-XGBoost model can be attributed to two primary factors: the autonomous feature extraction capability of the CAE and the ensemble-based classification approach of XGBoost. The CAE effectively captures complex fault patterns irrespective of the fault type directly from raw data, eliminating the need for manual feature engineering, while XGBoost incorporates its ensemble nature to combine multiple decision trees, ensuring precise classification and resilience to overfitting. Therefore, these results demonstrate the superior classification capability and generalization ability of the CAE-XGBoost model in accurately diagnosing both turbine fault and boiler leakage, making it the most reliable hybrid model for diverse fault detection scenarios in power plants.
Table 5 provides a comparison of the proposed approach with existing popular methods for boiler leakage detection, emphasizing the advantages of the hybrid models. The CAE-XGBoost model achieved the highest accuracy (93.20%), surpassing traditional methods such as SVM (90.50%), KNN (88.10%), Naive Bayes (85.70%), and Discriminant Analysis (88.10%). These results underscore the effectiveness of combining autonomous feature extraction with advanced classifiers for improved detection accuracy in boiler leakage scenarios.

5. Conclusions

This study proposed a novel autonomous fault detection framework that integrates CAEs, PCA, and XGBoost to address critical challenges in power plant fault detection. The framework was developed through a systematic process, starting with the identification of key fault scenarios—boiler waterwall tube leakage and turbine motor-driven oil pump failure—and the acquisition of high-quality sensor data representing both healthy and faulty states. To ensure consistency and eliminate biases in the data, normalization was applied as the sole preprocessing step, transforming raw sensor readings into a standardized range for effective analysis. Autonomous feature extraction was performed using CAEs, which captured complex fault-related patterns directly from the raw sensor data, eliminating the need for manual feature engineering. PCA was employed to reduce dimensionality while preserving critical fault-related information, and XGBoost provided robust classification, leveraging its ensemble learning capabilities to achieve high accuracy, precision, recall, and F1-scores. Comparative analysis demonstrated its superiority over traditional models, including ANNs and CAE-SVM hybrids, in terms of classification performance and generalization ability. This structured and iterative development approach ensured that the framework met the dual objectives of early fault detection and practical applicability. The CAE-XGBoost model was shown to significantly improve operational efficiency, making it a scalable and robust solution for critical industrial applications. Future work will focus on addressing the challenge of model explainability, which remains a significant limitation in deep learning-based approaches. Developing interpretable methods to provide insights into the decision-making process of the proposed framework will enhance its transparency and foster greater trust for deployment in critical industrial applications.

Author Contributions

Conceptualization, H.S.K. and S.K.; methodology, S.K. and M.M.A.; software, S.K. and M.M.A.; formal analysis, S.K.; resources, H.S.K.; writing—original draft preparation, S.K. and M.M.A.; writing—review and editing, S.K. and H.S.K.; supervision, H.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2024-00405691).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Eguchi, S.; Takayabu, H.; Lin, C. Sources of Inefficient Power Generation by Coal-Fired Thermal Power Plants in China: A Metafrontier DEA Decomposition Approach. Renew. Sustain. Energy Rev. 2021, 138, 110562. [Google Scholar] [CrossRef]
  2. Khalil, E. Steam Power Plants. WIT Trans. State Art Sci. Eng. 2008, 42, 99–139. [Google Scholar]
  3. Tumanovskii, A.G.; Shvarts, A.L.; Somova, E.V.; Verbovetskii, E.K.; Avrutskii, G.D.; Ermakova, S.V.; Kalugin, R.N.; Lazarev, M.V. Review of the Coal-Fired, over-Supercritical and Ultra-Supercritical Steam Power Plants. Therm. Eng. 2017, 64, 83–96. [Google Scholar] [CrossRef]
  4. Omosanya, A.J.; Akinlabi, E.T.; Okeniyi, J.O. Overview for Improving Steam Turbine Power Generation Efficiency. J. Phys. Conf. Ser. 2019, 1378, 032040. [Google Scholar] [CrossRef]
  5. Che, C.; Qian, G.; Yang, X.; Liu, X. Fatigue Damage of Waterwall Tubes in a 1000 MW USC Boiler. In Proceedings of the 7th International Conference on Fracture Fatigue and Wear, Gent, Belgium, 9–10 July 2018; Abdel Wahab, M., Ed.; Lecture Notes in Mechanical Engineering. Springer: Singapore, 2019; pp. 314–324, ISBN 978-981-13-0410-1. [Google Scholar]
  6. Singh, P.M.; Mahmood, J. Stress Assisted Corrosion of Waterwall Tubes in Recovery Boiler Tubes: Failure Analysis. J. Fail. Anal. Prev. 2007, 7, 361–370. [Google Scholar] [CrossRef]
  7. Barella, S.; Bellogini, M.; Boniardi, M.; Cincera, S. Failure Analysis of a Steam Turbine Rotor. Eng. Fail. Anal. 2011, 18, 1511–1519. [Google Scholar] [CrossRef]
  8. Gałka, T. Vibration-Based Diagnostics of Steam Turbines. Mech. Eng. 2012, 34, 315–340. [Google Scholar]
  9. Wang, C.; Zhang, D.; Xie, Y. Research on Fault Diagnosis of Steam Turbine Rotor Unbalance and Parallel Misalignment Based on Numerical Simulation and Convolutional Neural Network. In Proceedings of the ASME Turbo Expo 2021: Turbomachinery Technical Conference and Exposition, Virtual, Online, 7–11 June 2021; American Society of Mechanical Engineers: New York, NY, USA, 2021; Volume 85017, p. V008T22A019. [Google Scholar]
  10. Niu, J.; Lu, S.; Liu, Y.; Zhao, J.; Wang, Q. Intelligent Bearing Fault Diagnosis Based on Tacholess Order Tracking for a Variable-Speed AC Electric Machine. IEEE Sens. J. 2019, 19, 1850–1861. [Google Scholar] [CrossRef]
  11. Dias, C.G.; Pereira, F.H. Broken Rotor Bars Detection in Induction Motors Running at Very Low Slip Using a Hall Effect Sensor. IEEE Sens. J. 2018, 18, 4602–4613. [Google Scholar] [CrossRef]
  12. Rao, S.G.; Lohith, S.; Gowda, P.C.; Singh, A.; Rekha, S.N. Fault Analysis of Induction Motor. In Proceedings of the 2019 IEEE International Conference on Intelligent Techniques in Control, Optimization and Signal Processing (INCOS), Tamilnadu, India, 11–13 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar]
  13. Khalid, S.; Song, J.; Raouf, I.; Kim, H.S. Advances in Fault Detection and Diagnosis for Thermal Power Plants: A Review of Intelligent Techniques. Mathematics 2023, 11, 1767. [Google Scholar] [CrossRef]
  14. Haghighat-Shishavan, B.; Firouzi-Nerbin, H.; Nazarian-Samani, M.; Ashtari, P.; Nasirpouri, F. Failure Analysis of a Superheater Tube Ruptured in a Power Plant Boiler: Main Causes and Preventive Strategies. Eng. Fail. Anal. 2019, 98, 131–140. [Google Scholar] [CrossRef]
  15. Huang, Y.-C.; Liu, C.-Y.; Huang, C.-M. Intelligent Approaches for Vibration Fault Diagnosis of Steam Turbine-Generator Sets. In Proceedings of the 2nd International Conference on Intelligent Technologies and Engineering Systems (ICITES2013), 12–14 December 2013, Kaohsiung, Taiwan; Juang, J., Chen, C.-Y., Yang, C.-F., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 585–591. [Google Scholar]
  16. Kang, S.J.; Moon, J.C.; Choi, D.-H.; Choi, S.S.; Woo, H.G. A Distributed and Intelligent System Approach for the Automatic Inspection of Steam-Generator Tubes in Nuclear Power Plants. IEEE Trans. Nucl. Sci. 1998, 45, 1713–1722. [Google Scholar] [CrossRef]
  17. Lu, B.; Upadhyaya, B.R. Monitoring and Fault Diagnosis of the Steam Generator System of a Nuclear Power Plant Using Data-Driven Modeling and Residual Space Analysis. Ann. Nucl. Energy 2005, 32, 897–912. [Google Scholar] [CrossRef]
  18. Lu, B.; Upadhyaya, B.R.; Perez, R.B. Structural Integrity Monitoring of Steam Generator Tubing Using Transient Acoustic Signal Analysis. IEEE Trans. Nucl. Sci. 2005, 52, 484–493. [Google Scholar]
  19. Li-Juan, G.; Chun-Hui, Z.; Min, H.; Yong, Z. Vibration Analysis of the Steam Turbine Shafting Caused by Steam Flow. ℡KOMNIKA Indones. J. Electr. Eng. 2013, 11, 4422–4432. [Google Scholar] [CrossRef]
  20. Caruso, F.T. Thermal Imaging for the Nuclear Power Industry. In Proceedings of the Thermosense VIII: Thermal Infrared Sensing for Diagnostics and Control, Cambridge, MA, USA, 17–19 September 1985; SPIE: Bellingham, WA, USA, 1986; Volume 581, pp. 122–127. [Google Scholar]
  21. Li, F.; Upadhyaya, B.R.; Coffey, L.A. Model-Based Monitoring and Fault Diagnosis of Fossil Power Plant Process Units Using Group Method of Data Handling. ISA Trans. 2009, 48, 213–219. [Google Scholar] [CrossRef] [PubMed]
  22. Odgaard, P.F.; Lin, B.; Jorgensen, S.B. Observer and Data-Driven-Model-Based Fault Detection in Power Plant Coal Mills. IEEE Trans. Energy Convers. 2008, 23, 659–668. [Google Scholar] [CrossRef]
  23. David, N.P.; Swaminathan, B. Modeling, Identification and Detection of Faults in Industrial Boiler. In Proceedings of the 2015 IEEE Technological Innovation in ICT for Agriculture and Rural Development (TIAR), Chennai, India, 10–12 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 197–201. [Google Scholar]
  24. Sivathanu, A.K.; Vaidyanathan, K.; Murugan, N. Extended Kalman Filter Based Tube Leak Detection for Thermal Power Plant Reheater. In Proceedings of the AIP Conference Proceedings, Advances in Mechanical Engineering (ICAME-2022), Chennai, India, 24–26 March 2022; AIP Publishing: Melville, NY, USA, 2023; Volume 2813. [Google Scholar]
  25. Afgan, N.; Coelho, P.J.; Carvalho, M.G. Boiler Tube Leakage Detection Expert System. Appl. Therm. Eng. 1998, 18, 317–326. [Google Scholar] [CrossRef]
  26. Holbert, K.E.; Lin, K. Nuclear Power Plant Instrumentation Fault Detection Using Fuzzy Logic. Sci. Technol. Nucl. Install. 2012, 2012, 421070. [Google Scholar] [CrossRef]
  27. Choi, H.; Kim, C.-W.; Kwon, D. Data-Driven Fault Diagnosis Based on Coal-Fired Power Plant Operating Data. J. Mech. Sci. Technol. 2020, 34, 3931–3936. [Google Scholar] [CrossRef]
  28. Choi, M.-G.; Kim, J.-Y.; Jeong, I.; Kim, Y.-H.; Kim, J.-M. A Real-Time Monitoring System for Boiler Tube Leakage Detection. In Proceedings of the Hybrid Intelligent Systems, Porto, Portugal, 13–15 December 2018; Madureira, A.M., Abraham, A., Gandhi, N., Varela, M.L., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 106–114. [Google Scholar]
  29. Ukil, A.; Braendle, H.; Krippner, P. Distributed Temperature Sensing: Review of Technology and Applications. IEEE Sens. J. 2012, 12, 885–892. [Google Scholar] [CrossRef]
  30. Yu, J.; Yoo, J.; Jang, J.; Park, J.H.; Kim, S. A Novel Plugged Tube Detection and Identification Approach for Final Super Heater in Thermal Power Plant Using Principal Component Analysis. Energy 2017, 126, 404–418. [Google Scholar] [CrossRef]
  31. Swiercz, M.; Mroczkowska, H. Multiway PCA for Early Leak Detection in a Pipeline System of a Steam Boiler—Selected Case Studies. Sensors 2020, 20, 1561. [Google Scholar] [CrossRef] [PubMed]
  32. Ajami, A.; Daneshvar, M. Data Driven Approach for Fault Detection and Diagnosis of Turbine in Thermal Power Plant Using Independent Component Analysis (ICA). Int. J. Electr. Power Energy Syst. 2012, 43, 728–735. [Google Scholar] [CrossRef]
  33. Khalid, S.; Lim, W.; Kim, H.S.; Oh, Y.T.; Youn, B.D.; Kim, H.-S.; Bae, Y.-C. Intelligent Steam Power Plant Boiler Waterwall Tube Leakage Detection via Machine Learning-Based Optimal Sensor Selection. Sensors 2020, 20, 6356. [Google Scholar] [CrossRef] [PubMed]
  34. Jaswanth, C.; Shiva, G.P.; Raj, G.N.S.R.; M, T. Pipeline Leak Detection System Using Machine Learning. In Proceedings of the 2024 IEEE International Conference on Information Technology, Electronics and Intelligent Communication Systems (ICITEICS), Bangalore, India, 28–29 June 2024; pp. 1–6. [Google Scholar]
  35. Liang, Z.; Zhang, L.; Wang, X. A Novel Intelligent Method for Fault Diagnosis of Steam Turbines Based on T-SNE and XGBoost. Algorithms 2023, 16, 98. [Google Scholar] [CrossRef]
  36. Que, Z.; Xu, Z. A Data-Driven Health Prognostics Approach for Steam Turbines Based on Xgboost and Dtw. IEEE Access 2019, 7, 93131–93138. [Google Scholar] [CrossRef]
  37. Wu, Z.; Zhou, M.; Lin, Z.; Chen, X.; Huang, Y. Improved Genetic Algorithm and XGBoost Classifier for Power Transformer Fault Diagnosis. Front. Energy Res. 2021, 9, 745744. [Google Scholar] [CrossRef]
  38. Khan, A.; Kim, H.S. A Brief Overview of Delamination Localization in Laminated Composites. Multiscale Sci. Eng. 2022, 4, 102–110. [Google Scholar] [CrossRef]
  39. Azad, M.M.; Shah, A.U.R.; Prabhakar, M.N.; Kim, H.S. Deep Learning-based Fracture Mode Determination in Composite Laminates. J. Comput. Struct. Eng. Inst. Korea 2024, 37, 225–232. [Google Scholar] [CrossRef]
  40. Kim, H.; Ko, J.U.; Na, K.; Lee, H.; Kim, H.; Son, J.; Yoon, H.; Youn, B.D. Opt-TCAE: Optimal Temporal Convolutional Auto-Encoder for Boiler Tube Leakage Detection in a Thermal Power Plant Using Multi-Sensor Data. Expert Syst. Appl. 2023, 215, 119377. [Google Scholar] [CrossRef]
  41. Zhang, H.; Zuo, Z.; Li, Z.; Ma, L.; Liang, S.; Lü, Q.; Zhou, H. Leak Detection for Natural Gas Gathering Pipelines under Corrupted Data via Assembling Twin Robust Autoencoders. Process Saf. Environ. Prot. 2024, 188, 492–513. [Google Scholar] [CrossRef]
  42. Zhai, J.; Ye, J.; Cao, Y. An Unsupervised Fault Warning Method Based on Hybrid Information Gain and a Convolutional Autoencoder for Steam Turbines. Energies 2024, 17, 4098. [Google Scholar] [CrossRef]
  43. Barrera, J.M.; Reina, A.; Mate, A.; Trujillo, J.C. Fault Detection and Diagnosis for Industrial Processes Based on Clustering and Autoencoders: A Case of Gas Turbines. Int. J. Mach. Learn. Cybern. 2022, 13, 3113–3129. [Google Scholar] [CrossRef]
  44. Yu, J.; Zhou, X. One-Dimensional Residual Convolutional Autoencoder Based Feature Learning for Gearbox Fault Diagnosis. IEEE Trans. Ind. Inform. 2020, 16, 6347–6358. [Google Scholar] [CrossRef]
  45. Maćkiewicz, A.; Ratajczak, W. Principal Components Analysis (PCA). Comput. Geosci. 1993, 19, 303–342. [Google Scholar] [CrossRef]
  46. Kurita, T. Principal Component Analysis (PCA). In Computer Vision; Ikeuchi, K., Ed.; Springer International Publishing: Cham, Switzerland, 2021; pp. 1013–1016. ISBN 978-3-030-63415-5. [Google Scholar]
  47. Chen, T.; He, T.; Benetsy, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K.; Mitchell, R.; Cano, I.; Zhou, T.; et al. Extreme Gradient Boosting. R Package Version 1.3.2.1. 2021. Available online: https://cran.r-project.org/web/packages/xgboost/index.html (accessed on 19 January 2025).
  48. Cherif, I.L.; Kortebi, A. On Using eXtreme Gradient Boosting (XGBoost) Machine Learning Algorithm for Home Network Traffic Classification. In Proceedings of the 2019 Wireless Days (WD), Manchester, UK, 24–26 April 2019; pp. 1–6. [Google Scholar]
  49. Guenther, N.; Schonlau, M. Support Vector Machines. Stata J. Promot. Commun. Stat. Stata 2016, 16, 917–937. [Google Scholar] [CrossRef]
  50. Valkenborg, D.; Rousseau, A.-J.; Geubbelmans, M.; Burzykowski, T. Support Vector Machines. Am. J. Orthod. Dentofacial Orthop. 2023, 164, 754–757. [Google Scholar] [CrossRef]
  51. Jing, C.; Hou, J. SVM and PCA Based Fault Classification Approaches for Complicated Industrial Process. Neurocomputing 2015, 167, 636–642. [Google Scholar] [CrossRef]
  52. Zou, J.; Han, Y.; So, S.-S. Overview of Artificial Neural Networks. In Artificial Neural Networks; Livingstone, D.J., Ed.; Methods in Molecular BiologyTM; Humana Press: Totowa, NJ, USA, 2008; Volume 458, pp. 14–22. ISBN 978-1-58829-718-1. [Google Scholar]
  53. Chang, H.-S.; Tsai, J.-L. Predict Elastic Properties of Fiber Composites by an Artificial Neural Network. Multiscale Sci. Eng. 2023, 5, 53–61. [Google Scholar] [CrossRef]
  54. Masci, J.; Meier, U.; Cireşan, D.; Schmidhuber, J. Stacked Convolutional Auto-Encoders for Hierarchical Feature Extraction. In Proceedings of the Artificial Neural Networks and Machine Learning—ICANN 2011, Espoo, Finland, 14–17 June 2011; Honkela, T., Duch, W., Girolami, M., Kaski, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 52–59. [Google Scholar]
  55. Pawar, K.; Attar, V.Z. Assessment of Autoencoder Architectures for Data Representation. In Deep Learning: Concepts and Architectures; Pedrycz, W., Chen, S.-M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 101–132. ISBN 978-3-030-31756-0. [Google Scholar]
  56. Mazziotta, M.; Pareto, A. Normalization Methods for Spatio-temporal Analysis of Environmental Performance: Revisiting the Min–Max Method. Environmetrics 2022, 33, e2730. [Google Scholar] [CrossRef]
  57. Raju, V.N.G.; Lakshmi, K.P.; Jain, V.M.; Kalidindi, A.; Padma, V. Study the Influence of Normalization/Transformation Process on the Accuracy of Supervised Classification. In Proceedings of the 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT), Tirunelveli, India, 20–22 August 2020; pp. 729–735. [Google Scholar]
  58. Kim, S.; Azad, M.M.; Song, J.; Kim, H. Delamination Detection Framework for the Imbalanced Dataset in Laminated Composite Using Wasserstein Generative Adversarial Network-Based Data Augmentation. Appl. Sci. 2023, 13, 11837. [Google Scholar] [CrossRef]
  59. Khalid, S.; Hwang, H.; Kim, H.S. Real-World Data-Driven Machine-Learning-Based Optimal Sensor Selection Approach for Equipment Fault Detection in a Thermal Power Plant. Mathematics 2021, 9, 2814. [Google Scholar] [CrossRef]
Figure 1. Fault distribution in steam power plants based on forced outages and severity.
Figure 1. Fault distribution in steam power plants based on forced outages and severity.
Mathematics 13 00342 g001
Figure 2. Proposed methodology for autonomous fault detection in boiler and turbine systems.
Figure 2. Proposed methodology for autonomous fault detection in boiler and turbine systems.
Mathematics 13 00342 g002
Figure 3. Architecture of the CAE used for autonomous feature extraction.
Figure 3. Architecture of the CAE used for autonomous feature extraction.
Mathematics 13 00342 g003
Figure 4. Healthy and faulty raw vibration data from Vibration Bearing 1 (X direction, Sensor S1), Vibration Bearing 2 (X direction, Sensor S2), and Vibration Bearing 3 (X direction, Sensor S3) over 8 days for the steam turbine motor-driven oil pump fault case study.
Figure 4. Healthy and faulty raw vibration data from Vibration Bearing 1 (X direction, Sensor S1), Vibration Bearing 2 (X direction, Sensor S2), and Vibration Bearing 3 (X direction, Sensor S3) over 8 days for the steam turbine motor-driven oil pump fault case study.
Mathematics 13 00342 g004
Figure 5. Visualization of sensor data from SH-I, SH-II, and RH-I over 17.5 days for healthy and leakage conditions.
Figure 5. Visualization of sensor data from SH-I, SH-II, and RH-I over 17.5 days for healthy and leakage conditions.
Mathematics 13 00342 g005
Figure 6. Normalized vibration data for Sensor S1, illustrating healthy and faulty operating conditions over 8 days.
Figure 6. Normalized vibration data for Sensor S1, illustrating healthy and faulty operating conditions over 8 days.
Mathematics 13 00342 g006
Figure 7. Overview of the proposed model development methodology, including data preprocessing, random data splitting, training, validation, and testing phases for fault detection in steam power plant systems.
Figure 7. Overview of the proposed model development methodology, including data preprocessing, random data splitting, training, validation, and testing phases for fault detection in steam power plant systems.
Mathematics 13 00342 g007
Figure 8. The training, validation and testing accuracies for turbine fault detection in steam power plant.
Figure 8. The training, validation and testing accuracies for turbine fault detection in steam power plant.
Mathematics 13 00342 g008
Figure 9. The confusion matrix for the developed models for turbine fault detection using unseen test datasets.
Figure 9. The confusion matrix for the developed models for turbine fault detection using unseen test datasets.
Mathematics 13 00342 g009
Figure 10. The training, validation and testing accuracies for boiler fault detection in steam power plants using the best three hybrid models.
Figure 10. The training, validation and testing accuracies for boiler fault detection in steam power plants using the best three hybrid models.
Mathematics 13 00342 g010
Figure 11. The confusion matrix for the best three hybrid models for boiler fault detection using unseen test dataset.
Figure 11. The confusion matrix for the best three hybrid models for boiler fault detection using unseen test dataset.
Mathematics 13 00342 g011
Table 1. The computational resources required by each model in terms of computational time and model size.
Table 1. The computational resources required by each model in terms of computational time and model size.
ModelTraining Time (ms)Testing Time (ms)Model Size (Mbs)
ANN1973.44179.040.06
CAE-SVM-LK2807.62237.051.59
CAE-SVM-PK2807.62238.051.60
CAE-SVM-RK2806.62239.051.59
CAE-XGBoost2823.62239.051.67
Table 2. Performance evaluation of ANN, CAE-SVM and CAE-XGBoost models for turbine fault detection.
Table 2. Performance evaluation of ANN, CAE-SVM and CAE-XGBoost models for turbine fault detection.
ModelHealth StateAccuracy (%)Precision (%)Recall (%)F-1 Score (%)
ANNHealthy86.21100.0086.1092.59
Faulty100.0087.88100.0093.55
CAE-SVM-LKHealthy93.10100.0093.1096.43
Faulty100.0093.55100.0096.67
CAE-SVM-PKHealthy93.10100.0093.1096.43
Faulty100.0093.55100.0096.67
CAE-SVM-RKHealthy86.21100.0086.1092.59
Faulty100.0087.88100.0093.55
CAE-XGBoostHealthy93.10100.0093.1096.43
Faulty100.0093.55100.0096.67
Table 3. Comparison of the proposed approach with existing popular methods for turbine fault detection.
Table 3. Comparison of the proposed approach with existing popular methods for turbine fault detection.
ModelAccuracy (%)
SVM [59]88.10
KNN [59]86.80
Naive Bayes [59]93.00
ANN93.10
CAE-SVM-LK96.55
CAE-SVM-PK95.55
CAE-SVM-RK93.10
CAE-XGBoost96.55
Table 4. Performance evaluation of the best three hybrid models for boiler fault detection.
Table 4. Performance evaluation of the best three hybrid models for boiler fault detection.
ModelHealth StateAccuracy (%)Precision (%)Recall (%)F-1 Score (%)
CAE-SVM-LKHealthy92.0092.0092.0092.00
Leakage92.0092.0092.0092.00
CAE-SVM-PKHealthy85.8794.7185.8790.07
Leakage95.2087.0795.2090.06
CAE-XGBoostHealthy92.5393.7892.5393.15
Leakage93.8792.6393.8793.25
Table 5. Comparison of the proposed approach with existing popular methods for boiler leakage detection.
Table 5. Comparison of the proposed approach with existing popular methods for boiler leakage detection.
ModelAccuracy (%)
SVM [33]90.50
KNN [33]88.10
Naive Bayes [33]85.70
Discriminant analysis [33]88.10
CAE-SVM-LK92.00
CAE-SVM-PK90.53
CAE-XGBoost93.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khalid, S.; Azad, M.M.; Kim, H.S. A Generalized Autonomous Power Plant Fault Detection Model Using Deep Feature Extraction and Ensemble Machine Learning. Mathematics 2025, 13, 342. https://doi.org/10.3390/math13030342

AMA Style

Khalid S, Azad MM, Kim HS. A Generalized Autonomous Power Plant Fault Detection Model Using Deep Feature Extraction and Ensemble Machine Learning. Mathematics. 2025; 13(3):342. https://doi.org/10.3390/math13030342

Chicago/Turabian Style

Khalid, Salman, Muhammad Muzammil Azad, and Heung Soo Kim. 2025. "A Generalized Autonomous Power Plant Fault Detection Model Using Deep Feature Extraction and Ensemble Machine Learning" Mathematics 13, no. 3: 342. https://doi.org/10.3390/math13030342

APA Style

Khalid, S., Azad, M. M., & Kim, H. S. (2025). A Generalized Autonomous Power Plant Fault Detection Model Using Deep Feature Extraction and Ensemble Machine Learning. Mathematics, 13(3), 342. https://doi.org/10.3390/math13030342

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop