Next Article in Journal
Structural Properties of Dynamic Systems Biology Models: Identifiability, Reachability, and Initial Conditions
Next Article in Special Issue
Industrial Process Monitoring in the Big Data/Industry 4.0 Era: from Detection, to Diagnosis, to Prognosis
Previous Article in Journal
Special Issue “Real-Time Optimization” of Processes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Outlier Detection in Dynamic Systems with Multiple Operating Points and Application to Improve Industrial Flare Monitoring

1
Davidson School of Chemical Engineering, Purdue University, West Lafayette, IN 47907, USA
2
Analytical Technology Center, The Dow Chemical Company, Freeport, TX 77541, USA
3
Process System and Solutions, Emerson Process Management, Roundrock, TX 78681, USA
*
Author to whom correspondence should be addressed.
Processes 2017, 5(2), 28; https://doi.org/10.3390/pr5020028
Submission received: 25 March 2017 / Revised: 8 May 2017 / Accepted: 24 May 2017 / Published: 31 May 2017
(This article belongs to the Collection Process Data Analytics)

Abstract

:
In chemical industries, process operations are usually comprised of several discrete operating regions with distributions that drift over time. These complexities complicate outlier detection in the presence of intrinsic process dynamics. In this article, we consider the problem of detecting univariate outliers in dynamic systems with multiple operating points. A novel method combining the time series Kalman filter (TSKF) with the pruned exact linear time (PELT) approach to detect outliers is proposed. The proposed method outperformed benchmark methods in outlier removal performance using simulated data sets of dynamic systems with mean shifts. The method was also able to maintain the integrity of the original data set after performing outlier removal. In addition, the methodology was tested on industrial flaring data to pre-process the flare data for discriminant analysis. The industrial test case shows that performing outlier removal dramatically improves flare monitoring results through Partial Least Squares Discriminant Analysis (PLS-DA), which further confirms the importance of data cleaning in process data analytics.

1. Introduction

Modern process industries rely on dependable measurements from instrumentation in order to achieve efficient, reliable and safe operation. To this end, the concept of the Internet of Things is receiving wider acceptance in the industry. This is resulting in facilities with massively instrumented intelligent sensors, actuators, and other smart devices gathering real-time process knowledge at a high frequency. However, in order to take take full advantage of such abundant data streams, we need to extract useful information from them as outlined in the vision for future smart manufacturing platforms [1]. A typical application of processing large data streams is flare monitoring. Industrial flares are safety devices designed to burn waste gas in a safe and controlled manner [2], and they are commonly used in petroleum refining. The flares can be especially useful during non-routine situations—such as power outages, emergency conditions or plant maintenance activities. Conventionally, flare monitoring work relies on thermal imaging cameras to recognize the difference in the heat signature of a flare stack flame and the surrounding background [3]. In this paper, we will demonstrate that plant-wide historical data can also be used to monitor flares effectively.
Raw data from field instrumentation stored in historians is difficult to use directly for modeling until it is cleaned and processed. One of the most important aspects of data cleaning and pre-processing is to remove erroneous values (i.e., outliers in this paper) that are inconsistent with the expected behaviors of measured signals during that timeframe. In chemical processes, outliers could be generated through malfunction of sensors, erroneous signal processing by the control system, or human-related errors such as inappropriate specification of data collection strategies. Outliers can negatively affect main statistical analyses such as the t-test and ANOVA by violating corresponding distribution assumptions, masking the existence of anomalies and swamping effects. Outliers can further negatively impact downstream data mining and processing procedures such as system identification and parameter estimation [4].
Numerous methods have been proposed for outlier detection, and decent reviews can be found in the work done by experts from different fields such as computer science [5] and chemometrics [6,7]. Generally, based on whether we have knowledge of a process model a priori, we can categorize those methods into model-based and data-driven, and the latter can be further divided into four subcategories. First, from the perspective of regression theory, removing outliers from data sets is equivalent to estimating the underlying model of “meaningful” values. Several robust regression-based estimators in the presence of outliers are proposed, including L-estimator [8], R-estimator [9], M-estimators [10], S-estimators [11], etc. Although the estimators are simple to implement and take advantage of relations between variables, they do not work well when variables are independent, and the iterative procedures of deleting and refitting will significantly increase the computational cost.
Second, if we focus on estimating data location and scale robustly, we can apply several proximity-based methods including the general extreme studentized deviate (GESD) method [12,13], Hampel identifier [14], quartile-based identifier and boxplots [15] or minimum covariance determinant estimator [16]. It is important to point out that a critical assumption of proximity-based methods is that the data follow a well-behaved distribution. However, in the majority of cases, such an assumption does not hold in chemical processes data due to transient dynamics of measured signals. As a result, being able to discriminate outliers from normal process dynamics poses a challenge for end-users of process data. In current literature, a moving window technique is often used to account for process dynamics by performing statistical abnormality detection within a smaller local window [4]. However, such an approach does not always give satisfactory results when the time scales of variations in the datasets are non-uniform (fast and slow dynamics occurring in the same dataset).
Recently, machine learning methods have become increasingly popular in outlier detection. Typical examples include k-means clustering [17], k-nearest neighbor (kNN) [18], support vector machine (SVM) [19], principal component analysis (PCA) [20,21], and isolation forest [22]. The general advantage of machine learning algorithms lies in their capability to explore interactions between variables and computational efficiency. In this paper, we use isolation forest as the reference method, which isolates abnormal observations through randomly selecting a value between the maximum and minimum values, partitioning the data and constructing a series of trees. The average path length from the root node to the leaf node of a tree over the forest is used to measure the normality of an observation and outliers usually have shorter paths.
The final class of outlier detection methods is time series based including the time series Kalman filter (TSKF) [23]. The TSKF approximates normal process variations in the signal by univariate time-series models such as an autoregressive model (AR) and then identifies observations that are inconsistent with the expected time-series model behavior. The advantages of the TSKF method include robustness against ill-conditioned auto-covariance matrices (by using Burg’s model estimation method [24]) while maintaining the original dataset integrity. Based on case studies in [23], the TSKF method obtains superior performance with stationary time series; however, it does not perform well on non-stationary process data with distribution shifts like product grade transitions. In addition, considering the fact that continuous operations of many chemical plants, especially petrochemical ones, are scattered along different operating regions with frequent mean shifts for process variables [25], detecting outliers in such situations becomes quite challenging because data no longer follow a single well-behaved distribution. As a result, we have to find ways to help us quickly pinpoint the mean shift locations of each variable to improve the overall performance of univariate outlier detection within each operating region bracketed by changepoints.
Generally speaking, changepoint analysis focuses on identifying break points within a dataset where the statistical properties such as mean or variance show dramatic differences from previous periods [26]. Assuming we have a sequence of time series data, y 1 : n = y 1 , . . . , y n , and the data set includes m changepoints at τ 1 : m = τ 1 , . . . , τ m , where 1 τ 1 < τ i < τ m n - 1 . We define τ 0 = 0 and τ m + 1 = n , and assume that τ i are ordered, i.e., τ i < τ j if and only if i < j . As a result, the data y 1 : n are divided into m + 1 segments with the ith segment containing y τ i - 1 + 1 : τ i .
Changepoints can be identified by minimizing:
i = 1 m + 1 y τ i - 1 + 1 : τ i + β ,
where represents the cost function for a segment and β is a penalty to reduce over-fitting. The cost function is usually the negative log-likelihood given the density function f y | θ , as shown in Equation (2):
y τ i - 1 + 1 : τ i = - max θ i = τ i - 1 + 1 τ i log f y i | θ ,
where θ can be estimated by maximum likelihood given data within a single stage. The penalty is commonly chosen to be proportional to the number of changepoints, for example, β = m , and it grows with an increasing number of change points. For data y 1 : s , if we use F s to represent the optimal solution of Equation (1), and define T s = τ : 0 = τ 0 < τ 1 < < τ m < τ m + 1 = s to be the set of possible changepoints, the following recursion from optimal partitioning (OP) [27] can be used to give the minimal cost for data y 1 : t with t < s :
F s = min τ T s i = 1 m + 1 y τ i - 1 + 1 : τ i + β = min t min τ T t i = 1 m y τ i - 1 + 1 : τ i + β + y t + 1 : n + β = min t F t + y t + 1 : n + β ,
where F 0 = - β . The recursion can be solved in turn for s = 1 , 2 , . . . , n with a linear cost in s; as a result, the overall computational cost of finding F n using an optimal partitioning (OP) approach is O n 2 .
Other widely used changepoint search methods include binary segmentation (BS) [28] and segment neighborhood (SN) method [29] with O n l o g n and O Q n 2 computational costs (Q is the maximum number of change points), respectively. Although the BS method is computationally efficient, it does not necessarily find the global minimum of Equation (1) [26]. As for the SN method, since the number of change points Q increases linearly with n, the overall computational cost will become high with O n 3 . Bayesian approaches have also been reported [30,31] for changepoint detection, but the associated heavy computational cost cannot be overlooked. In this paper, because we deal with an enormous amount of high-frequency industrial data, a pruned exact linear (PELT) algorithm [26] with O n computational efficiency is adopted.
The article is organized as follows. In Section 2.1 to Section 2.3, after an introduction to both methods, we propose a strategy to integrate TSKF with PELT to improve TSKF’s performance in handling outlier detection in a dynamic data set with multiple operating points for each variable. In Section 2.4, the partial least squares discriminant analysis (PLS-DA) is briefly described to facilitate the understanding of how such a data-driven approach is applied in practice. In Section 3.1, the new outlier detection methodology framework is tested using simulated data set and compared with the conventional general extreme studentized deviate (GESD) and the latest isolation forest methods. Furthermore, The PELT-TSKF is applied to pre-processing both sediment and industrial flare data sets, and its efficacy in improving PLS-DA results are demonstrated in Section 3.2. Although this paper is mainly focused on building data-driven models for industrial flare monitoring, the sediment toxicity detection work is shown as an additional case study to demonstrate the proposed methodology.Discussion on the results of case studies is provided in Section 4. Finally, conclusions and future directions are given in Section 5.

2. Methodology

2.1. Pruned Exact Linear Time (PELT) Method

To improve the computational efficiency of the optimal partitioning approach, a PELT method [26] is proposed to prune τ s that can never be optimal at each iteration based on the assumption that there exists a constant K such that for all t < s < T ,
y t + 1 : s + y s + 1 : T + K y t + 1 : T .
As a result, at a future time T > s , if the following condition holds, there will never be an optimal last changepoint at t prior to T:
F t + y t + 1 : s + K F s .
In above formulations, if the cost function is represented by negative log-likelihood, K = 0. The overview of the PELT algorithm is shown in Figure 1. We first initialize the objective function F 0 = - β and define an empty changepoint set and a searching pool. Second, we iteratively find the minimal value of objective function F τ * and associated time point τ 1 , extend the change point set, and prune τ s that cannot reach optima by shrinking the searching pool R τ * . The pruning step shown in Figure 1 can significantly reduce the computational cost of the optimal partitioning approach especially when n is large.

2.2. Time Series Kalman Filter (TSKF)

Given a univariate data set y t t = 1 N , steps of an offline implementation of TSKF are as follows [23]:
  • Data partition: partition the data set into M subsets y t t = 1 N i , i = 1 , 2 , . . . , M .
  • Pre-whitening: for each subset y t t = 1 N i , i = 1 , 2 , . . . , M , pre-whiten the data using the reweighed minimum covariance determinant estimator [16], and centralize the data with robust center μ i .
  • Model fitting: based on the preliminary clean data y t c t = 1 N i ,
    3.1.
    (Optional) select the model order p according to Bayesian information criteria [32].
    3.2.
    Calculate the model coefficients Φ i based on Burg’s parameter estimation method [24].
  • Outlier detection: for each subset y t t = 1 N i , i = 1 , 2 , . . . , M :
    4.1
    Reformat:
    Y t = Θ Y t - 1 + U t , y t = H Y t ,
    where
    Y t T = y t , y t - 1 , . . . , y t - p + 1 1 × p ,
    U t T = ϵ ^ , 0 , . . . , 0 1 × p ; ϵ ^ N 0 , Q ,
    H = 1 , 0 , . . . , 0 1 × p ,
    Θ = Φ 1 Φ 2 Φ p - 1 Φ p 1 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 1 0 p × p .
    4.2
    Predict:
    Y ^ t | t - 1 = Θ Y ^ t - 1 | t - 1 ,
    P t | t - 1 = Θ P t - 1 | t - 1 Θ T + Q .
    4.3
    Update:
    E t = y t - H Y ^ t | t - 1 ,
    S t = H P t | t - 1 H T + τ I ,
    d t = E t T S t - 1 E t ,
    K t = P t | t - 1 H t T S t - 1 ,
    Y ^ t | t = Y ^ t | t - 1 + K t E t ,
    P t | t = I - K t H P t | t - 1 .
    4.4
    Detect:
    4.4.1.
    Set Δ = 0 .
    4.4.2.
    Find a number of n observations whose Mahalanobis distance d t Δ .
    4.4.3.
    Calculate the percentage of normal data:
    ξ = N i - n N i .
    4.4.4.
    If ξ γ , stop; otherwise, increase Δ by d Δ .
    4.4.5.
    The outliers correspond to observations with Mahalanobis distance d t Δ f i n a l .
    4.5
    Replace: replace the outliers with neighboring normal values.
Usually, the autoregressive model order p = 2 will suffice, and, in this study, we pick the threshold γ = 95 % , which means that we assume that the maximum amount of outliers in datasets would not exceed 5% of the total number of observations.

2.3. An Integrated Method for Outlier Detection in a Dynamic Data Set with Multiple Operating Points

To overcome the challenge of outlier detection in dynamic data sets with multiple operating points, we propose to combine the PELT with TSKF, and an integrated methodology overview is shown in Figure 2. For each variable y t t = 1 n , we first use the PELT algorithm to find mean change points that segment the data and then apply TSKF to outlier detection within each subset bracketed by change points. Finally, we replace the outliers with neighboring normal points.

2.4. Partial Least Squares Discriminant Analysis (PLS-DA)

Partial least squares discriminant analysis (PLS-DA) is a linear classification method that derives from partial least squares (PLS) [33,34] by building a regression model between the predictor matrix X and the predicted label matrix Y [35]. The standard mathematical formula for PLS-DA model is shown in Equation (20):
X = T P T + E = k = 1 a t k p k T + E Y = U Q T + F = k = 1 a u k q k T + F = X B + B 0 ,
where X is an n × m matrix of predictors; Y is an n × p one hot encoded matrix of class labels, and y i j represents the membership of the i-th sample to the j-th class expressed with binary code (1 or 0); T = t 1 , t 2 , . . . , t n T and U are n × a score matrices; P and Q are m × a and p × a orthogonal loading matrices, respectively; E and F are error matrices assumed to be independently and identically distributed; B = b 1 , b 2 , . . . , b p m × p , B 0 = b 0 , 1 T , b 0 , 2 T , . . . , b 0 , n T n × p T are the regression coefficients.
A distinctive feature of PLS-DA is that it maximizes the covariance between X and Y for each principal component spanning the reduced space, i.e., the covariance between T and U . In addition, advantages of PLS-DA include dimension reduction by using latent variables that are linear combinations of the original variables to model data variability and providing intuitive graphical visualization of data patterns and relations based on scores and loadings [36]. The PLS1 algorithm is commonly used to build PLS-DA models [37] in the two-class scenarios. PLS-DA returns estimated class values within (0,1). The classification of samples is conducted by choosing a label with the highest probability, and usually a threshold is used to determine which class a sample belongs to. In two-class scenarios with labels “0” and “1”, we assume that such a threshold equals 0.5 for simplicity, although the Bayes theorem can be used to give a more rigorous result [38].

3. Case Studies

In this section, PELT-TSKF’s capability of detecting univariate outliers in the dynamic data sets with multiple operating points is tested in simulated cases, and is compared with the general ESD (GESD) identifier [13] that detects outliers based on mean and standard deviation, and the isolation forest [22] that detects outliers based on finding points associated with the shortest average path lengths from the root node to the leaf node of a tree over the forest. (Note: the isolation forest is implemented in publicly available python package scikit-lean [39].). Moreover, we use an open-source sediment toxicity detection challenge as a toy problem to demonstrate the efficacy of PELT-TSKF in improving the PLS-DA model performance before applying the same methodology to industrial flare monitoring.

3.1. Simulated Case Study

In our simulated case studies, to illustrate the capability of outlier detection in dynamic data sets with mean shifts, performance of the new PELT-TSKF algorithm is compared with that of a GESD identifier combined with the PELT algorithm (the performance of GESD identifier will become much worse if without PELT). The intrinsic dynamic variation is approximated by a stationary autoregressive moving average (ARMA) (1,1) model:
1 - ϕ z - 1 x t = 1 - θ z - 1 ε t ,
where ϕ , θ are coefficients, ε is the white noise, ε t N 0 , 1 and z - 1 is a shift operator.
A total number of 1000 samples are simulated with two mean shifts at the 300th time point ( μ = 10 ) and the 600th time point ( μ = 20 ), respectively.
The contamination rate κ is defined as the percentage of outliers in the data:
κ = N o u t l i e r s N t o t a l d a t a .
In our simulation case, κ = 5 % , and outliers with equal probabilities being +Amp or −Amp (Amp is the amplitude of outliers) are added every 20th sample point except for the mean shift at the 300th and 600th time points.
The detection rate χ is defined as the percentage of outliers being successfully identified, and mis-identification rate β is defined as the percentage of normal data falsely tagged as outliers (type I error):
χ = N s u c c e s s f u l l y i d e n t i f i e d N t o t a l o u t l i e r s ,
β = N f a l s e a l a r m N n o r m a l d a t a .
A series of data sets with different values of ϕ , θ , A m p are simulated repetitively 100 times each, and the average detection rate χ ¯ and mis-identification β ¯ are given in Table 1.
Comparing the values of χ ¯ in Table 1, we can see that the PELT-TSKF works better than PELT-ESD and isolation forest algorithms when process autocorrelation increases with larger values of ( ϕ , θ ) . The result suggests that system dynamics negatively affect the GESD identifier and isolation forest significantly.
Moreover, a larger outlier size Amp will surely help the outlier detection by increasing the outlier detection rate χ , but it will not necessarily decrease the mis-identification rate β for PELT-GESD because the relative conservativeness of the GESD identifier leading to minor changes of the false alarms with different outlier sizes. Although the PELT-TSKF is generally more aggressive than the PELT-GESD with a higher mis-identification rate β , it has limited influence when abundant data are available coming from high-frequency sampling.
We provide an illustrative data set with ϕ = 0.5 , θ = - 0.5 , A m p = 4 . Figure 3 and Figure 4 show that the PELT algorithm can easily find two mean shifts at the 300th and 600th sample points marked by black dashed lines in both figures. In addition, additive outliers detected are highlighted by red squares, and we can see that the PELT-TSKF method is able to detect more additive outliers than the PELT-GESD identifier because there are more red squares in Figure 3 than in Figure 4. However, as indicated by the red circles shown in Figure 3, the PELT-TSKF cannot shield normal observations from swamping effects caused by outliers and leads to a few mis-identified instances—for example, the 21st point after an outlier at the 20th, and the 741st point after an outlier at the 740th time point are both mistaken for outliers because of previous anomalies.

3.2. Application of PELT-TSKF to PLS-DA Case Studies

3.2.1. Sediment Toxicity Detection

In this section, a sediment dataset is used. The X data include 1884 sediment samples and nine chemical variables that are the log-transformed concentrations (ppm) of metals (As, Cd, Cr, Cu, Pb, Hg, Ni, Ag, Zn) [40]. Sediment samples are divided into two classes on the basis of their toxicity, as shown in Table 2. While class 1 contains 1624 non-toxic samples, class 2 includes 260 toxic samples.
Figure 5 displays raw training X of the sediment data set. There are obvious mean shifts in variables like Cd, Cu, Ni, Ag, and Zn. Figure 6 and Figure 7 show cleaning results by PELT-TSKF on Cd and Ni data set. The new method is able to detect abnormal observations highlighted in red squares and find the mean shift points marked by black dash lines. In Figure 6, there is still an outlier at around the 1300th point that the PELT-TSKF fails to detect, mainly because there are not enough normal data on the second stage to provide a good estimation of autoregressive models. As a result, the outlier detection performance is compromised.
Both raw training and testing X data sets are cleaned by PELT-TSKF. Then, raw and clean data sets are both weight centered by subtracting the average means of two groups to remove biases caused by unequal class sizes [37] and scaled using associated standard deviations. PLS-DA models are built and tested based on resulting standardized raw and clean data sets. Three principal components are selected based on cross-validation and are able to capture 80% of total variance of X .
Three metrics are used to evaluate the model performance: non-error rate (NER, percentage of correctly predicted samples), toxicity sensitivity (Sn, ability to correctly recognize toxic samples as toxic in percentage), and non-toxicity specificity (Sp, ability to correctly recognize non-toxic samples as non-toxic, in percentage). The model testing results are shown in Table 3:
Results in Table 3 show that cleaning the outliers by PELT-TSKF improved the data quality and consequently improved the PLS-DA classification results as reflected through three metrics: N E R increased from 0.792 to 0.847, S n from 0.846 to 0.877 and S p from 0.783 to 0.842.

3.2.2. Industrial Flare Monitoring

In this section, the application of the proposed PELT-TSKF method and PLS-DA is tested with flare data sets taken from the petrochemical industry (for intellectual property reasons, the variables are named V1, V2, ...). Data are sampled at every 1 min. Flares associated with non-routine situations are tagged with “1” as opposed to those corresponding to normal operations with “0”. We randomly select two testing data sets and use observations that precede them as training data. Details on the sizes of training and testing sets are listed in Table 4.
In this study, out of 132 variables, we focus on critical ones including feed/outlet flows, temperature, water level, etc. Critical variables in both training and testing data are cleaned using PELT-TSKF methods, and part of the results are shown in Figure 8 and Figure 9. As highlighted by black dash lines and red squares, the new algorithm is able to find out mean shift points while detecting outliers masked by process variations.
Two metrics are used to evaluate the performance of non-routine flares detection: detection rate α and false alarm rate γ :
α = N s u c c e s s f u l l y i d e n t i f i e d N t o t a l n o n - r o u t i n e f l a r e s ,
γ = N f a l s e a l a r m s N n o r m a l d a t a .
Table 5 summarizes PLS-DA results based on raw and clean flare data sets, and we can see that cleaning outliers improves flare monitoring results in the Flare 1 case, increasing the detection rate α from 62.86% to 88.47%. In addition, although removing outliers does not affect the false alarm rate in the Flare 1 case, it helps reduce false alarm rates in the Flare 2 case, decreasing the false alarm rate β from 0.18% to 0.16% or reducing six false alarms in total.
Furthermore, Figure 10 and Figure 11 illustrate flare monitoring results obtained from PLS-DA models built on raw and clean data sets. While Y = 1 indicates the occurrence of non-routine flares, Y = 0 represents normal operation. Regarding Flare 1 results, comparing the middle plot with the bottom one shown in Figure 10, we can see that cleaning outliers will slightly increase the width of rectangular pulse, leading to a flare detection result close to the true measured flare events given by the plot on top.
As for Flare 2 results in Figure 11, we can easily find out that there are more spikes shown in the middle plot associated with raw data than the bottom one corresponding to a clean data set, indicating that the plant will suffer from a larger false alarm rate if no data cleaning is applied.

4. Discussion

From the results of simulated ARMA (1,1) data with mean shifts, we can see that PELT-TSKF is able to correctly identify the two stage changes and outperformed the general ESD and isolation forest approaches in outlier identification accuracy. This is especially apparent when the process dynamics become significant (shown by larger autocorrelation coefficients). However, it is worth pointing out that the isolation forest can also be used to identify multivariate outliers and is more efficient than the extension of TSKF in multivariate cases. The toy problem of sediment toxicity detection illustrates a successful application of PELT-TSKF in practice: cleaning outliers from concentration measurements and increasing the non-error rate, toxicity sensitivity and non-toxicity specificity calculated from the PLS-DA model as a result. In the flare monitoring case study, on the one hand, we demonstrate that applying data-driven approaches like PLS-DA can effectively assist process engineers; on the other hand, the efficacy of new PELT-TSKF has been proved in ameliorating the quality of industrial flaring data which contain multiple operating points, and improving the resulting flare monitoring performance by increasing the anomaly detection rate and reducing possibilities of false alarms.

5. Conclusions

Taking advantage of abundant raw plant data gathered by intelligent devices plays an important role in transforming modern plant operations from reactive to predictive. Data cleaning lays the foundation of knowledge discovery from data. In this paper, we focus on overcoming the challenges of outlier detection in dynamic systems with multiple operating points, and propose a novel method combining the time series Kalman filter (TSKF) with the pruned exact linear time (PELT) approach. The PELT-TSKF approach is able to pinpoint mean shifts and differentiate outliers from process variations at the same time. In addition, we demonstrate that applying data driven PLS-DA can assist process engineers in flare monitoring, and cleaning outliers using TSKF-PELT can further improve the performance of such a challenge task.
Although our method is able to deal with univariate outlier detection within dynamic system mean shifts, in practice, process operating points may be characterized by other factors other than mean, such as variance or frequency. In addition, operating point changes may affect individual variables in different ways: while a few of them show mean shifts, others exhibit variance changes. In addition, we only consider situations where operating point changes are significant enough that the impact of outliers can be ignored. When the impact of outliers on change points detection becomes significant, we can apply cross-validation to determine the optimal number of stage changes associated with the lowest cross validation error. Furthermore, multivariate outliers are also usually faced in process data, and how to efficiently detect multivariate outliers in the dynamic systems with multiple operating points is worth exploring in the future.

Acknowledgments

The authors would like to gratefully acknowledge the financial and technical support from the Emerson Process Management.

Author Contributions

Shu Xu and Bo Lu designed case studies; Shu Xu conducted the case studies, analyzed the data and wrote the paper; Noel Bell and Mark Nixon provided the industrial flare monitoring data and provided feedback on the results.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
TSKFtime series Kalman filter
PELTpruned exact linear time
PLS-DApartial least squares discriminant analysis
GESDgeneral extreme studentized deviate
ARMAautoregressive moving average
NERnon-error rate
Sntoxicity sensitivity
Spnon-toxicity specificity

References

  1. Davis, J.; Edgar, T.; Porter, J.; Bernaden, J.; Sarli, M. Smart manufacturing, manufacturing intelligence and demand-dynamic performance. Comput. Chem. Eng. 2012, 47, 145–156. [Google Scholar] [CrossRef]
  2. Torres, V.M.; Herndon, S.; Kodesh, Z.; Allen, D.T. Industrial flare performance at low flow conditions. 1. study overview. Ind. Eng. Chem. Res. 2012, 51, 12559–12568. [Google Scholar] [CrossRef]
  3. Bader, A.; Baukal, C.E., Jr.; Bussman, W. Selecting the proper flare systems. Chem. Eng. Prog. 2011, 107, 45–50. [Google Scholar]
  4. Xu, S.; Lu, B.; Baldea, M.; Edgar, T.F.; Wojsznis, W.; Blevins, T.; Nixon, M. Data cleaning in the process industries. Rev. Chem. Eng. 2015, 31, 453–490. [Google Scholar] [CrossRef]
  5. Hodge, V.J.; Austin, J. A survey of outlier detection methodologies. Artif. Intell. Rev. 2004, 22, 85–126. [Google Scholar] [CrossRef]
  6. Daszykowski, M.; Kaczmarek, K.; Heyden, Y.V.; Walczak, B. Robust statistics in data analysis—A review: Basic concepts. Chemometr. Intell. Lab. Syst. 2007, 85, 203–219. [Google Scholar] [CrossRef]
  7. Singh, A. Outliers and robust procedures in some chemometric applications. Chemometr. Intell. Lab. Syst. 1996, 33, 75–100. [Google Scholar] [CrossRef]
  8. Dielman, T.E. Least absolute value regression: Recent contributions. J. Stat. Comput. Simul. 2005, 75, 263–286. [Google Scholar] [CrossRef]
  9. Jaeckel, L.A. Estimating regression coefficients by minimizing the dispersion of the residuals. Ann. Math. Stat. 1972, 43, 1449–1458. [Google Scholar] [CrossRef]
  10. Huber, P.J. Robust estimation of a location parameter. Ann. Math. Stat. 1964, 35, 73–101. [Google Scholar] [CrossRef]
  11. Rousseeuw, P.; Yohai, V. Robust regression by means of S-estimators. In Robust and Nonlinear Time Series Analysis; Springer-Verlag: New York, NY, USA, 1984; pp. 256–272. [Google Scholar]
  12. Davies, L.; Gather, U. The identification of multiple outliers. J. Am. Stat. Assoc. 1993, 88, 782–792. [Google Scholar] [CrossRef]
  13. Rosner, B. Percentage points for a generalized ESD many-outlier procedure. Technometrics 1983, 25, 165–172. [Google Scholar] [CrossRef]
  14. Hampel, F.R. A general qualitative definition of robustness. Ann. Math. Stat. 1971, 42, 1887–1896. [Google Scholar] [CrossRef]
  15. Tukey, J.W. Exploratory Data Analysis (Behavior Science); Pearson: London, UK, 1977. [Google Scholar]
  16. Rousseeuw, P.J.; Driessen, K.V. A fast Algorithm for the minimum covariance determinant estimator. Technometrics 1999, 41, 212–223. [Google Scholar] [CrossRef]
  17. Allan, J.; Carbonell, J.; Doddington, G.; Yamron, J.; Yang, Y. Topic detection and tracking pilot study final report. In Proceedings of the DARPA Broadcast News Transcription and Understanding Workshop, Pittsburgh, PA, USA, 8–11 February 1998; pp. 194–218. [Google Scholar]
  18. Knorr, E.M.; Ng, R.T. Algorithms for mining distance based outliers in large datasets. In Proceedings of the International Conference on Very Large Data Bases; Gupta, A., Shmueli, O., Widom, J., Eds.; Morgan Kaufmann: New York, NY, USA, 1998; pp. 392–403. [Google Scholar]
  19. Tax, D.M.J.; Duin, R.P.W. Support vector data description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
  20. Chiang, L.H.; Russell, E.L.; Braatz, R.D. Fault Detection and Diagnosis in Industrial Systems; Springer-Verlag: London, UK, 2001. [Google Scholar]
  21. Kourti, T.; MacGregor, J.F. Process analysis, monitoring and diagnosis, using multivariate projection methods. Chemometr. Intell. Lab. Syst. 1995, 28, 3–21. [Google Scholar] [CrossRef]
  22. Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation Forest. In Proceedings of the ICDM Conference, Pisa, Italy, 15–19 December 2008. [Google Scholar]
  23. Xu, S.; Baldea, M.; Edgar, T.F.; Wojsznis, W.; Blevins, T.; Nixon, M. An improved methodology for outlier detection in dynamic datasets. AIChE J. 2015, 61, 419–433. [Google Scholar] [CrossRef]
  24. Kay, S.M. Modern Spectral Estimation: Theory and Application; Prentice Hall: Bergen County, NJ, USA, 1988. [Google Scholar]
  25. Dunia, R.; Edgar, T.F.; Blevins, T.; Wojsznis, W. Multistate analytics for continuous processes. J. Process Control 2012, 22, 1445–1456. [Google Scholar] [CrossRef]
  26. Killick, R.; Fearnhead, P.; Eckley, I.A. Optimal detection of changepoints with a linear computational cost. J. Am. Stat. Assoc. 2012, 107, 1590–1598. [Google Scholar] [CrossRef]
  27. Jackson, B.; Sargle, J.D.; Barnes, D.; Arabhi, S.; Alt, A.; Gioumousis, P.; Gwin, E.; Sangtrakulcharoen, P.; Tan, L.; Tsai, T.T. An algorithm for optimal partitioning of data on an interval. IEEE Signal Process. Lett. 2005, 12, 105–108. [Google Scholar] [CrossRef]
  28. Scott, A.J.; Knott, M. A cluster analysis method for grouping means in the analysis of variance. Biometrics 1974, 30, 507–512. [Google Scholar] [CrossRef]
  29. Auger, I.E.; Lawrence, C.E. Algorithms for the optimal identification of segment neighborhoods. Bull. Math. Biol. 1989, 51, 39–54. [Google Scholar] [CrossRef] [PubMed]
  30. Adams, R.P.; MacKay, D.J.C. Bayesian Online Changepoint Detection. 2007. Available online: http://arxiv.org/abs/0710.3742 (accessed on 3 March 2017).
  31. Lai, T.L.; Xing, H. Sequential Change-Point Detection When the Pre- and Post-Change Parameters are Unknown. Seq. Anal. 2010, 29, 162–175. [Google Scholar] [CrossRef]
  32. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  33. Geladi, P.; Kowalski, B.R. Partial least squares: A tutorial. Anal. Chim. Acta 1986, 185, 1–17. [Google Scholar] [CrossRef]
  34. Wold, S.; Sjostrom, M.; Eriksson, L. PLS-regression: A basic tool of chemometrics. Chemom. Intell. Lab. Syst. 2001, 58, 109–130. [Google Scholar] [CrossRef]
  35. Brereton, R.G. Chemometrics for Pattern Recognition; John Wiley and Sons: Chichester, UK, 2009. [Google Scholar]
  36. Ballabio, D.; Consonni, V. Classification tools in chemistry. Part 1: Linear models. PLS-DA. Anal. Method 2013, 5, 3790–3798. [Google Scholar] [CrossRef]
  37. Brereton, R.G.; Lloyd, G.R. Partial least squares discriminant analysis: Taking the magic away. J. Chemom. 2014, 28, 213–225. [Google Scholar] [CrossRef]
  38. Pérez, N.F.; Ferré, J.; Boqué, R. Calculation of the reliability of classification in discriminant partial least-squares binary classification. Chemom. Intell. Lab. Syst. 2009, 95, 122–128. [Google Scholar] [CrossRef]
  39. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  40. Alvarez-Guerra, M.; Ballabio, D.; Amigo, J.M.; Viguri, J.R.; Bro, R. A chemometric approach to the environmental problem of predicting toxicity in contaminated sediments. J. Chemom. 2010, 24, 379–386. [Google Scholar] [CrossRef]
Figure 1. Pruned exact linear time (PELT) algorithm overview.
Figure 1. Pruned exact linear time (PELT) algorithm overview.
Processes 05 00028 g001
Figure 2. An integrated outlier detection algorithm overview.
Figure 2. An integrated outlier detection algorithm overview.
Processes 05 00028 g002
Figure 3. Outlier detection by pruned exact linear time—time series Kalman filter (PELT-TSKF).
Figure 3. Outlier detection by pruned exact linear time—time series Kalman filter (PELT-TSKF).
Processes 05 00028 g003
Figure 4. Outlier detection by pruned exact linear time—general extreme studentized deviate (PELT-GESD).
Figure 4. Outlier detection by pruned exact linear time—general extreme studentized deviate (PELT-GESD).
Processes 05 00028 g004
Figure 5. Sediment training data set.
Figure 5. Sediment training data set.
Processes 05 00028 g005
Figure 6. Outlier detection result on Cd.
Figure 6. Outlier detection result on Cd.
Processes 05 00028 g006
Figure 7. Outlier detection result on Ni.
Figure 7. Outlier detection result on Ni.
Processes 05 00028 g007
Figure 8. PELT-TSKF for V1.
Figure 8. PELT-TSKF for V1.
Processes 05 00028 g008
Figure 9. PELT-TSKF for V2.
Figure 9. PELT-TSKF for V2.
Processes 05 00028 g009
Figure 10. Flare 1: Partial least squares-discriminant analysis (PLS-DA) based on raw and clean data.
Figure 10. Flare 1: Partial least squares-discriminant analysis (PLS-DA) based on raw and clean data.
Processes 05 00028 g010
Figure 11. Flare 2: PLS-DA analysis based on raw and clean data.
Figure 11. Flare 2: PLS-DA analysis based on raw and clean data.
Processes 05 00028 g011
Table 1. Additive outlier detection results for data from autoregressive moving average (ARMA) (1, 1) multistage processes at κ = 5 % .
Table 1. Additive outlier detection results for data from autoregressive moving average (ARMA) (1, 1) multistage processes at κ = 5 % .
Case No. ϕ θ AmpPELT-GESD (*)PELT-TSKF (**)Isolation Forest
β ¯ ( / % ) χ ¯ ( / % ) β ¯ ( / % ) χ ¯ ( / % ) β ¯ ( / % ) χ ¯ ( / % )
10.3−0.340.2753.870.8284.292.0064.40
20.3−0.350.3588.440.4791.421.7669.23
30.3−0.540.2739.730.9880.632.2260.19
40.3−0.550.2975.400.6587.601.9565.44
50.5−0.340.2937.600.9980.562.4056.63
60.5−0.350.2869.190.6986.692.0164.25
70.5−0.540.2925.901.2275.712.7349.98
80.5−0.550.3069.980.6887.082.4555.69
(*) PELT-GESD: pruned exact linear time—general extreme studentized deviate; (**) PELT-TSKF: pruned exact linear time—time series Kalman filter.
Table 2. Sediment dataset: sample partition in toxic and non-toxic classes.
Table 2. Sediment dataset: sample partition in toxic and non-toxic classes.
Non-Toxic (Class 1)Toxic (Class 2)Total
Training set12181951413
Test set40665471
Total16242601884
Table 3. Sediment test results summary.
Table 3. Sediment test results summary.
NER (*)SnSp
Raw data0.7920.8460.783
Clean data0.8470.8770.842
(*) NER: non-error rate.
Table 4. Sizes of training and testing sets.
Table 4. Sizes of training and testing sets.
ObservationsVariables
Flare training set 110,800132
Flare testing set 12200132
Flare training set 244,998132
Flare testing set 230,000132
Table 5. Partial least squares-discriminant analysis (PLS-DA) results on raw and clean data sets.
Table 5. Partial least squares-discriminant analysis (PLS-DA) results on raw and clean data sets.
α (/%) γ (/%)
Flare 1 raw data62.860.092
Flare 1 clean data88.470.092
Flare 2 raw data1000.18
Flare 2 clean data1000.16

Share and Cite

MDPI and ACS Style

Xu, S.; Lu, B.; Bell, N.; Nixon, M. Outlier Detection in Dynamic Systems with Multiple Operating Points and Application to Improve Industrial Flare Monitoring. Processes 2017, 5, 28. https://doi.org/10.3390/pr5020028

AMA Style

Xu S, Lu B, Bell N, Nixon M. Outlier Detection in Dynamic Systems with Multiple Operating Points and Application to Improve Industrial Flare Monitoring. Processes. 2017; 5(2):28. https://doi.org/10.3390/pr5020028

Chicago/Turabian Style

Xu, Shu, Bo Lu, Noel Bell, and Mark Nixon. 2017. "Outlier Detection in Dynamic Systems with Multiple Operating Points and Application to Improve Industrial Flare Monitoring" Processes 5, no. 2: 28. https://doi.org/10.3390/pr5020028

APA Style

Xu, S., Lu, B., Bell, N., & Nixon, M. (2017). Outlier Detection in Dynamic Systems with Multiple Operating Points and Application to Improve Industrial Flare Monitoring. Processes, 5(2), 28. https://doi.org/10.3390/pr5020028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop