Next Article in Journal
A Revised Hilbert–Huang Transform and Its Application to Fault Diagnosis in a Rotor System
Previous Article in Journal
Experimental Phantom-Based Security Analysis for Next-Generation Leadless Cardiac Pacemakers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Outlier Detection in Wireless Sensor Networks Using Model Selection-Based Support Vector Data Descriptions

1
School of Information Science & Engineering, Changzhou University, Changzhou 213164, China
2
School of IoT Enginering, Jiangnan University, Wuxi 214122, China
3
Research Center of IoT Technology Application Engineering (MOE), Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(12), 4328; https://doi.org/10.3390/s18124328
Submission received: 12 October 2018 / Revised: 18 November 2018 / Accepted: 22 November 2018 / Published: 7 December 2018
(This article belongs to the Section Sensor Networks)

Abstract

:
Wireless sensor networks (WSNs) are often deployed in harsh and unattended environments, which may cause the generation of abnormal or low quality data. The inaccurate and unreliable sensor data may increase generation of false alarms and erroneous decisions, so it’s very important to detect outliers in sensor data efficiently and accurately to ensure sound scientific decision-making. In this paper, an outlier detection algorithm (TSVDD) using model selection-based support vector data description (SVDD) is proposed. Firstly, the Toeplitz matrix random feature mapping is used to reduce the time and space complexity of outlier detection. Secondly, a novel model selection strategy is realized to keep the algorithm stable under the low feature dimensions, this strategy can select a relatively optimal decision model and avoid both under-fitting and overfitting phenomena. The simulation results on SensorScope and IBRL datasets demonstrate that, TSVDD achieves higher accuracy and lower time complexity for outlier detection in WSNs compared with existing methods.

1. Introduction

WSNs are composed of a large number of sensor nodes, which are equipped with wireless transceivers, low-power microcontrollers, energy sources and various sensors [1]. A wide variety of applications of WSNs can be found, ranging from personal spaces to the scientific, industrial, business, and military domains. However, sensor observations collected from sensor nodes often have low data quality and reliability due to the limited capability of sensor nodes in terms of energy, memory, computational power, bandwidth, and the harshness of the deployment environment [2,3]. The use of low quality sensor data in any data analysis and decision-making process limits the possibilities for reliable and real-time situation-awareness. A solution to ensure the quality of sensor data is outlier detection. In addition, outlier detection can help diagnose the health condition of WSNs and identify the environmental events (such as forest fire, air pollution, etc.) [4,5,6]. Therefore, it is very important to find an effective and efficient outlier detection technique for WSNs, which should be able to identify outliers with high accuracy and a low false alarm rate, while satisfying the constraints in terms of memory and computational complexity [7].
In WSNs, outliers can be defined as ‘those measurements that significantly deviate from the normal pattern of sensed data’ [8]. In recent years, researchers have shown increased interest in applying machine learning approaches for outlier detection in WSNs. Zhang [9] classified outlier detection techniques into statistical-based, nearest neighbor-based, clustering-based, classification-based, spectral decomposition-based and other types. SVDD is a one-class classification technique, its main idea is to find a minimum hyper-sphere such that all or most acceptable data samples are enclosed in the hyper-sphere [10,11]. The boundary of the hyper-sphere is the decision boundary, which can be used to identify outlier data. To reduce the high computational complexity of SVDD, Platt [12] and Fan [13] proposed sequential minimal optimization (SMO) algorithms to calculate the quadratic optimization required in the SVDD algorithm, effectively reducing the time complexity to O(n2). Chang [14] proposed an SMO optimization method for resolving dual quadratic optimization problems by using decomposition methods. Liu [15] and Feng [16] proposed to directly find a hyper-sphere or hyper-ellipsoid preimage of the feature vector, and then used a simple relationship between this feature vector and the SVDD sphere center to re-express the center with a single vector. Although the above-mentioned methods can reduce the computational complexity of the SVDD, none of them solves the problem of excessive calculation required for kernel functions. In 2007, Rahimi [17,18] mapped the input data to a randomized low-dimensional feature space and then applied existing fast linear learning methods. A significant speed up can be achieved by computing random features. However, it needs to maintain a high accuracy when the feature space dimension is high enough. Sutherland [19] improved the uniform error bound of paper [17], as well as gave a novel understanding of the embedding’s variance, approximation error, and use in some machine learning methods. Aman [20] extends the randomized-feature approach to the task of learning a kernel (via its associated random features), and presents an efficient optimization problem that learns a kernel in a supervised manner. They proved the consistency of the estimated kernel as well as generalization bounds for the class of estimators induced by the optimized kernel. Andrea [21] proposed an explicit description of the reproducing kernel Hilbert space (RKHS) induced by the approximated Gaussian kernel. They demonstrated that the approximations had indistinguishable performance from the full kernels yet greatly reduce the train/test times of SVMs.
The majority of existing outlier detection methods for WSNs does not take into account multi- dimensional data and assume the sensor data is univariate. They ignore the fact that the attributes together can display anomaly while in some cases none of the attributes individually has an anomalous value. However, each sensor node may be equipped with multiple sensors and also certain correlations may exist among attributes of sensor data. The attributes together can display anomaly while in some cases none of the attributes individually has an anomalous value. Thus, outlier detection techniques for WSNs should be able to analyze multi-dimensional data and identify whether the attributes together display anomaly [9].
For outlier detection in WSNs, it is necessary to maintain a high accuracy at a low feature space dimension. To meet this goal, this paper proposes an outlier detection method (TSVDD) using model selection-based SVDD, and this method can analyze multi-dimensional sensor data and improve the accuracy of outlier detection. The novelty of this method lies in the model selection strategy, which can help select a relatively optimal decision model and avoid both under-fitting and overfitting phenomena. The remainder of the paper is organized as follows: First we introduce the basic idea of SVDD and the random Fourier feature in Section 2. In Section 3 we discuss the proposed outlier detection method (TSVDD). Experimental results will be shown in Section 4, and we provide our conclusions in Section 5.

2. Support Vector Data Description and Random Fourier Feature

In this section, we introduce the principle behind the SVDD algorithm and the random Fourier feature, which are related to our research.

2.1. Support Vector Data Description

Given a target training set with n data objects, the main idea of SVDD [10,11] is to find a minimum-volume sphere with center a and radius R such that all or most of the target training data can be enclosed by the sphere. The optimization problem can be formulated as:
min R R 2 + C i = 1 n ξ i s . t . ( x i a ) ( x i a ) T R 2 + ξ i
where ξi are the slack variables, C is the penalty weight, which gives the trade-off between the two error terms: volume of the sphere and the number of target objects rejected. The geometric model of SVDD is shown in Figure 1. The black points in Figure 1 are the data objects.
Incorporating the constraints in Equation (1), we construct the Lagrangian:
L ( R , a , α i , ξ i ) = R 2 + C i = 1 n ξ i i = 1 n α i ( R 2 + ξ i ( x i 2 2 a x i + a 2 ) ) i = 1 n γ i ξ i
Using Lagrange multipliers αi ≥ 0 and γi ≥ 0, setting L / R = 0 , L / a = 0 , and L / i = 0 , and then substituting the results back into Equation (2), we obtain the dual problem:
L = i = 1 n α i ( x i x i ) i = 1 n j = 1 n α i α j ( x i x j )
However, this method only has good performance for data sets whose input space is spherically distributed. To find a more flexible method, the data objects can be transformed into a higher dimensional feature space. The inner products in Equation (3) can be replaced by a kernel function K (xi, xj). Here, we select the radial basis function as the kernel function:
K ( x i , x j ) = ϕ ( x i ) ϕ ( x j ) = exp ( ( x i x j ) 2 2 δ 2 )
For this Gaussian kernel function K (xi, xi) ≡ 1, hence Equation (3) can be transformed as:
W = max α i = 1 n j = 1 n α i α j K ( x i , x j ) 1 s . t . i = 1 n α i = 1 ,   0 α i C ( i = 1 ,   2 , , n )
Equation (5) is a typical quadratic optimization problem, the target training data can be classified into the three categories: (i) αi = 0, the data that are inside the sphere; (ii) 0 < αi < C, the data that are on the boundary of the sphere; and (iii) αi = C, the data that are outside the sphere. Hence, we have the SVDD decision function:
f ( x i ) = sgn ( | | ϕ ( x i ) a | | 2 R 2 )
if f(xi) = −1, xi is accepted as a target data; otherwise xi is labeled as an outlier data.

2.2 Random Fourier Feature

The first set of random features consists of random Fourier bases cos(ωx + b) where ωRD and bR are random variables. These mappings first project data points on a randomly chosen line, and then pass the resulting scalar through a sinusoidal function (see Figure 2). Drawing the direction of these lines from an appropriate distribution guarantees that the product of two transformed points will approximate a desired shift-invariant kernel [17].
For example, in Figure 2, each component of the feature map z(x) projects x and y onto a random direction ω drawn from the Fourier transform p(ω) of k(Δ), and wraps this line onto the unit circle in R2. After transforming two points x and y in this way, their inner product is an unbiased estimator of k(x, y) [17]. The map z(x) = cos(ω’x + b) additionally rotates this circle by a random amount b and projects the points onto the interval [0, 1].
Theorem 1.
(Bochner [22]) A continuous kernel k(x,y) = k(x − y) on RD is positive definite if and only if k(δ) is the Fourier transform of a non-negative measure. If k(δ) is properly scaled, Bochner’s theorem guarantees that its Fourier transform p(w) is a proper probability distribution [17].
Defining ς w ( x ) = e j w x , we have:
k ( x y ) = R d p ( w ) e j w ( x y ) d w = E w [ ς w ( x ) ς w ( y ) * ]
From Equation (8), we know that ς w ( x ) w ( y ) * is an unbiased estimator of k(x,y) when w is drawn from p(w). We can lower the variance of ς w ( x ) ς w ( y ) * by concatenating D randomly chosen ςw into a column vector ς and normalizing each component by D . The inner product of points characterized by the 2D-dimensional random feature ς, ς w ( x ) ς w ( y ) * = 1 D j = 1 D ς w j ( x ) ς w j ( y ) * is a sample average of ς w j ( x ) ς w j ( y ) , and is therefore a lower variance approximation to the expectation (8) [17].

3. Outlier Detection Algorithm Using Model Selection Based Support Vector Data Description

SVDD is an excellent one-class classification algorithm. However, kernel function calculation is complex. This paper proposes an outlier detection algorithm using model selection-based SVDD (TSVDD), which can reduce the computational complexity, while maintaining high detection accuracy. TSVDD consists of random Fourier feature mapping and model selection.

3.1. Toeplitz Random Fourier Feature Mapping in Support Vector Data Description (TRFF)

In WSNs, the distribution of sensor data is often irregular. The linear SVDD algorithm is not suitable for outlier detection. Therefore, this paper chooses the kernel-based SVDD, and the radial basis function is used as the kernel function. Considering the resource limitation of sensor nodes, we apply the Toeplitz random Fourier feature mapping to reduce the computational complexity of kernel function-based SVDD algorithm (kernel_SVDD).
Definition 1.
A Toeplitz matrix is a matrix in which each descending diagonal from left to right is constant. For instance, matrix T given in Equation (9) is a Toeplitz matrix.
T = [ a b c d e a b c f e a b ]
A circulant matrix is a special kind of Toeplitz matrix, where each row vector is rotated one element to the right relative to the preceding row vector. Equation (10) shows an instance:
T = [ a f e e a f f e a ]
Lemma 1.
Suppose the D-dimensional vector T(1)~N(0, ID2), and the circulant matrix TD is constructed by Toeplitz transform using T(1). Then TD satisfies the following condition:
E ( φ ( x i ) φ ( x j ) ) = k ( x i , x j )
where, φ ( x i ) = 1 D e i T D x i , k ( x i , y j ) = e x p ( x i y j 2 2 δ 2 ) .
Therefore, the algorithm only needs to store the first column vector so that we can reconstruct the whole matrix, and the space complexity is only O(n), so we use the Toeplitz random matrix to substitute the random Fourier feature matrix W for random feature mapping, and reduce the computational complexity of the SVDD algorithm. Actually, SVDD typically suffers from cubic complexity since it needs to solve convex quadratic programming problems, and it’s very difficult to be used directly in resource-constrained WSNs. Using the Toeplitz random Fourier feature to approximate the radial basis function, we propose the Toeplitz random Fourier feature SVDD algorithm (TRFF), which is described as follows.
Step 1:
Initialize the radial basis function parameter δ and the feature dimension D.
Step 2:
Draw samples T(1) from N (0, ID/δ2);
Step 3:
Use the Toeplitz transformation to obtain the D-dimensional matrix TD;
Step 4:
Compute the approximate radial basis function KM_RFF by Equation(11);
Step 5:
Solve the QP problem using the SMO algorithm for KM_RFF;
Step 6:
Construct the decision function f ( x ) = s g n ( e i T D a 2 R 2 ) of the TRFF algorithm.

3.2. Model Selection

The traditional random feature mapping algorithm has poor stability in low dimensional feature spaces. Thus, the phenomenon of over-fitting or under-fitting often happens in the decision model. In this study, we combine the model selection strategy with the TRFF algorithm to avoid the over-fitting and under-fitting in the SVDD.
Claim 1.
Uniform convergence of Fourier features [17]:
P [ ( ς w ( x ) ς w ( y ) * k ( x , y ) ) > ε ] exp ( 2 ε 2 D 2 )
Proof. 
Suppose x1, x2, …, xn are independent random variables, a i x i b i , x ¯ = x 1 + x 2 + + x n n , then for any δ > 0, Hoeffding’s inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount: P { X ¯ E ( X ) δ } exp ( 2 δ 2 N 2 ) . Since ς w ( x ) ς w ( y ) * is an unbiased estimator of k(x,y), and E [ ς w ( x ) ς w ( y ) * ] = k ( x , y ) , we can conclude that P [ ( ς w ( x ) ς w ( y ) * k ( x , y ) ) > ε ] exp ( 2 ε 2 D 2 ) . □
If the error between the radial basis function and its unbiased estimator is ε = 0.1, and the random feature mapping dimension is D = 10, then P [ ( ς w ( x ) ς w ( y ) * k ( x , y ) ) > 0.1 ] 0.1353 , hence the confidence level is 86.47%. If the error between the radial basis function and its unbiased estimator is ε = 0.01, and the random feature mapping dimension is D = 10, then:
P [ ( ς w ( x ) ς w ( y ) * k ( x , y ) ) > 0.01 ] 0.9801
Hence, the confidence level is 1.99%. Given the error bound ε and random feature dimension D, the approximate random Fourier feature mapping with a difference from k(x,y) less than ε can be found at the confidence level. This proves that when the random feature dimension D is low, there is a relatively optimal unbiased estimator that approximates the kernel function. Therefore, the goal of model selection is to select a relatively optimal model.

3.2.1 Under-Fitting Error

Definition 2.
An under-fitted model is a model that cannot adequately capture the underlying structure of the data, where some parameters or terms that would appear in a correctly specified model are missing. The under-fitting model is shown as follow in Figure 3b.
Figure 3 shows the decision model trained by the two algorithms under the random data set with 181 data objects generated by the Gaussian function N(0,1), where, all the data are 2D-dimentional, 160 data objects are normal and 21 data objects are outliers. In addition, the dimension of the random feature space is D = 10. Figure 3a shows the trained decision model of the kernel_SVDD algorithm. The support vector points accurately reflect the training dataset region. The contour line is the hyper-sphere in the feature space. Figure 3b is the trained decision model of TRFF algorithm under the random feature dimension D = 10. There is a mismatch between the model and the training dataset region, which does not reflect the characteristics of the region where the training dataset is located. If this model is used for outlier detection, it will inevitably lead to false positive alarms. Contrasting Figure 3a,b, there exists a big difference between the support vectors of the kernel_SVDD and the TRFF decision model.

3.2.2 Over-Fitting Error

Definition 3.
The over-fitted model: The production of a model that corresponds too closely or exactly to a particular set of data, and may therefore fail to fit additional data or predict future observations reliably. The over-fitting model is shown in Figure 4b.
Figure 4 shows the decision models trained by the two algorithms using the same random data set used in Figure 3. Figure 4a gives the decision model of kernel_SVDD, and Figure 4b shows the decision model of TRFF trained under the random feature mapping when the dimensionality is D = 10. Its outer boundary is basically the same as that of the kernel_SVDD decision model. Hence, it correctly displays the training dataset region. The outer support vector point is also basically consistent with the kernel_SVDD. However, there have some internal support vector points shown in Figure 4b, which reflect that the data samples surrounding the internal support vector points are outliers, so this model will also lead to false negative alarms during the detection process as the under-fitting model. Model selection needs to avoid the internal support vector points.
To design an optimal decision model, we should avoid both over-fitting and under-fitting. Algorithm 1 gives the pseudocode of the model selection strategy for kernel_SVDD, where T(1) is a column vector drawn from a Gaussian distribution, TD is a D-dimensional random feature matrix from the Toeplitz transformation, and error_underτ is the given threshold of under-fitting error.
Algorithm 1 Model selection for Kernel_SVDD
Input: Training dataset Train = {x1, x2, …, xn}
   Support vector SVS of kernel_SVDD
Process:
1: while (1) do
2: Sample T(1)~N(0, ID/δ2);
3: Apply the Toeplitz transformation of T(1) to form a D-dimensional feature matrix TD;
4: Train the training set Train to obtain decision model TRFF_f using TRFF algorithm;
5: Calculate the over-fitting error: error_over
6: if error_over = 0
7:   Calculate the under-fitting error error_under;
8:   if error_under < error_underτ
9:    break;
10:   else
11:    continue;
12:   end if;
13: else
14:  continue;
15: end if;
16: end while;
Output: Random feature matrix of optimal model TD
Once the optimal model is derived, we can detect the outliers using this optimal decision model and SVDD. The proposed outlier detection algorithm (TSVDD) is shown in Algorithm 2.
Algorithm 2 TSVDD algorithm
Input: Training dataset Train
   Testing dataset: Test = {x1, x2, …, xn}
Process:
1: Derive the decision model f of SVDD use training dataset;
2: While (Testφ) do
3:  if f(xi) > 0, (i = 1, 2, …, n)
4:   xi is marked as an outlier;
5:  else
6:   xi is marked as an adequate sample;
7:  end if;
8: end while;
Output: the outlier set

4. Experimental Results

To evaluate the performance of the proposed outlier detection algorithm, we carried out several simulation experiments on two WSN data sets, and compared the results of TSVDD with those of FastFood [23], RFF [17] and the traditional kernel_SVDD. All algorithms are implemented using Matlab 2014a, in a PC with equipped with an Intel (R) corei3 dual-core 3.6 GHz CPU, 4 G memory, and the Windows 7 operating system.

4.1. Data Sets

4.1.1. IBRL Dataset

The IBRL dataset was collected from 54 sensors deployed in the Intel Berkeley Research lab between February 28th and April 5th, 2004. Mica2Dot sensors with weather boards collected time stamped topology information, along with humidity, temperature, light and voltage values once every 31 s. Data was collected using the TinyDB in-network query processing system, built on the TinyOS platform. Considering the data integrity and continuity, we chose the humidity data and temperature data from Node 51 as the experimental dataset. We use IBRL_51 to represent this dataset. It is notable that there are some missing epochs in IBRL.

4.1.2. SensorScope Dataset

The SensorScope system [24] was deployed at the Grand-St-Bernard pass at 2400 m between Switzerland and Italy. The SensorScope dataset was collected from 13 September 2007 to 26 October 2007 and includes ambient temperature, surface temperature and relative humidity data. The sampling period was 2 min. Similarly, considering the data integrity and continuity, we selected the ambient temperature data and surface temperature data from Node 12 as the experimental dataset. We use SS_12 to represent this data set. Similarly, there are some missing data in SS_12.
Table 1 lists the abovementioned datasets used for our experiments. To evaluate the robustness of the proposed algorithm, we chose three sub-datasets (numbered 1, 2, 3, respectively) from different periods for each sensor. For example, IBRL_51-1, IBRL_51-2 and IBRL_51-3 consisted of the sensor data of 2 days, 4 days and 5 days, respectively, SS_12_1, SS_12_2, and SS_12_3 consisted of the sensor data of 3 days, 6 days and 9 days, respectively, and then we divided them into training sets and testing sets with almost the same ratio. By preliminarily statistical analysis, we found that the abnormal data account for 8% to15% of the testing datasets.

4.2. Performance Metrics

For the one-class classification problem, the samples can be classified into four types: true positive (TP), false positive (FP), true negative (TN) and false negative (FN), according to the combination of their true category and decided category by the algorithms, which can be described by a confusion matrix [20], as shown in Table 2.
The following four metrics can be calculated from the confusion matrix: True Positive Rate (TPR), True Negative Rate (TNR), False Positive Rate (FPR), and False Negative Rate (FNR). The calculation formulas are as follows:
T P R = T P T P + F N
F P R = F P F P + T N
Since TPR + FNR = 1, TNR + FPR = 1, we only use TPR and FPR as the performance evaluation metrics in this paper.

4.3 Performance Comparison Among Different Outlier Detection Algorithms

In our experiments, we set the radial basis function parameter δ = 0.8, and the regularization parameter is C = 1. Considering the randomness, we finished 100 runs of each algorithm for outlier detection, and then compared the average results. Table 3 shows the average and standard deviations of TPR and FPR of 100 experiments for five different algorithms, where RFF_30 and RFF_300 represent the RFF algorithm [17] with random feature dimension D = 30 and D = 300 respectively, TSVDD is the proposed algorithm in this study, and FastFood is the algorithm given in [23]. Because the kernel_SVDD has no random feature, its experimental results are unchanged, then we used these results as the references for comparison. The bold number represents the smallest standard deviation under the current data set. The smaller standard deviation reflects the more stable algorithm. Table 3a,b give the results of under the SS_12 dataset. It can be seen that, for FastFood and RFF_30, their average values of TPR and FPR are relatively high compared with the other algorithms. For RFF_300 and TSVDD, the standard deviation is very low, but TPR and FPR are very close to those of kernel_SVDD algorithm. It shows that RFF_300 and TSVDD completed the random feature mapping to approximate the kernel function precisely. Table 3c,d show the results under IBRL_51 dataset. Comparatively, RFF_300 and TSVDD have lower standard deviation values, so their results are more stable. In the case of low random feature dimension, TSVDD has almost the same FPR and TPR as kernel_SVDD.
Figure 5 shows the training decision models of kernel_SVDD, linear_SVDD, FastFood, RFF and TSVDD on SS_12-3 dataset. Figure 5a gives the model trained by kernel_SVDD, which accurately describes the edge of the data. The models represented by Figure 5e,f are trained by RFF_300 and TSVDD, both models are similar with that of kernel_SVDD, and their data edges are basically accurate. From Figure 5c,d, it can be seen that the models trained by FastFood and RFF_30 are under-fitting models, which will inevitably lead to false positives.
Figure 6 presents the run time for outlier detection on the given six datasets, which does not include the model training time. Here, the data size is the sample number of the testing dataset. From Figure 6, we can see that the run time of TSVDD on each dataset is not more than that of the other algorithms. For large datasets, the run times of kernel_SVDD and RFF_30 are longer than the other algorithms. For IBRL_51-3 dataset, its data size is 9694, and the run time of kernel_SVDD is almost three times that of TSVDD. When the data size increases, the time difference between RFF_300 and TSVDD will become increasingly greater. Due to the higher dimensionality of RFF_300 random feature mapping, the dataset size has a greater impact on its run time, so TSVDD has the highest efficiency while maintaining a high detection accuracy compared with the other algorithms.
The Receiver Operating Characteristic (ROC) curve is often applied to judge the performance of outlier detection algorithms, the area under the ROC curve is called Area under ROC Curve (AUC). Figure 7 shows the ROC curves of kernel_SVDD, linear_SVDD, FastFood, RFF, and TSVDD for the IBRL_51-3 dataset. From Figure 7 it can be seen that TPR of TSVDD is better than FastFood and RFF for the same random feature dimensions. Meanwhile, the TPR values of TSVDD and kernel_SVDD are almost equal to 1, so TSVDD is very accurate for outlier detection. From our experiments, we found that the ROC curves have the same trends for all other datasets. Generally, in low random feature dimensional cases, TSVDD has a higher TPR and a lower FPR, and it can remain maintain a stable performance for different datasets. Compared with kernel_SVDD and high dimensional random feature mapping algorithms, TSVDD has higher efficiency.

5. Conclusions

This study proposes an outlier detection algorithm (TSVDD) for wireless sensor networks. TSVDD aims to solve two issues of traditional SVDD algorithms. The first issue is the high computational cost of radial basis function calculations. To reduce the computational complexity, a Toeplitz random feature mapping with circulant matrix projection is used for approximating the radial basis function. The second issue is the poor stability of the traditional random feature mapping in low dimension feature space. A model selection strategy for avoiding over-fitting and under-fitting errors is proposed to guarantee the stability at low random feature dimensions. Simulation results on different datasets show that TSVDD has higher detection accuracy, a lower false alarm rate and higher efficiency than other traditional algorithms.

Author Contributions

Conceptualization, Z.H. and C.W.; Methodology, G.L.; Software, C.W.; Validation, G.L. and Z.H.; Investigation, C.W.; Writing-Original Draft Preparation, Z.H. and C.W.; Writing-Review & Editing, G.L. and Z.H.; Supervision, G.L.

Funding

This study was supported in part by the National Natural Science Foundation of China (Grant No. 61472368 and 61772248), the Key Research and Development Project of Jiangsu Province (Grant No. BE2016627), the Fundamental Research Funds for the Central Universities (Grant No. RP51635B), and the International Cooperative Science and Technology R &D Project of Wuxi City (Grant No. CZE02H1706).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tomić, I.; Mccann, J.A. A Survey of Potential Security Issues in Existing Wireless Sensor Network Protocols. IEEE Int. Things J. 2017, 4, 1910–1923. [Google Scholar] [CrossRef] [Green Version]
  2. Rawat, P.; Singh, K.D.; Chaouchi, H.; Bonnin, J.M. Wireless sensor networks: a survey on recent developments and potential synergies. J. Supercomput. 2014, 68, 1–48. [Google Scholar] [CrossRef]
  3. Zhang, Z.; Mehmood, A.; Shu, L.; Huo, Z.; Zhang, Y.; Mukherjee, M. A survey on Fault Diagnosis in Wireless Sensor Networks. IEEE Access 2016, 6, 11349–11364. [Google Scholar] [CrossRef]
  4. Qiu, T.; Qiao, R.; Wu, D.O. EABS: An Event-Aware Backpressure Scheduling Scheme for Emergency Internet of Things. IEEE Trans. Mob. Comput. 2018, 17, 72–84. [Google Scholar] [CrossRef]
  5. Qiu, T.; Chen, N.; Li, K.; Atiquzzaman, M.; Zhao, W. How can heterogeneous Internet of Things build our future: A survey. IEEE Commun. Surv. Tutorials 2018, 20, 2011–2027. [Google Scholar] [CrossRef]
  6. Qiu, T.; Zheng, K.; Song, H.; Han, M.; Kantarci, B. A local-optimization emergency scheduling scheme with self-recovery for smart grid. IEEE Trans. Ind. Inf. 2017, 13, 3195–3205. [Google Scholar] [CrossRef]
  7. Dhurgadevi, M.; Devi, P.M. An Analysis of Energy Efficiency Improvement through Wireless Energy Transfer in Wireless Sensor Network. Wirel. Pers. Commun. 2018, 98, 3377–3391. [Google Scholar] [CrossRef]
  8. Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. ACM Comput. Surv. 2009, 41, 1–58. [Google Scholar] [CrossRef]
  9. Zhang, Y.; Meratnia, N.; Havinga, P. Outlier Detection Techniques for Wireless Sensor Networks: A Survey. IEEE Commun. Surv. Tutor. 2010, 12, 159–170. [Google Scholar] [CrossRef] [Green Version]
  10. Tax, D.M.J.; Duin, R.P.W. Support vector domain description. Pattern Recognit. Lett. 1999, 11, 1191–1199. [Google Scholar] [CrossRef]
  11. Tax, D.M.J.; Duin, R.P.W. Support Vector Data Description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef] [Green Version]
  12. Platt, J.C. Fast Training of Support Vector Machines Using Sequential Minimal Optimization; MIT Press: Cambridge, MA, USA, 1999; pp. 185–208. [Google Scholar]
  13. Fan, R.E.; Chen, P.H.; Lin, C.J. Working Set Selection Using Second Order Information for Training Support Vector Machines. J. Mach. Learn. Res. 2005, 6, 1889–1918. [Google Scholar]
  14. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  15. Liu, Y.H.; Liu, Y.C.; Chen, Y.J. Fast support vector data descriptions for novelty detection. IEEE Trans. Neural Netw. 2010, 21, 1296–1313. [Google Scholar] [PubMed]
  16. Feng, Z.; Fu, J.; Du, D.; Li, F.; Sun, S. A new approach of anomaly detection in WSNs using support vector data description. Int. J. Distrib. Sens. Netw. 2017, 13, 1120–1123. [Google Scholar] [CrossRef]
  17. Rahimi, A.; Recht, B. Random features for large-scale kernel machines. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; pp. 1177–1184. [Google Scholar]
  18. Rahimi, A.; Recht, B. Weighted sums of random kitchen sinks: replacing minimization with randomization in learning. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–10 December 2008; pp. 1313–1320. [Google Scholar]
  19. Sutherland, D.J.; Schneider, J. On the error of random Fourier features. In Proceedings of the Conference on Uncertainty in Artificial Intelligence, Amsterdam, The Netherlands, 12–16 July 2015; pp. 862–871. [Google Scholar]
  20. Aman, S.; John, C.D. Learning kernels with random features. In Proceedings of the Neural Information Processing Systems (NIPS), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  21. Vedaldi, A.; Zisserman, A. Efficient Additive Kernels via Explicit Feature Maps. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 480–492. [Google Scholar] [CrossRef] [PubMed]
  22. Rudin, W. Fourier Analysis on Groups; Wiley-Interscience: New York, NY, USA, 1994. [Google Scholar]
  23. Le, Q.V.; Sarlos, T.; Smola, A.J. Fastfood: Approximate Kernel Expansions in Loglinear Time. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 244–525. [Google Scholar]
  24. Ingelrest, F.; Barrenetxea, G.; Schaefer, G.; Vetterli, M.; Couach, O.; Parlange, M. SensorScope: Application-Specific Sensor Network for Environmental Monitoring. ACM Trans. Sens. Netw. 2010, 6, 1–32. [Google Scholar] [CrossRef]
Figure 1. Geometry model of SVDD.
Figure 1. Geometry model of SVDD.
Sensors 18 04328 g001
Figure 2. Random Fourier Feature map.
Figure 2. Random Fourier Feature map.
Sensors 18 04328 g002
Figure 3. Under-fitting model (a) Model trained by kernel_SVDD; (b) Model trained by TRFF.
Figure 3. Under-fitting model (a) Model trained by kernel_SVDD; (b) Model trained by TRFF.
Sensors 18 04328 g003
Figure 4. Over-fitted model (a) Trained model of kernel_SVDD algorithm; (b) Trained model of TRFF algorithm.
Figure 4. Over-fitted model (a) Trained model of kernel_SVDD algorithm; (b) Trained model of TRFF algorithm.
Sensors 18 04328 g004
Figure 5. Decision models trained by different algorithm on SS_12-3 dataset (a) kernel_SVDD model; (b) linear_SVDD model; (c) FastFood model; (d) RFF_30 model; (e) RFF_300 model; (f) TSVDD model.
Figure 5. Decision models trained by different algorithm on SS_12-3 dataset (a) kernel_SVDD model; (b) linear_SVDD model; (c) FastFood model; (d) RFF_30 model; (e) RFF_300 model; (f) TSVDD model.
Sensors 18 04328 g005aSensors 18 04328 g005b
Figure 6. Comparison of run time.
Figure 6. Comparison of run time.
Sensors 18 04328 g006
Figure 7. ROC curve.
Figure 7. ROC curve.
Sensors 18 04328 g007
Table 1. Experimental datasets.
Table 1. Experimental datasets.
Dataset SourceData TypeDataset NumberSize of Training DatasetSize of Testing Dataset
SensorScopeTemperature & humiditySS_12-17171473
SS_12-214402877
SS-12-321574303
IBRLambient temperature & surface temperatureIBRL_51-118223562
IBRL_51-235625816
IBRL_51-350689694
Table 2. Confusion matrix of classification results.
Table 2. Confusion matrix of classification results.
True categoryDecision category
OutlierNormal
OutlierTPFN
NormalFPTN
Table 3. Performance comparison of various algorithms under different data sets.
Table 3. Performance comparison of various algorithms under different data sets.
(a) TPR comparison of SS_12 dataset.
Data SetKernel_SVDD (%)FastFoodRFF_30RFF_300TSVDD
Avg (%)stdAvg (%)stdAvg (%)stdAvg (%)std
SS_12-199.7999.770.002299.780.003399.710.002299.820.0013
SS_12-299.2197.840.014498.290.012998.680.008298.550.0112
SS_12-396.8296.870.019196.920.015397.320.006496.580.0086
(b) FPR comparison of SS_12 dataset
Data SetKernel_SVDD (%)FastFoodRFF_30RFF_300TSVDD
Avg (%)stdAvg (%)stdAvg (%)stdAvg (%)std
SS_12--110.8128.220.142727.300.142814.320.036116.760.0444
SS_12-20.001.620.01391.210.01220.120.00160.720.0031
SS_12-30.758.800.08687.260.08791.280.01041.460.0108
(c) TPR comparison of IBRL_51 dataset
Data SetKernel_SVDD (%)FastFoodRFF_30RFF_300TSVDD
Avg (%)StdAvg (%)StdAvg (%)StdAvg (%)Std
IBRL_51-199.7697.580.036497.420.032399.140.011699.280.0066
IBRL_51-299.4398.150.015297.920.015298.190.011597.960.0019
IBRL_51-399.7399.260.009299.490.006299.490.004499.370.0090
(d) FPR comparison of IBRL_51 dataset
Data SetKernel_SVDD (%)FastFoodRFF_30RFF_300TSVDD
Avg (%)StdAvg (%)StdAvg (%)StdAvg (%)Std
IBRL_51-10.553.610.04482.610.03191.890.01471.880.0236
IBRL_51-20.062.720.02242.270.02130.680.00850.410.0022
IBRL_51-30.254.780.03525.870.04111.280.01561.140.0046

Share and Cite

MDPI and ACS Style

Huan, Z.; Wei, C.; Li, G.-H. Outlier Detection in Wireless Sensor Networks Using Model Selection-Based Support Vector Data Descriptions. Sensors 2018, 18, 4328. https://doi.org/10.3390/s18124328

AMA Style

Huan Z, Wei C, Li G-H. Outlier Detection in Wireless Sensor Networks Using Model Selection-Based Support Vector Data Descriptions. Sensors. 2018; 18(12):4328. https://doi.org/10.3390/s18124328

Chicago/Turabian Style

Huan, Zhan, Chang Wei, and Guang-Hui Li. 2018. "Outlier Detection in Wireless Sensor Networks Using Model Selection-Based Support Vector Data Descriptions" Sensors 18, no. 12: 4328. https://doi.org/10.3390/s18124328

APA Style

Huan, Z., Wei, C., & Li, G. -H. (2018). Outlier Detection in Wireless Sensor Networks Using Model Selection-Based Support Vector Data Descriptions. Sensors, 18(12), 4328. https://doi.org/10.3390/s18124328

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop