Next Article in Journal
EEG-Based Multi-Modal Emotion Recognition using Bag of Deep Features: An Optimal Feature Selection Approach
Next Article in Special Issue
Blind Spoofing GNSS Constellation Detection Using a Multi-Antenna Snapshot Receiver
Previous Article in Journal
Off-Axis Cavity-Enhanced Absorption Spectroscopy of 14NH3 in Air Using a Gain-Switched Frequency Comb at 1.514 μm
Previous Article in Special Issue
An Empirical Study on V2X Enhanced Low-Cost GNSS Cooperative Positioning in Urban Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Impact of Different Kernel Functions on the Performance of Scintillation Detection Based on Support Vector Machines †

Departments of Electronics and Telecommunications, Politecnico di Torino, 10129 Torino, Italy
*
Author to whom correspondence should be addressed.
This paper is an extended version of “Comparative Performance Study of Linear and Gaussian Kernel SVM Implementations for Phase Scintillation Detection” published in the Proceedings of the 2019 9th International Conference on Localization and GNSS (ICL-GNSS), Nuremberg, Germany, 4–6 June 2019.
Sensors 2019, 19(23), 5219; https://doi.org/10.3390/s19235219
Submission received: 21 October 2019 / Revised: 22 November 2019 / Accepted: 25 November 2019 / Published: 28 November 2019

Abstract

:
Scintillation caused by the electron density irregularities in the ionospheric plasma leads to rapid fluctuations in the amplitude and phase of the Global Navigation Satellite Systems (GNSS) signals. Ionospheric scintillation severely degrades the performance of the GNSS receiver in the signal acquisition, tracking, and positioning. By utilizing the GNSS signals, detecting and monitoring the scintillation effects to decrease the effect of the disturbing signals have gained importance, and machine learning-based algorithms have been started to be applied for the detection. In this paper, the performance of Support Vector Machines (SVM) for scintillation detection is discussed. The effect of the different kernel functions, namely, linear, Gaussian, and polynomial, on the performance of the SVM algorithm is analyzed. Performance is statistically assessed in terms of probabilities of detection and false alarm of the scintillation event. Real GNSS signals that are affected by significant phase and amplitude scintillation effect, collected at the South African Antarctic research base SANAE IV and Hanoi, Vietnam have been used in this study. This paper questions how to select a suitable kernel function by analyzing the data preparation, cross-validation, and experimental test stages of the SVM-based process for scintillation detection. It has been observed that the overall accuracy of fine Gaussian SVM outperforms the linear, which has the lowest complexity and running time. Moreover, the third-order polynomial kernel provides improved performance compared to linear, coarse, and medium Gaussian kernel SVMs, but it comes with a cost of increased complexity and running time.

1. Introduction

Trans-ionospheric communication of the radio waves while traveling from transmitter to user is affected by the ionosphere that is highly variable and dynamic in both time and space [1]. The ionosphere, highly varied propagation medium, has an irregular structure due to plasma instabilities and scintillation is basically random fluctuations of the parameters of trans-ionospheric waves. Observation of the scintillation has been used in many research fields such as astronomy, geophysics, atmospheric physics, ocean acoustics, telecommunications, etc. so as to identify the irregular structure of the propagation medium [1]. Moreover, it has been found that the atmospheric structure has the basic characteristics of fluid turbulence in the equilibrium range, which is characterized by the well-known properties of Kolmogorov turbulence [2]. It has led to the development of the theoretical characterization of the scintillation effect [3,4] besides the observational point of view of the scintillation effect. However, although scintillation phenomena are well understood and widely exploited, it is difficult to find a definitive treatment of the theory of scintillation [2], and it is a fact of life for a number of communication and radar systems that have to operate through the auroral or equatorial ionosphere [5].
Global Navigation Satellite Systems (GNSS) signals also undergo severe propagation effects such as phase shifts and amplitude variations while they propagate through the Earth’s upper atmosphere, the ionosphere [6]. Ionospheric irregularities affect the GNSS signals in two ways, and both refractive and diffractive effects are gathered under the name of scintillation since they cause large-scale variations in both signal power and phase [7]. Ionospheric scintillation severely may degrade GNSS receiver performance by causing signal power loss and increasing measurement noise level, causing loss of lock of the tracking of GNSS signals [8,9]. In case of strong amplitude and phase scintillation events, it has been observed that even acquisition of the signals can be prevented [10].
Under a scintillation event, proper countermeasures could be undertaken at signal processing level, enabling either more robust signal acquisition and tracking or alternate resources to decrease the effect of disturbed signal propagations. Therefore, detecting and monitoring the scintillation effects in order to estimate the ionospheric scintillation in its early stages and measure the scintillation parameters gains importance. In this sense, GNSS signals provide an excellent means for measuring scintillation effects due to the fact that they are available all the time and can be acquired through many points of the ionosphere simultaneously [11]. With the coming new GNSS systems, a greater number of signals are available for monitoring signals and the advantages of new modernized signals and constellations for scintillation monitoring can be found in [12]. Moreover, an interested reader can find useful material about the design of monitoring stations in [6,13,14,15,16] and alternative receiver architectures for monitoring of ionospheric scintillations by means of GNSS signals in [17,18].
With the evolving of the artificial intelligence world, machine learning (ML) algorithms started to be applied also for scintillation detection. One of the proposed methods is based on a support vector machines (SVM) algorithm [19] that belongs to the class of supervised machine learning algorithms [20]. Supervised algorithms require large data sets to properly train the algorithm so that it is able to recognize the scintillation presence during the analysis of new measurements. In [21], another type of supervised learning algorithm, namely, the decision tree, is applied to enable early scintillation alerts. In [22], the performance of SVM implementations for phase and amplitude scintillation detection have been evaluated. The main weakness of these works is that the impact of the design parameters of the different algorithms on the achieved performance has not been carefully analyzed in detail as witnessed by the limited literature addressing the application of machine learning to GNSS. Since the success of the SVM algorithm can be attributed to the joint use of a robust classification procedure and of a versatile way of pre-processing the data, the parameters of the machine learning phase must be carefully chosen [23].
The SVM algorithm is the most widely used kernel learning algorithm [19]. The kernel method enables the SVM algorithm to find a hyperplane in the kernel space by mapping the data from the feature space into higher dimensional kernel space and leading to achievement of nonlinear separation in the kernel space [24]. Kernel representations offer an alternative solution to increase the computational power of the linear learning machines [25]. In SVM implementations, the kernel functions are linear, Gaussian radial basis function (RBF), and polynomial are widely used. Hence, the problem of choosing an architecture for an ML-based application is equivalent to the problem of choosing a suitable kernel for an SVM implementation [25]. Moreover, when training an SVM algorithm, besides choosing a suitable kernel function, a number of decisions should also be made in the preparation of the data, by labeling them, and setting the parameters of the SVM [26]. Otherwise, uninformed choices might result in degraded performance [27].
In this paper, we extend our previous work [28] which involves the performance comparison of linear and Gaussian kernels for phase scintillation detection versus the analysis of both amplitude and phase scintillation events. Linear and Gaussian kernels, the implementation of different order polynomial kernels, and the performance comparison on the cross-validation results are the original core of this paper. The impact of the kernel function on the scintillation detection performance by considering the related design parameters (e.g., scale parameter, polynomial order) is discussed. Performance is assessed by exploiting the receiver operating characteristics (ROC) curves, confusion matrix results and the related performance metrics. This study is performed using real GNSS signals that are affected by significant phase and amplitude scintillation effect, collected at the South African Antarctic research base (SANAE IV, 71 . 67 S, 2 . 84 W) and Vietnam (Hanoi, 21 . 00 N, 105 . 85 E).
The paper is organized as in the following. In Section 2, we introduce the scintillation measurements and the analysis of the real data. In Section 3, we provide an overview of the SVM algorithm. Then, implementation and performance analysis of the SVM algorithm with different kernel functions for scintillation detection and the comparison of the test results are discussed in Section 4. Finally, Section 5 draws conclusions.

2. Ionospheric Scintillation Data and Analysis

The amount of amplitude and phase scintillation that affect the GNSS signal can be monitored and measured by exploiting the signal tracking stage correlator outputs in the GNSS receiver. Specialized Ionospheric Scintillation Monitoring Receivers (ISMRs) or software defined radio (SDR) based receivers can be used for monitoring purposes [6,13,14,15,16,29].
Figure 1 shows an example of scintillation monitoring and data collection setup. A part of the data that are used throughout this paper were collected in the Antarctic station and in an Equatorial site by means of data grabbers of this kind. The data-collecting setup is a custom-designed solution based on a multi-constellation and multi-frequency GNSS data grabber and a software-defined radio receiver [15,30]. Data sets acquired in Antarctica grow constantly, and several levels of the data availability such as public access, limited access, and data license are offered [30].
Fourtune, which is the type of multiband data grabber shown in Figure 1, is a multi-frequency data collection unit, and it is able to perform medium-complexity signal processing (e.g., decimation, digital filtering, quantization, etc.). It has been developed by the researchers at Joint Research Center (JRC) of the European Commission in Italy and an interested user can find more information about its architecture in [16]. Scintillated data sets that are collected via Fourtune and used in this paper are summarized in Table 1. Pseudo-Random Noise (PRN) codes are ranging code components of the transmitted satellite signals and are unique for each satellite signal. In Table 1, they refer to the satellites from which the scintillation effect observed in the received signals.
Typically, there are two parameters to indicate the amount of scintillation effect. Amplitude scintillation is monitored by computing the S 4 index which corresponds to the standard deviation of the detrended signal intensity [15]. In an ISMR, S 4 metric is calculated as [16]
S 4 = I 2 T I T 2 I T 2 ,
where I is the detrended signal intensity, · T is the average operation over a fixed period T. Thus, as to compute the scintillation effect on the signal amplitude and to remove the variations due to other effects, detrending operations that commonly correspond to processing the signal intensity through cascaded low-pass filters are employed. On the other hand, phase scintillation monitoring is achieved by computing the σ ϕ index corresponding to the standard deviation of the detrended phase measurements [16]:
σ ϕ = φ 2 T φ T 2 ,
where φ is the detrended phase measurement that can be obtained by processing the carrier phase measurements through three cascaded second-order high pass filters or Butterworth high-pass filter [16]. Moreover, useful material in the design of filters for phase and amplitude detrending algorithms can be found in [11].
Figure 2a,b show an example of the computed amplitude and phase scintillation indices of GPS L1 signal that belongs to the data collected at Antarctic station on 21 January 2016. In Figure 2b, a sharp increase starting around 12:50 a.m. can be noticed, which indicates that the GPS signal broadcast from PRN-14 satellite experiences strong phase scintillation. On the other hand, in Figure 2a, there is no increase observed in the computed S 4 indices, which could be an indicator for amplitude scintillation. Due to polar location of SANAE IV, phase scintillation statistically occurs more often than amplitude scintillation [31]. Owing to the fact that the occurrence of the ionospheric scintillation depends on the several factors such as solar and geomagnetic activity, geographic location, the season of the year, and local time [15], it is not straightforward to model the occurrence of the event, and statistical analyses have been exploited in order to characterize the intensity, duration, and occurrence frequency of the scintillation events observed in different locations [31,32].
Figure 3a,b show an example of the computed amplitude and phase scintillation indices of GPS L1 signal that is broadcast from the PRN-20 satellite. It is seen that both amplitude and phase scintillations occur at the same time in this data that were collected in Hanoi on 16 April 2013. In the scintillation events observed at the equatorial region, typically both phase and amplitude scintillations occur with faster and deeper signal power fadings that take longer durations [32].
After having analyzed the scintillated data, the detection of the scintillation and the test results will be explained in the following sections.

3. Overview of Support Vector Machines Algorithm

An SVM algorithm classifies the data by finding the best hyper-plane that separates all the data of one class from those of the other class [33]. Figure 4 shows a pictorial example of two data sets that can be separable into two classes. However, as it can be seen in Figure 4a, there could be an infinite number of separating hyper-planes. The classes can be separated by the linear boundaries as well as nonlinear boundaries. SVM approaches this problem through the concept of the smallest distance between the decision boundary and any of the data samples [34]. In Figure 4b, as an example of linear classifiers, the best hyper-plane that corresponds to the one providing the largest margin between the classes is depicted. The margin is the maximum width of the slice, parallel to the hyper-plane that has no data samples within [33]. The data samples that are closest to the separating hyper-plane are called support vectors as shown in Figure 4b.
In SVM linear classification, the idea is to take the projection of an unknown vector x i along vector ω ¯ , which has to be perpendicular to the decision boundary medium (e.g., it has to be in the third dimension if the decision boundary is spanned in two dimensions), and to check whether it crosses the boundary or not in order to decide the classification. The implementation starts with the derivation of the optimal hyper-plane of SVM.

3.1. Derivation of the Optimum Hyper-Plane

With the given input data-set x 1 , x 2 , . . . , x N where x i R n , the two-class data classification problem using linear models are written in the following form [34]:
y i = f ( x i ) = ω 0 T x i + b 0 ,
where ω 0 T x corresponds to the projection of x over the direction spanned by ω 0 . ω 0 and b 0 are parameters of the optimum hyperplane [20]. Corresponding target values t 1 , t 2 , . . . , t N to the input values are decided according to y i :
c l a s s ( x i ) = C 1 , ω 0 T x i b 0 t i = + 1 , C 2 , ω 0 T x i < b 0 t i = 1 ,
where t i 1 , + 1 defines the class labels. If all the points are classified correctly, it leads to t i y i > 0 for all i. The distance of x i from the hyperplane is computed as [35]
d ( x i ) = f ( x i ) ω = t i ω 0 T x i + b 0 ω .
Up to this point, there are not enough constraints to fix specific values for b and ω . Therefore, in order to define a decision rule and have an optimum hyper-plane, the parameters should be optimized by considering the maximum width for the margin that can be obtained by solving [34]
arg max ω 0 , b 0 = 1 ω min i t i w T x i + b .
The direct optimization problem would be very complex and this constrained optimization problem is solved by using Lagrange Multipliers [34]. The cost function is written as
L = 1 2 ω 2 i = 1 N a i t i ω T x i + b 1 ,
where a = a 1 , a 2 , . . . a N T are Lagrangian multipliers with a i 0 . The solutions are saddle points of the cost function L. There is a minus sign before the multipliers because saddle points are obtained by minimizing with respect to ω and b and maximizing with respect to a [20,34]. The derivatives of the cost function with respect to ω and b are then set to zero:
L ω = 0 ω 0 = i = 1 N α i t i x i ,
L b = 0 i = 1 N α i t i = 0 .
Replacing Equations (8) and (9) in Equation (7) leads to the dual representation of the maximum margin problem [34]
L ˜ ( a ) = i = 1 N a i 1 2 i = 1 N j = 1 N a i a j t i t j κ ( x i , x j ) subject to
a i 0 , i = 1 , . . . , N ,
where κ ( x i , x j ) = x i T x j is kernel function. A necessary and sufficient condition for a function κ ( x i , x j ) to be a valid kernel, the function κ ( x i , x j ) should be positive semi-definite for all the possible choices, and this also ensures that the Lagrangian function L ˜ ( a ) is bounded [34]. The Lagrange multipliers are obtained by solving
a ^ = arg max a L ˜ ( a ) ,
with respect to a subject to the constraints
a ^ 0 , i = 1 , 2 , . . . , N ,
i = 1 N a ^ i t i = 0 ,
where all the points that have a ^ 0 are on the margin; in other words, they are support vectors [24]. For all the points that do not lie on the margin, the corresponding Lagrange multiplier is a ^ = 0 . Now, y i , defined by Equation (3), can be evaluated as
y i = f ( x i ) = j = 1 N a ^ j t j κ ( x i , x j ) + b .
The optimization of this form satisfies the following Karush–Kuhn–Tucker (KKT) conditions that are both necessary and sufficient for optimum SVM solution [33,34,35]:
α ^ i 0 , i , t i y i 1 0 , i , α ^ i t i y i 1 = 0 , i .
After obtaining the Lagrange multipliers a ^ by solving Equation (12), the optimal parameters w 0 and b 0 can be computed by considering the set of support vectors S because any data point for which a ^ = 0 plays no role in making predictions for the new data points. Having obtained the a ^ and support vector x i values that satisfy KKT conditions and Equation (15), the value of the threshold parameter b can be determined [34]
w 0 = i ϵ S a ^ i t i x i ,
and
b 0 = 1 N S i ϵ S t i j ϵ S a ^ j t j κ ( x i , x j ) ,
where N S is the total number of support vectors in the set S and averaging over all support vectors provides a numerically more stable solution for b 0 [34]. Eventually, the decision function of SVM is computed as
y ^ ( x i ) = sign j = 1 N s a ^ i t i κ ( x j , x i ) + b 0 .

3.2. Kernel Extension

A kernel is a function κ that for all x , y X satisfies [36]
κ x , y = ϕ ( x ) , ϕ ( y ) ,
where ϕ is a mapping from X to F κ that is an inner product feature space associated with the kernel κ :
ϕ : x X ϕ ( x ) = κ x , · F κ .
Any finite subset of the space X is positive semi-definite, and the kernel function satisfies a positive semi-definite condition as mentioned. Actually, corresponding space F κ is referred to as Reproducing Kernel Hilbert Space (RKHS), which is a Hilbert space containing the Cauchy sequence limit condition [36]. The theory of reproducing kernels was published by Aronszajn in 1950, and detailed theory can be found in [37]. Moreover, the kernel concept was introduced into the pattern recognition field by Aizerman in 1964 [34].
In most cases, the samples may not be linearly separable. As it is seen in Equation (10), linear kernel can be expressed as
κ ( x i , x j ) = x i T x j .
If the classification problem is not linearly separable, SVM can be powered up by a proper kernel function. The kernel method enables SVM to find a hyperplane in the kernel space and hence nonlinear separation can be achieved in that feature space [24]. An example of nonlinear kernels is Gaussian RBF, which can be written as
κ ( x i , x j ) = exp x i x j 2 2 σ 2 = exp γ x i x j 2 ,
where σ defines the width of the kernel. If the parameter σ is close to zero, SVM tends to over-fitting, which means all the training instances are used as support vectors [24]. Assigning a bigger value to σ may cause under-fitting, leading all the instances to be classified into one class. Therefore, a proper value must be selected for the kernel width. In the same manner, kernel scale parameter corresponds to γ parameter in the RBF definition as being different from the σ representation.
Another most commonly used kernel function is the polynomial that can be represented as
κ ( x i , x j ) = 1 + x i T x j p ,
where p is the order of the polynomial kernel. The lowest degree polynomial corresponds to the linear kernel, and it is not preferred in case of having a nonlinear relationship between the features. The degree of the polynomial kernel controls the flexibility of the classifier and higher-degree allows a more flexible decision boundary compared to linear boundaries [38].
The performance of linear, Gaussian, and polynomial kernels are compared in our implementation in the next section.

4. Experimental Tests

In this section of the paper, the implementation and performance analysis of the scintillation detection based on the SVM method that exploits different kernel functions are provided through the collected data.

4.1. Training Data Preparation and Labeling

The preparation of the data is the most important step in the machine-learning implementations. Amplitude and phase scintillation indices have to be put into a format so that the SVM algorithm can detect the scintillation in the correct way. As it is mentioned, it is difficult to model the occurrence of scintillation due to temporal and spatial variabilities of the ionosphere [32]. Statistical analysis has been highly benefited and also chosen fixed period (T) for the computations of S 4 and σ ϕ indices is quite important, given in Equations (1) and (2). Generally, T is adjusted to 60 s in ISMR receivers. In this case, by considering only one value to feed the algorithm, early detection of the scintillation seems not to be possible. Moreover, according to performed analysis of high-latitude and equatorial ionospheric scintillation events in [32], phase scintillation lasts around 5 . 6 min at high latitude regions and 10 . 2 min in the equatorial region. On the other hand, it has been observed that amplitude scintillation events last around 3 . 1 min at the high-latitude region and 12 . 4 min in the equatorial region. Therefore, so as to enable early scintillation detection, the training data are put into a format by partitioning the data into three-minutes blocks via a moving time window.
In this work, only two class labels, namely, scintillation and no-scintillation, are assigned as follows:
c l a s s ( x i ) = C 1 Scintillation , t i = 1 ( σ ϕ 0 . 3 or S 4 0 . 4 ) , C 2 No Scintillation , t i = 0 ( σ ϕ < 0 . 3 or S 4 < 0 . 4 ) .
The class definitions in Equation (25) have been decided according to the limit values of σ ϕ and S 4 indices given in [39,40] that are observed in high-latitude and equatorial scintillation events. The list of the training data segments is reported in Table 1. The data collected on different days at SANAE IV and Hanoi stations are put in the defined structure and labeled according to the class definition above. Figure 5a,b show an example of labeling for both amplitude and phase scintillation data sets.
In Figure 5a, S 4 indices computed from the received GPS L1 signal broadcast from the satellite PRN-11 are plotted. It belongs to the data collected on 10 April 2013 in Hanoi. The figure indicates that amplitude scintillation events occurred starting around 1:20 p.m.
Another event that is observed in the data collected on 21 January 2016 at the Antarctic station SANAE IV is analyzed, and the computed σ ϕ indices are plotted in Figure 5b. The figure shows the phase scintillation event occurred starting from 12:40 a.m. as denoted by the sharp increase in the indices. Although the σ ϕ values go to values lower than 0 . 3 around 12:50 a.m., such a time interval is still considered to be part of the scintillation event and then labeled accordingly. In fact, the data portion between the consecutive scintillation events is affected anyway by a residual scintillation effect that can be observed in the receiver tracking outputs and GNSS measurement observables. Therefore, labeling has been done manually, by inspection, for all the training data sets.

4.2. Cross Validation

After class labels are assigned to the data sets for amplitude and phase scintillation events, SVM methods with different kernel functions are trained. In this section, the performance of validation of the methods is evaluated in terms of ROC curves.
Figure 6 shows an example of the ROC graph. It is a two-dimensional plot of a classifier indexed in one dimension by the false positive rate (FPR) and in the other by the true positive rate (TPR). An ROC graph depicts relative trade-offs that a classifier makes between benefits (true positives) and costs (false positives) [41]. True positive rate (i.e., sensitivity) and the true negative rate (i.e., specificity) are the terms that split the predictive performance of the classifier into the proportion of positives and negatives correctly classified, respectively.
In ROC space, the point ( 0 , 1 ) represents the perfect classification as shown in Figure 6. The dashed diagonal line represents the case of random assignment of an element to a class. If one point is closer to the upper left corner (i.e., higher TPR, lower FPR), its classification performance is better than another. Classifiers appearing on the left side of the ROC space are called conservative such that they make classification with strong evidence so they have small FPR, but they generally have low TPR [42]. On the other hand, the classifiers on the right side of the ROC space are thought as liberal. They make classifications with low evidence so they classify all the positives correctly with a drawback of high FPR. Each point on the ROC curve represents a trade-off, in other words, a cost ratio [43]. Cost ratio is equal to the slope of the line tangent to the ROC curve at a given point.
SVM implementation for both phase and amplitude scintillation detection has been done by employing MATLAB’s statistics and machine learning toolbox (R2016b, The Mathworks, Inc., Natick, MA, USA) [44]. In order to evaluate the performance of the implemented SVM methods, a 10-fold cross-validation technique has been applied. In this technique, the partitions are put into 10 randomly chosen subsets of equal size. Then, each subset is used to validate the model by using the trained remaining nine subsets. This process is repeated 10 times so that each subset is used once for the validation.
Figure 7a,b show the ROC curves of SVM methods with different kernel functions for amplitude and phase scintillation, respectively. In general, it shows that, in an amplitude scintillation case, the performance of the SVM methods is better than the phase scintillation case. However, due to the fact that SVM performance depends on the data sets, we evaluate each scintillation case separately and compare the performance of the kernel functions under the same conditions.
In the analysis, Gaussian RBF kernel scale parameter γ in Equation (23) is adjusted to different values according to the following assumptions [45]:
γ f G = n / 4 , for fine Gaussian , γ m G = n , for medium Gaussian , γ c G = 4 n , for coarse Gaussian ,
where n is the number of features or the dimension size of x i in Equation (3). Moreover, second and third-order polynomial kernels are included in the analysis. The different colored plus symbols on the figures indicate the operating points of each method in Figure 7. Since the ROC curves are close to each other and it is not easy to differentiate the differences between each method, the results are summarized in Table 2 and Table 3.
In Table 2 and Table 3, we compare the time complexity values that are dedicated time to both training and testing. The complexity of a classifier is divided into two kinds of complexity, namely, time complexity and space complexity [46]. While time complexity deals with the time spent on the execution of the algorithm, space complexity considers the amount of memory used by the algorithm [46]. For example, run-time complexity of linear and RBF kernels differ from each other. While the complexity of the RBF kernel is shown to be O n S V × d , which is dependent on the number of support vectors ( n S V ) and the input dimension (d), linear method has O d prediction complexity [47]. Therefore, as it is seen in both Table 2 and Table 3, running time increases in the cases of the kernel functions in which the samples are uplifted into higher dimensions, and it also becomes dependent on the number of samples.
The parameter, Area Under Curve (AUC), represents the estimated area under the ROC curve, and it is used as a peformance measure for the machine learning algorithms [48]. AUC is accepted as an indicator for the overall accuracy of the classifier; both Table 2 and Table 3 show the importance of kernel scale parameter. While the overall accuracy of coarse and medium Gaussian kernel SVM methods are less than the linear SVM, fine Gaussian SVM outperforms the linear. Moreover, the third-order polynomial kernel provides improved performance compared to linear, coarse, and medium Gaussian kernel SVMs, but it comes with a cost of increased complexity and time.

4.3. Tests and Evaluation

In this section of the paper, as a performance cross-check, we also used the collected data which are not included in the training sets to evaluate the performance of scintillation detection methods. Figure 8a,b show the decisions of the different kernel SVM methods for the data sets for both amplitude and phase scintillation.
In order to evaluate the performance comparison, the confusion matrix technique, which is a two-dimensional matrix indexed in one dimension by the true class of an object and in the other by the class that the classifier assigns [49] as it is seen in Table 4, is applied. A confusion matrix represents the dispositions of a set of instances (i.e., test data) according to a defined classification model that maps the set of instances to predicted classes [42]. Actually, each different point in a ROC curve corresponds to a confusion matrix.
In the two-class data classification problem, the four cells of the confusion matrix correspond to the values of true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN). By considering the numbers of the measures, this matrix forms the basis for the terms, namely, accuracy, precision, sensitivity, specificity, and error rate. Accuracy is the ratio of the correct predictions (i.e., the sum of true positives and true negatives) to the total predictions made by the classifier. In the same sense, the error rate is the ratio of the incorrectly classified objects to the total objects. In Table 5, the performance of different kernels for the data sets in terms of accuracy and error rate are summarized.
As it is expected, the results in Table 5 are consistent with the cross-validation results. With the correct setting of the kernel scale parameter, the fine Gaussian SVM method outperforms in terms of accuracy rate. Furthermore, although the differences in the accuracy of the methods seem not to be at a considerable level, in terms of early detection, the method that provides higher accuracy gains importance. For example, in the cases of the computation rate of the scintillation indices being around one minute, the method having a higher accuracy rate will provide quite advantageous conditions in terms of early detection. However, the third order polynomial kernel provides an improvement in the accuracy compared to the linear kernel, but its performance should be evaluated with increased time and space complexity.

5. Conclusions

In this paper, we provided performance analysis comparing different kernel methods of SVM and we reviewed and analyzed the linear, Gaussian, and polynomial kernel SVM algorithms for both phase and amplitude scintillation detection. Performance comparison was assessed by exploiting the ROC curves, confusion matrix results, and the performance metrics associated with the confusion matrix. It has been observed that, if the kernel scale parameter of Gaussian RBF kernel SVM algorithm is optimized, the performance of the RBF kernel SVM method outperforms the linear kernel SVM method in terms of overall accuracy. Moreover, although third order polynomial kernel SVM performs better than the linear kernel, it comes with a cost of increased time and space complexity.
For a possible scintillation detection implementation that could be based on unsupervised algorithms in the future, the effect of the multipath has to be considered because the presence of the multipath causes an increase in the amplitude scintillation index values indicating scintillation alarm falsely.

Author Contributions

Conceptualization, F.D.; Formal analysis, C.S.; Funding acquisition, F.D.; Investigation, C.S.; Methodology, C.S. and F.D.; Supervision, F.D.; Writing—original draft, C.S.; Writing—review and editing, F.D.

Funding

This research work is undertaken under the framework of the TREASURE project (Grant Agreement No. 722023) funded by European Union’s Horizon 2020 Research and Innovation Programme within Marie Skłodowska-Curie Actions (MSCA) Innovative Training Network (ITN).

Acknowledgments

The authors thank the DemoGRAPE project for providing the recorded data collected in the Antarctic region. The authors also wish to thank the Joint Research Center (JRC) for providing the scintillated real data that were collected in Hanoi.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

GNSSGlobal Navigation Satellite Systems
SVMSupport Vector Machines
ROCReceiver Operating Characteristics
RBFRadial Basis Function
TPTrue Positive
FPFalse Positive
TNTrue Negative
FNFalse Negative
TPRTrue Positive Rate
FPRFalse Positive Rate
AUCArea Under the Curve

References

  1. Priyadarshi, S. A review of ionospheric scintillation models. Surv. Geophys. 2015, 36, 295–324. [Google Scholar] [CrossRef] [PubMed]
  2. Rino, C.L. The Theory of Scintillation with Applications in Remote Sensing; IEEE Press, Wiley, Inc. Publication: Hoboken, NJ, USA, 2011. [Google Scholar]
  3. Tatarskii, V.I. The Effects of the Turbulent Atmosphere on Wave Propagation; National Technical Information Service: Springfield, VA, USA, 1971. [Google Scholar]
  4. Yeh, K.C.; Liu, C.-H. Radio wave scintillation in the ionosphere. IEEE Proc. 1982, 70, 324–360. [Google Scholar]
  5. Crane, R.K. Ionospheric scintillation. IEEE Proc. 1977, 65, 180–199. [Google Scholar] [CrossRef]
  6. Cristodaro, C.; Dovis, F.; Linty, N.; Romero, R. Design of a configurable monitoring station for scintillations by means of a GNSS software radio receiver. IEEE Geosci. Remote. Sens. Lett. 2018, 15, 325–329. [Google Scholar] [CrossRef]
  7. Kintner, P.M.; Humphreys, T.E.; Hinks, J.C. GNSS and ionospheric scintillation – how to survive the next solar maximum. Inside GNSS 2009, 22–29. Available online: http://www.insidegnss.com (accessed on 15 September 2019).
  8. Savas, C.; Dovis, F.; Falco, G. Performance evaluation and comparison of GPS L5 acquisition methods under scintillations. In Proceedings of the 31st International Technical Meeting of The Satellite Division of the Institute of Navigation (ION GNSS+ 2018), Miami, FL, USA, 24–28 September 2018; pp. 3596–3610. [Google Scholar]
  9. Fugro Satellite Positioning—The Effect of Space Weather Phenomena on Precise GNSS Applications; White Paper; Fugro N.V.: Leidschendam, The Netherlands, 2014; Available online: https://fsp.support (accessed on 26 November 2019).
  10. Savas, C.; Falco, G.; Dovis, F. A comparative analysis of polar and equatorial scintillation effects on GPS L1 and L5 tracking loops. In Proceedings of the International Technical Meeting of the Institute of Navigation, Reston, VA, USA, 28–31 January 2019; pp. 632–646. [Google Scholar]
  11. Van Dierendonck, A.J.; Klobuchar, J.; Hua, Q. Ionospheric scintillation monitoring using commercial single frequency C/A code receivers. In Proceedings of the 6th International Technical Meeting of Satellite Division of the institute of Navigation (ION GPS 1993), Salt Lake City, UT, USA, 22–24 September 1993; pp. 1333–1342. [Google Scholar]
  12. Romero, R.; Linty, N.; Cristodaro, C.; Dovis, F.; Alfonsi, L. On the use and performance of new Galileo signals for ionospheric scintillation monitoring over Antarctica. In Proceedings of the International Technical Meeting of The Institute of Navigation, Monterey, CA, USA, 25–28 January 2016; pp. 989–997. [Google Scholar]
  13. Curran, J.T.; Bavaro, M.; Morrison, A.; Fortuny, J. Developing a multi-frequency for GNSS-based scintillation monitoring receiver. In Proceedings of the 27th International Technical Meeting of The Satellite Division of the Institute of Navigation (ION GNSS+ 2014), Tampa, FL, USA, 8–12 September 2014; pp. 1142–1152. [Google Scholar]
  14. Peng, S.; Morton, Y. A USRP2-based reconfigurable multi- constellation multi-frequency GNSS software receiver front end. GPS Solut. 2013, 17, 89–102. [Google Scholar] [CrossRef]
  15. Linty, N.; Dovis, F.; Alfonsi, L. Software-defined radio technology for GNSS scintillation analysis: Bring Antarctica to the lab. GPS Solut. 2018, 22, 96. [Google Scholar] [CrossRef]
  16. Curran, J.T.; Bavaro, M.; Fortuny, J. Developing an Ionospheric Scintillation Monitoring Receiver. Inside GNSS 2014, 60–72. Available online: http://www.insidegnss.com (accessed on 15 September 2019).
  17. Curran, J.T.; Bavaro, M.; Fortuny, G.J.J. An open-loop vector receiver architecture for GNSS-based scintillation monitoring. In Proceedings of the European Navigation Conference (ENC-GNSS), Rotterdam, The Netherlands, 15–17 April 2014; pp. 1–12. [Google Scholar]
  18. Linty, N.; Dovis, F. An open-loop receiver architecture for monitoring of ionospheric scintillations by means of GNSS signals. Appl. Sci. 2019, 9, 2482. [Google Scholar] [CrossRef]
  19. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  20. Jiao, Y.; Hall, J.J.; Morton, Y.T. Automatic equatorial GPS amplitude scintillation detection using a machine learning algorithm. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 405–418. [Google Scholar] [CrossRef]
  21. Linty, N.; Farasin, A.; Favenza, A.; Dovis, F. Detection of GNSS ionospheric scintillations based on machine learning decision tree. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 303–317. [Google Scholar] [CrossRef]
  22. Jiao, Y.; Hall, J.J.; Morton, Y.T. Performance evaluation of an automatic GPS ionospheric phase scintillation detector using a machine-learning algorithm. NAVIGATION: J. Inst. Navig. 2017, 64, 391–402. [Google Scholar] [CrossRef]
  23. Bousquet, O.; Herrmann, D.J.L. On the complexity of learning the kernel matrix. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 9–14 December 2002; pp. 399–406. [Google Scholar]
  24. Liu, Z.; Zuo, M.J.; Zhao, X.; Xu, H. An analytical approach to fast parameter selection of Gaussian RBF kernel for support vector machine. J. Inf. Sci. Eng. 2015, 31, 691–710. [Google Scholar]
  25. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other KERNEL-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
  26. Asa, B.H.; Weston, J. A user’s guide to support vector machines. Methods Mol. Biol. 2010, 609, 223–239. [Google Scholar]
  27. Hsu, C.-W.; Chang, C.-C.; Lin, C.-J. A Practical Guide to Support Vector Classification; Technical Report; Department of Computer Science, National Taiwan University: Taipei City, Taiwan, 2003. [Google Scholar]
  28. Savas, C.; Dovis, F. Comparative performance study of linear and gaussian kernel SVM implementations for phase scintillation detection. In Proceedings of the IEEE International Conference on Localization and GNSS (ICL-GNSS), Nuremberg, Germany, 4–6 June 2019; pp. 1–6. [Google Scholar]
  29. Linty, N.; Romero, R.; Dovis, F.; Alfonsi, L. Benefits of GNSS software receivers for ionospheric monitoring at high latitudes. In Proceedings of the 1st URSI Atlantic Radio Science Conference (URSI AT-RASC 2015), Las Palmas, Spain, 16–24 May 2015; pp. 1–6. [Google Scholar]
  30. Alfonsi, L.; Cilliers, P.J.; Romano, V.; Hunstad, I.; Correia, E.; Linty, N.; Dovis, F.; Terzo, O.; Ruiu, P.; Ward, J.; et al. First, observations of GNSS ionospheric scintillations from DemoGRAPE project. Space Weather 2016, 14, 704–709. [Google Scholar] [CrossRef]
  31. Jacobsen, K.S.; Dahnn, M. Statistics of ionospheric disturbances and their correlation with GNSS positioning errors at high latitudes. J. Space Weather Space Clim. 2014, 4, 1–10. [Google Scholar] [CrossRef]
  32. Jiao, Y.; Morton, Y.; Taylor, S. Comparative studies of high-latitude and equatorial ionospheric scintillation characteristics of GPS signals. In Proceedings of the IEEE/ION Position, Location and Navigation Symposium—PLANS, Monterey, CA, USA, 5–8 May 2014; pp. 37–42. [Google Scholar]
  33. Gutierrez, D.D. Machine Learning and Data Science: An Introduction to Statistical Learning Methods with R; Technics Publications: Basking Ridge, NJ, USA, 2015. [Google Scholar]
  34. Bishop, C.M. Pattern Recognition and Machine Learning (Information Science and Statistics); Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  35. Laface, P.; Cumani, S. Support Vector Machines; Class Lecture; Machine Learning for Pattern Recognition (01SCSIU), Politecnico di Torino: Turin, Italy, 19 March 2018. [Google Scholar]
  36. Shawe-Taylor, J.; Cristianini, N. Kernel Methods for Pattern Analysis; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  37. Aronszajn, N. Theory of reproducing kernels. Trans. Am. Math. Soc. 1950, 68, 337–404. [Google Scholar] [CrossRef]
  38. Asa, B.-H.; Ong, C.S.; Sonnenburg, S.; Scholkopf, B.; Ratsch, G. Support vector machines and kernels for computational biology. PLoS Comput. Biol. 2008, 4, e1000173. [Google Scholar]
  39. Kinrade, J.; Mitchell, C.N.; Yin, P.; Smith, N.; Jarvis, M.J.; Maxfield, D.J.; Rose, M.C.; Bust, G.S.; Weatherwax, A.T. Ionospheric scintillation over Antarctica during the storm of 5–6 April 2010. J. Geophys. Res. 2012, 117, A05304. [Google Scholar] [CrossRef]
  40. Moraes, A.O.; Costa, E.; De Paula, E.R.; Perrella, W.J.; Monico, J.F.G. Extended ionospheric amplitude scintillation model for GPS receivers. Radio Sci. 2014, 49, 315–329. [Google Scholar] [CrossRef]
  41. Provost, F.; Fawcett, T. Data Science for Business: What You Need to Know about Data Mining and Data-Analytic Thinking; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2013. [Google Scholar]
  42. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  43. Japkowicz, N.; Shah, M. Evaluating Learning Algorithms: A Classification Perspective; Cambridge University Press: New York, NY, USA, 2011. [Google Scholar]
  44. The Mathworks, Inc. MATLAB and Statistics and Machine Learning Toolbox R2016b; The Mathworks, Inc.: Natick, MA, USA, 2016. [Google Scholar]
  45. Wang, T. Development of the Machine Learning-Based Approach to Access Occupancy Information through Indoor CO2 Data. Master’s Thesis, Delft University of Technology, Delft, The Netherlands, 2017. Available online: http://repository.tudelft.nl/ (accessed on 15 September 2019).
  46. Abdiansah, A.; Wardoyo, R. Time complexity analysis of support vector machines (SVM) in LibSVM. Int. J. Comput. Appl. 2015, 128, 28–34. [Google Scholar] [CrossRef]
  47. Claesen, M.; De Smet, F.; Suykens, J.A.; De Moor, B. Fast prediction with SVM models containing RBF kernels. arXiv 2014, arXiv:1403.0736v3. [Google Scholar]
  48. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef] [Green Version]
  49. Ting, K.M. Confusion Matrix. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2011. [Google Scholar]
Figure 1. Experimental scintillated GNSS data collection configuration.
Figure 1. Experimental scintillated GNSS data collection configuration.
Sensors 19 05219 g001
Figure 2. Scintillation index values of GPS L1 C/A PRN-14 signal—21 January 2016 (SANAE IV).
Figure 2. Scintillation index values of GPS L1 C/A PRN-14 signal—21 January 2016 (SANAE IV).
Sensors 19 05219 g002
Figure 3. Scintillation index values of GPS L1 C/A PRN-20 signal—16 April 2013 (Hanoi).
Figure 3. Scintillation index values of GPS L1 C/A PRN-20 signal—16 April 2013 (Hanoi).
Sensors 19 05219 g003
Figure 4. An overview sketch of Support Vector Machines (SVM) algorithm linear classifier.
Figure 4. An overview sketch of Support Vector Machines (SVM) algorithm linear classifier.
Sensors 19 05219 g004
Figure 5. Labeling of the amplitude and phase scintillation index values in the training data sets.
Figure 5. Labeling of the amplitude and phase scintillation index values in the training data sets.
Sensors 19 05219 g005
Figure 6. An example of Receiver Operating Characteristics (ROC) space for classification evaluation.
Figure 6. An example of Receiver Operating Characteristics (ROC) space for classification evaluation.
Sensors 19 05219 g006
Figure 7. Receiver Operating Characteristics (ROC) curves of Support Vector Machines (SVM) methods with different kernel functions.
Figure 7. Receiver Operating Characteristics (ROC) curves of Support Vector Machines (SVM) methods with different kernel functions.
Sensors 19 05219 g007
Figure 8. Scintillation detection results based on Support Vector Machines (SVM) with different kernel functions. “1” corresponds to the points in which the related method points out the scintillation event and “0” means no-scintillation event is detected. Both amplitude and phase scintillation indices synchronized in time can be evaluated as ground truth in the graph to evaluate the performances of different kernels.
Figure 8. Scintillation detection results based on Support Vector Machines (SVM) with different kernel functions. “1” corresponds to the points in which the related method points out the scintillation event and “0” means no-scintillation event is detected. Both amplitude and phase scintillation indices synchronized in time can be evaluated as ground truth in the graph to evaluate the performances of different kernels.
Sensors 19 05219 g008
Table 1. Specifications of the scintillated data sets with visible Pseudo-Random Noise (PRN) codes and location of the stations.
Table 1. Specifications of the scintillated data sets with visible Pseudo-Random Noise (PRN) codes and location of the stations.
DatesPRNsStationCoordinates
121 January 2016
3 February 2016
8 February 2016
17 August 2016
11 , 14 , 22
3 , 6
1 , 12 , 14
9
South African Antarctic
Research Base (SANAE-IV), Antarctic
Lat.: 71 . 67278 ° S
Long.: 2 . 840556 ° W
210 April 2013
12 April 2013
16 April 2013
4 October 2013
11
17
1 , 20 , 28
15 , 21 , 24
Hanoi, VietnamLat.: 21 . 004556 ° N
Long.: 105 . 843917 ° E
Table 2. Phase scintillation detection performance comparison in terms of complexity, True Positive Rate (TPR), False Positive Rate (FPR), and Area Under Curve (AUC) under 10-fold cross-validation test.
Table 2. Phase scintillation detection performance comparison in terms of complexity, True Positive Rate (TPR), False Positive Rate (FPR), and Area Under Curve (AUC) under 10-fold cross-validation test.
SVM MethodKernel
Scale
Running
Time
Validation
Accuracy (%)
Operating PointAUC
(%)
TPRFPR
Linear1 t p 86.010.67720.046891.98
Coarse Gaussian6.91.28 t p 86.290.67550.048290.10
Medium Gaussian1.71.55 t p 86.160.67510.048090.85
Fine Gaussian0.431.70 t p 85.950.68900.053093.16
Polynomial (Order:2)11.37 t p 86.260.67680.048589.38
Polynomial (Order:3)13.20 t p 86.040.67790.048892.67
Table 3. Amplitude scintillation detection performance comparison in terms of complexity, True Positive Rate (TPR), False Positive Rate (FPR), and Area Under Curve (AUC) under 10-fold cross-validation test.
Table 3. Amplitude scintillation detection performance comparison in terms of complexity, True Positive Rate (TPR), False Positive Rate (FPR), and Area Under Curve (AUC) under 10-fold cross-validation test.
SVM MethodKernel
Scale
Running
Time
Validation
Accuracy (%)
Operating PointAUC
(%)
TPRFPR
Linear1 t a 90.440.89900.046095.37
Coarse Gaussian6.91.02 t a 90.720.90040.048795.18
Medium Gaussian1.71.18 t a 91.650.90040.048795.19
Fine Gaussian0.431.33 t a 91.560.90180.037896.01
Polynomial (Order:2)11.08 t a 91.420.89200.026595.61
Polynomial (Order:3)11.80 t a 91.560.89270.029295.88
Table 4. Confusion matrix.
Table 4. Confusion matrix.
CONFUSION
MATRIX
ACTUAL
ScintillationNo-Scintillation
PREDICTIONScintillationTrue Positive (TP)False Positive (FP)
No-ScintillationFalse Negative (FN)True Negative (TN)
Table 5. Accuracy and error rate performances of different kernel SVMs for scintillation detection
Table 5. Accuracy and error rate performances of different kernel SVMs for scintillation detection
Phase ScintillationAmplitude Scintillation
AccuracyError RateAccuracyError Rate
Linear 90 . 98 % 9 . 02 % 94 . 46 % 5 . 54 %
Coarse Gaussian 92 . 61 % 7 . 39 % 94 . 18 % 5 . 82 %
Medium Gaussian 92 . 61 % 7 . 39 % 94 . 18 % 5 . 82 %
Fine Gaussian 92 . 50 % 7 . 50 % 95 . 02 % 4 . 98 %
Polynomial (Order:2) 92 . 28 % 7 . 72 % 94 . 46 % 5 . 54 %
Polynomial (Order:3) 93 . 59 % 6 . 41 % 94 . 74 % 5 . 26 %

Share and Cite

MDPI and ACS Style

Savas, C.; Dovis, F. The Impact of Different Kernel Functions on the Performance of Scintillation Detection Based on Support Vector Machines. Sensors 2019, 19, 5219. https://doi.org/10.3390/s19235219

AMA Style

Savas C, Dovis F. The Impact of Different Kernel Functions on the Performance of Scintillation Detection Based on Support Vector Machines. Sensors. 2019; 19(23):5219. https://doi.org/10.3390/s19235219

Chicago/Turabian Style

Savas, Caner, and Fabio Dovis. 2019. "The Impact of Different Kernel Functions on the Performance of Scintillation Detection Based on Support Vector Machines" Sensors 19, no. 23: 5219. https://doi.org/10.3390/s19235219

APA Style

Savas, C., & Dovis, F. (2019). The Impact of Different Kernel Functions on the Performance of Scintillation Detection Based on Support Vector Machines. Sensors, 19(23), 5219. https://doi.org/10.3390/s19235219

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop