Next Article in Journal
Mapping Rural Road Networks from Global Positioning System (GPS) Trajectories of Motorcycle Taxis in Sigomre Area, Siaya County, Kenya
Next Article in Special Issue
Journey-to-Crime Distances of Residential Burglars in China Disentangled: Origin and Destination Effects
Previous Article in Journal
Automation of Building Permission by Integration of BIM and Geospatial Data
Previous Article in Special Issue
Grid-Based Crime Prediction Using Geographical Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identifying Modes of Driving Railway Trains from GPS Trajectory Data: An Ensemble Classifier-Based Approach

School of Traffic and Transportation, Beijing Jiaotong University, No. 3 Shang Yuan Cun, Hai Dian District, Beijing 100044, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2018, 7(8), 308; https://doi.org/10.3390/ijgi7080308
Submission received: 18 June 2018 / Revised: 24 July 2018 / Accepted: 30 July 2018 / Published: 1 August 2018
(This article belongs to the Special Issue Human-Centric Data Science for Urban Studies)

Abstract

:
Recognizing Modes of Driving Railway Trains (MDRT) can help to solve railway freight transportation problems in driver behavior research, auto-driving system design and capacity utilization optimization. Previous studies have focused on analyses and applications of MDRT, but there is currently no approach to automatically and effectively identify MDRT in the context of big data. In this study, we propose an integrated approach including data preprocessing, feature extraction, classifiers modeling, training and parameter tuning, and model evaluation to infer MDRT using GPS data. The highlights of this study are as follows: First, we propose methods for extracting Driving Segmented Standard Deviation Features (DSSDF) combined with classical features for the purpose of improving identification performances. Second, we find the most suitable classifier for identifying MDRT based on a comparison of performances of K-Nearest Neighbor, Support Vector Machines, AdaBoost, Random Forest, Gradient Boosting Decision Tree, and XGBoost. From the real-data experiment, we conclude that: (i) The ensemble classifier XGBoost produces the best performance with an accuracy of 92.70%; (ii) The group of DSSDF plays an important role in identifying MDRT with an accuracy improvement of 11.2% (using XGBoost). The proposed approach has been applied in capacity utilization optimization and new driver training for the Baoshen Railway.

1. Introduction

1.1. Background

With the increasing demands of diversification and individuation, quality improvements in railway freight transportation services have received an unprecedented amount of attention from operators of railway systems. From a planning perspective, quality railway freight transportation services must be built on precise information or knowledge that reflects the characteristics of the railway system [1,2]. For this reason, data mining has become one of the most important areas in the field of railway research. In these repositories of information or knowledge, the operators eagerly want to know the Modes of Driving Railway Trains (MDRT) [3].
MDRT are the results of driving autonomy. For the purpose of ensuring the robustness of railway systems, operators of railways give autonomy to drivers in the driving processes (This phenomenon even exists in the highly information-influenced High-speed Railway System of China). Under these circumstances, drivers are able to drive the trains according to their experience and basic constraints. Although robustness is ensured, the resulting driving diversity makes railway operations unpredictable since different types of driving modes have different impacts on the efficiency, feasibility, etc. of the railway system [3,4,5,6]. For example, MDRT existing in sections (the connecting parts of adjacent stations) make the running time multivariate, even in the same running plan. If operators have little understanding of MDRT (e.g., setting running time as a constant value instead of a multivariate dataset with different contexts), the resulting plans are inaccurate and remain largely unimplemented.
Although a good understanding of MDRT can contribute to the study of driver behavior habits [3,7,8,9], auto-driving system design [3], and capacity utilization optimization [10,11], building a MDRT set is difficult in the context of big data; traditional methods have certain limitations in terms of scale and continuous learning. This paper is motivated by finding an approach to identifying MDRT automatically. Once the train has completed a portion of the work, its trajectory data will be immediately identified and stored. After a certain period of accumulation, the operator can analyze the MDRT in terms of proportion and internal characteristics. The information obtained can be used for multivariate running time calculations (for planning) or driver driving strategy extraction (for new driver training).
In China, locomotives now have built-in GPS devices with the deployment of a set of railway controlling systems (e.g., Train Operation Dispatching Command System (TDCS), and Central Traffic Control (CTC), etc.). Benefiting from this development of railway information systems, we can use GPS trajectory data as proxies to identify MDRT [12]. Recognizing train modes from GPS trajectory data has many advantages, such as being low cost [13], having few spatiotemporal limitations [14], abundant information [12,13,14], and maturing techniques [15,16,17,18]. This paper focuses on finding a proper approach (including data preprocessing, feature extraction, classifiers modeling, training and parameter tuning, and model evaluation) to identify modes of driving railway trains from GPS trajectory data.

1.2. Related Works

In existing studies, the analyses and applications of concepts similar to MDRT are hot spots in railway data mining [3,7,8,9,10,11]. A multi-phase analyzing framework [3] was built to support many complex railway studies such as large-scale micro-simulation [7] and timetable optimizations [9]. However, these studies focused on the applications of MDRT, and there was a lack of an approach to automatically and effectively identify MDRT in a big data context. Inspired by a relatively new area called transportation mode detection, we will propose a data-mining-based approach to fill this gap.
Transportation mode detection is actually a group of applications of supervised-learning classification methods [19]. A technical framework for researching detection includes four components [12,13,14,19,20,21,22,23,24,25]: (i) Data preprocessing; (ii) Feature extraction and selection; (iii) Classifier identifications; (iv) Evaluations and analyses. A proper combination of methods in these four components determines the quality of identification.
A set of discriminative features plays a vital role in identifying modes. In existing studies, global features [19,21,25,26,27,28,29], local features [12,30], time-domain features [26], frequency-domain features [25,26] and specific features [19,28,31,32,33] were extracted through corresponding methods. Among these features, global features focus on describing whole characteristics of trajectory parameters (e.g., speed, acceleration, etc.) including average values (i.e., mean, absolute mean, median, mode), variance, standard deviation, percentiles, skewness, and kurtosis [21] Conversely, local features focus on describing local characteristics of trajectory parameters. For example, Deng et al. [30] proposed a random forest based method to split the trajectory into segments and obtain features including the mean, standard deviation, and slope from the interval part of the trajectory. In addition, feature sinuosity and deviation, which are extracted from each parameter profile of segment data by a profile decomposition algorithm, were proposed by Dodge et al. [12]. The combination of global features and local features has shown a high level of performance (only 0.0923 misclassification rate) in the research [21].
Moreover, from the point of view of signal analysis, time-domain and frequency-domain features were used in mode detections [25,26]. Two time-domain features, zero-crossing rate and number of peaks extracted from the time-domain, were proposed by Elhoushi et al. [26]. As time domain features are easily affected by noise and interference, the current research has focused more on the frequency domain. Frequency-domain features were obtained through DFT (Discrete Fourier Transform) [25], Short-time Fourier Transform (STFT) [26] or Fast Orthogonal Search (FOS) [26]. The results of these studies show that features in the time domain and frequency domain could help to improve accuracy.
In addition, some specific features were proposed for the purposes of improving accuracy of classification, such as average rail location closeness [19], estimated horizontal accuracy uncertainty [20] and average proximity to bus, tram, or train network [28]. The design of a specific features extracting method should be based on characteristics of transportation options. A good combination of the features mentioned above could help improve accuracy of detection.
As to classifiers, many studies have introduced different classifiers into transportation mode detection, such as Decision Tree [14,19,24,26,34], K-Nearest-Neighbor (KNN) [21,34,35], Support Vector Machines (SVM) [12,14,21,34], Artificial Neural Network (ANN) [20,28,33], Fuzzy Logic [13,28], Hidden Markov Models [25,27], and Bayesian Network [19,36,37]. In addition, ensemble classifiers have become a hotspot in the field of transportation mode detections since ensemble classifiers are a combination set of weak learners, which are superior to a single stronger learner [21,38]. Representative ensemble classifiers include XGBoost [39,40], Random Forest (RF) [19,21,30,34], Adaboost [41,42,43]. Another advantage of using ensemble classifiers is that these methods can conduct feature selection, which performed better than the PCA method [21].
Although prior studies achieved a great deal, our goal faces certain challenges; the most significant one is as follows: The existing studies focused on detecting modes from different transportation options. However, our study aims to identify different driving modes in one method of transportation, where the commonness and characteristics of modes are much more complex than those in existing studies. Thus, for the best performance in identifying MDRT, we want the following: (i) A method to extract new features according to railway characteristics; (ii) the most suitable classifier based on performance comparisons of different classifiers.
The proposed approach has been applied in capacity utilization optimization and new driver training of Baoshen railway: (i) the approach constructed a MDRT collection where obtained a multivariate parameter set of running time, so that operators can make distinct plan for different situations; (ii) new drivers can learn different modes and enhance their ability to cope in different situations.
The remainder of this paper is organized as follows: In Section 2, methods of data preprocessing, feature extraction, classifiers modeling, training and parameter tuning, and model evaluation are elaborated in detail. Among these, one of the highlights of this study is that we propose a method for extracting DSSDF combined with classical features for the purpose of improving performance identification. In Section 3, a real-data experiment is conducted, where we fulfill the following tasks: (i) Finding the optimal value of n s , i.e., determining the most suitable extraction method for DSSDF; (ii) Evaluating performances of classifiers by different indicators and finding the most suitable classifier for identifying MDRT; (iii) Evaluating and comparing performances of different combinations of features and classifiers. In Section 4, discussions and conclusions about the experiment, as well as key points for future research, are presented.

2. Methodologies

In this paper, we propose an integrated approach to identify MDRT using GPS data. The main modules of our identification methodologies are shown in Figure 1 and include data preprocessing (discussed in Section 2.1), feature extraction (discussed in Section 2.2), classifiers modeling (discussed in Section 2.3), training and parameter tuning (discussed in Section 2.3), and model evaluation (discussed in Section 2.4). Raw GPS data were first preprocessed (including data cleaning and representation) into data segments, and then four groups of features were extracted before using the ensemble classifiers (AdaBoost, Random Forest, Gradient Boosting Decision Tree, and eXtreme Gradient Boosting) and single classifiers (K-Nearest Neighbor and Support Vector Machine) to identify MDRT. In order to obtain more reliable, stable and accurate models, techniques such as k-fold cross validation and Bayesian optimization were utilized. In addition, an evaluation system was built to represent the performances of features and classifiers. Each step is detailed in the following sections.

2.1. Preprocessing

To achieve a better performance, two data preprocessing techniques were employed in this study. First, we removed the duplicate data in the dataset as some GPS points were recorded more than once due to recording errors on the GPS device. Second, according to common sense, we removed some outlier trajectories that were deemed abnormal. For instance, if the average speed of a trajectory exceeds the designed speed, we identified it as an abnormal trajectory and removed it from the dataset.

2.1.1. Modes of Driving Railway Train

MDRT are utilized in the field of railway transportation, which is a relatively niche area as compared to existing areas of transportation mode detection. To obtain a more intuitive understanding, Figure 2 is used to illustrate MDRT by trajectories of parameter speed/distance.
Figure 2 shows 12 different modes in MDRT as examples, where the horizontal axis represents distance (measured from the starting station) and the vertical axis represents speed (km/h). These distinct modes are summarized from the actual production processes, which carry information of actual driving habits, reactions to signals and device characteristics. For example, mode 11 is a type of driving that starts from a station, does not speed up to the maximum speed, and passes by the next station by the main track. The details of other modes will be illustrated in Section 3.1. Compared to existing transportation modes, MDRT are more inclined to describe differences and commence at the level of microcosmic driving operation. Thus, the number of modes in MDRT is much more than that of existing mode detections.
These modes are difficult to identify. In the actual production process, operators once used the running time to distinguish MDRT. However, this method is inefficient; many modes are very close in running time. In addition, the inner phases of these modes are complex. For example, mode 8 has two unfixed-location parts of acceleration, which need to be labeled manually. For this reason, methods based on driving phase analyses [3,7,9] in studies could not be used to identify MDRT directly; the traditional method requires a large amount of manual processing, which will restrict pattern recognition under the condition of large data size.
To improve the accuracy of identification of MDRT, we propose a method to extract specific features named Driving Segmented Standard Deviation Features (DSSDF). The idea of extracting DSSDF is from the measures of driving deviations from running plans. From Figure 2, we find two groups of modes: 0–3 stop in both Station 1 and Station 2 (stop-stop). 4–11 stop only in Station1 and pass Station2 (stop-pass). We call these two stop strategies Running plan 1 and Running plan 2, respectively. From the aspect of planning, each running plan corresponds to an idea driving trajectory called Planed Traction Curve (PTC). The proposed features are designed depending on the difference between PTC and MDRS. The details of extracting DSSDF are shown in Section 2.2.2.

2.1.2. Representation and Segments of Trajectory Data

We proposed a three-level representation framework to formulate trajectory data for our study, which contains not only a set of notations but also clear definitions for the inner-relationship between elements according to mode identification. In this framework, tracks are the most primitive forms of GPS trajectory data submitted from GPS sensors. Segments are the data the proposed identifications focus on. In addition, trips are the connections between tracks and segments. We use Figure 3 to illustrate the three-level (track-trip-segment [13,14]) framework for representing trajectory data.
As to the track, it is in large volume and continuous [25] since the GPS sensors in locomotives work ubiquitously. A track is a sequence of GPS points denoted by t r a c k = { P 1 , P 2 , , P m } , where GPS point P i is formatted as l a t , l o n , t , v , h , a c c , where: l a t represents the latitude; l o n represents longitude; t represents the timestamp of the sensor data submitting; v represents the current ground speed of the device; h represents the direction of travel; and a c c represents the accuracy level of the latitude and longitude coordinates. Distance, speed, and acceleration can be obtained using the Haversine formula [36]. In the context of railway freight transportation, the track is the whole GPS trajectory of train movements including running in sections, operating in station, and repaired in depot. A track is too large-scale and complex to be analyzed, and therefore, decompositions of trajectory data are necessary.
After decomposing, we obtain components of the track: trips and activities [13]. We define t r a c k = { t r i p s ,   a c t i v i t i e s } . Trips by railway are the GPS sub-trajectories between two terminal stations corresponding to moving tasks (e.g., an operation line on timetable), and activities are the connections between trips (activities are corresponding to tasks in terminal stations). The key of detecting trips is finding activities. In the view of GPS data, if without signal loss, activities are defined as GPS points whose speeds are close to zero [18] or bundles of GPS points that are very close to each other [44,45] or a large change in the direction of the trajectory. Otherwise, if there are single losses (e.g., at an underground station, etc.), activities are defined as silence intervals, where no GPS points are summited. Therefore, there are three criteria for detection of activities in situations with or without signal loss, from aspects of density, speed-satisfied-time or time difference between two consecutive GPS points [13,14].
As to what constitutes a segment, it is defined as a subset of trip. In the railway context, segments correspond to data in sections (between two adjacent stations). We denote segments as S e = { P a , P a + 1 P b } = { d 1 , d 2 , d N } , where b a + 1 = N and N is the number of data points in a segment trajectory. The threshold based rule [19] and single loss based rule [13,19,46] are utilized to generate segments. To differentiate the stops in stations and the stops in signal range outside stations, we preset space ranges for each station. If the rules above are activated in the range, the train stops in stations. Otherwise, the train stops at signals. To describe the pattern recognition method more straightforwardly, in the remainder of this study, we also call segments data samples.
Based on segments, parameters (or profiles) such as speed, acceleration, and turning angle can be obtained. These parameters are key carriers of knowledge that could characterize the movement behavior and physics of the driving. Each parameter from a segment can be expressed as a time series X = { x 1 , x 2 , x N } . In this study, we used two parameters, speed and acceleration, for extracting features, which is discussed in the following Section 2.2.

2.2. Feature Extraction

Feature extraction plays an important part in Feature engineering, which have great impacts on the performance of classifiers. In this subsection, we first presented classical extracted features including: Global Statistical Features (GSF) and Local Statistical Features (LSF), as well as Time-Domain and Frequency-Domain Features (TDFDF). Then, we proposed a new group of features named Driving Segmented Standard Deviation Features (DSSDF). DSSDF are based on railway transportation characteristics, which are designed for improving identification accuracy. The following explains each group of features in detail.

2.2.1. Classical Features

GSF [21,24,29] refer to descriptive statistics for data samples, which makes the samples more comparable. LSF [12] is extracted by profile decomposition and reveals more detail in movement behavior. TDFDF can represent characteristics from time-domain analysis and frequency-domain analysis [25,26]. We illustrate these features in Table 1.

2.2.2. Driving Segmented Standard Deviation Features (DSSDF)

Although the classical features were proved efficient in the field of transportation mode detection [21], these features still cannot effectively represent MDRT; modes in MDRT are differentiated by microscopic driving details (e.g., some modes have two acceleration phases, and some modes have two breaking phases, etc.). Specific features with rich information are needed in this study. Thus, we designed DSSDF, which measures deviations between modes and each Planed Traction Curve (PTC) of running plans (as discussed in Section 2.1.1) to improve the quality of identification of MDRT.
As a complex system, railway requires many plans or benchmarks to ensure the efficiency and feasibility of operation. In terms of train driving, the PTC is the most important benchmark (PTC is the ideal driving curve designed by the railway design department by means of an experience summary or Newtonian mechanics calculation [5]). Trajectories in each mode would obey a certain distribution in parts or whole scales. If we measure the differences between a trajectory and the PTC by a certain way, the identifications might obtain more information and thus be more accurate.
Driving Segmented Standard Deviation Features s s d ( s i , p j ) measures the differences between a sub-segment and the sub-PTC corresponding to the same distance sub-range s i of running plan p j .
s s d ( s i , p j ) = D T W ( X ( s i ) , P T C ( s i , p j ) ) i X ( s i ) = X ,   i P T C ( s i , p j ) = P T C p j ,   and   s i = S
where X ( s i ) is the sub-segment and S is the distance range corresponding to a target parameter. s i is the ith ( 0 i < n s , n s represents the number of sub-segments) sub-range of the distance range S . P T C ( s i , p j ) is the sub-PTC of P T C p j . P T C p j is the PTC of running plan p j . ( 0 j < n p , n p represents the number of running plans). D T W ( T ( s i ) , P T C ( s i , p j ) ) is a function measuring similarity between X ( s i ) and P T C ( s i , p j ) by the method of Dynamic Time Warping (DTW). The details of DTW can be found in Keogh et al. [47]. The number of PTCs depends only on existing running plans, which can be found in railway plans.
We illustrate the measurements of s s d ( s i , p j ) with 4-folds ( n s = 4 ) and 1 PTC in Figure 4. In this illustration, two parameter trajectories, speed and time, are taken into account. From the aspect of distance range, trajectories and PTCs are divided into 4 sub-ranges (distance sub-ranges 1, 2, 3, 4). Therefore, we extract 8 DSSDF from this example (e.g., in distance sub-range 1, there are two features calculated by D T W ( T 1 ( 1 ) , P T C ( 1 , 1 ) ) corresponding to origin-cube curve and blue-cube curve as well as D T W ( T 1 t ( 1 ) , P T C t ( 1 , 1 ) ) corresponding to yellow-triangle curve and gray-triangle curve) and so on for the other 6 DSSDF. If there is more than one PTC (multiple running plans), differences between a trajectory and any PTC should be measured.
If we just set n s = 1 , and only 2 DSSDF could be extracted, the result might be inaccurate. Conversely, if we set n s too large, the calculation cost might be very high. As a key control parameter, the number of sub-segments n s plays an important part in identifications of MDRT. A proper value of n s should be found before training classifiers.
As to the criteria of determining the sub-segment, we used signal ranges as the basic units for starting a new sub-segment. In addition, we used a greedy search method that assumes that drivers have more details about driving when leaving and entering stations than when cruising in the middle of a section to generate sub-segments. After enter n s , this search method begins by setting the signal range nearest to the arrival station as a new sub-segment; after that, set the signal range nearest the departure station. In addition, set the second nearest signal range to the arrival station as the third sub-segment. This iteration does not stop until n s is achieved. Although this method is rough, its efficiency and interpretability are acceptable in practical applications. The determination process of n s will be discussed in Section 3.3.

2.2.3. Features Summary

A summary of the features discussed above is shown in Table 2. In Figure 5, we plotted the distribution of six example-features (of 12 modes) mentioned above. However, we found it difficult to differentiate between the different modes using only one feature. Nevertheless, these features can provide useful information in identifying MDRT when they are combined.

2.3. Classifiers Modeling and Parameter Tuning

2.3.1. Classifiers Modeling

In this study, Ensemble classifiers (AdaBoost [48,49], Random Forest (RF) [50], Gradient Boosting Decision Tree (GBDT) [51,52], and eXtreme Gradient Boosting (XGBoost) [39,40]) and single classifiers (K-Nearest Neighbor (KNN) [21,34,35,53] and Support Vector Machine (SVM) [54,55]) were modeled to identify MDRT. The details can be found in their respective references.
As to the implementations of classifiers, the KNN, SVM, AdaBoost, RF, and GBDT models were implemented in the Python software along with “scikit-learn” package. In addition, the XGBoost model was implemented in the Python software along with “XGBoost” package.

2.3.2. Parameter Tuning

In machine learning, parameter tuning (also called hyperparameter optimization) is the problem of choosing a set of optimal hyperparameters for a learning algorithm [38,56]. Grid search [57,58], Random search [59,60], Bayesian optimization [61,62,63,64], and Gradient-based optimization [65] are four existing methods of tuning parameters. In practice, Bayesian optimization has been shown to obtain better results in fewer evaluations compared to grid search and random search due to the ability to reason with respect to the quality of experiments before they are run [61,62,63,64]. Therefore, we chose Bayesian optimization to conduct parameter tuning. The details of Bayesian optimization can be found in references [61,62,63,64].
The tuning parameters of each classifier and the setting of parameter ranges for parameter tuning are shown in Table 3.

2.4. Evaluation Methods and Cross-Validation

2.4.1. Evaluation Indicators

An appropriate evaluation framework could help to estimate performance. Basically, samples in classification problems are divided into four categories: true positives; true negatives; false positives; and false negatives. These four categories are shown in Figure 6 as (3), (4), (5), and (6), respectively. Based on these four categories, an evaluation system is established. The arrangement of the indicators in Figure 6 portrays the relationship between them. For example, indicator Recall (10) is calculated from condition positive (1) and true positive (3). Thus it is arranged in row 3 column 1. Indicators in this evaluation system were widely used in existing studies to represent performance [19,21,25,27,28].
In this study, we used PPV, TPR, F1-score, ACC, confusion matrix and Receiver Operating Characteristic curve (ROC, whose x-axis is FPR and y-axis is TPR, in Figure 6), which are marked blue in Figure 6 to evaluate performances of proposed features and classifiers.

2.4.2. K-Fold Cross-Validation

K-fold cross validation is a technique used to obtain more reliable, stable and accurate models in machine learning. In k-fold cross validation, the original samples are randomly partitioned into k equal parts; of the k parts, a single part is used as the validation dataset, and the remaining k-1 subparts are used as the training dataset to construct the model. This procedure is repeated k times, where a different validation dataset is chosen each time, before the final accuracy of the model is equal to the average accuracy obtained each time. All samples are used for both training and validation, and each sample is used for validation exactly once [21]. For example, a dataset has 10 samples. If we use five-fold cross validation to train a model, this dataset will be divided into 5 parts (each with 2 samples), named 1–5. The training process has 5 rounds. In the first round, part 1 is used as the validation dataset, and parts 2–4 are used as the training dataset to construct the model. In the remaining rounds, parts 2–4 are used as validation dataset, respectively. The final accuracy of the model is equal to the average accuracy obtained each round.
The k-fold cross-validation was embedded in the classifier training processes of our experiment, which can guide parameter tuning processes by estimated generalization performance [60].

3. Results and Discussion

3.1. Experiment Data

A dataset of trajectories over a period (of seven years from January 2009 to January 2013) from a section (Dongsheng–Aobaogou) in Baoshen Railway, Inner Mongolia, China was collected for validating the performance of the models (shown in Figure 7). This dataset contains 20,349 trajectories (approximately 13.94 trajectories per day) produced by 103 locomotives. Almost all the trajectories were recorded densely; i.e., the locomotives use an event-based recording mechanism to record trajectories. That is, if an event occurs (e.g., speed change, gear change), the locomotive submit one record (one data point) to the database.
In this experiment, 12 modes (the representative trajectories are shown in Figure 7b) were identified, and 2 PTCs (i.e., PTC i and PTC ii) corresponding to two running plans (shown in Figure 7c) are used to extract DSSDF. Modes 0–3 stop in both stations (stop-stop). 4–11 stop only in departure station and pass arrival station (stop-pass). The details of these 12 modes are summarized as follows: (1) Mode 0 tries to conduct PTC i; (2) mode 1 slows down in advance; (3) mode 2 slowly moves forward, and the speed does not increase to the speed limit; (4) mode 3 first moves forward at low speed and then accelerates to the speed limit; (5) model 4 tries to conduct PTC ii and passes through the main track; (6) mode 5 passes the station after a large proportion of deceleration; (7) mode 6 slows down a little in advance and passes the station; (8) mode 7 slowly moves forward, and the speed does not increase to the speed limit. In addition, it then passes the station; (9) mode 8 accelerates to the speed limit and starts to decelerate and then re-accelerates and passes the main track; (10) mode 9 accelerates to the speed limit and starts to decelerate and passes at low speed; (11) mode 10 passes the station through a side line; (12) mode 11 slowly accelerates and then passes through the main track.
In this dataset, 1/3 of trajectories were labeled with their modes in each segment. The dataset was separated into a training set (consisting of 70% of the data) and a testing set (30% of the data). The distribution of the number of different categories is shown in Table 4. As the distribution of the different categories was unbalanced, it was unreasonable to base our evaluation on accuracy alone, despite this being commonly done in the previous literature. Furthermore, it was difficult to evaluate the performance of the model based on a single measure as these measures are sometimes the same. Thus, in our study, different evaluation elements were used, including PPV, TPR, F1-Score, ROC curve and confusion matrix. We used accuracy (ACC), which is obtained from five-fold Cross-Validation, to tune the parameters (by Bayesian optimization) [21].

3.2. Experiment Scheme

We extracted four groups of features from 6783 trajectories, including GSF (indexed by 1), LSF (indexed by 2), TDFDF (indexed by 3), and DSSDF (indexed by 4), and then used six types of classifiers (KNN, SVM, Ada, RF, GBDT, and XGBoost) to identify MDRT. In this experiment, we need to fulfill the following tasks: (i) Finding the optimal value of n s , i.e., determining the most suitable extraction method for DSSDF; (ii) Evaluating performances of classifiers by different indicators; (iii) Evaluating and comparing performances of different combinations of features and classifiers.
As to task (i), we extracted DSSDF and used XGBoost to determine the optimal value of n s for the purpose of improving performance identification in the following tests. A candidate set of n s { 1 , 2 , 3 , 4 , 5 , 6 } was used to extract a set of DSSDF and fed to XGBoost. A curve of indicator ACC was presented for selecting the optimal value of n s . In the processes of trainings with different features, Bayesian optimization and five-fold cross-validation were used to tune parameters so that the XGBoost classifier was in a good situation. The optimal value of n s would be used in the following trainings.
As to task (ii), we analyzed the performances of six classifiers including two types of single classifiers and four types of ensemble classifiers aggregated with the complete feature group combinations (i.e., { GSF , LSF , TDFDF , DSSDF } , denoted by “1234”) by indicators ROC(AUC), ACC, PPV, TPR, F1-score. Two single classifiers formed the control group as a basis for measuring the performances of the ensemble classifiers. Bayesian Optimization and five-fold cross-validation were used to tune the parameters of each classifier.
As to task (iii), we evaluated different combinations of the six classifiers and feature groups (i.e., the non-empty power set of set { GSF , LSF , TDFDF , DSSDF } ) by indicator ACC to analyze matching relationships between features and classifiers. In addition, the efficiency of DSSDF was evaluated quantitatively here. Bayesian Optimization and five-fold cross-validation were used to tune parameters.
As to parameter tuning, we used the Bayesian Optimization method in the Python software along with the “Bayesian-Optimization” package. The parameters were set as: “n_iter”: 25 and “init_points”: 5, which means the tuning has an initialization with 5 parameter points and 25 iterations. When tuning parameters, “int” type parameters were converted to “real” type fed to Gaussian Process of Bayesian optimization and were converted back to “int” type fed to classifiers.

3.3. Determination the Optimal Value of n s

As discussed in Section 2.2, the extractions of DSSDF are controlled by parameter n s . In this subsection, we conducted an experiment that fed different DSSDF with different n s to XGBoost and then selected the optimal n s according to identification performance (evaluated by indicator ACC).
We measured different DSSDF from the n s collection { 1 , 2 , 3 , 4 , 5 , 6 } (in this section, there are 6 signal ranges) and then trained XGBoost (tuning parameters by Bayesian Optimization with indicator ACC). The Results are shown in Figure 8. We find that 4 is the best value of n s .

3.4. Classifiers Performance Evaluations

3.4.1. K-Nearest Neighbor

Figure 9a and Table 5 show the changes in ACC in the process of parameter tuning, which was obtained using Bayesian optimization and five-fold cross-validation. From Table 5, we can see that the highest ACC was achieved when K was 5.43, 5.28, 5.18, 5.75 5.74, 5.00 (5), and the value was 74.1%. The ROC curve is plotted in Figure 9b, and the confusion matrix is presented in Table 6. The weighted mean of AUC was 0.85. The F1-score was 74.9%. The PPV was 77.7%. The TPR was 72.9%. These indicators show the poor performance of KNN, which led to the conclusion that this classifier is not suitable for differentiating MDRT.

3.4.2. Support Vector Machines

Figure 10a and Table 7 show the changes in ACC in the process of parameter tuning, which was obtained using Bayesian optimization and five-fold cross-validation. From Table 7, we can see that the highest ACC was achieved when “C” was 0.001 and “gamma” was 0.01, and the corresponding value was 88.9%. The ROC curve is plotted in Figure 10b, and the confusion matrix is presented in Table 8. The weighted mean of AUC was 0.97. The F1-score was 88.2%. The PPV was 88.0%. The TPR was 88.5%. These indicators show that although the SVM model performed better than the KNN model, it still cannot effectively identify MDRT.

3.4.3. AbaBoost

Figure 11a and Table 9 show the changes in ACC in the process of parameter tuning, which was obtained using Bayesian optimization and five-fold cross-validation. From Table 9, we can see that the highest ACC was achieved when “n_estimators” was 233.10 (233) and “learning_rate” was 0.02, and the corresponding value was 84.3%. The ROC curve is plotted in Figure 11b, and the confusion matrix is presented in Table 10. The weighted mean of AUC was 0.95. The F1-score was 84.4%. The PPV was 91.5%. The TPR was 79.6%. These indicators show that the AdaBoost is weaker than the SVM model when identifying MDRT.

3.4.4. Random Forest

Figure 12a and Table 11 show the changes in ACC in the process of parameter tuning, which was obtained using Bayesian optimization and five-fold cross-validation. From Table 11, we can see that the highest ACC was achieved when “n_estimators” was 151.51 (151), “min_samples_split” was 9.24 and “max_features” was 0.40, and the corresponding value was 90.9%. The ROC curve is plotted in Figure 12b, and the confusion matrix is presented in Table 12. The weighted mean of AUC was 0.98. The F1-score was 90.9%. The PPV was 93.7%. The TPR was 88.8%. These indicators show that the RF has remarkably outperformed the KNN, SVM and Aba models. However, from the confusion matrix, we find that when identifying class2, the RF is weaker than the other classifiers. It still cannot effectively identify MDRT.

3.4.5. Gradient Boosting Decision Tree

Figure 13a and Table 13 show the changes in ACC in the process of parameter tuning, which was obtained using Bayesian optimization and five-fold cross-validation. From Table 13, we can see that the highest ACC was achieved when “n_estimators” was 181.32 (181), “learning_rate” was 0.10 and “subsample” was 0.90, and the corresponding value was 86.9%. The ROC curve is plotted in Figure 13b, and the confusion matrix is presented in Table 14. The weighted mean of AUC was 0.97. The F1-score was 85.1%. The PPV was 88.5%. The TPR was 82.8%. The results imply a small improvement over the KNN model, but GBDT is weaker than SVM and RF model.

3.4.6. Extreme Gradient Boosting

Figure 14a and Table 15 show the changes in ACC in the process of parameter tuning, which was obtained using Bayesian optimization and five-fold cross-validation. From Table 15, we can see that the highest ACC was achieved when “min_child_weight” was 3.55 (3), “colsample_bytree” was 1.00, “max_depth” was 10.00 (10), “subsample” was 1.00, “gamma” was 0, “alpha” was 0 and “eta” was 0.10, and the value was 92.7%. The ROC curve is plotted in Figure 14b, and the confusion matrix is presented in Table 16. The weighted mean of AUC was 0.99. The F1-score was 91.2%. The PPV was 92.5%. The TPR was 90.0%. The confusion matrix also shows that although XGBoost outperformed other classifiers when identifying modes 0, 1, 3, etc., it is not good at identifying mode 2 with only 76.5% accuracy (much less than the figure of 88.2% achieved by SVM). SVM might be better at dealing with small samples (mode 2 has only 42 labeled trajectories) than XGboost. Overall, these indicators show that the XGBoost classifier has remarkably outperformed the other classifiers when identifying MDRT.

3.5. The Comparison of Features and Classifiers

We tested different combinations of features and classifiers by ACC, which is shown in Figure 15 and Table 17. From this comparison, we can conclude the following: (i) In general, combinations of feature groups perform better than a single feature group; (ii) Different classifiers have their own most suitable feature combinations; (iii) Overall, XGBoost is the most suitable classifier for identifying MDRT with features GSF, LSF, TDFDF, and DSSDF; (iv) The DSSDF (4) is important since all the best performances of different classifiers are related to it; (v) By comparing the results of “123” (XGBoost) and “1234” (XGBoost), we find that DSSDF results in an accuracy improvement of 11.2%.
Features in DSSDF are important mainly because the modes in MDRT mainly differ in the details of microscopic driving. It is difficult to capture these in general statistical methods. With PTCs as benchmarks, the differences between different modes can be demonstrated by segmentation and DTW calculations.
The XGBoost has a complete theoretical system that makes the modeling and tuning process clearer. At the same time, the construction process of its tree structure is more accurate and robust. These advantages make it better in this study.

4. Conclusions

An approach including data preprocessing, feature extraction, classifiers modeling, training and parameter tuning, and model evaluation was developed to infer Modes of Driving Railway Trains (MDRT) using only GPS data. To obtain better performance, we proposed four groups of features with corresponding extraction methods and used different classifiers including four ensemble classifiers and two single classifiers to identify MDRT. The experiment revealed the following: (i) Combinations of feature groups make sense when identifying MDRT, and proposed Driving Segmented Standard Deviation Features play an important role in improving identification performance with an accuracy improvement of 11.2% (using XGBoost); (ii) In general, combinations of feature groups perform better than a single feature group; (iii) Different classifiers have their own most suitable feature combinations; (iv) The XGBoost classifier is the best according to the evaluation indicators, with the highest accuracy (92.70%) from five-fold cross-validation and Bayesian optimization.
The research results of this paper have strong application value in the field of railway transportation. First, in the field of driver behavior habits, further analysis of the internal characteristics of the MDRT can give operators a large deal of micro-decision-support information, such as the driving strategy structure adopted by the driver after the delay, and the proportion of different driving behaviors in the face of signal constraints [3]. The information can help to correct parameters of planning and is of great significance for further optimizing the railway transportation plans [8]. Second, in the field of automatic driving, MDRT can provide the auto-driving system with a decision-making basis for different conditions in accordance with historical reality, reduce the optimization calculation scale and improve the feasibility of the driving scheme. Finally, in the field of capacity utilization optimization, combined with the random occupancy time model, the research results of this paper can be combined with micro-simulation models [3,7,9] and conflict detection systems [10,11] in order to improve the efficiency and feasibility of the optimization process.
In this paper, we used single classifiers and ensemble classifiers to identify MDRS. Testing more classifiers (deep learning and Bayesian network, etc.) is one part of our future work. In addition, analyzing the internal characteristics of MDRT is an important basic of implementing MDRT, which is planned for further work. How to explain the ensemble model more intuitively is also a part of our future work. Another aspect of the research that must be addressed is how to use fusion data to identify MDRS. The utilization of fusion data might improve accuracy of identification. On one hand, we need to strengthen the relevance of our research to existing railway data fusion systems [3,7,8,9,10,11]. On the other hand, approaches to collecting fusion data from other sensors in locomotives need to be mastered.

Author Contributions

Methodology, H.Z.; Software, H.Z. and Z.C.; Supervision, X.Z.; Writing—original draft, H.Z.; Writing—review and editing, X.Z.

Funding

The research is supported by National Natural Science Foundation of China (Grant No. U1734204).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, J.; Zhang, X.; Cai, H.; Zheng, Y. A monitoring data mining based approach to measuring and correcting timetable parameters. Procedia Soc. Behav. Sci. 2012, 43, 644–652. [Google Scholar] [CrossRef]
  2. Wang, J.; Zhang, X.; Yi, Z.; Chen, J. Method for the measurement and correction of train diagram parameters based on monitoring data mining. China Railw. Sci. 2011, 32, 117–121. [Google Scholar]
  3. Longo, G.; Medeossi, G.; Nash, A. Estimating train motion using detailed sensor data. In Proceedings of the Transportation Research Board 91st Annual Meeting, Washington, DC, USA, 22–26 January 2012; pp. 1–6. [Google Scholar]
  4. Zhou, L.; Tong, L.; Chen, J.; Tang, J.; Zhou, X. Joint optimization of high-speed train timetables and speed profiles: A unified modeling approach using space-time-speed grid networks. Transp. Res. Part B Methodol. 2017, 97, 157–181. [Google Scholar] [CrossRef]
  5. Bešinović, N. Integrated Capacity Assessment and Timetabling Models for Dense Railway Networks; Netherlands TRAIL Research School: Delft, The Netherlands, 2017. [Google Scholar]
  6. Bešinović, N.; Goverde, R.M.P.; Quaglietta, E.; Roberti, R. An integrated micro–macro approach to robust railway timetabling. Transp. Res. Part B Methodol. 2016, 87, 14–32. [Google Scholar] [CrossRef] [Green Version]
  7. Fabris, S.D.; Longo, G.; Medeossi, G. Automated analysis of train event recorder data to improve micro-simulation models. In Proceedings of the COMPRAIL 2010 Conference, Beijing, China, 31 August–2 September 2010; pp. 575–583. [Google Scholar]
  8. Powell, J.P.; Palacín, R. Driving Style for Ertms Level 2 and Conventional Lineside Signalling: An Exploratory Study; ResearchGate: Berlin, Germany, 2016. [Google Scholar]
  9. Medeossi, G.; Longo, G.; Fabris, S.D. A method for using stochastic blocking times to improve timetable planning. J. Rail Transp. Plan. Manag. 2011, 1, 1–13. [Google Scholar] [CrossRef]
  10. Goverde, R.M.P.; Daamen, W.; Hansen, I.A. Automatic identification of route conflict occurrences and their consequences. In Computers in Railways XI; WIT Press: Southampton, UK, 2008; pp. 473–482. [Google Scholar]
  11. Albrecht, T.; Goverde, R.M.P.; Weeda, V.A.; Luipen, J.V. Reconstruction of train trajectories from track occupation data to determine the effects of a driver information system. In Proceedings of the COMPRAIL 2006 Conference, Prague, Czech Republic, 31 August–2 September 2006; pp. 207–216. [Google Scholar]
  12. Dodge, S.; Weibel, R.; Forootan, E. Revealing the physics of movement: Comparing the similarity of movement characteristics of different types of moving objects. Comput. Environ. Urban Syst. 2009, 33, 419–434. [Google Scholar] [CrossRef] [Green Version]
  13. Schuessler, N.; Axhausen, K.W. Processing GPS Raw Data without Additional Information; Transportation Research Board: Washington, DC, USA, 2008. [Google Scholar]
  14. Zheng, Y.; Liu, L.; Wang, L.; Xie, X. Learning transportation mode from raw gps data for geographic applications on the web. In Proceedings of the International Conference on World Wide Web (WWW 2008), Beijing, China, 21–25 April 2008; pp. 247–256. [Google Scholar]
  15. Wagner, D.P. Lexington Area Travel Data Collection Test: GPS for Personal Travel Surveys; Elsevier: Amsterdam, The Netherlands, 1997. [Google Scholar]
  16. Yalamanchili, L.; Pendyala, R.; Prabaharan, N.; Chakravarthy, P. Analysis of global positioning system-based data collection methods for capturing multistop trip-chaining behavior. Transp. Res. Rec. J. Transp. Res. Board 1999, 1660, 58–65. [Google Scholar] [CrossRef]
  17. Draijer, G.; Kalfs, N.; Perdok, J. Global positioning system as data collection method for travel research. Opt. Express 2000, 1719, 147–153. [Google Scholar] [CrossRef]
  18. Wolf, J.L. Using GPS Data Loggers to Replace Travel Diaries in the Collection of Travel Data. Ph.D. Thesis, School of Civil and Environmental Engineering, Georgia Institute of Technology, Atlanta, GA, USA, 2000. [Google Scholar]
  19. Stenneth, L.; Wolfson, O.; Yu, P.S.; Xu, B. Transportation mode detection using mobile phones and gis information. In Proceedings of the ACM Sigspatial International Symposium on Advances in Geographic Information Systems (ACM-GIS 2011), Chicago, IL, USA, 1–4 November 2011; pp. 54–63. [Google Scholar]
  20. Gonzalez, P.A.; Weinstein, J.S.; Barbeau, S.J.; Labrador, M.A.; Winters, P.L.; Georggi, N.L.; Perez, R. Automating mode detection using neural networks and assisted gps data collected using gps-enabled mobile phones. In Proceedings of the 15th World Congress on Intelligent Transport Systems and ITS America’s 2008 Annual Meeting, New York, NY, USA, 16–20 November 2008. [Google Scholar]
  21. Xiao, Z.; Wang, Y.; Fu, K.; Wu, F. Identifying different transportation modes from trajectory data using tree-based ensemble classifiers. ISPRS Int. J. Geo-Inf. 2017, 6, 57. [Google Scholar] [CrossRef]
  22. Patterson, D.J.; Liao, L.; Fox, D.; Kautz, H. Inferring High-Level Behavior from Low-Level Sensors; Springer: Berlin, Germany, 2003; pp. 73–89. [Google Scholar]
  23. Lin, L.; Fox, D.; Kautz, H. Learning and inferring transportation routines. In Proceedings of the 19th National Conference on Artifical Intelligence, San Jose, CA, USA, 25–29 July 2004; pp. 348–353. [Google Scholar]
  24. Zheng, Y.; Li, Q.; Chen, Y.; Xie, X.; Ma, W.Y. Understanding mobility based on gps data. In Proceedings of the 10th International Conference on Ubiquitous Computing, Seoul, Korea, 21–24 September 2008; pp. 312–321. [Google Scholar]
  25. Reddy, S.; Min, M.; Burke, J.; Estrin, D.; Hansen, M.; Srivastava, M. Using mobile phones to determine transportation modes. ACM Trans. Sensor Netw. 2010, 6, 13. [Google Scholar] [CrossRef]
  26. Elhoushi, M.; Georgy, J.; Noureldin, A.; Korenberg, M. Online motion mode recognition for portable navigation using low-cost sensors. Navigation 2016, 62, 273–290. [Google Scholar] [CrossRef]
  27. Widhalm, P.; Nitsche, P.; Brändle, N. Transport mode detection with realistic smartphone sensor data. In Proceedings of the 2012 21st International Conference on Pattern Recognition (ICPR), Tsukuba, Japan, 11–15 November 2012; pp. 573–576. [Google Scholar]
  28. Das, R.D.; Winter, S. Detecting urban transport modes using a hybrid knowledge driven framework from gps trajectory. Int. J. Geo-Inf. 2016, 5, 207. [Google Scholar] [CrossRef]
  29. Mardia, K.V.; Jupp, P.E. Directional Statistics; Wiley: Chichester, UK, 2000. [Google Scholar]
  30. Deng, H.; Runger, G.; Tuv, E.; Vladimir, M. A time series forest for classification and feature extraction. Inf. Sci. 2013, 239, 142–153. [Google Scholar] [CrossRef] [Green Version]
  31. Zhang, J.; Wang, Y.; Zhao, W. An improved hybrid method for enhanced road feature selection in map generalization. Int. J. Geo-Inf. 2017, 6, 196. [Google Scholar] [CrossRef]
  32. Qian, H.; Lu, Y. Simplifying gps trajectory data with enhanced spatial-temporal constraints. Int. J. Geo-Inf. 2017, 6, 329. [Google Scholar] [CrossRef]
  33. Ma, C.; Zhang, Y.; Wang, A.; Wang, Y.; Chen, G. Traffic command gesture recognition for virtual urban scenes based on a spatiotemporal convolution neural network. ISPRS Int. J. Geo-Inf. 2018, 7, 37. [Google Scholar] [CrossRef]
  34. Jahangiri, A.; Rakha, H.A. Applying machine learning techniques to transportation mode recognition using mobile phone sensor data. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2406–2417. [Google Scholar] [CrossRef]
  35. Bagnall, A.; Lines, J.; Bostrom, A.; Large, J.; Keogh, E. The great time series classification bake off: A review and experimental evaluation of recent algorithmic advances. Data Min. Knowl. Discov. 2017, 31, 606–660. [Google Scholar] [CrossRef]
  36. Feng, T.; Timmermans, H.J.P. Transportation mode recognition using gps and accelerometer data. Tramsp. Res. Part C Emerg. Technol. 2013, 37, 118–130. [Google Scholar] [CrossRef]
  37. Xiao, G.; Juan, Z.; Zhang, C. Travel mode detection based on gps track data and bayesian networks. Comput. Environ. Urban Syst. 2015, 54, 14–22. [Google Scholar] [CrossRef]
  38. Bishop, C.M. Pattern Recognition and Machine Learning (Information Science and Statistics); Springer: New York, NY, USA, 2006; p. 049901. [Google Scholar]
  39. Nielsen, D. Tree Boosting with Xgboost—Why Does Xgboost win “ Every” Machine Learning Competition? Master’s Thesis, Norwegian University of Science and Technology, Trondheim, Norway, 2016. [Google Scholar]
  40. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  41. Guo, L.; Ge, P.S.; Zhang, M.H.; Li, L.H.; Zhao, Y.B. Pedestrian detection for intelligent transportation systems combining adaboost algorithm and support vector machine. Expert Syst. Appl. 2012, 39, 4274–4286. [Google Scholar] [CrossRef]
  42. Kowsari, T.; Beauchemin, S.S.; Cho, J. Real-time vehicle detection and tracking using stereo vision and multi-view adaboost. In Proceedings of the International IEEE Conference on Intelligent Transportation Systems, Washington, DC, USA, 5–7 October 2011; pp. 1255–1260. [Google Scholar]
  43. Khammari, A.; Nashashibi, F.; Abramson, Y.; Laurgeau, C. Vehicle detection combining gradient analysis and adaboost classification. In Proceedings of the International IEEE Conference on Intelligent Transportation Systems, Vienna, Austria, 16–16 September 2005; pp. 66–71. [Google Scholar]
  44. Stopher, P.; Jiang, Q.; Fitzgerald, C. Processing gps data from travel surveys. In Proceedings of the 2nd International Colloqium on the Behavioural Foundations of Integrated Land-Use and Transportation Models: Frameworks, Models and Applications, Toronto, ON, Canada, 13–14 June 2005. [Google Scholar]
  45. Jun, J.; Guensler, R.; Ogle, J. Smoothing methods to minimize impact of global positioning system random error on travel distance, speed, and acceleration profile estimates. Transp. Res. Rec. J. Transp. Res. Board 2006, 1972, 141–150. [Google Scholar] [CrossRef]
  46. Prelipcean, A.C.; Gidofalvi, G.; Susilo, Y.O. Measures of transport mode segmentation of trajectories. Int. J. Geogr. Inf. Sci. 2016, 30, 1763–1784. [Google Scholar] [CrossRef]
  47. Keogh, E.; Ratanamahatana, C.A. Exact indexing of dynamic time warping. Knowl. Inf. Syst. 2005, 7, 358–386. [Google Scholar] [CrossRef] [Green Version]
  48. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. In Proceedings of the European Conference on Computational Learning Theory, London, UK, 13–15 March 1995; pp. 23–37. [Google Scholar]
  49. Freund, Y.; Schapire, R.E. Experiments with a new boosting algorithm. In Proceedings of the Thirteenth International Conference on International Conference on Machine Learning, Bari, Italy, 3–6 July 1996; pp. 148–156. [Google Scholar]
  50. Liaw, A.; Wiener, M. Classification and regression by randomforest. R News 2002, 23, 18–22. [Google Scholar]
  51. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  52. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef] [Green Version]
  53. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans.Inf. Theory 2002, 13, 21–27. [Google Scholar] [CrossRef]
  54. Cortes, C. Support vector network. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  55. Liu, L.; Shen, B.; Wang, X. Research on Kernel Function of Support Vector Machine; Springer: Dordrecht, The Netherlands, 2014; pp. 827–834. [Google Scholar]
  56. Claesen, M.; Moor, B.D. Hyperparameter search in machine learning. arXiv, 2015; arXiv:1502.02127. [Google Scholar]
  57. Hsu, C.W. A Practical Guide to Support Vector Classification; National Taiwan University: Taibei, Taiwan, 2010; Volume 67. [Google Scholar]
  58. Chicco, D. Ten quick tips for machine learning in computational biology. Biodata Min. 2017, 10, 35. [Google Scholar] [CrossRef] [PubMed]
  59. Wang, Z.; Hutter, F.; Zoghi, M.; Matheson, D.; De Freitas, N. Bayesian optimization in a billion dimensions via random embeddings. Comput. Sci. 2016. [Google Scholar] [CrossRef]
  60. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  61. Bergstra, J.; Bengio, Y. Algorithms for hyper-parameter optimization. In Proceedings of the International Conference on Neural Information Processing Systems, Granada, Spain, 12–15 December 2011; pp. 2546–2554. [Google Scholar]
  62. Hutter, F.; Hoos, H.H.; Leyton-Brown, K. Sequential model-based optimization for general algorithm configuration. In Proceedings of the International Conference on Learning and Intelligent Optimization, Rome, Italy, 17–21 January 2011; pp. 507–523. [Google Scholar]
  63. Thornton, C.; Hutter, F.; Hoos, H.H.; Leytonbrown, K. Auto-weka: Combined selection and hyperparameter optimization of classification algorithms. Comput. Sci. 2012, 847–855, 847–855. [Google Scholar] [CrossRef]
  64. Snoek, J.; Larochelle, H.; Adams, R.P. Practical bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. 2012, 4, 2951–2959. [Google Scholar]
  65. Chapelle, O.; Vapnik, V.; Bousquet, O.; Mukherjee, S. Choosing multiple parameters for support vector machines. Mach. Learn. 2002, 46, 131–159. [Google Scholar] [CrossRef]
Figure 1. Main Modules of the Methodologies for Identifying Modes of Driving Railway Trains (MDRT).
Figure 1. Main Modules of the Methodologies for Identifying Modes of Driving Railway Trains (MDRT).
Ijgi 07 00308 g001
Figure 2. Examples of MDRT.
Figure 2. Examples of MDRT.
Ijgi 07 00308 g002
Figure 3. Illustration of Representation of Trajectory Data.
Figure 3. Illustration of Representation of Trajectory Data.
Ijgi 07 00308 g003
Figure 4. Illustration of 4-fold Driving Segmented Standard Deviation Features (DSSDF). This figure shows the correspondences between sub-segments.
Figure 4. Illustration of 4-fold Driving Segmented Standard Deviation Features (DSSDF). This figure shows the correspondences between sub-segments.
Ijgi 07 00308 g004
Figure 5. Distributions of some features: (a) u ; (b) l p ( 1 ) ; (c) k ; (d) s t f t ; (e) s s d ( 1 , 1 ) ; and (f) d i s .
Figure 5. Distributions of some features: (a) u ; (b) l p ( 1 ) ; (c) k ; (d) s t f t ; (e) s s d ( 1 , 1 ) ; and (f) d i s .
Ijgi 07 00308 g005
Figure 6. Evolution System of Classification. The four blue ones are our focused indicators.
Figure 6. Evolution System of Classification. The four blue ones are our focused indicators.
Ijgi 07 00308 g006
Figure 7. Illustration of Experiment Data. (a) The distribution of trajectory data. The Data in section Dongsheng–Aobaogou are used in this experiment; (b) Examples of parameters (profiles) of used data; (c) Planed Traction Curves (PTCs) i and ii in this experiment, corresponding to two running plans: stop-stop and stop-pass.
Figure 7. Illustration of Experiment Data. (a) The distribution of trajectory data. The Data in section Dongsheng–Aobaogou are used in this experiment; (b) Examples of parameters (profiles) of used data; (c) Planed Traction Curves (PTCs) i and ii in this experiment, corresponding to two running plans: stop-stop and stop-pass.
Ijgi 07 00308 g007
Figure 8. The Impact of Number of Sub-segments n s on ACC (XGBoost).
Figure 8. The Impact of Number of Sub-segments n s on ACC (XGBoost).
Ijgi 07 00308 g008
Figure 9. Parameter Tuning Process and Receiver Operating Characteristic (ROC) of K-Nearest-Neighbor (KNN), where (a) Description of the parameter tuning process of KNN, where y-axis is ACC and x-axis is iteration; (b) ROC of KNN.
Figure 9. Parameter Tuning Process and Receiver Operating Characteristic (ROC) of K-Nearest-Neighbor (KNN), where (a) Description of the parameter tuning process of KNN, where y-axis is ACC and x-axis is iteration; (b) ROC of KNN.
Ijgi 07 00308 g009
Figure 10. Parameter Tuning Process and ROC of Support Vector Machines (SVM), where (a) Description of the parameter tuning process of SVM, where y-axis is ACC and x-axis is iteration; (b) ROC of SVM.
Figure 10. Parameter Tuning Process and ROC of Support Vector Machines (SVM), where (a) Description of the parameter tuning process of SVM, where y-axis is ACC and x-axis is iteration; (b) ROC of SVM.
Ijgi 07 00308 g010
Figure 11. Parameter Tuning Process and ROC of AdaBoost, where (a) Description of the parameter tuning process of AdaBoost, where y-axis is ACC and x-axis is iteration; (b) ROC of AdaBoost.
Figure 11. Parameter Tuning Process and ROC of AdaBoost, where (a) Description of the parameter tuning process of AdaBoost, where y-axis is ACC and x-axis is iteration; (b) ROC of AdaBoost.
Ijgi 07 00308 g011
Figure 12. Parameter Tuning Process and ROC of Random Forest (RF), where (a) Description of the parameter tuning process of RF, where y-axis is ACC and x-axis is iteration; (b) ROC of RF.
Figure 12. Parameter Tuning Process and ROC of Random Forest (RF), where (a) Description of the parameter tuning process of RF, where y-axis is ACC and x-axis is iteration; (b) ROC of RF.
Ijgi 07 00308 g012
Figure 13. Parameter Tuning Process and ROC of Gradient Boosting Decision Tree (GBDT), where (a) Description of the parameter tuning process of GBDT, where y-axis is ACC and x-axis is iteration; (b) ROC of GBDT.
Figure 13. Parameter Tuning Process and ROC of Gradient Boosting Decision Tree (GBDT), where (a) Description of the parameter tuning process of GBDT, where y-axis is ACC and x-axis is iteration; (b) ROC of GBDT.
Ijgi 07 00308 g013
Figure 14. Parameter Tuning Process and ROC of XGBoost, where (a) Description of the parameter tuning process of XGboost, where y-axis is ACC and x-axis is iteration; (b) ROC of XGBoost.
Figure 14. Parameter Tuning Process and ROC of XGBoost, where (a) Description of the parameter tuning process of XGboost, where y-axis is ACC and x-axis is iteration; (b) ROC of XGBoost.
Ijgi 07 00308 g014
Figure 15. Comparison of Features and Classifiers by ACC.
Figure 15. Comparison of Features and Classifiers by ACC.
Ijgi 07 00308 g015
Table 1. Features of Global Statistical Features (GSF), Local Statistical Features (LSF) and Time-Domain and Frequency-Domain Features (TDFDF).
Table 1. Features of Global Statistical Features (GSF), Local Statistical Features (LSF) and Time-Domain and Frequency-Domain Features (TDFDF).
IDG-Name (ID) 1NameNotationR 2
1GSF (1)Mean u = y i / N [21,29]
2Standard deviation σ = 1 N i ( x i u ) 2
3Mode m o = M o d e ( X )
4Median m e = M e d i a n ( X )
5Max 3 values
Min 3 values
m a 3 = M a x 3 ( X )
m i 3 = M i n 3 ( X ) 3
6Value Range v r = M a x 1 ( X ) M i n 1 ( X ) 3
7Percentile tuple p e r c = ( p e r c 25 , p e r c 75 ) 4
8Interquartile Range | p e r c | = p e r c 75 p e r c 25 4
9Skewness s = N ( N 1 ) ( N 2 ) ( i = 1 N ( x i u ) σ ) 3
10Kurtosis k = 1 N ( i = 1 N ( x i u ) σ ) 4 3
11Coefficient of variation c v = σ / u
12Autocorrelation coefficient a u t o = c 1 / c 0
c k = 1 N i = 1 N k ( x i u ) ( x i + k u )
13Stop rate s r = | p s | / d i s [24]
14Velocity change rate v c r = 1 d i s i = 2 N I { ( x i x i 1 ) ξ s } 5
15Trajectory length d i s = d ( X )
16LSF (2)Mean length of each decomposition class l u 0 3 = 1 n x s 0 3 i = 1 n x s 0 3 L e n ( x s i 0 3 ) 6[12]
17Length standard deviation of each decomposition class l σ 0 3 = 1 n x s 0 3 i ( L e n ( x s i 0 3 ) l u 0 3 ) 2 6
18Proportion of each decomposition class l p 0 3 = 1 N i = 1 n x s 0 3 L e n ( x s i 0 3 ) 6
19Change times c t = n x s 0 3 1 6
20TDFDF (3)Median crossover rate m c r = 1 N 1 i = 1 N 1 I { ( x i m e ) × ( x i 1 m e ) < 0 } 5[25,26]
21Number of peaks p e a k = i = 1 N 2 I { s i g n ( x i x i 1 ) > s i g n ( x i + 1 x i ) } 5
s i g n ( a ) = { + 1   ( a > 0 ) 0   ( a = 0 ) 1   ( a < 0 )
22Short-time Fourier transform s t f t k = i x i w i e j 2 π k i N
w i = { 1   ( 0 < i N 1 ) 0   o t h e r w i s e
1 Group Name (ID); 2 References; 3 M a x k ( X ) and M i n k ( X ) are the maximal and minimal k values of points in X ; 4 p e r c k is the kth percentile of X ; 5 I { x } is the indicator function; 6 n x s 0 3 represents the number of decomposition segments of classes 0 3 . x s i 0 3 represents the ith decomposition segments. L e n ( x s i 0 3 ) calculates the length of decomposition x s i 0 3 .
Table 2. Feature Summary.
Table 2. Feature Summary.
IDDescriptorsFeature DescriptionNumber
1GSF u , σ , m o , m e , m a 3 (3), m i 3 (3), v r , p e r c (2), | p e r c | , s , k , c v , a u t o . For each parameter (2). 18 × 2 = 36
s r , v c r , d i s .3
2LSF l u (4), l σ (4), l p (4), c t . For each parameter (2). ( 4 + 4 + 4 + 1 ) × 2 = 26
3TDFDF m c r , p e a k , s t f t . For each parameter (2). 3 × 2 = 6
4DSSDF s s d ( n s ). For each parameter (2). n s × n p × 2 = 2 · n p · n s
SUM ( 36 + 3 ) + 26 + 6 + 2 n s
= 71 + 2 · n p · n s
Table 3. Parameter Ranges for Parameter Tuning.
Table 3. Parameter Ranges for Parameter Tuning.
ClassifierParameter RangeNotes
KNN“n_neighbors”: (1, 15)n_neighbors (int): Number of neighbors to get.
SVM“C”: (0.001, 100)
“gamma”: (0.0001, 0.1)
C (float): Penalty parameter C of the error term.
gamma (float): Kernel coefficient for ‘rbf’, ‘poly’ and ‘sigmoid’.
Ada“n_estimators”: (10, 250)
“learning_rate”: (0.001, 0.1)
n_estimators (int): The maximum number of estimators at which boosting is terminated.
learning_rate (float): Learning rate shrinks the contribution of each classifier by ‘learning_rate’. There is a trade-off between “learning_rate” and “n_estimators”.
RF“n_estimators”: (10, 200)
“min_samples_split”: (2, 15)
“max_features”: (0.1, 0.999)
n_estimators (int): number of trees in the forest.
min_samples_split (int or float): The minimum number (int) or percentage (float) of samples required to split an internal node.
max_features (int or float): The number or percentage of features to consider when looking for the best split.
GBDT“n_estimators”: (10, 250)
“learning_rate”: (0.1, 0.999)
“subsample”: (0.1, 0.9)
n_estimators (int): The number of boosting stages to perform.
learning_rate (float): learning rate shrinks the contribution of each tree by ‘learning_rate’.
subsample (float): The fraction of samples to be used for fitting the individual base learners.
XGBoost“min_child_weight”: (1, 10)
“colsample_bytree”: (0.1, 1)
“max_depth”: (5, 10)
“subsample”: (0.5, 1)
“gamma”: (0, 1)
“alpha”: (0, 1)
“eta”: (0.001, 0.1)
min_child_weight (int): Minimum sum of instance weight (Hessian) needed in a child.
colsample_bytree (float): Subsample ratio of columns when constructing each tree.
max_depth (int): Maximum tree depth for base learners.
subsample (float): Subsample ratio of the training instance.
gamma (float): Minimum loss reduction required to make a further partition on a leaf node of the tree.
alpha (float): L1 regularization term on weights
eta (float): Boosting learning rate.
Table 4. The distribution of MDRT in this experiment.
Table 4. The distribution of MDRT in this experiment.
Mode IDTrajectory NumberMode IDTrajectory Number
012186441
1102971050
2428546
34629252
484101176
529411189
SUM6783
Table 5. Parameter Tuning Process of KNN Classifier.
Table 5. Parameter Tuning Process of KNN Classifier.
I 1n_n(int) 2ACCI 1n_n(int) 2ACCI 1n_n(int) 2ACC
05.91 (5) 74.1%613.97 (13) 65.6%161.69 (1) 60.1%
02.06 (2) 68.3%74.76 (4) 68.6%175.28 (5) 74.1%
010.50 (10) 66.8%811.64 (11) 66.1%185.18 (5) 74.1%
03.83 (3) 69.4%96.72 (6) 68.1%195.75 (5) 74.1%
07.68 (7) 68.0%105.43 (5)74.1%205.74 (5) 74.1%
115.00 (15) 65.9%118.41 (8) 67.5%216.25 (6) 68.1%
212.83 (12) 65.6%129.83 (9) 67.5%225.00 (5) 74.1%
37.74 (7) 68.0%134.25 (4) 68.6%235.62 (5) 74.1%
41.00 (1) 60.1%1414.53 (14) 65.4%245.18 (5) 74.1%
59.10 (9) 67.5%1512.23 (12) 65.6%255.75 (5) 74.1%
1 Iteration; 2 n_neighbors (the value converted to type “int”).
Table 6. The Confusion Matrix of the KNN Classifier.
Table 6. The Confusion Matrix of the KNN Classifier.
KNNA
01234567891011PPV
P080.714.6 11.8 4.9 2.9 0.0 4.0 0.7 0.5 0.0 1.9 13.2 77.7
110.5 75.511.8 22.7 2.9 6.8 5.7 0.7 0.9 0.0 0.4 1.3
20.6 0.5 76.50.5 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0
31.8 7.3 0.0 68.60.0 0.0 0.6 0.5 0.9 2.0 0.2 1.3
40.0 0.0 0.0 0.0 67.60.8 0.0 0.0 0.0 0.0 0.0 0.0
50.2 0.0 0.0 0.0 8.8 72.92.3 0.0 0.9 3.0 0.0 0.0
61.0 0.7 0.0 0.0 2.9 6.8 60.21.0 8.3 2.0 3.8 3.9
71.4 0.2 0.0 0.0 0.0 2.5 1.7 69.53.2 1.0 14.9 10.5
80.4 0.5 0.0 2.7 14.7 5.9 10.2 2.4 76.64.0 2.1 3.9
90.0 0.0 0.0 0.0 0.0 2.5 3.4 0.0 0.0 87.10.0 0.0
102.9 0.7 0.0 0.0 0.0 1.7 9.7 25.0 7.8 1.0 76.65.3
110.4 0.0 0.0 0.5 0.0 0.0 2.3 0.2 0.9 0.0 0.0 60.5
TPR72.9 F1: 74.9
Table 7. Parameter Tuning Process of SVM Classifier.
Table 7. Parameter Tuning Process of SVM Classifier.
I 1CGammaACCI 1CGammaACCI 1CGammaACC
043.936 0.03 50.9%60.001 0.03 74.3%160.002 0.08 71.2%
05.636 0.09 88.6%799.999 0.10 78.4%1799.996 0.04 85.4%
07.577 0.07 64.6%80.007 0.03 68.5%180.004 0.07 65.7%
043.207 0.08 50.0%90.004 0.03 87.5%19100.000 0.00 75.0%
079.405 0.04 52.1%10100.000 0.09 59.9%20100.000 0.00 51.1%
199.999 0.09 63.8%110.002 0.02 88.6%210.005 0.04 86.8%
20.001 0.04 82.2%120.008 0.07 87.1%220.001 0.07 88.6%
399.999 0.07 76.4%130.004 0.07 88.7%230.002 0.07 88.7%
40.004 0.10 88.6%140.003 0.01 88.8%240.0010.0188.9%
599.996 0.09 86.3%150.002 0.08 69.5%250.003 0.06 86.2%
1 Iteration.
Table 8. The Confusion Matrix of the SVM Classifier.
Table 8. The Confusion Matrix of the SVM Classifier.
SVMA
01234567891011PPV
P092.26.8 0.0 2.7 0.0 0.0 2.8 1.0 0.0 0.0 0.4 3.9 88.0
15.7 88.15.9 5.9 2.9 1.7 2.3 1.0 0.9 1.0 0.4 0.0
20.2 0.0 88.20.5 8.8 0.0 0.0 0.0 0.0 0.0 0.0 0.0
30.2 3.4 5.9 88.10.0 0.8 0.0 0.2 0.0 0.0 0.2 1.3
40.0 0.0 0.0 0.0 85.32.5 2.3 0.0 0.0 0.0 0.0 0.0
50.0 0.0 0.0 0.0 2.9 90.70.0 0.0 0.9 3.0 0.0 0.0
60.6 0.2 0.0 0.0 0.0 0.0 86.41.0 1.4 1.0 1.3 0.0
70.2 0.2 0.0 0.0 0.0 0.0 1.7 86.90.9 0.0 7.7 1.3
80.0 0.2 0.0 0.5 0.0 1.7 1.1 0.7 92.22.0 1.9 7.9
90.0 0.0 0.0 0.5 0.0 2.5 0.6 0.0 0.0 91.10.0 1.3
100.6 0.5 0.0 0.0 0.0 0.0 2.8 9.3 3.2 0.0 88.11.3
110.2 0.5 0.0 1.6 0.0 0.0 0.0 0.0 0.5 2.0 0.0 82.9
TPR88.5 F1: 88.2
Table 9. Parameter Tuning Process of AdaBoost Classifier.
Table 9. Parameter Tuning Process of AdaBoost Classifier.
I 1n_e(int) 2l_r 3ACCIn_e(int)l_rACCIn_e(int)l_rACC
0130.42 (130) 0.07 82.5%678.22 (78) 0.00 81.2%16160.28 (160)0.00 75.9%
089.48 (89) 0.06 83.8%794.43 (94) 0.00 75.4%17183.11 (183) 0.00 79.3%
0216.15 (216) 0.07 82.2%8192.64 (192)0.10 81.6%1818.81 (18) 0.10 81.4%
0230.86 (230) 0.08 82.3%929.19 (29) 0.10 80.1%1984.10 (84) 0.10 81.1%
0233.10 (233)0.0284.3%10150.89 (150) 0.10 81.0%2065.29 (65) 0.10 80.9%
110.00 (10) 0.01 75.6%11119.18 (119) 0.10 81.1%2171.37 (71) 0.10 80.4%
2250.00 (250) 0.05 81.3%12204.05 (204) 0.00 80.0%22113.07 (113) 0.00 75.6%
3171.39 (171) 0.00 75.2%1339.52 (39) 0.10 81.1%23125.01 (125) 0.00 75.7%
450.67 (50) 0.00 75.3%14140.94 (140) 0.00 75.8%24210.52 (210) 0.10 81.4%
5108.26 (108) 0.00 82.1%15242.32 (242) 0.10 82.0%2558.44 (58) 0.10 81.8%
1 Iteration; 2 n_estimators (the value converted to type “int”); 3 learning_rate.
Table 10. The Confusion Matrix of the AdaBoost Classifier.
Table 10. The Confusion Matrix of the AdaBoost Classifier.
AdaA
01234567891011PPV
P096.511.2 5.9 5.9 0.0 0.8 5.1 2.9 0.9 0.0 4.7 7.9 91.5
11.8 85.911.8 16.8 8.8 11.0 5.1 1.0 9.2 16.8 0.2 14.5
20.0 0.0 82.40.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0
30.0 0.5 0.0 76.20.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0
40.0 0.0 0.0 0.0 82.40.0 0.0 0.0 0.0 0.0 0.0 0.0
50.0 0.0 0.0 0.0 0.0 72.00.0 0.0 0.0 0.0 0.0 0.0
60.0 0.0 0.0 0.0 8.8 3.4 80.70.0 4.6 1.0 0.0 0.0
70.0 0.0 0.0 0.0 0.0 0.0 0.0 75.70.0 0.0 1.3 2.6
80.0 0.2 0.0 0.0 0.0 11.9 4.5 0.0 78.48.9 0.4 3.9
90.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 73.30.0 0.0
101.6 2.2 0.0 1.1 0.0 0.8 4.5 20.5 6.9 0.0 93.42.6
110.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 68.4
TPR79.6 F1: 84.4
Table 11. Parameter Tuning Process of Random Forest Classifier.
Table 11. Parameter Tuning Process of Random Forest Classifier.
I 1n_e(int) 2m_s 3m_f 4ACCI 1n_e(int) 2m_sm_fACC
0109.17 (109) 14.51 0.66 88.8%1160.86 (60) 2.00 0.10 86.2%
0128.93 (128) 11.59 0.16 88.9%12188.44 (188) 8.18 0.11 88.1%
082.95 (82) 7.90 0.76 88.1%1310.00 (10) 2.00 1.00 85.5%
0173.56 (173) 3.06 0.55 90.2%14163.27 (163) 8.70 0.11 86.9%
0151.51 (151)9.240.4090.9%1570.61 (70) 15.00 1.00 87.4%
1199.83 (199) 14.90 0.40 88.9%16118.81 (118) 4.19 1.00 88.3%
2158.00 (158) 2.04 0.40 88.7%17188.43 (188) 2.26 0.99 88.2%
310.00 (10) 15.00 0.40 79.7%1894.25 (94) 14.94 0.99 88.2%
4170.47 (170) 14.97 0.40 88.3%19144.86 (144) 8.30 1.00 87.9%
551.23 (51) 15.00 0.40 85.2%2081.62 (81) 14.96 0.12 86.8%
6199.98 (199) 2.05 0.40 89.6%21182.48 (182) 14.97 0.10 87.1%
7138.00 (138) 2.03 0.40 87.7%2229.87 (29) 15.00 1.00 86.6%
829.14 (29) 2.00 0.40 82.8%2375.74 (75) 2.00 1.00 87.4%
9144.82 (144) 14.97 0.40 87.1%24120.71 (120) 14.92 0.16 88.9%
10104.86 (104) 2.00 0.40 87.7%25199.81 (199) 8.60 0.97 88.5%
1 Iteration; 2 n_estimators (the value converted to type “int”); 3 min_samples_split; 4 max_features.
Table 12. The Confusion Matrix of the Random Forest Classifier.
Table 12. The Confusion Matrix of the Random Forest Classifier.
RFA
01234567891011PPV
P094.75.6 5.9 2.2 0.0 0.8 2.3 0.0 0.9 0.0 1.3 7.9 93.7
13.9 90.511.8 10.3 2.9 2.5 2.8 0.2 0.5 0.0 0.0 1.3
20.0 0.0 76.50.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0
30.4 1.5 0.0 87.60.0 0.0 0.0 0.0 0.0 1.0 0.0 1.3
40.0 0.2 0.0 0.0 82.40.0 0.0 0.0 0.0 0.0 0.0 0.0
50.2 0.0 0.0 0.0 11.8 93.21.1 0.0 2.3 0.0 0.0 1.3
60.2 0.2 0.0 0.0 0.0 1.7 88.10.0 1.8 0.0 0.4 5.3
70.4 0.7 0.0 0.0 0.0 0.0 0.6 92.60.9 0.0 7.0 5.3
80.0 0.5 0.0 0.0 0.0 0.8 1.7 0.0 89.40.0 0.4 3.9
90.0 0.0 5.9 0.0 2.9 0.8 0.0 0.0 0.5 99.00.0 0.0
100.2 0.7 0.0 0.0 0.0 0.0 3.4 7.1 3.2 0.0 90.91.3
110.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.5 0.0 0.0 72.4
TPR88.8 F1: 90.9
Table 13. Parameter Tuning Process of GBDT Classifier.
Table 13. Parameter Tuning Process of GBDT Classifier.
I 1n_e(int) 2l_r 3SubsampleACCI 1n_e(int) 2l_rSubsampleACC
0218.76 (218) 0.96 0.35 25.3%11113.85 (113) 0.10 0.10 64.8%
0240.10 (240) 0.82 0.77 34.9%1230.64 (30) 0.10 0.10 70.4%
0123.81 (123) 0.37 0.88 81.9%13176.71 (176) 1.00 0.10 34.9%
017.96 (17) 0.68 0.56 71.3%14151.52 (151) 0.10 0.90 85.4%
0137.47 (137) 0.76 0.18 38.0%15187.40 (187) 0.10 0.10 59.4%
172.03 (72) 0.10 0.10 69.2%16207.85 (207) 0.10 0.90 85.9%
299.14 (99) 1.00 0.10 24.4%17156.08 (156) 1.00 0.10 32.7%
344.94 (44) 1.00 0.90 30.3%18145.83 (145) 1.00 0.90 22.5%
4181.32 (181)0.100.9086.2%19229.98 (229) 0.10 0.10 53.2%
5164.68 (164) 1.00 0.90 24.1%20129.97 (129) 0.10 0.10 31.7%
6196.56 (196) 1.00 0.10 32.3%2137.08 (37) 1.00 0.90 25.2%
710.00 (10) 0.10 0.10 72.9%2224.64 (24) 1.00 0.90 24.5%
8250.00 (250) 0.10 0.10 27.2%2352.90 (52) 0.10 0.10 69.0%
961.47 (61) 1.00 0.90 23.5%24106.94 (106) 1.00 0.90 27.3%
1082.64 (82) 1.00 0.90 26.2%25119.00 (119) 1.00 0.90 22.5%
1 Iteration; 2 n_estimators (the value converted to type “int”); 3 learning_rate.
Table 14. The Confusion Matrix of the GBDT Classifier.
Table 14. The Confusion Matrix of the GBDT Classifier.
GBDTA
01234567891011PPV
P092.48.3 5.9 3.8 2.9 0.0 1.1 0.2 0.0 0.0 1.1 6.6 88.5
16.0 86.417.6 9.2 2.9 0.8 0.0 0.0 0.5 1.0 0.0 0.0
20.0 0.2 58.80.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0
30.2 1.5 11.8 84.90.0 0.0 0.6 0.0 0.9 4.0 0.0 0.0
40.0 0.2 0.0 0.0 88.20.8 0.0 0.0 0.0 1.0 0.0 0.0
50.0 0.0 0.0 0.0 5.9 90.72.3 0.0 1.8 3.0 0.0 0.0
60.4 0.5 0.0 0.0 0.0 1.7 80.10.5 3.2 1.0 1.3 3.9
70.4 1.7 0.0 0.5 0.0 0.8 2.3 86.42.3 0.0 12.8 6.6
80.0 0.5 0.0 0.5 0.0 2.5 7.4 0.5 88.54.0 1.1 1.3
90.0 0.0 5.9 0.5 0.0 1.7 0.0 0.0 0.5 85.10.0 2.6
100.6 0.5 0.0 0.0 0.0 0.8 5.7 12.4 2.3 1.0 83.89.2
110.0 0.2 0.0 0.5 0.0 0.0 0.6 0.0 0.0 0.0 0.0 69.7
TPR82.8 F1: 85.1
Table 15. Parameter Tuning Process of XGBoost Classifier.
Table 15. Parameter Tuning Process of XGBoost Classifier.
I 1m_c_w(int) 2c_b 3m_d(int) 4SubsampleGammaAlphaetaACC
04.25 (4) 0.16 9.03 (9) 0.50 0.77 0.66 0.01 61.9%
07.57 (7) 0.54 8.77 (8) 0.72 0.31 0.34 0.04 91.7%
09.37 (9) 0.30 9.79 (9) 0.68 0.95 0.42 0.05 91.1%
09.77 (9) 0.43 6.35 (6)0.75 0.20 0.57 0.02 91.5%
08.61 (8) 0.86 5.66 (5) 0.84 0.17 0.12 0.08 91.2%
110.00 (10) 1.00 10.00 (10) 1.00 0.00 1.00 0.10 60.8%
210.00 (10) 0.10 10.00 (10) 0.50 0.00 0.00 0.00 69.9%
31.00 (1) 1.00 5.00 (5) 1.00 1.00 1.00 0.10 92.1%
46.01 (6) 0.10 5.00 (5) 1.00 1.00 1.00 0.10 88.7%
58.46 (8) 1.00 7.31 (7) 0.50 1.00 1.00 0.00 65.0%
61.00 (1) 1.00 10.00 (10) 1.00 0.00 0.00 0.10 92.6%
71.00 (1) 0.10 5.00 (5) 1.00 0.00 0.00 0.10 87.6%
86.02 (6) 1.00 10.00 (10) 1.00 1.00 0.00 0.10 91.8%
91.00 (1) 0.10 10.00 (10) 1.00 1.00 1.00 0.10 88.7%
1010.00 (10) 0.10 5.00 (5) 1.00 1.00 0.00 0.10 88.7%
111.00 (1) 1.00 7.46 (7) 1.00 0.00 1.00 0.10 72.6%
126.91 (6) 0.10 10.00 (10) 1.00 0.00 1.00 0.10 86.5%
131.00 (1) 1.00 7.54 (7) 1.00 1.00 0.00 0.10 92.5%
143.91 (3) 1.00 5.00 (5) 1.00 0.00 0.00 0.10 91.8%
1510.00 (10) 1.00 8.12 (8) 1.00 1.00 0.00 0.10 92.2%
1610.00 (10) 1.00 5.00 (5) 1.00 0.00 1.00 0.10 91.3%
179.23 (9) 0.10 8.51 (8) 1.00 1.00 1.00 0.10 88.4%
181.00 (1) 1.00 10.00 (10) 0.50 1.00 1.00 0.00 89.9%
198.29 (8) 1.00 10.00 (10) 1.00 1.00 1.00 0.00 85.6%
206.18 (6) 0.10 5.00 (5) 0.50 0.00 0.00 0.10 86.4%
212.72 (2) 0.10 5.00 (5) 0.50 1.00 1.00 0.10 88.6%
225.34 (5) 1.00 7.52 (7) 1.00 0.00 0.00 0.10 91.8%
231.00 (1) 0.10 8.99 (8) 0.50 0.00 0.00 0.10 85.5%
243.55 (3)1.0010.00 (10)1.000.000.000.1092.7%
251.00 (1) 1.00 5.00 (5) 0.50 0.00 1.00 0.10 90.8%
1 Iteration; 2 min_child_weight (the value converted to type “int”); 3 colsample_bytree; 4 max_depth (the value converted to type “int”).
Table 16. The Confusion Matrix of the XGBoost Model.
Table 16. The Confusion Matrix of the XGBoost Model.
XGBA
01234567891011PPV
P095.14.9 0.0 1.6 0.0 0.0 0.6 0.2 0.0 0.0 1.7 2.6 92.5
14.1 93.211.8 5.9 0.0 0.0 1.1 0.2 0.9 0.0 0.0 0.0
20.0 0.0 76.50.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0
30.0 1.2 5.9 91.90.0 0.0 0.0 0.0 0.0 1.0 0.0 0.0
40.0 0.0 0.0 0.0 85.32.5 0.0 0.0 0.0 0.0 0.0 0.0
50.0 0.0 0.0 0.0 8.8 95.81.1 0.0 0.9 0.0 0.2 0.0
60.0 0.0 0.0 0.0 0.0 0.0 90.90.0 2.8 0.0 0.6 1.3
70.2 0.2 0.0 0.0 0.0 0.0 1.1 93.10.5 0.0 5.5 3.9
80.0 0.2 0.0 0.0 5.9 0.8 2.3 0.0 92.72.0 0.9 2.6
90.0 0.0 5.9 0.5 0.0 0.8 0.0 0.0 0.0 97.00.0 0.0
100.6 0.2 0.0 0.0 0.0 0.0 2.8 6.4 2.3 0.0 91.13.9
110.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 85.5
TPR90.0F1: 91.2
Table 17. Impacts of Features and Classifiers on ACC.
Table 17. Impacts of Features and Classifiers on ACC.
KNNSVMAdaboostRFDBGTXGBoost
132.50 56.30 48.30 50.50 58.80 57.90 Combinations
220.70 21.40 24.40 40.00 24.00 25.00
329.70 46.00 32.80 35.00 33.00 33.90
448.50 60.20 47.40 63.60 61.10 62.20
1234.40 59.00 49.60 81.10 70.10 78.70
1336.10 60.00 51.30 83.40 69.00 81.00
1447.10 70.30 64.90 82.50 88.10 82.10
2330.20 29.50 33.80 37.00 36.10 36.00
2453.60 58.70 51.50 67.60 65.20 66.00
3463.10 68.00 52.90 73.30 70.00 73.70
12336.10 49.50 52.40 81.90 67.70 81.50
13448.50 57.00 66.00 88.60 89.6080.20
12445.30 70.30 65.40 91.6087.70 90.20
23455.00 74.20 52.00 74.60 71.60 75.00
123474.1388.9184.3090.94 86.22 92.70
Classifiers

Share and Cite

MDPI and ACS Style

Zheng, H.; Cui, Z.; Zhang, X. Identifying Modes of Driving Railway Trains from GPS Trajectory Data: An Ensemble Classifier-Based Approach. ISPRS Int. J. Geo-Inf. 2018, 7, 308. https://doi.org/10.3390/ijgi7080308

AMA Style

Zheng H, Cui Z, Zhang X. Identifying Modes of Driving Railway Trains from GPS Trajectory Data: An Ensemble Classifier-Based Approach. ISPRS International Journal of Geo-Information. 2018; 7(8):308. https://doi.org/10.3390/ijgi7080308

Chicago/Turabian Style

Zheng, Han, Zanyang Cui, and Xingchen Zhang. 2018. "Identifying Modes of Driving Railway Trains from GPS Trajectory Data: An Ensemble Classifier-Based Approach" ISPRS International Journal of Geo-Information 7, no. 8: 308. https://doi.org/10.3390/ijgi7080308

APA Style

Zheng, H., Cui, Z., & Zhang, X. (2018). Identifying Modes of Driving Railway Trains from GPS Trajectory Data: An Ensemble Classifier-Based Approach. ISPRS International Journal of Geo-Information, 7(8), 308. https://doi.org/10.3390/ijgi7080308

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop