Next Article in Journal
A Novel Approach for Lidar-Based Robot Localization in a Scale-Drifted Map Constructed Using Monocular SLAM
Next Article in Special Issue
Evaluating and Diagnosing Road Intersection Operation Performance Using Floating Car Data
Previous Article in Journal
Design and Implementation of an Integrated IoT Blockchain Platform for Sensing Data Integrity
Previous Article in Special Issue
Design and Evaluation of a Surface Electromyography-Controlled Steering Assistance Interface
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Autoencoder Neural Networks for Short-Term Traffic Congestion Prediction of Transportation Networks

1
Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu 610041, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
School of Big Data, Guizhou Institute of Technology, Guiyang 550003, China
4
School of Mechanical Engineering, Guizhou University, Guiyang 550003, China
5
College of Big Data Statistics, GuiZhou University of Finance and Economics, Guiyang 550025, China
6
Department of Computer Science and Engineering, University of South Carolina, Columbia, SC 29201, USA
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(10), 2229; https://doi.org/10.3390/s19102229
Submission received: 15 April 2019 / Revised: 9 May 2019 / Accepted: 9 May 2019 / Published: 14 May 2019
(This article belongs to the Special Issue Advance in Sensors and Sensing Systems for Driving and Transportation)

Abstract

:
Traffic congestion prediction is critical for implementing intelligent transportation systems for improving the efficiency and capacity of transportation networks. However, despite its importance, traffic congestion prediction is severely less investigated compared to traffic flow prediction, which is partially due to the severe lack of large-scale high-quality traffic congestion data and advanced algorithms. This paper proposes an accessible and general workflow to acquire large-scale traffic congestion data and to create traffic congestion datasets based on image analysis. With this workflow we create a dataset named Seattle Area Traffic Congestion Status (SATCS) based on traffic congestion map snapshots from a publicly available online traffic service provider Washington State Department of Transportation. We then propose a deep autoencoder-based neural network model with symmetrical layers for the encoder and the decoder to learn temporal correlations of a transportation network and predicting traffic congestion. Our experimental results on the SATCS dataset show that the proposed DCPN model can efficiently and effectively learn temporal relationships of congestion levels of the transportation network for traffic congestion forecasting. Our method outperforms two other state-of-the-art neural network models in prediction performance, generalization capability, and computation efficiency.

1. Introduction

Onsets of freeway congestion reduce efficiency and capacity of transportation networks and should be forecast to take measures to prevent its formation in an accurate and timely manner in most situations [1,2]. Compared with flow or volume prediction [3,4,5,6], traffic congestion forecasting is more intuitive to both road travelers and transportation administrators. It can help road travelers to make better route selections, reduce pollution such as emission of greenhouse gases, and improve transportation operation efficiency. For these reasons, many prediction methods have been proposed and evaluated for traffic congestion prediction [7,8,9].
Various kinds and amounts of traffic data have been used by researchers in recent years for traffic condition prediction and related research of intelligent transportation systems. Most of these works use data sources such as road sensors, induction loops, automatic vehicle identification systems, remote traffic microwave sensors, in-road reflectors, floating car data, and simulation [3,4,10,11,12,13]. Images from cameras installed on roads, aerial photographs and remote sensing images as another kind of data source have also been used [14,15]. These kinds of data are either difficult to access due to special requirement of permissions or often outdated. On the other hand, many traffic administrative departments [16,17] and online map service providers [18,19,20] provide real-time or near real-time online traffic congestion condition maps to the general public for free. Such services use and integrate various data sources, for example induction loops and more generally location-based data originating from apps active on Global Positioning System (GPS)-enabled smartphones carried in running cars. Such maps provide a new kind of real-time traffic condition data source which differs vastly in timeliness, accessibility, availability, and coverage from aforementioned kinds of data sources. Yet there are few studies based on such data sources for traffic congestion prediction. This paper proposes a systematic method to collect and use this new kind of data source for traffic congestion forecasting.
Furthermore, traffic congestion maps provided by traffic administration departments and online map services often cover areas with road networks of various scales instead of a single road. It is more significant and interesting to understand and forecast traffic congestion in an entire road network rather than on a single road. It is also more helpful to present complete traffic congestion information for travelers to choose routes wisely and for traffic administrators to manage road networks and allocate resources systematically. However, it is more demanding to predict congestion for a road network and thus needs abilities to deal with more computational complexity introduced by the network topology, to generate more timely and effective prediction from a three-dimensional perspective, and to output more accurately predicted congestion levels. Unfortunately, conventional traffic congestion prediction models, which represent each road or road segment as an element in a sequence ignoring impacts of road length [7] or only consider a limited number of road links, do not meet aforementioned requirements due to limitations in input and output representation and normalization, improper hypotheses and assumptions, and inability to cope with the curse of dimensionality [21]. Thus, existing models may fail to predict large-scale network traffic congestion.
Motivated by hierarchical feature extraction and motion prediction, this paper represents congestion levels of a traffic network as a 2D-matrix and introduces an image-based deep learning approach for congestion forecasting. A deep learning architecture inspired by deep autoencoders (DAE) [22] is proposed to first learn low-dimensional vector representations of features and relationships embodied in images obtained from aforementioned data sources and then predict future congestion levels in a supervised way based on these learned representations. An autoencoder is a neural network consisting of an encoding module (encoder) and a decoding module (decoder) designed to learn representational properties of the input data [23]. Furthermore, multiple encoders and decoders can transform high-dimensional data into low-dimensional vector representations for further applications [22,24,25].
This paper tries to forecast future congestion levels of a transportation network using only historical temporal information and correlations of congestion levels of the network. In particular, we examine whether it is feasible to directly predict and output road network congestion images through the examination of just a sequence of previous congestion images for that same transportation network in an end-to-end approach with the help of a simple yet efficient deep learning neural network architecture inspired by DAEs.
The contributions of the paper are summarized as follows:
  • We propose an accessible and general approach to collect, transform, and represent snapshots of transportation network maps marked with traffic congestion conditions for roads inside, which are publicly available from transportation administrative departments and online traffic map service providers. Based on this approach, we have built and released a long-span traffic congestion dataset.
  • We develop a deep neural network model for efficient end-to-end prediction of transportation network congestion levels by using hierarchical feature extraction. Our end-to-end learning model directly outputs prediction results presented visually and intuitively in the same road network structure and form as inputs are, thus eliminating the need for manual feature selection and engineering.
  • Our extensive experiments on a transportation network in the Seattle area demonstrates effectiveness and efficiency of the proposed approach.
The remainder of this paper is organized as follows: Section 2 discusses the related literature on traffic prediction. Section 3 presents a systematic approach to collect and transform snapshots of transportation network maps with congestion levels, introduces a grid-based representation method for traffic congestion levels, and shows the architecture of our deep neural network to learn temporal feature representation and correlations for traffic congestion prediction. In Section 4 a transportation network in the Seattle area, Washington state is used to build a traffic congestion dataset and to test the effectiveness of the proposed model. To evaluate the performance of our proposed neural network model, we compare two state-of-the-art deep learning models titled convolutional long short-term memory (LSTM) Network (ConvLSTM) [26] and Spatiotemporal Recurrent Convolutional Network (SRCN) [6]. Then in Section 5 we discuss implications and limitations of our work. The conclusions and future studies are presented in the final section.

2. Related Work

There are two main categories of approaches for prediction of traffic related variables such as speed, volume, and density, namely parametric and nonparametric approaches [21,27].
Being a parametric approach, the Auto Regressive Integrated Moving Average (ARIMA) model was proposed to construct models from time series of historical states to predict future values. Parameters of the ARIMA model could be configured via the Box-Jenkins method [28]. Smith and Williams used the ARIMA model for the first time to predict traffic flow at a single point [28]. A family of ARIMA-based models, such as seasonal ARIMA (SARIMA) models [29,30,31], KARIMA models [32], ARIMAX models [33], and CTM-SARIMA models [34], have been deployed for traffic forecasting since then. These parametric approaches share common requirements which demand predetermined structures of models according to theoretical or physical assumptions and tuning a set of parameters to reflect the evolution of traffic conditions in the real world as much as possible [35,36]. Especially SARIMA models are found to incur high computational cost [28].
The limitations of parametric algorithms have led the shift to nonparametric approaches for traffic prediction such as nonparametric regression, Support Vector Machines (SVM), and Artificial Neural Networks. As a nonparametric approach, k-nearest neighbors (KNN) models have been used to forecast traffic speeds and flows, with both univariate and multivariate cases [37,38,39,40,41]. Supporting vector machine (SVM) and its variants such as SVR, seasonal SVM, and Online-SVM have been explored to improve traffic prediction performance due to their capabilities to generalize well and capture the high dynamics and sensitivity of traffic data [42,43,44]. Artificial neural networks (ANNs), with their advantages such as capability to work with multi-dimensional data, implementation flexibility, generalization ability, and good forecasting performance, have been applied to traffic prediction problems [45]. Kumar et al. applied an ANN to predict traffic volume using time information in addition to past traffic related data such as volume, speed, and density [46]. Kashi and Akbarzadeh used wavelet transformation to remove unimportant fluctuation from the flow signal and then an ANN to train on past data to predict future flow on different highways and locations [47].
Although these methods can model the non-linearity and extract spatial-temporal relationship in the traffic data to achieve better results than parametric approaches, they require significant prior domain knowledge and extensive preprocessing work such as feature engineering. With traffic density increasing, wide adoption of sensors and cameras, and popularization of GPS-enabled navigation apps for smartphones, the big data paradigm has emerged from transportation related data. Such data explosion introduces a problem well-known as the curse of dimensionality, which cannot be handled efficiently by traditional approaches [48,49]. To gain insight from big traffic data, deep neural networks have become popular in recent years to learn deep correlations inherent in data with little or no prior knowledge and need for manual feature engineering [13,50]. Stacked autoencoder models have been used to exploit temporal and spatiotemporal information on real-world or simulated datasets to predict traffic flow [5,51]. Recurrent neural networks (RNNs) especially LSTM have been used to predict traffic flow, speed, and congestion because of their built-in memory cells enabling learning temporal knowledge and thus being suitable for time-series analytics [52,53,54]. Convolutional neural networks (CNNs), with its special strength in extracting spatial correlations, have been adopted for traffic speed prediction on image generated from speed data [4,55]. Deep belief networks (DBNs) have also been used for traffic flow prediction due to their capability to learn effective representative features from data in an unsupervised way [50,56,57]. Furthermore, there have been research efforts which combine more than one kind of deep neural networks. Yu et al. [6] proposed a deep learning model named SRCN combining CNNs and LSTMs to capture and learn both spatial dependencies of different roads and long-term temporal dependency of each road and to predict traffic speed on a certain set of roads, using in-house synthesized road maps with floating car GPS data.
However, these attempts mainly focus on prediction of traffic related properties such as flow, speed, and time on a single road segment, several number of roads, or a small network region [21,58,59,60,61]. One major reason for lack of such studies of traffic congestion prediction is the challenge to obtain large data set. One recent work by Ma et al. [7] used the congestion information collected from the GPS data from taxi to model and predict traffic congestion evolution. However, this dataset has time intervals of 30 minutes and 60 minutes, which is too sparse for training models for real-world congestion prediction. It also spans only four weeks.

3. Methodology

In this section, we describe our accessible and general approach for obtaining and processing raw snapshots of traffic congestion maps from online map service providers. In our approach, traffic congestion data are extracted automatically using an image mask for highways and a customized map-reduce implementation is used to process and transform these extracted congestion data originally represented as pixels into float numbers. Our method can extract traffic congestion levels from raw traffic congestion maps or snapshots provided publicly and for free by online map and traffic service providers. Then we propose a deep neural network (DNN) architecture for traffic congestion prediction.
Inspired by research findings of computer vision and deep learning in motion prediction which estimate future trajectories of objects via sequences of scenes generated by itself [62], we first collect time series of snapshots of network-wide traffic congestion map from Washington State Department of Transportation (WSDOT) [17] with the help of web browsers and web crawlers, and then build a dataset from these snapshots. The approach can be easily extended to other online map service providers to build more datasets. This dataset is then used to train and back-test different deep learning models for congestion prediction. Based on the theory of city management grid modeling [63], we segment the transportation network covered by these snapshots into different grids. Each snapshot is divided into non-overlapping tiled 8 × 8 pixel grids where each grid corresponds to an area of about 80 × 80 m 2 in the real world. Each grid has a congestion level calculated as the average of traffic congestion levels represented by all pixels in that grid.

3.1. Representation of Congestion Level of the Transportation Network

Since we want to forecast congestion levels inside every grid in a traffic network, we first retain only the road network and congestion conditions in the transportation network by removing all other pixels with the help of an image mask as shown by the transformation from Figure 1a to Figure 1b.
Then we segment network-only images into R × C grids, as is shown by Figure 1c. For each grid located at ( i , j ) where 1 i R and 1 j C , { c t , i , j k 1 k 64 } denote the set of congestion levels as represented by all pixels in that grid at time t. Calculated from c t , i , j k according to Equation (1), c ¯ i , j t denotes the averaged congestion level of the grid at the time t.
c ¯ i , j t = k = 1 64 c t , i , j k k = 1 64 [ c t , i , j k > 0 ]
Based on Equation (1), the final congestion level representation for the segmented road network at time t is expressed by the matrix in Equation (2), where R stands for number of grids latitudinally and C longitudinally. Figure 1d provides a colored visualization of a matrix representation as expressed by Equation (2). Each pixel in Figure 1d stands for an area of 80 × 80 m 2 and is rendered according to a custom linear color map [64] generated from a sequence of RGB values ( ( 0 , 0 , 0 ) , ( 0 , 255 , 0 ) , ( 255 , 255 , 0 ) , ( 255 , 0 , 0 ) , ( 139 , 0 , 0 ) ).
R t = c ¯ 1 , 1 t c ¯ 1 , 2 t c ¯ 1 , C t c ¯ 2 , 1 t c ¯ 2 , 2 t c ¯ 2 , C t c ¯ R , 1 t c ¯ R , 2 t c ¯ R , C t
Suppose that we need to predict traffic congestion levels for the road network at time points in ( t + 1 , t + 2 , , t + h ) where h is the prediction horizonand historical congestion levels in the time range ( t - n , , t - 2 , t - 1 , t ) are used as inputs. When we arrange historical records of network traffic congestion representation by time, we get the time-series sequence ( R t - n , , R t - 2 , R t - 1 , R t , R t + 1 , R t + 2 , , R t + h ) .
Because multiple time intervals are used as input for forecasting, the traffic congestion prediction task can be regarded as a time-series sequence prediction problem. When all grids in the network are predicted at the same time, it is known as the multi-dimensional sequence learning problem.

3.2. Temporal Features

Sequences of snapshots of traffic congestion levels across a road network delimited by a fixed time interval in chronological order are very similar to natural language sentences consisting of words separated by spaces. Future congestion levels might be affected by earlier congestion levels to varying degrees due to the temporal dependency inherent in traffic data. In our work we try to use such temporal correlation for prediction of future congestion levels.

3.3. Deep Congestion Prediction Network

We propose to use a deep learning model titled Deep Congestion Prediction Network (DCPN) inspired by DAEs for transportation network congestion prediction. It is designed to learn and represent temporal features and correlations of traffic congestion levels among roads in the transportation network. Our proposed model consists of two components. The first component contains an encoder and a decoder. The encoder first obtains a vector representation of historical congestion levels of a transportation network and their correlations using four encoding layers. Next the decoder builds a representation of the congestion levels for a future time point using four decoding layers. The architecture of this first component uses symmetrical layers for the encoder and the decoder, while in DAEs the encoder and the decoder shares the inner most layer [22]. The second component of DCPN uses two dense layers to construct congestion levels for each grid in that transportation network at that future time point. These two dense layers take the output F t from the decoder in the first component and calculate a vector representation of predicted traffic congestion levels, as shown by Equation (3). W h and b h represent the weight matrix and the bias between the last layer of the DAE-like architecture and hidden dense layer, while W o and b o between the hidden dense layer and the output dense layer. To avoid overfitting, a dropout layer is added between the two dense layers. The prediction vector R ^ t + h has the same number of elements as each of input series and is further reshaped to have the same shape. Thus, in this way we train the proposed model from end to end. To integrate DAE-like architectures and dense layers for traffic congestion prediction, we propose to use the architecture in Figure 2 for DCPN to forecast traffic congestion on a network scale. A DCPN model includes one input layer, one flattened layer, one deep autoencoder, two dense layers with one dropout layer in between, and one reshaping layer.
R ^ t + h = W o × ( W h × F t + b h ) + b o

4. Experiments and Results Analysis

4.1. Data Source

Real-time congestion levels of roads in a transportation network are provided publicly by various online map service providers (such as Google map [18], Bing maps [20] and Autonavi map [19]) and transportation administration departments (for example WSDOT and Beijing Traffic Management Bureau [16]), and can be taken as or have been provided as a series of snapshots of different transportation networks including freeways, interchanges, intersections, ramps, and other elements. Figure 1a provides an example of a raw snapshot of a highway transportation network with traffic congestion levels in Seattle area from WSDOT. Each such raw snapshot is 557 ×1199 pixels in size, covers a large area of about 40 km × 96 km, and embodies congestion levels for mostly interstate and state highways. Service providers derive and infer congestion levels from different data sources, among which are mobility data generated collaboratively by mobile map apps running on smartphones used by commuters and sampling data from loop detectors.
We focus on congestion levels forecasting during morning rush hours from 07:00 a.m. to 10:00 a.m. on workdays. For that purpose, we collect snapshots of traffic congestion map for a transportation network in Seattle area in the aforementioned morning period from 1 January 2016 to 28 February 2017 and all missing data are left as is (a copy of these raw snapshots is available at https://github.com/senzhangcas/traffic-congestion with permission from the original data provider). After removing weekends and days without any snapshots, there are 301 days left. 283 of these days have 19 samples extracted from the time period between 7:00 a.m. and 10:00 a.m. with an interval of 10 min and 18 days have less samples due to missing data. Samples in the year 2016 are used as the initial train set and samples in the year 2017 are used as the test set via back-testing. We name the dataset built from criteria listed above as the Seattle Area Traffic Congestion Status (SATCS) dataset.

4.1.1. Data Comprehension, Preprocessing and Representation

As can be seen from Figure 1a, traffic congestion status snapshots usually contain other elements such as signs and labels which are currently not considered for congestion prediction. We manually make a road network mask using image processing tools and develop shell scripts to filter out such unneeded elements and apply a preprocessing step to convert these raw snapshots to images with only the road network in concern (as shown in Figure 1b) and then these images are arranged chronologically into sequences (as shown in Figure 3) for further transformations as described in the next paragraph.
Each pixel in every image of such a sequence is converted to a float number based on its color (transparent, green, yellow, red, or black) for different congestion levels as shown by the legend in Figure 1e (non-road area, wide open, moderate, heavy, or stop-and-go). Inspired by the work in [65], to make the transformation easy and accurate, the HSL (Hue, Saturation, Lightness) color space is used. Furthermore, based on the theory of city management grid modeling, the 2D-matrix is first segmented (see Figure 1c) and then shrunk (see Figure 1d) via grids of 8 × 8 pixels according to Equation (1). After these preprocessing steps, sequences of 2D-matrices are provided as inputs to all the neural network models we experiment with.

4.2. Performance Comparison and Metrics

To evaluate performance of DCPN and its parameter configuration for congestion prediction, two state-of-the-art deep learning neural network-based models are selected as baselines for comparison. The first model is the SRCN neural network proposed to predict traffic speed using both spatial and temporal information [6]. This model consists of multiple CNN layers and LSTM layers. The second baseline is a ConvLSTM neural network consists of multiple ConvLSTM layers proposed by [26,66] for traffic prediction. ConvLSTM combines advantages of CNN and LSTM by using convolutional structures in both the input-to-state and state-to-state transitions. To perform a wider coverage of time horizons for comparison and analysis, we select three time intervals, namely 10, 30, and 60 min for prediction. We use two metrics—mean absolute error (MAE) and weighted mean squared error (wMSE) as defined in Equations (4) and (5)—to evaluate performance of traffic congestion forecasting in this paper. c i j t , c ^ i j t and w i j t denote the true traffic congestion level, the predicted level and the penalty weight at time t in the grid as designated by the point ( i , j ) in the 2D-matrix representation of the highway transportation network. To combat imbalanced distributions of different congestion levels, w i j t is defined heuristically to put more penalty on incorrect prediction of non-smooth congestion levels. W and H denote numbers of grids along horizontal and vertical dimensions.
M A E = 1 W × H i = 1 W j = 1 H | c i j t - c ^ i j t |
w M S E = 1 W × H i = 1 W j = 1 H w i j t × ( c i j t - c ^ i j t ) 2

4.3. Implementation of the DCPN Model

Previous work by Koesdwiady etc. has determined an optimized time lag of 120 min for traffic prediction [56]. Since samples in the dataset built by us have an interval of 10 min, we use 12 samples initially for DCPN’s parameter selection. We adopt a grid search of 5 configurations of number of hidden units for layers of DCPN to find a proper one. Results of the grid search are shown in Table 1. As can be seen, for prediction horizons of 60 min the 2nd configuration achieves minimum metric values for both wMSE and MAE which is followed very closely by the 1st one, while the 1st configuration perform the best in the 30 min case. With prediction horizon of 10 minutes, the 1st configuration archives minimum value on MAE and the 2nd one on wMSE. Further investigation using Welch’s t-test shows that there are no significant differences between both wMSE and MAE means produced by these two configurations, as shown in Table 2. However, the 1st configuration is more computationally efficient because fewer hidden units and connections are involved. So, we use the 1st configuration in our implementation of DCPN.
Furthermore, we experiment with both time lags of 120 and 110 min with the hope of reducing computing resource requirement. Results are shown in Table 3. Our experiment results show that the DCPN model performed better with the time lag of 110 minutes in terms of wMSE across all three prediction horizons. Thus, to reduce computational demand, for all three models and prediction horizons 11 historical samples before each testing point (as represented in 2D-matrices in Section 3.1) are used as inputs.
Detailed configuration of parameters for the DCPN model is listed in Table 4. The RMSprop optimizer [67] is used to train DCPN. The decay parameter is set to 0.9 and the learning rate is set to 0.001 and the batch size is set to a dynamic number equal to the count of samples in each day, which is provided by a custom data generator. We use a customized mean squared error algorithm with different weights for each congestion level as the loss function. Back-testing is deployed to predict traffic congestion for each workday during the morning rush hours in the year 2017 and samples from most recent 259 days before that day are used as the training dataset. Also, we use early stopping and a dropout layer to keep the model from overfitting [68].
For comparison, we also trained prediction models of two neural networks which can mine both spatial and temporal correlations according to existing literature: SRCN [6] and ConvLSTM [26]. We performed a grid search for both models. Results of the grid search show that the configuration [16, 32, 64, 128] achieves the best metrics for both models. So we adopt the SRCN model using the same configuration parameters as in [6] except the dropout rate is reduced from 0.2 to 0.1. For the ConvLSTM model, we use four stacked ConvLSTM layers with consecutive filter sizes set to 16, 32, 64, and 128; after each of these layers we add one max pooling layer, one activation layer with ReLU activation, and one batch normalization layer; finally one flatten layer, one dense layer, one dropout with a rate of 0.1, and one reshape layer are added to obtain the prediction result. For all three models, 11 historical samples before each testing point are used as inputs. In addition, all three models are implemented using the Keras deep learning library [69] on an Ubuntu 16.04 machine with 2 NVIDIA TITAN Xp Graphics Cards without a NVIDIA SLI connection.

4.4. Back-Testing

Instead of using cross validation for training and testing or fixed train/test split, we use back-testing for time-series prediction to collect metrics of applying models to the SATCS dataset for performance comparison. The widely used method of K-fold cross validation systematically splits the dataset into K groups and gives each group a chance to be a held-out test set. However, K-fold cross validation cannot be directly used with time-series data, because it assumes independence among the observations which is false for time-series data due to highly probable dependency among observations in temporal order imposed by the time dimension. Another common approach is to divide the samples into train and test datasets based on a specific time point as in [7], which does not meet the real-world requirement of daily or even near real-time prediction.

4.5. Results and Analysis

Figure 4 and Figure 5 show the daily total MAEs and wMSE errors for each of the 42 days from 2 January 2017 to 8 February 2017 of three compared models for network congestion prediction. A 10 min prediction horizon is used here and the daily total MAE and wMSE are calculated and shown in Figure 4 and Figure 5. As can be seen from these two figures, the DCPN model has the lowest prediction errors for most of the days.
For more details, Table 5 presents prediction performance per day for the whole transportation network by different models applied to the SATCS dataset across three prediction horizons. The DCPN model performs the best most of the time in terms of achieving the lowest values for both MAE and wMSE for the horizon of 10 minutes. Out of the 42 days for back-testing, it achieves the lowest values of MAE and wMSE for 32 and 33 days, respectively. While for horizons of 30 and 60 min, although SRCN is slightly better than DCPN when only MAE is considered, namely 22 against 20 days, and 21 against 21 days respectively for the two horizons, DCPN is overwhelmingly better than both SRCN and ConvLSTM by achieving the lowest values of wMSE on 41 and 40 days. The metric wMSE is pragmatically more significant and meaningful to travelers and administration departments because it concerns more about non-smooth congestion levels such as moderate, heavy, or stop-and-go ones through applying more penalties for incorrect prediction of such levels. So, it may be inferred that DCPN can be a stable candidate model for practical application of short-term traffic congestion prediction across all three prediction intervals. Furthermore, the averaged values of these two prediction error metrics—with the minimum averaged MAE of 0.0106, 0.0102 and 0.0095 and the minimum wMSE of 0.0579, 0.0536 and 0.0449 for horizons of 10, 30, and 60 min respectively—indicate that our proposed DCPN model achieves the best prediction performance. It can be seen that DCPN outperforms generally the other two on samples during the days for back-testing, suggesting that DCPN has a better generalization capability to new samples.
Both SRCN and ConvLSTM do not perform quite well for traffic congestion prediction on the SATCS dataset, while both have been reported to be able to use spatial information via CNNs [6,26]. On the one hand, characteristics of spatial structures of roads in transportation networks differ significantly from those of handwritten digits in various shapes by different people or images for the same category of objects. The spatial structure of roads stays unchanged while in the latter case each digit or object image has their own unique structure. On the other hand, roads in a metropolitan transportation network interact with each other differently than those in an urban transportation network do. In an urban transportation network, traffic congestion at a certain location can very likely cause congestion at other locations [6]. While in the SATCS dataset which covers the Seattle metropolitan area transportation network consisting of mostly interstate and state highways, we doubt that there exists such a spatial correlation of congestion among highways due to the vast area encompassed and sparse distribution of highways, so CNNs would have no spatial correlation to exploit to improve performance of congestion prediction. However it deserves further investigation to determine how and to what extent convolution operations in CNNs affect performance for congestion prediction in metropolitan transportation networks.
Also, the DCPN model is more computationally efficient than the other two in terms of computation time when applied to the Seattle area transportation network. Table 6 lists the training epochs and running time for the tested models. All three models are run on the same experiment equipment. Although DCPN needs more total number of epochs to train, it uses the least total amount of training time to converge, thus it is computationally most efficient and more suitable for real-world applications. We doubt that multiple convolutional operations inside both SRCN and ConvLSTM models applied to the Seattle metropolitan area transportation network consisting of mostly interstate and state highways would incur higher computing requirements. As an interesting phenomenon observed during our experiment, it suggests further work to investigate demand for computing resources by transportation networks with different sizes and distributions of roads.
As an example of the end-to-end prediction of traffic congestion levels by the DCPN model, Figure 6 shows both the ground truth and predicted congestion levels on the day 17 January 2017 with a prediction horizon of 10 min. Ground truth congestion levels at each time point are created using the same steps as described in Section 4.1.1. It can be seen that the DCPN model outputs visually intuitive prediction results in the same structure and form as the ground truth.

5. Discussion

Lack of large-scale traffic data related datasets with open access hinders research of intelligent transport systems. Data used in previous work—for example GPS and floating car data captured by taxis, images from surveillance CCTV cameras—either requires prior complex or special processes of application by users such as researchers and grant from authorities such as governmental agencies and companies. Nevertheless, there have been transportation network maps marked with congestion indicators or levels provided online for free by transportation administration departments and map service providers. For example, WSDOT has started providing such traffic congestion maps since at least as early as 2012 according to the Internet Archive [70]. To ease building of large-scale datasets for and to promote traffic congestion related research, we present a convenient and general workflow to create datasets for prediction of traffic congestion based on raw data from these free online service providers. Specifically, first we have made available an archived dataset of raw snapshots of the highway transportation network in the Seattle area from 1 January 2016 to 28 February 2017 as collected from WSDOT, most of which are no longer provided on the WSDOT website. Then based on this raw dataset, we build a dataset named SATCS and release it for future traffic congestion related research by us and others.
Even though now there is the SATCS dataset for traffic congestion research, some information is lost when congestion levels are represented in our current work. We must shrink matrix representations of congestion levels in the Seattle area highway transportation network using a grid-based scheme due to limitation imposed by GPU memory size of our current experiment equipment. As a result, each grid covers an area of 80 m × 80 m and the average of congestion levels represented by all pixels in a grid is used as the congestion level for that grid. During this shrinking process information is lost and geographical accuracy is reduced. It is yet unclear what information is lost and how loss of information affects prediction performance. We will deal with this information loss problem in our future work to predict traffic congestion at a finer granularity.
Furthermore, there is still room for improvement of computation efficiency. In the current representation of traffic congestion levels, most values equal 0 . 0 which is for grids containing no roads at all in the vast background area. However, DCPN still uses such values when training and testing, and thus incurs unnecessary computation. In our future work, we will try to exclude such background information to further improve efficiency.

6. Conclusions

In this work, we first propose an accessible and general approach to collect, transform and represent snapshots of road networks marked with congestion levels. We then apply it to build a dataset named SATCS for traffic congestion research. We develop a deep learning model DCPN by combining a DAE-inspired feature learning architecture and dense layers to learn representational features and temporal correlations from historical traffic congestion data for prediction of future congestion levels in a transportation network near the Seattle area, Washington state, USA. To evaluate the effectiveness of the proposed DCPN model for short-term traffic congestion forecasting, we compare its prediction performance with that of two state-of-the-art deep learning neural network models using the back-testing technique. Results over the SATCS benchmark dataset show that our proposed DCPN is more effective and computationally efficient for short-term traffic congestion forecasting.
This study focuses only on prediction of traffic congestion levels using traffic congestion snapshots from a single data source, and is limited by our experiment equipment. However, more extensive traffic forecast solutions are possible by covering travel time, volume, speed, and occupancy, and using other information such as weather conditions, which may be more accurate and meaningful for travelers, commuters, and administration departments. In future work, we will try to enhance computing capability of our experiment equipment to perform more thorough trials, experiment with snapshots from other service providers, and fuse multiple types of data from different sources, in order to build traffic forecast models for predicting aforementioned traffic condition related properties.
Availability: The Seattle area traffic congestion status dataset will be released publicly at https://mekhub.cn/Madsen/SATCS-dataset upon acceptance of this manuscript.

Author Contributions

Conceptualization, S.Z., J.H. (Jianjun Hu); methodology, S.Z., Y.Y. and J.H. (Jie Hu); software, S.Z. and Y.Y.; validation, S.Z.,Y.Y. and J.H. (Jie Hu); investigation, S.Z., Y.Y., and J.H. (Jie Hu); resources, S.L.; data curation, S.Z. and J.H. (Jie Hu); writing—original draft preparation, S.Z., Y.Y., J.H. (Jianjun Hu), and Y.Z.; writing—review and editing, S.Z., J.H. (Jianjun Hu); supervision, J.H. (Jianjun Hu) and S.L.; project administration, J.H. (Jie Hu) and S.L.; funding acquisition, S.L.

Funding

This research was funded by The National Natural Science Foundation of China under Grant No. 91746116 and 51741101, and Science and Technology Project of Guizhou Province under Grant Nos. [2018]5788, Talents [2015]4011 and [2016]5013, Collaborative Innovation [2015]02.

Acknowledgments

We gratefully acknowledge the support of the NVIDIA Corporation with the donation of the Titan X Pascal GPU used for this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tseng, F.; Hsueh, J.; Tseng, C.; Yang, Y.; Chao, H.; Chou, L. Congestion Prediction With Big Data for Real-Time Highway Traffic. IEEE Access 2018, 6, 57311–57323. [Google Scholar] [CrossRef]
  2. Chen, M.; Yu, X.; Liu, Y. PCNN: Deep Convolutional Networks for Short-Term Traffic Congestion Prediction. IEEE Trans. Intell. Transp. Syst. 2018, 19, 3550–3559. [Google Scholar] [CrossRef]
  3. Coifman, B. Freeway Detector Assessment: Aggregate Data from Remote Traffic Microwave Sensor. Transp. Res. Record J. Transp. Res. Board 2005, 1917, 149–163. [Google Scholar] [CrossRef]
  4. Fouladgar, M.; Parchami, M.; Elmasri, R.; Ghaderi, A. Scalable Deep Traffic Flow Neural Networks for Urban Traffic Congestion Prediction. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 2251–2258. [Google Scholar]
  5. Lv, Y.; Duan, Y.; Kang, W.; Li, Z.; Wang, F.Y. Traffic Flow Prediction with Big Data: A Deep Learning Approach. IEEE Trans. Intell. Transp. Syst. 2015, 16, 865–873. [Google Scholar] [CrossRef]
  6. Yu, H.; Wu, Z.; Wang, S.; Wang, Y.; Ma, X. Spatiotemporal Recurrent Convolutional Networks for Traffic Prediction in Transportation Networks. Sensors 2017, 17, 1501. [Google Scholar] [CrossRef]
  7. Ma, X.; Yu, H.; Wang, Y.; Wang, Y. Large-Scale Transportation Network Congestion Evolution Prediction Using Deep Learning Theory. PLoS ONE 2015, 10, e0119044. [Google Scholar] [CrossRef]
  8. Zhang, X.; Onieva, E.; Perallos, A.; Osaba, E.; Lee, V.C. Hierarchical fuzzy rule-based system optimized with genetic algorithms for short term traffic congestion prediction. Transp. Res. Part C Emerg. Technol. 2014, 43, 127–142. [Google Scholar] [CrossRef]
  9. Lopez-Garcia, P.; Onieva, E.; Osaba, E.; Masegosa, A.D.; Perallos, A. A Hybrid Method for Short-Term Traffic Congestion Forecasting Using Genetic Algorithms and Cross Entropy. IEEE Trans. Intell. Transp. Syst. 2016, 17, 557–569. [Google Scholar] [CrossRef]
  10. Ben-Akiva, M.; Bierlaire, M.; Koutsopoulos, H.; Mishalani, R. DynaMIT: A Simulation-Based System for Traffic Prediction; DACCORD Short Term Forecasting Workshop: Delft, The Netherlands, 1998; pp. 1–12. [Google Scholar]
  11. Dailey, D.J.; Trepanier, T. The Use of Weather Data to Predict Non-Recurring Traffic Congestion; Technical Report; Washington Department of Transportation: Seattle, Washington, 2006. [Google Scholar]
  12. Lima, A.M.; Campos, J. Evaluating the Use of Traffic Information from Web Map Services to Analyze the Impact of Non-Recurrent Events. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 2077–2082. [Google Scholar] [CrossRef]
  13. Ali, U.; Mahmood, T. Using Deep Learning to Predict Short Term Traffic Flow: A Systematic Literature Review. In First International Conference on Intelligent Transport Systems; Springer: Cham, Switzerland, 2017; pp. 90–101. [Google Scholar]
  14. Kahaki, S.M.; Nordin, M.J. Highway Traffic Incident Detection Using High-Resolution Aerial Remote Sensing Imagery. J. Comput. Sci. 2011, 7, 949. [Google Scholar] [CrossRef]
  15. Kahaki, S.M.M.; Nordin, M.J. Vision-Based Automatic Incident Detection System Using Image Sequences for Intersections. In Proceedings of the 2011 International Conference on Pattern Analysis and Intelligence Robotics, Putrajaya, Malaysia, 28–29 June 2011; Volume 1, pp. 3–7. [Google Scholar] [CrossRef]
  16. Beijing Traffic Management Bureau. Available online: http://eye.bjjtw.gov.cn/Web-T_bjjt_new/Main.html (accessed on 7 October 2018).
  17. WSDOT—Seattle Washington Cameras. Available online: http://www.wsdot.com/traffic/seattle/default.aspx (accessed on 23 June 2018).
  18. Google Maps. Available online: https://www.google.com/maps (accessed on 6 October 2018).
  19. AutoNavi Map. Available online: https://ditu.amap.com (accessed on 6 March 2018).
  20. Bing Maps. Available online: https://www.bing.com/maps (accessed on 6 October 2018).
  21. Ermagun, A.; Levinson, D. Spatiotemporal Traffic Forecasting: Review and Proposed Directions. Transp. Rev. 2018, 38, 786–814. [Google Scholar] [CrossRef]
  22. Hinton, G.E.; Salakhutdinov, R.R. Reducing the Dimensionality of Data with Neural Networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
  23. Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press: Cambridge, UK, 2016; Volume 1. [Google Scholar]
  24. Krizhevsky, A.; Hinton, G.E. Using Very Deep Autoencoders for Content-Based Image Retrieval. In 19th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning; Ciaco: Bruges, Belgium, 2011. [Google Scholar]
  25. Feng, X.; Zhang, Y.; Glass, J. Speech Feature Denoising and Dereverberation via Deep Autoencoders for Noisy Reverberant Speech Recognition. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 1759–1763. [Google Scholar]
  26. Wang, D.; Yang, Y.; Ning, S. DeepSTCL: A Deep Spatio-Temporal ConvLSTM for Travel Demand Prediction. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar] [CrossRef]
  27. Mori, U.; Mendiburu, A.; Álvarez, M.; Lozano, J.A. A Review of Travel Time Estimation and Forecasting for Advanced Traveller Information Systems. Transp. A Transp. Sci. 2015, 11, 119–157. [Google Scholar] [CrossRef]
  28. Smith, B.L.; Williams, B.M.; Oswald, R.K. Comparison of Parametric and Nonparametric Models for Traffic Flow Forecasting. Transp. Res. Part C Emerg. Technol. 2002, 10, 303–321. [Google Scholar] [CrossRef]
  29. Williams, B.M.; Hoel, L.A. Modeling and Forecasting Vehicular Traffic Flow as a Seasonal ARIMA Process: Theoretical Basis and Empirical Results. J. Transp. Eng. 2003, 129, 664–672. [Google Scholar] [CrossRef]
  30. Ghosh, B.; Basu, B.; O’Mahony, M. Time-Series Modelling for Forecasting Vehicular Traffic Flow in Dublin. In Proceedings of the 84th Annual Meeting of the Transportation Research Board, Washington, DC, USA, 9 January 2005. [Google Scholar]
  31. Tran, Q.T.; Ma, Z.; Li, H.; Hao, L.; Trinh, Q.K. A Multiplicative Seasonal ARIMA/GARCH Model in EVN Traffic Prediction. Int. J. Commun. Netw. Syst. Sci. 2015, 8, 43. [Google Scholar] [CrossRef]
  32. Van Der Voort, M.; Dougherty, M.; Watson, S. Combining Kohonen Maps with Arima Time Series Models to Forecast Traffic Flow. Transp. Res. Part C Emerg. Technol. 1996, 4, 307–318. [Google Scholar] [CrossRef]
  33. Williams, B. Multivariate Vehicular Traffic Flow Prediction: Evaluation of ARIMAX Modeling. Transp. Res. Record J. Transp. Res. Board 2001, 1776, 194–200. [Google Scholar] [CrossRef]
  34. Szeto, W.Y.; Ghosh, B.; Basu, B.; O’Mahony, M. Multivariate Traffic Forecasting Technique Using Cell Transmission Model and SARIMA Model. J. Transp. Eng. 2009, 135, 658–667. [Google Scholar] [CrossRef]
  35. Van Lint, J.W.C.; Van Hinsbergen, C. Short-Term Traffic and Travel Time Prediction Models. Artif. Intell. Appl. Crit. Transp. Issues 2012, 22, 22–41. [Google Scholar]
  36. Vlahogianni, E.I.; Karlaftis, M.G.; Golias, J.C. Short-Term Traffic Forecasting: Where We Are and Where We’re Going. Transp. Res. Part C Emerg. Technol. 2014, 43, 3–19. [Google Scholar] [CrossRef]
  37. Davis, G.A.; Nihan, N.L. Nonparametric Regression and Short-Term Freeway Traffic Forecasting. J. Transp. Eng. 1991, 117, 178–188. [Google Scholar] [CrossRef]
  38. Clark, S. Traffic Prediction Using Multivariate Nonparametric Regression. J. Transp. Eng. 2003, 129, 161–168. [Google Scholar] [CrossRef]
  39. Chang, H.; Lee, Y.; Yoon, B.; Baek, S. Dynamic Near-Term Traffic Flow Prediction: Systemoriented Approach Based on Past Experiences. IET Intell. Transp. Syst. 2012, 6, 292–305. [Google Scholar] [CrossRef]
  40. Liu, Z.; Du, W.; Yan, D.M.; Chai, G.; Guo, J.H. Short-Term Traffic Flow Forecasting Based on Combination of K-Nearest Neighbor and Support Vector Regression. J. Highway Transp. Res. Dev. (Engl. Ed.) 2018, 12, 89–96. [Google Scholar] [CrossRef]
  41. Zheng, L.; Zhu, C.; Zhu, N.; He, T.; Dong, N.; Huang, H. Feature Selection-Based Approach for Urban Short-Term Travel Speed Prediction. IET Intell. Transp. Syst. 2018, 12, 474–484. [Google Scholar] [CrossRef]
  42. Wu, C.H.; Ho, J.M.; Lee, D.T. Travel-Time Prediction with Support Vector Regression. IEEE Trans. Intell. Transp. Syst. 2004, 5, 276–281. [Google Scholar] [CrossRef]
  43. Castro-Neto, M.; Jeong, Y.S.; Jeong, M.K.; Han, L.D. Online-SVR for Short-Term Traffic Flow Prediction under Typical and Atypical Traffic Conditions. Expert Syst. Appl. 2009, 36, 6164–6173. [Google Scholar] [CrossRef]
  44. Hong, W.C.; Dong, Y.; Zheng, F.; Lai, C.Y. Forecasting Urban Traffic Flow by SVR with Continuous ACO. Appl. Math. Model. 2011, 35, 1282–1291. [Google Scholar] [CrossRef]
  45. Karlaftis, M.G.; Vlahogianni, E.I. Statistical Methods versus Neural Networks in Transportation Research: Differences, Similarities and Some Insights. Transp. Res. Part C Emerg. Technol. 2011, 19, 387–399. [Google Scholar] [CrossRef]
  46. Kumar, K.; Parida, M.; Katiyar, V.K. Short Term Traffic Flow Prediction in Heterogeneous Condition Using Artificial Neural Network. Transport 2015, 30, 397–405. [Google Scholar] [CrossRef]
  47. Kashi, S.O.M.; Akbarzadeh, M. A Framework for Short-Term Traffic Flow Forecasting Using the Combination of Wavelet Transformation and Artificial Neural Networks. J. Intell. Transp. Syst. 2018, 1–12. [Google Scholar] [CrossRef]
  48. Donoho, D.L. High-Dimensional Data Analysis: The Curses and Blessings of Dimensionality. AMS Math Chall. Lecture 2000, 1, 32. [Google Scholar]
  49. Verleysen, M.; François, D. The Curse of Dimensionality in Data Mining and Time Series Prediction. In International Work-Conference on Artificial Neural Networks; Springer: Barcelona, Spain, 2005; pp. 758–770. [Google Scholar]
  50. Huang, W.; Song, G.; Hong, H.; Xie, K. Deep Architecture for Traffic Flow Prediction: Deep Belief Networks With Multitask Learning. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2191–2201. [Google Scholar] [CrossRef]
  51. Duan, Y.; Lv, Y.; Wang, F.Y. Performance Evaluation of the Deep Learning Approach for Traffic Flow Prediction at Different Times. In Proceedings of the 2016 IEEE International Conference on Service Operations and Logistics, and Informatics (SOLI), Beijing, China, 10–12 July 2016; pp. 223–227. [Google Scholar]
  52. Tian, Y.; Pan, L. Predicting Short-Term Traffic Flow by Long Short-Term Memory Recurrent Neural Network. In Proceedings of the 2015 IEEE International Conference on Smart City/SocialCom/SustainCom (SmartCity), Chengdu, China, 19–21 December 2015; pp. 153–158. [Google Scholar] [CrossRef]
  53. Ma, X.; Tao, Z.; Wang, Y.; Yu, H.; Wang, Y. Long Short-Term Memory Neural Network for Traffic Speed Prediction Using Remote Microwave Sensor Data. Transp. Res. Part C Emerg. Technol. 2015, 54, 187–197. [Google Scholar] [CrossRef]
  54. Chen, Y.y.; Lv, Y.; Li, Z.; Wang, F.Y. Long Short-Term Memory Model for Traffic Congestion Prediction with Online Open Data. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 132–137. [Google Scholar]
  55. Ma, X.; Dai, Z.; He, Z.; Ma, J.; Wang, Y.; Wang, Y. Learning Traffic as Images: A Deep Convolutional Neural Network for Large-Scale Transportation Network Speed Prediction. Sensors 2017, 17, 818. [Google Scholar] [CrossRef]
  56. Koesdwiady, A.; Soua, R.; Karray, F. Improving Traffic Flow Prediction with Weather Information in Connected Cars: A Deep Learning Approach. IEEE Trans. Veh. Technol. 2016, 65, 9508–9517. [Google Scholar] [CrossRef]
  57. Soua, R.; Koesdwiady, A.; Karray, F. Big-Data-Generated Traffic Flow Prediction Using Deep Learning and Dempster-Shafer Theory. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 3195–3202. [Google Scholar] [CrossRef]
  58. Yildirimoglu, M.; Geroliminis, N. Experienced travel time prediction for congested freeways. Transp. Res. Part B Methodol. 2013, 53, 45–63. [Google Scholar] [CrossRef] [Green Version]
  59. Lv, L.; Chen, M.; Liu, Y.; Yu, X. A plane moving average algorithm for short-term traffic flow prediction. In Advances in Knowledge Discovery and Data Mining; Springer: Ho Chi Minh City, Vietnam, 2015; pp. 357–369. [Google Scholar]
  60. Habtemichael, F.G.; Cetin, M. Short-term traffic flow rate forecasting based on identifying similar traffic patterns. Transp. Res. Part C Emerg. Technol. 2016, 66, 61–78. [Google Scholar] [CrossRef]
  61. Yao, B.; Chen, C.; Cao, Q.; Jin, L.; Zhang, M.; Zhu, H.; Yu, B. Short-Term Traffic Speed Prediction for an Urban Corridor. Comput.-Aided Civ. Infrastruct. Eng. 2017, 32, 154–169. [Google Scholar] [CrossRef]
  62. Lu, C.; Hirsch, M.; Schölkopf, B. Flexible Spatio-Temporal Networks for Video Prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6523–6531. [Google Scholar]
  63. Liu, Y.; Yan, X.; Wang, Y.; Yang, Z.; Wu, J. Grid Mapping for Spatial Pattern Analyses of Recurrent Urban Traffic Congestion Based on Taxi GPS Sensing Data. Sustainability 2017, 9, 533. [Google Scholar] [CrossRef]
  64. Custom Colormaps. Available online: https://github.com/CSlocumWX/custom_colormap (accessed on 9 June 2018).
  65. Tostes, A.I.J.; de LP Duarte-Figueiredo, F.; Assunção, R.; Salles, J.; Loureiro, A.A. From Data to Knowledge: City-Wide Traffic Flows Analysis and Prediction Using Bing Maps. In Proceedings of the 2nd ACM SIGKDD International Workshop on Urban Computing, Chicago, IL, USA, 11 August 2013; p. 12. [Google Scholar]
  66. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In Advances in Neural Information Processing Systems; Curran Associates: Montreal, QC, Canada, 2015; pp. 802–810. [Google Scholar]
  67. Tieleman, T.; Hinton, G. Lecture 6.5-Rmsprop: Divide the Gradient by a Running Average of Its Recent Magnitude. COURSERA Neural Netw. Mach. Learn. 2012, 4, 26–31. [Google Scholar]
  68. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  69. Available online: https://github.com/fchollet/keras (accessed on 22 June 2018).
  70. Internet Archive Wayback Machine. Available online: https://web.archive.org (accessed on 9 May 2018).
Figure 1. (a) A snapshot with traffic congestion levels from WSDOT. (b) Only the road network is retained after preprocessing. (c) The road network is segmented into grids. (d) The congestion of each grid is calculated, normalized, and colored for visualization. (e) Legend of congestion levels for Figure 1a as from WSDOT.
Figure 1. (a) A snapshot with traffic congestion levels from WSDOT. (b) Only the road network is retained after preprocessing. (c) The road network is segmented into grids. (d) The congestion of each grid is calculated, normalized, and colored for visualization. (e) Legend of congestion levels for Figure 1a as from WSDOT.
Sensors 19 02229 g001
Figure 2. The architecture of our proposed deep neural network for traffic congestion prediction.
Figure 2. The architecture of our proposed deep neural network for traffic congestion prediction.
Sensors 19 02229 g002
Figure 3. A sample of a sequence of images with only road networks organized chronologically for further transformations.
Figure 3. A sample of a sequence of images with only road networks organized chronologically for further transformations.
Sensors 19 02229 g003
Figure 4. Daily total wMSE errors with a prediction horizon of 10 min on 42 days evaluated with back-testing.
Figure 4. Daily total wMSE errors with a prediction horizon of 10 min on 42 days evaluated with back-testing.
Sensors 19 02229 g004
Figure 5. Daily total MAE errors with a prediction horizon of 10 min on 42 days evaluated with back-testing.
Figure 5. Daily total MAE errors with a prediction horizon of 10 min on 42 days evaluated with back-testing.
Sensors 19 02229 g005
Figure 6. Ground truth congestion levels (first row) vs. predicted congestion levels (second row) on 17 January 2017 with a prediction horizon of 10 min.
Figure 6. Ground truth congestion levels (first row) vs. predicted congestion levels (second row) on 17 January 2017 with a prediction horizon of 10 min.
Sensors 19 02229 g006
Table 1. Comparison of prediction metrics by different configurations of DCPN. Minimum wMSE values marked in bold.
Table 1. Comparison of prediction metrics by different configurations of DCPN. Minimum wMSE values marked in bold.
#Prediction Horizon Averaged Metric Model Config10 min wMSEMAE30 min wMSEMAE60 min wMSEMAE
1st512_384_256_1280.0588730.0106350.0542980.0100280.0456380.009572
2nd640_512_384_2560.0583570.0107370.0543140.0101250.0454140.009245
3rd768_640_512_3840.0618180.0127960.0583840.0128380.0491120.011893
4th896_768_640_5120.0692790.0163290.0647610.0162800.0551380.016076
5th1024_896_768_6400.0692270.0163380.0646630.0162290.0549090.015983
Table 2. Significance of difference between the top 2 configurations of DCPN using Welch’s t-test.
Table 2. Significance of difference between the top 2 configurations of DCPN using Welch’s t-test.
Prediction Horizon Averaged Metric Test Results10 min wMSEMAE30 min wMSEMAE60 min wMSEMAE
t stat0.073076−0.220181−0.002431−0.1867720.0382660.734073
p-value0.9419280.8262910.9980670.8523130.9695710.465075
Table 3. Comparison of prediction metrics using time lags of 120 and 110 minutes. Minimum wMSE values marked in bold.
Table 3. Comparison of prediction metrics using time lags of 120 and 110 minutes. Minimum wMSE values marked in bold.
Prediction Horizon Averaged Metric Time Lag (minutes)10 min wMSEMAE30 min wMSEMAE60 min wMSEMAE
1100.0587300.0107050.0542240.0101300.0453050.009293
1200.0588730.0106350.0542980.0100280.0456380.009572
Table 4. Configuration of parameters for DCPN.
Table 4. Configuration of parameters for DCPN.
LayerNameChannelsShape
0Inputs1(11, 149, 69)
1Flattern1113,091
2Dense (ReLU)1512
3Dense (ReLU)1384
4Dense (ReLU)1256
5Dense (ReLU)1128
6Dense (ReLU)1128
7Dense (ReLU)1256
8Dense (ReLU)1384
9Dense (ReLU)1512
10Dense (Sigmoid)1 149 × 69
11Dropout (0.1)————
12Dense (Sigmoid)1 149 × 69
13Reshape1(149, 69)
Table 5. MAE and wMSE by day of the whole network at different prediction horizons of 10, 30, and 60 min through back-testing. Best performance values for each day are marked with a bold typeface.
Table 5. MAE and wMSE by day of the whole network at different prediction horizons of 10, 30, and 60 min through back-testing. Best performance values for each day are marked with a bold typeface.
10 min30 min60 min
MAEwMSEMAEwMSEMAEwMSE
DaySRCNConvLSTMDCPNSRCNConvLSTMDCPNSRCNConvLSTMDCPNSRCNConvLSTMDCPNSRCNConvLSTMDCPNSRCNConvLSTMDCPN
2017-01-020.00970.00680.00350.00170.00160.00100.00830.00690.00340.00140.00160.00090.01040.00720.00670.00370.00230.0025
2017-01-030.01050.01250.01070.06600.06510.05760.01150.01240.01090.05500.05900.05430.01030.01200.00880.04900.05000.0483
2017-01-040.01040.01420.01150.05270.06130.05040.01030.01230.01040.04600.06400.04710.00930.01370.00880.03870.04760.0382
2017-01-050.01030.01280.00860.04220.05250.04040.01120.01270.00890.03880.04910.03810.01100.01490.00860.03290.04050.0315
2017-01-060.01000.01160.00680.02510.03410.02280.01030.01110.00710.02740.03850.02340.01110.01310.00770.02640.02580.0214
2017-01-090.01020.01370.01010.05710.07730.05570.01100.01220.00850.05770.07870.05310.01200.01280.00930.05160.05380.0447
2017-01-100.01100.01540.01240.09270.13070.08810.01580.01410.01180.10510.13020.08140.01270.01460.01150.09170.10570.0676
2017-01-110.01100.01430.01120.05930.07650.05820.01670.01210.01080.07040.07950.05080.01080.01380.01110.04680.05340.0395
2017-01-120.01190.01430.01270.07220.10100.07250.01300.01460.01240.09000.11220.07240.01150.01310.01170.08120.08790.0683
2017-01-130.01210.01070.00800.03370.03990.02770.01000.01130.00870.03640.03750.02590.01030.01400.00810.03850.03910.0285
2017-01-160.01330.01040.00490.00890.00750.00500.01070.00930.00580.00890.00640.00480.00970.01220.00440.00670.00870.0036
2017-01-170.01350.01350.01200.09810.12820.09510.01240.01320.01180.11180.12620.09120.01100.01280.01060.09070.09830.0752
2017-01-180.01270.01510.01260.10890.14950.11060.01320.01620.01220.13180.16210.10240.01340.01420.01070.11690.12430.0866
2017-01-190.01160.01270.01360.07500.09650.07380.01100.01320.01120.08070.09430.06650.01120.01240.00940.06790.07550.0533
2017-01-200.01190.01090.00880.02630.03460.02550.00790.01240.00830.02920.03080.02570.00720.01080.00850.02080.02420.0179
2017-01-230.01190.01140.00980.03910.05170.03960.00940.01200.00980.04550.05190.03730.00790.01080.00880.03140.03370.0264
2017-01-240.01310.01530.01220.11910.17800.11820.01310.01460.01360.13300.17350.11180.01140.01410.01170.12610.12620.0850
2017-01-250.01260.01210.01140.05680.06930.05630.00970.01240.01230.06140.06910.05460.00820.01150.00920.04800.05120.0397
2017-01-260.01240.01270.01130.06420.08510.06580.01000.01220.01080.06970.08840.06050.00930.01180.00900.05780.05950.0433
2017-01-270.01350.01000.00970.02550.02640.02400.00790.00990.00820.02430.03000.02190.00790.01050.00750.02420.02530.0187
2017-01-300.01330.01070.00990.03770.04450.03630.00850.01050.00940.03120.03990.03100.00790.01030.00730.02620.02810.0224
2017-01-310.01280.01220.01080.05600.07310.05650.00930.01110.01070.05980.06930.04930.00730.00970.00850.04500.05270.0403
2017-02-010.01220.01180.01070.06540.08380.06750.01050.01160.01050.07230.09010.06190.00860.01090.00990.06180.06850.0500
2017-02-020.01230.01120.01050.05470.06940.05640.00950.01120.01050.05820.06620.05180.00850.01130.00930.04840.05020.0408
2017-02-030.01160.01100.01000.05390.06450.05480.00950.01070.01000.05500.06180.04690.00840.01030.00790.04010.04080.0331
2017-02-060.01320.00930.00710.03400.01850.02170.01230.00950.00970.02120.01890.01080.01270.01340.00610.02250.03080.0208
2017-02-070.01150.01100.01050.03900.04240.03800.00960.01020.00790.03610.03910.03240.00960.01060.00740.02990.03070.0245
2017-02-080.01290.01370.01260.11590.15910.10640.01210.01420.01250.13130.18120.10460.01130.01330.01220.13330.16740.0943
2017-02-090.01390.01500.01440.14170.19550.12510.01360.01570.01330.15620.21830.12210.01300.01390.01230.14640.16980.1097
2017-02-100.01430.01160.01140.05380.06790.05220.01030.01200.01050.05760.07180.04880.00900.01110.01010.04720.05460.0411
2017-02-130.01420.01230.01300.07220.10230.07020.01100.01240.01220.08090.09490.06120.00950.01180.01170.06940.07950.0532
2017-02-140.01440.01260.01320.09010.11230.08730.01220.01330.01340.10540.13810.08910.01120.01200.01150.10110.11370.0815
2017-02-150.01450.01510.01300.12140.16600.11730.01420.01520.01240.13660.16760.10990.01250.01410.01200.12130.14280.0943
2017-02-160.01420.01350.01340.08170.11440.08020.01180.01350.01160.08880.11500.07330.01000.01240.01120.07830.08950.0643
2017-02-170.01400.01050.00900.03060.03600.02880.00740.01010.00830.03080.03400.02650.00720.01000.00890.02890.03010.0244
2017-02-200.01370.00920.00600.00740.00520.00340.01010.00910.00490.00650.00700.00410.01120.01160.00600.00830.00960.0048
2017-02-210.01230.01270.01290.07160.10650.07070.01010.01250.01110.07690.10670.06400.00900.01090.01080.06300.07850.0515
2017-02-220.01220.01150.01100.04100.04720.03960.00940.01130.00960.04350.05190.03410.00780.01070.00890.03070.03630.0274
2017-02-230.01240.01190.01050.04510.05580.04430.00970.01140.00940.04350.05130.03650.00750.01020.00820.03060.03480.0269
2017-02-240.01210.01000.00800.02740.02910.02360.00810.00970.00850.02490.03080.02180.00750.00960.01130.01950.02350.0206
2017-02-270.01320.01540.01280.10780.15470.10510.01270.01660.01200.11420.16720.09370.01100.01340.01190.09100.12310.0753
2017-02-280.01180.01250.01170.05860.08040.05760.00960.01200.01130.06060.08000.05220.00820.01110.01180.04850.06040.0427
Average0.01240.01230.01060.06030.07850.05790.01080.01210.01020.06470.08060.05360.00990.01200.00950.05580.06310.0449
Table 6. Computing resources for training with prediction horizons of 10, 30, and 60 min.
Table 6. Computing resources for training with prediction horizons of 10, 30, and 60 min.
10 min
SRCNConvLSTMDCPN
metric description
total number of epochs to converge876719823
total training time (s)30,646.51770,125.47121,450.032
30 min
SRCNConvLSTMDCPN
metric description
total number of epochs to converge757631845
total training time (s)26,572.62961,677.39722,235.832
60 min
SRCNConvLSTMDCPN
metric description
total number of epochs to converge769690795
total training time (s)27,585.75566,646.38120,299.434

Share and Cite

MDPI and ACS Style

Zhang, S.; Yao, Y.; Hu, J.; Zhao, Y.; Li, S.; Hu, J. Deep Autoencoder Neural Networks for Short-Term Traffic Congestion Prediction of Transportation Networks. Sensors 2019, 19, 2229. https://doi.org/10.3390/s19102229

AMA Style

Zhang S, Yao Y, Hu J, Zhao Y, Li S, Hu J. Deep Autoencoder Neural Networks for Short-Term Traffic Congestion Prediction of Transportation Networks. Sensors. 2019; 19(10):2229. https://doi.org/10.3390/s19102229

Chicago/Turabian Style

Zhang, Sen, Yong Yao, Jie Hu, Yong Zhao, Shaobo Li, and Jianjun Hu. 2019. "Deep Autoencoder Neural Networks for Short-Term Traffic Congestion Prediction of Transportation Networks" Sensors 19, no. 10: 2229. https://doi.org/10.3390/s19102229

APA Style

Zhang, S., Yao, Y., Hu, J., Zhao, Y., Li, S., & Hu, J. (2019). Deep Autoencoder Neural Networks for Short-Term Traffic Congestion Prediction of Transportation Networks. Sensors, 19(10), 2229. https://doi.org/10.3390/s19102229

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop