Meta-Extreme Learning Machine for Short-Term Traffic Flow Forecasting
Abstract
:1. Introduction
- First, we reconceptualize the improvement of traffic flow forecasting models from a model perspective and exemplify a data-driven optimization algorithm to optimize the learning model.
- Second, based on the ELM end-to-end mechanism principle, we propose a combination of the ABC algorithm and the ELM.
- Third, through sufficient experiments on four benchmark datasets, the ABC-ELM model outperforms in comparison to the most advanced models.
- Fourth, the ABC algorithm used in the ELM not only did not increase the model complexity of the ELM but also unleashed the potential of ELM in short-term traffic flow forecasting.
2. Methodology
2.1. Artificial Bee Colony Algorithm
Algorithm 1 Artificial Bee Colony(ABC) algorithm | |
1: | Set ABC parameters; |
2: | Initialize food source location; |
3: | Construct the current global optimal solution BestSol; |
4: | while Reach the maximum number of iterations do |
5: | for All employed bees do |
6: | Select food source according to Roulette Wheel Selection algorithm; |
7: | Search for a new food source near the corresponding food source ; |
8: | Calculate fitness values for new food sources newbee.Cost; |
9: | if The fitness value of the original food source pop(i)..Cost then |
10: | ; |
11: | end if |
12: | end for |
13: | for All following bees do |
14: | Search for a new food source near the corresponding food source ; |
15: | calculate fitness values for new food sources newbee.Cost; |
16: | if The fitness value of the original food source pop(i)..Cost then |
17: | ; |
18: | end if |
19: | end for |
20: | Compare and update the global optimal solution BestSol; |
21: | for All detecting bees do |
22: | if neighborhood search times then |
23: | reinitialize food source ; |
24: | ; |
25: | end if |
26: | end for |
27: | end while |
2.2. Extreme Learning Machine
2.3. ABC-ELM for the Traffic Flow Forecasting
Algorithm 2 ABC-ELM | |
1: | Initialize the ELM network; |
2: | Set activation function f(x) and the number of hidden nodes k; |
3: | Initialize the Artificial Bee Colony; |
4: | Determine colony size n, the maximum of iterations I, the upper bound of abandon L and the acceleration coefficient upper bound a; |
5: | Data pre-processing; |
6: | Input the processed samples; |
7: | while termination conditions do |
8: | for each i∈ [1, I] do |
9: | for each i∈ [1, n] do |
10: | Recruited Bees Stage; |
11: | finding new positions randomly; |
12: | evaluation; |
13: | calculate fitness values and selection probabilities; |
14: | end for |
15: | for each i∈ [1, n] do |
16: | Onlooker Bees Stage; |
17: | finding new positions with roulette algorithm; |
18: | evaluation; |
19: | end for |
20: | for each i∈ [1, n] do |
21: | Scout Bees Stage; |
22: | Discard more than L honey source and search for new; |
23: | end for |
24: | Update and store best solution ever found; |
25: | end for |
26: | end while |
27: | Get input weight w and hidden layer bias b; |
28: | Calculate output weight of hidden layer; |
29: | Test and Error calculation; |
3. Experiments
3.1. Description of Data
3.2. Experimental Setup
3.3. Evaluation Criterion
3.4. Performance Evaluation
- Autoregression (AR) is a linear regression model widely cited for forecasting. Especially in the field of applied traffic flow, AR can perform well in the field of traffic flow, which is full of randomness, because it expresses the random variables of the latter through a linear combination of the previous random variables. The parameters of the AR model are the same as those of [27].
- Artificial Neural Network (ANN) is a non-parametric learning model that simulates neuronal activity with a mathematical model and is based on mimicking the structure and function of neural networks in the brain. In this paper, the number of hidden layers of this ANN is one, and the spread of radial basis functions is set as 2000 according to the [51].
- Gray Model (GM) is based on gray system theory and can perform series forecasting, catastrophe forecasting, seasonal catastrophe forecasting, topological forecasting, and integrated system forecasting. We use the most commonly used gray forecasting model GM(1,1) model, which represents a differential equation model of order one and one variable. The GM(1,1) model group, the metabolic model, is the most desirable model.
- Kalman filter (KF), as one of the most popular filtering algorithms, is a filtering process based on linear equations to eliminate the effects of noise and other effects in the data, including the system, i.e., the process of making optimal predictions about the system. As suggested in [52], we set the initial state to , where N is set as eight.
- Decision tree (DT) is an intuitive graphical analysis method for classification and regression problem solving, where different states are generated by each decision to form a decision tree, top-down from the root node to the leaf nodes to form different paths i.e., different solutions and to find the desired optimal path. In [53], the authors used the classification and regression tree (CART) to predict the traffic flow, which is robust to missing data.
- 1.
- Since the connection weights of the input and implicit layers and the threshold of the implicit layer of ELM can be set randomly and do not need to be adjusted after setting, the connection weights do not need to be adjusted iteratively. The above are determined once by solving a system of equations, which makes the learning speed of ELM significantly higher than that of ANN.
- 2.
- ANN, as a traditional gradient-based learning algorithm, also faces more complex problems, such as falling into local optimum, overfitting, etc. Overcoming these problems often requires some cost, while ELM does not need to overcome these problems, which leads to a much simpler network structure for ELM than ANN.
- 3.
- ELM can use non-differentiable functions as activation functions, but ANN is only applicable to differentiable functions.
3.5. Ablation Study
3.6. Discussion
4. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Zhou, T.; Han, G.; Xu, X.; Han, C.; Huang, Y.; Qin, J. A learning-based multimodel integrated framework for Dynamic traffic flow forecasting. Neural Process. Lett. 2019, 49, 407–430. [Google Scholar] [CrossRef]
- Cai, L.; Yu, Y.; Zhang, S.; Song, Y.; Xiong, Z.; Zhou, T. A Sample-rebalanced Outlier-rejected k-nearest Neighbour Regression Model for Short-Term Traffic Flow Forecasting. IEEE Access 2020, 8, 22686–22696. [Google Scholar] [CrossRef]
- Zhou, T.; Jiang, D.; Lin, Z.; Han, G.; Xu, X.; Qin, J. Hybrid dual Kalman filtering model for short-term traffic flow forecasting. Iet Intell. Transp. Syst. 2019, 13, 1023–1032. [Google Scholar] [CrossRef]
- Pei, M.; Lin, P.; Liu, R.; Ma, Y. Flexible transit routing model considering passengers’ willingness to pay. Iet Intell. Transp. Syst. 2019, 13, 841–850. [Google Scholar] [CrossRef]
- Li, S.; Lyu, D.; Huang, G.; Zhang, X.; Gao, F.; Chen, Y.; Liu, X. Spatially varying impacts of built environment factors on rail transit ridership at station level: A case study in Guangzhou, China. J. Transp. Geogr. 2020, 82, 102631. [Google Scholar] [CrossRef]
- Li, S.; Zhuang, C.; Tan, Z.; Gao, F.; Lai, Z.; Wu, Z. Inferring the trip purposes and uncovering spatio-temporal activity patterns from dockless shared bike dataset in Shenzhen, China. J. Transp. Geogr. 2021, 91, 102974. [Google Scholar] [CrossRef]
- Cai, L.; Lei, M.; Zhang, S.; Yu, Y.; Zhou, T.; Qin, J. A noise-immune LSTM network for short-term traffic flow forecasting. Chaos 2020, 30, 023135. [Google Scholar] [CrossRef]
- Qin, J.; He, Z.S. A SVM face recognition method based on Gabor-featured key points. In Proceedings of the 2005 International Conference on Machine Learning and Cybernetics, Guangzhou, China, 18–21 August 2005; Volume 8, pp. 5144–5149. [Google Scholar]
- Polikar, R. Ensemble learning. In Ensemble Machine Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–34. [Google Scholar]
- Cui, Z.; Huang, B.; Dou, H.; Cheng, Y.; Guan, J.; Zhou, T. A Two-Stage Hybrid Extreme Learning Model for Short-term Traffic Flow Forecasting. Mathematics 2022, 10, 2087. [Google Scholar] [CrossRef]
- Wei, Y.; Zheng, S.; Yang, X.; Huang, B.; Tan, G.; Zhou, T. A noise-immune extreme learning machine for short-term traffic flow forecasting. In Proceedings of the International Conference on Smart Transportation and City Engineering, Chongqing, China, 6–8 August 2021; Volume 12050, pp. 599–604. [Google Scholar]
- Cai, L.; Chen, Q.; Cai, W.; Xu, X.; Zhou, T.; Qin, J. SVRGSA: A hybrid learning based model for short-term traffic flow forecasting. Iet Intell. Transp. Syst. 2019, 13, 1348–1355. [Google Scholar] [CrossRef]
- Cai, W.; Yang, J.; Yu, Y.; Song, Y.; Zhou, T.; Qin, J. PSO-ELM: A Hybrid Learning Model for Short-term Traffic Flow Forecasting. IEEE Access 2020, 8, 6505–6514. [Google Scholar] [CrossRef]
- Zheng, S.; Zhang, S.; Song, Y.; Lin, Z.; Dazhi, J.; Zhou, T. A noise-immune boosting framework for short-term traffic flow forecasting. Complexity 2021, 2021, 5582974. [Google Scholar] [CrossRef]
- Cui, Z.; Huang, B.; Dou, H.; Tan, G.; Zheng, S.; Zhou, T. GSA-ELM: A Hybrid Learning Model for Short-Term Traffic Flow Forecasting. Iet Intell. Transp. Syst. 2022, 16, 41–52. [Google Scholar] [CrossRef]
- Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
- Huang, D.R.; Song, J.; Wang, D.C.; Cao, J.Q.; Li, W. Forecasting model of traffic flow based on ARMA and wavelet transform. Comput. Eng. Appl. 2006, 42, 191–194. [Google Scholar]
- Han, C.; Song, S.; Wang, C.h. A real-time short-term traffic flow adaptive forecasting method based on ARIMA model. Acta Simulata Syst. Sin. 2004, 7, 3. [Google Scholar]
- Zhang, S.; Song, Y.; Jiang, D.; Zhou, T.; Qin, J. Noise-identified Kalman filter for short-term traffic flow forecasting. In Proceedings of the 2019 15th International Conference on Mobile Ad-Hoc and Sensor Networks (MSN), Shenzhen, China, 11–13 December 2019; pp. 462–466. [Google Scholar] [CrossRef]
- Cai, L.; Zhang, Z.; Yang, J.; Yu, Y.; Zhou, T.; Qin, J. A noise-immune Kalman filter for short-term traffic flow forecasting. Phys. A Stat. Mech. Its Appl. 2019, 536, 122601. [Google Scholar] [CrossRef]
- Shi, G.; Guo, J.; Huang, W.; Williams, B.M. Modeling seasonal heteroscedasticity in vehicular traffic condition series using a seasonal adjustment approach. J. Transp. Eng. 2014, 140, 04014012. [Google Scholar] [CrossRef]
- Williams, B.M.; Hoel, L.A. Modeling and forecasting vehicular traffic flow as a seasonal ARIMA process: Theoretical basis and empirical results. J. Transp. Eng. 2003, 129, 664–672. [Google Scholar] [CrossRef] [Green Version]
- Li, L.; Qin, L.; Qu, X.; Zhang, J.; Wang, Y.; Ran, B. Day-ahead traffic flow forecasting based on a deep belief network optimized by the multi-objective particle swarm algorithm. Knowl.-Based Syst. 2019, 172, 1–14. [Google Scholar] [CrossRef]
- Zhang, Y.; Ye, Z. Short-term traffic flow forecasting using fuzzy logic system methods. J. Intell. Transp. Syst. 2008, 12, 102–112. [Google Scholar] [CrossRef]
- Arqub, O.A.; Al-Smadi, M.; Momani, S.; Hayat, T. Application of reproducing kernel algorithm for solving second-order, two-point fuzzy boundary value problems. Soft Comput. 2017, 21, 7191–7206. [Google Scholar] [CrossRef]
- Li, L.; Qu, X.; Zhang, J.; Wang, Y.; Ran, B. Traffic speed prediction for intelligent transportation system based on a deep feature fusion model. J. Intell. Transp. Syst. 2019, 23, 605–616. [Google Scholar] [CrossRef]
- Zhou, T.; Han, G.; Xu, X.; Lin, Z.; Han, C.; Huang, Y.; Qin, J. δ-agree AdaBoost stacked autoencoder for short-term traffic flow forecasting. Neurocomputing 2017, 247, 31–38. [Google Scholar] [CrossRef]
- Lu, H.; Huang, D.; Youyi, S.; Jiang, D.; Zhou, T.; Qin, J. ST-TrafficNet: A Spatial-Temporal Deep Learning Network for Traffic Forecasting. Electronics 2020, 9, 1474. [Google Scholar] [CrossRef]
- Lu, H.; Ge, Z.; Song, Y.; Jiang, D.; Zhou, T.; Qin, J. A temporal-aware lstm enhanced by loss-switch mechanism for traffic flow forecasting. Neurocomputing 2021, 427, 169–178. [Google Scholar] [CrossRef]
- Fang, W.; Zhuo, W.; Yan, J.; Song, Y.; Jiang, D.; Zhou, T. Attention Meets Long Short-term Memory: A Deep Learning Network for Traffic Flow Forecasting. Phys. A Stat. Mech. Its Appl. 2022, 587, 126485. [Google Scholar] [CrossRef]
- Fang, W.; Zhuo, W.; Song, Y.; Yan, J.; Zhou, T.; Qin, J. Δfree-LSTM: An Error Distribution Free Deep Learning for Short-term Traffic Flow Forecasting. Neurocomputing 2023. [Google Scholar]
- Huang, B.; Dou, H.; Luo, Y.; Li, J.; Wang, J.; Zhou, T. Adaptive Spatiotemporal Transformer Graph Network for Traffic Flow Forecasting by IoT Loop Detectors. IEEE Internet Things J. 2022. [Google Scholar] [CrossRef]
- Li, H.; Yang, S.; Luo, Y.; Li, J.; Song, Y.; Zhou, T. Spatial Dynamic Graph Convolutional Network for Traffic Flow Forecasting. Appl. Intell. 2022. [Google Scholar] [CrossRef]
- Yang, S.; Li, H.; Luo, Y.; Li, J.; Song, Y.; Zhou, T. Spatiotemporal Adaptive Fusion Graph Network for Short-Term Traffic Flow Forecasting. Mathematics 2022, 10, 1594. [Google Scholar] [CrossRef]
- Zhou, T.; Dou, H.; Tan, J.; Song, Y.; Wang, F.; Wang, J. Small dataset solves big problem: An outlier-insensitive binary classifier for inhibitory potency prediction. Knowl.-Based Syst. 2022. [Google Scholar] [CrossRef]
- Dou, H.; Tan, J.; Wei, H.; Wang, F.; Yang, J.; Ma, X.G.; Wang, J.; Zhou, T. Transfer inhibitory potency prediction to binary classification: A model only needs a small training set. Comput. Methods Progr. Biomed. 2022, 215, 106633. [Google Scholar] [CrossRef] [PubMed]
- Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar]
- Ahila, R.; Sadasivam, V.; Manimala, K. An integrated PSO for parameter determination and feature selection of ELM and its application in classification of power system disturbances. Appl. Soft Comput. 2015, 32, 23–37. [Google Scholar] [CrossRef]
- Yuan, Y.; Quan, T.; Song, Y.; Guan, J.; Zhou, T.; Wu, R. Noise-immune Extreme Ensemble Learning for Early Diagnosis of Neuropsychiatric Systemic Lupus Erythematosus. IEEE J. Biomed. Health Inform. 2022, 26, 3495–3506. [Google Scholar] [CrossRef]
- Ozturk, C.; Karaboga, D. Hybrid artificial bee colony algorithm for neural network training. In Proceedings of the 2011 IEEE Congress of Evolutionary Computation (CEC), New Orleans, LA, USA, 5–8 June 2011; pp. 84–88. [Google Scholar]
- Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man, Cybern. Part B 2011, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Akay, B.; Karaboga, D. A modified artificial bee colony algorithm for real-parameter optimization. Inf. Sci. 2012, 192, 120–142. [Google Scholar] [CrossRef]
- Karaboga, D.; Akay, B. A comparative study of artificial bee colony algorithm. Appl. Math. Comput. 2009, 214, 108–132. [Google Scholar] [CrossRef]
- Raza, M.Q.; Khosravi, A. A review on artificial intelligence based load demand forecasting techniques for smart grid and buildings. Renew. Sustain. Energy Rev. 2015, 50, 1352–1372. [Google Scholar] [CrossRef]
- Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
- Mahmood, S.F.; Marhaban, M.H.; Rokhani, F.Z.; Samsudin, K.; Arigbabu, O.A. FASTA-ELM: A fast adaptive shrinkage/thresholding algorithm for extreme learning machine and its application to gender recognition. Neurocomputing 2017, 219, 312–322. [Google Scholar] [CrossRef]
- Feng, X.; Ling, X.; Zheng, H.; Chen, Z.; Xu, Y. Adaptive multi-kernel SVM with spatial–temporal correlation for short-term traffic flow prediction. IEEE Trans. Intell. Transp. Syst. 2018, 20, 2001–2013. [Google Scholar] [CrossRef]
- Wang, Y.; Van Schuppen, J.H.; Vrancken, J. Prediction of traffic flow at the boundary of a motorway network. IEEE Trans. Intell. Transp. Syst. 2013, 15, 214–227. [Google Scholar] [CrossRef]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Hsieh, W.W. Machine Learning Methods in the Environmental Sciences: Neural Networks and Kernels; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
- Zhu, J.Z.; Cao, J.X.; Zhu, Y. Traffic volume forecasting based on radial basis function neural network with the consideration of traffic flows at the adjacent intersections. Transp. Res. Part Emerg. Technol. 2014, 47, 139–154. [Google Scholar] [CrossRef]
- Gallo, C.; Conto, F.; Fiore, M. A neural network model for forecasting CO2 emission. Agris-Line Pap. Econ. Inform. 2014, 6, 31–36. [Google Scholar]
- Xu, Y.; Kong, Q.J.; Liu, Y. Short-term traffic volume prediction using classification and regression trees. In Proceedings of the 2013 IEEE Intelligent Vehicles Symposium (IV), Gold Coast, Australia, 23–26 June 2013; pp. 493–498. [Google Scholar]
- Weerakody, P.B.; Wong, K.W.; Wang, G.; Ela, W. A review of irregular time series data handling with gated recurrent neural networks. Neurocomputing 2021, 441, 161–178. [Google Scholar] [CrossRef]
Abbreviation | Description | Value |
---|---|---|
nVar | number of decision variables | 500 |
varSize | decision variables matrix size | (1500) |
varMin | decision variables lower bound | −1 |
varMax | decision variables upper bound | 1 |
maxIt | maximum number of iteration | 200 |
nPop | population size(colony size) | 100 |
nOnlooker | number of onlooker bees | 100 |
a | acceleration coefficient upper bound | 1.2 |
Abbreviation | Description | Value |
---|---|---|
N | number of hidden layer nodes | 50 |
Models | Data A1 | Data A2 | Data A4 | Data A8 |
---|---|---|---|---|
SVR | 14.34 | 12.22 | 12.23 | 12.47 |
AR | 13.57 | 11.59 | 12.70 | 12.71 |
ANN | 12.61 | 10.89 | 12.49 | 12.53 |
GM | 12.49 | 10.90 | 13.22 | 12.89 |
KF | 12.46 | 10.72 | 12.62 | 12.63 |
DT | 12.08 | 10.86 | 12.34 | 13.62 |
ELM | 11.92 | 10.32 | 12.09 | 12.58 |
GA-ELM | 11.86 | 10.30 | 11.87 | 12.26 |
GSA-ELM | 11.69 | 10.25 | 11.72 | 12.05 |
ABC-ELM | 11.40 | 9.95 | 11.26 | 11.90 |
Models | Data A1 | Data A2 | Data A4 | Data A8 |
---|---|---|---|---|
SVR | 329.09 | 259.74 | 253.66 | 190.30 |
AR | 301.44 | 214.22 | 226.12 | 166.71 |
ANN | 299.64 | 212.95 | 225.86 | 166.50 |
GM | 347.94 | 261.36 | 275.35 | 189.57 |
KF | 322.03 | 239.87 | 250.51 | 187.48 |
DT | 316.57 | 224.79 | 243.19 | 238.35 |
ELM | 300.67 | 208.84 | 224.54 | 172.69 |
GA-ELM | 291.42 | 211.43 | 228.57 | 169.25 |
GSA-ELM | 287.89 | 203.04 | 221.39 | 163.24 |
ABC-ELM | 286.25 | 200.42 | 220.07 | 163.67 |
Models | Data A1 | Data A2 | Data A4 | Data A8 |
---|---|---|---|---|
ELM | 15.429 | 15.544 | 15.168 | 14.267 |
GA-ELM | 15.416 | 15.522 | 15.136 | 14.161 |
GSA-ELM | 15.324 | 15.431 | 15.066 | 14.151 |
ABC-ELM | 15.323 | 15.429 | 14.992 | 14.136 |
Aggregation Intervals | Data A1 | Data A2 | Data A4 | Data A8 |
---|---|---|---|---|
5 min | 12.68 | 11.50 | 11.68 | 13.03 |
10 min | 11.40 | 9.95 | 11.26 | 11.90 |
20 min | 13.44 | 12.98 | 15.68 | 15.79 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, X.; Li, L.; Huang, B.; Dou , H.; Yang, X.; Zhou, T. Meta-Extreme Learning Machine for Short-Term Traffic Flow Forecasting. Appl. Sci. 2022, 12, 12670. https://doi.org/10.3390/app122412670
Li X, Li L, Huang B, Dou H, Yang X, Zhou T. Meta-Extreme Learning Machine for Short-Term Traffic Flow Forecasting. Applied Sciences. 2022; 12(24):12670. https://doi.org/10.3390/app122412670
Chicago/Turabian StyleLi, Xin, Linfeng Li, Boyu Huang, Haowen Dou , Xi Yang, and Teng Zhou. 2022. "Meta-Extreme Learning Machine for Short-Term Traffic Flow Forecasting" Applied Sciences 12, no. 24: 12670. https://doi.org/10.3390/app122412670
APA StyleLi, X., Li, L., Huang, B., Dou , H., Yang, X., & Zhou, T. (2022). Meta-Extreme Learning Machine for Short-Term Traffic Flow Forecasting. Applied Sciences, 12(24), 12670. https://doi.org/10.3390/app122412670