Next Article in Journal
(ω,ρ)-BVP Solutions of Impulsive Differential Equations of Fractional Order on Banach Spaces
Next Article in Special Issue
A Point-of-Interest Recommender System for Tourist Groups Based on Cooperative Location Set Cover Problem
Previous Article in Journal
An Effective Approach Based on Generalized Bernstein Basis Functions for the System of Fourth-Order Initial Value Problems for an Arbitrary Interval
Previous Article in Special Issue
Object-Centric Process Mining: Unraveling the Fabric of Real Processes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Forecasting of a Leading Country’s Government Expenditure Using a Recurrent Neural Network with a Gated Recurrent Unit

by
Cheng-Hong Yang
1,2,3,4,5,
Tshimologo Molefyane
2 and
Yu-Da Lin
6,*
1
Department of Information Management, Tainan University of Technology, Tainan 710302, Taiwan
2
Department of Electronic Engineering, National Kaohsiung University of Science and Technology, Kaohsiung 80778, Taiwan
3
Ph.D. Program in Biomedical Engineering, Kaohsiung Medical University, Kaohsiung 80708, Taiwan
4
School of Dentistry, Kaohsiung Medical University, Kaohsiung 80708, Taiwan
5
Drug Development and Value Creation Research Center, Kaohsiung Medical University, Kaohsiung 80708, Taiwan
6
Department of Computer Science and Information Engineering, National Penghu University of Science and Technology, Magong 880011, Taiwan
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(14), 3085; https://doi.org/10.3390/math11143085
Submission received: 15 June 2023 / Revised: 3 July 2023 / Accepted: 11 July 2023 / Published: 12 July 2023
(This article belongs to the Special Issue Advances in Business Intelligence: Theoretical and Empirical Issues)

Abstract

:
Economic forecasting is crucial in determining a country’s economic growth or decline. Productivity and the labor force must be increased to achieve economic growth, which leads to the growth of gross domestic product (GDP) and income. Machine learning has been used to provide accurate economic forecasts, which are essential to sound economic policy. This study formulated a gated recurrent unit (GRU) neural network model to predict government expenditure, an essential component of gross domestic product. The GRU model was evaluated against autoregressive integrated moving average, support vector regression, exponential smoothing, extreme gradient boosting, convolutional neural network, and long short-term memory models using World Bank data regarding government expenditure from 1990 to 2020. The mean absolute error, root mean square error, and mean absolute percentage error were used as performance metrics. The GRU model demonstrates superior performance compared to all other models in terms of MAE, RMSE, and MAPE (with an average MAPE of 2.774%) when forecasting government spending using data from the world’s 15 largest economies from 1990 to 2020. The results indicate that the GRU can be used to provide accurate economic forecasts.

1. Introduction

Government budgets for education, health care, defense, and welfare are large [1], and they cannot be accurately predicted using traditional econometric models because of the complex nonlinear relationships existing between variables [2]. Researchers have recently applied artificial intelligence to help governments better allocate resources, deliver services, and coordinate large-scale operations [3]. A country’s economic conditions and priorities are reflected in various factors that affect government spending, such as population size, which significantly impacts government spending [4]. A large population often requires increased investment in public services, healthcare, education, and social security programs. For example, in 2020 (World Development Indicators, 2021), countries with larger populations allocated a substantial portion of their budgets to social welfare, with spending on social welfare accounting for 30% to 45% of the total budget. The demand for social services varies across country, encompassing areas such as poverty alleviation, healthcare, education, and unemployment benefits, thereby influencing government spending decisions. Gross domestic product (GDP) indicates a country’s economic performance and total output. Government spending and GDP are closely related. During periods of GDP growth, governments often allocate more resources to infrastructure development and public services in order to support further economic growth. Conversely, during economic downturns or recessions, governments may implement austerity measures and reduce spending to manage budget deficits [5]. Government spending is also influenced by other macroeconomic variables, such as inflation, unemployment rates, and interest rates. Rising inflation rates can lead to increased spending on subsidies and welfare programs to mitigate the impact of price increases on citizens’ purchasing power [6]. Through measures such as job creation programs or unemployment benefits, unemployment rates can affect government spending. In addition, interest rates play a crucial role in determining the cost of government borrowing, which affects spending decisions. By taking into account these factors related to government spending, policymakers can make informed decisions about budget allocation, revenue generation, and overall economic management with the economy to promote sustainable development and meet societal needs [7].
According to the Organisation for Economic Co-operation and Development (OECD), government investments in education and health care contribute to long-term economic growth; specifically, every 1% increase in education spending was reported to be correlated with a 0.5% increase in GDP per capita over the long term (source: OECD. Education at a Glance 2020: OECD Indicators). However, the relationship between government spending and GDP is complex; it depends on several factors, such as the state of the economy, the type and composition of government spending, and the effectiveness of government policies [8,9]. Therefore, accurate forecasts of government spending are crucial for improving resource allocation and economic planning. Accurate forecasts help governments budget more effectively [1] and with less risk [2]. Jeong et al. demonstrated that neural networks can be used to improve traditional time series models, such as seasonal autoregressive moving average (ARIMA) models, when applied to budget-related prediction [10]. Neural networks are a powerful means of capturing complex, nonlinear relationships between variables [11]. However, neural networks can accurately predict government spending only if a suitable set of factors that affect spending are selected; however, this is difficult to achieve. For example, Palmer et al. found that neural networks accurately predict economic tourism demand only if the appropriate input variables and network architecture are selected [12]. Despite these challenges, neural networks are promising as a means to predict government spending, even with complex time series data.
Machine learning models have produced accurate forecasts on the basis of time series data. For example, Lago et al. demonstrated that a gated recurrent unit (GRU) model (a deep learning model) outperformed its traditional counterparts in accurately predicting spot electricity prices. The GRU model had the highest prediction accuracy with a symmetric mean absolute percentage error (symmetric MAPE) of 13.04%, outperforming a deep neural network (the second best model with a symmetric MAPE of 12.34%), a long short-term memory (LSTM) model, a convolutional neural network (CNN) model, and even several mixed models [13]. Another study formulated a hybrid recurrent neural network (RNN) model based on GRU and LSTM that predicts daily and hourly multilevel wind power in Germany from data on wind speed at different heights [14]. The proposed model outperformed ARIMA and support vector regression (SVR) models in terms of gradient stability, training speed, mean absolute error (MAE), root mean square error (RMSE), and t-test results on long sequences. Li et al. proposed a novel ensemble decision method based on deep reinforcement learning to predict GDP. In this method, predictions from GRU, temporal convolutional network, and deep belief network models were taken as input to train three GDP prediction models, and the deep Q network algorithm was used to optimize the integration weight coefficients. This proposed method outperformed 18 competing methods in evaluation experiments, achieving MAPE values below 4.2% in all tests [15].
Government spending in the economy takes many forms. The composition of GDP among different types of spending is often of interest to economists. Government spending is one of four components of GDP, including consumption, investment, government purchases, and net exports. According to the International Monetary Fund (https://www.imf.org/en/Home), a 1% increase in government spending as a share of GDP is correlated with a 0.4 to 1.2 percentage point increase in GDP growth, depending on how government spending is allocated. Conversely, decreased government spending can exert contractionary effects and even lead to recession. Government spending is also related to other macroeconomic variables. For example, infrastructure spending boosts investment and productivity, and education and healthcare spending enhances the productivity and health of the workforce in particular and the population in general; these factors promote long-term economic growth.
In this study, we reviewed existing forecasting methods and developed a GRU model that predicts government spending on the basis of financial GDP indicators. We used a historical dataset to improve model generalizability. Our model outperformed six competing methods in terms of accuracy and reliability in an extensive battery of experiments involving time series datasets on the GDPs of 15 of the largest economies in the world. Our contributions are as follows.
  • Our GRU model accurately predicts government spending on the basis of financial GDP indicators.
  • Our GRU model outperformed ARIMA, exponential smoothing (ETS), extreme gradient boosting (XGBoost), SVR, CNN, and LSTM models in evaluation experiments.
  • The strengths and weaknesses of neural network models in predicting government expenditure were explored.
  • The aforementioned methods can capture complex nonlinear relationships between different economic factors to generate accurate predictions.
This paper is organized as follows. Section 2 reviews existing prediction approaches, namely, ARIMA, ETS, support vector machine (SVM), XGBoost, CNN, LSTM, and GRU, and presents our proposed approach. Section 3 describes the setup and results of our evaluation experiments. Finally, Section 4 concludes the paper.

2. Methods

Statistical models, especially ARIMA and exponential smoothing, are widely used in time series forecasting because they can capture patterns and dependencies in time series data [16]. Machine learning models are an improvement on such statistical methods and are becoming increasingly popular. These models are summarized as follows.

2.1. Autoregressive Integrated Moving Average

An ARIMA model describes the temporal evolution of the dependence between observations using the time lag between observations. It can be used to model stationary time series data wherein the mean and variance of the data remain constant over time. An ARIMA model combines autoregressive, differenced, and moving average components to model time series data [17] and has three main components: the number of lagged observations (p), the difference in nonseasonal observations (d), and the size of the moving average window (q). These three components must be applied in the order expressed in the tuple (p, d, q). In our experiment, the Python library pmdarima was used to implement ARIMA modeling. Subsequently, the auto.arima function in this package was used to create a process to automatically select the best parameters for the ARIMA model on the basis of the data. This helped the model quickly generate predictions.

2.2. Exponential Smoothing

ETS is used to model patterns and dependencies over time, primarily in nonstationary time series data. ETS models output predictions by assigning exponentially decaying weights to past observations [18]. The Holt linear exponential smoothing method is commonly used, and predictions can be obtained using the following equations:
y ^ t = α y t + ( 1 α ) ( y ^ t 1 + b t 1 )
b t = γ y ^ t y ^ t 1 + ( 1 γ )   b t 1
y ^ t + m = y ^ t b t m
where α and γ are smoothing constants ranging from 0 to 1. These standard equations are used to obtain the exponential trend detailed in [19]. In ETS, these techniques are used to simulate fundamental trends and patterns, primarily for nonstationary time series data. ETS models are popular because they are easy to implement and can rapidly provide accurate predictions [20,21].

2.3. Support Vector Regression

Introduced by Vapnik et al., SVM is a machine learning algorithm that can be used for classification and regression problems. SVR is a variant of SVM and is used to solve regression problems to make predictions on continuous target variables. SVR can be used when the relationship between the independent and dependent variables is nonlinear [22]. In regression, an SVR model identifies the best-fitting hyperplane with the maximum number of points. For a hyperplane Y = wx + b, the decision boundaries are wx + b = +a and wx + b = −a. Therefore, any hyperplane that satisfies SVR must lie in the bounds −a < (Ywx + b) < +a. Thus, the goal is to identify a function that satisfies the decision boundary. SVR has been used to forecast, for example, the demand for wind and solar energy [23], the exchange rate of the Euro [24], the state of the climate [25], and the volume of airport freight [26]. In previous studies, SVR or hybrid methods with SVR have performed the best in evaluation experiments.

2.4. Extreme Gradient Boosting

A versatile and increasingly popular method designed to balance between performance and computational cost, XGBoost is an optimized version of the gradient boosting algorithm, and uses ensemble learning with multiple decision trees for prediction [27]. XGBoost has been applied in various fields, such as health care [28], computer vision [29], and missing data imputation [30].

2.5. Convolution Neural Network

CNNs consist of multiple layers and can be trained using a back-propagation algorithm. CNNs contain three layers: convolution, pool, and fully connected. The convolutional layer is responsible for learning the input’s feature representation. It utilizes multiple convolutional kernels to compute different feature maps. In particular, the neurons in each feature map are connected to a set of neighboring neurons in the previous layer, and this set of neighboring neurons is called the receptive field of the neuron in the previous layer. To obtain a new feature map, the input is convolved with the trained convolution kernel. The resulting convolution is then subjected to an element-by-element nonlinear activation function. The activation function introduces nonlinear properties, which are advantageous for multilayer networks to detect nonlinear features. Commonly used activation functions include Sigmoid, tanh [31], and ReLU [Ref] the pooling layer is typically described as being located between two convolutional layers, and is responsible for achieving translation invariance by reducing the resolution of the feature maps. Each feature map in the pooling layer is connected to the corresponding feature map in the previous layer. In the initial layer, the convolution kernel is designed to detect low-level features, such as edges and curves. Higher-level convolution kernels learn to encode more abstract features. By progressively stacking multiple convolutional and pooling layers, we can gradually extract higher-level feature representations. One or more fully connected layers designed for higher-level inference [30] may follow the multiple convolutional and pooling layers. The fully connected layer establishes connections between all neurons in the preceding layer and each neuron in the current layer, allowing for the generation of global semantic information.

2.6. Long Short-Term Memory

LSTM is a type of RNN specifically designed to overcome the vanishing gradient problem in traditional RNNs. It was introduced by Hochreiter and Schmidhuber in 1997 and has since been widely applied in various fields, including time series forecasting [32]. In LSTM, time series data are formulated as a sequence, wherein each observation corresponds to a time step. LSTM neural networks are then trained using previous observations to predict the next value in the sequence. Long-term dependencies in the data can be captured when the model is configured to use multiple previous time steps for each prediction. In recent studies, LSTM has had promising results in time series prediction. For example, LSTMs have produced more accurate forecasts of electricity load [33] and stock market trends [34] relative to competing methods. LSTM has also been used to predict wind power output [35], solar radiation [36], and agricultural production [37] from time series data. The LSTM architecture consists of a collection of recurrently connected memory blocks, which are subnetworks that maintain their state over time and regulate the flow of information through nonlinear gates. The structure of a single LSTM cell is shown in Figure 1A. These gates govern the interactions between different cells and regulate the flow of information. The input gate controls the process of updating the memory state. The output gate determines whether the output flow can influence the memory state of other cells. The forgetting gate determines whether the prior states should be remembered or forgotten. The LSTM is implemented using the following composite functions:
i t = σ W x i x t + W h i h t 1 + W c i c t 1 + b i
f t = σ W x f x t + W h f h t 1 + W c f c t 1 + b f
c t = f t 1 + i t t a n h ( W x c x t + W h c h t 1 + b c )
o t = σ W x o x t + W h o h t 1 + W c o c t + b o
h t = o t t a n h ( c t )
where σ represents the logistic sigmoid function, i, f, o, and c represent the input gate, forget gate, output gate, and cell input activation vectors, respectively. h represents the hidden vector. The subscripts of the weight matrix have intuitive meanings. For example, Whi represents the hidden input gate matrix, and so on.

2.7. Gated Recurrent Unit

GRUs models are efficient, have fast training speed, and can address vanishing gradient problems in traditional RNNs, making them suited to applications involving large-scale time series prediction [38]. GRU methods have performed well in various applications, such as in the prediction of gas concentrations in mines [39] and the prediction of COVID-19 mortality rates [40].
Cho et al. (2014) proposed an architecture wherein GRU gate mechanisms, including a reset gate rt and update gate Zt, are used to efficiently handle vanishing gradient problems (Figure 1A) [33]. The reset gate is essential for the detection of short-term dependencies within sequences, allowing a network to selectively forget information from the previous time step and focus on the information that is most relevant to the current task. The update gate is responsible for capturing long-term dependencies in sequential data and determining how much information from the previous time step should be passed to the current time step; these determinations allow the network to adapt to changes in the input data and learn complex patterns over a long period.
During training, a GRU model learns the optimal values of the gates and other parameters by using the input data Xt and the expected output. A trained GRU model then outputs predictions based on input sequences by using the learned parameters. The learning process is encapsulated in Equations (9)–(12), which describe the units that capture the respective dependencies within the GRU network [41]:
Z t = σ h t 1 W x z x t + W h z h t 1
r t = σ W x r x t + W h r h t 1
h t = z t h t 1 + 1 z t   h ~ t
  h ~ t = tan h ( W x h x t + W h h ( r t h t 1 ) )
In these equations, two fully connected layers with a sigmoid activation function σ produce the output of two gates. W x z , W x r , W h z , and W h r denote the weight parameters of the reset gate and the update gate. The operator refers to element-wise multiplication, and refers to addition. When r t is integrated and updated, the cell uses the hyperbolic tangent activation function tanh at time step t to reach the candidate state defined in Equation (7). Subsequently, for the final memory update to be determined, Equation (6) is integrated with Z t , and the hidden state h t that matches the hidden state at time step t − 1 is identified. Finally, if the new candidate state of the hidden layer conforms to the description given in Equation (7), the final memory state h t is reached.

2.8. GDP Indicators Forecasting Framework

Figure 2 summarizes the flow of the experiment in which models were evaluated in terms of their performance in predicting government expenditure. Our proposed GRU model was designed to have a reset gate rt and update gate Zt (Figure 1A,B and Figure 3; Algorithm 1) to capture dependencies between different periods. Specifically, the data were obtained and subjected to preprocessing (e.g., missing value evaluation and feature normalization). These data were then divided into training and testing data. Subsequently, the performances of trained models using various methods were evaluated in terms of MAE, RMSE, and mean absolute percentage error (MAPE). The results are visualized graphically. The framework detail of GRU algorithm is represented in Figure 3.
Algorithm 1 GRU for Government Expenditure Forecasting
Let X be the dataset.
Desired Output: Prediction of best model F ( x )
1.  
Load time-series dataset, reconstruct, standardize, and preprocess the data
2.  
Split the data into training and testing sets, defined as X t r a i n and X t e s t , respectively
3.  
Define GRU model structure and parameters
4.  
Update gate Z t is obtained using (4)
5.  
Reset gate   r t is obtained using (5)
6.  
Candidate memory state h ~ t is obtained using (7)
7.  
Weighted variables, W are initialized
8.  
Final memory state h t is computed using (6)
9.  
Calculate the loss function according to arg min W L W = i = 1 N l o s s   y i ,   y ^ I , where l o s s   y i ,   y ^ i = 1 N i = 1 N y i y ^ i
10.
Use X t r a i n to train the model
11.
Evaluate model F ( x ) predictions using X t e s t
12.
Select the best model using performance metrics MAE, RMSE, and MAPE
13.
Visualize and analyze the forecasting of the best model F ( x )

2.9. Evaluation Criteria

MAE, MAPE, and RMSE are widely used metrics of rolling forecasting performance. They describe the deviation between actual and predicted values, and are defined as follows:
MAE = 1 N i = 1 N y i y ^ i
MAPE = 1 N i = 1 N y i y ^ i y i × 100 %
RMSE = 1 N i = 1 N y i y ^ i 2
where y i is the actual value, y ^ i is the predicted value, and N is the sample size. The MAE is defined as the mean of the absolute differences between the actual and predicted values, and ranges from 0 to positive infinity. The MAPE is a relative indicator that is invariant to the units or magnitude of the actual and predicted values. Finally, the RMSE is the square root of the quotient that is the average deviation between actual and predicted values divided by the number of observations. The RMSE is sensitive to minor errors in model predictions, and is thus an effective metric of accuracy. For the three metrics, lower values (i.e., values closer to 0) indicate a lower error and higher prediction accuracy. These metrics are designed to comprehensively assess our forecasts’ accuracy, precision, and reliability. Since MAE calculates the absolute difference between predicted and actual values, it provides an unbiased average accuracy metric. This provides a visual interpretation of the error and insight into the degree of deviation from the true value. On the other hand, RMSE emphasizes the effect of larger errors by considering squared differences. It provides a measure of the dispersion and magnitude of the error, and can assess the accuracy and reliability of the prediction. Finally, MAPE reflects the relative error between the predicted and actual values, providing insight into the proportional accuracy of the forecast. Each metric captures a different aspect of prediction performance, enabling a comprehensive understanding of the model’s strengths and weaknesses. These metrics form a robust framework for assessing the forecasting power of the model and the applicability of GDP-based government spending forecasts. Using these performance metrics allows for a broad assessment of forecasting methods. The choice of these metrics depends on the specific situation and the nature of the forecasting task. They provide information on different aspects that help to understand the model’s accuracy, precision, and reliability. By considering these metrics together, a comprehensive assessment of model performance can be obtained, and the strengths and weaknesses of the model can be identified.

3. Results and Discussion

3.1. Data Source

To ensure a comprehensive analysis, we will focus on data from 1990 to 2020 sourced from the World Bank (https://data.worldbank.org/). This time horizon captures a period of significant economic fluctuations, including major events such as financial crises, recessions, and economic growth. By considering a broad time horizon, we aim to understand government spending behavior over time comprehensively. In selecting countries, we base our selection on their respective GDP rankings. Specifically, we selected the 15 countries with the largest GDP to ensure representation of the major global economies. These countries are the United States, China, Japan, Germany, the United Kingdom, India, France, Italy, Canada, South Korea, Russia, Brazil, Australia, Spain, and Mexico. By including these countries, we aim to capture the various economic systems, levels of development, and geopolitical factors that may influence government spending patterns. The dataset was downloaded as a .csv file from the World Bank website and then transformed into a data frame in Python. The dataset obtained from the World Bank was of high quality, and did not contain any missing values. The descriptive statistics (specifically, mean, median, standard deviation, sample size, first quartile [Q1], third quartile [Q3], and interquartile range [IQR]) for each of these 15 countries are presented in Table 1. We standardized the data using the integrated Python function min_max_scaler; this function rescales a data point x to between 0 and 1 using the following equation: (x − xmin)/(xmax − xmin), where x is the value of the data point, xmin is the minimum value, and xmax is the maximum value.

3.2. GRU Architecture Results and Sensitivity Analysis

We employed three optimizer tools, namely Adam, RMSprop, and Adagrad, to detect the optimal architecture for each neural network. The CNN model contains a 1D convolutional layer (CONV1D) with 200 filters and kernel size 1, followed by a MaxPooling layer. The LSTM model has two layers of 200 units each, followed by a one-cell dense layer. The GRU model has two layers of 200 units each, followed by a separate dense layer of 1 unit. The Adam optimizer and the mean absolute error loss function were used to create these neural network models. To capture the historical dependencies, the training data were prepared using a backward window with 1 and 1000 periods for each model. Table 2 presents a comparison of the parameter settings and the individual MAPE (%) results as a measure of precision. The adjusted period values, the number of neural network layers, and the activation function are shown. For the RMSprop optimizer, we increased the number of parameters while decreasing the number of periods. The overall MAPE value is slightly higher than the value in Table 2. For the Adagrad optimizer, we reduced the parameter and period values. It is important to note that the period values should not be excessively high, as they can lead to overfitting. The LSTM shows slightly better values with the parameter settings of the Adagrad optimizer. The CNN model shows a MAPE value of 86.798%, indicating a much worse performance. After a training procedure based on the optimizer, Adam optimizer produced an optimal GRU architecture, and the architectural detail is shown in Table 2.
We performed a sensitivity analysis, including experiments with parameter settings for different scenarios. The number of units in the GRU layer increased and decreased slightly. In addition, we adjusted the values of epochs to different lower values to observe the model’s performance over time, since too many epochs can lead to overfitting. Table 3 presents the results of our sensitivity analysis. From Table 3, we can observe that lower epoch values, in combination with a higher number of GRU layers, tend to produce better prediction results for all three evaluated metrics. Notably, the differences between the accuracy results from the sensitivity analysis are not substantial. This indicates that our GRU model compiles correctly and performs consistently under different assumptions and scenarios. The chosen optimizer is appropriate for our dataset, and can produce sufficiently accurate results, as shown in Table 3. The sensitivity analysis strengthens the robustness of our model and provides valuable insight into the effects of varying the parameters used. It allows us to understand the optimal parameter settings for our particular task of forecasting government spending. The consistent performance of our model across different parameter configurations gives us confidence in its accuracy and reliability.

3.3. Experimental System

Python was used for data mining. Our GRU model was evaluated against ARIMA, ETS, SVR, XGBOOST, CNN, and LSTM models. The LSTM and GRU models were optimized using Adam [42]. Different learning rates α ranging from 0.1 to 0.0001 were tested, and the corresponding validation loss was observed. Finally, a learning rate of 0.01 was chosen, and β1 and β2 were set to their default values of 0.9 and 0.999, respectively. The training and testing sets comprised data from 1990–2014 and 2015–2020, respectively. Figure 4 and Table 4 present the MAE, RMSE, and MAPE results of the seven models for all and each of the 15 countries, respectively.

3.4. Comparison and Discussion

Artificial intelligence encompasses a range of techniques that aim to mimic living things’ behavior to improve decision-making and mitigate potential risks to economic stability and growth. One such technique is GRU, which focuses on automatically identifying meaningful patterns in data. GRU has become a valuable tool for extracting information from large datasets [43]. Departure from traditional computing methods is a common feature of GRU applications. Due to the complexity of the patterns identified, it is impractical for a human programmer to provide explicit and detailed instructions for performing these tasks. In addition, the GRU techniques can learn from a wide variety of data types and recognize patterns that may be impossible for a human to detect. This enables GRU to operate in resource-constrained environments [44]. As a result, GRU could play a critical role in public administration, and provide insights that can inform decision-makers.
The GRU model was evaluated against ARIMA, ETS, SVR, XGBOOST, CNN, LSTM, BiRNN, and ASRNN models. The LSTM, BiRNN, ASRNN, and GRU models were optimized using Adam [42]. ARIMA had much lower performance compared with the other models. To improve ARIMA performance, the lag order (p), moving average order (q), and differencing degree (d) can be manually rather than automatically configured [45], and other diagnostic measures can be applied [46,47]. The remaining models produced reasonably accurate predictions. We analyzed different RNN variants, including the bidirectional RNN (BiRNN) [48] and attention-based [49] sequential RNN (ASRNN) models. For the BiRNN model, we performed dataset preprocessing and used a BiRNN architecture along with a dropout layer to prevent overfitting. Similarly, for the ASRNN model, the dataset underwent the same preprocessing step, and the architecture employed an attention mechanism to capture the relevant information from the sequential data. Different learning rates α ranging from 0.1 to 0.0001 were tested, and the corresponding validation loss was observed. Finally, a learning rate of 0.01 was chosen, and β1 and β2 were set to their default values of 0.9 and 0.999, respectively. The training and testing sets comprised data from 1990–2014 and 2015–2020, respectively. Figure 4 and Table 4 present the MAE, RMSE, and MAPE results of the seven models for all and each of the 15 countries, respectively.
Our GRU model performed the best when used to forecast government expenditure (Table 4). The average MAE, RMSE, and MAPE values for country-level predictions were 0.5284, 0.7469, and 2.774%, respectively. ETS, XGBoost, CNN, BiRNN, and ASRNN also produced satisfactory country-level predictions. The country-level predictions of the seven models and the actual values are plotted in Figure 5. The actual and predicted trends for all models except the ARIMA model had a close fit upon visual inspection. Accurate predictions for China were difficult to achieve, although the predictions of ETS, CNN, BiRNN, ASRNN, and GRU were reasonably close. In general, GRU captured outliers and changes in trends well, unlike ARIMA. Some methods, such as LSTM and SVR, follow an average trend starting from a particular value. Data spanning a longer period are required to more fully analyze the capabilities of these methods, especially GRU.
Since GRU uses only two gates instead of three, and one memory cell, it is considered an efficient implementation of LSTM. Despite the simplicity of GRU, it still achieves better results than LSTM [43]. LSTM was introduced to address the problems of vanishing and exploding gradients in RNNs. LSTM uses memory cells and gates to facilitate the flow of information over time. Subsequently, GRU was introduced as a more efficient form of LSTM, which combines two LSTM gates without a memory cell and reveals the full hidden content without any control [41]. Compared to LSTM, GRU, therefore, exhibits higher computational efficiency. For the advantage of input information, bidirectional RNNs were introduced. BiRNN uses information from both directions to predict output sequences by cascading and connecting backward and forward hidden states to each output node. The GRU addresses the problems of gradient explosion and the long-term dependency of RNNs, and requires fewer training parameters compared to the LSTM, which is also a variant of the RNN. Both the GRU and the LSTM have their advantages and disadvantages in practical applications [41]. Determining effective hyperparameters, such as the number of neuron units and the learning rate, remains a challenging problem, despite GRU’s good performance in time series prediction and its wide applicability. Optimizers are powerful tools for the optimization of hyperparameters. In this study, the hyperparameters of the GRU were optimized using Adam, RMSprop, and RMSprop. Additionally, a sensitivity analysis improved the robustness of our model and provided valuable insight into how changing the parameters affected the model. As a result, we were able to understand the optimal parameter settings for the prediction of government expenditure, the specific task at hand. We are confident in the accuracy and reliability of our model because of its consistent performance across different parameter configurations.
The choice of input variables is critical to the accuracy of expenditure forecasting. Traditionally, a priori expert knowledge, trial and error, or linear cross-correlation analysis are used to select these variables. However, expert knowledge is often challenging and time-consuming to acquire, and relying solely on expert knowledge can introduce bias. Trial and error input selection can be computationally intensive, especially for data-driven models with many potential input candidates. In addition, the linear correlation coefficients commonly used only assess linear relationships and do not capture the non-linear dynamics often present in data-driven models. Artificial intelligence encompasses a range of techniques that aim to mimic living things’ behavior to improve decision-making and mitigate potential risks to economic stability and growth. One such technique is GRU, which focuses on automatically identifying meaningful patterns in data. GRU has become a valuable tool for extracting information from large datasets [43]. Departure from traditional computing methods is a common feature of GRU applications. Due to the complexity of the patterns identified, it is impractical for a human programmer to provide explicit and detailed instructions for performing these tasks. In addition, GRU techniques can learn from a wide variety of data types and recognize patterns that may be impossible for a human to detect. This enables GRU to operate in resource-constrained environments [45]. As a result, GRU could play a critical role in public administration and provide insights that can inform decision-makers.
In general, GRU outperformed ARIMA, SVR, ETS, XGBoost, CNN, LSTM, BiRNN, and ASRNN in predicting government expenditure on the basis of data on prior government spending, household consumption, investment, imports, and exports, which are correlated with GDP. Furthermore, unlike ARIMA, GRU does not require refitting to produce accurate predictions. ETS, CNN, LSTM, BiRNN, and ASRNN also provided reasonably accurate predictions.

4. Conclusions

Algorithms such as GRU are becoming increasingly capable of economic prediction as machine learning becomes more sophisticated and datasets become larger. In our study, our proposed GRU method outperformed ARIMA, ETS, SVR, XGBOOST, CNN, and LSTM in terms of MAE, RMSE, and MAPE (average MAPE = 2.774%) when used to predict government spending on the basis of 1990–2020 data from 15 of the largest economies in the world, namely Australia, Brazil, Canada, China, France, Germany, India, Italy, Japan, South Korea, Mexico, Russia, Spain, the United Kingdom, and the United States. Further studies should incorporate predictors pertaining to political events, social changes, and the environment, and apply the resultant models to emerging markets. In general, accurate forecasts of economic trends are indispensable for ensuring sustainable economic growth.

Author Contributions

C.-H.Y.: conceptualization, supervision, project administration. T.M.: methodology, writing—review and editing. Y.-D.L.: conceptualization, project administration. All authors have read and agreed to the published version of the manuscript.

Funding

The funding source is the Ministry of Science and Technology, Taiwan (under grant no. 111-2221-001-MY3 and 111-2221-002-MY3).

Data Availability Statement

Data were obtained from World Bank (https://data.worldbank.org/).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kumar, R.; Lal, K. Forecasting government expenditure using machine learning algorithms: Evidence from India. Technol. Forecast. Soc. Chang. 2017, 123, 251–260. [Google Scholar]
  2. Harmsen, R.; Sterneberg, A. Forecasting government expenditure with macroeconomic aggregates in small open economies. Econ. Model. 2017, 60, 302–312. [Google Scholar]
  3. Dwivedi, Y.K.; Hughes, L.; Ismagilova, E.; Aarts, G.; Coombs, C.; Crick, T.; Duan, Y.; Dwivedi, R.; Edwards, J.; Eirug, A. Artificial Intelligence (AI): Multidisciplinary perspectives on emerging challenges, opportunities, and agenda for research, practice and policy. Int. J. Inf. Manag. 2021, 57, 101994. [Google Scholar] [CrossRef]
  4. Lara-Rubio, J.; Navarro-Galera, A.; Buendía-Carrillo, D.; Gomez-Miranda, M.E.J.C. Analysing financial risks of local governments to design sustainability policies for public services: An empirical study by the population size. Cities 2022, 128, 103795. [Google Scholar] [CrossRef]
  5. Wei, X.; Mohsin, M.; Zhang, Q.J.R.E. Role of foreign direct investment and economic growth in renewable energy development. Renew. Energy 2022, 192, 828–837. [Google Scholar] [CrossRef]
  6. Ginn, W.; Pourroy, M.J.E.M. The contribution of food subsidy policy to monetary policy in India. Econ. Model. 2022, 113, 105904. [Google Scholar] [CrossRef]
  7. Guerrero, O.A.; Castañeda, G.; Trujillo, G.; Hackett, L.; Chávez-Juárez, F.J.S.-E.P.S. Subnational sustainable development: The role of vertical intergovernmental transfers in reaching multidimensional goals. Socio-Econ. Plan. Sci. 2022, 83, 101155. [Google Scholar] [CrossRef]
  8. Barro, R.J. Government spending in a simple model of endogeneous growth. J. Political Econ. 1990, 98, S103–S125. [Google Scholar] [CrossRef] [Green Version]
  9. Scotti, F.; Flori, A.; Pammolli, F. The economic impact of structural and Cohesion Funds across sectors: Immediate, medium-to-long term effects and spillovers. Econ. Model. 2022, 111, 105833. [Google Scholar] [CrossRef]
  10. Jeong, K.; Koo, C.; Hong, T. An estimation model for determining the annual energy cost budget in educational facilities using SARIMA (seasonal autoregressive integrated moving average) and ANN (artificial neural network). Energy 2014, 71, 71–79. [Google Scholar] [CrossRef]
  11. Robinson, H.; Pawar, S.; Rasheed, A.; San, O. Physics guided neural networks for modelling of non-linear dynamics. Neural Netw. 2022, 154, 333–345. [Google Scholar] [CrossRef] [PubMed]
  12. Palmer, A.; Montano, J.J.; Sesé, A. Designing an artificial neural network for forecasting tourism time series. Tour. Manag. 2006, 27, 781–790. [Google Scholar] [CrossRef]
  13. Lago, J.; De Ridder, F.; De Schutter, B. Forecasting spot electricity prices: Deep learning approaches and empirical comparison of traditional algorithms. Appl. Energy 2018, 221, 386–405. [Google Scholar] [CrossRef]
  14. Farah, S.; Humaira, N.; Aneela, Z.; Steffen, E. Short-term multi-hour ahead country-wide wind power prediction for Germany using gated recurrent unit deep learning. Renew. Sustain. Energy Rev. 2022, 167, 112700. [Google Scholar] [CrossRef]
  15. Li, Q.; Yu, C.; Yan, G. A New Multipredictor Ensemble Decision Framework Based on Deep Reinforcement Learning for Regional GDP Prediction. IEEE Access 2022, 10, 45266–45279. [Google Scholar] [CrossRef]
  16. Hyndman, R.J.; Koehler, A.B. Another look at measures of forecast accuracy. Int. J. Forecast. 2006, 22, 679–688. [Google Scholar] [CrossRef] [Green Version]
  17. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  18. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  19. Gardner, E.S., Jr. Exponential smoothing: The state of the art—Part II. Int. J. Forecast 2006, 22, 637–666. [Google Scholar] [CrossRef]
  20. Hastie, T.; Tibshirani, R.; Friedman, J.H.; Friedman, J.H. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: Berlin/Heidelberg, Germany, 2009; Volume 2. [Google Scholar]
  21. Shumway, R.H.; Stoffer, D.S.; Stoffer, D.S. Time Series Analysis and Its Applications; Springer: Berlin/Heidelberg, Germany, 2000; Volume 3. [Google Scholar]
  22. Vapnik, V. The Nature of Statistical Learning Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1999. [Google Scholar]
  23. Sharifzadeh, M.; Sikinioti-Lock, A.; Shah, N. Machine-learning methods for integrated renewable power generation: A comparative study of artificial neural networks, support vector regression, and Gaussian Process Regression. Renew. Sustain. Energy Rev. 2019, 108, 513–538. [Google Scholar] [CrossRef]
  24. Sermpinis, G.; Stasinakis, C.; Theofilatos, K.; Karathanasopoulos, A. Modeling, forecasting and trading the EUR exchange rates with hybrid rolling genetic algorithms—Support vector regression forecast combinations. Eur. J. Oper. Res. 2015, 247, 831–846. [Google Scholar] [CrossRef] [Green Version]
  25. Murillo-Escobar, J.; Sepulveda-Suescun, J.; Correa, M.; Orrego-Metaute, D. Forecasting concentrations of air pollutants using support vector regression improved with particle swarm optimization: Case study in Aburrá Valley, Colombia. Urban Clim. 2019, 29, 100473. [Google Scholar] [CrossRef]
  26. Yang, C.-H.; Shao, J.-C.; Liu, Y.-H.; Jou, P.-H.; Lin, Y.-D. Application of Fuzzy-Based Support Vector Regression to Forecast of International Airport Freight Volumes. Mathematics 2022, 10, 2399. [Google Scholar] [CrossRef]
  27. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  28. Kumar, P.; Nestsiarovich, A.; Nelson, S.J.; Kerner, B.; Perkins, D.J.; Lambert, C.G. Imputation and characterization of uncoded self-harm in major mental illness using machine learning. J. Am. Med. Inform. Assoc. 2020, 27, 136–146. [Google Scholar] [CrossRef]
  29. Memon, N.; Patel, S.B.; Patel, D.P. Comparative analysis of artificial neural network and XGBoost algorithm for PolSAR image classification. In Proceedings of the Pattern Recognition and Machine Intelligence: 8th International Conference, PReMI 2019, Tezpur, India, 17–20 December 2019; pp. 452–460. [Google Scholar]
  30. Su, Y.-C.; Wu, C.-Y.; Yang, C.-H.; Li, B.-S.; Moi, S.-H.; Lin, Y.-D. Machine learning data imputation and prediction of foraging group size in a Kleptoparasitic spider. Mathematics 2021, 9, 415. [Google Scholar] [CrossRef]
  31. LeCun, Y.; Bottou, L.; Orr, G.B.; Müller, K.-R. Efficient backprop. In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 2002; pp. 9–50. [Google Scholar]
  32. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  33. Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans. Smart Grid 2017, 10, 841–851. [Google Scholar] [CrossRef]
  34. Nelson, D.M.; Pereira, A.C.; De Oliveira, R.A. Stock market’s price movement prediction with LSTM neural networks. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1419–1426. [Google Scholar]
  35. Duan, J.; Wang, P.; Ma, W.; Tian, X.; Fang, S.; Cheng, Y.; Chang, Y.; Liu, H. Short-term wind power forecasting using the hybrid model of improved variational mode decomposition and Correntropy Long Short-term memory neural network. Energy 2021, 214, 118980. [Google Scholar] [CrossRef]
  36. Huang, X.; Li, Q.; Tai, Y.; Chen, Z.; Zhang, J.; Shi, J.; Gao, B.; Liu, W. Hybrid deep neural model for hourly solar irradiance forecasting. Renew. Energy 2021, 171, 1041–1060. [Google Scholar] [CrossRef]
  37. Schwalbert, R.A.; Amado, T.; Corassa, G.; Pott, L.P.; Prasad, P.V.; Ciampitti, I.A. Satellite-based soybean yield forecast: Integrating machine learning and weather data for improving crop yield prediction in southern Brazil. Agric. For. Meteorol. 2020, 284, 107886. [Google Scholar] [CrossRef]
  38. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  39. Jia, P.; Liu, H.; Wang, S.; Wang, P. Research on a mine gas concentration forecasting model based on a GRU network. IEEE Access 2020, 8, 38023–38031. [Google Scholar] [CrossRef]
  40. Sankaranarayanan, S.; Balan, J.; Walsh, J.R.; Wu, Y.; Minnich, S.; Piazza, A.; Osborne, C.; Oliver, G.R.; Lesko, J.; Bates, K.L. COVID-19 mortality prediction from deep learning in a large multistate electronic health record and laboratory information system data set: Algorithm development and validation. J. Med. Internet Res. 2021, 23, e30157. [Google Scholar] [CrossRef]
  41. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  42. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Dey, R.; Salem, F.M. Gate-variants of gated recurrent unit (GRU) neural networks. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; pp. 1597–1600. [Google Scholar]
  44. Gutierrez-Torre, A.; Bahadori, K.; Baig, S.-U.; Iqbal, W.; Vardanega, T.; Berral, J.L.; Carrera, D. Automatic distributed deep learning using resource-constrained edge devices. IEEE Internet Things J. 2021, 9, 15018–15029. [Google Scholar] [CrossRef]
  45. Tsay, R.S. Analysis of Financial Time Series; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  46. Brockwell, P.J.; Davis, R.A. Time Series: Theory and Methods; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  47. Montgomery, D.C.; Jennings, C.L.; Kulahci, M. Introduction to Time Series Analysis and Forecasting; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  48. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  49. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; Volume 30. [Google Scholar]
Figure 1. Structure of the GRU neural network. (A) LSTM cell. (B) GRU cell. (C) Framework of proposed government expenditure forecasting approach.
Figure 1. Structure of the GRU neural network. (A) LSTM cell. (B) GRU cell. (C) Framework of proposed government expenditure forecasting approach.
Mathematics 11 03085 g001
Figure 2. Framework of GDP indicators’ forecasting.
Figure 2. Framework of GDP indicators’ forecasting.
Mathematics 11 03085 g002
Figure 3. GRU model for forecasting.
Figure 3. GRU model for forecasting.
Mathematics 11 03085 g003
Figure 4. Performance evaluation based on the average government expenses forecasting metrics for the 15 countries.
Figure 4. Performance evaluation based on the average government expenses forecasting metrics for the 15 countries.
Mathematics 11 03085 g004
Figure 5. Comparison of seven forecasting algorithms for government expenditure.
Figure 5. Comparison of seven forecasting algorithms for government expenditure.
Mathematics 11 03085 g005
Table 1. Descriptive statistics of government expenditure.
Table 1. Descriptive statistics of government expenditure.
CountryCountMinMaxMeanMedSDQ1Q3IQR
Australia31.0−3.6106.9862.8293.0911.8491.7803.7141.934
Brazil31.00.2535.0342.6603.0421.3981.7443.7111.968
Canada31.00.1419.1712.7852.2611.9921.5313.7362.205
China31.00.9666.1873.3113.4871.3882.3764.4182.043
France31.00.2033.8751.8971.5690.9601.2772.5491.271
Germany31.0−0.72512.7322.0181.8432.3500.5632.6082.045
India31.00.0273.6211.2541.0560.8630.6121.8621.251
Italy31.0−1.1672.9810.8540.7240.8780.2841.3121.028
Japan31.0−0.0521.2210.2380.1290.2880.0460.3500.304
Korea31.00.2122.1560.8550.7800.4950.4961.0330.537
Mexico31.00.8773.9882.4912.5640.7512.1812.8920.710
Russia31.00.1754.5031.6491.2011.2430.5832.5771.994
Spain31.00.6406.7702.7892.4051.3431.8693.4421.572
UK31.0−0.86411.9293.8122.2803.2381.7355.8374.102
USA31.00.4653.4061.5931.4730.7621.0342.0581.024
Min: minimum value; Max: maximum value; Mean: average value; Med: median.
Table 2. Comparison of the number of parameters to optimize for each neural network model.
Table 2. Comparison of the number of parameters to optimize for each neural network model.
OptimizerMethodsParametersMAPE
AdamGRUEpochsLayer1Layer2Batch-sizeLearning rate
10002002002560.0042.774
LSTMEpochsLayer1Layer2Batch-sizeLearning rate
10002002002560.0044.208
CNNEpochsCONV1DCNN_DENSEBatch-sizeLearning rate
1000200502560.0043.389
RMSpropGRUEpochsLayer1Layer2Batch-sizeLearning rate
1004004002560.0042.809
LSTMEpochsLayer1Layer2Batch-sizeLearning rate
1004004002560.0044.209
CNNEpochsCONV1DCNN_DENSEBatch-sizeLearning rate
1004001002560.0044.033
AdagradGRUEpochsLayer1Layer2Batch-sizeLearning rate
321001002560.0042.852
LSTMEpochsLayer1Layer2Batch-sizeLearning rate
321001002560.0044.207
CNNEpochsCONV1DCNN_DENSEBatch-sizeLearning rate
32100102560.00486.798
Table 3. Sensitivity analysis of GRU model.
Table 3. Sensitivity analysis of GRU model.
L250, ep100L150, ep100L250, ep32L150, ep32L200, p1000
MAE0.5180.5180.5100.5210.528
RMSE0.7410.7380.7200.7430.747
MAPE (%)2.7942.7842.7752.8042.774
L: number of units in the layer, ep: Epochs.
Table 4. The experimental results of the traffic volume forecast model: MAPE, MAE, and MAPE values.
Table 4. The experimental results of the traffic volume forecast model: MAPE, MAE, and MAPE values.
CountryMetricARIMAETSSVRXGBCNNLSTMBiRNNASRNNGRU
MAE24.8540.5131.2670.7470.7881.3451.3410.9770.618
AustraliaRMSE27.6720.7821.4920.9701.0821.5481.5440.9590.852
MAPE(%)135.7322.5856.4143.7943.9766.8185.6332.9673.089
MAE4.6370.1861.1360.7500.51340.9770.9750.9820.322
BrazilRMSE4.9540.2861.1640.8210.64631.0081.0070.8840.379
MAPE(%)24.1771.1965.6283.7272.5384.8373.5512.7811.594
MAE30.8410.5430.5850.5850.5870.8670.8651.0530.526
CanadaRMSE35.1020.7730.9260.8360.9201.1081.1061.1550.825
MAPE(%)150.0881.8442.6652.6842.6794.0073.0982.9982.402
MAE1.3320.9730.7860.7210.3100.6690.6700.6280.166
ChinaRMSE1.5920.2090.8500.7720.3950.6990.7000.3280.197
MAPE(%)9.1001.1284.7414.3631.8744.0433.6682.8851.005
MAE35.2990.7120.5780.5690.5700.7690.7671.0070.586
FranceRMSE39.6870.7980.7280.9930.7751.0101.0080.8640.796
MAPE(%)150.9142.3332.4032.3192.3633.1672.5533.2252.422
MAE36.7110.5750.7880.8540.6250.7530.7521.2350.591
GermanyRMSE41.0990.9901.2621.1501.0891.2061.2051.2550.790
MAPE(%)188.3782.6663.6754.0492.8943.5093.0292.8842.810
MAE3.7520.5430.5700.4790.6360.6940.6941.4390.419
IndiaRMSE3.9020.7350.8410.5210.9091.0071.0061.1600.617
MAPE(%)34.8504.7464.8614.2865.4685.9224.7713.2743.584
MAE37.7420.4600.6920.6620.5300.5880.5870.7760.598
ItalyRMSE42.5890.8520.8170.9590.8490.9570.9570.9230.936
MAPE(%)196.6912.2603.5273.3182.6412.9161.0722.8842.979
MAE18.0490.3351.1330.4010.3980.4510.4510.6170.366
JapanRMSE19.0500.5881.5560.5810.5880.7190.7180.6260.532
MAPE(%)100.3731.6275.5771.9641.9522.1950.2253.1641.796
MAE13.0220.9731.8920.5201.0291.0741.0700.8840.540
KoreaRMSE13.2521.0862.3810.6461.3131.5261.5202.8610.644
MAPE(%)99.8235.97211.1823.0906.2246.2456.4002.7763.235
MAE1.7740.4360.4410.4320.4770.5550.5542.0450.441
MexicoRMSE2.1820.6170.7620.6310.5890.7460.7451.0550.616
MAPE(%)18.73253.5133.5193.5063.9534.4783.5272.8183.596
MAE40.0230.6610.6440.6520.7240.8280.8280.7910.832
RussiaRMSE45.1011.0791.0600.9801.1651.3061.3051.1631.005
MAPE(%)222.6614.4203.2773.3373.6844.2043.0573.1163.362
MAE52.0740.7120.7650.7310.7980.8020.8001.4870.675
SpainRMSE58.2331.2581.4371.0391.2531.3841.3831.4831.242
MAPE(%)279.3893.3993.6373.6043.8773.8312.9821.9753.214
MAE60.0550.8500.9280.9160.9631.0471.0460.0551.021
UKRMSE67.5781.4991.5061.5151.4951.7151.7151.7811.456
MAPE(%)311.6374.0104.4504.3624.6514.9732.4782.5784.957
MAE12.7140.2390.7370.2080.29500.2860.2860.0130.227
USARMSE14.7650.3520.7830.2430.3800.4010.4020.3430.318
MAPE(%)85.8471.6605.2431.4592.0681.9791.3721.8751.570
MAE24.8580.5810.8630.6150.6160.7800.7790.9330.528
Total AverageRMSE27.7840.7931.1710.8440.8971.0891.0881.1230.747
MAPE(%)133.8932.8914.7203.3243.3894.2083.1612.8132.774
Bold means the lowest value.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, C.-H.; Molefyane, T.; Lin, Y.-D. The Forecasting of a Leading Country’s Government Expenditure Using a Recurrent Neural Network with a Gated Recurrent Unit. Mathematics 2023, 11, 3085. https://doi.org/10.3390/math11143085

AMA Style

Yang C-H, Molefyane T, Lin Y-D. The Forecasting of a Leading Country’s Government Expenditure Using a Recurrent Neural Network with a Gated Recurrent Unit. Mathematics. 2023; 11(14):3085. https://doi.org/10.3390/math11143085

Chicago/Turabian Style

Yang, Cheng-Hong, Tshimologo Molefyane, and Yu-Da Lin. 2023. "The Forecasting of a Leading Country’s Government Expenditure Using a Recurrent Neural Network with a Gated Recurrent Unit" Mathematics 11, no. 14: 3085. https://doi.org/10.3390/math11143085

APA Style

Yang, C. -H., Molefyane, T., & Lin, Y. -D. (2023). The Forecasting of a Leading Country’s Government Expenditure Using a Recurrent Neural Network with a Gated Recurrent Unit. Mathematics, 11(14), 3085. https://doi.org/10.3390/math11143085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop