Next Article in Journal
Visualization of WiFi Signals Using Programmable Transfer Functions
Previous Article in Journal
Augmented Reality, Virtual Reality, and Game Technologies in Ophthalmology Training
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accurate Air-Quality Prediction Using Genetic-Optimized Gated-Recurrent-Unit Architecture

1
School of Computer Science and Technology, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
2
Shaanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi’an 710121, China
3
Xi’an Key Laboratory of Big Data and Intelligent Computing, Xi’an 710121, China
4
Shaanxi Key Laboratory of Speech & Image Information Processing (SAIIP), School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710129, China
5
National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, Xi’an 710129, China
*
Author to whom correspondence should be addressed.
Information 2022, 13(5), 223; https://doi.org/10.3390/info13050223
Submission received: 20 March 2022 / Revised: 15 April 2022 / Accepted: 17 April 2022 / Published: 26 April 2022
(This article belongs to the Topic Big Data and Artificial Intelligence)

Abstract

:
Air pollution is becoming a serious concern with the development of society and urban expansion, and predicting air quality is the most pressing problem for human beings. Recently, more and more machine-learning-based methods are being used to solve the air-quality-prediction problem, and gated recurrent units (GRUs) are a representative method because of their advantage for processing time-series data. However, in the same air-quality-prediction task, different researchers have always designed different structures of the GRU due to their different experiences. Data-adaptively designing a GRU structure has thus become a problem. In this paper, we propose an adaptive GRU to address this problem, and the adaptive GRU structures are determined by the dataset, which mainly contributes with three steps. Firstly, an encoding method for the GRU structure is proposed for representing the network structure in a fixed-length binary string; secondly, we define the reciprocal of the sum of the loss of each individual as the fitness function for the iteration computation; thirdly, the genetic algorithm is used for computing the data-adaptive GRU network structure, which can enhance the air-quality-prediction result. The experiment results from three real datasets in Xi’an show that the proposed method achieves better effectiveness in RMSE and SAMPE than the existing LSTM-, SVM-, and RNN-based methods.

1. Introduction

Air pollution has become one of the most crucial environmental issues in the world [1], contributing to a set of serious outcomes for human beings. The greenhouse gases emitted by anthropogenic activities consistently affect the global ecosystem [2], and the amount of air pollutants is increasing obviously, which indirectly affects the health of people by permeating into agricultural food products [3,4] and even influences the living comfort of citizens [5]. The degree of air pollution depends on the different types of air pollutants, such as sulfur dioxide (SO2), carbon monoxide (CO), nitrogen oxide (NOx), and particulate matter (PM2.5, PM10), the concentrations of which in air are closely related to human health [6,7,8]. Air-pollutant concentration is an index to quantitatively measure the content of certain pollutants in the air. When the pollutant value is lower, the air-pollution degree is lower and the impact on human health is smaller, and vice versa. In recent years, with the increase in types and accuracy of sensors, environmental monitoring technology is constantly being upgraded and more and more air-quality-prediction methods have emerged.
Traditional air-quality-prediction methods are mainly based on numerical prediction and generally simulate the movement of air pollutants in the air through a combination of mathematical models and physical knowledge, predict the actual diffusion degree of air pollutants in the atmosphere, and then obtain the actual concentration of air pollutants through simulation calculation [9,10,11,12,13]. For example, the Danish Eulerian model is a powerful air-pollution model which calculates sensitivity indices by taking physical and chemical processes into consideration [14]. However, the prediction process requires complex formula derivation and programming to reproduce real physical processes, which is time-consuming and laborious, and the prediction effect is mediocre. In addition to numerical prediction-based methods, some statistics-based methods are also applied for air-quality prediction. Decision tree, a classic algorithm based on statistics, uses the nonparametric supervised learning method to realize the decision and classification of air quality by optimizing the Gini coefficient [15,16]. On this basis, the decision-tree algorithm was taken as the basic unit of the random forest algorithm, and the sampling method with replacement was used to generate the training set; the prediction results were obtained after repeated iterations [17]. Some scholars also proposed to use support vector machine (SVM) [18] to solve the problem of air-quality prediction and obtain prediction results by dividing the hyper plane of the feature space of the data. Gao et al. proposed the MFO-SVM method [19], further optimized the SVM algorithm by solving the linear equations, and achieved good prediction results. At the same time, a K-means algorithm [20] was also applied to the problem of air-quality prediction and has achieved acceptable results.
In recent years, with the active development of machine-learning theory, more and more artificial-intelligence technologies are applied for air-quality prediction [21,22,23,24], especially methods based on deep learning. For example, an artificial neural network (ANN) was proposed in the 1980s that can simulate the activity of human neurons to achieve an effect similar to human numerical calculation. This method has also been well-applied in the field of air-quality prediction [25]. The appearance of the BP neural network [26] improved the computing power of artificial neural networks, and after optimization using the KNN algorithm, it was also applied for the prediction of air quality [27]. However, due to the stochasticity of initialization of the back-propagation neural network (BP neural network), the network weight could converge slowly. The genetic algorithm [28] has also been used to optimize neural networks for obtaining a better network initial value and threshold value to improve the training speed. Under the condition of better initialization of the network, its weight will still float within a certain range after training and convergence. Li et al. [29] proposed to use the discrete HopField network structure for air-quality prediction, which can fix the weight after the network training and stability so as to improve the storage capacity of the network and obtain a better result. In order to improve the prediction accuracy of the neural network, scholars selected the optimal particle size in a dynamic wavelet neural network [30] and tried to change the solution space of air-quality prediction, which achieved certain prediction results.
Although deep-learning-based methods have achieved acceptable results in the air-quality-prediction task, the characteristic of the air-quality-prediction task needs to combine the previous data to predict the future data, i.e., air-quality data generally contain a time series, which is hardly captured by traditional networks such as BP, CNN, etc. Equally, most of the existing air-quality-prediction methods based on deep learning do not have strong time-series memory, which affects the prediction accuracy. Thus, the support of a strong model with time-series memory is crucial for air-quality prediction.
Mitigating the above problems, the recurrent neural network (RNN)-based air-quality-prediction method [31] improved the shortcoming of slow convergence of traditional machine-learning algorithms and the prediction accuracy by combining the time-series correlation of the data itself. However, it is inefficient for an RNN to deal with long-term memory due to the gradient vanishing. Hence, an air-quality-prediction method based on a long short-term memory network (LSTM) [32,33,34] improved the efficiency of extracting continuous time features and the prediction effect of the RNN. Some scholars proposed a mixed model of RNN-LSTM [35] to further improve the accuracy of air-quality prediction. These methods have made a lot of changes to the classical convolution neural network and fully connected network.
Compared with RNN and LSTM, the GRU has demonstrated its better performance in many real tasks. Therefore, we use a gated recurrent unit (GRU) [36] as our air-quality-prediction model, which can effectively leverage time-series data of air quality and has a longer memory of time-series data. Furthermore, we note that its structures are manually designed, which limits the flexibility of the model and influences the prediction accuracy.
This paper presents a data-adaptively designed GRU structure for air-quality prediction, which mainly contributes in three aspects. Firstly, a new encoding method is utilized to encode the GRU network structure, which adjusts its characteristics. Secondly, a new fitness function is used for the genetic algorithm process, which uses the reciprocal of the loss function as the fitness function and solves the evaluation problem of the candidate structure in the genetic algorithm. Finally, the genetic algorithm is utilized to data-adaptively design the GRU network structure for air-quality prediction.
Using the above algorithm, a GRU with a data-adaptively designed structure will be obtained to forecast the air quality of Xi’an city, and only the values of PM2.5, PM10, NO2, SO2, O3, and CO need to be observed or predicted; then, the air-quality index can be predicted by the model, thus allowing relevant departments to carry out environmental assessment and protection. Of course, the method can also be used to establish models to predict the air quality of other regions and even the whole country.
The paper is organized as follows: Section 2 introduces the details of related works. Section 3 illustrates the proposed method. The experimental results and analysis are presented in Section 4, and in Section 5 the conclusions of this paper are presented.

2. Related Works

2.1. Gated Recurrent Unit

The gated-recurrent-unit network model is a neural network model that combines the unit state and hidden layer state of the long short-term memory (LSTM) [37]. The network model could improve on the shortcomings of LSTM, i.e., long training time, high number of parameters, and complex internal calculation. The GRU combines the forget gate and input gate into a single update gate and has a reset gate. By combining the cell state and hidden state, the GRU is a new method of calculating new information at the current moment based on LSTM that is different from LSTM but maintains the effect of the LSTM model. However, it has a simpler structure, fewer parameters, and a better convergence model. The basic unit structure of the gated recurrent neural network is shown in Figure 1.
In the figure, "×" and "+" represent matrix multiplication and matrix addition, respectively; σ and Tanh are the Sigmoid activation function and Tanh activation function, respectively; r t is the reset gate; z t is the update gate; H ˜ is the candidate hidden state; H t 1 and H t are the hidden state; and X t denotes the input.
As seen in Figure 1, the GRU has only two gates, namely the update gate r t and the reset gate z t . The update gate r t is used to control the degree to which the state information of the previous moment is brought into the current state. When the update gate r t is larger, more state information of the previous moment is brought into the current state. The reset gate z t is used to control the degree to which the state information of the previous moment is ignored. The smaller the value of the reset gate z t is, the more the state information is ignored. Through the mechanism of these two gates, the GRU can adjust the flow of information to reduce short-term memory problems. Therefore, this paper proposes an air-quality-prediction method based on the GRU to achieve long-term and continuous data prediction.
The activation functions Sigmoid and Tanh are used in the GRU to process the input values. The Sigmoid function is used to convert the input value to 0~1, as shown in Formula (1).
σ ( x ) = 1 1 + e x
The Tanh function is similar to the Sigmoid function, converting the input value to between −1 and 1 and retaining a nonlinear monotonic relationship between the input and output, as shown in Formula (2).
T a n h ( x ) = e x e x e x + e x
In this paper, the number of hidden units in the GRU network is h, and the number of hidden layers is denoted by L. At a given time step t, the input is X t , whose batch size is n, and the sample number of each batch is d.
The steps of GRU forward propagation are as follows.
First of all, the hidden state of the last time step is H t 1 R n × h . The calculations of the reset gate r t R n × h and update gate z t R n × h are shown in Formulas (3) and (4), respectively.
r t = σ ( X t W x r + H t 1 W h r + b r )
z t = σ ( X t W x z + H t 1 W h z + b z )
where W h r , W h z R h × h are the weight matrix; b r , b z R 1 × h are the bias matrix; and W x r , W x z , b r , and b z are the parameters that must be updated.
In addition, the candidate hidden state is computed by the reset gate H t ˜ R n × h , where Θ represents element-wise multiplication, as shown in Formula (5):
H t ˜ = t a n h ( X t W x h + ( r t Θ H t 1 ) W h h + b h )
where W x h R d × h and W h h R h × h are the weight matrices, and b h R 1 × h is a bias matrix.
Finally, the hidden state is computed by the result of the reset gate, update gate, and candidate hidden state H t R n × h , as shown in Formula (6):
H t = z t Θ H t 1 + ( 1 z t ) Θ H t ˜

2.2. Genetic Algorithm

The genetic algorithm is an optimized algorithm based on the mechanism of natural selection and population inheritance. It simulates reproduction, hybridization, and mutation in the process of natural selection and inheritance and uses these bioinspired operators to generate effective solutions to optimization and search problems [38,39,40,41]. When using the genetic algorithm to solve a problem, individuals constitute every possible solution of the problem which could be encoded as a "chromosome", and the population is the solution domain, which is composed of all possible individuals. Evidently, a typical genetic algorithm generally needs to consider two prerequisites, namely the genetic representation of the solution domain and the design of the fitness function to evaluate the competitiveness of each candidate—e.g., the traveling salesman problem [42] aiming to find the optimal Hamiltonian path of the N-node graph, whose fitness function is the total cost of the path, and each feasible solution is represented as {1,2,…,N}.
The genetic algorithm starts by randomly producing individuals; the fitness value of each individual is given by being evaluated according to a predetermined fitness function, and some individuals are selected to produce the next generation based on this fitness value. Selection allows us to keep the strong ones and eliminate the weak ones. The selected individuals then produce a new generation through crossover and mutation operators, and the method of mutation and crossover varies from case to case, usually based on the properties of the particular problem. The individuals in the new generation inherit some of the good traits of the previous generation, and their performance is therefore better than that of the previous generation, thus gradually evolving towards the optimal solution. Therefore, some previous work applied the genetic algorithm to explore efficient neural network architecture [43,44,45,46,47,48].

3. The Proposed Method

This section illustrates the data preprocessing before the air-quality-prediction based on an adaptive GRU using the genetic algorithm and the GRU network structure. We also present a genetic algorithm customized for obtaining a more competitive GRU network structure. Firstly, we depict how to encode the network structure into a fixed-length binary string. Secondly, we define several genetic operations, i.e., selection, mutation, and crossover, through which we can search the adaptive GRU structure. Finally, the training and evaluation method is discussed.

3.1. Data Preprocessing

To demonstrate the data preprocessing, we have taken real-time report data of the air-quality index from 2018 to 2020 recorded at the Central Square station of Xincheng District, Xi’an as an example, which include 25,569 data points. Hourly average concentrations of fine particulate matter (PM2.5), inhalable particulate matter (PM10), sulfur dioxide (SO2), nitrogen dioxide (NO2), ozone (O3), and carbon monoxide (CO) selected from 2018 to 2020 are used as an original dataset, which is denoted by D . Several parts of dataset D are shown in Table 1.
The original dataset D is divided into a training set D t r a i n , a validation set D v a l , and a testing set D t e s t , which are based on a certain proportion mentioned in the following section. Then, the normalization operation is applied to the three sets to balance the influence of different types of air pollutants on the fitness. Taking the training set D t r a i n as an example, the method of normalization is shown in Formula (7).
x t r a i n i = x t r a i n i μ t r a i n σ t r a i n
Here, x t r a i n i represents the data in D t r a i n ; i { 1 , 2 , , i , , n } ; n is the number of training samples; μ t r a i n is the mean of data in D t r a i n ; σ t r a i n is the standard deviation of data in D t r a i n ; and x t r a i n i represents the data of D t r a i n which is normalized from D t r a i n .
The dataset D t r a i n r is shown in Table 2.
Then, each continuous 25 h of the normalized dataset D t r a i n is represented as a sample. The previous 24 h (i.e., the data in order 0–23 h) of the sample will serve as the input, and the last hour (i.e., the data in order 24 h) denotes the label of the input data; this operation will work in the training set, validation set, and testing set. The sample is shown in Table 3.

3.2. The Network Structure of GRU

In this paper, we use the genetic algorithm to search the solution domain of the GRU network structures, and the best solution in this paper includes a feature in the hidden state and b hidden layers; its network structure is shown in Figure 2. x X B × 144 is the input of the network with B batch size, where 144 is the number of data points of 6 air pollutants in 24 h, and the nth element of x is denoted as x n . G l m represents the mth structural unit of the lth layer of the network. h l m X B × 6 denotes the output of G l m , while h l 0 is the initial hidden state of the lth layer. The fully connected network is denoted as FC. A schematic diagram of the proposed GRU neural networks is shown in Figure 2.

3.3. Binary-Network Representation

We provide a binary-string representation for a network structure. We firstly note that the number of layers and the number of features in hidden state is variable, which mainly affect the effectiveness of the GRU, while the size of input and output data is unchanged after being defined.
In this paper, we use a binary string of length 20 as an example. Figure 3 shows the binary string with random numbers that may occur in our experiment.
Here, the i-th number of the string L is s i { 0 , 1 } , i = 1,2,…,20. The first 10 numbers of the string L represent the code of the number of layers L 1 , which is denoted as L 1 ; the remaining numbers represent the number of features in the hidden state L 2 , which is denoted as L 2 .
L 1 decodes to the solution L 1 as shown in Formula (7).
L 1 = 1 8 i = 1 10 ( s i × 2 10 i ) ,  
Similarly, L 2 decodes to the solution L 2 as shown in Formula (8).
L 2 = 1 8 i = 11 20 ( s i × 2 20 i ) ,
Here, we take L = 0 110110111 | 0000010100 as an example, where L 1 = 0110110111 and L 2 = 0000010100 . The final numbers of layers L 1 and features L 2 are 55 and 2, respectively.

3.4. Genetic Operation

The genetic algorithm starts with the initialization of N random individuals. We perform T generations of the whole genetic process—i.e., we repeat the operations of selection, crossover, and mutation T times. Then, the fitness of each individual is obtained by training the reference dataset. The detailed genetic algorithm is shown in the following algorithm steps.

3.4.1. Initialization

First, we randomly initialize a group of models { M 0 n } n = 1 N . The number of layers and the number of features in the hidden state of each model of the group are represented by a binary string of length 20. Each bit, b 0 , n l , l = 1 , 2 , , 20 , of the binary string independently follows a Bernoulli distribution: b 0 , n l B ( 1 , 0.5 ) . Then, we obtain the fitness of each initial model by the fitness function. The fitness function in this paper is shown in Formula (11).
1 k = 1 6 L o s s k + 1 e 10
Here, L o s s k is the loss generated after evaluating the model which has pretrained on the reference dataset for the kth air pollutant.

3.4.2. Selection

We then perform selection at the beginning of each generation. At the beginning of the t-th generation, the fitness r t 1 n of individual M t 1 n at the (t − 1)th generation or initial generation is given by the fitness function. Here, r t 1 n affects the probability of M t 1 n being selected in the selection process.
Rank selection is used to determine which individuals survive the selection process. Firstly, at the beginning of generation T, the population is sorted according to fitness values. Each chromosome is then assigned selection probabilities based on its rank [43]. Individuals are selected according to their selection probability, and each individual from the previous generation can be selected multiple times in order to keep the number of individuals constant.
In rank selection, the sum of ranks is computed and the probability of each individual is computed, as shown in Formulas (10) and (11), respectively:
R s u m = i = 1 N R i j ,  
P r a n k = R i j R s u m ,  
where R i j denotes the jth individual of the ith layer.

3.4.3. Mutation

We give each bit of code of the M l n a probability p M to change, and the individuals with the probability q M are selected to perform this process. In general, p M is very small. For example, p M = 0.1 is used in our experiment. Although the mutation may not have a great impact on the individual, its purpose is to provide new possibilities while preserving the excellent genotype of the surviving individual.

3.4.4. Crossover

Crossovers involve genotypic changes in both individuals, and q c denotes the number of individuals that will be selected to perform this process (Algorithm 1). The basic operated object is a stage of the genotype, rather than a single gene, with the aim of preserving the local structure of a good genotype. Similar to mutations, each pair of corresponding stages is exchanged with a small probability p c . In this paper, we have adopted the single-point crossover method. First, two crossover points are randomly set in the two individual coding strings that correspond with each other. Then, the two individuals swap parts of their chromosomes between the two designated intersections.
Algorithm 1 The Genetic Process for Network
1: Input: the reference dataset D, the number of generations T, the number of individuals in each generation N, the mutation probability p M , the crossover probabilities p c , the mutation parameter q M , and the crossover parameter q c .
2: Initialization: randomly generating a group of models { M 0 n } n = 1 N and computing their fitness;
3: for t = 1, 2, 3,…, T do
4:      Selection: generating a new generation { M t n } n = 1 N using rank selection;
5:      Crossover: performing crossover with probability q M and parameter q c ;
6:      Mutation: performing mutation on each individual { M t n } n = 1 N with probability p M and parameter q M ;
7:      Evaluation: computing the fitness for each individual { M t n } n = 1 N ;
8: end for
9: Output: the final generation { M T n } n = 1 N .

3.5. Training and Evaluation

For each dataset, the genetic algorithm is implemented in the training set corresponding with the dataset. In the genetic process, we set the number of generations as 100 and the number of individuals as 20; the fitness of each individual in the validation dataset is achieved after training 200 epochs, and this fitness is the evaluation index fitness. Finally, we obtain an optimal GRU network structure i.e., the GRU structure with the best fitness in all individuals at the final genetic generation.
Next, we use the adaptive network structure to train the corresponding dataset in this experiment to obtain the final optimal predicted result. The network after 1000 training iterations is obtained via the training set, and the predicted results are obtained by inputting the testing data. The predicted results will be evaluated by the evaluation index.

4. Experiments and Analysis

In order to verify the effectiveness of air-quality prediction based on adaptive GRU using genetic algorithm, the actual data of three observation stations in Xi’an city were used for the experiments. The following is an introduction to the dataset used for the experiments, the experimental environment, the experimental evaluation index, the experimental results, and analysis.

4.1. Datasets

This paper adopted the real-time air-quality information collected by the following three air-quality observatories in Xi’an city, China, as the experimental datasets: Dataset 1, Xi’an Xincheng Center Square Station from 2018 to 2020; Dataset 2, Xi’an Caotang Base; Dataset 3, Xi’an Gaoxin West Station.
The air-quality dataset of Xi‘an Xincheng Center Square Station from 2018 to 2020 includes 25,569 data points, that of Xi’an Gaoxin West Station includes 14,567 data points, and that of Xi’an Caotang base includes 14,568 data points. The quantity of data in the three datasets is shown in Table 4.
In this paper, the data from every 25 consecutive hours are considered as a sample. The data of the previous 24 h are considered as the input data, and the data of the last 1 h are considered as label data. The stride is designed as 1, and these processed datasets are made into reconstructed datasets. The sample quantity of the reconstructed datasets is shown in Table 5.
In addition, the datasets were divided into training sets and testing sets at a rate of 7:3, and the last 10% of the training set was then taken as a validation set—i.e., the first 63% of each reconstructed dataset was taken as the training set, the following 7% as the validation set, and the last 30% as the testing set.
The sample numbers in the training set, validation set, and testing set obtained from each reconstructed dataset are shown in Table 6.

4.2. The Experimental Environment

The code was run on a computer with an Intel(R) Core (TM) i9-10900K CPU @3.70 GHz, NVIDIA GeForce RTX 2080, 128 GB RAM, 1 T SSD, Python3.6, and PyTorch 1.4.1.
The number of genetic-process generations was 100. The number of iterations was set to 1000. The learning rate was 0.02. The batch size was 512. Two fully connected layers were included: the first layer had 256 hidden nodes, and the second layer had 512 hidden nodes.

4.3. The Experimental Evaluation Index

In order to evaluate the effectiveness of the air-quality-prediction model proposed in this paper, two evaluation indexes were adopted, namely root-mean-square error (RMSE) and symmetric mean absolute percentage error (SMAPE), which were used to analyze the error between the prediction results and the real data. In general, the smaller the RMSE and SMAPE values are, the smaller the deviation between the predicted results and the true values is and the better the model is. The calculation formulas of RMSE and SMAPE are shown in Equations (13) and (14), respectively.
R M S E ( y , y ^ ) = 1 m i = 1 m ( y y ^ ) 2  
Here, y is the prediction of the model, y ^ represents the real data taken from the real-time air-quality measurements, and m is the sum of the number of data points used in calculation.
S M A P E = 100 % n i = 1 n 2 | f f ^ | ( | f | + | f ^ | )  
Here, f is the prediction of the model, f ^ represents the real data taken from the real-time air-quality measurements, and n is the sum of the number of data points used in calculation.

4.4. The Adaptive GRU Structure Using Genetic Algorithm

In order to prove the efficiency of the genetic algorithm used in this paper in searching the adaptive GRU network structure, we applied the genetic algorithm process to the GRU network structure in three datasets.
The generations and corresponding best fitness in Dataset 1 are shown in Figure 4.
The best fitness and corresponding network structure at each genetic generation are shown in Table 7.
The generations and corresponding best fitness in Dataset 2 are shown in Figure 5, and the best fitness and corresponding network structure at each genetic generation are shown in Table 8.
The generations and corresponding best fitness in Dataset 3 are shown in Figure 6.
The best fitness and corresponding network structure at each genetic generation are shown in Table 9.
As shown in Figure 4, Figure 5 and Figure 6, we observed, in detail, a change in the best fitness and corresponding generations.
Table 7, Table 8 and Table 9 demonstrate that the best fitness always emerges in the final generation, so the best network structure is also in the final generation. Finally, the optimal number of hidden layers of the GRU and the number of features of the GRU were 54 and 6, respectively, with Dataset 1; 55 and 2, respectively, with Dataset 2; and 61 and 2, respectively, with Dataset 3.

4.5. The Adaptive GRU Structure Compared with the Manually Designed GRU Structure

To prove the effectiveness of the adaptive GRU structure, we compared it with two manually designed structures with the three datasets. These two GRU network structures were manually designed as follows:
GRU1: 3 features in hidden state and 10 hidden layers;
GRU2: 256 features in hidden state and 256 hidden layers;
GRU_GA: the adaptive GRU structure.
The results for Dataset 1 are shown in Table 10.
As shown in Table 10, the adaptive GRU network structure obtained the best prediction results for five air pollutants. For O3, the GRU_GA structure achieved the second-best prediction results, which are almost equal to the best one.
The results for Dataset 2 are shown in Table 11.
As shown in Table 11, the adaptive GRU network structure achieved the best prediction results compared with the other two manually designed GRU network structures with Dataset 2, except for CO.
The results for Dataset 3 are shown in Table 12.
Table 12 shows that the adaptive GRU network structure performed better than the other two manually designed GRU network structures for four air pollutants in Dataset 3, and GRU1 had the best prediction results for CO and O3.
It should be noted that the adaptive GRU network structure derived from the proposed method with the three datasets could perform better than the manually designed GRU network structures for most air pollutants.

4.6. The Adaptive GRU Compared with Other Air-Quality-Prediction Methods

In order to prove the effectiveness of the proposed method in air-quality prediction, it was compared with SVM, RNN, and LSTM methods for air-quality prediction, and the prediction capability of the proposed method was verified in three datasets. The prediction results of the air-quality-prediction methods based on SVM, RNN, LSTM, and GRU_GA for each pollutant in Dataset 1 are shown in Table 13.
As shown in Table 13, the RMSE and SMAPE values obtained by the method proposed in this paper for the air-quality prediction of six pollutants were the best in Dataset 1, with RMSE values for PM10, PM2.5, SO2, NO2, CO, and O3 of 0.0024, 0.0078, 0.0035, 0.0068, 0.0001, and 0.0008, respectively, which are the lowest of the inferred methods. The SMAPE values with GRU_GA for PM10, PM2.5, SO2, NO2, CO, and O3 are 0.0855, 0.0678, 0.0143, 0.0959, 0.0007, and 0.0761, respectively, which are also lower than those obtained with other methods.
A comparison between the predicted values and true values of different pollutants with the adaptive GRU network in Dataset 1 is shown in Figure 7.
The air-quality-prediction results of SVM, RNN, LSTM, and GRU_GA for Dataset 2 are shown in Table 14.
As can be seen from Table 14, GRU_GA achieved the best RMSE values for five air pollutants and the lowest SMAPE values for all pollutants in Dataset 2. The RMSE values from GRU_GA for PM10, PM2.5, SO2, NO2, and O3 were 0.0339, 0.0057, 0.0023, 0.0075, and 0.0115, respectively, which are lower than those obtained with other methods. However, the RMSE value obtained with SVM for CO was the best out of the models, at 0.0003. The SMAPE values obtained with the proposed method for the six pollutants were the best, with values for PM10, PM2.5, SO2, NO2, CO, and O3 of 0.0854, 0.0701, 0.0160, 0.0971, 0.0015, and 0.0895, respectively.
A comparison between the predicted values and true values of different pollutants from the adaptive GRU network in Dataset 2 are shown in Figure 8.
The RMSE and SMAPE values of the air-quality-prediction methods based on SVM, RNN, LSTM, and GRU for each pollutant in Dataset 3 are shown in Table 15.
As demonstrated in Table 15, the RMSE values obtained by the GRU_GA in the air-pollutant predictions for PM10, PM2.5, SO2, NO2, and O3 were the best at 0.0208, 0.0078, 0.0028, 0.0084, and 0.0100, respectively, but compared with GRU_GA, the SVM and RNN performed better for CO with RMSE values of 0.0003. The SMAPE values with GRU_GA for the six pollutants were the best, with values for PM10, PM2.5, SO2, NO2, CO, and O3 of 0.0854, 0.0701, 0.0160, 0.0955, 0.0015, and 0.0895, respectively.
A comparison between the predicted values and true values of different pollutants with the adaptive GRU network in Dataset 3 is shown in Figure 9.
The experimental results demonstrate that air quality prediction based on the adaptive GRU using a genetic algorithm proposed in this paper is superior to that based on SVM, RNN, and LSTM methods and can obtain more accurate prediction results.

5. Conclusions

In this paper, in order to better solve air-quality data with time-sequence information, we chose a GRU to address the task of air quality prediction. Inspired by the genetic algorithm, the adaptive GRU network structure was obtained via genetic processing, and optimal prediction results of air quality were achieved. Compared with other, previously used air-quality-prediction methods, our proposed method showed a better performance in the air-quality-prediction task with three datasets. By applying the proposed method, the effective prediction of air quality can provide the government and relevant departments with the changing trend of air quality in time, which is conducive to improving the ability of environmental-protection departments to study and judge the risk information of air pollution and providing early warnings. In addition, there are many factors influencing the air quality, including not only meteorological factors but also the social environment, human factors, and the geographical environment, etc., but due to the limited data acquirement, there are insufficient data related to local production and living in Xi‘an, such as the distribution of polluting enterprises, people’s life customs, etc. Therefore, if we want to consider the effect of the whole index, a series of more detailed and effective work needs to be conducted in the later stage.

Author Contributions

Conceptualization, C.D. and Z.Z.; methodology, C.D., L.Z. and X.X.; validation, Y.Z., C.D., S.Z. and D.W; formal analysis, L.Z. and X.X.; investigation, Z.Z. and C.D.; resources, C.D. and D.W.; data curation, C.D. and Z.Z.; writing—original draft preparation, C.D. and Z.Z.; writing—review and editing, C.D., Y.Z., L.Z. and D.W; supervision, Y.Z., X.X. and L.Z.; project administration, Y.Z. and X.W.; funding acquisition, C.D., D.W. and L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundations of China (grant no. 61901369, grant no. 62101454, grant no. 61834005, and grant no. 61772417), the Foundation of National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology (grant no. 20200203), and the National Key Research and Development Project of China (No. 2020AAA0104603).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data can only be used for research and verify the effectiveness of the method, which cannot be used for financial gain.

Acknowledgments

We acknowledge the government of Xi’an city for gathering the air-quality-pollutant information.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; and in the decision to publish the results.

References

  1. Aliyu, Y.A.; Botai, J.O. Reviewing the local and global implications of air pollution trends in Zaria, northern Nigeria-ScienceDirect. Urban Clim. 2018, 26, 51–59. [Google Scholar] [CrossRef] [Green Version]
  2. Aliyu, Y.A.; Musa, I.J.; Jeb, D.N. Geostatistics of pollutant gases along high traffic points in Urban Zaria, Nigeria. Int. J. Geomat. Geosci. 2014, 5, 19–31. [Google Scholar]
  3. Alegria, A.; Barbera, R.; Boluda, R.; Errecalde, F.; Farré, R.; Lagarda, M.J. Environmental cadmium, lead and nickel contamination: Possible relationship between soil and vegetable content. Fresenius J. Anal. Chem. 1991, 339, 654–657. [Google Scholar] [CrossRef]
  4. Ercilla-Montserrat, M.; Muoz, P.; Montero, J.I.; Gabarrell, X.; Rieradevall, J. A study on air quality and heavy metals content of urban food produced in a Mediterranean city (Barcelona). J. Clean. Prod. 2018, 195, 385–395. [Google Scholar] [CrossRef]
  5. Wang, B.; Hong, G.; Qin, T.; Fan, W.R.; Yuan, X.C. Factors governing the willingness to pay for air pollution treatment: A case study in the Beijing-Tianjin-Hebei region. J. Clean. Prod. 2019, 235, 1304–1314. [Google Scholar] [CrossRef]
  6. Cohen, A.J.; Brauer, M.; Burnett, R.; Anderson, H.R.; Frostad, J.; Estep, K.; Balakrishnan, K.; Brunekreef, B.; Dandona, L.; Dandona, R.; et al. Estimates and 25-year trends of the global burden of disease attributable to ambient air pollution: An analysis of data from the Global Burden of Diseases Study 2015. Lancet 2017, 389, 1907–1918. [Google Scholar] [CrossRef] [Green Version]
  7. Yin, P.; He, G.; Fan, M.; Chiu, K.Y.; Fan, M.; Liu, C.; Xue, A.; Liu, T.; Pan, Y.; Mu, Q.; et al. Particulate air pollution and mortality in 38 of China’s largest cities: Time series analysis. BMJ Clin. Res. 2017, 356, j667. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. United Nations. The World’s Cities in 2018 Data Booklet, 1009 Economics & Social Affairs. 2018. Available online: https://www.un.org/development/desa/pd/content/worlds-cities-2018-data-booklet (accessed on 13 April 2020).
  9. Afzali, A.; Rashid, M.; Afzali, M.; Younesi, V. Prediction of Air Pollutants Concentrations from Multiple Sources Using AERMOD Coupled with WRF Prognostic Model. J. Clean. Prod. 2017, 166, 1216–1225. [Google Scholar] [CrossRef]
  10. Todorov, V.; Dimov, I.; Ostromsky, T.; Apostolov, S.; Georgieva, R.; Dimitrov, Y.; Zlatev, Z. Advanced stochastic approaches for Sobol’ sensitivity indices evaluation. Neural Comput. Appl. 2021, 33, 1999–2014. [Google Scholar] [CrossRef]
  11. Dimov, I.; Todorov, V.; Sabelfeld, K. A study of highly efficient stochastic sequences for multidimensional sensitivity analysis. Monte Carlo Methods Appl. 2022, 28, 1–12. [Google Scholar] [CrossRef]
  12. Roeva, O.; Fidanova, S.; Paprzycki, M. Population Size Influence on the Genetic and Ant Algorithms Performance in Case of Cultivation Process Modeling. In Recent Advances in Computational Optimization. Studies in Computational Intelligence; Fidanova, S., Ed.; Springer: Cham, Switzerland, 2015; Volume 580. [Google Scholar]
  13. Zhou, X.; Xu, J.; Zeng, P.; Meng, X. Air pollutant concentration prediction based on GRU method. J. Phys. Conf. Ser. 2019, 1168, 032058. [Google Scholar] [CrossRef]
  14. Christensen, J.H. The Danish eulerian hemispheric model—A three-dimensional air pollution model used for the arctic. Atmos. Environ. 1997, 31, 4169–4191. [Google Scholar] [CrossRef]
  15. Zhu, J.Y.; Sun, C. Data Classification for Air Quality on Wireless Sensor Network Monitoring System Using Decision Tree Algorithm. In Proceedings of the 2016 2nd International Conference on Science and Technology-Computer (ICST), Yogyakarta, Indonesia, 27–28 October 2016. [Google Scholar]
  16. Kujaroentavon, K.; Kiattisin, S.; Leelasantitham, A.; Thammaboosadee, S. Air quality classification in Thailand based on decision tree. In Proceedings of the 7th 2014 Biomedical Engineering International Conference, Fukuoka, Japan, 26–28 November 2014. [Google Scholar]
  17. Zhang, C.; Yuan, D. Fast Fine-Grained Air Quality Index Level Prediction Using Random Forest Algorithm on Cluster Computing of Spark. In Proceedings of the 2015 IEEE 12th International Conference on Ubiquitous Intelligence and Computing and 2015 IEEE 12th International Conference on Autonomic and Trusted Computing and 2015 IEEE 15th International Conference on Scalable Computing and Communications and Its Associated Workshops (UIC-ATC-ScalCom), Beijing, China, 10–14 August 2015. [Google Scholar]
  18. Ghaemi, Z.; Farnaghi, M.; Alimohammadi, A. Hadoop-Based Distributed System for Online Prediction of Air Pollution Based on Support Vector Machine, The International Archives of the Photogrammetry. Remote Sens. Spat. Inf. Sci. 2015, 40, 215–219. [Google Scholar]
  19. Gao, S.; Hu, H.; Li, Y.; Bai, Y. Prediction of Ait Quality Index Based on MFO-SVM. J. North Univ. China Nat. Sci. Ed. 2018, 39, 373–379. [Google Scholar]
  20. Kingsy, R.; Grace, R.; Manimegalai, M.S.; Geetha Devasena, S.; Rajathi, K.; Usha, N.; Raabiathul, B. Air Pollution Analysis using Enhanced K-Means Clustering Algorithm for Real Time Sensor Data. In Proceedings of the IEEE Region 10 Conference, Singapore, 22–25 November 2016. [Google Scholar] [CrossRef]
  21. Taghavifar, H.; Mardani, A.; Mohebbi, A.; Khalilarya, S.; Jafarmadar, S. Appraisal of artificial neural networks to the emission analysis and prediction of CO2, soot, and NOx of n-heptane fueled engine. J. Clean. Prod. 2016, 112, 1729–1739. [Google Scholar] [CrossRef]
  22. Taylan, O. Modelling and analysis of ozone concentration by artificial intelligent techniques for estimating air quality. Atmos. Environ. 2017, 150, 356–365. [Google Scholar] [CrossRef]
  23. Gao, M.; Yin, L.; Ning, J. Artificial neural network model for ozone concentration estimation and Monte Carlo analysis. Atmos. Environ. 2018, 184, 129–139. [Google Scholar] [CrossRef]
  24. Nieto, P.; García-Gonzalo, E.; Sánchez, A.B.; Rodríguez Miranda, A.A. Air Quality Modeling Using the PSO-SVM-Based Approach, MLP Neural Network, and M5 Model Tree in the Metropolitan Area of Oviedo (Northern Spain). Environ. Model. Assess. 2018, 23, 229–247. [Google Scholar] [CrossRef]
  25. Gholizadeh, M.H.; Darand, M. Forecasting the Air Pollution with using Artificial Neural Networks: The Case Study; Tehran City. J. Appl. Sci. 2009, 9, 3882–3887. [Google Scholar] [CrossRef]
  26. Ai, H.; Shi, Y. Study on Prediction of Haze Based on BP Neural Network. Comput. Simul. 2015, 32, 402–405. [Google Scholar]
  27. Zhao, W.Y.; Xia, L.S.; Gao, G.K.; Cheng, L. PM2.5 prediction model based on weighted KNN-BP neural network. J. Environ. Eng. Technol. 2019, 9, 14–18. [Google Scholar]
  28. Ma, C.; Zu, J.; Fu, Q.; Luo, L. Air visibility forecast based on genetic neural network model. Chin. J. Environ. Eng. 2015, 9, 1905–1910. [Google Scholar]
  29. Li, K.; Zhou, R.; Xu, H. Based on Hopfield neural network to determine the air quality levels. In Proceedings of the 2011 International Conference on Business Management and Electronic Information, Guangzhou, China, 13–15 May 2011. [Google Scholar]
  30. Wang, X.; Zhang, Y.; Zhao, S.; Zhang, L. Air quality forecasting based on dynamic granular wavelet neural network. Comput. Eng. Appl. 2013, 49, 221–224. [Google Scholar]
  31. Fan, J.; Li, Q.; Zhu, Y.; Hou, J.; Feng, X. Aspatio-temporal prediction framework for air polution based on dep RNN. Sci. Surv. Mapp. 2017, 4, 15. [Google Scholar]
  32. Yang, C.; Wang, Y.; Shu, Z.; Liu, J.; Xie, N. Application of LSTM Model Based on TensorFlow in Air Quality Index Prediction. Digit. Technol. Appl. 2021, 39, 203–206. [Google Scholar]
  33. Li, X.; Peng, L.; Yao, X.; Cui, S.; Hu, Y.; You, C.; Chi, T. Long short-term memory neural network for air pollutant concentration predictions: Method development and evaluation. Environ. Pollut. 2017, 231, 997–1004. [Google Scholar] [CrossRef]
  34. Zhao, J.; Deng, F.; Cai, Y.; Chen, J. Long short-term memory-fully connected (LSTM-FC) neural network for PM2.5concentration prediction. Chemosphere 2019, 220, 486–492. [Google Scholar] [CrossRef]
  35. Lu, H.; Xie, M.; Wu, Z.; Liu, B.; Gao, Y.; Chen, G.; Li, Z. Chengyu region machine learning WRF-CMAQ PM2.5 concentration numerical air quality forecast. Acta Sci. Circumstantiae 2020, 40, 4419–4431. [Google Scholar]
  36. Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  37. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  38. Houck, C.; Joines, J.; Kay, M. A Genetic Algorithm for Function Optimization: A Matlab Implementation; Technical Report; North Carolina State University: Raleigh, NC, USA, 2009. [Google Scholar]
  39. Reeves, C. A Genetic Algorithm for Flowshop Sequencing. Comput. Oper. Res. 1995, 22, 5–13. [Google Scholar] [CrossRef]
  40. Beasley, J.; Chu, P. A Genetic Algorithm for the Set Covering Problem. Eur. J. Oper. Res. 1996, 94, 392–404. [Google Scholar] [CrossRef]
  41. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  42. Grefenstette, J.; Gopal, R.; Rosmaita, B.; van Gucht, D. Genetic Algorithms for the Traveling Salesman Problem. In Proceedings of the International Conference on Genetic Algorithms and Their Applications, Pittsburg, PA, USA, 24–26 July 1985. [Google Scholar]
  43. Yao, X. Evolving Artificial Neural Networks. Proc. IEEE 1999, 87, 423–1447. [Google Scholar]
  44. Snyder, L.; Daskin, M. A Random-Key Genetic Algorithm for the Generalized Traveling Salesman Problem. Eur. J. Oper. Res. 2006, 174, 38–53. [Google Scholar] [CrossRef]
  45. Bayer, J.; Wierstra, D.; Togelius, J.; Schmidhuber, J. Evolving Memory Cell Structures for Sequence Learning. In Proceedings of the International Conference on Artificial Neural Networks, Limassol, Cypros, 14–17 September 2009. [Google Scholar]
  46. Ding, S.; Li, H.; Su, C.; Yu, J.; Jin, F. Evolutionary Artificial Neural Networks: A Review. Artif. Intell. Rev. 2013, 39, 251–260. [Google Scholar] [CrossRef]
  47. Xie, L.; Yuille, A. Genetic CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), IEEE Computer Society, Venice, Italy, 22–29 October 2017. [Google Scholar]
  48. Baker, J.E. Adaptive selection methods for genetic algorithms. In Proceedings of the International Conference on Genetic Algorithms and Their Applications, Pittsburg, PA, USA, 24–26 July 1985; pp. 101–111. [Google Scholar]
Figure 1. The basic unit structure of the gated recurrent neural network.
Figure 1. The basic unit structure of the gated recurrent neural network.
Information 13 00223 g001
Figure 2. Schematic diagram of the proposed GRU neural networks.
Figure 2. Schematic diagram of the proposed GRU neural networks.
Information 13 00223 g002
Figure 3. Example of a binary string for the network architecture.
Figure 3. Example of a binary string for the network architecture.
Information 13 00223 g003
Figure 4. The generations and corresponding best fitness in Dataset 1.
Figure 4. The generations and corresponding best fitness in Dataset 1.
Information 13 00223 g004
Figure 5. The generations and corresponding best fitness in Dataset 2.
Figure 5. The generations and corresponding best fitness in Dataset 2.
Information 13 00223 g005
Figure 6. The generations and corresponding best fitness in Dataset 3.
Figure 6. The generations and corresponding best fitness in Dataset 3.
Information 13 00223 g006
Figure 7. Comparison of the predicted values and real values of different pollutants with the GRU network in Dataset 1: (a) PM10; (b) PM2.5; (c) SO2; (d) NO2; (e) CO; (f) O3.
Figure 7. Comparison of the predicted values and real values of different pollutants with the GRU network in Dataset 1: (a) PM10; (b) PM2.5; (c) SO2; (d) NO2; (e) CO; (f) O3.
Information 13 00223 g007aInformation 13 00223 g007b
Figure 8. Comparison of predicted values and real values of different pollutants with the GRU network in Dataset 2: (a) PM10; (b) PM2.5; (c) SO2; (d) NO2; (e) CO; (f) O3.
Figure 8. Comparison of predicted values and real values of different pollutants with the GRU network in Dataset 2: (a) PM10; (b) PM2.5; (c) SO2; (d) NO2; (e) CO; (f) O3.
Information 13 00223 g008
Figure 9. Comparison of the predicted values and real values of different pollutants with the GRU network in Dataset 3: (a) PM10; (b) PM2.5; (c) SO2; (d) NO2; (e) CO; (f) O3.
Figure 9. Comparison of the predicted values and real values of different pollutants with the GRU network in Dataset 3: (a) PM10; (b) PM2.5; (c) SO2; (d) NO2; (e) CO; (f) O3.
Information 13 00223 g009aInformation 13 00223 g009b
Table 1. Air-pollutant-data examples at the station of Xincheng Central Square in Xi’an from 2018 to 2020.
Table 1. Air-pollutant-data examples at the station of Xincheng Central Square in Xi’an from 2018 to 2020.
TimePM10
(μg/m3)
PM2.5
(μg/m3)
SO2
(μg/m3)
NO2
(μg/m3)
CO
(mg/m3)
O3
(μg/m3)
2018-01-01 01:0043620127852.25
2020-12-31 23:0026721821822.47
Table 2. The examples of the dataset D .
Table 2. The examples of the dataset D .
OrderTimePM10PM2.5SO2NO2COO3
02018-01-01 01:005.61612.1780−0.36760.4809−0.7304−0.6894
12018-01-01 02:005.13333.0558−0.41150.3639−0.7260−0.6748
26,2762020-12-31
23:00
3.14362.4267−0.45530.4370−0.7275−0.6602
Table 3. The example of the sample.
Table 3. The example of the sample.
OrderTimePM10PM2.5SO2NO2COO3
02018-01-01 01:005.61612.1780−0.36760.4809−0.7304−0.6894
12018-01-01 02:005.13333.0558−0.41150.3639−0.7260−0.6748
232018-01-02 00:006.14283.3777−0.04570.8613−0.7158−0.6894
242018-01-02 01:006.75735.3528−0.17740.7589−0.7099−0.6309
Table 4. The quantity of data in three datasets.
Table 4. The quantity of data in three datasets.
Dataset NameXi‘an Xincheng Center Square StationXi‘an Caotang BaseXi‘an Gaoxin West Station
Dataset NumberDataset 1Dataset 2Dataset 3
Data Quantity25,56914,56714,568
Table 5. The quantity of samples in the reconstructed datasets.
Table 5. The quantity of samples in the reconstructed datasets.
Dataset NameXi‘an Xincheng Center Square StationXi‘an Caotang BaseXi‘an Gaoxin West Station
Dataset NumberDataset 1Dataset 2Dataset 3
Sample Quantity25,54514,54314,544
Table 6. Number of samples in training set, validation set, and testing set of the reconstructed datasets.
Table 6. Number of samples in training set, validation set, and testing set of the reconstructed datasets.
DatasetNumber of Samples
Training SetValidation SetTesting Set
Dataset 116,09317887664
Dataset 2915710174360
Dataset 3915710184360
Table 7. The best fitness and corresponding network structure in Dataset 1.
Table 7. The best fitness and corresponding network structure in Dataset 1.
GenerationsBest FitnessNetwork Structure
00514.130110011000|0001001001
01514.130110011000|0001001001
03536.760110011100|0001001001
05548.550110011101|0000101001
08548.550110011101|0000101001
10548.550110011101|0000101001
30551.220111011101|0000101011
50574.820111100010|0000101010
80574.820111100010|0000101010
100575.370110110010|0000101111
Table 8. The best fitness and corresponding network structure in Dataset 2.
Table 8. The best fitness and corresponding network structure in Dataset 2.
GenerationsBest FitnessNetwork Structure
001722.220010111110|0000001101
011909.790101111000|0000010101
031919.860010111101|0001011110
051967.900111000111|0001011101
082043.250111110011|0000010110
102043.250111110011|0000010110
302043.250111110011|0000010110
502053.950111110111|0000010110
802058.130110110111|0000010100
1002058.130110110111|0000010100
Table 9. The best fitness and corresponding network structure in Dataset 3.
Table 9. The best fitness and corresponding network structure in Dataset 3.
GenerationsBest FitnessNetwork Structure
002125.730010111000|0001000010
012204.670011101111|0000101111
032213.150010010110|0000110010
052273.150011101011|0000101011
082311.430011101011|0000100001
102311.430110010110|0000101001
302381.330111011101|0000100010
502382.010111110011|0000011010
802404.840111100111|0000011011
1002404.840111100111|0000011011
Table 10. The result of the adaptive GRU network structure compared with manually designed GRU network structures with Dataset 1.
Table 10. The result of the adaptive GRU network structure compared with manually designed GRU network structures with Dataset 1.
Air PollutantRMSESMAPE
GRU1GRU2GRU_GAGRU1GRU2GRU_GA
PM100.02640.06470.02240.08700.19370.0855
PM2.50.01360.03920.00780.06990.30410.0678
SO20.00670.00650.00350.01540.27340.0143
NO20.00760.02020.00680.09740.38350.0959
CO0.00880.00030.00010.00070.00140.0007
O30.00870.03590.00880.07590.36080.0761
Table 11. The result of the adaptive GRU network structure compared with manually designed GRU network structures with Dataset 2.
Table 11. The result of the adaptive GRU network structure compared with manually designed GRU network structures with Dataset 2.
Air PollutantRMSESMAPE
GRU1GRU2GRU_GAGRU1GRU2GRU_GA
PM100.03880.07220.03390.12860.23070.1297
PM2.50.00660.03300.00570.07770.36500.0692
SO20.00250.00290.00230.01360.01500.0127
NO20.00780.01860.00750.10450.21950.1035
CO0.00040.00030.00040.00200.00160.0022
O30.01190.04410.01150.10030.21640.0978
Table 12. The result of the adaptive GRU network structure compared with manually designed GRU network structures with Dataset 3.
Table 12. The result of the adaptive GRU network structure compared with manually designed GRU network structures with Dataset 3.
Air PollutantRMSESMAPE
GRU1GRU2GRU_GAGRU1GRU2GRU_GA
PM100.02730.06520.02080.08650.21360.0854
PM2.50.00850.04250.00780.07400.32700.0701
SO20.00450.00540.00280.01950.02850.0160
NO20.00820.02460.00840.10350.38590.0971
CO0.00030.00030.00220.00140.00160.0015
O30.00990.04300.01000.08740.31240.0895
Table 13. The prediction results of GRU_GA compared with other air-quality-prediction methods with Dataset 1.
Table 13. The prediction results of GRU_GA compared with other air-quality-prediction methods with Dataset 1.
Air PollutantRMSESMAPE
SVMRNNLSTMGRU_GASVMRNNLSTMGRU_GA
PM100.67320.64710.64700.02240.19360.19370.18370.0855
PM2.50.40310.39230.39210.00780.30310.30630.30410.0678
SO20.00780.00660.00650.00350.02900.02850.02850.0143
NO20.02150.02030.02020.00680.37110.37250.36170.0959
CO0.00030.00030.00030.00010.00150.00120.00140.0007
O30.03550.03580.03590.00880.37270.36010.34030.0761
Table 14. The prediction results of GRU_GA compared with other air-quality-prediction methods with Dataset 2.
Table 14. The prediction results of GRU_GA compared with other air-quality-prediction methods with Dataset 2.
Air PollutantRMSESMAPE
SVMRNNLSTMGRU_GASVMRNNLSTMGRU_GA
PM100.07220.07220.07220.03390.23060.23060.23070.0854
PM2.50.03310.03300.03300.00570.35890.36590.36500.0701
SO20.00300.00290.00310.00230.01510.01500.01500.0160
NO20.01860.01870.01860.00750.21920.23410.21950.0971
CO0.00030.00040.00070.00040.00160.00330.00160.0015
O30.04400.04410.04410.01150.21670.21730.21640.0895
Table 15. The prediction results of GRU_GA compared with other air-quality-prediction methods with Dataset 3.
Table 15. The prediction results of GRU_GA compared with other air-quality-prediction methods with Dataset 3.
Air PollutantRMSESMAPE
SVMRNNLSTMGRU_GASVMRNNLSTMGRU_GA
PM100.06510.06510.06520.02080.21300.21280.21330.0854
PM2.50.04250.04250.04250.00780.32600.32700.32770.0701
SO20.00540.00540.00550.00280.02840.02840.02910.0160
NO20.02460.02460.02470.00840.16730.16810.28370.0955
CO0.00030.00030.00060.00220.00160.00160.00280.0015
O30.04300.04300.04310.01000.31320.31390.31500.0895
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ding, C.; Zheng, Z.; Zheng, S.; Wang, X.; Xie, X.; Wen, D.; Zhang, L.; Zhang, Y. Accurate Air-Quality Prediction Using Genetic-Optimized Gated-Recurrent-Unit Architecture. Information 2022, 13, 223. https://doi.org/10.3390/info13050223

AMA Style

Ding C, Zheng Z, Zheng S, Wang X, Xie X, Wen D, Zhang L, Zhang Y. Accurate Air-Quality Prediction Using Genetic-Optimized Gated-Recurrent-Unit Architecture. Information. 2022; 13(5):223. https://doi.org/10.3390/info13050223

Chicago/Turabian Style

Ding, Chen, Zhouyi Zheng, Sirui Zheng, Xuke Wang, Xiaoyan Xie, Dushi Wen, Lei Zhang, and Yanning Zhang. 2022. "Accurate Air-Quality Prediction Using Genetic-Optimized Gated-Recurrent-Unit Architecture" Information 13, no. 5: 223. https://doi.org/10.3390/info13050223

APA Style

Ding, C., Zheng, Z., Zheng, S., Wang, X., Xie, X., Wen, D., Zhang, L., & Zhang, Y. (2022). Accurate Air-Quality Prediction Using Genetic-Optimized Gated-Recurrent-Unit Architecture. Information, 13(5), 223. https://doi.org/10.3390/info13050223

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop