Next Article in Journal
Identification of Paddy Varieties from Landsat 8 Satellite Image Data Using Spectral Unmixing Method in Indramayu Regency, Indonesia
Previous Article in Journal
A GIS Pipeline to Produce GeoAI Datasets from Drone Overhead Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Passenger Flow Prediction of Scenic Spots in Jilin Province Based on Convolutional Neural Network and Improved Quantile Regression Long Short-Term Memory Network

School of Mathematics and Statistics, Changchun University of Technology, Changchun 130012, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2022, 11(10), 509; https://doi.org/10.3390/ijgi11100509
Submission received: 30 July 2022 / Revised: 24 September 2022 / Accepted: 26 September 2022 / Published: 30 September 2022

Abstract

:
Passenger flow is an important benchmark for measuring tourism benefits, and accurate tourism passenger flow prediction is of great significance to the government and related tourism enterprises and can promote the sustainable development of China’s tourism industry. For daily passenger flow time series data, a passenger flow forecasting method based on convolutional neural network (CNN) and improved quantile regression long short-term memory network (QRLSTM), denoted as CNN-IQRLSTM, is proposed with reconstructed correlation features and in the form of sliding windows as inputs. First, four discrete variables such as whether the day is a weekend and holiday are created by time; then, a sliding window of width 42 is used to pass the passenger flow data into the network sequentially; finally, the loss function of the sparse Laplacian improved QRLSTM is introduced for passenger flow prediction, and the point prediction and interval prediction results under different quartiles are obtained. The application of quantile regression captures the overall picture of the data, enhances the robustness, fit, predictive power and nonlinear processing capability of neural networks, and fills the gap between quantile regression and neural network methods in the field of passenger flow prediction. CNN can effectively handle complex input data, and the improved nonlinear QR model can provide passenger flow quantile prediction information. The method is applied to the tourism traffic prediction of four 5A scenic spots in Jilin Province, and the effectiveness of the method is verified. The results show that the method proposed in this paper fits best in point prediction and has higher prediction accuracy. The MAPE of the Changbai Mountain dataset was 0.07, the MAPE of the puppet palace museum dataset was 0.05, the fit of the Sculpture Park dataset reached 93%, and the fit of the net moon lake dataset was as high as 99%. Meanwhile, the interval prediction results show that the method has a larger interval coverage as well as a smaller interval average width, which improves the prediction efficiency. In 95% of the interval predictions, the interval coverage of Changbai Mountain data is 99% and the interval average width is 0.49. It is a good reference value for the management of different scenic spots.

1. Introduction

In this wonderful state of national harmony and domestic tranquility, not only do people have a certain economic base, but the face of cities has also changed like never before, making travel a necessary form of relaxation. In recent years, China has placed increased emphasis on tourism and has issued a string of relevant documents one after another in an effort to become a world power in tourism [1]. Jilin Province has made tourism a strategic choice and a decisive step in revitalizing its development, transformation, and upgrading. Over the past few years, ice and snow festivals such as the Changchun Ice and Snow Festival, the Jilin Fog Festival, the Chagan Lake Winter Fishing Festival, and the Changbai Mountain Powder Snow Festival have been held year after year, attracting visitors from all directions and increasing the flow of visitors to Jilin’s attractions. Therefore, having the correct passenger flow pattern can provide technical support to assist in decision making by setting reasonable ticket prices, setting up relevant tourism activities at attractions, and optimizing the configuration of attractions. This study uses the Jilin Province region as the research object to construct an accurate passenger flow prediction model using temporal features. A combination of quantile regression and machine learning is used for point prediction and interval prediction. The prediction results provide an important reference for scenic spots to implement development planning, maintenance and repair, and provide intelligent tourism services.
As early as the 1960s, some scholars began to study the prediction of tourism passenger flow and proposed many prediction models, such as time series models, artificial intelligence models, econometric models, and deep neural network models [2,3,4,5,6]. Many research results have been achieved in theory and practice. For example, Lim et al. [7] constructed a seasonal ARMA model for tourism flow forecasting to analyze the seasonality of passenger flows from Hong Kong, Malaysia, and Singapore to Australia to test and estimate the monthly seasonal patterns of international tourism time series. The model captures the distinct seasonal characteristics in tourism passenger flows. Niu et al. [8] established a railroad passenger flow prediction model based on time series analysis, combining long-term trend factors, seasonal factors, and weather factors. The passenger flow change pattern under different conditions was studied, and the railroad passenger flow was predicted for the next two weeks, and the corresponding vehicle configuration optimization and station docking scheme was proposed. If the influence of weather conditions on passenger flow is studied in depth, it will make the railway passenger flow prediction model more in line with the actual situation. Li [9] analyzed the prediction progress of scenic area passenger flow using big data analysis and used the ARIMA model and BP neural network to model scenic area passenger flow. Their prediction results were weighted, and the fitting and prediction accuracy of scenic passenger flow by big data analysis were improved compared with the classical model. Cui et al. [10] constructed a tourism passenger flow prediction model based on EMD-GRU with the Heihe valley scenic area as an example, and the prediction results showed that the prediction efficiency of the EMD-GRU model was higher compared with RNN and LSTM, and the prediction model could effectively improve the accuracy of the original data prediction model after using EMD denoising. However, more explanatory variables were not explored as the influencing factors of scenic traffic to construct the prediction model. Lu et al. [11] established a prediction method (GA-CNN-LSTM) combining CNN and LSTM and optimized by genetic algorithm (GA), which was more accurate than other intelligent algorithms in terms of MAPE and R predicting daily visitor flows more accurately than other intelligent algorithms. Although the GA-CNN-LSTM algorithm has higher accuracy in peak hours than other algorithms, the overall prediction accuracy for peak hours is still insufficient. Xu [12] designed a regression-analysis-based model for predicting cultural tourism flows in the Yangtze River Delta, with the competitiveness of flows as the core, and selected 28 indicators from four aspects: cultural tourism brand resources, cultural tourism support and protection, and urban tourism market income, to build an evaluation index system for the influencing factors of flows, and the designed model has a promising fit. Chen et al. [13] combined residual networks with fully connected networks to provide an enhanced Quad-ResNet model for predicting the regional tourist flow of rural tourism. This method can predict the regional passenger flow of rural tourism based on pedestrian location data, weather and holiday data to find hotspots of tourist attractions. However, since the regional passenger flow of rural tourism in this study was obtained from pedestrian data, there are still some aspects that need to be improved.
When we study regression models, we usually consider different loss functions [14]. The most commonly used one is the squared loss, i.e., the least squares estimate. However, it is sensitive to heavy-tailed distributions and outliers and is not robust. QR models [15] can not only handle heteroskedasticity and outliers but can also construct confidence intervals for parameters using empirical likelihood inference methods without the need to estimate asymptotic variance. Zou et al. [16] first proposed a composite QR method by considering the losses of multiple quantile points simultaneously, which greatly extended the theoreticality and stability of the estimation results. In 2020, Rodrigues et al. [17] proposed a multi-output multi-quantile deep neural network structure based on convolutional long- and short-term memory layers, which can solve the embarrassing quantile crossover problem by approximating the quantile regression problem from a multi-task learning perspective using two large-scale datasets in transportation, while significantly outperforming existing quantile regression methods. Yu et al. [18] combined the advantages of hybrid neural networks and QR and proposed a spatio-temporal QR (SQR) algorithm for short-term nonparametric probabilistic prediction of regional wind power, and SQR enhanced the prediction effect with highly reliable performance. In the SQR-based forecasting process, all inputs are mapped directly to regional wind power, avoiding the cumulative errors caused by combining forecasts from individual wind farms.
At present, passenger flow forecasting methods are mainly divided into quantitative forecasting and qualitative forecasting. Qualitative forecasting is usually based on qualitative analysis combined with empirical judgment, and the forecasting accuracy is low. Quantitative forecasting is mainly based on mathematical methods to establish quantitative forecasting models. However, in the actual forecasting process, many methods only consider a single or a few factors affecting passenger flow, which lacks comprehensiveness and leads to bias in passenger flow forecasting. Algorithms such as random forest and support vector machine are widely used in passenger flow forecasting, but each algorithm has its drawbacks. Random forest tends to perform poorly when it encounters noisy data. Support vector machines also have the problem of slow computing speed. QR models can not only handle heteroskedasticity and outliers but can also construct confidence intervals for the parameters using empirical likelihood inference methods without estimating the asymptotic variance. The loss function of QR plays a crucial role in the prediction results, and this paper improves the loss function to fill the problem of low prediction accuracy. On the other hand, the application of QR can capture the overall situation of the data and enhance the robustness, fit, predictive power, and nonlinear processing ability of neural networks, which fills the gap between QR and neural network methods in the field of passenger flow prediction.
Scientific and reasonable prediction of tourist flow is an important guiding significance for the effective use of tourism resources and local economic development [19]. However, there is a lack of point prediction models with high accuracy, and there is less research on interval prediction of tourism passenger flow. In this paper, we take the tourist passenger flow of 5A scenic spots in Jilin Province as the research object, construct a deep network framework consisting of a convolutional layer, pooling layer, and improved QRLSTM, extend LSTM to QR, and compare it with the tree model, RNN, LSTM, GRU, and other models, and we can find that the proposed method significantly outperforms the baseline model. The results of the study contribute to the formation of a knowledge system for scenic area management, which can be used as a decision-making tool for tourism managers.
The main contributions of this paper can be summarized as follows.
(1).
This study extends the proposed sparse Laplacian quantile loss function to LSTM, adopts the network structure constraint as the penalty term of the objective function, and smoothens the deviation degree of the network weights in the iterative correction process according to the sparse Laplacian in order to improve the robustness of the prediction.
(2).
An IQRLSTM deep network framework model combined with CNN is proposed for point prediction and interval prediction for scenic tourist passenger flow data in Jilin Province, providing a reliable basis for uncertainty analysis of passenger flow.
(3).
Four relevant data features are added for the date attribute, combined with the sliding window extracted features as input data to obtain more information about the passenger flow, providing a new perspective and idea for the accurate prediction of tourism passenger flow.
(4).
The CNN-IQRLSTM is critically evaluated on four scenic spot passenger flow datasets, and by comparing multiple tree models and neural network models, it is shown that the method in this paper significantly outperforms other baseline models.
The remainder of this paper is organized as follows: the material and methods of the paper are described in detail in Section 2. Section 3 of the paper presents the findings of the paper. Section 4 contains the discussion and conclusions of the paper.

2. Materials and Methods

2.1. Materials

The methods involved in this study are CNN, LSTM, and QR. The CNN model is one of the most popular and widely used models in the field of deep learning in recent years. The one-dimensional convolutional structure is simpler, with less weight input, and the complexity of data reconstruction and feature extraction is reduced, according to which the model extracts time-series features and deeply explores the connection between input data. LSTM is a kind of artificial neural network with the ability to be responsible for calculating the dependence between individual observations in the time series, and also has the inherent ability to quickly adapt to the sharp changes in the trend and has a high degree of fit to the passenger flow change trend. Compared with ordinary least squares estimation, QR estimation can more accurately describe the effect of the explanatory variables on the range of variation of the explained variables and the shape of the conditional distribution. It provides a more comprehensive characterization of the distribution, leading to a comprehensive analysis. Although LSTM is very good at processing time series data for predicting visitor flow, it cannot mine the effective information of data distribution. To solve this problem, a new model based on LSTM is built.
In this paper, daily historical data from August 2017 to December 2021 for four scenic spots, namely Changbai Mountain, the puppet palace museum, Sculpture Park, and net moon lake, were selected as the raw data, and all data were provided by the Jilin Provincial Tourism Information Center. Since the sample size of the dataset was not very large, this study initially divided the training and test sets into the ratios of 8:2 and 9:1, respectively, and the results showed that the prediction accuracy of dividing in the ratio of 8:2 was lower than that of dividing in the ratio of 9:1. Therefore, we decided to choose the ratio of 9:1. Figure 1 shows the time series of tourist passenger flow for the four scenic spots. Changbai Mountain is a tourist destination with the reputation of “sacred mountain, holy water, strange forest and immortal fruit”, and is one of the 10 most famous mountains in China. The puppet palace museum of the False Manchus is a unique humanistic scenic spot that combines the False Manchus Palace, red tourism, cultural and leisure areas, and tourism and commercial services. Changchun World Sculpture Park is a famous practice base for sculpture research and teaching in China. Net moon lake is a natural landscape formed by 100 square kilometers of artificial forest surrounded by a pool of beautiful water and is famous at home and abroad for its strong tourism resources and superior ecological environment.
As the original data is time series data and contains less information, the general forecasting method is to use a sliding window forecast, for example, window 1 August 2017 to 11 September 2017 and the next window is 2 August 2017 to 12 September 2017, with the overall moving one unit to the right instead of one window. In this paper, the features of the data were expanded to include four discrete variables of the day of the week, week ordinal of the current month, whether it is a weekend, and whether it is a holiday, and then combined with the sliding window as the final feature to forecast the scenic passenger flow on a case-by-case basis, and Table 1 shows the partial input data of Changbai Mountain. Table 1 shows the passenger flow data for Changbai Mountain from 10 to 24 August 2017. The “Date” column indicates the date; the “Flow” column indicates the daily passenger flow. “Weekday” indicates the day of the week, ranging from 1 to 7, with 6 and 7 for Saturday and Sunday. “Week” indicates the week of the month, ranging from 1 to 5. “Weekend” indicates whether the day is a weekend, if yes, filled with 1, otherwise filled with 0; “Holiday” indicates whether the day is a holiday, if yes, filled with 1, otherwise filled with 0.

2.2. Methods

2.2.1. CNN Model

The CNN was proposed by LeCun [20]. The CNN uses local connectivity and weight sharing to extract features from the original data and build a dense and complete feature vector. This study uses CNN to extract data features. The original data is processed at a higher level and more abstractly through convolutional and pooling layers to obtain the internal features in the data, to deeply mine the connections between the input data, and finally to pass the features into the LSTM network after processing. Figure 2 shows the structure of the CNN.
Convolutional layer: convolutional operation is performed on passenger traffic data using convolutional layer to extract hidden features. In the deep learning process, the problem is solved by sharing the parameters of convolutional kernels using the weight sharing process of CNN. Here, each convolutional kernel has an acceptance domain for extracting local neurons from the previous layer. Additionally, the neurons between different layers are locally connected [21]. The feature mapping of the convolutional layer is obtained by computing the dot product of the feature mapping of the previous layer with the convolutional kernel, and then it is nonlinearized by the activation function as follows [21]:
C j l = f ( i N i I i l 1 w i , j l + b j l )
where f ( ) is the activation function. I i l 1 denotes the feature mapping in layer l 1 , is the convolution operation, N i denotes the input set of the feature mapping, w i , j l is the weight of feature mapping i in layer l 1 with respect to feature mapping j in layer l , b j l denotes the bias of feature mapping i in layer l 1 to feature mapping j in layer l , and C j l denotes feature mapping j in layer l .
Pooling layer: The effect of pooling is to downsample. The pooling layer can reduce the dimensionality of the kernel by preserving the significant features and increasing the perceptual field of the kernel. The pooling layer can reduce the dimensionality of the extracted feature information, which makes the feature map smaller, simplifies the computational complexity of the network, and avoids overfitting to a certain extent; on the other hand, feature compression is performed to extract the main features.

2.2.2. Long Short-Term Memory Network (LSTM)

LSTM is an improved model based on Recurrent Neural Network (RNN), which can be an effective solution to the emergent long-range dependency problem. LSTM cleverly preserves long and short-term memory through memory units and gating mechanisms, and its basic unit architecture is shown in Figure 3. LSTM units consist of forgetting gates, input gates and output gates. The forgetting gate controls the extent to which historical information is forgotten, the input gate controls the extent to which new information is accepted, and the output gate determines the final output [22].
Given the current input x t , the implicit layer state h t 1 and the stored state C t 1 at the previous moment, the details are calculated as follows [23].
i t = σ ( W i [ x t , h t 1 ] Τ + b i )
f t = σ ( W f [ x t , h t 1 ] Τ + b f )
O t = σ ( W o [ x t , h t 1 ] Τ + b o )
C ˜ t = tanh ( W C [ x t , h t 1 ] Τ + b C )
C t = C t 1 f t + i t C ˜ t
h t = O t tanh ( C t )
where x t is the input at the current moment; h t 1 is the output of the LSTM at the previous moment; f t , i t , O t are the results of the forgetting gate, input gate, and output gate state operations, respectively; σ is the Sigmoid activation function; W f , W i , W O are the forgetting gate, input gate, and output gate weight matrices, respectively; b f , b i , b O are the forgetting gate, input gate, and output gate bias terms, respectively; C t is the unit state of the input at the moment of t ; W C is the input unit state weight matrix; b C is the input unit state bias item [23].

2.2.3. Quantile Regression (QR)

QR is a useful and popular alternative to mean regression, particularly for biased outcome data. In addition, quartiles at multiple quantile levels can provide information for capturing the distribution of the variable of interest. The advantages over mean reversion are as follows: first, the effect of the explanatory variables on the entire conditional distribution of the response variable can be carefully portrayed. Second, no distributional assumptions need to be made about the random perturbation terms of the model, enhancing the robustness of the model construction. Third, monotonic transformability is achieved for the response variable. Fourth, the parameter estimates are asymptotically good under large sample theory. For a random variable Y , the τ th quantile of Y is generally defined as Equation (8) [24].
Q Y ( τ ) = inf { y : Pr ( Y y ) τ } , τ ( 0 , 1 )
where Q Y ( τ ) is a function on τ that gives a complete description of the distribution of the random variable Y . Thus, given the covariates, the τ th conditional quantile of Y can be defined as Equation (9).
Q Y | X ( τ ) = inf { y : Pr ( Y y | X ) τ } , τ ( 0 , 1 )
The linear QR model is defined as follows:
Q Y ( τ | X ) = X Τ β ( τ ) , τ ( 0 , 1 )
where β ( τ ) = ( β 1 ( τ ) , β 2 ( τ ) , , β n ( τ ) ) Τ is the quantile coefficient that depends on τ . The QR model can be used to estimate the regression coefficient β ( τ ) . First, define Equation (11).
ρ τ ( u ) = u ( τ I ( u 0 ) ) = { u τ                           u > 0 u ( τ 1 )       u 0
where I ( ) is the indicator function. The following QR model is developed:
min i = 1 n ρ τ ( y i X i T β )

2.2.4. QRLSTM Model

Considering the temporal and non-linear nature of passenger forecasting, the LSTM is used as a conditional quantile function for passenger flow and by optimizing the objective function a little, the parameters of the QRLSTM model can be estimated [25].
L Q R L S T M = min W ( τ i ) , b ( τ i ) i = 1 N ρ τ ( Y i f ( X i , W ( τ i ) , b ( τ i ) ) ) = i | Y i f ( X i , W ( τ i ) , b ( τ i ) τ | Y i f ( X i , W ( τ i ) , b ( τ i ) | + i | Y i < f ( X i , W ( τ i ) , b ( τ i ) ( 1 τ ) | Y i f ( X i , W ( τ i ) , b ( τ i ) |
where L Q R L S T M is the loss function of the QRLSTM model at quantile τ i ; Y i is the actual value of the sample; N is the number of samples; f ( X i , W ( τ i ) , b ( τ i ) is the output value of the LSTM network; W ( τ i ) is the weight matrix of the LSTM network; and b ( τ i ) is the bias item of the LSTM network.
When an estimate W ^ ( τ i ) , b ^ ( τ i ) of the best weight term is obtained, the conditional quartiles of the dependent variable can be treated using the following equation.
Q ^ Y ( τ | X ) = f ( X , W ^ ( τ i ) , b ^ ( τ i ) )
where Q ^ Y is the conditional quantile of Y at τ ( 0 , 1 ) .

2.2.5. Improved QRLSTM Model

Although LSTM models enhance the ability to adjust the feedback on the intrinsic features of the data compared to traditional machine learning, they are also inevitably affected by historical anomalous perturbations during the training process due to their inherent sensitivity, making the model less generalizable [26]. In this paper, the network structure constraint is used as the penalty term of the objective function to smooth the deviation of the network weights during the iterative correction process according to the sparse Laplacian in order to improve the robustness of the prediction. According to the network structure theory, each feature is assumed to be a node, and if there is a relationship between every two nodes, it means that there is an edge between these two nodes, and the weights can be derived accordingly, and the larger the weight indicates the stronger the correlation between these two variables. In this paper, the Pearson correlation coefficient is used to construct a data matrix between features and between features and labels, denoted as A = [ a i j ] p × p , called the adjacency matrix. In a grid structure where orientation is not considered, the A matrix has symmetry and the element a i j of it measures the degree of similarity between nodes i and j . The adjacency matrix is calculated as in Equation (15).
a i j = | r i j | [ I ( | r i j | > r ) ] [ I ( | r i y | > r ) ] [ I ( | r j y | > r ) ]
where r i j is the Pearson correlation coefficient between features i and j . r i y is the Pearson correlation coefficient between features i and y . r j y is the Pearson correlation coefficient between features j and y . r is the threshold value, determined according to the Fisher transformation and calculated as in Equation (16).
r = exp ( 2 c / n 3 ) 1 exp ( 2 c / n 3 ) + 1
where c is the threshold value of n 3 f i j . Establish the statistic f i j = 0.5 log ( ( 1 + r i j ) / ( 1 r i j ) ) . If the correlation between X i and X j is 0, n 3 f i j approximately obeys the standard normal distribution N ( 0 , 1 ) . The value of c is determined using hypothesis testing. The higher the value of c , the larger the threshold r , the higher the sparsity of the adjacency matrix, and vice versa the lower the sparsity. Similarly, the 0.95,0.975,0.995 quantile of the standard normal distribution can be taken, and in this paper the 0.995 quantile is taken to be 2.58.
The matrix is sparse since some elements of A can be taken to zero. The direction of the correlation is not taken into account and is applicable for both positive and negative correlations. In this study, a semi-positive definite matrix L = D A is used to construct sparse Laplace smoothing (SLS) grid constraints as shown in Equation (17).
β τ Τ L β τ = 1 i < j p | a i j | ( β τ , i s i j β τ , j ) 2
where s i j = sgn ( r i j ) , a i j = a j i , 1 i , j p . Let L = D A , D be diagonal matrices, D = d i a g ( d 1 , d 2 , , d p ) , d i = j = 1 p | a i j | . Typically, a i j is an edge and d i is a degree, indicating the connectivity between nodes. The matrix L is associated with the weighted graph ξ = ( V , ε ) , the set of nodes V = { 1 , 2 , , p } , and the set of edges ε = { ( i , j ) : ( i , j ) V × V } [27,28]. As the correlation between X i and X j becomes stronger, the larger a i j is, the more β τ , i s i j β τ , j is compressed. In this paper, the grid constraint is extended to the QR loss function.
arg min β i = 1 n ρ τ ( y i Q y ( τ | x ) ) + β τ Τ L β τ
Based on this, this study proposes a sparse Laplace quantile regression long short-term memory network model and applies the above loss function to the QRLSTM model to further improve the model training efficiency and prediction capability.

2.3. Algorithm Implementation

In this paper, GBDT, XGBoost, LSTM, and RNN models will be introduced as benchmark models to compare the predictive power of CNN-IQRLSTM models. Before performing validation, the hyperparameters of the basic models are first set. All models are trained on the training set, and to reduce the complexity of training and improve the training efficiency of the models, a normalization method is used to normalize all data to a value between ( 0 , 1 ) . Each method is iterated for 200 epochs to ensure convergence of the loss function when the training iterations are stopped and to achieve the best prediction results. Once the prediction is complete, the normalized data is inverted to obtain the true prediction.
The operating system used in this paper is Windows 10. The running software is Python version 3.8.8 and the running tool is Jupyter Notebook. Scikit-learn version is Sklearn 0.24.1. the processor is Intel(R) Core(TM) i7-10700 CPU @ 2.90 GHz 2.90 GHz and the RAM is 16.0 GB. All working environments have no costs and the software has no license fees.
Tourism has slowly become an important part of the local and national economy. How to manage scenic spots scientifically and efficiently is an urgent problem for scenic spot management. Passenger flow prediction is the premise of passenger flow management. Only on the premise of accurate prediction can scenic area management make reasonable allocation of scenic area resources and ensure sustainable development of scenic areas. Taking the famous 5A scenic spot in Jilin Province as an example, this study proposes a prediction method that extends LSTM into improved QR, effectively fuses the loss functions of CNN and IQRLSTM, fits the model to the training set with the best parameters, and evaluates it on the test set. The prediction results of the model proposed in this paper and different neural network methods are compared, and the performance of the model is evaluated comprehensively by quantitative analysis of the evaluation metrics of point prediction and interval prediction to find the best prediction model.
The general framework diagram of this paper is shown in Figure 4. This study was divided into three stages. The first stage includes data preprocessing. The left side of the first frame in Figure 4 shows the four datasets, and the right side shows some input data from the puppet palace museum, and we use the added data features and the features generated by the sliding window as model inputs. The second stage is model construction. The left side of the second frame in Figure 4 shows the improved quantile loss function, and the right side shows the flow of the CNN-IQRLSTM model. The third stage is model evaluation. The left side of the third frame of Figure 4 lists the comparison methods for point prediction and interval prediction, respectively, corresponding to the 5 plus 2 evaluation metrics on the right side.

2.4. Evaluation Metric

2.4.1. Evaluation Metric of Point Prediction

When we evaluate the prediction model, the main measure is the difference between the predicted value y ^ i and the observed value y i . In this paper, five important metrics, mean absolute error (MAE), root mean square error (RMSE), mean absolute percentage error (MAPE), symmetric mean absolute percentage error (SMAPE), and coefficient of determination ( R 2 ), are used to evaluate the point prediction model. Meanwhile, three important indicators, interval coverage (PICP), interval width (WS), and coverage width criterion (MC), are used to evaluate the interval prediction model [29].
The specific formulae for MAE, RMSE, RMSE, SMAPE, and R 2 are shown below.
M A E = 1 n i = 1 n | y i y ^ i |
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
M A P E = 1 n i = 1 n | y i y ^ i y i |
S M A P E = 1 n i = 1 n | y i y ^ i | ( | y ^ i | + | y i | ) / 2
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
where y ¯ denotes the average of the actual values.

2.4.2. Evaluation Metric of Interval Prediction

The performance of interval prediction is usually evaluated using two metrics. The first one is called the prediction interval coverage probability (PICP) [30] and is expressed as Equation (24).
P I C P = 1 n i = 1 n c i
where n is the number of samples. c i is the number of samples whose observations fall within the prediction interval (PI). If the actual observation of the i th sample falls within the prediction interval, then c i = 1 . Otherwise, c i = 0 .
Sharpness is measured by the width of PI and reflects the concentration of the predicted distribution. Its mathematical expression is Equation (25).
W S = 1 n R i = 1 n ( U i ( x ) L i ( x ) )
where R is the difference between the maximum and minimum values of the sample. U i ( x ) and L i ( x ) are the upper and lower interval boundaries of the i th sample, respectively.
The prediction interval was constructed to have a high PICP and a low WS, so the composite indicator of the interval prediction was defined as MC [29]. The smaller the MC, the more appropriate the prediction interval. MC is defined as Equation (26).
M C = W S P I C P

3. Results

In this study, the above methods are applied to multiple datasets to predict scenic passenger flow. To avoid the overfitting phenomenon, the dataset is divided into two parts: 90% of the data for model training and 10% for evaluating GBDT [31], XGBoost [32], LightGBM [33], RNN, LSTM, Gated Recurrent Unit (GRU) [34], Feedforward Neural Network (FNN) [35], QRGBDT, QRXGBoost, QRLightGBM, QRRNN, QRLSTM, QRGRU, QRFNN, CNN-LSTM, CNN-QRLSTM, and CNN-IQRLSTM models for prediction accuracy.
This study did not include the year 2022 for passenger flow forecasting, and as can be seen from the time series graphs of each scenic area, the passenger flow data is showing a certain periodicity. Therefore, the thesis framework is still applicable to the year 2022 today, and even to the forecast of passenger traffic in different regions. The processing of the dataset can still be undertaken in the same way as the current one. Meanwhile, existing model parameters can also be used.

3.1. Point Prediction Results Evaluation

The point prediction result evaluation is used to verify the prediction accuracy of CNN-IQRLSTM and to compare seven original tree models and neural network methods, as well as eight methods that incorporate QR. The point prediction evaluation metrics of the proposed method and the other seven methods on the four datasets are shown in Table 2. The point prediction evaluation metrics with QR methods are shown in Table 3. The best metrics in each dataset are highlighted in black font. Three results can be revealed from Table 2 and Table 3 as follows.
(1) In the Changbai Mountain dataset in Table 2, the MAPEs of GBDT, XGBoost, LightGBM, RNN, LSTM, GRU, FNN, and CNN-LSTM are 0.07, 0.08, 0.08, 0.12, 0.12, 0.13, 0.09, and 0.08, respectively. Additionally, the MAPE of the CNN-IQRLSTM model proposed in this paper is 0.07. Although the MAPE is the same as the result of the GBDT method, the remaining four evaluation metrics show that the prediction accuracy of this paper’s method is significantly higher than that of the seven methods, which has great superiority. Second, the original tree model shows better performance than the neural network method. The MAPE and R 2 of the tree model on the Changbai Mountain data and the Sculpture Park data were able to match the proposed model. The same conclusion can be obtained by comparing the point prediction evaluation metrics for the remaining three datasets.
(2) A further comparison of QRGBDT, QRXGBoost, QRLightGBM, QRRNN, QRLSTM, QRGRU, QRFNN, CNN-QRLSTM, and CNN-IQRLSTM was undertaken. The results of SMAPE in the net moon lake dataset are 6.89, 5.38, 6.09, 11.65, 21.59, 21.53, 13.00, 6.47, and 4.83, respectively. The results for R 2 were 0.92, 0.96, 0.90, 0.75, 0.36, 0.37, 0.88, 0.97, and 0.99, respectively. The results show that the prediction performance of the original four neural network methods is far inferior to the proposed method, especially the R 2 of QRRNN and QRGRU is only 0.36 and 0.37. The proposed method is optimal in all evaluation metrics of the four datasets. In particular, the R 2 of the method in the net moon lake data reached 0.99. In this study, the results of RMSE are as high as several hundreds or even thousands, the reason is that different attractions show different bases of passenger flow and tourist passenger flow itself is a larger dataset, especially attractions with characteristics that increase to tens of thousands of passengers per day. Therefore, the results of RMSE are reasonable. By comparing the point prediction metrics, CNN-IQRLSTM still has the highest prediction accuracy in the Changbai Mountain dataset, the puppet palace museum dataset, and the Sculpture Park dataset, which shows that the method is a competitive passenger flow prediction method in terms of prediction accuracy.
(3) Combined with Table 2 and Table 3, the method with QR can achieve better forecasting performance. The conditional distribution of the dependent variable can be estimated using QR, which extends the existing passenger flow forecasting system for probabilistic forecasting. The method proposed in this study can handle complex problems with time-series and nonlinear points capability. Additionally, the LSTM module incorporates dropout to reduce the probability of overfitting. Table 4 shows the comparison results of the point prediction difference between the method with QR and the method without QR. The values of MAE, RMSE, MAPE, and SMAPE are less than or equal to 0, which means that the method with QR performs better, otherwise the method in Table 2 performs better. The opposite is true for R 2 . A value greater than or equal to 0 means that the method with QR performs better, otherwise the method in Table 2 is better. MAE, MAPE, and SMAPE verify the conclusion in most of the methods. Although some negative values were obtained for R 2 , most of them differed within 10%. For the indicator RMSE, it measures the standard deviation of the residuals and is more influenced by outliers. We can see in the time series plot of the four datasets that the passenger flow is very high on one day, which may be one of the reasons for the increase in RMSE. Five methods showed excellent performance in the Sculpture Park data, while the rest of the dataset needs to be adjusted with better parameters to increase the model performance. In addition, Table 5 shows the comparison results of CNN-LSTM, CNN-QRLSTM, and CNN-IQRLSTM methods. The results show that the CNN-IQRLSTM method outperforms the CNN-LSTM method in all five metrics including RMSE. Except for the puppet palace museum dataset where the RMSE of the CNN-QRLSTM method is higher than the RMSE of the CNN-LSTM method, all the other metrics show that the CNN-QRLSTM method outperforms the CNN-LSTM, further validating the effectiveness of the QR method.
In conclusion, CNN-IQRLSTM can capture more distribution information without reducing the prediction accuracy. The same conclusion can be drawn from the comparison graph of point prediction results, as shown in Figure 5. The green line in the figure indicates the predicted value and the orange line indicates the true value. From the figure, it can be seen that the data of net moon lake has a high degree of fit, which basically satisfies the travel demand. However, the QRGBDT, QRXGBoost, QRLightGBM, QRRNN, QRLSTM, QRGRU, QRFNN, CNN-QRLSTM, and CNN-IQRLSTM algorithms are all inadequate in predicting the number of passengers during peak hours. The prediction of peak hour passenger numbers is generally low in Changbai Mountain data and the puppet palace museum data. The predicted values for the rest of the datasets also have the same oscillation trend as the true values, and the prediction models have a relatively excellent fit for both the trained and new samples. The comparison plots of the true and predicted values of the other compared methods for the Changbai Mountain dataset are shown in Figure 6. A plot of the true versus predicted values for the other comparison methods for the puppet palace museum dataset is shown in Figure 7. Point prediction comparison plots for other comparison methods for the Sculpture Park and net moon lake datasets are shown in Appendix A and Appendix B. (a) plot of prediction curves for the GBDT, XGBoost, and LightGBM methods; (b) plot of prediction curves for the RNN, LSTM, and GRU methods; (c) plot of prediction curves for the FNN, CNN-LSTM, and CNN-QRLSTM methods; (d) plot of QRGBDT, QRXGBoost, and QRLightGBM methods prediction curves for the QRRNN, QRLSTM, QRGRU, and QRFNN methods; (e) plotted prediction curves for the QRRNN, QRLSTM, QRGRU, and QRFNN methods. The fluctuations of RNN, LSTM, GRU, and the methods with QR added are larger than the rest of the methods, but they are also in line with the general trend. The tree model approach is relatively smooth.
There may be the following reasons for the poor prediction results of peak data. For deep learning models, the more data in the training set, the better the effect. In this paper, due to the limited data obtained by objective conditions, the fitting effect of the model needs to be further improved. Secondly, the later period of the prediction data selection period may be affected by the new crown epidemic.

3.2. Interval Prediction Results Evaluation

The interval prediction results are evaluated to verify the interval coverage probability and the average width of the interval, and thus to determine whether the time interval is appropriate. In this study, 90% and 95% interval prediction results are provided, and Table 6 and Figure 8 show the different interval prediction evaluation metrics for each method on the four datasets. The best metrics are highlighted in black font for each dataset. Two results were analyzed from Table 6 and Figure 8 as follows.
(1) Taking Changbai Mountain data as an example, the PICP, WS, and MC of the CNN-IQRLSTM model are 0.98, 0.37, and 0.38, respectively. Compared with the CNN-QRLSTM model before improvement, the PICP, WS, and MC of the CNN-IQRLSTM model are improved by 1%, 2%, and 3%, respectively. Compared with the PICP of the QRRNN model, the improvement is 18%. Compared with the WS and MC of the QRLSTM model, the improvement is 19% and 29%, respectively. The scalability and practicality of the CNN-QRLSTM method can be seen. In the puppet palace museum dataset, the CNN-IQRLSTM model does not perform optimally in all evaluation metrics, with PICP and WS just 3% and 5% lower. Similarly, the above conclusions can be obtained for the remaining two datasets. Although the WS of the QRFNN method in the net moon lake data is the optimal result, the PICP is only 0.58, indicating that the smaller interval width leads to a lower interval coverage. From these tables, it can be concluded that the prediction results of QRRNN, QRLSTM, QRGRU, and QRFNN algorithms are more volatile than the amount of scenic tourist traffic, and there is a large gap between the predicted trends. By observing the comprehensive index MC, all four datasets perform optimally on the model proposed in this paper, which can better describe the changing characteristics of scenic passenger flow, and the comparison results validate the superiority of the scenic passenger flow prediction model combining QR and machine learning methods.
(2) Figure 8 represents the 95% prediction interval performance histogram of different models, with indicator PICP in green, indicator WS in orange, and indicator MC in blue. Specific numerical results are shown in Appendix C. The coverage probability of the nine methods on the four datasets is close to 99% as seen in the figure, indicating that the interval prediction of the nine methods is reasonable. In the Changbai Mountain dataset, CNN-IQRLSTM has the highest PICP value of 0.9913, and the smallest WS and MC values of 0.4886 and 0.4929, respectively, which shows that the prediction ability of this model is significantly better than other models. The worst performer was the QRRNN model, which ranked last with the smallest PICP and highest WS and MC. The second-ranked model is CNN-QRLSTM without improvement, which indicates that the deep learning framework using CNN for feature filtering combined with LSTM prediction has important application value for scenic spots passenger flow. For the puppet palace museum dataset, the PICP, WS, and MC of CNN-IQRLSTM model are 0.9661, 0.408, and 0.4223, respectively. WS and MC perform best among all methods. The PICP of the QRLightGBM model is the best with 0.9813. As a witness of Changchun’s history, the pseudo-Manchu architecture is an artistic symbol that can be appreciated. Therefore, it is important to accurately forecast its patronage. In the Sculpture Park dataset, the PICP, WS, and MC of CNN-IQRLSTM model are 0.9914, 0.5187, and 0.5233, respectively. The PICP, WS, and MC of the CNN-QRLSTM model are 1, 0.5815, and 0.5815, respectively. Although the interval coverage of the CNN-QRLSTM model is 100%, its interval width is also 6.28% higher than that of the CNN-IQRLSTM model. Except for the QRRNN, QRLSTM, and QRGUR models, the rest of the MCs are between 0.5 and 0.6, reflecting the advantage of the tree model in this type of dataset. In the net moon lake dataset, the PICP, WS, and MC of the CNN-IQRLSTM model are 0.9915, 0.4841, and 0.4883, respectively. The PICP, WS, and MC of the QRFNN model are 0.6496, 0.4197, and 0.6461, respectively.
In summary, the 95% prediction interval will cover more true values than the 90% prediction interval, which makes it probable that our estimated interval will contain true values in the future actual prediction. It can greatly improve the accuracy of scenic spots’ passenger flow prediction and achieve better prediction results regardless of whether in the peak tourist season or holidays. However, only to predict a larger interval has no practical significance, we need to make the prediction interval smaller under the condition of guaranteeing the accuracy rate, prompting the scenic spots management to make a reasonable allocation of scenic spots to ensure the sustainable development of scenic resources.
Figure 9 and Figure 10 show plots of the 90% and 95% prediction interval results for the four datasets. The 90% prediction interval is composed of the quantile predicted by the 0.95 quantile and the quantile predicted by the 0.05 quantile. The 95% prediction interval is composed of the quantile predicted by the 0.975 quantile and the quantile predicted by the 0.025 quantile. The 90% prediction interval results for the GBDT, XGBoost, and LightGBM methods are shown in Figure 11. The 95% prediction interval results for GBDT, XGBoost, and LightGBM methods are plotted in Appendix D.
Figure 9 shows the 90% PIs, which were obtained through the proposed methodology from 17 July to 31 December 2021, with a forecast time resolution of daily. The actual tourist flow curves are also plotted in Figure 9. It can be seen that the PIs obtained using the proposed model can cover the actual tourist flow curve well, which visually illustrates the effectiveness of the proposed method for probabilistic forecasting of tourist flow in scenic areas. Figure 10 shows the 95% PIs, and it can be clearly seen that the 95% PIs are wider than the 90% PIs. The 95% over 90% PICP and WS for the four datasets improved by 0.0087, 0.1144, 0.0254, 0.0451, 0.0259, 0.1419, 0.0611, and 0.1653, respectively. Although the differences in the evaluation index results of different models are small, the superiority of the CNN-IQRLSTM model can still be seen, which further illustrates that the method proposed in this study is the best scenic tourist flow prediction model.
The training time of the neural network is shown in Table 7. It includes point prediction, 90% interval prediction, and 95% interval prediction. As seen in the results, the CNN-QRLSTM model and the CNN-IQRLSTM model show excellent performance. In the interval prediction, the training time of QRRNN, QRLSTM, and QRGRU is as much as tens of times that of the CNN-IQRLSTM model. In this study, we train all samples at once before updating the parameters and use the gradient descent method for calculation.
For the new dataset, after feature filling and sliding window processing, it can be directly input into the neural network for training. The results of parameters such as hidden layers, neurons, and epochs for different models are shown in Table 8. In the defined CNN-IQRLSTM network, first, all features pass through a 1D convolutional layer with a convolutional window of length 3 and a step size of 1. Second, they pass through a max-pooling layer with a window size of 2 and a step size of 1. Finally, after the LSTM layer, the loss function is set to quantile loss with grid constraints.
The accuracy of the prediction model for the four datasets before feature selection was low and the error was large. Therefore, we added data features considering the dimensionality of the dataset. We found a significant improvement in the accuracy after subjecting the data to the increased features. The simulation experiment results show that the model in this paper is a type of scenic area passenger flow average prediction model with high accuracy and excellent generality, which has wide application prospects.

4. Discussion of the Proportion of Datasets

This section provides a detailed description of the division of the dataset. Since the sample size of the dataset was not very large, this study initially divided the training and test sets into ratios of 8:2 and 9:1, respectively. In the results for the Changbai Mountain data, it was shown that the prediction accuracy of dividing the data at a ratio of 8:2 was lower than the prediction accuracy of dividing the data at a ratio of 9:1. Therefore, it was decided to choose a ratio of 9:1 for data partitioning in this paper. Table 9 in the paper shows the point prediction results for dividing the data at 8:2 and dividing the data at 9:1. Regardless of which of the four indicators MAE, RMSE, MAPE, and SMAPE was used, the result was the lowest error in dividing the dataset by 9:1.
Using 90% as a training set and 10% as a test set increases the risk of overfitting, but this is not absolute. Figure 12 in the paper shows the loss of the training and test sets for point prediction using the CNN-IQRLSTM method on the Changbai Mountain data, and the results show that no overfitting occurs using 90% of the data as the training set.

5. Discussion and Conclusions

With the continuous reform and opening up and the rapid development of the national economy, the economic ability and living standard of Chinese people have been improving. Nowadays, more and more Chinese people are focusing on better quality of life and higher level of spiritual pursuit [11]. Tourism around the world has not only changed the way information is disseminated but also profoundly influenced people’s life, work, and entertainment, changing the way people receive information and their way of thinking, and people have higher requirements for information [36]. The modeling and prediction of scenic area passenger flow can help scenic area managers understand the changing dynamics of scenic area passenger flow, develop more reasonable management measures, and improve scenic area management. Therefore, scenic spots passenger flow prediction has become a hot issue in the field of economic research. With the booming development of tourism, how to bring good benefits to tourist attractions has become a topic worth thinking about. Choosing the appropriate method to process the type features plays an important role in improving the universality and prediction accuracy of the prediction model.
Accurate passenger flow prediction is essential to ensure the proper operation of scenic spots. However, a single model does not effectively capture the characteristics of the data. Daily tourist flow data has strong non-linear characteristics and needs to be accurately predicted. Accordingly, this study took Changbai Mountain, the puppet palace museum, the Sculpture Park, and the net moon lake as the research objects to realize the daily passenger flow prediction of scenic spots in Jilin Province. Aiming at the drawbacks of the current poor scenic spots passenger flow prediction and improving the scenic spots passenger flow prediction results, this paper constructed a deep network framework consisting of a convolutional layer, pooling layer, and improved QRLSTM to extend LSTM to QR. A sparse Laplacian smoothing grid constraint was proposed as a penalty term for quantile loss to improve the robustness of the prediction. Four relevant data features were added for the date attribute and combined with the sliding window extracted features as the input data to obtain more information about the passenger flow and provide a new perspective and idea for the accurate prediction of tourism passenger flow.
Unlike previous passenger flow forecasting, this paper creatively proposed a nonparametric probabilistic forecasting method for scenic tourist flow forecasting. The QR model was applied to the scenic tourist passenger flow prediction, and the CNN-IQRLSTM model was built on this basis to improve the prediction performance. Based on the actual data of four 5A scenic spots in Jilin Province, this paper used GBDT, XGBoost, LightGBM, RNN, LSTM, GRU, FNN, QRGBDT, QRXGBoost, QRLightGBM, QRRNN, QRLSTM, QRGRU, QRFNN, and CNN-QRLSTM as the benchmark to calculate 120 days before the passenger flow forecast and analyze MAE, RMSE, MAPE, SMAPE, PICP, WS, and MC, respectively, to verify the performance of the CNN-IQRLSTM model. The CNN-IQRLSTM model is the most accurate, reliable, and sharpest model among the 15 models, proving the potential of the method to be widely used in practice. The research framework in this paper is not proposed only for Jilin Province passenger flow and can be validated on cross-datasets. The method proposed in this study is applicable for any dataset framework. When applied to other domain datasets, we need to adjust the corresponding parameters to make the prediction accuracy ideal.
In the point prediction results, the methods with QR and without QR were compared separately. First, in the method without QR, the MAPE of the CNN-IQRLSTM model for Changbai Mountain data is 0.07. Although the MAPE is the same as the result of the GBDT method, the remaining four evaluation metrics show that the prediction accuracy of this paper’s method is significantly higher than that of the seven methods, which is vastly superior. Secondly, the original tree model shows better performance than the neural network method. Secondly, among the methods with QR, CNN-IQRLSTM still has the highest prediction accuracy in the Changbai Mountain dataset, the puppet palace museum dataset, and the Sculpture Park dataset. Finally, methods with QR achieve better prediction performance, and MAE, MAPE, and SMAPE validate this conclusion in most methods. In the interval prediction results, by observing the composite index MC, all four datasets perform optimally on the model proposed in this paper, which can better characterize the change in scenic traffic, and the coverage probability of the nine methods on the four datasets is close to 99%. Meanwhile, the 95% prediction interval will cover more true values than the 90% prediction interval, and the 95% PI of the four datasets improve the PICP and WS by 0.0087, 0.1144, 0.0254, 0.0451, 0.0259, 0.1419, 0.0611, and 0.1653, respectively, compared with the 90% PI. In a subsequent study, we will try to use GPUs supporting Computational Unified Device Architecture (CUDA) to achieve accelerated training times. CUDA is a novel hardware and software architecture for issuing and administering computations on GPUs as data parallel computing devices on GPUs without mapping them to image APIs.
This study combined QR and neural network techniques to provide a new framework to enhance the analysis and prediction of tourist passenger flow in different scenic spots by fusing multidimensional features. Compared with previous methods, the advantages of the proposed CNN-IQRLSTM model are as follows. (1) Compared with the traditional QR model, the QR prediction model constructed based on CNN and LSTM can better capture the dynamic features of passenger flow changes and obtain higher prediction accuracy. (2) CNN-IQRLSTM can predict multiple quantile prediction results at the same time, and it is trained with appropriate parameters during the training process, which significantly improves the prediction efficiency while ensuring the prediction effect. (3) Compared with the CNN-QRLSTM model, the addition of grid constraints effectively avoids the crossover between quantile prediction values, and the sparse Laplacian smoothing grid constraint makes the prediction results more reasonable and significantly improves the reliability of point prediction and interval prediction. In summary, the method in this paper not only has high accuracy of point prediction, but also can obtain reasonable interval prediction results, which can provide more accurate and rich information for scientific decision-making of scenic spots managers.
Since the neural network will be affected by abnormal historical disturbance during the training process, the generalization of the model is reduced. In this paper, we adopted the network structure constraint as the penalty term of the objective function to improve the robustness of prediction by smoothing the deviation degree of network weights in the iterative correction process according to the sparse Laplacian. However, there are still some limitations. First, the validation dataset does not have a long enough period to study the effects of longer spatial and temporal factors, such as year, season, etc. Second, the effects of external factors, such as financial crisis, climate, etc., were not taken into account. Third, the dynamic spatial characteristics of passenger flow were not studied in this paper, and the rich spatio-temporal correlation and non-linear information would improve the validity of the prediction. Finally, the model running time is improved over some methods, but not significantly and substantially. In the future, we will further explore the influence of external and spatial features and investigate the application of CNN-IQRLSTM in more complex datasets. This study also intends to investigate the extension of composite quantile regression in deep learning models to flexibly exploit the properties of neural networks to explore nonlinear relationships between variables, and multiple regression quantile features can be exploited to improve estimation efficiency and predictive power. Finally, a framework of swarm intelligence optimization algorithms can be established for the parameter problems in neural networks to perform the optimization search process, aiming to develop a more effective and accurate and reliable prediction model. It has excellent prospects for tourism management research and application and can promote the healthy and sustainable development of the tourism industry.
The factors considered in this paper are not representative of all influencing factors and do not take into account unexpected events, financial crises, economic collapse, energy crises, or unknown causes. Therefore, if other influencing factors can be explored and analyzed, the forecasting model can be enhanced and the forecasting accuracy can be further improved. For potential anomalies, such as coronavirus, there are numerous impacts on the tourism industry. We can divide the passenger flow forecast for tourist attractions in Jilin province into pre- and post-epidemic, considering the actual situation of each attraction, as well as the availability of travel in each province. Next, future passenger flow can be further predicted based on the duration of the virus that has been predicted.

Author Contributions

Conceptualization, Xiwen Qin, Xiaogang Dong and Dongmei Yin; methodology, Dongmei Yin; software, Dongxue Chen; validation, Dongmei Yin, Dongxue Chen and Shuang Zhang; formal analysis, Dongmei Yin; investigation, Xiwen Qin and Xiaogang Dong; resources, Xinwen Qi; data curation, Dongmei Yin; writing—original draft preparation, Dongmei Yin; writing—review and editing, Dongmei Yin; visualization, Dongxue Chen and Shuang Zhang; supervision, Xiwen Qin and Xiaogang Dong; project administration, Xiwen Qin; funding acquisition, Xiwen Qin. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Science and Technology Department Project of Jilin Province (Grant No. 20200403182SF), the National Natural Science Foundation of China (Grant No. 12026430), and Education Department Project of Jilin Province (Grant No. JJKH20210716KJ).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not publicly available due to research data ownership issues.

Acknowledgments

Thanks to Jilin Provincial Tourism Information Center for providing data for this study. The authors thank the editors for their comments and suggestions.

Conflicts of Interest

No conflict of interest exists in the submission of this manuscript.

Appendix A

Figure A1. Comparison of true and predicted values of other comparison methods for the Sculpture Park dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Figure A1. Comparison of true and predicted values of other comparison methods for the Sculpture Park dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Ijgi 11 00509 g0a1

Appendix B

Figure A2. Comparison of true and predicted values of other comparison methods for the net moon lake dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Figure A2. Comparison of true and predicted values of other comparison methods for the net moon lake dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Ijgi 11 00509 g0a2aIjgi 11 00509 g0a2b

Appendix C

Table A1. Comparison of 95% prediction interval results.
Table A1. Comparison of 95% prediction interval results.
Scenic SpotsMetricQRGBDTQRXGBoostQRLightGBMQRRNNQRLSTMQRGRUQRFNNCNN-QRLSTMCNN-IQRLSTM
Changbai MountainPICP0.950.950.960.870.970.930.880.980.99
WS0.701.150.782.720.810.820.620.580.49
MC0.741.210.823.130.840.880.710.590.49
The puppet palace museumPICP0.960.970.980.960.950.920.830.970.97
WS0.690.640.801.701.330.670.460.610.41
MC0.720.660.821.771.400.730.560.630.42
Sculpture ParkPICP0.960.970.980.870.980.980.921.000.99
WS0.550.520.581.591.221.280.530.580.52
MC0.570.540.591.821.251.300.570.580.52
Net moon lakePICP0.910.930.960.950.960.970.650.940.99
WS0.711.000.762.741.421.680.420.560.48
MC0.781.070.792.891.491.740.650.600.49
Note: boldface indicates the best result for each indicator.

Appendix D

Figure A3. Plot of 90% interval prediction results for GBDT, XGBoost, and LightGBM methods.
Figure A3. Plot of 90% interval prediction results for GBDT, XGBoost, and LightGBM methods.
Ijgi 11 00509 g0a3

References

  1. Fan, X.H. The Evaluation on Tourism Climate Comfort in Chongqing; Southwest University: Chongqing, China, 2017. [Google Scholar]
  2. Lee, H.; Lee, J.; Chung, N.; Koo, C. Tourists’ happiness: Are there smart tourism technology effects? Asia Pac. J. Tour. Res. 2018, 23, 486–501. [Google Scholar] [CrossRef]
  3. Chen, N.; Liang, Y. A Tourist Flow Prediction Model for Scenic Areas Based on Particle Swarm Optimization of Neural Network. Rev. D’Intell. Artif. 2020, 34, 395–402. [Google Scholar] [CrossRef]
  4. Lin, V.S.; Liu, A.; Song, H. Modeling and Forecasting Chinese Outbound Tourism: An Econometric Approach. J. Travel Tour. Mark. 2015, 32, 34–49. [Google Scholar] [CrossRef]
  5. Gustavsson, P.; Jonas, N. The Impact of Seasonal Unit Roots and Vector ARMA Modelling on Forecasting Monthly Tourism Flows. Tour. Econ. 2001, 7, 117–133. [Google Scholar] [CrossRef]
  6. Chen, C.; Lai, M.; Yeh, C. Forecasting tourism demand based on empirical mode decomposition and neural network. Knowl. Based Syst. 2012, 26, 281–287. [Google Scholar] [CrossRef]
  7. Lim, K.H.; Chan, J.; Karunasekera, S.; Leckie, C. Tour recommendation and trip planning using location-based social media: A survey. Knowl. Inf. Syst. 2019, 60, 1247–1275. [Google Scholar] [CrossRef]
  8. Niu, K.; Long, H.Y.; Yu, X.T.; Liu, M.Y. Railway Passenger Flow Forecast and Ticket Allocation Optimization Based on Time Series. Sci. Technol. Eng. 2020, 20, 9937–9942. [Google Scholar]
  9. Li, D. Study on the Prediction Model of Tourist Flow in Scenic Spots Based on Big Data Analysis. Microcomput. Appl. 2021, 37, 117–119. [Google Scholar] [CrossRef]
  10. Cui, H.R.; Yang, X.X.; Yu, Y.L. Prediction of Tourists Flow Based on EMD-GRU Model: A Case Study of Black Valley Scenic Area in Chongqing. J. China West Norm. Univ. (Nat. Sci.) 2022, 43, 1–9. [Google Scholar]
  11. Lu, W.; Rui, H.; Liang, C.; Jiang, L.; Zhao, S.; Li, K. A Method Based on GA-CNN-LSTM for Daily Tourist Flow Prediction at Scenic Spots. Entropy 2020, 22, 261. [Google Scholar] [CrossRef]
  12. Xu, J. Design of a Cultural Tourism Passenger Flow Prediction Model in the Yangtze River Delta Based on Regression Analysis. Sci. Program. 2021, 2021, 9913468. [Google Scholar] [CrossRef]
  13. Chen, X.; Cong, D.T. Application of Improved Algorithm Based on Four-Dimensional ResNet in Rural Tourism Passenger Flow Prediction. J. Sens. 2022, 2022, 9675647. [Google Scholar] [CrossRef]
  14. Qin, X.W.; Yin, D.M.; Dong, X.G.; Chen, D.X.; Zhang, S. Survival prediction model for right-censored data based on improved composite quantile regression neural network. Math. Biosci. Eng. 2022, 19, 7521–7542. [Google Scholar] [CrossRef]
  15. Koenker, R.; Bassett, G.W. Regression Quantiles. Econometrica 1978, 46, 33–50. [Google Scholar] [CrossRef]
  16. Zou, H.; Yuan, M. Composite Quantile Regression and the Oracle Model Selection Theory. Ann. Stat. 2008, 36, 1108–1126. [Google Scholar] [CrossRef]
  17. Rodrigues, F.; Pereira, F.C. Beyond Expectation: Deep Joint Mean and Quantile Regression for Spatiotemporal Problems. IEEE Trans. Neural Netw. Learn. Syst. 2020, 12, 5377–5389. [Google Scholar] [CrossRef]
  18. Yu, Y.; Han, X.; Yang, M.; Yang, J. Probabilistic Prediction of Regional Wind Power Based on Spatiotemporal Quantile Regression. IEEE Trans. Ind. Appl. 2020, 56, 6117–6127. [Google Scholar] [CrossRef]
  19. Ni, T.; Wang, L.; Zhang, P.; Wang, B.; Li, W. Daily tourist flow forecasting using SPCA and CNN-LSTM neural network. Concurr. Comput. Pract. Exp. 2021, 33, e5980. [Google Scholar] [CrossRef]
  20. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  21. Yu, Y.; Wang, M.; Yan, F.; Yang, M.; Yang, J. Improved convolutional neural network-based quantile regression for regional photovoltaic generation probabilistic forecast. IET Renew. Power Gener. 2020, 14, 2712–2719. [Google Scholar] [CrossRef]
  22. Li, D.; Zhang, Y.H.; Yang, B.H.; Wang, Q. Short Time Power Load Probabilistic Forecasting Based on Constrained Parallel-LSTM Neural Network Quantile Regression Mode. Power Syst. Technol. 2021, 45, 1356–1363. [Google Scholar] [CrossRef]
  23. Dai, J.J.; Song, H.; Sheng, G.H. Study on the operation state prediction method of power transformer based on LSTM network. High Volt. Eng. 2018, 44, 1099–1106. [Google Scholar] [CrossRef]
  24. Jia, Y.; Jeong, J.H. Deep learning for quantile regression under right censoring: DeepQuantreg. Comput. Stat. Data Anal. 2022, 165, 107323. [Google Scholar] [CrossRef]
  25. Qin, X.W.; Sheng, H.; Dong, X.G. Interval Wind-Speed Forecasting Model Based on Quantile Regression Bidirectional Minimal Gated Memory Network and Kernel Density Estimation. Arab. J. Sci. Eng. 2022, 47, 1–15. [Google Scholar] [CrossRef]
  26. Yi, S.Z.; Liu, Y.K.; Yang, F.; Zhang, J.W.; Jia, H.J.; Peng, X.G. Short-term Load Probability Forecasting Based on Improved Quantile Regression of Gated Recurrent Unit. Smart Power 2021, 49, 84–90. [Google Scholar]
  27. Huang, J.; Ma, S.; Li, H.; Zhang, C.H. The Sparse Laplacian Shrinkage Estimator for High-Dimensional Regression. Ann. Stat. 2011, 39, 2021–2046. [Google Scholar] [CrossRef] [PubMed]
  28. Deng, C.; He, X.; Hu, Y. Learning a Spatially Smooth Subspace for Face Recognition. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–7. [Google Scholar] [CrossRef]
  29. Zhang, Z.D.; Qin, H.; Liu, Y.Q.; Yao, L.Q.; Yu, X.; Lu, J.T.; Jiang, Z.Q.; Feng, Z.K. Wind speed forecasting based on quantile regression minimal gated memory network and kernel density estimation. Energy Convers. Manag. 2019, 196, 1395–1409. [Google Scholar] [CrossRef]
  30. Khosravi, A.; Nahav, S.; Creighton, D. A prediction interval based approach to determine optimal structures of neural network metamodels. Expert Syst. Appl. 2010, 37, 2377–2387. [Google Scholar] [CrossRef]
  31. Zhou, L. GBDT-SVM Credit Risk Assessment Model and Empirical Analysis of Peer-to-Peer Borrowers under Consideration of Audit Information. Open J. Bus. Manag. 2018, 6, 362–372. [Google Scholar] [CrossRef]
  32. Chen, T.Q.; Carlos, G. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  33. Qi, M. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 3149–3157. [Google Scholar]
  34. Kyunghyun, C.; Bart, V.M.; Caglar, G.; Dzmitry, B.; Fethi, B.; Holger, S.; Yoshua, B. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar] [CrossRef]
  35. Zhang, W.; Du, T.; Wang, J. Deep Learning over Multi-Field Categorical Data-A Case Study on User Response Prediction; ECIR: Welcome, NC, USA, 2016. [Google Scholar] [CrossRef]
  36. Wang, X.R.; Wang, W.L.; Zhang, W.L. Digital Dissemination of Scene Art in Changbai Mountain Area of Visual Sensor Images. Wirel. Commun. Mob. Comput. 2022, 2, 2468799. [Google Scholar] [CrossRef]
Figure 1. Time-series diagram of tourist passenger flow in four scenic spots.
Figure 1. Time-series diagram of tourist passenger flow in four scenic spots.
Ijgi 11 00509 g001
Figure 2. Schematic diagram of the CNN model structure.
Figure 2. Schematic diagram of the CNN model structure.
Ijgi 11 00509 g002
Figure 3. Schematic diagram of the LSTM model structure.
Figure 3. Schematic diagram of the LSTM model structure.
Ijgi 11 00509 g003
Figure 4. Framework diagram of the proposed CNN-IQRLSTM model.
Figure 4. Framework diagram of the proposed CNN-IQRLSTM model.
Ijgi 11 00509 g004
Figure 5. Comparison of the actual and predicted values of the four datasets. (a) Actual and predicted values of Changbai Mountain data; (b) actual and predicted values of the puppet palace museum data; (c) actual and predicted values of Sculpture Park data; (d) actual and predicted values of net moon lake data.
Figure 5. Comparison of the actual and predicted values of the four datasets. (a) Actual and predicted values of Changbai Mountain data; (b) actual and predicted values of the puppet palace museum data; (c) actual and predicted values of Sculpture Park data; (d) actual and predicted values of net moon lake data.
Ijgi 11 00509 g005aIjgi 11 00509 g005b
Figure 6. Comparison of true and predicted values of other comparison methods for the Changbai Mountain dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Figure 6. Comparison of true and predicted values of other comparison methods for the Changbai Mountain dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Ijgi 11 00509 g006aIjgi 11 00509 g006b
Figure 7. Comparison of true and predicted values of other comparison methods for the puppet palace museum dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Figure 7. Comparison of true and predicted values of other comparison methods for the puppet palace museum dataset. (a) GBDT/XGBoost/LightGBM; (b) RNN/LSTM/GRU; (c) FNN/CNN-LSTM/CNN-QRLSTM; (d) QRGBDT/QRXGBoost/QRLightGBM; (e) QRRNN/QRLSTM/QRGRU/QRFNN.
Ijgi 11 00509 g007aIjgi 11 00509 g007b
Figure 8. Performance results of different models for the 95% prediction interval.
Figure 8. Performance results of different models for the 95% prediction interval.
Ijgi 11 00509 g008
Figure 9. The 90% prediction intervals for four datasets. (a) PI of Changbai Mountain data; (b) PI of the puppet palace museum data; (c) PI of Sculpture Park data; (d) PI of net moon lake data.
Figure 9. The 90% prediction intervals for four datasets. (a) PI of Changbai Mountain data; (b) PI of the puppet palace museum data; (c) PI of Sculpture Park data; (d) PI of net moon lake data.
Ijgi 11 00509 g009
Figure 10. The 95% prediction intervals for four datasets. (a) PI of Changbai Mountain data; (b) PI of the puppet palace museum data; (c) PI of Sculpture Park data; (d) PI of net moon lake data.
Figure 10. The 95% prediction intervals for four datasets. (a) PI of Changbai Mountain data; (b) PI of the puppet palace museum data; (c) PI of Sculpture Park data; (d) PI of net moon lake data.
Ijgi 11 00509 g010
Figure 11. Plot of the 90% interval prediction results for GBDT, XGBoost, and LightGBM methods.
Figure 11. Plot of the 90% interval prediction results for GBDT, XGBoost, and LightGBM methods.
Ijgi 11 00509 g011
Figure 12. Training set and test set losses for point prediction of the Changbai Mountain dataset.
Figure 12. Training set and test set losses for point prediction of the Changbai Mountain dataset.
Ijgi 11 00509 g012
Table 1. Partial input data of Changbai Mountain.
Table 1. Partial input data of Changbai Mountain.
DateFlowWeekdayWeekWeekendHoliday
10 August 201724,2064200
11 August 201724,1695200
12 August 201730,4496210
13 August 201728,5787210
14 August 201724,0911200
15 August 201722,6342300
16 August 201722,4853300
17 August 201723,0824300
18 August 201722,8625300
19 August 201726,8556310
20 August 201718,0847310
21 August 201720,7331300
22 August 201719,3432400
23 August 201718,1613400
24 August 201717,9484400
Table 2. Comparison results of point prediction between the original tree model and the neural network method.
Table 2. Comparison results of point prediction between the original tree model and the neural network method.
Scenic SpotsMetricGBDTXGBoostLightGBMRNNLSTMGRUFNNCNN-LSTM
Changbai MountainMAE889.95934.99958.521224.881432.781454.471076.79973.77
RMSE2034.231943.641963.182105.742608.472502.482114.242029.40
MAPE0.070.080.080.120.120.130.090.08
SMAPE7.438.047.8912.5212.6012.329.328.25
Rsquare0.860.880.870.850.780.790.850.86
The puppet palace museumMAE152.28164.87156.04270.05355.66326.05318.31221.47
RMSE218.33237.84230.12385.64573.94480.37425.60316.54
MAPE0.060.070.060.110.140.130.130.09
SMAPE6.286.765.8910.8913.8812.5312.458.80
Rsquare0.950.940.950.850.670.770.820.09
Sculpture ParkMAE300.79260.36263.34482.05393.04368.62294.36266.24
RMSE443.91358.09364.12624.44486.56512.22392.87362.43
MAPE0.080.070.070.120.110.090.080.07
SMAPE7.807.027.0213.1910.689.617.997.21
Rsquare0.890.930.920.780.860.850.910.92
Net moon lakeMAE668.21688.44785.901362.142337.982593.191461.58727.35
RMSE957.48962.711094.551973.833289.314092.602092.17942.36
MAPE0.070.070.070.130.210.220.140.08
SMAPE6.927.127.4013.2921.3222.9913.558.04
Rsquare0.970.970.960.870.630.430.850.97
Table 3. Comparison results of the point prediction with QR methods.
Table 3. Comparison results of the point prediction with QR methods.
Scenic SpotsMetricQRGBDTQRXG BoostQRLightGBMQRRNNQRLSTMQRGRUQRFNNCNN-QRLSTMCNN-IQRLSTM
Changbai MountainMAE881.51859.14906.921248.901315.311452.751024.65879.72819.19
RMSE2179.632043.632206.632285.872800.543481.842103.182020.031913.09
MAPE0.070.070.070.110.100.100.090.070.07
SMAPE6.806.916.9611.0011.0211.188.997.106.68
Rsquare0.840.860.840.830.740.600.850.870.88
The puppet palace museumMAE153.79192.24179.81244.59323.53361.61308.16209.21117.33
RMSE318.09414.62431.95413.19639.92757.30458.70333.04168.02
MAPE0.060.060.050.090.110.120.120.080.05
SMAPE5.856.575.738.9911.4312.8011.848.055.04
Rsquare0.900.830.820.830.590.430.790.890.97
Sculpture ParkMAE256.10270.87265.50412.51368.69366.88280.02255.59248.89
RMSE371.24398.90377.31551.00484.73481.49370.22360.82354.04
MAPE0.070.070.070.100.100.100.080.070.07
SMAPE6.776.946.9810.839.639.657.706.776.53
Rsquare0.920.910.920.830.870.870.920.930.93
Net moon lakeMAE757.20588.22809.901402.382545.552543.181297.89608.41447.03
RMSE1492.931065.441746.622680.744311.774291.961844.19883.66662.39
MAPE0.070.050.060.110.210.220.130.060.05
SMAPE6.895.386.0911.6521.5921.5313.006.474.83
Rsquare0.920.960.900.750.360.370.880.970.99
Note: boldface indicates the best result for each indicator.
Table 4. Comparison results of point prediction differences with and without QR methods.
Table 4. Comparison results of point prediction differences with and without QR methods.
Scenic SpotsMetricQRGBDT
GBDT
QRXGBoost
XGBoost
QRLightGBM
LightGBM
QRRNN
RNN
QRLSTM
LSTM
QRGRU
GRU
QRFNN
FNN
CNN-LSTM
CNN-QRLSTM
Changbai MountainMAE−8.44−75.85−51.6024.03−117.47−1.71−52.13−94.05
RMSE145.3999.99243.45180.13192.07979.36−11.06−9.37
MAPE−0.01−0.01−0.01−0.01−0.02−0.02−0.01−0.01
SMAPE−0.62−1.14−0.93−1.52−1.58−1.14−0.33−1.15
Rsquare−0.02−0.01−0.03−0.03−0.03−0.190.000.00
The puppet palace museumMAE1.5127.3723.77−25.45−32.1335.57−10.15−12.26
RMSE99.76176.79201.8427.5565.99276.9433.1016.50
MAPE−0.010.000.00−0.02−0.03−0.01−0.01−0.01
SMAPE−0.44−0.19−0.17−1.90−2.450.27−0.61−0.75
Rsquare−0.05−0.11−0.13−0.02−0.08−0.34−0.030.80
Sculpture ParkMAE−44.6810.512.16−69.54−24.35−1.74−14.34−10.65
RMSE−72.6740.8113.19−73.44−1.83−30.73−22.65−1.62
MAPE−0.010.000.00−0.02−0.010.00−0.010.00
SMAPE−1.03−0.08−0.04−2.36−1.050.05−0.29−0.44
Rsquare0.03−0.02−0.010.050.000.020.010.00
Net moon lakeMAE88.99−100.2224.0040.24207.56−50.01−163.69−118.94
RMSE535.44102.73652.07706.911022.46199.36−247.98−58.70
MAPE0.00−0.02−0.01−0.02−0.01−0.010.00−0.02
SMAPE−0.03−1.74−1.32−1.630.26−1.47−0.55−1.57
Rsquare−0.04−0.01−0.06−0.11−0.27−0.060.030.00
Table 5. Comparison results of CNN-LSTM, CNN-QRLSTM, and CNN-IQRLSTM.
Table 5. Comparison results of CNN-LSTM, CNN-QRLSTM, and CNN-IQRLSTM.
Scenic SpotsMetricCNN-LSTMCNN-QRLSTMCNN-IQRLSTM
Changbai MountainMAE973.77879.72819.19
RMSE2029.42020.031913.09
MAPE0.080.070.07
SMAPE8.257.16.68
Rsquare0.860.870.88
The puppet palace museumMAE221.47209.21117.33
RMSE316.54333.04168.02
MAPE0.090.080.05
SMAPE8.88.055.04
Rsquare0.90.890.97
Sculpture ParkMAE266.24255.59248.89
RMSE362.43360.82354.04
MAPE0.070.070.07
SMAPE7.216.776.53
Rsquare0.920.930.93
Net moon lakeMAE727.35608.41447.03
RMSE942.36883.66662.39
MAPE0.080.060.05
SMAPE8.046.474.83
Rsquare0.970.970.99
Note: boldface indicates the best result for each indicator.
Table 6. Comparison of 90% prediction interval results.
Table 6. Comparison of 90% prediction interval results.
Scenic SpotsMetricQRGBDTQRXGBoostQRLightGBMQRRNNQRLSTMQRGRUQRFNNCNN-QRLSTMCNN-IQRLSTM
Changbai MountainPICP0.920.930.940.800.830.910.860.970.98
WS0.410.430.470.400.560.500.480.390.37
MC0.450.460.500.500.670.550.550.410.38
The puppet palace museumPICP0.890.950.880.930.860.890.720.970.94
WS0.430.420.511.580.470.470.310.490.36
MC0.480.440.571.700.550.530.430.500.39
Sculpture ParkPICP0.910.920.950.790.900.970.850.950.97
WS0.430.360.401.230.380.640.400.430.38
MC0.470.390.421.550.430.660.470.460.39
Net moon lakePICP0.870.910.870.880.900.890.580.920.93
WS0.500.600.530.631.061.310.310.560.32
MC0.580.660.610.711.181.470.530.610.34
Note: boldface indicates the best result for each indicator.
Table 7. Neural network training time (s).
Table 7. Neural network training time (s).
Scenic SpotsMetricRNNLSTMGRUCNN-LSTMQRRNNQRLSTMQRGRUCNN-QRLSTMCNN-IQRLSTM
Changbai MountainPoint prediction143.00235.00208.0058.50248.00203.00204.0065.0892.00
90% PI————————530.004679.00601.0045.5742.00
95% PI————————459.002479.002776.0045.5752.40
The puppet palace museumPoint prediction51.10235.00208.00234.00243.00185.00187.0042.4097.00
90% PI————————748.003638.003119.00118.0082.00
95% PI————————775.003536.003882.0045.70373.00
Sculpture ParkPoint prediction52.00201.00314.0087.60280.00236.00199.0051.4016.90
90% PI————————960.104860.80422.0044.10156.00
95% PI————————618.004224.00570.00101.6061.01
Net moon lakePoint prediction42.60168.00153.00275.00227.00227.00202.0043.4096.00
90% PI————————883.004348.00816.0044.80106
95% PI————————767.002920.002682.00150.00118.00
Note: boldface indicates the best result for each indicator.
Table 8. Summary of parameters of neural networks.
Table 8. Summary of parameters of neural networks.
ModelParameterValue
GBDT QRGBDTDefault parameters——
XGBoost QRXGBoostDefault parameters——
LightGBM QRLightGBMDefault parameters——
RNN QRRNNNumber of hidden layer nodes32
LSTM QRLSTMEpochs of training100
GRU QRGRUNumber of hidden layers1
FNN QRFNNNumber of hidden layer nodes128
Activation functionsigmoid
Epochs of training100
Number of hidden layers3
CNN-LSTM CNN-QRLSTM
CNN-IQRLSTM
Number of hidden layer nodes128
Activation functionelu
Epochs of training100
Number of hidden layers2
Dropout0.1
Table 9. Point prediction results for data divided by 8:2 and data divided by 9:1 for the Changbai Mountain dataset.
Table 9. Point prediction results for data divided by 8:2 and data divided by 9:1 for the Changbai Mountain dataset.
ProportionsMetricQRGBDTQRXG BoostQRLightGBMQRRNNQRLSTMQRGRUQRFNNCNN-QRLSTMCNN-IQRLSTM
9:1MAE881.51859.14906.921248.91315.311452.751024.65879.72819.19
RMSE2179.632043.632206.632285.872800.543481.842103.182020.031913.09
MAPE0.070.070.070.110.10.10.090.070.07
SMAPE6.86.916.961111.0211.188.997.16.68
8:2MAE1114.431138.041210.611487.631619.961778.791442.061198.221190.85
RMSE2274.992186.642341.932468.293183.7393183.922444.4872255.282246.74
MAPE0.070.080.080.110.10450.120.10330.080.08
SMAPE7.577.88.091111.241212.5410.638.248.16
Note: boldface indicates the best result for each indicator.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qin, X.; Yin, D.; Dong, X.; Chen, D.; Zhang, S. Passenger Flow Prediction of Scenic Spots in Jilin Province Based on Convolutional Neural Network and Improved Quantile Regression Long Short-Term Memory Network. ISPRS Int. J. Geo-Inf. 2022, 11, 509. https://doi.org/10.3390/ijgi11100509

AMA Style

Qin X, Yin D, Dong X, Chen D, Zhang S. Passenger Flow Prediction of Scenic Spots in Jilin Province Based on Convolutional Neural Network and Improved Quantile Regression Long Short-Term Memory Network. ISPRS International Journal of Geo-Information. 2022; 11(10):509. https://doi.org/10.3390/ijgi11100509

Chicago/Turabian Style

Qin, Xiwen, Dongmei Yin, Xiaogang Dong, Dongxue Chen, and Shuang Zhang. 2022. "Passenger Flow Prediction of Scenic Spots in Jilin Province Based on Convolutional Neural Network and Improved Quantile Regression Long Short-Term Memory Network" ISPRS International Journal of Geo-Information 11, no. 10: 509. https://doi.org/10.3390/ijgi11100509

APA Style

Qin, X., Yin, D., Dong, X., Chen, D., & Zhang, S. (2022). Passenger Flow Prediction of Scenic Spots in Jilin Province Based on Convolutional Neural Network and Improved Quantile Regression Long Short-Term Memory Network. ISPRS International Journal of Geo-Information, 11(10), 509. https://doi.org/10.3390/ijgi11100509

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop