Next Article in Journal
A Dual-Path Cross-Modal Network for Video-Music Retrieval
Next Article in Special Issue
Binned Data Provide Better Imputation of Missing Time Series Data from Wearables
Previous Article in Journal
Spectra Reconstruction for Human Facial Color from RGB Images via Clusters in 3D Uniform CIELab* and Its Subordinate Color Space
Previous Article in Special Issue
3-D Data Interpolation and Denoising by an Adaptive Weighting Rank-Reduction Method Using Multichannel Singular Spectrum Analysis Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

State Causality and Adaptive Covariance Decomposition Based Time Series Forecasting

1
College of Computer Science, Sichuan University, Chengdu 610065, China
2
College of Innovation and Entrepreneurship, Shijiazhuang Institute of Technology, Shijiazhuang 050228, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(2), 809; https://doi.org/10.3390/s23020809
Submission received: 28 November 2022 / Revised: 4 January 2023 / Accepted: 5 January 2023 / Published: 10 January 2023
(This article belongs to the Special Issue Advances in Time Series Analysis)

Abstract

:
Time series forecasting is a very vital research topic. The scale of time series in numerous industries has risen considerably in recent years as a result of the advancement of information technology. However, the existing algorithms pay little attention to generating large-scale time series. This article designs a state causality and adaptive covariance decomposition-based time series forecasting method (SCACD). As an observation sequence, the majority of time series is generated under the influence of hidden states. First, SCACD builds neural networks to adaptively estimate the mean and covariance matrix of latent variables; Then, SCACD employs causal convolution to forecast the distribution of future latent variables; Lastly, to avoid loss of information, SCACD applies a sampling approach based on Cholesky decomposition to generate latent variables and observation sequences. Compared to existing outstanding time series prediction models on six real datasets, the model can achieve long-term forecasting while also being lighter, and the forecasting accuracy is improved in the great majority of the prediction tasks.

1. Introduction

Time series forecasting has long been a research focus. With the development of information technology, a significant amount of time series is generated in various production activities. As the output of the observation system, the time series, which is characterized by a large scale, objectively records the system’s information at each point in time, such as exchange rates and energy load. It has become a hot research topic to accurately mine the generation pattern and achieve high-precision long-term forecasting in large-scale time series. Traditional time series forecasting methods are based on statistical knowledge. Box and Jenkins [1] illustrated that the ARIMA model is theoretically applicable to time series analysis in various production activities.
In recent years, RNNs [2,3], which are essentially nonlinear exponential smoothing [4], have achieved satisfying results in sequence prediction for their ability to fit nonlinear relationships in short-term series. However, cumulative errors appear as the major drawback in long-term forecasting for large-scale time series in the models discussed above. Transformers [5,6] have fixed the problem of cumulative errors to some extent by optimizing the attention calculation method and decomposition-based strategy. However, it is firmly dependent on the periodicity of the series. In summary, the correlation between local data is crucial to the prediction results.
It shows that non-stationary properties are evident [7] from a survey of large-scale time series on exchange rates, diseases, and electrical loads. Moreover, similar to HMM [8], the observation sequence is dominated by hidden states. For example, temperature changes are influenced by weather and seasonal factors. VSMHN [9] and DGM2 [10] identify the hidden states by VAE [11,12], which applies neural networks to encode the latent variables [13] but is only suitable for the single-step generation of sparse sequences. It can be concluded that achieving long-term prediction for large-scale time series has the following challenges:
  • Challenge 1: Large-scale time series have evident non-stationary properties, and nonlinear models, such as neural networks, rely heavily on data periodicity. As a result, it is a challenge to investigate the typical generation rule of large-scale series and improve the model’s generalization;
  • Challenge 2: Hidden states influence the observation value of each moment, and the time span of a large-scale time series is larger. Therefore, it is a key issue to extend the hidden state estimation of time series from a certain moment to a period of time;
  • Challenge 3: Time series depend on the values observed before. Another challenging assignment for the long-term forecasting of large-scale time series is how to retain temporal dependence in order to address the cumulative error problem.
To meet the above challenges, this paper proposes the State Causality and Adaptive Covariance Decomposition-based Time Series Forecasting (SCACD) model for long-term forecasting of the large-scale time series. SCACD uses latent variables to designate hidden states, builds the neural network for estimating the distribution of latent and observed variables in large-scale scenarios drawing on the idea of VAE, and finally generates prediction sequences. The main works are as follows:
  • For Challenges 1 and 2, this paper first extracts sub-sequences by sliding window, secondly extends time points to larger scale sequences, then finally, designs an adaptive estimation method to encode the latent variables corresponding sub-sequences, and finally uses causal convolutional networks to predict future latent variables.
  • For Challenge 3, SCACD employs the Cholesky [14,15] decomposition of the covariance matrix to maintain the temporal dependence, thus generating the latent variables. Based on the latent variables, SCACD infers the prior distribution of the observed variables and generates an observation sequence with the same approach.
  • SCACD’s effectiveness was validated and analyzed on six publicly available standard large-scale datasets.

2. Related Work

With the development of deep learning techniques, time series forecasting algorithms are also constantly improving. There are three main research subtopics of time series forecasting tasks, from short-term forecasting of early sparse time series to long-term forecasting of large-scale time series: autoregressive models, Transformer-based models, and causal convolution-based models.

2.1. Autoregressive Model

Shallow autoregression: This category contains ARIMA, exponential smoothing [4], and Kalman filtering [2,16,17], as well as Hidden Markov Model-based models (HMM) [8] for identifying hidden states. Box-Jenkins [1] illustrates that ARIMA could achieve high prediction accuracy by estimating the model’s unknown parameters after converting the series into a stationary series. The principle of exponential smoothing is to decompose a time series into numerous components, model them separately, and then combine them to obtain the forecast series. Kalman filtering considers the deviation between the true values and observation values, utilizing multidimensional known information to infer more essential data. To forecast future hidden states and time series, the HMM [18,19] estimates the state transition probability matrix and observation matrix.
Deep autoregression: Deep models, such as RNN [20], have achieved better results compared to shallow models in sequence tasks. Although RNNs can fit nonlinear short-term series, they could not be more effective in the long-term prediction of large-scale time series. In addition, deep autoregressive models also include generation-based models such as DGM, VSMHN, DeepAR [21], and others, among which DeepAR is the most representative, which predicts the value of each moment by predicting the corresponding distribution. Therefore, we can conclude that the correlation between the time series plays a key role in the time-series prediction. However, the above models do not address the cumulative error of long-term forecasting for large-scale time series. Consequently, although the autoregressive model is applicable to most forecasting scenarios, long-term forecasting of large-scale time series still has a low prediction accuracy.

2.2. Transformer-Based Model

Transformer [6] has been utilized successfully in the field of sequences. Differing from autoregressive models, Transformer adds attention and positional encoding to achieve parallel computing. Recent research has revealed that the Transformer-based models perform exceptionally well in long-term time series prediction.
Thus far, the enhanced Transformer-based models for long-term forecasting include FEDformer [5], Autoformer [22], Informer [23], Reformer [24], LST [25], LogTrans [26], etc. This kind of algorithm primarily improves the attention calculation method, thus increasing the accuracy and reducing the time complexity. In addition, the position encoding is replaced by the timestamp encoding in the interim, increasing forecast accuracy. FEDformer calculates the attention based on the frequency components of time series by Fourier analysis, which significantly improves the results of long-term series prediction, but it is computationally expensive. Autoformer considers series autocorrelation to calculate the attention score in the Transformer module based on the decomposition strategy. Informer employs a distilled self-attention mechanism and multi-scale timestamp encoding. Reformer increases the time complexity through locality-sensitive hashing attention and space complexity through RevNet. The logTrans adds convolutional self-attention layers, reducing memory from O ( L 2 ) to O ( L ( log L ) 2 ) .
These models show that sequence dependencies can be modeled for long-term forecasting [27], but rely heavily on periodicity. In general, our findings suggest that the correlation between subsequences is crucial.

2.3. Causal Convolution-Based Model

General neural network prediction algorithms for time series, such as LSTM, DRU, RNN, etc., model historical data and then implement multi-step prediction, resulting in significant cumulative error. Causal convolution [28,29] and dilated convolution [30] solve this problem to some extent, which captures the autocorrelation of the time series. The dilation mechanism [31] is introduced to expand the field of perception of the causal network by adjusting the distance between the convolution kernel elements or increasing the dilation rate. The primary sequence methods based on causal convolution are Wavenet [29] and TCN [32]. Unlike RNNs, TCNs are convolutional neural networks that serve as convolutional prediction models for time series. TCNs overcome the cumulative error problem for long-term time series prediction to some extent by combining causal and dilated convolution. The essential reason is to implement causal convolutions at multiple scale sequences from bottom to top, thus extracting autocorrelation across multiple scale sequences and expanding the field of perception. TCN models the lowest scale data, therefore, its prediction scale is limited to the number of dilation convolution layers, and as the scale of the time series increases, the applicability decreases. In conclusion, while causal convolution can capture sequence dependency with fewer parameters, existing models continue to struggle with a long-term prediction of large-scale time series.

3. Materials and Methods

Problem Definition: Given a time series X = ( x 1 , , x i , , x L ) , x i R , the goal is to learn a mapping that predicts X ^ = ( x L + 1 , , x L + n ) , i.e., X ^ = f ( X ) . Since X is generated by some random process involving unobserved continuous random latent variable a. The process consists of two steps [11]: (1) a i is generated from a prior distribution p ( a i ) ; (2) x i is generated from a conditional distribution p ( x i | a i ) . To complete the forecasting based on these two steps, we propose an end-to-end SCACD composed of neural network components for large-scale time series. For long-term forecasting, a sliding window [1], which has two hyper-parameters: the size of the window (l) and the sliding step ( s t e p ), is employed to extract the subsequences S = ( s 1 , , s i , , s T ) , s i R l . s i corresponds to the latent variable z i R h , where h is the dimension of z i , i.e., S corresponds to Z = ( z 1 , , z i , , z T ) .
First, SCACD adaptively estimates the posterior distribution p ( z i s i ) of the latent variable z i , assuming z i N ( μ z , i , Σ z , i ) , and the ( μ i , Σ z , i ) serves as the feature encoding of z i . Furthermore, SCACD employs causal convolution to predict the p ( z T + 1 ) of the future latent variable z T + 1 and generates z T + 1 after executing the Cholesky decomposition of the covariance matrix Σ z , T + 1 . Finally, SCACD again adaptively estimates the conditional distribution p ( s T + 1 z T + 1 ) for the future sequence s T + 1 , s T + 1 N ( μ s , T + 1 , Σ s , T + 1 ) . The s T + 1 , similar to z T + 1 , is generated by sampling after Cholesky decomposition. The diagram of the SCACD model is shown in Figure 1.

3.1. Adaptive Estimation for p ( z i s i )

Based on the observed sequence s i , we infer the posterior distribution p ϑ ( z i s i ) of the latent variable, whose distribution parameters are μ z , i and Σ z , i . However, the existing predecessors [9,10] assume that the dimensions of the latent variable are independent of each other. Although this hypothesis saves computational effort, it fails to account for the correlation between dimensions, resulting in information loss. For this problem, SCACD designs an adaptive algorithm for estimating the mean and covariance matrix. The processes are shown in the formulas:
H z , i = M L P ( s i )
μ z , i = M L P ( s i )
Σ z , i = W ( H z , i μ z , i ) · ( W ( H z , i μ z , i ) ) T
Through the adaptive estimation module, the encoding sequences of Z are obtained: E μ = ( μ z , 1 , , μ z , i , , μ z , T ) , μ z , i R h , E Σ = ( Σ z , 1 , , Σ z , i , , Σ z , T ) , Σ z , i R h × h . The schematic of the adaptive estimation is shown in Figure 2.

3.2. State-Causal Convolution

Following the idea of second-order Markov chains [33], we use the encoding of the first two states to forecast the next state, i.e., the T is set to 2. Specifically, utilizing causal convolutions, the encodings of z T 1 and z T are used as input to estimate the mean and covariance matrix of z T + 1 : μ z , T + 1 and Σ z , T + 1 . The convolution processes are in the following formulas, and the diagram is shown in Figure 3.
μ z , T + 1 = C o n v 1 D ( E μ )
Σ z , T + 1 = C o n v 3 D ( E Σ )

3.3. Sequence Generation

3.3.1. Decomposition and Sampling

SCACD infers the prior distribution p π ( s T + 1 z T + 1 ) to generate future sequences, where z T + 1 is generated as p φ ( z T + 1 ) . Hence z T + 1 is crucial in putting the forecast into practice. Since z T + 1 is assumed to follow the Gaussian distribution: z T + 1 N ( μ z , T + 1 , Σ z , T + 1 ) , SCACD generates z T + 1 via Cholesky decomposition [14,15], resulting in efficient sampling. Positive semidefiniteness is necessary for Cholesky decomposition, Σ z , T + 1 is multiplied by Σ z , T + 1 T . The specific procedures are shown in Equation (6) and Figure 4.
z T + 1 = ξ · C + μ z , T + 1 Σ ^ z , T + 1 = C h o l e s k y C C T Σ ^ z , T + 1 = Σ z , T + 1 · Σ z , T + 1 T
where C belongs to the lower triangular matrix, ξ N ( 0 , I ) .

3.3.2. Sequence Prediction

Based on z T + 1 , SCACD utilizes MLP [11] to estimate the prior conditional distribution p π ( s T + 1 z T + 1 ) , as follows:
Σ s , T + 1 = W ( H s , T + 1 μ s , T + 1 ) · ( W ( H s , T + 1 μ s , T + 1 ) ) T μ s , T + 1 = M L P ( H s , T + 1 ) H s , T + 1 = M L P ( z T + 1 )
Finally, the forecasting sequence s ^ T + 1 is then obtained by re-implementing Cholesky-based sampling and adding an attention layer. The final objective function is as follows:
L ( ϑ , φ , π ) = 1 N n = 1 N ( s ^ T + 1 , n s T + 1 , n 2 2 )
s ^ T + 1 = 1 K k = 1 K s T + 1 k
where n denotes the sample serial number, and s T + 1 k is sampled K times. ϑ , φ , and π are the parameters to be optimized.

4. Results and Discussion

We conducted extensive experiments to evaluate the effectiveness of SCACD: comparison of prediction performance, covariance matrix analysis, analysis for the dimension of z, and analysis of efficiency. Moreover, six real datasets were calculated to certify the generality of SCACD.

4.1. Hyperparameter for SCACD

Pytorch 1.5, NVIDIA GeForce GTX 1060, and the Core i7-7700HQ CPU are the environments for implementation. The h is set as 16, the hidden layers for MLP are set as 3, the number of convolution layers is 1, and the K ranges from 5 to 50. SCACD was optimized by Adam optimizer [34], with the learning rate decaying by 50% every 100 epochs to avoid the local optimum. The parameters are suitable for most prediction tasks.

4.2. Introduction of the Models

  • FEDformer [5], being a Transformer-based model, computes attention coefficients in the frequency domain in order to represent point-wise interactions. Currently, FEDformer is the best model for long-term prediction, regardless of processing efficiency.
  • ARIMA [35] is a linear model that includes fewer endogenous variables and is mainly used for short-term forecasting of stationary time series.
  • LSTM [36] is a deep model for sequence prediction, which can capture the nonlinear relationship and solve the problem of RNN gradient disappearance by the memory unit.
  • TCN [28] utilizes causal and dilated convolution to capture the multi-scale temporal features of time series, which can build deep networks. The method is effective in time series multi-step prediction.
  • DeepAR [21] does not directly output definite prediction values but estimates the probability distribution of the future values to predict the observation sequence.
  • GM11 [37] is obtained by constructing and solving the differential equation of the cumulative series, which has the advantage of fewer prior constraints on the series.

4.3. Introduction to Datasets

The datasets for validating the algorithms include Electricity Transformer Temperature (ETT) [38], Exchange [25], Electricity, Weather, Traffic and Influenza-Like Illness (ILI) [22]. The details of the datasets are shown in Table 1, we divide the six datasets into training: test: val in a 7:2:1 ratio.

4.4. Comparison of Prediction Performance

Recently, Transformer-based models, such as FEDformor, Autoformer, and Informer, have performed best in long-term forecasting tasks. In order to verify the effectiveness of SCACD and make a fair comparison, the experiment set up the same six prediction tasks from the FEDformer to evaluate the efficiency. Two historical sliding windows are used to predict the future window sequence. Therefore, the sliding step size is the actual prediction length. The (l, s t e p ) corresponding tasks are, respectively, (800, 720), (350, 336), (200, 192), and (100, 96) for the datasets except for ILI, which is set as (62, 60), (50, 48), (38, 36), and (26, 24). The results are shown in Table 2. In general, SCACD has the best performance in most of the prediction tasks.
Retrospective Analysis. The FEDformer performed better in ETT-96, ETT-192, ETT-336, Electric-192, and Electricity-336. The investigation discovered that ETT and Electricity are more cyclical and that FEDformer reinforces the embedding with the timestamp, which improved prediction accuracy.
ARIMA is effective for short-term stationary time-series prediction, but it cannot handle the accumulative error problem of long-term non-stationary time-series prediction. TCN designs causal convolution and extended causal convolution to extract multi-scale features for a short-term forecast, but the generational rule of the time series is not mined. DeepAR mainly focuses on single-point prediction, ignoring the influence of the hidden state on the observation series, resulting in a poor long-term prediction effect. The GM is obtained by solving the differential equation of the cumulative series, which has fewer prior constraints on the series. However, it is less effective for long-term prediction.
Ablation Analysis. The SCACD-NC (SCACD without Cholesky decomposition) in Table 2 is the strategy assuming that the dimensions of the latent variable are independent. The results show that SCACD is superior to SCACD-NC except for Exchange-336, Weather-336, and ILI-48. The energy loss caused by the assumption of independence limits SCACD-NC’s generating ability. Furthermore, when compared to other models, SCACD-NC shows significant advantages. It exhibits the rationale and robustness of the state causality-based sequence generation algorithm for long-time series prediction tasks.
The advantages of SCACD are: (1) The covariance matrix is adaptively estimated to capture the correlation between the dimensions of the variable, which improves the generation accuracy of variables; (2) SCACD predicts the distribution of the future variables with fewer parameters by state causal convolution. In addition, in datasets with weak periodicity, SCACD’s prediction effect outperforms other methods. The visualizations of the predictions are presented in Figure A1.

4.5. Covariance Matrix Analysis

SCACD outperforms other deep models in terms of prediction efficiency and interpretability, and it investigates the inference behind the adaptive covariance matrix and causal convolution. Taking the ILI data set as an example, Figure 5 shows the covariance matrices of z and s at each scale. First, the heat of the sub-diagonal region of each matrix is darker, indicating the topological relationship between the dimensions of similar latent variables and the higher correlation between variables with a similar relationship.
In addition, there are highly correlated elements in the off-diagonal region. For example, the correlation coefficient between dimension 10 and dimension 12 in z T 1 , Figure 5a, in which a total of 106 values are displayed, is 0.9 , ranking second. The necessity of estimating covariance matrices for z T 1 and z T is verified. Meanwhile, the rank changes smoothly. Consider again the 10–12 pairs mentioned above; the ranks in z T 1 , z T , and z T + 1 are 2, 1, and 2, respectively. The index conforms to the concept drift principle, verifying the rationality and effectiveness of the state causal convolutional network. For larger scale time series, taking ETT as an example, the heat maps are shown in Figure A2.
The covariance matrix of the observation series is shown on the far right in Figure 5, and it has similar characteristics to the covariance matrix of the latent variable. The high correlation in off-diagonal regions also demonstrates the necessity of the covariance matrix-based sampling strategy.

4.6. Analysis for the Dimension of z

SCACD obtained excellent performance for long-term forecasting. The latent variable is essential for inferring the distribution of the observed sequence, so we conducted experiments to investigate the effect of the dimension of z on prediction precision. The dimensions of z are set as 8, 16, 24, and 32, and SCACD was implemented, respectively. The corresponding results in Table 3 show that the efficiency of the model with 16-dimensional z was better overall compared to the other settings. Generally, the prediction results for all four settings are good, indicating a fair and robust model.

4.7. Analysis of Efficiency

In comparison to the other deep models, SCACD obtains outstanding performance based on a simple construction. This experiment illustrates the proposed model’s indicators in the implementing processes. Table 4 shows the number of model parameters and the time required for the forward and backward processes under the same conditions.
It is evident that SCACD has the fewest parameters and takes the least time. Furthermore, SCACD converges swiftly in all the prediction tasks shown in Figure 6, which demonstrates the model’s stability and robustness.

5. Conclusions

This work proposes a time-series prediction model (SCACD) based on state convolution and adaptive covariance matrix decomposition to handle the challenges of long-term prediction of large-scale time series. SCACD is a large-scale generation strategy to solve the cumulative error problem. First, MLPs are utilized to adaptively encode the hidden states. Secondly, a causal convolution is implemented to predict the distribution of future variables, and finally, a decomposition-based sampling is executed to complete the forecasting. SCACD infers the generational law of large-scale time series through two steps, as opposed to existing methods that rely on calculating the attention of the observation sequence. As it turned out, SCACD has significant advantages over baseline models, particularly in weak cycle datasets such as Exchange and ILI, where SCACD outperforms the current SOTA model by more than 500%. Furthermore, the extensive experiments demonstrate that the proposed model is highly interpretable and performs well.
Possible directions for extending this work include finding algorithms for extracting subsequences based on concept drift and multi-level causal convolution strategies. In this work, we used a single-level causal convolution network because of the fixed length of the subsequence and the small number of subsequences. Foreseeably, as the scale of the time series increases further, one could construct personalized multi-scale prediction models by combining adaptive subsequences, causal convolution, and dilated convolution [28].

Author Contributions

Conceptualization, J.W. and J.P.; methodology, J.W. and F.H.; software, J.W. and T.G.; validation, J.W.; formal analysis, Z.H., F.H. and P.Y.; investigation, J.W. and F.H.; resources, T.G., J.P. and F.H.; data curation, J.P., Z.H. and P.G.; writing—original draft preparation, J.W., Z.H., F.H. and J.P.; visualization, Z.H.; supervision, J.P.; project administration, T.G., J.P. and F.H.; funding acquisition, J.P., T.G. and F.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant No. 61906126; the Intelligent Terminal Key Laboratory of Sichuan Province grant No. SCITLAB-20001; the Sichuan Science and Technology Program grant No. 2022YFG0034; the Cooperative Program of Sichuan University and Yibin grant No.2020CDYB-30; Post doctoral Interdisciplinary Innovation Fund (10822041A2137).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets’ sources are fully annotated in the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Prediction Curves

The prediction results prove that SCACD has mastered the time-series generation law, which considerably decreases the accumulation of long-term prediction errors in large-scale time series. The prediction curves are shown in Figure A1.
Figure A1. The forecasting curves of the six datasets.
Figure A1. The forecasting curves of the six datasets.
Sensors 23 00809 g0a1

Appendix A.2. Heatmaps for ETT

Due to the large-scale characteristic, the covariance matrix heat map of s has poor resolution, so only the covariance matrices of the latent variables are shown in Figure A2.
Figure A2. The covariance matrices for all the states for the ETT.
Figure A2. The covariance matrices for all the states for the ETT.
Sensors 23 00809 g0a2aSensors 23 00809 g0a2b

References

  1. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  2. Engle, R.F. Autoregressive conditional heteroscedasticity with estimates of the variance of united kingdom inflation. Econom. J. Econom. Soc. 1982, 50, 987–1007. [Google Scholar] [CrossRef]
  3. Petneházi, G. Recurrent neural networks for time series forecasting. arXiv 2019, arXiv:1901.00069. [Google Scholar]
  4. Gardner, E.S., Jr. Exponential smoothing: The state of the art—Part II. Int. J. Forecast. 2006, 22, 637–666. [Google Scholar] [CrossRef]
  5. Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. arXiv 2022, arXiv:2201.12740. [Google Scholar]
  6. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. Available online: https://papers.nips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html (accessed on 27 November 2022).
  7. Cheng, L.; Zang, H.; Xu, Y.; Wei, Z.; Sun, G. Augmented convolutional network for wind power prediction: A new recurrent architecture design with spatial-temporal image inputs. IEEE Trans. Ind. 2021, 17, 6981–6993. [Google Scholar] [CrossRef]
  8. Deng, Q.; Söffker, D. A review of the current hmm-based approaches of driving behaviors recognition and prediction. IEEE Trans. Intell. Veh. 2021. [Google Scholar] [CrossRef]
  9. Li, L.; Zhang, J.; Yan, J.; Jin, Y.; Zhang, Y.; Duan, Y.; Tian, G. Synergetic learning of heterogeneous temporal sequences for multi-horizon probabilistic forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021. [Google Scholar]
  10. Wu, Y.; Ni, J.; Cheng, W.; Zong, B.; Song, D.; Chen, Z.; Liu, Y.; Zhang, X.; Chen, H.; Davidson, S.B. Dynamic gaussian mixture based deep generative model for robust forecasting on sparse multivariate time series. Proc. AAAI Conf. Artif. Intell. 2021, 35, 651–659. [Google Scholar] [CrossRef]
  11. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  12. Chung, J.; Kastner, K.; Dinh, L.; Goel, K.; Courville, A.C.; Bengio, Y. A recurrent latent variable model for sequential data. Adv. Neural Inf. Process. Syst. 2015, 28, 2980–2988. [Google Scholar]
  13. Lyhagen, J. The exact covariance matrix of dynamic models with latent variables. Stat. Probab. Lett. 2005, 75, 133–139. [Google Scholar] [CrossRef]
  14. Pourahmadi, M. Cholesky decompositions and estimation of a covariance matrix: Orthogonality of variance–correlation parameters. Biometrika 2007, 94, 1006–1013. [Google Scholar] [CrossRef]
  15. Lv, J.; Guo, C.; Yang, H.; Li, Y. A moving average cholesky factor model in covariance modeling for composite quantile regression with longitudinal data. Comput. Stat. Data Anal. 2017, 112, 129–144. [Google Scholar] [CrossRef]
  16. Liu, J.; Hu, J.; Li, Z.; Sun, Q.; Ma, Z.; Zhu, J.; Wen, Y. Dynamic estimation of multi-dimensional deformation time series from insar based on kalman filter and strain model. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
  17. de Bézenac, E.; Rangapuram, S.S.; Benidis, K.; Bohlke-Schneider, M.; Kurle, R.; Stella, L.; Hasson, H.; Gallinari, P.; Januschowski, T. Normalizing kalman filters for multivariate time series analysis. Adv. Neural Inf. Process. Syst. 2020, 33, 2995–3007. [Google Scholar]
  18. Abe, K.; Miyatake, H.; Oguri, K. A study on switching ar-hmm driving behavior model depending on driver’s states. In Proceedings of the 2007 IEEE Intelligent Transportation Systems Conference, Bellevue, WA, USA, 30 September–3 October 2007; pp. 806–811. [Google Scholar]
  19. Barber, C.; Bockhorst, J.; Roebber, P. Auto-regressive hmm inference with incomplete data for short-horizon wind forecasting. Adv. Neural Inf. Process. Syst. 2010, 23, 136–144. [Google Scholar]
  20. Zaremba, W.; Sutskever, I.; Vinyals, O. Recurrent neural network regularization. arXiv 2014, arXiv:1409.2329. [Google Scholar]
  21. Salinas, D.; Flunkert, V.; Gasthaus, J.; Januschowski, T. Deepar: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
  22. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  23. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. Proc. AAAI Conf. Artif. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  24. Kitaev, N.; Kaiser, Ł.; Levskaya, A. Reformer: The efficient transformer. arXiv 2020, arXiv:2001.04451. [Google Scholar]
  25. Lai, G.; Chang, W.-C.; Yang, Y.; Liu, H. Modeling long-and short-term temporal patterns with deep neural networks. In Proceedings of the The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 95–104. [Google Scholar]
  26. Li, S.; Jin, X.; Xuan, Y.; Zhou, X.; Chen, W.; Wang, Y.-X.; Yan, X. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. Adv. Neural Inf. Process. Syst. 2019, 32, 5243–5253. [Google Scholar]
  27. Zeng, A.; Chen, M.; Zhang, L.; Xu, Q. Are transformers effective for time series forecasting? arXiv 2022, arXiv:2205.13504. [Google Scholar]
  28. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  29. Oord, A.V.D.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  30. Yu, F.; Koltun, V.; Funkhouser, T. Dilated residual networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 8–14 December 2017; pp. 472–480. [Google Scholar]
  31. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  32. Lara-Benítez, P.; Carranza-García, M.; Luna-Romera, J.M.; Riquelme, J.C. Temporal convolutional networks applied to energy-related time series forecasting. Appl. Sci. 2020, 10, 2322. [Google Scholar] [CrossRef] [Green Version]
  33. Wilinski, A. Time series modeling and forecasting based on a markov chain with changing transition matrices. Expert Syst. Appl. 2019, 133, 163–172. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Gilbert, K. An arima supply chain model. Manag. Sci. 2005, 51, 305–310. [Google Scholar] [CrossRef]
  36. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  37. Kayacan, E.; Ulutas, B.; Kaynak, O. Grey system theory-based models in time series prediction. Expert Syst. Appl. 2010, 37, 1784–1789. [Google Scholar] [CrossRef]
  38. Zhou, J.; Lu, X.; Xiao, Y.; Su, J.; Lyu, J.; Ma, Y.; Dou, D. Sdwpf: A dataset for spatial dynamic wind power forecasting challenge at kdd cup 2022. arXiv 2022, arXiv:2208.04360. [Google Scholar]
Figure 1. The schema of SCACD. It mainly includes three parts: Adaptive Estimation, State Causal Convolution, and Cholesky Decomposition.
Figure 1. The schema of SCACD. It mainly includes three parts: Adaptive Estimation, State Causal Convolution, and Cholesky Decomposition.
Sensors 23 00809 g001
Figure 2. The phase of adaptive estimation. (1) Find the embedding of s i ; (2) Estimate μ z , i and H i ; (3) Get Σ z , i .
Figure 2. The phase of adaptive estimation. (1) Find the embedding of s i ; (2) Estimate μ z , i and H i ; (3) Get Σ z , i .
Sensors 23 00809 g002
Figure 3. State causal convolution. The mean and covariance matrix of z T + 1 are estimated using 1D and 3D convolutions, respectively.
Figure 3. State causal convolution. The mean and covariance matrix of z T + 1 are estimated using 1D and 3D convolutions, respectively.
Sensors 23 00809 g003
Figure 4. Cholesky decomposition. The covariance matrix is decomposed to the diagonal matrix, followed by sampling z T + 1 .
Figure 4. Cholesky decomposition. The covariance matrix is decomposed to the diagonal matrix, followed by sampling z T + 1 .
Sensors 23 00809 g004
Figure 5. The covariance matrices for all the states of the prediction processes for the four tasks. Taking (a), for example, z T 1 , z T , and z T + 1 are the adaptive covariance matrices for each period, and the heat map of prediction series is s T + 1 .
Figure 5. The covariance matrices for all the states of the prediction processes for the four tasks. Taking (a), for example, z T 1 , z T , and z T + 1 are the adaptive covariance matrices for each period, and the heat map of prediction series is s T + 1 .
Sensors 23 00809 g005aSensors 23 00809 g005b
Figure 6. The convergence process of SCACD. The abscissa represents the number of epochs, and the ordinate represents the loss.
Figure 6. The convergence process of SCACD. The abscissa represents the number of epochs, and the ordinate represents the loss.
Sensors 23 00809 g006
Table 1. The details of the datasets.
Table 1. The details of the datasets.
Strongly Periodic Datasets
ETTETT is a key time-series indicator for long-term power deployment, which contains electricity data recorded every 15 min for two different counties in China from July 2016 to July 2018.
ElectricityThis dataset contains the hourly electricity consumption of 321 customers from 2012 to 2014. It records electricity usage in kilowatt-hours and is timestamped every 15 min.
WeatherIn order to verify the effect of the algorithm, we selected weather data containing 21 meteorological indicators (such as air temperature and humidity) in 2020 from the public dataset, and its time series sampling frequency was 10 min.
Weekly periodic datasets
ExchangeExchange records the daily exchange rates of eight different countries as a type of classical time series, with frequencies being daily, and the dataset covers exchange rate data from 1990 to 2016.
TrafficTraffic is a collection of hourly data covering the highway system in all major urban areas of California, recording road occupancy rates measured by different sensors at different times of the day.
ILITo verify the robustness of the algorithm on multiple time series datasets, we selected this dataset as the final part, which includes weekly Influenza-Like Illness (ILI) patient data recorded from 2002 to 2021, with the time series frequency being the weekly frequency data, which describes the ratio of ILI patients to the total number of patients.
Table 2. The best results are shown in bold. Compared with the current state-of-the-art FEDformer, the experimental effect of SCACD is improved by 10–100% in blue font, 100–500% in brown font, and greater than 500% in red font. The results for TCN are from Autoformer [22].
Table 2. The best results are shown in bold. Compared with the current state-of-the-art FEDformer, the experimental effect of SCACD is improved by 10–100% in blue font, 100–500% in brown font, and greater than 500% in red font. The results for TCN are from Autoformer [22].
ModelsSCACDSCACD-ncFEDformerARIMALSTMTCNDeepARGM11
LenMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
ETT960.0660.1860.0630.1840.0350.1460.2420.3674.8652.1293.0411.3301.4851.0520.0410.150
1920.0750.2150.0820.2250.0660.2020.1470.2844.0941.9523.0721.3393.8711.7720.8860.243
3360.1000.2370.1310.2880.0880.2340.2000.3123.3441.7133.1051.3481.0450.7830.3700.286
7200.0730.2110.1390.2800.1060.2520.1380.2493.4131.6763.1351.3540.9980.8250.0960.236
Electricity960.2130.3400.2520.3590.2510.3670.9220.7661.0340.7940.9850.8130.5730.6150.5940.624
1920.3780.4630.3800.4630.0560.1841.0470.8010.7230.6680.9960.8210.5020.5763.3750.806
3360.1600.2890.2240.3410.0820.2250.9910.7491.4280.8901.0000.8240.5820.6190.9670.369
7200.2080.3020.5660.5450.4030.4733.2980.9900.5590.6221.4380.7840.6710.6751.1660.636
Exchange960.0720.2020.0730.2030.1700.3190.0590.1800.2470.4323.0041.4321.3681.0950.1430.129
1920.0540.1910.0540.1820.3110.4310.2590.3700.4120.5143.0481.4440.2750.4310.0870.166
3360.0710.2130.0660.2040.5990.5920.6680.5020.4450.5243.1131.4590.7130.7320.2640.304
7200.1870.3291.3310.9801.4320.9241.5360.9340.9450.8263.1501.4580.2820.4240.8390.550
Weather960.0550.1660.0810.2040.2110.3230.5000.5790.9600.8551.4380.7840.9730.8371.0290.880
1920.1180.2320.1310.2470.2230.3280.7470.7270.9360.8241.4630.7941.1090.8690.9430.848
3360.0570.1710.0590.1270.2200.3260.7340.7120.9640.8391.4790.7990.8340.8070.9250.839
7200.1020.2300.1480.2900.2330.3330.8590.8010.9080.7771.4990.8040.8230.7910.9240.844
Traffic960.0020.0300.0050.0570.0050.0560.0050.0520.0400.1530.6150.5890.1250.2820.4140.062
1920.0010.0220.0080.0680.0050.0570.0060.0500.0830.2480.6290.6000.0960.2470.7910.089
3360.0020.0290.0010.0260.0050.0541.6660.0950.1790.3990.6390.6080.1240.2800.7980.094
7200.0020.0350.0050.0520.0040.0461.7440.0590.0770.2580.6390.6100.3210.4510.5600.072
ILI240.1110.2840.1580.3340.7010.6290.1930.3411.0050.8426.6241.8300.8930.7500.7650.481
360.1810.3560.1870.3430.5810.6130.2580.3991.0980.8526.8581.8790.9480.7630.2790.371
480.0620.1960.0540.1680.7130.6990.3430.4270.9290.7296.9681.8921.0450.8030.2960.361
600.0650.1930.1970.3270.8400.7680.2290.3790.7890.6657.1271.9181.0940.8180.8510.443
10–100% 100–500% >=500%
Table 3. The dimensions of the latent variable z in SCACD are taken as 8, 16, 24, and 32. The other hyperparameters are as in Section 4.1.
Table 3. The dimensions of the latent variable z in SCACD are taken as 8, 16, 24, and 32. The other hyperparameters are as in Section 4.1.
The Dim of LatentSCACD-8SCACD-16SCACD-24SCACD-32
LenMSEMAEMSEMAEMSEMAEMSEMAE
ETT960.0640.1850.0660.1860.0880.2160.0630.180
1920.1400.2960.0750.2150.1260.2800.1160.262
3360.0990.2500.1000.2370.1480.2900.1380.288
7200.1890.3440.0730.2110.2250.3680.1490.310
Electricity960.2910.4030.2130.3400.3170.4100.2520.366
1920.4580.5230.3780.4630.3870.4730.3790.460
3360.2120.3330.1600.2890.2610.3650.2550.380
7200.3750.4660.2080.3020.5040.4990.4830.527
Exchange960.0780.2060.0720.2020.0440.1650.0570.197
1920.1380.2970.0540.1910.2610.3690.0630.195
3360.4050.5010.0710.2130.1280.2800.5640.581
7200.7570.6240.9080.7800.7960.8440.5090.597
Traffic960.0950.2030.0550.1660.0870.1970.0870.199
1920.1000.2170.1180.2320.0770.2010.1650.275
3360.0350.1170.0570.1710.0440.1300.0470.133
7200.1320.2620.1020.2300.1680.2810.1440.267
Weather960.0010.0230.0020.0300.0060.0560.0180.075
1920.0040.0500.0010.0220.0010.0230.0040.050
3360.0050.0710.0020.0290.0030.0450.0030.044
7200.0050.0390.0020.0350.2150.0700.0360.144
ILI240.0940.2600.1110.2840.1290.2940.0940.242
360.1700.3360.1810.3560.1870.3580.1860.346
480.1350.3040.0620.1960.1670.3440.1170.276
600.0880.2310.0650.1930.1170.2610.0290.132
Table 4. The values correspond to the prediction scales. The hyperparameters are set as in Section 4.1. P: the number of the parameters; B: time (s/batch) for the backward; F: time (s/batch) for the forward.
Table 4. The values correspond to the prediction scales. The hyperparameters are set as in Section 4.1. P: the number of the parameters; B: time (s/batch) for the backward; F: time (s/batch) for the forward.
96192336720
PBFPBFPBFPBF
SCACD 4.85 × 10 4 2.74 × 10 2 1.18 × 10 2 1.74 × 10 5 2.68 × 10 2 1.17 × 10 2 5.13 × 10 5 3.03 × 10 2 1.15 × 10 2 2.61 × 10 6 4.10 × 10 2 1.50 × 10 2
Autoformer 1.1 × 10 7 6.63 × 10 2 6.59 × 10 2 1.05 × 10 7 1.04 × 10 1 1.23 × 10 1 1.05 × 10 7 1.91 × 10 1 2.36 × 10 1 1.05 × 10 7 1.07 × 10 1 1.17 × 10 1
FEDformer 1.63 × 10 7 2.54 × 10 1 5.50 × 10 2 1.63 × 10 7 2.51 × 10 1 5.83 × 10 2 1.68 × 10 7 2.87 × 10 1 6.73 × 10 2 1.63 × 10 7 2.93 × 10 1 6.69 × 10 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; He, Z.; Geng, T.; Huang, F.; Gong, P.; Yi, P.; Peng, J. State Causality and Adaptive Covariance Decomposition Based Time Series Forecasting. Sensors 2023, 23, 809. https://doi.org/10.3390/s23020809

AMA Style

Wang J, He Z, Geng T, Huang F, Gong P, Yi P, Peng J. State Causality and Adaptive Covariance Decomposition Based Time Series Forecasting. Sensors. 2023; 23(2):809. https://doi.org/10.3390/s23020809

Chicago/Turabian Style

Wang, Jince, Zibo He, Tianyu Geng, Feihu Huang, Pu Gong, Peiyu Yi, and Jian Peng. 2023. "State Causality and Adaptive Covariance Decomposition Based Time Series Forecasting" Sensors 23, no. 2: 809. https://doi.org/10.3390/s23020809

APA Style

Wang, J., He, Z., Geng, T., Huang, F., Gong, P., Yi, P., & Peng, J. (2023). State Causality and Adaptive Covariance Decomposition Based Time Series Forecasting. Sensors, 23(2), 809. https://doi.org/10.3390/s23020809

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop