Next Article in Journal
Linear Computation Coding: A Framework for Joint Quantization and Computing
Previous Article in Journal
Adaptive IDS for Cooperative Intelligent Transportation Systems Using Deep Belief Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Informer-WGAN: High Missing Rate Time Series Imputation Based on Adversarial Training and a Self-Attention Mechanism

1
Faculty of Computing, Harbin Institute of Technology, Harbin 150000, China
2
Beijing Institute of Control Engineering, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Algorithms 2022, 15(7), 252; https://doi.org/10.3390/a15070252
Submission received: 28 April 2022 / Revised: 13 July 2022 / Accepted: 15 July 2022 / Published: 20 July 2022

Abstract

:
Missing observations in time series will distort the data characteristics, change the dataset expectations, high-order distances, and other statistics, and increase the difficulty of data analysis. Therefore, data imputation needs to be performed first. Generally, data imputation methods include statistical imputation, regression imputation, multiple imputation, and imputation based on machine learning methods. However, these methods currently have problems such as insufficient utilization of time characteristics, low imputation efficiency, and poor performance under high missing rates. In response to these problems, we propose the informer-WGAN, a network model based on adversarial training and a self-attention mechanism. With the help of the discriminator network and the random missing rate training method, the informer-WGAN can efficiently solve the problem of multidimensional time series imputation. According to the experimental results under different missing rates, the informer-WGAN model achieves better imputation results than the original informer on two datasets. Our model also shows excellent performance on time series imputation of the key parameters of a spacecraft control moment gyroscope (CMG).

1. Introduction

Recently, with the continuous acceleration of industrialization, industrial sensors are used in all walks of life to obtain various data and are widely used in many fields such as weather forecasting, medical treatment, power statistics, and aerospace telemetry [1]. A time series is the data collected at a certain frequency of industrial physical quantities monitored by sensors, since multiple sensors in an industrial scene are often correlated and sampled from multiple industrial sensors to obtain a multidimensional time series. However, due to the problem of the hardware device, the multidimensional time series in the actual scene may have the problem of missing values. Not only that, but for some fields, such as weather forecasting and stock forecasting, multidimensional time series forecasting is also a very valuable research problem [2], both of which can be defined as imputation problems for missing data. For imputation of multidimensional time series, sufficient analysis of the time series is required.
In order to deal with the imputation problem of multidimensional time series, researchers have provided many traditional algorithms, forming the initial solutions. For example, the autoregressive integrated moving average (ARIMA) treats the imputation of time series as the process of calculating the unknown dependent variable from the known independent variables [3]. Schafter applied the Markov chain Monte Carlo method (MCMC method) to data imputation for processing data with multivariate arbitrary missing patterns [4]. However, these methods often require some prior knowledge, such as the difference order and autocorrelation lag order of the ARIMA [5]. At the same time, for complex multidimensional time series with high missing rates, due to the incomplete utilization of time series characteristics, the performance of these methods is generally poor [6].
In recent years, many machine learning methods have been used for multidimensional time series imputation problems. Rhaman et al. used the EM algorithm to combine the decision tree and build a decision forest. After classifying and merging the data, the EM algorithm was used to implement data imputation. Through experiments on six public datasets with missing values and three artificially made missing value datasets, as well as a comparison with the DMI and SiMI algorithms, the effectiveness of this method for the imputation of missing data in attribute classes was verified [7]. In 2015, Folguera et al. proposed applying a self-organizing map neural network (SOM) to the data imputation task to predict the physical and chemical parameters of water samples. An SOM uses a nonlinear method to automatically fill in missing values in the data matrix. Folguera verified the SOM method in data imputation through experiments on surface water data (17–39% missing rate) sampled from the Reconquista River in Argentina. Regarding task effectiveness [8], Langkvist provided a review of the methods related to building prediction models for time series through deep learning. The two models that perform well in sequence modeling problems are based on self-attention mechanism models and recurrent neural network-based models [9]. DeepAR, proposed by Salinas et al., utilizes an encoder-decoder structure combined with an autoregressive RNN model for time series prediction [10]. These machine learning methods have all conducted valuable explorations for the multidimensional time series imputation problem. Inspired by Langkvist’s review, we adopted a network model based on the self-attention mechanism to impute time series.
The main contributions of this paper include the following two aspects:
  • We propose a model for multidimensional time series imputation, which is based on the WGAN-GP framework and uses the informer part of the network structure to form the generator and the discriminator. The networks proposed outperform the original informer model and the GAN-based AST model in both real-world datasets.
  • We propose a random missing rate training method for the time series imputation problem which improves the accuracy of the data imputation model for time series imputation with different missing rates.

2. Related Work

2.1. Time Series Imputation

The data imputation method is to solve the problem of missing data, which refers to the phenomenon where the dataset is incomplete due to a missing value for one or more attributes. Missing data will not only distort the data characteristics and increase the difficulty of analysis but also change the statistics such as expectations and high-order distances, causing serious deviations in the analysis methods based on statistical models. For the imputation of missing data in time series, many researchers have proposed solutions before. The ARIMA model proposed by Box-Jenkins uses the autocorrelation of the time series lag k order to construct an autoregressive model [3] for time series forecasting. To predict the time series which exhibit long-range dependence or persistence in their observations, Javier E. Contreras-Reyes compared several autoregressive fractionally integrated moving average (ARFIMA) models [5]. Augustine Pwasong and Saratha Sathasivam [11] combined the ARFIMA model and a layer recurrent neural network (LRNN) to form a hybrid forecasting model. The ARIMA-LRNN model achieved a very robust and dependable result. Liu et al. combined the online learning algorithm with the ARIMA algorithm and proposed the online ARIMA algorithm. The online ARIMA algorithm is mainly used to solve the difficult problems of parameter estimation in the traditional ARIMA algorithm, such as lag order and interference terms. Liu reduced the ARIMA model to an online parameter optimization problem and verified through experiments that the online ARIMA algorithm performed better than the traditional ARIMA algorithm [12]. The network model of time series imputation based on an RNN is also a major research direction [13]. DeepAR was proposed by Salinas et al. by using the encoder-decoder structure combined with the autoregressive RNN model for time series prediction [10]. LSTNet is another RNN-based autoregressive model that utilizes long short-term memory (LSTM) for time series prediction [14]. However, due to the natural disadvantages of the RNN network model, parallel operations cannot be used to improve the imputation efficiency, especially when the missing rate is high, at which point it shows poor performance.

2.2. Generative Adversarial Networks

A generative adversarial neural network is a kind of generative deep neural network proposed by Ian et al. in 2014 [15] through the adversarial training of a generator and discriminator to learn the distribution characteristics of the original dataset. Generative adversarial networks have achieved great success in computer vision and natural language processing in recent years, and at the same time, several studies have also been applied to generate sequence data [16]. After using the method of machine learning to complete the data imputation, we need to train the model by minimizing the likelihood function, but the likelihood function is limited to the description of the sequence characteristics, meaning that the discriminator is essentially fitting a better likelihood function to identify the complete sequence after imputation.
In 2019, Yoon et al. used a GAN and unsupervised learning to generate time series data and proposed TimeGAN, which achieved the best results for time series generation tasks. This model is not directly applicable to data imputation, as it was only used to model historical time series data [17]. Luo et al. proposed a method using E2EGAN for multidimensional time series imputation using a gate recurrent unit (GRU) as a generator and a discriminator network and achieved the best results on multiple real-world datasets. At the same time, the E2EGAN model is superior to the two-stage method in terms of time consumption. Combining the idea of a GAN and the self-attention mechanism, Wu et al. proposed an adversarial sparse transformer (AST) model, which achieved optimal results on multiple real-world datasets. The AST model uses a sparse attention encoder and decoder as the generator network and uses a linear fully connected layer as the discriminator network, which improves the imputation performance of the model compared with the RNN-based algorithm. At the same time, Wu also proved through ablation experiments that adversarial training helps to obtain better data imputation models. However, the AST model adopts the step-by-step generation method, which has low imputation efficiency and has the problem of error accumulation [6].

2.3. Attention Mechanism

The attention mechanism has been shining in the field of time series processing and prediction since it was proposed in 2017 [18], as it solves the problem that the RNN model can by not only predicting in a single step in time series prediction but can also improve the efficiency of the time series processing. Especially in long sequence prediction, the attention mechanism has a tendency to replace RNN methods such as LSTM. Some recent works also used sparse attention [19,20] based on the attention mechanism. Due to the softmax mechanism, the traditional attention mechanism gives an attention score to each unit in the embedding even if the correlation is extremely low, resulting in distraction and reduced computing efficiency. The sparse attention mechanism sets the irrelevant units to zero, which will make the model more focused. Experiments have shown that the sparse attention-based model is effective for time series imputation tasks on some datasets. In 2021, the informer proposed by Zhou et al. used convolution and pooling operations to perform knowledge distillation on the embedded data, and they were combined with the prob attention (i.e., sparse attention) to improve the performance of the model [21].

3. Materials and Methods

3.1. Materials

Dataset

We used three public real-world datasets (electricity, ETTh1, and CMG) for our evaluation. The electricity dataset (Table 1), which consists of hourly electricity consumption data collected from 370 customers, was used to evaluate the AST model. The ETTh1 dataset is hourly power load index data collected from several counties in China, and it was used for the informer model. The cmg dataset is the real telemetry time series data of the key parameters of a spacecraft’s control moment gyro (CMG).

3.2. Informer-WGAN Model

In this section, we will present the details of the proposed time series imputation method. The generative adversarial network (GAN) consists of a generator network and a discriminator network. The generator network takes low-dimensional random noise as its input and outputs high-dimensional fake data according to different tasks. The discriminator network takes fake data and real data as the input and outputs a number in the range of (0, 1) as the evaluation of the input data. The discriminator network will try to score real data close to 1 and fake data close to 0, while the generator network will try to generate fake data that makes the discriminator network give a high score. We would obtain an excellent generator that generates realistic data taking advantage of adversarial training of the generator and discriminator. In order to avoid the mode collapse problem in GAN training, we adopted the WGAN-GP model (please see Figure 1) as the neural network framework while making some modifications to the original loss function.
Our network structure refers to the adversarial sparse transformer (AST) model, and we redesigned and implemented the generator network and the discriminator network based on it.

3.2.1. Problem Definition

The problem is multidimensional time series imputation, where d represents the dimension of the time series, t represents the length of the time series, and thus the time series is defined as x R d × n . According to the time feature of the sequence, we generated a covariate matrix through the sequence timestamp information c o v , c o v R d c o v × n , where d c o v describes the information of dimensions such as m o n t h , d a y , h o u r , and  w e e k d a y in the sequence timestamp. We decided the specific d c o v according to the feature of the actual time series, such as the period and trend.

3.2.2. Generator

In our proposed generator network Figure 2, the informer is mainly used as the generator network. Different from the specified missing length and token length of the original informer, the random missing rate and data window are used to calculate the missing length, which increases the robustness of the network for imputation under different missing rates of time series.
First, according to the missing rate r, the data window size L, and the formulas t 0 = L ( 1 r ) / 100 and τ = L r / 100 , we calculated the known time series and the unknown part. The encoder part receives x 0 : t 0 and c o v 0 : t 0 as inputs, where x 0 : t 0 is the real value of the time series from 0 to t 0 and c o v 0 : t 0 is the part of the covariate matrix from 0 to t 0 , calculated according to the time characteristics of the time series mentioned above. After the operation of n encoder parts, the network outputs the memory module z. The input of the decoder includes the output of the encoder z, d e c _ i n p u t , and c o v t 0 + 1 t o k e n _ l e n : t 0 + τ to participate in the self-attention operation of the decoder part, where d e c _ i n p u t = c o n c a t ( x t 0 + 1 t o k e n _ l e n : x t 0 , 0 t 0 + 1 : t 0 + τ ) . After n decoders, the generator outputs a fake sequence x ^ t 0 + 1 t o k e n _ l e n : t 0 + τ , which is spliced with the real value of the time series from 0 to t 0 + τ , to obtain the full time series forecast from 0 to t 0 + τ x ^ .
The loss function of the generator is a bit different from that of the original WGAN-GP:
L G = α D ( x ^ ) + Q 50 ( G ( z ) , x t 0 + 1 : t 0 + τ )
Since the discriminator network evaluates the entire complete sequence, we paid more attention to the sequence part of the actual imputation, which was achieved by the q u a n t i l e -50 loss.

3.2.3. Discriminator

Different from the AST model, which uses a multi-layer fully connected layer as the discriminator, the discriminator network Figure 3 we propose uses the encoder part of the informer model to extract the features of the time series. The time series features are extracted through a multi-layer encoder, and the discrimination results are output through a multi-layer fully connected layer.
The discriminator is used to distinguish the real value x of the time series from the fake sequence x ^ generated by the generator. The goal is to give a high score to the real value of the sequence and a low score to the fake sequence, so the loss function is as follows:
L D = D ( x ^ ) D ( x ) + g r a d i e n t _ p e n a l t y

3.2.4. Random Training

According to the problem definition, in order to enhance the robustness of the model and enable the model to generate high-quality imputation data at various missing rates, especially at high missing rates, we propose a network training process based on random missing rates. The process is as follows Algorithms 1:
Algorithm 1 GAN training process based on random missing rate.
1:
for each training iteration do
2:
   randomly sample [X,Y] from dataset by w i n d o w _ s i z e
3:
   set m i s s i n g _ r a t e = r a n d o m ( 10 , 80 )
4:
   set m i s s i n g _ l e n = w i n d o w _ s i z e m i s s i n g _ r a t e / / 100
5:
   compute d e c o d e r _ i n p u t by m i s s i n g _ r a t e and t o k e n _ l e n
6:
   compute Y f a k e by G e n e r a t o r G with Y f a k e = G ( y k n o w n , c o v k n o w n , d e c o d e r _ i n p u t , c o v m i s s i n g )
7:
   update G e n e r a t o r and D i s c r i m i n a t o r
8:
end for

4. Results

4.1. Evaluation Indicators

There are generally two types of evaluation indicators for time series imputation results. The first is the evaluation indicators using regression models, such as MAE and MSE. The other is to use correlation coefficients, such as cosine similarity, to calculate time series as vectors. We evaluated the time series imputation results with reference to the root mean square error (RMSE) and Pearson correlation coefficient (PCC). The RMSE and Pearson correlation coefficient formulas are as follows:
R M S E ( x ^ , x ) = 1 m i = 1 m ( x ^ i x i )
P C C s ( x ^ , x ) = N x ^ i x i x i ^ x i N x ^ i 2 ( x ^ i ) 2 N x i 2 ( x i ) 2
The smaller the root mean square error, the better, and the larger the Pearson correlation coefficient, the better. Therefore, we present the comprehensive evaluation indicators as follows:
p _ R M S E = R M S E ( x ^ , x ) P C C s

4.2. Implementation Details

We used an Nvidia RTX3080 to train the model for 100 epochs on the dataset for experimentation and result comparison. The discriminator network was trained using the RMSProp algorithm, and the generator network was trained using the Adam algorithm. For both datasets, 80% was used as the training set, 10% was used as the validation set, and 10% was used as the test set.

4.3. Performance Comparison

Figure 4 shows the actual imputation effect of our proposed model on the electricity dataset. The upper half is the time series imputation value, the lower half is the time series real value, and the red dots describe the missing data. Under the high missing rate, our model can still complete high-quality time series imputation, which fully demonstrates the performance of the informer-WGAN model. We also applied the informer-WGAN model on the CMG dataset, and Figure 5 shows the imputation results on the CMG dataset at the missing rates of 20% and 50%. Our proposed method revealed trends in the CMG time series. The pRMSE was 0.583 at the 20% missing rate and still reached 0.729 at the 80% missing rate. This shows that informer-WGAN is suitable for CMG time series imputation.
Figure 6 and Figure 7 show the frequency hist of the imputation results and real-time series on the electricity dataset and ETTh1 dataset. The imputation series demonstrates obvious similarity with the real-time series. The frequency hist shows the effect of our imputation model.
Figure 8 shows the actual results of different imputation methods on the ETTh1 dataset at missing rates of 50% and 80%. Obviously, our informer-WGAN performed better than the other imputation methods such as ARIMA and informer, especially when the missing rate was high.
In order to show the actual effect of the model proposed, we chose and compared the traditional data imputation methods, including the ARIMA, the AST model, and the original informer, to compare the data imputation of each model under different missing rates. We also modified the original AST model and compared its effect, and we named it as ST-WGAN. The supplementary results were quantitatively evaluated by the p_RMSE value proposed in the previous section.
Table 2 and Table 3 show the imputation results of our informer-WGAN model on the electricity and ETTh1 datasets. We chose five indicators to evaluate the distance and the Pearson correlation coefficient to evaluate the similarity between the imputaion results and the real data. With the increase in the missing rate, our model demonstrated a slight decrease with respect to all six indicators, which fit our expectation. This means all six indicators can reveal the effect of the imputation task well.
Table 4 and Table 5 show the comparison results of our informer-WGAN model and the other four imputation models on two datasets. Our informer-WGAN model showed a better effect with respect to the different indicators.
As shown in Table 6 and Table 7, the informer-WGAN model we proposed achieved the best results on both the electricity dataset and the ETTh1 dataset. The traditional ARIMA method had poor imputation performance. Compared with the deep learning method, the traditional method had limited utilization of the temporal features, resulting in poor imputation results. However, the accuracy of the AST method decreased significantly when the missing rate increased. Because the sparse transformer model adopted a step-by-step method during decoding (that is, the value of the next time was based on the imputation value of the previous time), it could not be forwarded at one time. The operation was obtained, and the problem of error accumulation occurred. Therefore, the AST model is not suitable for imputation problems with high missing rates. In the informer model, the concept of generative decoder is proposed, and all the imputation was completed through one forward calculation process. The model we proposed also adopts a similar design, which makes the model suitable for imputation under high missing rates. According to the p _ R M S E formula in the previous section, when the quality of the generated sequence is poor, due to the introduction of the Pearson correlation coefficient, the performance gap of the model is more significant.
Table 8 and Table 9 show the RMSE and Pearson correlation coefficient of the imputation results and the ground truth on the electricity and ETTh1 datasets. We note that in the table, the ARIMA algorithm obtained the best RMSE value under the 20% missing rate, but its actual imputation result was not close to the true value, so the Pearson correlation coefficient result was poor. The PRMSE value obtained by combining these two statistics was more conducive to obtaining time series imputation results that met the requirements.

5. Discussion

5.1. Performance Comparison

5.1.1. Informer Discriminator vs. Linear Discriminator

In order to verify the influence of our proposed informer decoder-based discriminator on the multi-dimensional time series imputation results, we used the fully connected layer of AST and the informer encoder network we proposed as the discriminator of the model and conducted experiments on the electricity dataset. The experimental results (Table 10) show that when using the informer encoder, the time series imputation at each missing rate was better than using the fully connected layer as the discriminator, which had a significant effect on training a better time series imputation model. The goal of the discriminator is to fit a likelihood function that evaluates the imputation results. When the informer is used as the discriminator, it can make full use of the characteristics of the time series and learn more reasonable evaluation criteria than the linear discriminator. Therefore, the model of the discriminator is the practical choice.

5.1.2. Random Training Process

We also conducted studies on a randomized training process to verify the effect of this design on improving model performance. The experimental results are shown in the Table 11. The experiment used for comparison only trained the time series data with a missing rate of 20%. It was foreseeable that the performance of the model under the high missing rate was significantly worse than the random training process. For the informer-WGAN model, as the missing rate increased, the gap was more significant. Not only that, but under the 20% missing rate, the model using the random training process was also significantly better than the model trained with a fixed missing rate. The results show that random training made the model more focused on the time series features, while fixed missing rate training made it easier for the model stuck in a local optimum.

6. Conclusions

We proposed a new network model for multidimensional time series imputation based on the informer and WGAN methods. We applied the random missing rate training process to train our network in order to obtain a suitable model for the time series imputation task with high missing rates. The results of the imputation experiments on real-world datasets demonstrated the effectiveness of this model in time series imputation, being superior to the ARIMA model, the AST model, and the informer model. At the same time, according to the experimental results, we also verified that the self-attention decoder discriminator was better than the linear discriminator, implying the necessity of the selection of the discriminator model. Our next step is to optimize the informer model used by the generator to improve the running efficiency of the model.

Author Contributions

Conceptualization, Y.Q.; methodology, Y.Q.; software, Y.Q.; validation, S.Z. and R.W.; formal analysis, Y.Q.; investigation, Y.Q., S.Z. and R.W.; resources, R.W., L.T. and B.Z.; data curation, Y.Q. and R.W.; writing—original draft preparation, Y.Q.; writing—review and editing, S.Z.; visualization, Y.Q.; supervision, R.W.; project administration, Y.Q.; funding acquisition, L.T. and B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Civil Aerospace Technology Advance Research Program (A0201) of the State Administration of Science, Technology and Industry for National Defense.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data needed to reproduce these findings cannot be shared at this time, as these data are also part of further research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chatfield, C. The Analysis of Time Series: An Introduction; Chapman and Hall: London, UK; CRC: Boca Raton, FL, USA, 2003. [Google Scholar]
  2. Luo, Y.; Zhang, Y.; Cai, X.; Yuan, X. E2gan: End-to-end generative adversarial network for multivariate time series imputation. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3094–3100. [Google Scholar]
  3. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  4. Schafer, J.L. Analysis of Incomplete Multivariate Data; CRC Press: Boca Raton, FL, USA, 1997. [Google Scholar]
  5. Contreras-Reyes, J.E.; Palma, W. Statistical analysis of autoregressive fractionally integrated moving average models in R. Comput. Stat. 2013, 28, 2309–2331. [Google Scholar] [CrossRef] [Green Version]
  6. Wu, S.; Xiao, X.; Ding, Q.; Zhao, P.; Wei, Y.; Huang, J. Adversarial sparse transformer for time series forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17105–17115. [Google Scholar]
  7. Rahman, M.G.; Islam, M.Z. Missing value imputation using decision trees and decision forests by splitting and merging records: Two novel techniques. Knowl.-Based Syst. 2013, 53, 51–65. [Google Scholar] [CrossRef]
  8. Folguera, L.; Zupan, J.; Cicerone, D.; Magallanes, J.F. Self-organizing maps for imputation of missing data in incomplete data matrices. Chemom. Intell. Lab. Syst. 2015, 143, 146–151. [Google Scholar] [CrossRef]
  9. Längkvist, M.; Karlsson, L.; Loutfi, A. A review of unsupervised feature learning and deep learning for time-series modeling. Pattern Recognit. Lett. 2014, 42, 11–24. [Google Scholar] [CrossRef] [Green Version]
  10. Salinas, D.; Flunkert, V.; Gasthaus, J.; Januschowski, T. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
  11. Pwasong, A.; Sathasivam, S. Forecasting comparisons using a hybrid ARFIMA and LRNN models. Commun. Stat.-Simul. Comput. 2018, 47, 2286–2303. [Google Scholar] [CrossRef]
  12. Liu, C.; Hoi, S.C.; Zhao, P.; Sun, J. Online arima algorithms for time series prediction. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  13. Zhang, J.; Man, K.F. Time series prediction using RNN in multi-dimension embedding phase space. In Proceedings of the SMC’98 Conference Proceedings—1998 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No. 98CH36218), San Diego, CA, USA, 14 October 1998; Volume 2, pp. 1868–1873. [Google Scholar]
  14. Lai, G.; Chang, W.C.; Yang, Y.; Liu, H. Modeling long-and short-term temporal patterns with deep neural networks. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 95–104. [Google Scholar]
  15. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  16. Luo, Y.; Cai, X.; Zhang, Y.; Xu, J.; Yuan, X. Multivariate time series imputation with generative adversarial networks. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  17. Yoon, J.; Jarrett, D.; Van der Schaar, M. Time-series generative adversarial networks. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  18. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  19. Correia, G.M.; Niculae, V.; Martins, A.F. Adaptively sparse transformers. arXiv 2019, arXiv:1909.00015. [Google Scholar]
  20. Martins, A.; Astudillo, R. From softmax to sparsemax: A sparse model of attention and multi-label classification. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; pp. 1614–1623. [Google Scholar]
  21. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21), Virtual, 2–9 February 2021. [Google Scholar]
Figure 1. Informer-WGAN structure.
Figure 1. Informer-WGAN structure.
Algorithms 15 00252 g001
Figure 2. Generator structure.
Figure 2. Generator structure.
Algorithms 15 00252 g002
Figure 3. Discriminator structure.
Figure 3. Discriminator structure.
Algorithms 15 00252 g003
Figure 4. Imputation results of informer-WGAN method on electricity dataset at missing rates of 20% and 50%.
Figure 4. Imputation results of informer-WGAN method on electricity dataset at missing rates of 20% and 50%.
Algorithms 15 00252 g004
Figure 5. Imputation results of informer-WGAN method on CMG dataset at missing rates of 20% and 50%.
Figure 5. Imputation results of informer-WGAN method on CMG dataset at missing rates of 20% and 50%.
Algorithms 15 00252 g005
Figure 6. Frequency hist of imputation results and real-time series on electrity dataset.
Figure 6. Frequency hist of imputation results and real-time series on electrity dataset.
Algorithms 15 00252 g006
Figure 7. Frequency hist of imputation results and the real time series on ETTh1 dataset.
Figure 7. Frequency hist of imputation results and the real time series on ETTh1 dataset.
Algorithms 15 00252 g007
Figure 8. Imputation results of different methods on ETTh1 dataset at missing rates of 50% and 80%.
Figure 8. Imputation results of different methods on ETTh1 dataset at missing rates of 50% and 80%.
Algorithms 15 00252 g008
Table 1. Dataset data, where F represents the sampling frequency, L represents the number of samples, and D represents the data dimension.
Table 1. Dataset data, where F represents the sampling frequency, L represents the number of samples, and D represents the data dimension.
DatasetFLD
electricityhourly32,304370
ETTh1hourly17,4206
Table 2. The results of the informer-WGAN on the electricity dataset.
Table 2. The results of the informer-WGAN on the electricity dataset.
Missing Ratemaemsermsemapemspep_corr
20%0.826156.709042.393740.199310.440890.95462
40%0.869587.502902.611230.213570.513830.95277
60%0.899817.642852.631680.236510.709460.94944
80%0.972508.133612.725100.271410.793670.93473
Table 3. The results of the informer-WGAN on the ETTh1 dataset.
Table 3. The results of the informer-WGAN on the ETTh1 dataset.
Missing Ratemaemsermsemapemspep_corr
20%1.9083510.734793.272810.534579.757830.77216
40%1.9712612.075603.471730.5400610.141370.75321
60%2.0137612.425903.523900.5841311.622960.74823
80%2.0550312.640883.554640.5770910.304200.74734
Table 4. The imputation results of different models on electricity dataset at a missing rate of 60%.
Table 4. The imputation results of different models on electricity dataset at a missing rate of 60%.
Imputation Modelmaemsermsemapemspep_corr
ARIMA2.4768931.010614.436820.200560.062200.86310
AST2.22961181.734609.4698791.1547611.272900.48822
Informer1.7475019.679884.216870.610753.345360.84613
ST-WGAN1.3594860.821385.7733260.860768.811880.84930
Informer-WGAN0.899697.647342.613950.236240.705430.94789
Table 5. The imputation results of different model on electricity dataset at the missing rate of 60%.
Table 5. The imputation results of different model on electricity dataset at the missing rate of 60%.
Imputation Modelmaemsermsemapemspep_corr
ARIMA2.4803419.274674.350740.524694.694850.65881
AST7.8729594.891179.707222.4460514.578770.38196
Informer2.1148811.825243.436850.594899.800830.74467
ST-WGAN2.2320721.707394.606130.6964424.207820.65263
Informer-WGAN2.0143912.374963.514920.5691010.934390.74725
Table 6. The PRMSE (the smaller, the better) results of the informer-WGAN and other imputation results on electricity dataset.
Table 6. The PRMSE (the smaller, the better) results of the informer-WGAN and other imputation results on electricity dataset.
Missing RateARIMAInfomerInformer-WGAN
20%2.7111.7911.119
40%4.2471.8301.157
60%4.6122.0191.188
80%12.6092.0621.283
Table 7. The PRMSE (the smaller, the better) results of the informer-WGAN and other imputation results on ETTh1 dataset.
Table 7. The PRMSE (the smaller, the better) results of the informer-WGAN and other imputation results on ETTh1 dataset.
Missing RateARIMAASTInfomerInformer-WGAN
20%2.7092.4772.3112.309
40%3.0367.4872.3392.338
60%3.1569.9632.4382.437
80%3.73619.5672.5742.569
Table 8. The RMSE/Pearson correlation results of the informer-WGAN and other imputation results on electricity dataset.
Table 8. The RMSE/Pearson correlation results of the informer-WGAN and other imputation results on electricity dataset.
Missing RateARIMAInfomerInformer-WGAN
20%0.937/0.3451.337/0.7471.029/0.920
40%1.081/0.2541.348/0.7371.055/0.911
60%0.858/0.1861.404/0.6961.073/0.902
80%1.132/0.0891.430/0.6941.130/0.880
Table 9. The RMSE/Pearson correlation results of the informer-WGAN and other imputation results on ETTh1 dataset.
Table 9. The RMSE/Pearson correlation results of the informer-WGAN and other imputation results on ETTh1 dataset.
Missing RateARIMAInfomerInformer-WGAN
20%2.066/0.7621.795/0.7731.796/0.774
40%2.240/0.7371.804/0.7671.847/0.756
60%2.311/0.7321.832/0.7501.867/0.751
80%2.439/0.6521.903/0.7391.883/0.748
Table 10. The ablation experiment results (PRMSE) of different discriminators on electricity dataset.
Table 10. The ablation experiment results (PRMSE) of different discriminators on electricity dataset.
Missing RateInformer-WGAN
(Linear Discriminator)
Informer-WGAN
(Informer Discriminator)
20%1.7101.119
40%1.7081.157
60%1.7981.188
80%1.8661.283
Table 11. The ablation experiment results (PRMSE) of different training methods on electricity dataset.
Table 11. The ablation experiment results (PRMSE) of different training methods on electricity dataset.
Missing RateInformer-WGAN (20% Train)Informer-WGAN (Random Train)
20%1.5601.119
40%2.3261.157
60%2.5821.188
80%2.6891.283
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qian, Y.; Tian, L.; Zhai, B.; Zhang, S.; Wu, R. Informer-WGAN: High Missing Rate Time Series Imputation Based on Adversarial Training and a Self-Attention Mechanism. Algorithms 2022, 15, 252. https://doi.org/10.3390/a15070252

AMA Style

Qian Y, Tian L, Zhai B, Zhang S, Wu R. Informer-WGAN: High Missing Rate Time Series Imputation Based on Adversarial Training and a Self-Attention Mechanism. Algorithms. 2022; 15(7):252. https://doi.org/10.3390/a15070252

Chicago/Turabian Style

Qian, Yufan, Limei Tian, Baichen Zhai, Shufan Zhang, and Rui Wu. 2022. "Informer-WGAN: High Missing Rate Time Series Imputation Based on Adversarial Training and a Self-Attention Mechanism" Algorithms 15, no. 7: 252. https://doi.org/10.3390/a15070252

APA Style

Qian, Y., Tian, L., Zhai, B., Zhang, S., & Wu, R. (2022). Informer-WGAN: High Missing Rate Time Series Imputation Based on Adversarial Training and a Self-Attention Mechanism. Algorithms, 15(7), 252. https://doi.org/10.3390/a15070252

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop