Predicting the Remaining Time before Earthquake Occurrence Based on Mel Spectrogram Features Extraction and Ensemble Learning
Abstract
:1. Introduction
2. Data Preprocessing
2.1. Datasets
2.2. Removing Noise
2.3. Selecting Features
3. Earthquake Prediction Models
3.1. The Deep Neural Network Model
- Input Layer. The input of the model is a tensor with the shape of (1, N), where N represents the feature dimension of the earthquake signal data. We introduced the mel spectrogram in Section 2.2 of the article, and here N = 30.
- Batch Normalization Layer. In the second layer of the model, we employed a batch normalization layer to normalize the input data. Batch normalization is a technique that improves the training performance of neural networks. It can accelerate mode convergence, enhance training stability, and reduce the sensitivity of the model to the initialization of weights. The basic idea of batch normalization is to standardize each feature on each mini-batch of data so that the mean of the data is 0 and the variance is 1. Specifically, by calculating the mean and variance of the mini-batch, we can perform centering and standardization on the input data , obtaining the normalized data. Subsequently, the normalized data are scaled and shifted using the learnable scale factor and shift factor to obtain the final output result . This process helps to improve the training performance and stability of neural networks and reduce the reliance on parameter initialization. Batch normalization applies the following transformation to each feature i of the input data.
- LSTM Layer. We employed a long short-term memory (LSTM) network to capture the long-term and short-term temporal dependencies in the earthquake signal data. LSTM is a special type of recurrent neural network (RNN) that addresses the vanishing and exploding gradient problems that traditional RNN may encounter when processing long sequences by using gating units. This makes LSTM more suitable for handling data with complex temporal dependencies, such as earthquake signals. In this model, we used an LSTM layer with 128 neurons to fully extract the temporal information in the input data. To further process the temporal features in subsequent layers, we set the return-sequences = True parameter, allowing the LSTM layer to output a sequence with the same length as the input. In this way, the LSTM layer can capture not only short-term fluctuations in the earthquake signal but also long-term trends in the earthquake sequence. By combining the LSTM layer and subsequent convolutional layers, our model can better capture and understand the dynamic characteristics of earthquake signal data, providing a more accurate and reliable basis for earthquake prediction.
- 1D-CNN Branch. After the LSTM layer, we constructed a subnetwork containing two 1D convolutional layers to extract local features from the earthquake signal. The first 1D convolutional layer has 128 convolution kernels, a kernel size of 2, and uses the ReLU activation function. The second 1D convolutional layer has 256 convolution kernels, a kernel size of 2, and also uses the ReLU activation function. Next, we employed a Gaussian dropout layer to prevent overfitting, and flattened the result for subsequent processing. Flattening is a data preprocessing operation in neural networks used to convert a multi-dimensional tensor into a one-dimensional tensor. This operation is very common in convolutional neural networks (CNN) because the feature maps usually have a multi-dimensional structure after multiple layers of convolution and pooling operations. In order to connect these feature maps to subsequent fully connected layers for further processing, we need to flatten them into a one-dimensional vector. Specifically, the operation of the flatten layer can be expressed with the following formula:
- 2D-CNN Branch. It is worth noting that after the input layer, we constructed a subnetwork consisting of a fully connected layer and multiple 2D convolutional layers. First, the data were reshaped into a tensor with a shape of (16, 16, 1). Next, we built two 2D convolutional layers with different kernel sizes, and then concatenated the outputs of the two 2D convolutional layers. Finally, the result was flattened for subsequent processing. 2D-CNN have certain advantages when processing time-series data, as they can capture local features in both the time and frequency domains simultaneously. Specifically, 2D-CNN can identify specific patterns at different times and frequencies, revealing spatial (time-domain) and frequency features in the signal. This is significant for analyzing complex time-series data, such as earthquake signals. Using convolution kernels of different sizes can help the model capture features at different scales. Smaller kernels (e.g., 5 × 5) can capture local, detailed features in the signal, while larger kernels (e.g., 7 × 7) can capture a broader spatial range, thus identifying larger-scale patterns in the signal. This multi-scale feature extraction method helps improve the model’s performance and generalization capabilities. Additionally, we used a technique called “dilated convolution” to increase the receptive field range. By increasing the dilation rate, the model can expand the receptive field range without increasing the kernel size. This helps capture more extensive contextual information while maintaining relatively low computational complexity. Finally, the feature maps after the convolution operation are concatenated through the concatenate layer, combining the different scale features captured by the two groups of convolutional layers. In summary, by using multi-scale convolution kernels and feature fusion strategies, the 2D-CNN branch can simultaneously capture rich features in the time and frequency domains, providing strong support for earthquake prediction tasks.
- Feature Concatenation Layer. After processing the outputs of the 1D-CNN and 2D-CNN branches, we concatenate them, achieving feature fusion. This design aims to combine the advantages of 1D-CNN and 2D-CNN to fully extract the features of the earthquake signal data.
- Dense Layer. After the feature fusion, we added a fully connected layer with 128 neurons and a ReLU activation function for further processing of the feature information.
- Outputs Layer. The model has three outputs, which are used to predict the time to failure (TTF), time since failure (TSF), and a binary classification result (to determine whether TTF is less than 0.5 s). The TTF output uses a fully connected layer with one neuron and a ReLU activation function. The TSF output also employs a fully connected layer with one neuron. For the binary classification output, a fully connected layer with one neuron and a sigmoid activation function is used.
3.2. Stacking
4. Experiments
4.1. Model Training
4.2. Experimental Results of Model Performance
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Aguirre, L.; Bataille, K.; Novoa, C.; Peña, C.; Vera, F. Kinematics of subduction processes during the earthquake cycle in Central Chile. Seismol. Res. Lett. 2019, 90, 1779–1791. [Google Scholar] [CrossRef]
- Hammond, A.L. Earthquake prediction and control. Science 1971, 173, 316. [Google Scholar] [CrossRef]
- Healy, J.H. Recent highlights and future trends in research on earthquake prediction and control. Rev. Geophys. 1975, 13, 361–364. [Google Scholar] [CrossRef]
- Scholz, C. A physical interpretation of the Haicheng earthquake prediction. Nature 1977, 267, 121–124. [Google Scholar] [CrossRef] [PubMed]
- Chen, Y.; Xue, R.; Zhang, Y. House price prediction based on machine learning and deep learning methods. In Proceedings of the 2021 International Conference on Electronic Information Engineering and Computer Science (EIECS), Changchun, China, 23–26 September 2021; pp. 699–702. [Google Scholar] [CrossRef]
- Mehtab, S.; Sen, J.; Dutta, A. Stock Price Prediction Using Machine Learning and LSTM-Based Deep Learning Models. arXiv 2020, arXiv:abs/2009.10819. [Google Scholar] [CrossRef]
- Ma, N.; Yin, H.; Wang, K. Prediction of the Remaining Useful Life of Supercapacitors at Different Temperatures Based on Improved Long Short-Term Memory. Energies 2023, 16, 5240. [Google Scholar] [CrossRef]
- Li, W.; Zhang, L.C.; Wu, C.H.; Wang, Y.; Cui, Z.; Niu, C. A data-driven approach to RUL prediction of tools. Adv. Manuf. 2023, 1–13. [Google Scholar] [CrossRef]
- Laurenti, L.; Tinti, E.; Galasso, F.; Franco, L.; Marone, C. Deep learning for laboratory earthquake prediction and autoregressive forecasting of fault zone stress. Earth Planet. Sci. Lett. 2022, 598, 117825. [Google Scholar] [CrossRef]
- Asim, K.; Martínez-Álvarez, F.; Basit, A.; Iqbal, T. Earthquake magnitude prediction in Hindukush region using machine learning techniques. Nat. Hazards 2017, 85, 471–486. [Google Scholar] [CrossRef]
- Cao, C.; Wu, X.; Yang, L.; Zhang, Q.; Wang, X.; Yuen, D.A.; Luo, G. Long short-term memory networks for pattern recognition of synthetical complete earthquake catalog. Sustainability 2021, 13, 4905. [Google Scholar] [CrossRef]
- Wang, K.; Johnson, C.W.; Bennett, K.C.; Johnson, P.A. Predicting fault slip via transfer learning. Nat. Commun. 2021, 12, 7319. [Google Scholar] [CrossRef] [PubMed]
- Wang, X.; Wang, Z.; Wang, J.; Miao, P.; Dang, H.; Li, Z. Machine learning based ground motion site amplification prediction. Front. Earth Sci. 2023, 11, 1053085. [Google Scholar] [CrossRef]
- Sang, K.H.; Yin, X.Y.; Zhang, F.C. Machine learning seismic reservoir prediction method based on virtual sample generation. Pet. Sci. 2021, 18, 1662–1674. [Google Scholar] [CrossRef]
- Feng, T.; Zhang, M.; Xu, L.; Wu, J.; Fang, L. Machine learning-based earthquake catalog and tomography characterize the middle-northern section of the Xiaojiang fault zone. Seismol. Soc. Am. 2022, 93, 2484–2497. [Google Scholar] [CrossRef]
- Wrona, T.; Pan, I.; Gawthorpe, R.L.; Fossen, H. Seismic facies analysis using machine learning. Geophysics 2018, 83, O83–O95. [Google Scholar] [CrossRef]
- Rouet-Leduc, B.; Hulbert, C.; Lubbers, N.; Barros, K.; Humphreys, C.J.; Johnson, P.A. Machine learning predicts laboratory earthquakes. Geophys. Res. Lett. 2017, 44, 9276–9282. [Google Scholar] [CrossRef]
- Hulbert, C.; Rouet-Leduc, B.; Johnson, P.A.; Ren, C.X.; Rivière, J.; Bolton, D.C.; Marone, C. Similarity of fast and slow earthquakes illuminated by machine learning. Nat. Geosci. 2019, 12, 69–74. [Google Scholar] [CrossRef]
- Zhang, W.; Yu, J.; Zhao, A.; Zhou, X. Predictive model of cooling load for ice storage air-conditioning system by using GBDT. Energy Rep. 2021, 7, 1588–1597. [Google Scholar] [CrossRef]
- Aymerich, E.; Sias, G.; Pisano, F.; Cannas, B.; Fanni, A.; the-JET-Contributors. CNN disruption predictor at JET: Early versus late data fusion approach. Fusion Eng. Des. 2023, 193, 113668. [Google Scholar] [CrossRef]
- Hu, R.; Xiang, S. CNN prediction based reversible data hiding. IEEE Signal Process. Lett. 2021, 28, 464–468. [Google Scholar] [CrossRef]
- Allam, J.P.; Samantray, S.; Sahoo, S.P.; Ari, S. A deformable CNN architecture for predicting clinical acceptability of ECG signal. Biocybern. Biomed. Eng. 2023, 43, 335–351. [Google Scholar] [CrossRef]
- Johnson, P.A.; Rouet-Leduc, B.; Pyrak-Nolte, L.J.; Beroza, G.C.; Marone, C.J.; Hulbert, C.; Howard, A.; Singer, P.; Gordeev, D.; Karaflos, D.; et al. Laboratory earthquake forecasting: A machine learning competition. Proc. Natl. Acad. Sci. USA 2021, 118, e2011362118. [Google Scholar] [CrossRef]
- Cao, J.; Li, Z.; Li, J. Financial time series forecasting model based on CEEMDAN and LSTM. Phys. A Stat. Mech. Appl. 2019, 519, 127–139. [Google Scholar] [CrossRef]
- Sagheer, A.; Kotb, M. Time series forecasting of petroleum production using deep LSTM recurrent networks. Neurocomputing 2019, 323, 203–213. [Google Scholar] [CrossRef]
- Taghizadeh-Mehrjardi, R.; Schmidt, K.; Amirian-Chakan, A.; Rentschler, T.; Zeraatpisheh, M.; Sarmadian, F.; Valavi, R.; Davatgar, N.; Behrens, T.; Scholten, T. Improving the spatial prediction of soil organic carbon content in two contrasting climatic regions by stacking machine learning models and rescanning covariate space. Remote Sens. 2020, 12, 1095. [Google Scholar] [CrossRef]
- Shaw, R.; Lokshin, A.E.; Miller, M.C.; Messerlian-Lambert, G.; Moore, R.G. Stacking machine learning algorithms for biomarker-based preoperative diagnosis of a pelvic mass. Cancers 2022, 14, 1291. [Google Scholar] [CrossRef]
- Chatzimparmpas, A.; Martins, R.M.; Kucher, K.; Kerren, A. StackGenVis: Alignment of data, algorithms, and models for stacking ensemble learning using performance metrics. IEEE Trans. Vis. Comput. Graph. 2020, 27, 1547–1557. [Google Scholar] [CrossRef]
- Tian, C.; Zheng, M.; Zuo, W.; Zhang, B.; Zhang, Y.; Zhang, D. Multi-stage image denoising with the wavelet transform. Pattern Recognit. 2023, 134, 109050. [Google Scholar] [CrossRef]
- Baldazzi, G.; Sulas, E.; Urru, M.; Tumbarello, R.; Raffo, L.; Pani, D. Wavelet denoising as a post-processing enhancement method for non-invasive foetal electrocardiography. Comput. Methods Programs Biomed. 2020, 195, 105558. [Google Scholar] [CrossRef]
- Wang, L.; Zheng, W.; Ma, X.; Lin, S. Denoising speech based on deep learning and wavelet decomposition. Sci. Program. 2021, 2021, 8677043. [Google Scholar] [CrossRef]
- Renfei, T.; Tao, L.; Min, O. Method for wavelet denoising of multi-angle prestack seismic data. Acta Geophys. 2023, 71, 1515–1524. [Google Scholar] [CrossRef]
- Sui, Y.; Ma, J. A nonstationary sparse spike deconvolution with anelastic attenuation. Geophysics 2019, 84, R221–R234. [Google Scholar] [CrossRef]
- Gao, Z.; Hu, S.; Li, C.; Chen, H.; Jiang, X.; Pan, Z.; Gao, J.; Xu, Z. A deep-learning-based generalized convolutional model for seismic data and its application in seismic deconvolution. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–17. [Google Scholar] [CrossRef]
- Mirel, M.; Cohen, I. Multichannel Semi-blind Deconvolution (MSBD) of seismic signals. Signal Process. 2017, 135, 253–262. [Google Scholar] [CrossRef]
- Ravasi, M.; Selvan, T.; Luiken, N. Stochastic multi-dimensional deconvolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
- Li, Y.; Mei, L.; Li, R.; Wu, C. Using noise level to detect frame repetition forgery in video frame rate up-conversion. Future Internet 2018, 10, 84. [Google Scholar] [CrossRef]
- Sahu, S.; Singh, H.V.; Kumar, B.; Singh, A.K. De-noising of ultrasound image using Bayesian approached heavy-tailed Cauchy distribution. Multimed. Tools Appl. 2019, 78, 4089–4106. [Google Scholar] [CrossRef]
- Xie, T.; Zheng, X.; Zhang, Y. Seismic facies analysis based on speech recognition feature parameters. Geophysics 2017, 82, O23–O35. [Google Scholar] [CrossRef]
- Zhou, Q.; Tong, G.; Xie, D.; Li, B.; Yuan, X. A seismic-based feature extraction algorithm for robust ground target classification. IEEE Signal Process. Lett. 2012, 19, 639–642. [Google Scholar] [CrossRef]
- Alghamdi, N.S.; Zakariah, M.; Hoang, V.T.; Elahi, M.M. Neurogenerative Disease Diagnosis in Cepstral Domain Using MFCC with Deep Learning. Comput. Math. Methods Med. 2022, 2022, 4364186. [Google Scholar] [CrossRef]
- Noda, J.J.; Travieso-González, C.M.; Sanchez-Rodriguez, D.; Alonso-Hernández, J.B. Acoustic classification of singing insects based on MFCC/LFCC fusion. Appl. Sci. 2019, 9, 4097. [Google Scholar] [CrossRef]
- Fahad, M.S.; Deepak, A.; Pradhan, G.; Yadav, J. DNN-HMM-based speaker-adaptive emotion recognition using MFCC and epoch-based features. Circuits Syst. Signal Process. 2021, 40, 466–489. [Google Scholar] [CrossRef]
- Borwankar, S.; Verma, J.P.; Jain, R.; Nayyar, A. Improvise approach for respiratory pathologies classification with multilayer convolutional neural networks. Multimed. Tools Appl. 2022, 81, 39185–39205. [Google Scholar] [CrossRef] [PubMed]
- Coelho, G.; Matos, L.M.; Pereira, P.J.; Ferreira, A.; Pilastri, A.; Cortez, P. Deep autoencoders for acoustic anomaly detection: Experiments with working machine and in-vehicle audio. Neural Comput. Appl. 2022, 34, 19485–19499. [Google Scholar] [CrossRef]
Index | Feather Name | Feather Description |
---|---|---|
1 | NN_Mels_mean4 | original signal in the 5th Mel frequency band |
2 | NN_Mels_mean5 | original signal in the 6th Mel frequency band |
3 | NN_Mels_mean18 | original signal in the 19th Mel frequency band |
4 | NN_LGBM_Mels_mean9 | original signal in the 10th Mel frequency band |
5 | NN_LGBM_Mels_mean11 | original signal in the 12th Mel frequency band |
6 | NN_LGBM_Mels_mean16 | riginal signal in the 17th Mel frequency band |
7 | NN_Mels_mean_denoise_6 | denoised signal in the 7th Mel frequency band |
8 | NN_Mels_mean_denoise_7 | denoised signal in the 8th Mel frequency band |
9 | NN_Mels_mean_denoise_13 | denoised signal in the 14th Mel frequency band |
Model | Train-MAE | Test-MAE |
---|---|---|
LGB | 1.89379 | 2.27931 |
NN-1D | 1.86133 | 2.28126 |
NN-2D | 1.86481 | 2.27835 |
Stacking | 1.85483 | 2.24514 |
Model | Train-MAE | Test-MAE |
---|---|---|
LGB | 1.89362 | 2.27851 |
NN-1D | 1.85528 | 2.27628 |
NN-2D | 1.86463 | 2.27388 |
Stacking | 1.85374 | 2.23831 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, B.; Xu, T.; Chen, W.; Zhang, C. Predicting the Remaining Time before Earthquake Occurrence Based on Mel Spectrogram Features Extraction and Ensemble Learning. Appl. Sci. 2023, 13, 12268. https://doi.org/10.3390/app132212268
Zhang B, Xu T, Chen W, Zhang C. Predicting the Remaining Time before Earthquake Occurrence Based on Mel Spectrogram Features Extraction and Ensemble Learning. Applied Sciences. 2023; 13(22):12268. https://doi.org/10.3390/app132212268
Chicago/Turabian StyleZhang, Bo, Tao Xu, Wen Chen, and Chongyang Zhang. 2023. "Predicting the Remaining Time before Earthquake Occurrence Based on Mel Spectrogram Features Extraction and Ensemble Learning" Applied Sciences 13, no. 22: 12268. https://doi.org/10.3390/app132212268
APA StyleZhang, B., Xu, T., Chen, W., & Zhang, C. (2023). Predicting the Remaining Time before Earthquake Occurrence Based on Mel Spectrogram Features Extraction and Ensemble Learning. Applied Sciences, 13(22), 12268. https://doi.org/10.3390/app132212268