Next Article in Journal
A Hybrid Image Augmentation Technique for User- and Environment-Independent Hand Gesture Recognition Based on Deep Learning
Previous Article in Journal
Quantum Machine Learning for Credit Scoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Bitcoin Price Volatility Estimator Predictions: A Four-Step Methodological Approach Utilizing Elastic Net Regression

by
Georgia Zournatzidou
1,
Ioannis Mallidis
2,*,
Dimitrios Farazakis
3,4 and
Christos Floros
1
1
Department of Accounting and Finance, Hellenic Mediterranean University, 71410 Heraklion, Greece
2
Department of Statistical and Insurance Science, University of Western Macedonia, 50100 Kozani, Greece
3
Institute of Applied and Computational Mathematics, Foundation for Research and Technology Hellas (FORTH), 71110 Heraklion, Greece
4
Department of Mathematics, University of Western Macedonia, 52100 Kastoria, Greece
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(9), 1392; https://doi.org/10.3390/math12091392
Submission received: 5 March 2024 / Revised: 24 April 2024 / Accepted: 28 April 2024 / Published: 2 May 2024

Abstract

:
This paper provides a computationally efficient and novel four-step methodological approach for predicting volatility estimators derived from bitcoin prices. In the first step, open, high, low, and close bitcoin prices are transformed into volatility estimators using Brownian motion assumptions and logarithmic transformations. The second step determines the optimal number of time-series lags required for converting the series into an autoregressive model. This selection process utilizes random forest regression, evaluating the importance of each lag using the Mean Decrease in Impurity (MDI) criterion and optimizing the number of lags considering an 85% cumulative importance threshold. The third step of the developed methodological approach fits the Elastic Net Regression (ENR) to the volatility estimator’s dataset, while the final fourth step assesses the predictive accuracy of ENR, compared to decision tree (DTR), random forest (RFR), and support vector regression (SVR). The results reveal that the ENR prevails in its predictive accuracy for open and close prices, as these prices may be linear and less susceptible to sudden, non-linear shifts typically seen during trading hours. On the other hand, SVR prevails for high and low prices as these prices often experience spikes and drops driven by transient news and intra-day market sentiments, forming complex patterns that do not align well with linear modelling.

1. Introduction

In the rapidly evolving landscape of financial markets, the accurate prediction of cryptocurrency prices is critical for a wide range of financial stakeholders, due to their highly volatile behaviour.
Volatility estimators, which are critical tools necessary for analysing cryptocurrency price movements, constitute a key component used to quantify the intrinsic uncertainty and risk present in the market [1]. However, their prediction encompasses inherent challenges associated with the stochastic behaviour of financial time series, the high frequency of outlier occurrences, the impact of unforeseen global events which can disrupt established patterns, and their nonstationary behaviour that exhibits periods of high volatility clustering [2].
ENR stands out as a highly effective tool for addressing the challenges emerging from financial time series analysis. With respect to the stochastic behaviour of financial data and through its L2 penalty, ENR stabilizes the model’s predictions by reducing the magnitude of the impacts of correlated predictors, which is particularly useful when data encompass extreme variations [3]. Additionally, by penalizing the size of the coefficients, ENR minimizes the impact of outliers that could affect predictions in more simplified regression models. This adjustment reduces the probability of overfitting to noise and outliers, which is in turn critical for maintaining predictive accuracy [4]. In the context of global unforeseen events, ENR’s focus on core and influential variables allows the model to adapt more effectively. By avoiding overfitting and focusing on significant predictors, the model pertains higher degrees of freedom to adapt to major changes [5]. Finally, for non-stationarity and volatility clustering issues, ENR addresses relevant predictors and reduces the influence of less pertinent volatilities, ensuring continued effectiveness during high-volatility periods, which is a common feature in financial markets [6].
ENR is however, also characterized by significant limitations: (a) ENR’s parameters (alpha and lambda) are not optimized during the model’s training process and must be specified in advance, often requiring extensive cross-validation to determine their optimal values [4]; (b) while ENR handles multicollinearity and feature selection well, it struggles in high-dimensional spaces. This situation can lead to computational difficulties and the need for dimensionality reduction techniques [7]; and (c) although ENR generally offers improved interpretability due to its feature-selection and noise-elimination capabilities, it may lag in predictive accuracy, especially when the underlying model is complex or when balancing bias and variance is challenging. This trade-off between interpretability and accuracy is particularly crucial in applications where accurate predictions are more important than model simplicity and interpretability [8].
To address the first limitation, we employ Python’s RandomizedSearchCV method to optimize hyperparameters (alpha and lambda). This method involves a non-deterministic search that randomly samples parameter values from a defined distribution over a specified number of iterations, avoiding the exhaustive characteristic of traditional grid searches and potentially enhancing the model’s performance significantly [9]. For addressing the second limitation, and as autoregressive lags are key variables in financial time series, we employ the Random Forest Regression (RFR) methodological approach developed by Polyzos and Siriopoulos [10] to optimize the selection of time lags in autoregressive models using an ensemble of decision trees. To this end, and for addressing the third limitation, we provide a comparative analysis of ENR’s accuracy against traditional machine learning regression models such as DTR, RFR, and SVR. DTR offers a straightforward and interpretable model structure, making it a useful baseline for comparison [11], while RFR is known for its robustness and consistently high accuracy across various datasets, serving as an excellent benchmark for evaluating the accuracy of ENR [12]. Finally, SVR excels at handling complex, non-linear relationships, allowing us to assess how well ENR performs compared to these sophisticated techniques in capturing predictive accuracy [13].
The remainder of this paper is organized as follows: A literature review of the predictive models utilized for forecasting high-frequency data is presented in Section 2. Subsequently, the employed methodology is outlined in Section 3. The model’s numerical implementation and derived results are presented in Section 4. Section 5 provides a discussion of the paper’s objectives and insights, while Section 6 concludes with the project’s results and future research perspectives.

2. Literature Review

Predictive models play a critical role guiding trading decisions in financial markets with intraday data. Numerous forecasting methods have been developed to handle the specificities of these markets. Studies focusing on cryptocurrency market trends have led to the utilization of predictive models that mainly leverage machine learning and deep learning techniques focusing on cryptocurrency prices solely [14].
On this basis Long Short-Term Memory (LSTM) networks have proven effective for capturing the complexities of cryptocurrency price fluctuations with Kumar et al. [15] assessing Long Short-Term Memory (LSTM) networks for predicting prices, while emphasizing the model’s ability to adjust to the cryptocurrency price fluctuations. The authors employed numerous market data features such as high, low, open, close, and market cap values. These values were then transformed into a multi-dimensional array to effectively utilize the LSTM’s capability to process sequential data. The dataset included approximately 78,922 rows of cryptocurrency data, with a specific focus on Bitcoin and Ethereum. The LSTM model’s performance was then compared with other models like the autoregressive (AR), the moving average (MA), and the Autoregressive Integrated Moving Average Model (ARIMA), with the LSTM achieving the highest accuracy of about 71%.
Similarly, Jethani et al. [16] employed the LSTM network for the prediction of stock market trends. The authors evaluated the model’s performance compared to traditional models and advanced deep learning techniques. The results obtained from their study indicated that the proposed LSTM model can predict more accurately compared to alternative models, in cases of highly volatile financial time series. The authors employed feature engineering techniques with features like opening, close, high, low prices, and volumes of traded stocks. The dataset employed, involved data from January 2018 to April 2021, covering diversified Indian companies and capturing the impacts of major economic events, including the occurrence of the COVID-19 pandemic.
Lumoring et al. [17] highlighted the significance of machine learning model implementation to predict trends in the cryptocurrency sector and stock market, focusing on the effectiveness of Support Vector Machines (SVM) and Long Short-Term Memory (LSTM) network models. The features considered typically included historical price data and trading volumes. Evaluation metrics such as accuracy and Mean Absolute Percentage Error (MAPE) were used, with LSTM models achieving the highest predictive accuracy.
Finally, Liu [18] compared deep learning techniques with traditional time series analysis in stock market prediction. The results have shown that the LSTM model attains higher predictive accuracy in AAPL stock compared to the classical ARIMA model. The employed dataset spans from September 2013 through to September 2023 and the prediction accuracy was measured using evaluation metrics such as the Mean Squared Error (MSE) and Mean Absolute Error (MAE).
Researchers have also delved into different frameworks of neural networks with the purpose of further improving the forecasting accuracy. The authors focused on stochastic neural networks (SNNs), which encompass stochastic processes that manage to accurately emulate the stochastic nature of the financial markets. As reported by Jay et al. [19], these models are an attempt to predict cryptocurrency price volatility with the use of processed market indicators and systematically outperform the more deterministic and simplistic models. The authors focus on key cryptocurrencies, including Bitcoin, Ethereum, and Litecoin, with the multitude of market indicators namely, price, volume, and even mining difficulty, as features. The dataset being considered ranged from mid-2017 to the end of 2019, while MAPE, MAE, and RMSE have been employed for assessing predictive accuracy. The findings from this study indicated that stochastic models adapted well to market dynamics and outperformed the deterministic models in a consistent manner.
Hemant et al. [20] proposed a method in which the kernel extreme learning machine (KELM) is coupled with a variational autoencoder (VAE) for predicting stock market trends. The approach of this paper is a combination of the CNN technique for extracting features and the VAE model for predictions towards improving the overall accuracy in stock price prediction. This single approach shows great potential in applying advanced machine learning algorithms for improving predictive accuracy. The model employs an exhaustive set of features, including the use of technical indicators and the history of the data transformed through CNN for deep feature extraction, further processed by VAE to add on to its prediction capabilities. The dataset involved daily price movements of stocks over a 2-year time horizon. The model’s accuracy was assessed through MAPE, MAE, and RMSE, with the findings revealing that the combined developed VAE-KELM approach greatly improves the performance of the standard prediction models.
Finally, Charandabi and Kamyar [21] provided a comparative analysis of multiple artificial neural network-based (ANN) approaches to predict cryptocurrency prices, while highlighting the pros and cons of each of the methods presented in terms of the time elapsed and their prediction accuracy. The datasets encompassed a range of historical cryptocurrency price data, with a particular focus on Bitcoin. Overall, their findings suggested that hybrid neural network models could predict cryptocurrency prices and time effectively and accurately, thus providing a foundational approach that supports the use of advanced machine learning techniques in financial market prediction.
The integration of sentiment analysis from social media with historical price data has opened new avenues for predictive accuracy. On this basis, Pathak and Kakkar [22] combine data and sentiment analysis from media using a new sentiment-based neural network model (SBNNM) to improve predictive modelling. Their approach involved extensive feature engineering, integrating sentiment analysis from social media with historical cryptocurrency price data to create a comprehensive feature set. The dataset used comprised real-time and historical data including pricing, volume, and market sentiment extracted from Twitter using advanced natural language processing techniques. The evaluation criteria focused on prediction accuracy and error metrics, with the neural network achieving a prediction accuracy of 77.89%.
Further emphasizing the breadth of methodologies, hybrid approaches involving both numerical and textual data have been employed to predict trends influenced by dynamic market news. Usmani and Shamsi [23] investigated the effectiveness of combining numerical and textual data to predict stock market trends influenced by news, introducing a framework that could also be applied for analysing cryptocurrency markets and emphasizing the increasing role of artificial neural networks (ANN) in improving prediction accuracy through diverse data integration. They employ sophisticated feature engineering, utilizing both structured and unstructured textual features, such as Bag-of-Words and event extraction techniques, alongside numerical data. This integration allows for a deep analysis of the news impact on stock trends. Their findings suggest that neural networks, capable of processing and learning from the complexity of combined data types, offer significant improvements in prediction accuracy, showcasing the potential for these methods in broader financial applications.
Finally, a comprehensive literature review was undertaken by Dopi et al. [24], which focused on the application of machine learning and deep learning techniques in the prediction of cryptocurrency market stock prices. Their study investigated the correlation between the stock and cryptocurrency markets, with a particular focus on the potential and extensive utility of predictive modelling tools in comprehending complex market dynamics. The review highlighted that the researchers predominantly utilize LSTM, MLP, RF, and SVM methods, with MLP showing the best performance at a 71.63% accuracy rate. Feature engineering was extensively used, incorporating both technical and fundamental analysis data, including sentiment analysis from social media. The datasets featured varied widely but commonly included historical price data, volume, and sentiment indicators from news and social media. The evaluation of the models was typically based on accuracy, precision, and loss metrics such as MAE and RMSE, reflecting a rigorous approach to assessing prediction performance. The findings from the review suggest that combining multiple forms of analysis and using advanced machine learning and deep learning techniques can significantly enhance the accuracy of stock price predictions in highly volatile markets like those of cryptocurrencies.
Table 1 provides a critical synthesis of academic research efforts based on the model type that these employs.
The findings from our critical synthesis of the literature highlight several key trends and gaps in the field of cryptocurrency price prediction using machine learning and deep learning techniques. Firstly, the review emphasizes a growing trend in the application of sophisticated machine learning and deep learning models, particularly LSTM, which are renowned for their precision in predicting cryptocurrency prices. Despite their accuracy, these models often lack interpretability and require extensive and complex training procedures. Secondly, it is evident that LSTM models dominate the research landscape, reflecting their effectiveness in capturing the dynamics of cryptocurrency prices. Thirdly, our review revealed a general absence of methods that quantitatively assess the impact of delays associated with the transposed volatility estimator of time series. This indicates a need for improved methodologies that can effectively incorporate temporal delays to enhance prediction accuracy. Fourthly, there appears to be a scarcity of research focused specifically on predicting cryptocurrency volatility estimators.
To address these identified gaps, our paper proposes several approaches:
o
ElasticNet Regression Approach: We propose the development and implementation of an ENR model, with optimized hyperparameter values, that balances computational efficiency with high prediction accuracy, providing a more interpretable model compared to deep learning methods.
o
Two-Step Methodological Framework for Lag Optimization: To tackle the challenge of quantifying the impacts of lags, our study employs a two-step methodology. The first step involves identifying the optimal number of autoregressive delays, which are then incorporated as additional independent variables in the dataset. This integration allows for a more comprehensive examination and enhances the precision of predictions.
o
Volatility Estimators from Bitcoin Prices: We focus on designing effective machine learning models, particularly utilizing volatility estimators derived from Bitcoin prices. This initiative aims to fill the gap in research concerning the prediction of cryptocurrency volatility estimators.
o
Comparative Analysis of Model Accuracy: We provide a comparative analysis of the predictive accuracy of the ENR model against other traditional models such as Decision Trees, Random Forests, and Support Vector Machines.

3. Methodology

This section provides the mathematical analysis employed for deriving the functions quantifying the volatility estimators of open, high, low, and close cryptocurrency prices.
The first volatility estimator employed is the Parkinson Volatility Estimator [25] of Equation (1).
  σ p 2 ^ = p m a x p m i n 2 4 l n 2 ,    
where p m a x = ln H i g h ln ( O p e n ) ,   p m i n = ln L o w ln ( O p e n ) and High, Open, and Low correspond to the High, Open, and Low cryptocurrency price data. In our analysis, we employ this function for exclusively quantifying the volatility estimators of open prices solely. The intuition behind the selection of the Parkinson Volatility Estimator solely for open prices hinges upon the fact that this estimator exhibits a unique ability to capture the market’s immediate reaction to overnight news and events at the opening bell [26].
Regarding high cryptocurrency prices, the second volatility estimator employed involves the Garman–Klass Estimator [27] of Equation (2)
σ G K 2 ^ = 0.5 · ( p m a x p m i n ) 2 ( 2 l n 2 1 ) · p c l o s e 2 ,
where p c l o s e = ln C l o s e l n ( O p e n ) , with Close corresponding to the close cryptocurrency prices. This volatility estimator is specifically tailored for high volatility assets like cryptocurrencies due to its inclusion of the full range of price movements (high, low, and close) within the trading day [27].
The third volatility estimator function employed for low cryptocurrency prices is mathematically expressed through Equation (3) [28]
σ M 2 ^ = 0.274 · σ 1 2 + 0.16 · σ s 2 + 0.365 · σ 3 2 + 0.2 · σ 4 2 ,
where σ 1 2 = 2 [ p m a x p c l o s e 2 + p l o w ] , σ s 2 ^ = p c l o s e 2 , σ 3 2 = 2 ( p m a x p c l o s e p m i n ) p c l o s e , σ 4 2 = ( p m a x p c l o s e ) p m i n 2 l n 2 5 4 and p c l o s e = p c l o s e , p m a x = p m a x , p m i n = p m i n   i f   p c l o s e > 0   a n d   p c l o s e = p c l o s e , p m a x = p m i n , p m i n = p m a x   i f   p c l o s e < 0 .
We selected the above volatility estimator for low cryptocurrency prices since this estimator has the capability to integrate σ 1 2 , σ s 2 ,   σ 3 2   a n d   σ 4 2 , in a single comprehensive measure. This approach is particularly effective in low-volatility scenarios, where traditional estimators might not capture the full extent of market dynamics or might overemphasize the impact of minor price changes [28].
The final fourth volatility estimator employed for close prices is the Rogers–Satchell (RS) estimator [29] of Equation (4).
σ R S 2 ^ = p m a x · p m a x p c l o s e + p m i n · p m i n p c l o s e ,
The reason for selecting this volatility estimator for close prices is mainly attributed to the fact that the estimator captures the entire range of intraday price movements in relation to both the opening and closing prices. Its formulation incorporates the high, low, open, and close prices, enabling it to reflect the volatility surrounding the closing price effectively [29].
Given the selection of the appropriate mathematical functions for quantifying the volatility estimators of cryptocurrency prices, the second step of our approach involves the determination of the optimal number of lags for capturing both the behaviour of the recent volatility coefficient values along with the inherent autocorrelation in the data [30].
We then employ the RFR model to optimize the selection of time-series lags in our predictive model. The approach begins by constructing a forest of decision trees, where each tree is built from a bootstrap sample of the data. The RFR model utilizes the Mean Decrease in Impurity (MDI) criterion to evaluate the importance of each lag in predicting the volatility estimator [10]. The optimal number of lags are the lags resulting to an 85% cumulative importance. This criterion has been empirically selected as it achieves a balance between model complexity and explanatory power, ensuring that the model retains the most significant predictors while avoiding overfitting by excluding less impactful variables.
Having selected the optimal number of lags, we finalized the dataset structure and split the dataset to an 80% train and 20% test set. We then formulated the ENR model and fit the model on the train set. The ENR model constitutes a combined statistical tool of Ridge and Lasso (Least Absolute Shrinkage and Selection Operator) regression, based on the OLS method. Ridge regression, Lasso regression, and Elastic Net regression are techniques used in the field of machine learning and statistics for regularization, which helps in reducing model complexity and preventing overfitting. The linear problem of the form [31] is presented in Equation (5).
y t ^ = β 0 + j = 1 J β j x t j + ε i , t T
Here, the vector x t j represents the value of the independent variable j J at time instance t T and β j is the regression coefficient vector of the independent variables. Then, the sum of squared residuals is defined through Equation (6) as follows:
S S R = t = 1 T y t y t ^ 2 ,
We then estimated the values of the coefficients by using the ordinary least square (OLS) model, and through the minimization of the sum of the squared residuals of Equation (7).
β ^ 0 , β j ^ O L S = a r g m i n β ^ 0 , β j ^ R J + 1 t = 1 T y t y t ^ 2 = a r g m i n β ^ 0 , β j ^ R J + 1 S S R ,
By incorporating penalization techniques into the OLS framework, we gain an understanding of the Ridge and Lasso regression. Ridge regression employs ridge constraints to restrict the size of certain coefficients, thereby regulating the selection of variables [32]. Ridge regression augments the loss function with a penalty equal to the square of the magnitude of the coefficients. By penalizing the inclusion of large coefficients, this method effectively mitigates the complexity of the model and aids in addressing multicollinearity, which occurs when independent variables are significantly correlated. It is particularly useful when dealing with data where the number of predictors (variables) is close to or exceeds the number of observations. The penalty term is controlled by a hyperparameter, often denoted by λ 0 , which determines the extent of regularization. The higher the value of λ , the greater the amount of shrinkage of the coefficients towards zero. The model is developed through Equation (8) as follows:
β ^ 0 , β j ^ R i d g e = a r g m i n β ^ 0 , β j ^ R J + 1 t = 1 T y t y t ^ 2 + λ j = 1 K β j 2 ,
We can notice that the difference between the estimations β ^ 0 , β j ^ O L S and β ^ 0 , β j ^ R i d g e is the l 2 penalty of the form:
β j 2 2 = j = 1 J β j 2 M ,
for M to be the upper bound for the sum of the coefficients. Lasso Regression also adds one more penalty (compared to the ridge regression) to the loss function, but unlike Ridge, the penalty is the absolute value of the magnitude of coefficients. This can lead to some coefficients being exactly zero when the penalty is large enough, effectively performing variable selection and producing models that are sparse [33]. This is particularly useful for models that benefit from variable reduction/selection. Similar to Ridge, the strength of the regularization is controlled by a hyperparameter, λ . Lasso regression satisfies the next optimization problem [34].
β ^ 0 , β j ^ L a s s o = a r g m i n β ^ 0 , β j ^ R J + 1 t = 1 T y t y t ^ 2 + λ j = 1 J β j ,
with l 1 penalty of the form
β j 1 = j = 1 J β j ,
Then, the combination of the two previous regression methods, Ridge and Lasso, yields the elastic net regression [4]. It adds both penalties (the square of the magnitude and the absolute value of the coefficients) to the loss function. This approach aims to leverage the benefits of both Ridge and Lasso regression. It is useful when there are multiple features correlated with each other. Elastic Net has two parameters to control the mix of Ridge and Lasso penalties, which controls the overall strength of the penalties [35]. The model for elastic net regression is then expressed through the following Equation (12) as follows:
β ^ 0 ,   β j ^ E N = a r g m i n β ^ 0 ,   β j ^ R J + 1 t = 1 T y t y t ^ 2 + λ ( 1 α ) 2 j = 1 J β j 2 + λ α j = 1 J β j ,
In order to optimize the selection of the Elastic Net Regression’s (ENR) hyperparameters, α and λ, which dictate the balance between the L1 and L2 penalties, we employ the RandomizedSearchCV method using the Python programming language on the train set. This methodological approach is implemented using libraries such as scikit-learn for the RandomizedSearchCV functionality and NumPy for numerical operations. We randomly select combinations from a predefined grid of α and λ values and evaluate the performance of each combination using k-fold cross-validation. This stochastic sampling method is not only computationally less intensive than an exhaustive grid search, but also provides a practical compromise between thoroughness and efficiency.
Table 2 provides the nomenclature Table of the examined elastic net model parameters.
To this end and for providing a comparative assessment of the ENR’s predictive accuracy, we compare ENR with traditional machine learning regression models such as the DTR, the RFR, and the SVR. Considering the optimally selected number of lags for each volatility estimator, these models are fitted on the 80% train dataset for determining their optimal parameter values and employed on predicting the 20% test set dependent variable value. The criterion for assessing the examined machine learning regression models is the Mean Absolute Error (MAE). The model parameters optimized under the decision tree regression model involve the depth and the leaves of each tree, while for the random forest, the average depth, and leaves of the forest trees [6]. Finally, for the SVR, the parameters optimized involve the regularization parameter C, the error margin epsilon, and the number of support vector regressors [36].
The following Figure 1 summarizes the process map and pseudocode of the employed methodology.

4. Numerical Analysis and Results

Table 3 summarizes the descriptive statistics of the volatility estimators associated with the open, high, low, and close data from September 2014 to January 2024.
All estimators have a data count of 3421 daily observations, illustrating a robust dataset. The mean values indicate the average level of volatility, showing a progression from minimal at open to more substantial at close, suggesting varying volatility levels at different trading times. The standard deviations are quite broad, especially for the close prices, pointing to significant variability and thus potential unpredictability in Bitcoin’s price movements. Minimum values near zero depict days of exceptionally low volatility, whereas the maximum values, which are drastically higher than the means, highlight days of extreme volatility, underscoring the erratic nature of cryptocurrency markets. The percentiles (25%, 50%, and 75%) suggest a right-skewed distribution, indicating that while most days witness lower volatility, a few experiences exceptionally high volatility. These comprehensive data underscore the substantial risk and the necessity for careful risk management in trading and investment strategies concerning Bitcoin.
With respect to the optimal lag selection decisions, the following Figure 2 depicts the lags that lead to a cumulative significance of 85% for each VE_1-4, using the random forest regression methodological approach, while Table 4 summarizes the optimal lags for each VE. As the volatility estimators are daily, we examined 30 lags as our initial lags number.
Having determined the optimal number of lags using the Random Forest Regression model, we proceed with modelling each Bitcoin price volatility estimator using ENR. Initially, the dataset is divided into training and test sets, with 80% allocated for training and the remaining 20% for testing. The ENR, which incorporates both L1 and L2 regularization, is then configured. To optimize the ENR’s hyperparameters efficiently, we employed RandomizedSearchCV. Once the optimal hyperparameters are identified, the next step involves the training of ENR on the training dataset. The trained model is then used to predict the volatility estimators in the test set, considering the identified lags for each volatility estimator as independent variables.
Table 5 provides the summary statistics of the ENR model.
The model results indicate that the intercepts and coefficients vary considerably across different estimators, suggesting unique underlying dynamics in each case. For VE1, the results show minimal coefficient shrinkage (λ = 0.0511) and a nearly pure ridge behavior (α ≈ 0), leading to a low mean absolute error (MAE), indicative of a robust model fit. In contrast, VE2 adopts a more LASSO-like approach (α = 0.9053) with a moderate λ (0.0940), resulting in a higher MAE, which might imply less predictive accuracy. VE 3 exhibits the highest regularization (λ = 0.1116) and a balanced α (0.7369), yet it scores the highest MAE among the models, possibly indicating issues with model suitability or data fit. Finally, VE4 demonstrates the highest degree of coefficient shrinkage (λ = 0.5613) with a predominance of ridge regression characteristics (α = 0.0762), which yields a relatively low MAE. These results highlight the critical importance of tuning the λ and α parameters in elastic net models to balance the trade-off between bias and variance effectively, particularly in financial datasets where volatility estimators can behave unpredictably.
To further evaluate the predictive power of elastic net regression compared to other regression models, we employed the same process for predicting the dependent target variable values of each bitcoin price volatility estimator considering DTR, RFR, and SVR. The derived optimized model parameter values along with their MAE values are summarized in the following Table 6.
The following Table 7 summarizes the MAD values of the examined regression methodologies on the train sets of each one of the examined volatility estimators
The analysis of Table 7 reveals that different regression models exhibit varied effectiveness across the volatility estimators of Bitcoin prices, corresponding to open, high, low, and close values. ENR performs exceptionally well with open and close prices, where the predictive relationships may be more linear and less susceptible to sudden, non-linear shifts typically seen during trading hours. The regularization in ENR helps to prevent overfitting while effectively capturing the linear trends that might govern the open and close price movements. On the other hand, SVR shows superior performance in estimating the volatility of high and low prices as these prices often experience spikes and drops driven by transient news and intra-day market sentiments [37]. SVR, with its capability to handle non-linear relationships through kernel transformations, can capture these dynamics more accurately [38].

5. Discussion

The primary objective of this study is to develop and employ a four-step methodological approach for accurately predicting the volatility estimators of open, high, low, and closed bitcoin prices. The first step of our methodological approach is to transform these prices into volatility estimators using motion assumptions and log price transformations. The second step involves the optimal selection of the number of lags required for accurately capturing the volatile nature of bitcoin prices, while the third step is the formulation of the ENR model on the dataset. The final fourth step provided a comparative analysis of ENR’s predictive accuracy with DTR, RFR, and SVR
Our findings reveal the paramount importance of immediate past volatility in predicting short-term fluctuations, highlighting the unpredictable nature of cryptocurrency markets that are prone to sudden shifts.
In comparison to other models tested in our study, the ENR model accuracy varied depending on the specific volatility estimator being analysed. More specifically, ENR exhibited exceptional performance with Open and Close prices, achieving the lowest Mean Absolute Error (MAE) values among the models for these price types. This performance highlights the model’s ability to capture the more linear and predictable trends at the opening and closing of trading, where the market dynamics are potentially less volatile and more susceptible to long-term factors.
On the other hand, SVR demonstrated superior performance in estimating the volatility of High and Low prices. These price points often encompass abrupt changes due to intra-day news and market sentiments, resulting in complex, non-linear patterns. SVR’s effectiveness in handling these dynamics, due to its capability to model non-linear relationships through kernel transformations, led to more accurate predictions of volatility under conditions of high unpredictability and transient market behaviours.
This analysis indicates that while the ENR effectively prevents overfitting and excellently captures linear relationships, its performance is not uniformly superior across all price volatility estimators. SVR’s superior handling of non-linear dynamics suggests that a combination of models may be necessary to fully address the diverse characteristics of Bitcoin’s price volatility. This comparative analysis underscores the necessity of employing tailored modelling approaches for different aspects of market behaviour, reflecting the multifaceted nature of cryptocurrency volatility.

6. Conclusions

This research significantly contributes to the field of cryptocurrency analysis by highlighting the efficacy of different regression models, including Elastic Net, in forecasting Bitcoin price volatility. Utilizing a robust dataset spanning from September 2014 to January 2024, this study emphasizes the importance of a methodologically sound approach to lag selection. Such methodologies enhance the model’s accuracy and relevance, as demonstrated by the meticulous selection of optimal lags using Random Forest, which is crucial in financial time series analysis. The ENR model excels in modelling the open and close price volatilities, showcasing its strength in capturing more linear and predictable market behaviours. However, the analysis also reveals that none of the examined regression models uniformly prevailed in terms of predictive accuracy.
Future research could involve assessing a broader range of cryptocurrencies to assess the generalizability of the proposed models across different digital currencies. Furthermore, the inclusion of additional financial and economic indicators could improve the models’ predictive capabilities, offering deeper insights into market behaviours.
Finally, the assessment of advanced machine learning and deep learning approaches could also provide valuable perspectives on nonlinear patterns in cryptocurrency volatility, potentially enhancing prediction accuracy.
In conclusion, while this study provides methodological advancements in predicting Bitcoin volatility, it also paves the way for further research to deepen our understanding of cryptocurrency markets and develop more sophisticated predictive models.

Author Contributions

Conceptualization, G.Z. and I.M.; Methodology, G.Z., I.M. and D.F.; Validation, I.M.; Resources, G.Z.; Writing—original draft, G.Z., I.M. and D.F.; Writing—review & editing, G.Z., I.M. and C.F.; Supervision, G.Z. and C.F.; Project administration, G.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sinha, N.; Yang, Y. Strategic Diversification for Asynchronous Asset Trading: Insights from Generalized Coherence Analysis of Cryptocurrency Price Movements. Ledger 2021, 6, 102–125. [Google Scholar] [CrossRef]
  2. Tsay, R.S. Analysis of Financial Time Series, 3rd ed.; Wiley: Hoboken, NJ, USA, 2010. [Google Scholar]
  3. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef]
  4. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd ed.; Springer: New York, NY, USA, 2009. [Google Scholar]
  5. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization Paths for Generalized Linear Models via Coordinate Descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef]
  6. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning with Applications in R; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  7. Bühlmann, P.; van de Geer, S. Statistics for High-Dimensional Data: Methods, Theory and Applications; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  8. Hastie, T.; Tibshirani, R.; Wainwright, M. Statistical Learning with Sparsity: The Lasso and Generalizations; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  9. James, B.; Bengio, Y. Random Search for Hyper-Parameter Optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  10. Polyzos, E.; Siriopoulos, C. Autoregressive Random Forests: Machine Learning and Lag Selection for Financial Research. Comput. Econ. 2023, 1–38. [Google Scholar] [CrossRef]
  11. Loh, W.Y. Classification and regression trees. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2011, 1, 14–23. [Google Scholar] [CrossRef]
  12. Liaw, A.; Wiener, M. Classification and regression by random Forest. R News 2002, 2, 18–22. [Google Scholar]
  13. Drucker, H.; Burges, C.J.C.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. Adv. Neural Inf. Process. Syst. 1997, 9, 155–161. [Google Scholar]
  14. de Prado Marcos, L. Advances in Financial Machine Learning; Wiley: Hoboken, NJ, USA, 2018. [Google Scholar]
  15. Kumar, A.S.; Gopirajan Pv, G.; Jackson, B. Machine Learning-Based Timeseries Analysis for Cryptocurrency Price Prediction A Systematic Review and Research. In Proceedings of the 2023 International Conference on Networking and Communications (ICNWC), Chennai, India, 5–6 April 2023; IEEE: Piscataway, NJ, USA, 2023. [Google Scholar]
  16. Jethani, L.; Patil, R.; Sanghvi, S.; Singh, R. Analysis of Machine Learning Models for Stock Market Prediction. In Proceedings of the 2023 International Conference on Integrated Intelligence and Communication Systems (ICIICS), Kalaburagi, India, 24–25 November 2023; Institute of Electrical and Electronics Engineers (IEEE): Piscataway, NJ, USA, 2023; pp. 1–8. [Google Scholar]
  17. Lumoring, N.; Chandra, D.; Gunawan, A.A.S. A Systematic Literature Review: Forecasting Stock Price Using Machine Learning Approach; IEEE: Piscataway, NJ, USA, 2023. [Google Scholar]
  18. Liu, L.A. Comparative Examination of Stock Market Prediction: Evaluating Traditional Time Series Analysis Against Deep Learning Approaches. Adv. Econ. Manag. Political Sci. 2023, 55, 196–204. [Google Scholar] [CrossRef]
  19. Jay, P.; Kalariya, V.; Parmar, P.; Tanwar, S.; Kumar, N.; Alazab, M. Stochastic neural networks for cryptocurrency price prediction. IEEE Access 2020, 8, 82804–82818. [Google Scholar] [CrossRef]
  20. Hemant; Parida, A.K.; Kumari, R.; Singh, A.R.; Bandyopadhyay, A.; Swain, S. Stock market prediction under a deep learning approach using Variational Autoencoder, and kernel extreme learning machine. In Proceedings of the 2023 OITS International Conference on Information Technology (OCIT), Raipur, India, 13–15 December 2023; pp. 156–161. [Google Scholar]
  21. Charandabi, S.E.; Kamyar, K. Prediction of Cryptocurrency Price Index Using Artificial Neural Networks: A Survey of the Literature. Eur. J. Bus. Manag. Res. 2021, 6, 17–20. [Google Scholar] [CrossRef]
  22. Pathak, S.; Kakkar, A. Cryptocurrency Price Prediction Based on Historical Data and Social Media Sentiment Analysis. Lect. Notes Netw. Syst. 2020, 103, 47–55. [Google Scholar]
  23. Usmani, S.; Shamsi, J.A. News sensitive stock market prediction: Literature review and suggestions. PeerJ Comput. Sci. 2021, 7, 1–36. [Google Scholar] [CrossRef]
  24. Dopi, G.Y.; Hartanto, R.; Fauziati, S. Systematic Literature Review: Stock Price Prediction Using Machine Learning and Deep Learning. In Proceedings of the International Conference on Management, Business, and Technology (ICOMBEST 2021), Jember, Indonesia, 12 October 2021; Volume 194, pp. 52–61. [Google Scholar]
  25. Parkinson, M. The Extreme Value Method for Estimating the Variance of the Rate of Return. J. Bus. 1980, 53, 61. [Google Scholar] [CrossRef]
  26. Robert, K. Opening and Closing Asymmetry: Empirical Analysis from ISE Xetra. Econ. Soc. Rev. 2008, 39, 55–78. [Google Scholar]
  27. Meilijson, I. The Garman–Klass Volatility Estimator Revisited. Revstat. Stat. J. 2008, 9, 199–212. [Google Scholar] [CrossRef]
  28. Robert, Ś.; Grzegorz, Z. High-Frequency and Model-Free Volatility Estimators. Available online: https://ssrn.com/abstract=2508648 (accessed on 19 October 2009). [CrossRef]
  29. Yang, D.; Zhang, Q. Drift-Independent Volatility Estimation Based on High, Low, Open, and Close Prices. J. Bus. 2000, 73, 477–492. [Google Scholar] [CrossRef]
  30. Hamilton, J.D. Time Series Analysis; Princeton University Press: Princeton, NJ, USA, 1994. [Google Scholar]
  31. Tay, J.K.; Narasimhan, B.; Hastie, T. Elastic Net Regularization Paths for All Generalized Linear Models. J. Stat. Softw. March 2023, 106, 1–13. [Google Scholar] [CrossRef] [PubMed]
  32. Xing, Y.; Li, D.; Li, C. Time series prediction via elastic net regularization integrating partial autocorrelation. Appl. Soft Comput. 2022, 129, 109640. [Google Scholar] [CrossRef]
  33. Roozbeh, M.; Arashi, M. Shrinkage ridge regression in partial linear models. Commun. Stat.—Theory Methods 2016, 45, 6022–6044. [Google Scholar] [CrossRef]
  34. Ghadeer, M.; Nadia, J.M.; Zahraa, A.-S. Regression shrinkage and selection variables via an adaptive elastic net model. J. Phys. Conf. Ser. 2021, 1879, 032014. [Google Scholar] [CrossRef]
  35. Tibshirani, R. Regression Shrinkage and Selection; Oxford University Press: Oxford, UK, 1996; Volume 58, pp. 267–288. [Google Scholar]
  36. Smola, A.J.; Schölkopf, B. A tutorial on support vector regression. Stat. Comput. 2004, 14, 199–222. [Google Scholar] [CrossRef]
  37. Schmidt, A.B. Quantitative Finance for Physicists: An Introduction; Academic Press: Cambridge, MA, USA, 2005. [Google Scholar]
  38. Thenmozhi, M. Support Vector Machines for Prediction of Futures Prices in Indian Stock Market. In Proceedings of the International Conference on Information Technology: Coding and Computing (ITCC’04), Las Vegas, NV, USA, 5–7 April 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 2, pp. 219–223. [Google Scholar]
Figure 1. Process map and pseudocode of the examined methodology.
Figure 1. Process map and pseudocode of the examined methodology.
Mathematics 12 01392 g001
Figure 2. Lags with an 85% cumulative feature importance.
Figure 2. Lags with an 85% cumulative feature importance.
Mathematics 12 01392 g002
Table 1. Critical synthesis of academic research effort.
Table 1. Critical synthesis of academic research effort.
ReferencesML Model(s)Feature EngineeringLag SelectionBitcoin Features
[15]LSTM, ARIMA, AR, MRYesNoprice, market cap
[16]LSTMYesNoprices
[17]LSTM, SVMYesNoprices, volumes
[18]ARIMA, LSTMYesNoprices
[19]SNNYesNoprices, volumes, and mining difficulty
[20]VAE, SNNYesNoprices, technical indicators
[21]ANNYesNoprices
[22]SBNNMYesNoprices, volume, and market sentiment
[23]ANNYesNoprices, volumes, and market sentiments
[24]LSTM, MLP, RF, SVMYesNoprices, volumes, and market sentiments
Table 2. Nomenclature of Elastic net Regression.
Table 2. Nomenclature of Elastic net Regression.
Model ParametersNomenclature
β 0 y intercept
β j Vector of the independent variable coefficients
y t Actual values of the prices at time t
y t ^ Predicted values of the prices at time t
T Dataset time periods (either train or test)
λ Regularization parameter. Higher λ means higher regularization
α Takes values of 0–1. Values of α higher than 0.5, means a higher impact of the Lasso regression regularization on the prediction, else a higher impact of Ridge regression
β j 1 Sum of absolute values of the coefficients
β j 2 2 Sum of squares of the coefficients
Table 3. Descriptive Statistics of Bitcoin Volatility Estimators.
Table 3. Descriptive Statistics of Bitcoin Volatility Estimators.
σ p 2 σ G K 2 σ M 2 σ R S 2
count3420.03420.03420.03420.0
mean0.0013750.9998840.9999830.268848
std0.0049133.8740173.0419940.963543
min4.34 × 10−157.87 × 10−105.19 × 10−88.95 × 10−13
25%0.0000330.0284110.0250950.006589
50%0.0002200.1535950.1564630.043649
75%0.0010530.6612570.7652130.209370
max0.207560125.595023100.63043042.454934
Table 4. Lags Selected per Volatility Estimator.
Table 4. Lags Selected per Volatility Estimator.
EstimatorOptimal No of LagsTop Significant Lags (Up to 85% CI)
VE_1185, 1, 21, 3, 10, 11, 24, 17, 15, 28, 8, 6, 7, 22, 20, 18, 9, 14
VE_21927, 28, 1, 11, 14, 22, 9, 10, 7, 12, 6, 26, 15, 23, 17, 8, 29, 18, 20
VE_3184, 15, 6, 9, 28, 22, 26, 7, 29, 24, 12, 16, 23, 13, 2, 10, 27, 3
VE_4185, 3, 1, 21, 10, 14, 24, 28, 22, 25, 18, 8, 13, 15, 7, 11, 20, 19
Table 5. Elastic net regression statistics.
Table 5. Elastic net regression statistics.
Coefficients V E 1   =   σ p 2 V E 2   =   σ G K 2 V E 3   =   σ M 2 V E 4   =   σ R S 2
β0 (Intercept)0.001071.0167 0.2515
Lag10.0171−0.0035 0.0000
Lag2
Lag30.0095 −0.0060254420.0000
Lag4 0.004396279
Lag50.0051 0.0063
Lag60.03040.000000.007677287
Lag70.0080−0.0039562390.0000000000.0614
Lag80.00000.000000000 0.0000
Lag90.0114−0.01242403−0.001456905
Lag100.00000.004415845 0.0000
Lag110.0000−0.006322465 0.0122
Lag12 −0.0039100960.000000000
Lag13 0.0000000000.0000
Lag140.0000−0.009497958 0.0000
Lag150.00050.0108476620.01145166670.0000
Lag16 −0.003135042
Lag170.0001 −0.008438236
Lag180.0046 0.000000000 0.0000
Lag19 0.0000
Lag200.00000.003699343 0.0000
Lag210.0000 0.0000
Lag220.00000.000000000.0079140590.0000
Lag23 −0.011457625−0.011388763
Lag240.0000 0.0046330000.0000
Lag25 0.0000
Lag26 0.000013700−0.007468076
Lag27 −0.002826818
Lag280.0000−0.0033253710.000000000.0000
Lag29 0.0083719660.002658218
Lag30
Parameter Values
λ (Lambda)0.05110.094050.11160.5613
α (Alpha)0.000007950.90530.73690.0762
MAE0.0015721.22301.28330.2076
Table 6. Model parameter and MAD values.
Table 6. Model parameter and MAD values.
Volatility EstimatorsModelParametersValues
1DTRDepth38
Leaves5407
MAE0.0020
RFRAverage_Depth40.64
Average_Leaves3414.1
MAE0.0018
SVRC (Regularization parameter)1
Epsilon (Error margin)0.1
Number_of_Support_Vectors17
MAE0.0839
2DTRDepth49
Leaves5423
MAE1.8767
RFRAverage_Depth43.75
Average_Leaves3425.3
MAE1.3853
SVRC (Regularization parameter)1
Epsilon (Error margin)0.1
Number_of_Support_Vectors2136
MAE0.8495
3DTRDepth56
Leaves5423
MAE2.0269
RFRAverage_Depth48.19
Average_Leaves3425.28
MAE1.5067
SVRC (Regularization parameter)1
Epsilon (Error margin)0.1
Number_of_Support_Vectors2230
MAE0.9483
4DTRDepth43
Leaves5425
MAE0.4224
RFRAverage_Depth43.13
Average_Leaves3426.6
MAE0.3359
SVRC (Regularization parameter)1
Epsilon (Error margin)0.1
Number_of_Support_Vectors1351
MAE0.2492
Table 7. MAE values per regression model and volatility estimator.
Table 7. MAE values per regression model and volatility estimator.
ModelVE1VE2VE3VE4
ENR0.00161.23661.28330.2076
DTR0.00201.87672.02690.4224
RFR0.00181.38531.50670.3359
SVR0.08390.84950.94830.2492
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zournatzidou, G.; Mallidis, I.; Farazakis, D.; Floros, C. Enhancing Bitcoin Price Volatility Estimator Predictions: A Four-Step Methodological Approach Utilizing Elastic Net Regression. Mathematics 2024, 12, 1392. https://doi.org/10.3390/math12091392

AMA Style

Zournatzidou G, Mallidis I, Farazakis D, Floros C. Enhancing Bitcoin Price Volatility Estimator Predictions: A Four-Step Methodological Approach Utilizing Elastic Net Regression. Mathematics. 2024; 12(9):1392. https://doi.org/10.3390/math12091392

Chicago/Turabian Style

Zournatzidou, Georgia, Ioannis Mallidis, Dimitrios Farazakis, and Christos Floros. 2024. "Enhancing Bitcoin Price Volatility Estimator Predictions: A Four-Step Methodological Approach Utilizing Elastic Net Regression" Mathematics 12, no. 9: 1392. https://doi.org/10.3390/math12091392

APA Style

Zournatzidou, G., Mallidis, I., Farazakis, D., & Floros, C. (2024). Enhancing Bitcoin Price Volatility Estimator Predictions: A Four-Step Methodological Approach Utilizing Elastic Net Regression. Mathematics, 12(9), 1392. https://doi.org/10.3390/math12091392

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop