Next Article in Journal
Financial Inclusion and Economic Growth in Sub-Saharan Africa—A Panel ARDL and Granger Non-Causality Approach
Next Article in Special Issue
A Cyclical Phenomenon among Stock & Commodity Markets
Previous Article in Journal
Corporate Finance, Governance, and Social Responsibility
Previous Article in Special Issue
Pricing Multi-Asset Bermudan Commodity Options with Stochastic Volatility Using Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Do Large Datasets or Hybrid Integrated Models Outperform Simple Ones in Predicting Commodity Prices and Foreign Exchange Rates?

Graduate School of Economics, Kobe University, 2-1 Rokkodai, Nada-Ku, Kobe 657-8501, Japan
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2023, 16(6), 298; https://doi.org/10.3390/jrfm16060298
Submission received: 27 April 2023 / Revised: 5 June 2023 / Accepted: 6 June 2023 / Published: 9 June 2023
(This article belongs to the Special Issue Commodity Market Finance)

Abstract

:
With the continuous advancement of machine learning and the increasing availability of internet-based information, there is a belief that these approaches and datasets enhance the accuracy of price prediction. However, this study aims to investigate the validity of this claim. The study examines the effectiveness of a large dataset and sophisticated methodologies in forecasting foreign exchange rates (FX) and commodity prices. Specifically, we employ sentiment analysis to construct a robust sentiment index and explore whether combining sentiment analysis with machine learning surpasses the performance of a large dataset when predicting FX and commodity prices. Additionally, we apply machine learning methodologies such as random forest (RF), eXtreme gradient boosting (XGB), and long short-term memory (LSTM), alongside the classical statistical model autoregressive integrated moving average (ARIMA), to forecast these prices and compare the models’ performance. Based on the results, we propose novel methodologies that integrate wavelet transformation with classical ARIMA and machine learning techniques (seasonal-decomposition-ARIMA-LSTM, wavelet-ARIMA-LSTM, wavelet-ARIMA-RF, wavelet-ARIMA-XGB). We apply this analysis procedure to the commodity gold futures prices and the euro foreign exchange rates against the US dollar.

1. Introduction

The increasing utilization of sentiment analysis (SA) for obtaining a sentiment index holds promise as an approach for predicting commodity prices and foreign exchange rates. By analyzing unstructured data such as social media posts, news articles, and other textual data, SA provides insights into public opinions and market sentiment, enabling price prediction (Smailović et al. 2013). Utilizing a sentiment index, rather than relying on a large dataset of indicators, offers several advantages, including simplifying the modeling process and reducing the risk of overfitting. SA also offers a more up-to-date perspective on market sentiment, as it captures real-time changes in public opinion and market sentiment (Philander and Zhong 2016). However, while a sentiment index proves valuable in predicting short-term fluctuations (Qiu et al. 2022) in commodity and foreign exchange markets, long-term trends in these markets are more significantly influenced by factors such as macroeconomic indicators and political events. Hence, while SA presents a promising approach to prediction, we must also consider its limitations and potential biases and supplement SA with other relevant data sources and indicators.
Meanwhile, research has demonstrated that advancements in machine learning and the availability of more data enhance the accuracy of price prediction in certain cases (Bakay and Ağbulut 2021; Bouktif et al. 2018; Wang and Wang 2016; Amat et al. 2018; Chatzis et al. 2018; Farsi et al. 2021; Zhang and Hamori 2020; Plakandaras et al. 2015; Luo et al. 2019; McNally et al. 2018; Phyo et al. 2022; Nguyen and Ślepaczuk 2022). These technologies aid in identifying patterns and correlations within large and complex datasets that may prove challenging for human analysts to discern. However, employing large datasets and machine learning algorithms does not guarantee accuracy as these techniques are susceptible to biases, overfitting, and the appropriateness of the model design. In certain scenarios, simple models may outperform more sophisticated ones (He 2018), particularly when limited data are available or the underlying relationships are straightforward. Decision making and risk management may, at times, derive greater benefit from simple models based on relevant facts and hypotheses.
Recent research has garnered significant interest from academics and practitioners due to the emergence of hybrid techniques that combine classical models with machine learning models. Hybrid prediction models have been utilized in various research fields, including meteorology, hydraulics, and exhaust emissions, for forecasting purposes (Chang et al. 2019; Liu et al. 2018; de O. Santos Júnior et al. 2019; McNally et al. 2018; Sadefo Kamdem et al. 2020; Selvin et al. 2017; Xue et al. 2022; Sun et al. 2022; Wu et al. 2021; Wu and Wang 2022; Yu et al. 2020; Zhang et al. 2018, 2022; Zolfaghari and Gholami 2021; Ma et al. 2019; Dave et al. 2021; Zhao et al. 2022; Moustafa and Khodairy 2023; Zolfaghari and Gholami 2021). This study proposes several approaches that integrate machine and deep learning models with conventional statistical models, based on the assumption that time series can be decomposed into linear and nonlinear components or into time-dependent sums of frequency components and noise.
Hence, the primary objectives of this research are as follows: First, to analyze whether sentiment indicators derived from sentiment analysis techniques can outperform a large dataset of indicators when employing machine learning and deep learning methods for prediction. Second, to verify whether machine learning models, which have gained considerable attention, genuinely exhibit better prediction capabilities than classical ARIMA models. Third, to apply our proposed hybrid model to commodity gold futures prices and foreign exchange rates, evaluate their prediction performance, and compare them with the aforementioned machine learning and classical statistical approaches.
This study is divided into three steps. In the first step, we perform sentiment analysis on the collected unstructured news headlines to obtain a sentiment index (referred to as the SI dataset). Then, we calculate technical indicators and collect other relevant indicators from stock markets, bond markets, commodity markets, and foreign exchange markets to create a multivariate dataset (referred to as the large dataset). In the second step, we apply moving window machine learning approaches (RF, XGB, and LSTM) and a classical statistical model (ARIMA) to these two datasets to evaluate their prediction performance using the root mean squared error (RMSE), mean absolute percentage error (MAPE), and mean absolute error (MAE). In the third step, we propose several decompositions and transformations integrated with statistical and machine learning approaches, such as seasonal-decomposition-ARIMA-LSTM, wavelet-ARIMA-LSTM, wavelet-ARIMA-RF, and wavelet-ARIMA-XGB. Specifically, we first transform and decompose the time series into linear and nonlinear parts or dynamic levels and noise parts. Then, we apply classical ARIMA to predict the linear and dynamic levels and use RF, XGB, and LSTM machine/deep learning approaches to predict the nonlinear and noise parts. We evaluate our proposed approaches using RMSE, MAPE, and MAE and compare the prediction results with the aforementioned forecasting. Additionally, we perform walk-forward testing to validate the effectiveness of the triple-combination approaches. To assess any statistically significant differences between our proposed approach and the ARIMA model, we utilize the modified Diebold–Mariano test statistic. This comprehensive testing methodology provides further insights into the performance and comparative analysis of the proposed approaches.
The main findings of this study are as follows: First, the combination of the sentiment indicator with the moving window LSTM machine learning model demonstrates outstanding forecasting performance. Second, the sentiment indicator dataset used in conjunction with the moving window machine learning and deep learning models does not surpass the performance of the traditional ARIMA model. Third, our proposed triple-combination approaches exhibit superior prediction performance compared to either the machine learning models or the ARIMA model when forecasting commodity gold futures prices and euro foreign exchange rates. Lastly, although the sentiment indicator dataset does not outperform the prediction accuracy of the ARIMA model, our empirical results indicate that the sentiment dataset is more accurate in predicting commodity prices and foreign exchange rates than the large dataset, which comprises various indicators.
To the best of our knowledge, this study is the first to investigate whether sentiment indicators can replace a large dataset of indicators in forecasting commodity prices and foreign exchange rates. Moreover, this study introduces a novel approach by combining data decomposition with machine learning models and classical statistical models to predict prices in commodity and foreign exchange markets. Additionally, the proposed triple-combination approaches demonstrate higher accuracy compared to the individual models. These findings offer new insights and potential predictors for investors and policymakers.
The rest of this paper is organized as follows: Section 2 reviews the literature. Section 3 provides a detailed description of the study’s data, methodologies, and evaluation measures. Section 4 presents and analyzes the empirical results. Finally, Section 5 concludes the study.

2. Literature Review

A wide range of valuable Internet data, particularly textual data such as news press releases, are being evaluated for forecasting purposes in various fields, thanks to the rapid expansion of the Internet and advancements in big data technologies. Consequently, researchers are actively working on improving sentiment analysis (SA) predictions and exploring the potential of SA to enhance time series forecasting performance in different markets (Bollen et al. 2011; Naeem et al. 2021; Deeney et al. 2015; Li et al. 2016; Das et al. 2018; Pai and Liu 2018; Razzaq et al. 2019; Bedi and Khurana 2019; Ito et al. 2019, 2020; Sivri et al. 2022; Seals and Price 2020; Xiang et al. 2021; Guo et al. 2020; Sharma et al. 2020; Mukta et al. 2022).
The contribution of Bedi and Khurana (2019) is focused on improving SA prediction for textual data by incorporating fuzziness with deep learning. Ito et al. (2019) and Ito et al. (2020) propose a novel neural network model called the contextual sentiment neural network (CSNN) model, which offers insights into the SA prediction process and utilizes an initialization propagation (IP) learning strategy. Leveraging SA on Twitter tweets, Naeem et al. (2021) suggest a machine learning-based strategy for forecasting exchange rates. Their findings demonstrate that SA can facilitate the prediction of foreign exchange rates, particularly the US dollar against the Pakistani rupee. Li et al. (2016) acknowledge the usefulness of online data, including news releases and social media networks such as Twitter, in forecasting price changes. Xiang et al. (2021) propose a Chinese Weibo SA algorithm that combines the BERT (Bidirectional Encoder Representations from Transformers) model and the Hawkes process to effectively monitor changes in users’ emotional states and perform SA on Weibo. However, limited studies have examined whether sentiment indicators can replace large sets of index data for forex prediction. If sentiment indicators can effectively replace a substantial amount of index datasets and achieve comparable or better forecasting performance, it could significantly enhance forecasting efficiency and provide valuable insights to investors and decision-makers.
Moreover, in recently published research, the use of rapidly developing machine and deep learning modeling techniques for forecasting time series is one of the most extensively researched topics in the academic literature (Bakay and Ağbulut 2021; Bouktif et al. 2018; Wang and Wang 2016; Amat et al. 2018; Chatzis et al. 2018; Farsi et al. 2021; Zhang and Hamori 2020; Plakandaras et al. 2015; Luo et al. 2019; McNally et al. 2018; Phyo et al. 2022). Specifically, Amat et al. (2018) demonstrate that fundamentals from simple exchange rate models (such as purchasing power parity (PPP) or uncovered interest rate parity (UIRP)) or Taylor-rule-based models improve exchange rate forecasts for major currencies when using machine learning models. Similarly, Zhang and Hamori (2020) find that integrating machine learning models with traditional foreign exchange rate models and Taylor’s rule foreign exchange rate models effectively predict foreign exchange rates. Phyo et al. (2022) train five of the best ML algorithms, including the extra trees regressor (ETR), random forest regressor (RFR), light gradient boosting machine (LGBM), gradient boosting regressor (GBR), and K neighbors regressor (KNN), to build the proposed voting regressor (VR) model. Li et al. (2020) propose a new dynamic ensemble forecasting system based on a multi-objective intelligent optimization algorithm to forecast the air quality index, which includes time-varying parameter weights and three main modules: a data preprocessing module, a dynamic integration forecasting module, and a system evaluation module. Plakandaras et al. (2015) predict daily and monthly exchange rates using machine learning techniques. Building on these empirical results, this paper considers the application of machine learning and deep learning methodologies to investigate whether sentiment indicator datasets can substitute for large datasets.
On the other hand, as a classical statistical model, ARIMA is used for long-term prediction (Darley et al. 2021). Many studies compare ARIMA and machine learning in forecasting time series (Shih and Rajendran 2019; Siami-Namini et al. 2018, 2019; He 2018; Yamak et al. 2019; Ribeiro et al. 2020; Liu et al. 2021). Siami-Namini et al. (2018) compare the ARIMA model with the LSTM model in forecasting time series and demonstrate that deep learning approaches such as LSTM outperform traditional models such as ARIMA. In contrast, He (2018) explores weekly crude oil price data from the U.S. Energy Information Administration between 2009 and 2017 to test the forecasting accuracy of time series models (simple exponential smoothing (SES), moving average (MA), and autoregressive integrated moving average (ARIMA)) against machine learning support vector regression (SVR) models. The main contribution of this study is to determine whether ARIMA provides more accurate forecasting results for crude oil prices than SVR models. Siami-Namini et al. (2019) conduct a behavioral analysis and comparison of BiLSTM and LSTM models and compare the two models with the ARIMA model. The results demonstrate that BiLSTM models provide better predictions compared to ARIMA and LSTM models. Yamak et al. (2019) conduct a comparison analysis between ARIMA, LSTM, and gated recurrent unit (GRU) for time series forecasting. Ribeiro et al. (2020) compare two benchmarks (autoregressive integrated moving average (ARIMA) and an existing manual technique used at the case site) against three deep learning models (simple recurrent neural networks (RNN), long short-term memory (LSTM), and gated recurrent unit (GRU)) and two machine learning models (support vector regression (SVR) and random forest (RF)) for short-term load forecasting (STLF) using data from a Brazilian thermoplastic resin manufacturing plant. Their empirical results show that the GRU model outperforms all other models. Liu et al. (2021) propose a seasonal autoregressive integrated moving average (SARIMA) model to predict hourly measured wind speeds in the coastal and offshore areas of Scotland. Motivated by the results of the prior literature and considering the limited literature comparing ARIMA models with machine learning and deep learning models for predicting gold prices and Euro FX prices, this study aims to fill this gap in the literature.
Since we are unable to demonstrate that machine learning and deep learning techniques outperform the traditional ARIMA model, we aim to enhance the accuracy of commodity price and foreign exchange rate predictions. In our literature research, we discover numerous studies in various fields, such as astronomy, hydraulics, exhaust emissions, and meteorology, that employ a combination of traditional models and other techniques such as machine learning, deep learning methodologies, and two-step models, which involve preprocessing the data before predicting time series. Some relevant studies include Chang et al. (2019), Liu et al. (2018), de O. Santos Júnior et al. (2019), McNally et al. (2018), Sadefo Kamdem et al. (2020), Selvin et al. (2017), Xue et al. (2022), Sun et al. (2022), Wu et al. (2021), Wu and Wang (2022), Yu et al. (2020), Zhang et al. (2018, 2022), Zolfaghari and Gholami (2021), Ma et al. (2019), Dave et al. (2021), Zhao et al. (2022), Moustafa and Khodairy (2023), and Zolfaghari and Gholami (2021).
To enhance prognostic accuracy, Ma et al. (2019) propose a data-fusion approach that combines long short-term memory (LSTM), recurrent neural network (RNN), and the autoregressive integrated moving average (ARIMA) method to forecast fuel cell performance. Chang et al. (2019) present an electricity price-prediction model based on a hybrid of the LSTM neural network and wavelet transform. Liu et al. (2018) attempt to forecast wind speed using a deep learning strategy with wavelet transform. Dave et al. (2021) aim to provide accurate predictions of Indonesia’s future exports by developing an integrated machine learning model with ARIMA. Zhou et al. (2022) propose a combined model based on complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), four deep learning (DL) models, and the autoregressive integrated moving average (ARIMA) model. Zhao et al. (2022) address the lack of using coupled models to separately model different frequency subseries of precipitation series for prediction and propose a coupled model based on ensemble empirical mode decomposition (EEMD), long short-term memory neural network (LSTM), and autoregressive integrated moving average (ARIMA) for month-by-month precipitation prediction. Moustafa and Khodairy (2023) implement four models, including long short-term memory (LSTM), autoregressive integrated moving average (ARIMA), seasonal autoregressive integrated moving average (SARIMA), and a hybrid model, to forecast the maximum sunspot number of cycles 25 and 26. Zolfaghari and Gholami (2021) employ a hybrid model that combines adaptive wavelet transform (AWT), long short-term memory (LSTM), and models from the ARIMAX-GARCH family to forecast stock indices for the Dow Jones Industrial Average (DJIA) and the Nasdaq Composite (IXIC). Chen and Wang (2019) integrate the LSTM and ARIMA models for predicting satellite time series data. Inspired by these studies, this investigation aims to propose hybrid approaches applicable to time series forecasting in commodity markets and foreign exchange markets.
To summarize, researchers have dedicated significant efforts to enhancing the accuracy of price prediction by utilizing machine learning techniques and internet-based information. The increasing availability of data sources, particularly textual data such as news articles, and advancements in big data technologies have led to the evaluation of various datasets for forecasting in different domains. However, in the context of time series forecasting in commodity and foreign exchange markets, there is a lack of literature that thoroughly compares the effectiveness of sentiment indicator datasets with large datasets containing diverse variables. Additionally, the recent academic literature extensively explores the application of rapidly evolving machine learning and deep learning modeling techniques for time series forecasting. Nevertheless, further investigation is required to determine whether machine learning and deep learning models outperform classical statistical methods, such as the ARIMA model, which have long been used for forecasting purposes in the commodity and foreign exchange markets. Therefore, in our study, we focus on improving forecasting accuracy by combining traditional models with other methods, including machine learning, deep learning techniques, and two-step models. We draw inspiration from previous studies conducted in fields such as astronomy, hydraulics, exhaust emissions, and meteorology, which have employed time series forecasting in their respective domains.

3. Data and Methodology

3.1. Data

3.1.1. Data Collection

Gold prices are widely regarded as a leading indicator of economic conditions, particularly inflation and market volatility, making it an extremely important commodity (Blose 2010; Livieris et al. 2020). As a result, gold is a popular investment asset (Ratner and Klein 2008) and is commonly used as a hedge against inflation and market volatility (Chua and Woodward 1982). Predicting gold prices can provide valuable insights for economic forecasts and assist policymakers and investors in making informed decisions (Raza et al. 2018). Additionally, many central banks maintain gold reserves as a means of preserving value and protecting against currency fluctuations (Aizenman and Inoue 2013).
On the other hand, foreign exchange rates have been utilized as leading indicators of economic growth and inflation (Razzaque et al. 2017). The foreign exchange market plays a crucial role in international trade (Latief and Lefen 2018), financial instrument settlement, inflation control, and overall economic development and currency stability. Accurate predictions of foreign exchange rates are essential for businesses and investors to develop effective hedging strategies that mitigate risks associated with currency fluctuations. Moreover, such predictions inform government policy decisions related to trade, monetary policy, and capital flows (Amato et al. 2005; Mussa 1976). Governments can use exchange rate predictions to anticipate the impacts of policy decisions on the economy and make necessary adjustments. It is also worth noting that the euro is the second-most traded currency globally, following the US dollar, and is extensively used by numerous European Union members. Given the widespread usage of the euro in international trade and its status as a major reserve currency, exchange rate fluctuations can significantly influence the costs and risks associated with international transactions. Therefore, forecasting euro exchange rates is vital for financial stability and effective hedging strategies. Consequently, this study selected gold futures prices from the commodity market and the EUR foreign exchange rate as the objects of forecasting.
Based on the concept of proposing a powerful alternative sentiment indicator to replace large datasets, this study applies sentiment analysis to unstructured data extracted from news headlines. The prediction objects selected for this study are gold futures prices and the euro exchange rate against the US dollar, sourced from invest.com. After preprocessing the dataset, a total of 3957 daily data points were obtained, covering the period from 3 February 2004 to December 2019. The prediction conducted in this study is one-day-ahead forecasting.
The large dataset used in this study consists of 22 different financial indicators obtained from various sources such as Bloomberg, Thomson Reuters Datastream, the Federal Reserve Bank, Investing.com, Yahoo! Finance, and Macrotrends. Specifically, the large dataset includes the stock market index, 10-year government bond yields, volatility indices, and significant commodity market indices such as oil, gas, corn, and wheat. Additionally, it incorporates 10 calculated technical indices, including moving averages, exponential weighted moving averages, Bollinger bands, moving average convergence divergence, and the relative strength index.

3.1.2. Sentiment Analysis and Sentiment Indicator

In this study, we conduct sentiment analysis to obtain a sentiment indicator as an input variable.
First, we utilize unstructured daily news headline text data from 19 February 2003 to 31 December 2020. The data consist of 1,226,258 news headlines collected from a reputable news source, the Australian Broadcasting Corporation (ABC). The news headline data are sourced from Harvard Dataverse, which was created by Kulkarni (2018). According to the authors’ notes, “with a volume of two hundred articles each day and a good focus on international news, we can be fairly certain that every event of significance has been captured here”.
For sentiment analysis on daily news headlines, we employ a Python natural language processing library called TextBlob. TextBlob is chosen for its ability to provide rules-based sentiment scores and assign polarity and subjectivity to words and phrases. These scores are derived from a pre-defined set of categorized words readily available from the Natural Language Toolkit (NLTK) database (Vijayarani and Janani 2016). The input data for sentiment analysis typically consist of a corpus, such as a collection of text documents. The output of sentiment analysis includes a sentiment polarity score (indicating positivity or negativity) and a subjectivity score (measuring opinionated-ness). The polarity score ranges from −1.0 to 1.0, where −1.0 represents strong negativity and 1.0 represents high positivity. The subjectivity score ranges from 0.0 to 1.0, where 0.0 denotes extreme objectivity or factual content, while 1.0 signifies high subjectivity.
The sentiment analysis procedure is described as follows:
  • Firstly, the NLTK is used to clean the unstructured text data.
  • Secondly, TextBlob is applied to classify the polarity and subjectivity of each news headline.
  • Thirdly, the total number of subjective, objective, negative, positive, and neutral news headlines is counted for each day, and then divided by the total number of news headlines on that day.
  • Fourthly, the sentiment analysis output data are obtained, which includes the percentage values for subjectivity, objectivity, negativity, neutrality, and positivity for each day.
  • Finally, following Henry’s finance-specific dictionary (Henry and Leone 2016), the sentiment can be evaluated using the formula below:
    S I t = N p ( H t ) N n ( H t ) N p ( H t ) + N n ( H t )
    where H t represents the collected news article headlines at time t , N p represents the total number of positive news headlines in H t , N n represents the total number of negative news headlines in H t , and S I t represents the corresponding sentiment indicator.
The sentiment indicator represents the percentage difference between the number of positive and negative news articles.

3.1.3. Sentiment Indicator Dataset and Large Indicator Dataset

After data processing, we obtain 3957 daily data points that contain 32 explanatory variables, covering a 15-year period from 3 February 2004 to 16 December 2019. The descriptions and sources of the data are elaborated in Table A1 of the Appendix A.
In this study, we use 85% of the daily data (3363 days) to train various models based on RF, XGBoost, and LSTM models. We then validate the remaining data (594 days) to conduct out-of-sample forecasting. Figure 1 illustrates the raw data of the gold futures prices, Figure 2 presents the prices of the euro rates multiplied by 100, and Figure 3 presents the calculated sentiment index based on the results of sentiment analysis. The dashed vertical line (14 July 2017) denotes the separation between the training and test data.
To test the hypothesis that the sentiment indicator can be a substitute for the large datasets of indicators in exchange rate prediction, we construct two datasets to evaluate the effectiveness of the sentiment indicator and compare their predictive performance. Detailed information regarding these variables is provided in Table 1.

3.2. Prediction Models and Proposed Approaches

This study applies the RF, XGB, and LSTM approaches in combination with the expanding moving window (EMW), and fixed moving window (FMW) methods to predict gold futures commodity prices and the euro foreign exchange rate. The initial parameters (Wysocki and Ślepaczuk 2022) are selected using the grid search method. Specifically, trained models with time-varying parameters are used to predict one-period-ahead prices, and the prediction performance of these models is evaluated using the remaining test datasets. The moving window technique proceeds iteratively with the prediction, where the size of the expanding moving window or fixed moving window is extended or shifted by one-time step in each iteration. Furthermore, the study employs the widely applied time series forecasting model ARIMA to validate the superiority of the sentiment indicator dataset. Additionally, triple-combination approaches are proposed, including wavelet-ARIMA-LSTM (wavelet-ARIMA-RF/wavelet-ARIMA-XGB) and seasonal-decomposition-ARIMA-LSTM.

3.2.1. Expanding Moving Window (EMW) and Fixed Moving Window (FMW)

This study employs two patterns of moving window techniques to predict one-period-ahead, aiming to investigate whether there is a difference in prediction performance when excluding historical data. One pattern is the fixed-length moving window (FMW) technique, and the other is the expanding-length moving window (EMW) technique.
The moving window statistics proceed iteratively with the prediction, extending or shifting the size of EMW or FMW by one time step in each iteration. Figure 4 illustrates the mechanism of EMW, while Figure 5 depicts the mechanism of FMW.
In terms of the expanding-length window, the initial window size is set to 3363, which is the same as the length of the validation data (there are 3957 observations from 3 February 2003 to 16 December 2020). When iterating the model fitting, the window size increases by one period. For example, the first window spans from 3 February 2003 to 16 July 2017, and is used to estimate 17 July 2017. The framework utilizes the dataset from period 1 to 3363 to train the model, then uses the trained model to forecast period 3364, and incorporates the extended training dataset from period 1 to 3364 to retrain the model. The updated model is then used to predict period 3365. This process is iterated until the last period of the time series. The expanding moving window technique is also employed in the model evaluation as walk-forward testing (Baranochnikov and Ślepaczuk 2022).
In terms of the fixed-length window, the window size is determined to be 3363. For instance, the first window spans from 3 February 2003 to 16 July 2017, and is used to estimate 17 July 2017. The model uses the dataset from period 1 to 3363 to train the model and utilizes this trained model to forecast period 3364. Then, the dataset from period 2 to 3364 is used to train the model, and the updated model is used to predict period 3365. This process is iterated until the last period of the time series.

3.2.2. Random Forest (RF)

The RF approach, introduced by Breiman (2001), is an ensemble machine learning method that incorporates multiple decision trees to improve prediction performance. By extending each tree from randomly selected features and building them from the primal sample, the RF method addresses the overfitting problem that can arise when adding more trees to the forest. This approach enhances prediction accuracy.
To maximize the forecasting performance of our model, we conducted a meticulous parameter-tuning process. We optimized several variables to achieve optimal results in our forecasting endeavor. The variables that underwent optimization included n_estimators (with values of 100, 200, 300, 400, and 500), max_depth (with values of 1, 3, 10, 20, 30, 40, and 50), bootstrap (with options of True and False), and min_samples_leaf (ranging from 1 to 10). After a thorough evaluation based on error metrics, we selected the following parameter values: n_estimators (300), max_depth (20), bootstrap (True), and min_samples_leaf (3). These parameter values were found to yield the best performance in our model, ensuring accurate and reliable forecasting outcomes.

3.2.3. Extreme Gradient Boosting (XGBoost)

XGBoost, an algorithm proposed by Chen and Guestrin (2016), is an ensemble machine learning model that enhances gradient boosting techniques (Friedman 2001). It employs an optimized platform for gradient boosting, leveraging parallel processing, tree pruning, and hardware optimization. XGBoost offers a variety of objective functions, including classification and regression, and combines weaker and simpler learner estimates (such as regression trees) to improve prediction accuracy. The model minimizes a subjective loss function through a penalty term for model complexity (i.e., regression tree functions) and a convex loss function. Iterative learning involves creating new trees and merging them with existing trees.
To enhance the predictive performance of our model, we conducted a meticulous parameter-tuning process. We optimized several variables to achieve optimal results in our forecasting endeavor. The variables that underwent optimization included n_estimators (ranging from 100 to 1000 in increments of 100), max_depth (with values of 1, 3, 5, and 10), learning_rate (with values of 0.001 and 0.01), and gamma (with values of 0, 0.001, and 0.01). Subsequently, based on the performance evaluation using error metrics, we selected the following parameter values: n_estimators (1000), max_depth (3), learning_rate (0.01), and gamma (0.01).

3.2.4. Long Short-Term Memory (LSTM)

The LSTM algorithm was first introduced by Hochreiter and Schmidhuber (1997). As a prominent model in deep learning, LSTM exhibits an external loop structure similar to that of RNN and an internal recurrent structure consisting of memory cells. Each memory cell possesses self-connected recurrent weights that interact with three types of gates, ensuring the preservation of signals over multiple time steps without suffering from exploding or vanishing gradients. Similar to RNN, LSTM can utilize more data at each time step, resembling the memory capacity of the LSTM unit. The network utilizes these gates to effectively manage the retention and forgetting of information for subsequent iterations.
To achieve optimal forecasting outcomes, we meticulously tuned the hyperparameters of our model. Various variables underwent optimization, including batch size (ranging from 10 to 200), number of epochs (ranging from 10 to 300), optimization technique (SGD, Adam, RMSprop), learning rate (0.001, 0.01, 0.1), dropout rate (ranging from 0.0 to 0.9), neuron activation function (relu, sigmoid), number of layers (ranging from 1 to 5), and number of neurons (16, 32, 46, 64, 128). During the training of the neural networks, we employed the traditional mean squared error (MSE) loss function, as utilized by Cao et al. (2019), Chimmula and Zhang (2020), and Livieris et al. (2020). This loss function is widely recognized and commonly used in the field. Following a comprehensive evaluation process, we selected the following parameter values that exhibited superior performance: a batch size of 15, 150 epochs, the Adam optimization technique, a learning rate of 0.001, no dropout (dropout rate of 0.0), relu activation function, 3 layers, and 46 neurons. These parameter values were determined to produce the most accurate and reliable forecasting results in our model.

3.2.5. AutoRegressive Integrated Moving Average (ARIMA)

ARIMA was developed in the 1970s by Box and Jenkins (1968) with the aim of mathematically characterizing variations in time series. Non-stationary data need to be differenced until stationarity is achieved, as ARIMA specifically works with stationary data. In ARIMA (p, d, q), where p represents the autoregressive terms, d represents the differencing order, and q represents the lagged errors, the best values for p, d, and q are determined using the Akaike information criterion to fit the data.
In this study, the selection of optimal (p, d, q) values for time series analysis is performed using the auto_arima function in Python. The auto_arima function employs a stepwise search method to minimize the Akaike Information Criteria (AIC). To ensure model parsimony, the maximum values for p and q are set to be less than 5. The determination of the optimal differencing parameter, d, is achieved through the application of the Augmented Dickey-Fuller test.

3.2.6. Wavelet-ARIMA-LSTM (Wavelet-ARIMA-RF/Wavelet-ARIMA-XGB)

The wavelet transform was first introduced by French scientist J. Morlet in 1974 (Morlet et al. 1982). Wavelet decomposition has been widely utilized as a preprocessing approach in various fields such as engineering, time series analysis, and medicine. By applying wavelet decomposition, time series data can be separated into approximation and detail components. In this study, we employ discrete wavelet decomposition (DWD) to decompose the gold futures prices and the euro exchange rate into multiple approximation and detail component series. Unlike previous research, we simplify the analysis by using the decomposed approximation series for forecasting one-period-ahead values using the ARIMA model. We then calculate the residuals and apply the LSTM model to predict the one-period-ahead residuals, and finally combine them.
In summary, the DWD technique is employed to decompose the price time series into linear approximation components and nonlinear residual components. The linear components are predicted using the ARIMA model, while the nonlinear parts are independently forecasted using the LSTM model, taking into account the intrinsic characteristics of these models.
Similarly, in the case of wavelet-ARIMA-RF and wavelet-ARIMA-XGB, the random forest model and extreme gradient boosting are applied, respectively, to predict the nonlinear components.

3.2.7. Seasonal-Decomposition-ARIMA-LSTM

Furthermore, we employed another preprocessing technique, known as traditional seasonal decomposition, for the time series models of gold futures prices and the euro exchange rate. According to the traditional concept of time series decomposition, a series is considered as a composite of level, trend, seasonality, and noise components. In this study, we regard the level, trend, and seasonality components as systematic components since they exhibit consistency or recurrence and can be described and modeled. Conversely, we classify the noise component as non-systematic due to its random variation nature. Diverging from previous research, we utilize the decomposed systematic components, including the trend series and seasonality series, to apply the ARIMA model for forecasting one-period-ahead values. Subsequently, we employ the decomposed non-systematic noise components to apply the LSTM model for predicting one-period-ahead noise, and ultimately aggregate these predicted values.
In summary, the traditional seasonal decomposition method is utilized to decompose the price time series into linear systematic components and nonlinear non-systematic components. The linear components are then forecasted using the ARIMA model, while the nonlinear components are separately predicted using the LSTM model.

3.3. Model Evaluation Measures

3.3.1. Root Mean Squared Error (RMSE)

The discrepancy between the expected and actual values is typically measured using the RMSE. The RMSE is typically computed as follows:
RMSE = i = 1 N ( x i x i ^ ) 2   N
where N is the number of non-missing data points, x i is the actual observation time series, and x i ^   is the estimated time series.

3.3.2. Mean Absolute Percentage Error (MAPE)

The accuracy of forecasting models is frequently assessed statistically using the mean absolute percentage error (MAPE). MAPE can be calculated as the average absolute percent error for each time period minus actual values divided by actual values. Generally speaking, the following equation defines MAPE:
MAPE = 1 n i = 1 N | x i x i ^ x i |
where i =  variable,  N =  number of non-missing data points,  x i =  actual observation time series,  x i ^ =  estimated time series.
This paper defined the MAPE accuracy (%) by  MAPE   ( % ) = 100 MAPE .

3.3.3. Mean Absolute Error (MAE)

The mean absolute error (MAE) is frequently used as a statistical measure of the average magnitude of the errors in a predicted dataset without considering their direction. It is the average over the test sample of the absolute differences between prediction and actual observation, where all individual differences have equal weight. Generally, MAE is defined by the following equation:
MAE = 1 n i = 1 N | x i x i ^ |
where i =  variable,  N =  number of non-missing data points,  x i =  actual observation time series,  x i ^ =  estimated time series.

3.3.4. Modified Diebold–Mariano Test

The DM test was originally introduced by Diebold and Mariano (1995). In empirical analyses, when there are two or more time series forecasting models, it is often a challenge to predict which model is more accurate or whether they are equally suitable. This test identifies whether the null hypothesis (i.e., that the competing model holds equivalent forecasting power as the base model) is statistically true. Assuming that the actual values { y t ; t = 1 , T ] , two forecasts { y ^ 1 t ; t = 1 , T ] , { y ^ 2 t ; t = 1 , T ] , and forecast error ε i t are as follows:
ε i t = y ^ i t y t , i = 1 , 2
where ε i t  denotes the forecast error and the loss function,  g ( ε i t ) , which is defined by the following function:
g ( ε i t )   = ( ε i t ) 2
Then, the loss differential d t is expressed as follows:
d t = g ( ε 1 t ) g ( ε 2 t )
Correspondingly, the statistic for the DM test is expressed using the following formula:
DM = d ¯ s N
where d ¯ , s ,   and   N  denote the mean loss differential, the variation of  d t , and the number of data points, respectively.
The null hypothesis is H 0 : E [ d t ] = 0 , t , meaning that the two forecast models hold equivalent forecasting performance. Meanwhile, the alternative hypothesis is H 1 : E [ d t ] 0 , t , which represents the difference in accuracy between these two forecasts. Under the null hypothesis, the statistics for the DM test are asymptotically N ( 0 , 1 ) normally distributed. The null hypothesis would be rejected when DM > 1.96 .
Harvey et al. (1997) proposed a modified DM test. They suggested that the modified DM test is more suitable when using a small sample. The statistic for the modified DM test is expressed as follows:
DM = [ n + 1 2 h + h ( h 1 ) ] n 1 DM
where h represents the horizon and DM refers to the original DM statistic. Here, we predicted one-period-ahead; hence, h = 1; hence,
DM = ( n 1 ) n 1 DM
Concerning how to interpret the DM test statistic results, since we set g ( ε 1 t )  as the target model,  g ( ε 2 t )  as the base model, the numerator is (target-base), therefore, if the DM test statistic is negative, that means the target model has a smaller variance than the base model; hence, the prediction performance of the target model is better than the base model. The p-value denotes the significance of this statistic.

4. Results

4.1. Empirical Results

4.1.1. Prediction Results of SI Dataset and Large Dataset

Firstly, this subsection presents the prediction performance results of the sentiment dataset and the large dataset to verify whether the sentiment dataset could replace the large dataset when predicting commodity gold prices and the euro foreign exchange rate.
Table 2 displays the prediction outcomes for gold futures prices utilizing the sentiment indicator dataset, while Table 3 presents the prediction results for gold futures prices employing the large dataset. Likewise, Table 4 lists the prediction results for the euro foreign exchange rate based on the sentiment indicator dataset, and Table 5 showcases the prediction results for the euro foreign exchange rate utilizing the large dataset. Overall, the prediction results indicate that the sentiment indicator dataset generally exhibits better forecasting performance than the large dataset. When comparing the performance metrics, namely RMSE, MAPE, and MSE, between the two datasets, it becomes evident that the fixed moving window LSTM approach using the SI dataset outperforms the alternative dataset and models considered. This finding suggests that combining the sentiment indicator with the moving window LSTM machine learning model yields the best results for predicting gold futures prices and euro exchange rates. These results align with the outcomes of previous studies by Plakandaras et al. (2015), Nwosu et al. (2021), and Dunis and Williams (2002), which suggest that neural network models or their proposed approaches, particularly when combined with neural networks, offer more accurate forecasts compared to other models. Furthermore, these results provide additional evidence supporting the superiority of the LSTM model’s complex loop structure. Turning to the forecasting results using the large dataset, the moving window RF results demonstrate the best performance. This may be attributed to the use of a large indicator dataset, which allows the RF classifier to effectively enhance the predictive power. Although our study employs a different data source for sentiment analysis compared to previous research (Naeem et al. 2021), our empirical results broadly align with the findings of Li et al. (2016) and Naeem et al. (2021) in terms of predicting gold futures and euro exchange rates, thus indicating that the sentiment dataset can serve as a viable substitute for the large dataset.

4.1.2. Prediction Results of ARIMA Model

However, when comparing with the classical statistical model, ARIMA, whether the conclusion holds robustness needs to be investigated. Therefore, we conducted the simple prediction by ARIMA, and the lags were chosen using the Akaike Information Criteria (AIC). The forecasting results are presented in Table 6.
Based on the above results, we are pleasantly surprised by the effectiveness of the powerful yet simple statistical model, ARIMA, in predicting time series. This finding aligns with the research reported by He (2018). However, it contradicts the studies conducted by Siami-Namini et al. (2018) and Siami-Namini et al. (2019). These results suggest that simplicity may be the key when it comes to designing prediction models for time series, despite the prevalence of complex models and fancy datasets. In contrast to the findings of Nwosu et al. (2021) and Dunis and Williams (2002), our results indicate that it is worth considering the use of simple traditional models in the design of prediction models.

4.1.3. Prediction Results of Proposed Approaches

However, it is worth noting that machine learning and deep learning models have been extensively validated in numerous studies for their superior effectiveness and accuracy in predicting time series compared to ARIMA models. Therefore, it is necessary to further verify the robustness of the simple statistical model, ARIMA. Inspired by Abdulrahman et al. (2021) and others, we propose a triple combination of wavelet-ARIMA-LSTM, wavelet-ARIMA-RF, and wavelet-ARIMA-XGB models, as well as the seasonal-decomposition-ARIMA-LSTM approach, to investigate this objective. The prediction results are summarized in Table 7 and Table 8.
Based on the results presented in Table 7 and Table 8, Figure 6 and Figure 7, our proposed triple-combination approach demonstrates superior prediction accuracy compared to individual ARIMA, machine learning, and deep learning approaches. This suggests that by decomposing time series into linear and nonlinear components and combining classical statistical models with machine learning approaches, we achieve more precise predictions. However, the best performing approach for both object time series, namely gold futures prices and the euro foreign exchange rate, is the SeasonalDecomposition_ARIMA_LSTM model. It is followed by Wavelet_ARIMA_XGB and Wavelet_ARIMA_RF. This finding suggests that the systematic and non-systematic decomposition combined with the ARIMA and LSTM models for predicting commodity prices and foreign exchange rates is preferable. These results align with previous studies (Chang et al. 2019; Chen and Wang 2019; Liu et al. 2018; Ma et al. 2019; Moustafa and Khodairy 2023), further supporting the effectiveness of the integrated multiple-model approach in prediction. Our empirical forecasting results provide additional evidence that the multiple-model integrated approach performs better in prediction.
In summary, first, the combination of the sentiment indicator with the fixed moving window LSTM machine learning model produces the best prediction results compared to the large dataset. This result demonstrates that sentiment indicators obtained through sentiment analysis outperform the large dataset in terms of prediction ability and can be utilized as a better alternative independent predictor. Second, based on the prediction results, the traditional and classical ARIMA model surprisingly outperforms both the sentiment indicator dataset and the large dataset combined with machine learning techniques. Finally, our proposed triple-combination techniques are superior to both machine learning models and the traditional statistical ARIMA model in terms of commodity price and foreign exchange rate prediction performance. The top three performing forecasting methods are the seasonal-decomposition_ARIMA_LSTM, the wavelet_ARIMA_XGB, and the wavelet_ARIMA_RF. In the first step, these approaches decompose the data into linear and nonlinear components by adopting seasonal decomposition or wavelet transformation. In the second step, they use the ARIMA model to predict the linear part and machine learning or deep learning models to predict the nonlinear part.

4.2. Model Evaluation Results

4.2.1. Walk-Forward Testing Results

In this study, we employ the walk-forward testing method as the chosen back-testing technique to validate the effectiveness of the proposed triple-combination approaches. To evaluate the performance of these models, we adopt an expanding moving window approach, focusing on the last 50 observations. The testing procedure involves conducting separate walk-forward tests on each decomposition component, followed by aggregating the results and comparing the error metrics against those obtained from the ARIMA model.
As we present in Table 9 and Table 10, the walk-forward testing results for gold futures prices and euro foreign exchange provide robust estimations for evaluating the effectiveness of our proposed triple-combination approaches. These results offer valuable insights into the performance and reliability of the models in predicting the respective market dynamics.

4.2.2. Diebold–Mariano Test Results

The Diebold–Mariano test is conducted to assess the predictive superiority of the triple-combination approaches compared to the ARIMA models. We present the results of this test in Table 10 and Table 11, offering insights into the relative performance of the proposed approaches. The DM test results for both gold futures prices and euro foreign exchange rates are analyzed.
From the results presented in Table 11 and Table 12, it is noteworthy that the proposed triple-combination approaches demonstrate a significant outperformance over the classical statistical model, the ARIMA model.

5. Conclusions and Policy Implications

As highlighted by Naeem et al. (2021) and Li et al. (2016), the rapid advancement of the Internet and big data technology has led to an abundance of online data, including textual data from sources such as Twitter and news releases, which can help to identify influential factors in specific markets. Motivated by this, our study aims to examine whether the sentiment indicator dataset obtained through sentiment analysis of unstructured online news headlines can serve as a substitute for the large dataset comprising various indicators in predicting commodity prices and foreign exchange rates.
In our empirical analysis, we employ sentiment analysis using the Python natural language processing library to process news headlines from ABC, which consists of 1,226,258 news headlines, to derive a sentiment indicator. Additionally, we collect 30 additional indicators to construct the large dataset. Subsequently, we utilize this sentiment indicator in conjunction with moving window machine learning and deep learning models, namely RF, XGBoost, and LSTM, to forecast commodity gold futures prices and the euro exchange rate. Alongside comparing the prediction performance of the datasets, we also conduct a prediction comparison between the classical statistical model, ARIMA, and time-varying parameter machine learning models.
Based on the results of the model comparisons, we cannot conclude that sentiment indicators combined with machine learning outperform the ARIMA model. However, from an alternative perspective, we propose triple-combination approaches that involve decomposing the time series data into linear and nonlinear components and subsequently forecasting the linear component using the robust statistical model, ARIMA, and the nonlinear component using machine learning models such as LSTM, XGB, and RF. This research sheds light on the issue of comparing the out-of-sample superiority of our proposed triple-combination approaches for foreign exchange rate prediction with the traditional powerful statistical model, ARIMA. Furthermore, we conduct walk-forward testing to validate the triple-combination approaches and employ the modified Diebold–Mariano test statistic to investigate statistically significant differences between the proposed approach and the ARIMA model.
The study’s primary conclusions are as follows: Firstly, the combination of the sentiment indicator with the moving window LSTM machine learning model demonstrates the best forecasting performance. These findings align with previous studies conducted by Plakandaras et al. (2015), Nwosu et al. (2021), and Dunis and Williams (2002). Secondly, the sentiment indicator dataset used by deep learning and moving window machine learning models does not surpass the classical ARIMA model, consistent with the findings reported by He (2018). This result contradicts the studies conducted by Siami-Namini et al. (2018) and Siami-Namini et al. (2019). Thirdly, the proposed triple-combination methods, which expand upon and derive from the approaches of Chang et al. (2019), Chen and Wang (2019), Liu et al. (2018), Ma et al. (2019), and Moustafa and Khodairy (2023), exhibit superior performance in predicting commodity prices and foreign exchange rates compared to both machine learning models and the ARIMA model. The seasonal-decomposition ARIMA-LSTM, wavelet-ARIMA-XGB, and wavelet-ARIMA-RF demonstrate the top three forecasting performances based on error metrics, walk-forward testing results, and Diebold–Mariano test results. In the first step, the data are decomposed into linear and non-linear components using wavelet transformation or seasonal decomposition. In the second step, the linear component is predicted using ARIMA, while the non-linear component is predicted using machine learning or deep learning models. Lastly, in addition to the aforementioned findings, the comparison of results between the sentiment indicator dataset and the large dataset indicate that sentiment indicators obtained through sentiment analysis possess superior forecasting capabilities compared to the large dataset consisting of various indicators. Consequently, they can be utilized as better alternative predictors. Our empirical results generally align with the findings of Li et al. (2016) and Naeem et al. (2021) in terms of predicting gold futures and euro exchange rates, further highlighting the potential of the sentiment dataset to enhance forecasting in time series prediction.
To the best of our knowledge, this study presents a pioneering investigation into the potential of sentiment indicators as a substitute for extensive datasets in forecasting commodity prices and foreign exchange rates. The novelty lies in proposing a novel integration of machine learning models, statistical models, and data decomposition techniques to enhance price predictions in these markets. Importantly, the results validate the superior accuracy of the proposed triple-combination approach compared to individual models. Furthermore, these findings offer valuable insights for investors and policymakers, providing them with fresh perspectives, predictive tools, and alternative forecasting approaches.
For investors, the research offers fresh perspectives on forecasting commodity prices and foreign exchange rates. It introduces new predictive tools and alternative approaches that enhance their decision-making processes and potentially lead to more accurate forecasts. Additionally, precise prediction of gold prices and euro exchange rates is crucial for informing hedging strategies aimed at mitigating risks arising from currency fluctuations.
For policymakers, these findings play a vital role in making informed investment decisions. Gold is widely utilized as a means to hedge against inflation and market volatility, and fluctuations in the euro exchange rate have a substantial impact on the costs and risks associated with international transactions as the second most traded currency globally. Moreover, improving the accuracy of gold price predictions is crucial for central banks that maintain gold reserves as a safeguard against currency fluctuations and as a store of value. Given that the euro is a major reserve currency used in international transactions and investments, precise prediction of euro exchange rates can bolster financial stability. Furthermore, gold prices and euro exchange rates are closely intertwined with the international economy and play a pivotal role in informing government policy decisions regarding trade, monetary policy, and capital flows. Therefore, our findings contribute to economic forecasting, empowering policymakers and investors to leverage these predictions for informed decision making, ensuring they are well-prepared to navigate and respond to evolving economic conditions.
Despite these findings, this study has its limitations. Since we only employ RF, XGB, and LSTM methods to compare forecasts with the ARIMA model, we cannot conclusively determine that ARIMA is superior to other machine learning and deep learning models. Further verification is necessary to address this point. Additionally, there are numerous other data decomposition methods that require testing to validate the conclusions.
In future research, it is recommended to explore alternative data decomposition methods, as well as additional machine learning and deep learning techniques, to expand the investigation to major commodity prices and currency exchange rates. This will help validate the rationality and robustness of the proposed approaches’ superiority. Furthermore, considering the potential of the sentiment indicator as a promising alternative dataset, empirical testing is planned to assess whether incorporating the proposed approaches with the additional sentiment indicator can further enhance the forecasting accuracy for commodity prices and foreign exchange rates.

Author Contributions

Investigation, J.S.; writing—original draft preparation, J.S.; writing—review and editing, S.H.; project administration, S.H.; funding acquisition, S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by JSPS KAKENHI [(grant number: 22K01424)].

Data Availability Statement

Not applicable.

Acknowledgments

We are grateful to four anonymous reviewers for their helpful comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The descriptions of and sources of the data are presented in Table A1.
Table A1. Descriptions and sources of the indicators used in this study.
Table A1. Descriptions and sources of the indicators used in this study.
VariableDescriptionSource
EUREuro against the US dollarInvesting.com
CADCanadian dollar against the US dollarInvesting.com
JPYJapanese yen against the US dollarInvesting.com
WTIf WTI Crude Oil futures pricesBloomberg
Brent_oilBrent Crude Oil futures pricesInvesting.com
Henryhub_gasHenry Hub Natural Gas futures pricesBloomberg
SP500Standard & Poor’s 500 Stock IndexFRB 1
FTSE100The Financial Times Stock Exchange Group:London Stock ExchangeFRB 1
NASDAQNASDAQ Composite IndexFRB 1
HangSengHong Kong Hang Seng Composite stock market indexMacrotrends
CAC40France’s CAC 40 stock market indexMacrotrends
GSPTSECanadian S&P/TSX Composite IndexInvesting.com
US10_BondUS 10-Year Treasury Constant Maturity RateYahoo! Finance
UK10_BondUnited Kingdom 10-Year Bond YieldInvesting.com
Germany10_BondGermany 10-Year Bond YieldInvesting.com
DAXGermany’s DAX 30 stock market indexMacrotrends
NIKKEITokyo Stock Exchange:Nikkei indexFRB 1
GoldGold futures pricesBloomberg
TWUSDITrade Weighted U.S. Dollar IndexFRB 1
FederalFundsFederal Funds RateMacrotrends
CORNCorn futures pricesDatastream 2
WHEATWheat futures pricesDatastream 2
RSIRelative Strength IndexCalculated
ma77-days Moving AverageCalculated
ma2121-days Moving AverageCalculated
26ema26-days Exponential Weighted Moving AverageCalculated
12ema12-days Exponential Weighted Moving AverageCalculated
MACDMoving Average Convergence/Divergence oscillatorCalculated
20sd20-days Standard DeviationCalculated
upper_bandBollinger BandsCalculated
lower_bandBollinger BandsCalculated
emaExponential Moving AverageCalculated
Note: 1 Federal Reserve Bank. 2 Thomson Reuters Datastream.

References

  1. Abdulrahman, Umar Farouk Ibn, Najim Ussiph, and Benjamin Hayfron-Acquah. 2021. A Hybrid Arima-Lstm Model for Stock Price Prediction. International Journal of Computer Engineering and Information Technology 12: 48–51. Available online: https://www.proquest.com/openview/288bcbf49b187672d89c8a93865cc9d0/1?pq-origsite=gscholar&cbl=2044551 (accessed on 25 April 2023).
  2. Aizenman, Joshua, and Kenta Inoue. 2013. Central Banks and Gold Puzzles. Journal of the Japanese and International Economies 28: 69–90. [Google Scholar] [CrossRef] [Green Version]
  3. Amat, Christophe, Tomasz Michalski, and Gilles Stoltz. 2018. Fundamentals and Exchange Rate Forecastability with Simple Machine Learning Methods. Journal of International Money and Finance 88: 1–24. [Google Scholar] [CrossRef]
  4. Amato, Jeffery D., Andrew J. Filardo, Gabriele Galati, Goetz von Peter, and Feng Zhu. 2005. Research on Exchange Rates and Monetary Policy: An Overview. SSRN Electronic Journal. [Google Scholar] [CrossRef] [Green Version]
  5. Bakay, Melahat Sevgül, and Ümit Ağbulut. 2021. Electricity Production Based Forecasting of Greenhouse Gas Emissions in Turkey with Deep Learning, Support Vector Machine and Artificial Neural Network Algorithms. Journal of Cleaner Production 285: 125324. [Google Scholar] [CrossRef]
  6. Baranochnikov, Illia, and Robert Ślepaczuk. 2022. A Comparison of LSTM and GRU Architectures with the Novel Walk-Forward Approach to Algorithmic Investment Strategy. No. 2022-21. Warsaw: QFRG. [Google Scholar]
  7. Bedi, Punam, and Purnima Khurana. 2019. Sentiment Analysis Using Fuzzy-Deep Learning. In Proceedings of ICETIT 2019: Emerging Trends in Information Technology. Berlin: Springer, pp. 246–57. [Google Scholar] [CrossRef]
  8. Blose, Laurence E. 2010. Gold Prices, Cost of Carry, and Expected Inflation. Journal of Economics and Business 62: 35–47. [Google Scholar] [CrossRef]
  9. Bollen, Johan, Huina Mao, and Xiaojun Zeng. 2011. Twitter Mood Predicts the Stock Market. Journal of Computational Science 2: 1–8. [Google Scholar] [CrossRef] [Green Version]
  10. Bouktif, Salah, Ali Fiaz, Ali Ouni, and Mohamed Serhani. 2018. Optimal Deep Learning LSTM Model for Electric Load Forecasting Using Feature Selection and Genetic Algorithm: Comparison with Machine Learning Approaches. Energies 11: 1636. [Google Scholar] [CrossRef] [Green Version]
  11. Box, George E. P., and Gwilym M. Jenkins. 1968. Some Recent Advances in Forecasting and Control. Applied Statistics 17: 91. [Google Scholar] [CrossRef]
  12. Breiman, Leo. 2001. Random forests. Machine Learning 45: 5–32. [Google Scholar] [CrossRef] [Green Version]
  13. Cao, Jian, Zhi Li, and Jian Li. 2019. Financial Time Series Forecasting Model Based on CEEMDAN and LSTM. Physica A: Statistical Mechanics and Its Applications 519: 127–39. [Google Scholar] [CrossRef]
  14. Chang, Zihan, Yang Zhang, and Wenbo Chen. 2019. Electricity Price Prediction Based on Hybrid Model of Adam Optimized LSTM Neural Network and Wavelet Transform. Energy 187: 115804. [Google Scholar] [CrossRef]
  15. Chatzis, Sotirios P., Vassilis Siakoulis, Anastasios Petropoulos, Evangelos Stavroulakis, and Nikos Vlachogiannakis. 2018. Forecasting Stock Market Crisis Events Using Deep and Statistical Machine Learning Techniques. Expert Systems with Applications 112: 353–71. [Google Scholar] [CrossRef]
  16. Chen, Tianqi, and Carlos Guestrin. 2016. Xgboost: A scalable tree boosting system. Paper presented at the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, Washington DC, USA, August 14–18; pp. 785–94. [Google Scholar]
  17. Chen, Yuwei, and Kaizhi Wang. 2019. Prediction of Satellite Time Series Data Based on Long Short Term Memory-Autoregressive Integrated Moving Average Model (LSTM-ARIMA). Paper presented at THE 2019 IEEE 4th International Conference on Signal and Image Processing (ICSIP), Wuxi, China, July 19–21. [Google Scholar]
  18. Chimmula, Vinay Kumar Reddy, and Lei Zhang. 2020. Time Series Forecasting of COVID-19 Transmission in Canada Using LSTM Networks. Chaos, Solitons & Fractals 135: 109864. [Google Scholar] [CrossRef]
  19. Chua, Jess, and Richard S. Woodward. 1982. Gold as an inflation hedge: A comparative study of six major industrial countries. Journal of Business Finance & Accounting 9: 191–97. [Google Scholar] [CrossRef]
  20. Darley, Olufunke G., Abayomi I. O. Yussuff, and Adetokunbo A. Adenowo. 2021. Price Analysis and Forecasting for Bitcoin Using Auto Regressive Integrated Moving Average Model. Annals of Science and Technology 6: 47–56. [Google Scholar] [CrossRef]
  21. Das, Sushree, Ranjan Kumar Behera, Mukesh Kumar, and Santanu Kumar Rath. 2018. Real-Time Sentiment Analysis of Twitter Streaming Data for Stock Prediction. Procedia Computer Science 132: 956–64. [Google Scholar] [CrossRef]
  22. Dave, Emmanuel, Albert Leonardo, Marethia Jeanice, and Novita Hanafiah. 2021. Forecasting Indonesia Exports Using a Hybrid Model ARIMA-LSTM. Procedia Computer Science 179: 480–87. [Google Scholar] [CrossRef]
  23. Deeney, Peter, Mark Cummins, Michael Dowling, and Adam Bermingham. 2015. Sentiment in Oil Markets. International Review of Financial Analysis 39: 179–85. [Google Scholar] [CrossRef] [Green Version]
  24. Diebold, Francis X, and Robert S Mariano. 1995. Comparing Predictive Accuracy. Journal of Business and Economic Statistics 13: 253–63. [Google Scholar] [CrossRef]
  25. Dunis, Christian, and Mark Williams. 2002. Modelling and trading the EUR/USD exchange rate: Do neural network models perform better? Derivatives Use, Trading and Regulation 8: 211–39. [Google Scholar]
  26. Farsi, Behnam, Manar Amayri, Nizar Bouguila, and Ursula Eicker. 2021. On Short-Term Load Forecasting Using Machine Learning Techniques and a Novel Parallel Deep LSTM-CNN Approach. IEEE Access 9: 31191–212. [Google Scholar] [CrossRef]
  27. Friedman, Jerome H. 2001. Greedy function approximation: A gradient boosting machine. Annals of Statistics 29: 1189–232. [Google Scholar] [CrossRef]
  28. Guo, Chenkai, Xiaoyu Yan, and Yan Li. 2020. Prediction of Student Attitude towards Blended Learning Based on Sentiment Analysis. Paper presented at THE 2020 9th International Conference on Educational and Information Technology, Oxford, UK, February 11–13; pp. 228–33. [Google Scholar]
  29. Harvey, David, Stephen Leybourne, and Paul Newbold. 1997. Testing the Equality of Prediction Mean Squared Errors. International Journal of Forecasting 13: 281–91. [Google Scholar] [CrossRef]
  30. He, Xin James. 2018. Crude Oil Prices Forecasting: Time Series vs. SVR Models. Journal of International Technology and Information Management 27: 25–42. [Google Scholar] [CrossRef]
  31. Henry, Elaine, and Andrew J. Leone. 2016. Measuring Qualitative Information in Capital Markets Research: Comparison of Alternative Methodologies to Measure Disclosure Tone. The Accounting Review 91: 153–78. [Google Scholar] [CrossRef]
  32. Hochreiter, Sepp, and Jürgen Schmidhuber. 1997. Long Short-Term Memory. Neural Computation 9: 1735–80. [Google Scholar] [CrossRef]
  33. Ito, Tomoki, Kota Tsubouchi, Hiroki Sakaji, Kiyoshi Izumi, and Tatsuo Yamashita. 2019. CSNN: Contextual Sentiment Neural Network. Paper presented at International Conference on Data Mining, Beijing, China, November 8–11. [Google Scholar]
  34. Ito, Tomoki, Kota Tsubouchi, Hiroki Sakaji, Tatsuo Yamashita, and Kiyoshi Izumi. 2020. Contextual Sentiment Neural Network for Document Sentiment Analysis. Data Science and Engineering 5: 180–92. [Google Scholar] [CrossRef]
  35. Júnior, Domingos S. de O. Santos, João F. L. de Oliveira, and Paulo S. G. de Mattos Neto. 2019. An Intelligent Hybridization of ARIMA with Machine Learning Models for Time Series Forecasting. Knowledge-Based Systems 175: 72–86. [Google Scholar] [CrossRef]
  36. Kulkarni, Rohit. 2018. A Million News Headlines. V6. Cambridge: Harvard Dataverse. Available online: https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/SYBGZL (accessed on 25 April 2023).
  37. Latief, Rashid, and Lin Lefen. 2018. The Effect of Exchange Rate Volatility on International Trade and Foreign Direct Investment (FDI) in Developing Countries along ‘One Belt and One Road. International Journal of Financial Studies 6: 86. [Google Scholar] [CrossRef] [Green Version]
  38. Li, Hongmin, Jianzhou Wang, and Hufang Yang. 2020. A Novel Dynamic Ensemble Air Quality Index Forecasting System. Atmospheric Pollution Research 11: 1258–70. [Google Scholar] [CrossRef]
  39. Li, Jian, Zhenjing Xu, Lean Yu, and Ling Tang. 2016. Forecasting Oil Price Trends with Sentiment of Online News Articles. Procedia Computer Science 91: 1081–87. [Google Scholar] [CrossRef] [Green Version]
  40. Liu, Hui, Xi-wei Mi, and Yan-fei Li. 2018. Wind Speed Forecasting Method Based on Deep Learning Strategy Using Empirical Wavelet Transform, Long Short Term Memory Neural Network and Elman Neural Network. Energy Conversion and Management 156: 498–514. [Google Scholar] [CrossRef]
  41. Liu, Xiaolei, Zi Lin, and Ziming Feng. 2021. Short-Term Offshore Wind Speed Forecast by Seasonal ARIMA—A Comparison against GRU and LSTM. Energy 227: 120492. [Google Scholar] [CrossRef]
  42. Livieris, Ioannis E., Emmanuel Pintelas, and Panagiotis Pintelas. 2020. A CNN–LSTM Model for Gold Price Time-Series Forecasting. Neural Computing and Applications 32: 17351–60. [Google Scholar] [CrossRef]
  43. Luo, Zhaojie, Xiaojing Cai, Katsuyuki Tanaka, Tetsuya Takiguchi, Takuji Kinkyo, and Shigeyuki Hamori. 2019. Can We Forecast Daily Oil Futures Prices? Experimental Evidence from Convolutional Neural Networks. Journal of Risk and Financial Management 12: 9. [Google Scholar] [CrossRef] [Green Version]
  44. Ma, Rui, Zhongliang Li, Elena Breaz, Chen Liu, Hao Bai, Pascal Briois, and Fei Gao. 2019. Data-Fusion Prognostics of Proton Exchange Membrane Fuel Cell Degradation. IEEE Transactions on Industry Applications 55: 4321–31. [Google Scholar] [CrossRef]
  45. McNally, Sean, Jason Roche, and Simon Caton. 2018. Predicting the Price of Bitcoin Using Machine Learning. Paper presented at THE 2018 26th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP), Cambridge, UK, March 21–23. [Google Scholar]
  46. Morlet, Jetal, G. Arens, Eliane Fourgeau, and D. Giard. 1982. Wave Propagation and Sampling Theory—Part II: Sampling Theory and Complex Waves. Geophysics 47: 222–36. [Google Scholar] [CrossRef] [Green Version]
  47. Moustafa, Sayed S. R., and Sara S. Khodairy. 2023. Comparison of different predictive models and their effectiveness in sunspot number prediction. Physica Scripta 98: 045022. [Google Scholar] [CrossRef]
  48. Mukta, Md Saddam Hossain, Md Adnanul Islam, Faisal Ahamed Khan, Afjal Hossain, Shuvanon Razik, Shazzad Hossain, and Jalal Mahmud. 2022. A Comprehensive Guideline for Bengali Sentiment Annotation. ACM Transactions on Asian and Low-Resource Language Information Processing 21: 1–19. [Google Scholar] [CrossRef]
  49. Mussa, Michael. 1976. The Exchange Rate, the Balance of Payments and Monetary and Fiscal Policy under a Regime of Controlled Floating. The Scandinavian Journal of Economics 78: 229. [Google Scholar] [CrossRef]
  50. Naeem, Samreen, Wali Khan Mashwani, Aqib Ali, M. Irfan Uddin, Marwan Mahmoud, Farrukh Jamal, and Christophe Chesneau. 2021. Machine Learning-Based USD/PKR Exchange Rate Forecasting Using Sentiment Analysis of Twitter Data. Computers, Materials & Continua 67: 3451–61. [Google Scholar] [CrossRef]
  51. Nguyen, Thi Thu Giang, and Robert Ślepaczuk. 2022. The Efficiency of Various Types of Input Layers of LSTM Model in Investment Strategies on S&P500 Index. (No. 2022-29). St. Louis: Research Papers in Economics. [Google Scholar]
  52. Nwosu, Ugochinyere Ihuoma, Chukwudi Paul Obite, Prince Henry Osuagwu, and Obioma Gertrude Onukwube. 2021. Modeling the British Pound Sterling to Nigerian Naira Exchange Rate During the COVID-19 Pandemic. Journal of Mathematics and Statistics Studies 2: 25–35. [Google Scholar] [CrossRef]
  53. Pai, Ping-Feng, and Chia-Hsin Liu. 2018. Predicting Vehicle Sales by Sentiment Analysis of Twitter Data and Stock Market Values. IEEE Access 6: 57655–62. [Google Scholar] [CrossRef]
  54. Philander, Kahlil, and YunYing Zhong. 2016. Twitter Sentiment Analysis: Capturing Sentiment from Integrated Resort Tweets. International Journal of Hospitality Management 55: 16–24. [Google Scholar] [CrossRef]
  55. Phyo, Pyae-Pyae, Yung-Cheol Byun, and Namje Park. 2022. Short-Term Energy Forecasting Using Machine-Learning-Based Ensemble Voting Regression. Symmetry 14: 160. [Google Scholar] [CrossRef]
  56. Plakandaras, Vasilios, Theophilos Papadimitriou, and Periklis Gogas. 2015. Forecasting Daily and Monthly Exchange Rates with Machine Learning Techniques. Journal of Forecasting 34: 560–73. [Google Scholar] [CrossRef]
  57. Qiu, Yue, Zhewei Song, and Zhensong Chen. 2022. Short-Term Stock Trends Prediction Based on Sentiment Analysis and Machine Learning. Soft Computing 26: 2209–24. [Google Scholar] [CrossRef]
  58. Ratner, Mitchell, and Steven Klein. 2008. The Portfolio Implications of Gold Investment. The Journal of Investing 17: 77–87. [Google Scholar] [CrossRef]
  59. Raza, Syed Ali, Nida Shah, and Muhammad Shahbaz. 2018. Does Economic Policy Uncertainty Influence Gold Prices? Evidence from a Nonparametric Causality-In-Quantiles Approach. Resources Policy 57: 61–68. [Google Scholar] [CrossRef]
  60. Razzaq, Abdul, Muhammad Asim, Zulqrnain Ali, Salman Qadri, Imran Mumtaz, Dost Muhammad Khan, and Qasim Niaz. 2019. Text Sentiment Analysis Using Frequency-Based Vigorous Features. China Communications 16: 145–53. [Google Scholar] [CrossRef]
  61. Razzaque, Mohammad A., Sayema Haque Bidisha, and Bazlul Haque Khondker. 2017. Exchange Rate and Economic Growth. Journal of South Asian Development 12: 42–64. [Google Scholar] [CrossRef]
  62. Ribeiro, Andrea Maria N. C., Pedro Rafael X. do Carmo, Iago Richard Rodrigues, Djamel Sadok, Theo Lynn, and Patricia Takako Endo. 2020. Short-Term Firm-Level Energy-Consumption Forecasting for Energy-Intensive Manufacturing: A Comparison of Machine Learning and Deep Learning Models. Algorithms 13: 274. [Google Scholar] [CrossRef]
  63. Sadefo Kamdem, Jules, Rose Bandolo Essomba, and James Njong Berinyuy. 2020. Deep Learning Models for Forecasting and Analyzing the Implications of COVID-19 Spread on Some Commodities Markets Volatilities. Chaos, Solitons & Fractals 140: 110215. [Google Scholar] [CrossRef]
  64. Seals, Ethan, and Steven R. Price. 2020. Preliminary Investigation in the Use of Sentiment Analysis in Prediction of Stock Forecasting Using Machine Learning. Paper presented at 2020 SoutheastCon, Raleigh, NC, USA, March 28–29. [Google Scholar]
  65. Selvin, Sreelekshmy, R. Vinayakumar, E. A. Gopalakrishnan, Vijay Krishna Menon, and K. P. Soman. 2017. Stock Price Prediction Using LSTM, RNN and CNN-Sliding Window Model. Paper presented at the 2017 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Manipal, India, September 13–16. [Google Scholar]
  66. Sharma, Urvashi, Rattan K. Datta, and Kavita Pabreja. 2020. Sentiment Analysis and Prediction of Election Results 2018. In Social Networking and Computational Intelligence. Berlin: Springer, pp. 727–39. [Google Scholar] [CrossRef]
  67. Shih, Han, and Suchithra Rajendran. 2019. Comparison of Time Series Methods and Machine Learning Algorithms for Forecasting Taiwan Blood Services Foundation’s Blood Supply. Journal of Healthcare Engineering 2019: 6123745. [Google Scholar] [CrossRef]
  68. Siami-Namini, Sima, Neda Tavakoli, and Akbar Siami Namin. 2018. A Comparison of ARIMA and LSTM in Forecasting Time Series. Paper presented at 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, December 17–18. [Google Scholar]
  69. Siami-Namini, Sima, Neda Tavakoli, and Akbar Siami Namin. 2019. The Performance of LSTM and BiLSTM in Forecasting Time Series. Paper presented at the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, December 9–12. [Google Scholar]
  70. Sivri, Mahmut Sami, Alp Ustundag, and Buse Sibel Korkmaz. 2022. Ensemble Learning Based Stock Market Prediction Enhanced with Sentiment Analysis. Paper presented at the INFUS 2021 Conference, Intelligent and Fuzzy Techniques for Emerging Conditions and Digital Transformation, Izmir, Turkey, August 24–26; vol. 2, pp. 446–54. [Google Scholar]
  71. Smailović, Jasmina, Miha Grčar, Nada Lavrač, and Martin Žnidaršič. 2013. Predictive Sentiment Analysis of Tweets: A Stock Market Application. Paper presented at the Human-Computer Interaction and Knowledge Discovery in Complex, Unstructured, Big Data, Maribor, Slovenia, July 1–3; pp. 77–88. [Google Scholar] [CrossRef]
  72. Sun, Chen, Yingxiong Nong, Zhibin Chen, Dong Liang, Ying Lu, and Yishuang Qin. 2022. The CEEMD-LSTM-ARIMA Model and Its Application in Time Series Prediction. Journal of Physics: Conference Series 2179: 012012. [Google Scholar] [CrossRef]
  73. Vijayarani, S., and R. Janani. 2016. Text Mining: Open Source Tokenization Tools—An Analysis. Advanced Computational Intelligence: An International Journal (ACII) 3: 37–47. [Google Scholar] [CrossRef]
  74. Wang, Jie, and Jun Wang. 2016. Forecasting Energy Market Indices with Recurrent Neural Networks: Case Study of Crude Oil Price Fluctuations. Energy 102: 365–74. [Google Scholar] [CrossRef]
  75. Wu, Junhao, and Zhaocai Wang. 2022. A Hybrid Model for Water Quality Prediction Based on an Artificial Neural Network, Wavelet Transform, and Long Short-Term Memory. Water 14: 610. [Google Scholar] [CrossRef]
  76. Wu, Xianghua, Jieqin Zhou, Huaying Yu, Duanyang Liu, Kang Xie, Yiqi Chen, Jingbiao Hu, Haiyan Sun, and Feng-Juan Xing. 2021. The Development of a Hybrid Wavelet-ARIMA-LSTM Model for Precipitation Amounts and Drought Analysis. Atmosphere 12: 74. [Google Scholar] [CrossRef]
  77. Wysocki, Maciej, and Robert Ślepaczuk. 2022. Artificial Neural Networks Performance in WIG20 Index Options Pricing. Entropy 24: 35. [Google Scholar] [CrossRef] [PubMed]
  78. Xiang, Nan, Qianqian Jia, and Yuedong Wang. 2021. Sentiment Analysis of Chinese Weibo Combining BERT Model and Hawkes Process. Paper presented at the 2021 5th International Conference on Deep Learning Technologies (ICDLT), Qingdao, China, July 23–25. [Google Scholar]
  79. Xue, Sheng, Hualiang Chen, and Xiaoliang Zheng. 2022. Detection and quantification of anomalies in communication networks based on LSTM-ARIMA combined model. International Journal of Machine Learning and Cybernetics 13: 3159–72. [Google Scholar] [CrossRef] [PubMed]
  80. Yamak, Peter T., Li Yujian, and Pius K. Gadosey. 2019. A Comparison between ARIMA, LSTM, and GRU for Time Series Forecasting. Paper presented at the 2019 2nd International Conference on Algorithms, Computing and Artificial Intelligence, Sanya, China, December 20–22. [Google Scholar]
  81. Yu, He, Li Ming, Ruan Sumei, and Zhao Shuping. 2020. A Hybrid Model for Financial Time Series Forecasting—Integration of EWT, ARIMA with the Improved ABC Optimized ELM. IEEE Access 8: 84501–18. [Google Scholar] [CrossRef]
  82. Zhang, Qiang, Feng Li, Fei Long, and Qiang Ling. 2018. Vehicle Emission Forecasting Based on Wavelet Transform and Long Short-Term Memory Network. IEEE Access 6: 56984–94. [Google Scholar] [CrossRef]
  83. Zhang, Xiaoyu, Stefanie Kuenzel, Nicolo Colombo, and Chris Watkins. 2022. Hybrid Short-Term Load Forecasting Method Based on Empirical Wavelet Transform and Bidirectional Long Short-Term Memory Neural Networks. Journal of Modern Power Systems and Clean Energy 10: 1216–28. [Google Scholar] [CrossRef]
  84. Zhang, Yuchen, and Shigeyuki Hamori. 2020. The Predictability of the Exchange Rate When Combining Machine Learning and Fundamental Models. Journal of Risk and Financial Management 13: 48. [Google Scholar] [CrossRef] [Green Version]
  85. Zhao, Jiwei, Guangzheng Nie, and Yihao Wen. 2022. Monthly Precipitation Prediction in Luoyang City Based on EEMD-LSTM-ARIMA Model. Water Science and Technology 87: 318–35. [Google Scholar] [CrossRef]
  86. Zhou, Yong, Li Wang, and Junhao Qian. 2022. Application of Combined Models Based on Empirical Mode Decomposition, Deep Learning, and Autoregressive Integrated Moving Average Model for Short-Term Heating Load Predictions. Sustainability 14: 7349. [Google Scholar] [CrossRef]
  87. Zolfaghari, Mehdi, and Samad Gholami. 2021. A Hybrid Approach of Adaptive Wavelet Transform, Long Short-Term Memory and ARIMA-GARCH Family Models for the Stock Index Prediction. Expert Systems with Applications 182: 115149. [Google Scholar] [CrossRef]
Figure 1. Historical data plotting for gold futures price. Note: This figure illustrates the raw data of gold futures prices and the dotted line represents the train/test data.
Figure 1. Historical data plotting for gold futures price. Note: This figure illustrates the raw data of gold futures prices and the dotted line represents the train/test data.
Jrfm 16 00298 g001
Figure 2. Historical data plotting for euro price. Note: This figure illustrates the raw data of the euro exchange rate multiplied by 100, and the dotted line represents the train/test data.
Figure 2. Historical data plotting for euro price. Note: This figure illustrates the raw data of the euro exchange rate multiplied by 100, and the dotted line represents the train/test data.
Jrfm 16 00298 g002
Figure 3. Sentiment index plotting. Note: This figure illustrates the calculated sentiment index based on the results of sentiment analysis.
Figure 3. Sentiment index plotting. Note: This figure illustrates the calculated sentiment index based on the results of sentiment analysis.
Jrfm 16 00298 g003
Figure 4. Mechanism of the EMW. Note: The figure illustrates the iterative mechanism of the EMW when adopting an initial window size of three periods for one-period-ahead forecasting.
Figure 4. Mechanism of the EMW. Note: The figure illustrates the iterative mechanism of the EMW when adopting an initial window size of three periods for one-period-ahead forecasting.
Jrfm 16 00298 g004
Figure 5. Mechanism of the FMW. Note: The figure illustrates the iterative mechanism of the FMW when adopting an initial window size of three periods for one-period-ahead forecasting.
Figure 5. Mechanism of the FMW. Note: The figure illustrates the iterative mechanism of the FMW when adopting an initial window size of three periods for one-period-ahead forecasting.
Jrfm 16 00298 g005
Figure 6. Prediction result plotting for gold future prices. Note: This figure illustrates the gold futures prices and the predicted values.
Figure 6. Prediction result plotting for gold future prices. Note: This figure illustrates the gold futures prices and the predicted values.
Jrfm 16 00298 g006
Figure 7. Prediction result plotting for euro foreign exchange rates. Note: This figure illustrates the euro foreign exchange rates and the predicted values.
Figure 7. Prediction result plotting for euro foreign exchange rates. Note: This figure illustrates the euro foreign exchange rates and the predicted values.
Jrfm 16 00298 g007
Table 1. Datasets used to predict gold futures prices and the euro exchange rates.
Table 1. Datasets used to predict gold futures prices and the euro exchange rates.
Containing VariablesNumber of Variables
SI datasetToday’s price + Sentiment Indicator2
Large datasetToday’s price + Collected/Calculated Indicators + Sentiment Indicator33
Note: SI dataset represents the dataset comprising of today’s price and sentiment indicator. Large dataset represents the dataset comprising of today’s price, sentiment indicator, and collected/calculated indicators.
Table 2. Results of the SI dataset for gold futures prices.
Table 2. Results of the SI dataset for gold futures prices.
DatasetEvaluationRF_EMWRF_FMWXGBoost_EMWXGBoost_FMWLSTM_EMWLSTM_FMW
SI datasetRMSE10.312210.42619.97119.985211.14619.3283
MAPE0.58100.58320.54800.54800.61600.5130
MSE7.71597.71597.30157.30568.20746.8072
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. The underline followed by EMW denotes the expanding moving window technique, while the underline followed by FMW denotes the fixed moving window technique. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error. The best performance in this set of prediction results is shown in bold.
Table 3. Results of the large dataset for gold futures prices.
Table 3. Results of the large dataset for gold futures prices.
DatasetEvaluationRF_EMWRF_FMWXGBoost_EMWXGBoost_FMWLSTM_EMWLSTM_FMW
Large datasetRMSE9.84629.875210.625910.637314.026711.1016
MAPE0.54500.54700.59100.59300.82300.6340
MSE7.25447.25447.91007.953410.82888.3960
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. The underline followed by EMW denotes the expanding moving window technique, while the underline followed by FMW denotes the fixed moving window technique. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error. The best performance in this set of prediction results is shown in bold.
Table 4. Results of the SI dataset for euro foreign exchange rate.
Table 4. Results of the SI dataset for euro foreign exchange rate.
DatasetEvaluationRF_EMWRF_FMWXGBoost_EMWXGBoost_FMWLSTM_EMWLSTM_FMW
SI datasetRMSE0.5500.5530.5960.6030.4790.47384
MAPE0.3700.3720.3960.4000.3220.3180
MSE0.4300.4300.4600.4650.3740.3693
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. The underline followed by EMW denotes the expanding moving window technique, while the underline followed by FMW denotes the fixed moving window technique. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error. The best performance in this set of prediction results is shown in bold.
Table 5. Results of the large dataset for euro foreign exchange rate.
Table 5. Results of the large dataset for euro foreign exchange rate.
DatasetEvaluationRF_EMWRF_FMWXGBoost_EMWXGBoost_FMWLSTM_EMWLSTM_FMW
Large datasetRMSE0.5180.5210.7750.8700.6770.583
MAPE0.3430.3460.4840.5460.4120.395
MSE0.3990.3990.5630.6360.4760.458
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. The underline followed by EMW denotes the expanding moving window technique, while the underline followed by FMW denotes the fixed moving window technique. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error. The best performance in this set of prediction results is shown in bold.
Table 6. Results of the ARIMA for gold futures prices and the euro foreign exchange rate.
Table 6. Results of the ARIMA for gold futures prices and the euro foreign exchange rate.
EvaluationGoldEuro
RMSE9.26580.47388
MAPE0.50900.3170
MSE6.75910.3687
Note: RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error.
Table 7. Results of the proposed approaches for gold futures prices.
Table 7. Results of the proposed approaches for gold futures prices.
EvaluationSeasonalDecomposition_ARIMA_LSTMWavelet_ARIMA_LSTMWavelet_ARIMA_XGBWavelet_ARIMA_RFARIMALSTMXGBRF
RMSE3.39168.44395.46105.46109.265812.260510.031110.8282
MAPE0.00200.48400.30600.30600.50900.76700.54200.6140
MSE2.68696.43764.05164.05166.75919.98417.22838.1704
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error. The best performance in this set of prediction results is shown in bold. SeasonalDecomposition denotes the seasonal decomposition. Wavelet represents the wavelet decomposition.
Table 8. Results of the proposed approaches for euro foreign exchange rate.
Table 8. Results of the proposed approaches for euro foreign exchange rate.
EvaluationSeasonalDecomposition_ARIMA_LSTMWavelet_ARIMA_LSTMWavelet_ARIMA_XGBWavelet_ARIMA_RFARIMALSTMXGBRF
RMSE0.16320.40830.18130.34430.47390.55780.59520.6526
MAPE0.11200.26870.12000.24000.31700.39600.39500.4410
MSE0.12980.31220.13890.27840.36870.45730.45950.5122
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error. The best performance in this set of prediction results is shown in bold. SeasonalDecomposition denotes the seasonal decomposition. Wavelet represents the wavelet decomposition.
Table 9. Results of Walk-Forward Testing for gold futures prices.
Table 9. Results of Walk-Forward Testing for gold futures prices.
EvaluationSeasonalDecomposition_ARIMA_LSTMWavelet_ARIMA_LSTMWavelet_ARIMA_XGBWavelet_ARIMA_RFARIMA
RMSE2.87653.13083.55655.04269.2658
MAPE0.14450.16380.18680.26420.5090
MSE2.13042.42652.76823.91706.7591
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. SeasonalDecomposition denotes the seasonal decomposition. Wavelet represents the wavelet decomposition. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error.
Table 10. Results of Walk-Forward Testing for euro foreign exchange rate.
Table 10. Results of Walk-Forward Testing for euro foreign exchange rate.
EvaluationSeasonalDecomposition_ARIMA_LSTMWavelet_ARIMA_LSTMWavelet_ARIMA_XGBWavelet_ARIMA_RFARIMA
RMSE0.12630.10280.12060.31420.4739
MAPE0.09370.07620.08860.22260.3172
MSE0.10370.08440.09800.24660.3687
Note: RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. SeasonalDecomposition denotes the seasonal decomposition. Wavelet represents the wavelet decomposition. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error.
Table 11. DM test results of gold futures prices.
Table 11. DM test results of gold futures prices.
Target ApproachBase Model (ARIMA)
DM Testp-Value
SeasonalDecomposition_ARIMA_LSTM−9.9779 0.000
Wavelet_ARIMA_LSTM−9.4216 0.000
Wavelet_ARIMA_XGB−9.9468 0.000
Wavelet_ARIMA_RF−7.1182 0.000
Note: DM test indicates the modified Diebold–Mariano test statistic. RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. SeasonalDecomposition denotes the seasonal decomposition. SeasonalDecomposition denotes the seasonal decomposition. Wavelet represents the wavelet decomposition. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error.
Table 12. DM test results of euro foreign exchange rate.
Table 12. DM test results of euro foreign exchange rate.
Target ApproachBase Model (ARIMA)
DM Testp-Value
SeasonalDecomposition_ARIMA_LSTM−12.9469 0.000
Wavelet_ARIMA_LSTM−4.5330 0.000
Wavelet_ARIMA_XGB−12.6462 0.000
Wavelet_ARIMA_RF−6.3385 0.000
Note: DM test indicates the modified Diebold–Mariano test statistic. RF represents random forest. XGBoost denotes eXtreme gradient boosting. LSTM denotes long short-term memory. SeasonalDecomposition denotes the seasonal decomposition. SeasonalDecomposition denotes the seasonal decomposition. Wavelet represents the wavelet decomposition. RMSE denotes the root mean squared error. MAPE denotes the mean absolute percentage error. MAE denotes the mean absolute error.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shang, J.; Hamori, S. Do Large Datasets or Hybrid Integrated Models Outperform Simple Ones in Predicting Commodity Prices and Foreign Exchange Rates? J. Risk Financial Manag. 2023, 16, 298. https://doi.org/10.3390/jrfm16060298

AMA Style

Shang J, Hamori S. Do Large Datasets or Hybrid Integrated Models Outperform Simple Ones in Predicting Commodity Prices and Foreign Exchange Rates? Journal of Risk and Financial Management. 2023; 16(6):298. https://doi.org/10.3390/jrfm16060298

Chicago/Turabian Style

Shang, Jin, and Shigeyuki Hamori. 2023. "Do Large Datasets or Hybrid Integrated Models Outperform Simple Ones in Predicting Commodity Prices and Foreign Exchange Rates?" Journal of Risk and Financial Management 16, no. 6: 298. https://doi.org/10.3390/jrfm16060298

APA Style

Shang, J., & Hamori, S. (2023). Do Large Datasets or Hybrid Integrated Models Outperform Simple Ones in Predicting Commodity Prices and Foreign Exchange Rates? Journal of Risk and Financial Management, 16(6), 298. https://doi.org/10.3390/jrfm16060298

Article Metrics

Back to TopTop