Next Article in Journal
Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting
Previous Article in Journal
An Adaptive Chirp Mode Decomposition-Based Method for Modal Identification of Time-Varying Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting ETF Performance: A Comparative Study of Deep Learning Models and the Fama-French Three-Factor Model

1
Doctorate Program in Intelligent Banking and Finance, CTBC Business School, Tainan 709, Taiwan
2
Department of Banking and Finance, Chinese Culture University, Taipei 111396, Taiwan
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(19), 3158; https://doi.org/10.3390/math12193158
Submission received: 13 September 2024 / Revised: 2 October 2024 / Accepted: 7 October 2024 / Published: 9 October 2024
(This article belongs to the Section E5: Financial Mathematics)

Abstract

:
The global financial landscape has witnessed a significant shift towards Exchange-Traded Funds (ETFs), with their market capitalization surpassing USD 10 trillion in 2023, due to advantages such as low management fees, high liquidity, and broad market exposure. As ETFs become increasingly central to investment strategies, accurately forecasting their performance has become crucial. This study addresses this need by comparing the efficacy of deep learning models against the traditional Fama-French three-factor model in predicting daily ETF returns. The methodology employs eight artificial neural network architectures, including ANN, LSTM, GRU, CNN, and their variants, implemented in Python and applied to data ranging from 2010 to 2020, while also exploring the impact of additional factors on forecast accuracy. Empirical results reveal that LSTM and the Fama-French three-factor model exhibit a superior performance in ETF return prediction. This study contributes to the literature on financial forecasting and offers practical insights into investment decision making. By leveraging advanced artificial intelligence techniques, this study aims to enhance the toolkit available for ETF performance analysis, potentially improving investment strategies in this dynamic market segment.

1. Introduction

In the financial market, all investors pay special attention to returns on assets, because they want to target more profits. As such, the variables that affect the rate of return have become research topics. The capital asset pricing model (CAPM) is the first model about variables that affect the rate of return [1] (pp. 425–442) and assumes that the expected return required by investors is composed of risk-free return (such as government bonds and fixed deposits) and risk return. Risks are systemic and non-diversifiable and have comprehensive effects on the whole market. In terms of non-systemic risks, differences in individual assets are reduced by holding investment portfolios, and the effects of non-systemic risk are reduced to zero as portfolios become more diversified [2] (pp. 77–91). Therefore, it is unnecessary to consider the effects of a single company’s poor operation or a single industry.
This study aims to compare the effectiveness of traditional financial models (such as the Fama-French three-factor model) and neural network models in predicting ETF returns. Our goal is not limited to determining which model performs better, but also explores the impact of different deep learning architectures on the accuracy of ETF return predictions. In addition, we investigate whether adding specific factors (e.g., size and momentum) to these models can improve their predictive capabilities in the context of the Taiwan ETF market.
The risks of the CAPM are obtained by multiplying the market risk premium by β, where β is used to measure the risk. However, there are basic assumptions on investors and capital market in studies on the CAPM, including that all investors are rational and aim to maximize assets; their expected returns on financial assets are the same; the capital market must be open and transparent, where investors can obtain the same information at no cost; the market is in perfect competition; and investors can buy any number of assets in the market, but do not have effects on the price of assets, without any cost or restriction. The following study is formed on these basic assumptions.
The CAPM has great effects, but the idea of explaining returns with a single variable is questionable. By an empirical analysis of stocks ranging from 1962 to 1972 based on the arbitrage pricing theory (APT), [3] (pp. 75–89) did not demonstrate which variable is related to returns, but found that there are three or four explanatory variables for the rate of return. [4] (pp. 3–18) presented the scale effect, and [5] (pp. 103–126) found that the book-to-market ratio leads to a high return, therefore leading to the three-factor model [6] (pp. 427–465). Many years later, Ref. [7] (pp. 1653–1678) found that the three-factor model is unable to explain profitability and investment. Hence, Ref. [8] (pp. 1–22) published the five-factor model, adding the profitability factor and investment factor on the basis of the three-factor model. However, the Fama-French five-factor model does not outperform the Fama-French three-factor model in every market. According to the study of [9] (pp. 199–222), the Fama-French three-factor model is effective in Latin America and Europe but does not have good results in the Asian market, and in fact, the Fama-French three-factor model is better.
Among the studies on the forecast of daily return, the most common ones are the comparison between the non-linear artificial neural network and the traditional linear theory and the comparison between the artificial neural network (ANN) and the arbitrage pricing model. By explaining the Colombian stock market with macroeconomic variables and stock market indices of various countries, it was found that artificial neural network models are more accurate [10] (pp. 41–73). The capital asset pricing model (CAPM) and the Fama-French three-factor model showed the same results ([11] (pp. 53–60); Ref. [12] (pp. 59–87)), and the forecasting ability of the Fama-French five-factor model is not as good as that of the Fama-French three-factor model [13]. In particular, according to the study of [14] (pp. 2499–2512) on the China stock market, the combination of the CAPM’s variables and the ANN has stronger forecasting abilities than the combination of the Fama-French three-factor model’s variables and the ANN. However, most studies show that the combination of the Fama-French three-factor model’s variables and the ANN had the best forecasting abilities.
According to the abovementioned studies, artificial neural networks using variables of traditional models have better results than traditional models, but simple artificial neural network models are used, and no comparison is made with other artificial neural network models. In terms of the forecasting results of artificial neural networks, ref. [15] (pp. 673–688) showed that artificial neural networks have better forecasting abilities in high-beta portfolios than in low-beta portfolios. Ref. [16] (pp. 168–191), using ANNs and variables of the Fama-French five-factor model, noted that artificial neural network models are highly accurate in forecasting assets with highly systemic risks, which is a good result for investors.
This study thus aims to achieve better results by combining the results of artificial neural network models suitable for time series with those of the traditional financial models, and the following questions are proposed. (1) Are the MAEs calculated by non-linear artificial neural network models less than those calculated by linear models? (2) Among the artificial neural network models for time series, which one has the lowest MAE in ex-plaining daily returns? (3) Will MAEs be reduced if other variables are added to artificial neural networks? (4) Will MAEs be reduced if artificial neural network models are mixed and stacked? Ref. [17] pointed out that using both MAE and RMSE within the same model can make it difficult to draw strong conclusions and may lead to judgment distortions. Ref. [18] argues that the mean absolute error (MAE) is a more natural and unambiguous measure of average model performance error compared to the root mean square error (RMSE), making it a preferable choice for dimensional evaluations and inter-comparisons.
First, it compares deep learning models (ANN, LSTM, GRU, and CNN) with the Fama-French three-factor model for Taiwan ETF return prediction. Second, it analyzes how different variables, like market risk premium and size factor, affect prediction accuracy. Third, the study evaluates model stacking techniques, finding limited benefits for daily return prediction. Fourth, it confirms that LSTM, combined with specific variables, outperforms other models. Lastly, it offers practical guidance by recommending a single-layer LSTM with key variables for improved forecasting.

2. Research Method

2.1. Research Samples

This study utilized data from the Taiwan Economic Journal (TEJ) database and multi-factor market models, covering the period ranging from 2010 to 2020. While this time frame provides a substantial dataset for analysis, it may not capture more recent market trends or changes in ETF behavior after 2020. The research design incorporates a 5-year training period followed by a 1-year forecast horizon. Drawing from [19] and subsequent studies, we considered a comprehensive set of variables, including the risk-free interest rate, market risk premium, book-to-market ratio, size factors (from three- and five-factor models), investment factor, profitability factor, dividend yield, short-term and long-term reversal factors, and momentum factor. Our dependent variables comprised the daily returns of six well-established ETFs traded in Taiwan: Yuanta Taiwan 50 (0050), Yuanta Mid-Cap 100 (0051), Yuanta Electronics (0053), Yuanta S&P Custom China Play 50 (0054), Yuanta MSCI Taiwan Financials (0055), and Yuanta Taiwan Dividend Plus (0056).
The choice of ETFs as forecasting targets aligns with current trends in finance research, particularly in the application of artificial neural networks (ANNs) for predicting next-day price movements and informing trading strategies [18,19,20]. Recent literature has identified Long Short-Term Memory (LSTM) networks, a type of recurrent neural network, as one of the most effective models for this purpose [21,22,23].

2.2. Development of Artificial Intelligence

The evolution of artificial intelligence (AI) and its subfields, machine learning (ML) and deep learning (DL), has significantly impacted modern financial forecasting methodologies. These interrelated domains offer distinct approaches to data analysis and prediction. Originating in 1956 [24], artificial intelligence (AI) initially focused on leveraging computing power to mimic human behavior. Early applications in finance included the use of genetic algorithms (GAs) for stock market forecasting and time-series pattern identification [25].
Furthermore, emerging in the 1980s, machine learning (ML) enhanced AI with data-driven decision-making capabilities. This era saw the development of various classifiers, such as support vector machines (SVMs), decision trees, and random forests. Artificial neural networks (ANNs) became particularly prominent for continuous value forecasting in finance [26,27,28], utilizing interconnected neurons to modify targets and extract data features. As a subset of ML, deep learning (DL) methods, including deep neural networks (DNNs), convolutional neural networks (CNNs), and recurrent neural networks (RNNs), have gained traction for processing complex financial data. RNNs, especially LSTM and gated recurrent units (GRUs), have proven particularly effective for time-series analysis, such as stock price forecasting.
The application of AI models in finance has expanded to various domains, including corporate bankruptcy prediction [29,30], financial asset performance forecasting [31,32,33], and trading strategy development [34,35,36]. These applications underscore the potential of non-linear algorithms in capturing complex financial market dynamics.
In this study, we selected a diverse array of neural network architectures to comprehensively evaluate their efficacy in predicting Exchange-Traded Fund (ETF) returns. Artificial neural networks (ANN) served as our baseline model, providing a crucial benchmark for more sophisticated architectures. To address the sequential nature of financial data, we incorporated Long Short-Term Memory (LSTM) networks and gated recurrent units (GRUs), which are specifically engineered to process time-series data and excel in identifying long-term dependencies characteristic of financial markets [37]. Convolutional neural networks (CNNs), despite their primary application in image processing, were included to explore their potential in detecting local patterns in financial time series. To investigate the impact of model complexity, we employed stacked LSTM and stacked GRU models, which theoretically allow for the learning of more intricate temporal dependencies. Lastly, we explored hybrid models, such as CNN-LSTM and CNN-GRU, which synergistically combine the feature extraction capabilities of CNNs with the temporal modeling prowess of recurrent networks. This comprehensive selection of models aims to provide a nuanced understanding of the relative strengths and limitations of various neural network architectures in ETF return prediction, contributing to the broader field of applied machine learning in finance.

2.3. ANN

The artificial neural network calculates targets by connecting multiple neurons and consists of input layers, hidden layers, and output layers, as shown in Equation (1).
Y j = σ ( W i , j X i b j ) ,
where Y j is the information of the jth input neuron, W i , j is the connection weight between the ith input and the jth neuron, X i is the ith input variable, and b j is the threshold value.
The artificial neural network adjusts weights to reduce the difference between the output value and the target value, and the difference is expressed as an energy function, as shown in Equation (2).
E = 1 2 j ( T j Y j ) .
Here, T j is the output value of the jth input neuron, and Y j is the target of the jth input neuron.
In the process of neural network learning, the weights of neurons are adjusted to minimize the energy function, as shown in Equation (3).
W i , j = η × δ j × X i ,
where η is the learning rate, δ j is the difference between the output value and the target value, X i is the ith input variable, W i , j is the weight adjustment of the ith input information in the jth neuron, and the threshold adjustment equation is shown in Equation (4).
b j = ( η ) × δ j ,
where η is the learning rate, δ j is the difference between the output value and the target value, and b j is the threshold of the jth neuron.

2.4. LSTM

LSTM is mainly used for correlated data, such as speech recognition and text parsing, and so it is used for time-series data and often used in the financial market. In studies on the financial market by LSTM, stock prices rise and fall [38] (pp. 1419–1426); Ref. [39] (pp. 1507–1515) and price [40] (pp. 288–292); Ref. [41] (pp. 1754–1756); Ref. [42] (pp. 1546–1551) and rate of return [40] (pp. 25–37); Ref. [43] (pp. 1–12) are taken as the forecasting targets. There are a lot of studies on stocks and indices. Open-high/close-low, trading volumes, and technical indicators are used in most studies, and special variables, such as text and network volume, are used in some studies [44] (pp. 1–6) [45] (pp. 13099–13111).
The recurrent neural network is a kind of model for processing time-series data. However, exploding gradient or vanishing gradient often occurs due to the need to memorize information over an extended period. Therefore, it is difficult to train the model and requires more samples [46] (pp. 1–70); Ref. [47] (pp. 157–166). Ref. [48] (pp. 1735–1780) proposed an LSTM, which classifies the original data into long-term memory and short-term memory. Long-term memory is slow in data updating, thus retaining early information. Long-term memory and short-term memory are separated to avoid long-term memory being covered by short-term memory and to reduce the difficulty of weight adjustment between neurons. LSTM is comprised of an input gate, forget gate, and output gate.
Long Short-Term Memory (LSTM) networks have emerged as a preeminent approach in financial time-series forecasting, demonstrating remarkable efficacy in capturing long-term dependencies inherent in market data. The distinctive architecture of LSTM, characterized by memory cells and specialized gates (input, forget, and output), provides a sophisticated mechanism for regulating information flow, enabling the retention of relevant information over extended temporal sequences. In the context of this study, LSTM’s superior performance is evidenced by its adept processing of complex temporal patterns in daily ETF returns, effectively integrating multifaceted variables such as market risk premium, size factor, and book-to-market ratio. This capacity for nuanced temporal analysis allows LSTM to outperform traditional forecasting methods and other neural network architectures, particularly in scenarios characterized by high volatility and complex market dynamics. Consequently, LSTM presents a compelling solution for deciphering the intricate and dynamic nature of financial returns, positioning it as an exceptionally powerful tool for ETF return prediction in an increasingly complex market environment.
In period t, x t and h t 1 are input into LSTM, and the two variables x t and h t 1 enter the forget gate. The information weight to be forgotten in the long-term memory of the current period is adjusted by the weighting matrix in the forget gate, as shown in Equation (5).
f t = σ x t U f + h t 1 W f + b f ,
where x t is the input variable of the current period, h t 1 is the short-term memory of the previous period, U f is the weight of the input variable in the forget layer, W f is the weight of the short-term memory of the previous period in the forget layer, and b f is the bias vector of the forget layer.
After x t   and   h t 1 are input, values are adjusted to be between 0 and 1 by tanh(x), and then adjusted to be the information needed for the current period by the weighting matrix, as shown in Equation (6).
c t ~ = tanh ( x t U c + h t 1 W c + b c ) .
where c ~ t is the information to be added to the long-term memory of the current period, U c is the weight of the input variable in the tanh layer, W c is the weight of the short-term memory of the previous period in the tanh layer, and b c is the bias vector of the tanh layer.
The input gate determines how much information in x t   a n d   h t 1 is useful, and so it is multiplied by the weight, as shown in Equation (7).
i t = σ ( x t U i + h t 1 W i + b i ) ,
where i t is the result of the input gate, U i is the weight of the input variable in the input gate, W i is the weight of the input gate, and b i is the bias vector of the input gate.
We also note that C t is the sum of the forget gate multiplied by C t 1 , and the input gate multiplied by C ~ t , and C t is retained in the long-term memory to be used for the forget gate and output for the next time, as shown in Equation (8).
C t = f t C t 1 i t c t ~ .
where is the multiplication of two corresponding elements, and in the addition of two corresponding elements.
Equation (9) shows the input gate, while x t and h t 1 are input into the input gate, and the input gate determines the weight of the output information of the current period.
O t = σ ( x t U o + h t 1 W o + b o ) ,
where O t is the input gate, U o is the weight of the output variable in the output gate, W o is the weight of the short-term memory of the previous period in the output gate, and b o is the bias vector of the output gate.
The weight determined by the output gate is multiplied by C t to obtain the output of the current period, as shown in Equation (10).
h t = O t tanh ( C t ) ,
where h t is the result of the current period, O t is the output gate, and C t is the information in the long-term memory of the current period.

2.5. GRU

A GRU, a type of recurrent neural network (RNN) introduced by [49], is recognized for its advantages in convergence rate and generalization over standard RNNs. While the GRU does not necessarily improve computing power compared to LSTM, it offers faster convergence and often performs better in specific contexts, particularly in time-series forecasting. Although a GRU is less studied than LSTM due to its more recent development in 2014, it has shown promising results in various financial applications.
Studies comparing GRUs with other models have yielded mixed results. For instance, Ref. [45] found that non-linear algorithms, like RNN-GRU, outperform linear models, such as linear regression (LR) and support vector machines (SVMs). Ref. [50] showed that LSTM remains the most effective model among RNNs, including GRUs. However, the choice between LSTM and GRUs may depend on the variables used. While LSTM often outperforms GRUs in models utilizing technical indicators, trading volumes, or open-high/close-low data, GRUs can excel when incorporating macroeconomic, sentiment, or news analysis [51]. Furthermore, Ref. [48] found that a double-layer GRU and a single-layer LSTM can produce similar results, suggesting their performance might be context-dependent. Additionally, Ref. [52] compared eight different models and found the GRU ranked second, surpassing LSTM in specific scenarios.
Structurally, a GRU has two gates—reset and update—that streamline the network’s operations, enhancing its computational efficiency. Unlike LSTM, which includes three gates (input, forget, and output), GRUs simplify the process by combining some of these functions, resulting in reduced computational time and memory use.
After x t and h t 1 are input and then adjusted by the reset gate, the information can be used for the current period as shown in Equation (11).
r t = σ ( x t U r + h t 1 W r + b r ) ,
where r t is the output of the reset gate, U r is the weight of the input variable x t in the reset gate, W r is the weight of the input variable h t 1 in the reset gate, b r is the bias vector of the reset gate, x t is the variable of the current period, and h t 1 is the information retained in the previous period.
After x t and h t 1 are input and then adjusted by the update gate, the short-term memory of the previous period determined to be retained is shown in Equation (12).
z t = σ ( x t U z + h t 1 W z + b z ) ,
where z t is the output of the update gate, U z is the weight of the input variable x t in the update gate, W z is the weight of the input variable h t 1 in the update gate, and b z is the bias vector of the update gate.
The available information of the previous period is added to the information of the current period to obtain the information to be updated in the current period, as shown in Equation (13).
h ~ t = t a n h x t U h + W h ( h t 1 r t + b h ) ,
where h ~ t is the information of the current period to be added to the short-term memory, r t is the short-term memory of the previous period to be retained, mainly determined by the update gate, W h is the weight of the information added to short-term memory, and b h is the bias vector of the information added to short-term memory.
The information of the current period is added to the short-term memory of the previous period to obtain the short-term memory of the current period, as shown in Equation (14).
h t = 1 z t h t 1 + z t h ~ t .
where h t is the short-term memory, 1 z t is the memory to be retained, and z t h ~ t is information to be added to the short-term memory.

2.6. CNN

CNNs, introduced by [53], are known for their powerful feature extraction and convergence abilities, making them well-suited for handling large datasets. While originally designed for image recognition and video analysis, CNNs have found various applications in the financial market due to their ability to process complex inputs efficiently.
In finance, CNNs have been utilized for different forecasting tasks. Ref. [54] employed CNNs to predict corporate bankruptcies using financial indicators, while [55] used them to classify customer evaluations as good or bad. CNNs are also frequently used for time-series analysis, particularly for forecasting stock price movements [56,57]. In recent years, studies have utilized CNNs to analyze candlestick charts and open-high/close-low (OHCL) data, often comparing their performance with LSTM models [57,58,59,60,61]. While CNNs are primarily used for classifying stock price movements and predicting performance or bankruptcy, their application in time-series classification remains relatively underexplored.
A typical CNN architecture includes input, convolutional, pooling, flatten, and output layers. The convolutional layer plays a crucial role in extracting features by multiplying input data with convolutional kernels, which are adjusted during training to capture essential patterns.
C i , j k = t a n h m = 1 M n = 1 N W m , n X i + m 1 , j + n 1 k + b k
where C i , j k is the neuron in row i and column j of the kth output feature map, W m , n is the weight of the convolutional kernel in row m and column n, M is the number of rows of the convolutional kernel, N is the number of columns of the convolutional kernel, X i + m 1 , j + n 1 k is row i + m − 1 and column j + n − 1 of the kth input feature map, and b k is the bias vector of the kth map.
The pooling layer is mainly to reduce the dimension of data, so as to reduce the amount of data and keep important information. There are the maximum pooling method and average pooling method. In simple terms, the maximum or average of the data within the range is taken.

2.7. Definitions of Input Variables

In addition to the Fama-French five-factor model, in 1996, Fama and French [19] used size factor, book-to-market ratio, dividend yield, short-term reversal factor, and long-term reversal factor for the values that the capital asset pricing model cannot solve. In this study, these variables are added to artificial neural networks to verify whether the test results can be improved.
(1)
Dividend yield premium
In finance, there is a long history of forecasting stock market returns or stock premiums by the total dividend yield ratio [19]. According to [62] (pp. 3–37), dividend yield premiums are mainly self-forecasting and only effective within the time range of more than 5 to 10 years. However, it is unknown whether they work in non-linear models.
(2)
Momentum factor
Ref. [63] (pp. 57–82) added momentum factors to the three-factor model to create the four-factor model and to capture the momentum effect in the coming year by momentum factors.
(3)
Short-term reversal factor
Portfolios of high-return stocks in the past have positive returns in the coming year except in the first month, the positive returns in the coming year are reduced by half, and there is no positive return in the second year in the future.
(4)
Long-term reversal factor
Ref. [64] (pp. 793–805) found that stocks with low long-term returns in the past are more likely to have high returns in the future. This result is consistent with the overreaction theory. In 36 months after the formation of portfolios, portfolios of losing stocks will outperform portfolios of winning stocks by 25%, and the portfolios of losing stocks will have substantial positive returns in the first month.
(5)
Profitability factor
Based on the value strategy proposed by [65], by holding companies whose market values are less than book values and shorting companies whose market values are more than book values, investors can obtain large book values for less money, indicating that more assets lead to more returns. Fama and French (2006) [66] found that book value and profitability positively relate to expected returns.
(6)
Investment factor
Stock prices often react well to announcements of major capital investments, but financing negatively relates to stock prices. According to Titman, abnormal capital investments negatively relate to future stock returns, and companies that increase their capital investments often have low stock returns over the next five years, and so the investment factor was added to the Fama-French five-factor model.

2.8. Model Parameter Setting

An artificial neural network model consists of input layers, hidden layers, and output layers. The hidden layers and the results are very important. The number of hidden layers and the number of neurons in the hidden layers are mainly set.
(1)
Number of hidden layers
There are LSTM and GRUs with single, double, or three hidden layers [67] (pp. 1837–1844), by comparing the number of hidden layers, found that LSTM with a single hidden layer and GRUs with double hidden layers are the most effective, and the results of LSTM, GRU, and GRU+LSTM with three hidden layers are ineffective. Hence, single and double hidden layers are used in this study.
(2)
Number of neurons in the hidden layers
Optimal parameters of the hidden layers for the artificial neural network models used in the stock market, including the number of input samples, the number of hidden layers, and the number of neurons in hidden layers. This study is set according to the equation obtained by them, as shown in Equation (16).
N = N i n + N p L
where N is the number of neurons in the optimal hidden layer, N i n is the type of input variables, N p is the number of input samples, and L is the number of hidden layers.

3. Empirical Results

In this section, the basic descriptive statistics of variables are first described. Second, variables of the CAPM, Fama-French three-factor model, and Fama-French five-factor model are added into the regression and ANN to verify whether artificial neural networks can reduce forecast errors. After that, based on the variables of the Fama-French three-factor model, eight models, such as ANN, LSTM, GRU and CNN, are combined with factors, such as the profitability factor, investment factor, short-term reversal factor, long-term reversal factor, and momentum factor, to verify whether artificial neural networks can be more effective.

Basic Descriptive Statistics

Table 1 shows the basic descriptive statistics of the returns on the full sample ranging from 2010 to 2019, including sample size, average value, standard deviation, minimum value, and maximum value. According to the standard deviation in descending order, the samples are Yuanta Mid-Cap 100 (0051), Yuanta S&P Custom China Play 50 (0054), Yuanta MSCI Taiwan Financials (0055), Yuanta Electronics (0053), Yuanta Taiwan 50 (0050), and Yuanta Taiwan Dividend Plus (0056). According to the average in descending order, the samples are Yuanta Taiwan 50 (0050), Yuanta Electronics (0053), Yuanta MSCI Taiwan Financials (0055), Yuanta Taiwan Dividend Plus (0056), Yuanta Mid-Cap 100 (0051), and Yuanta S&P Custom China Play 50 (0054).
Table 2 compares the linear traditional model with the non-linear artificial neural network model. The table shows that the non-linear artificial neural network model using variables of the CAPM, Fama-French three-factor model, and Fama-French five-factor model perform significantly better than the regression model, and the combination of the ANN and the Fama-French three-factor model variables is the most effective, followed by the combination of the ANN and the Fama-French five-factor model variables.
Table 3 compares the ANN and the artificial neural networks commonly used. In this study, LSTM and GRUs are combined with variables of the CAPM, Fama-French three-factor model, and Fama-French five-factor model. The results show that, regardless of the model used, combining it with the variables of the Fama-French three-factor model consistently outperforms combinations with the variables of the CAPM and Fama-French five-factor model. This indicates that using the Fama-French three-factor model variables provides a more effective basis for forecasting daily returns. When the variables of the three-factor model are combined with LSTM or GRUs, the average error is less than that of ANNs, further suggesting that time-series models can reduce prediction errors.
Based on the conclusion from Table 3 that the Fama-French three-factor model consistently achieves the minimum error, this study combines these variables with LSTM and adds other variables to explore potential improvements. According to Table 4, the results from 2015 to 2019 show that using the Fama-French three-factor model variables, with or without the addition of the short-term reversal factor, performs well. However, on average, the combination of the Fama-French three-factor model and LSTM yields the lowest error.
Based on the variables of the Fama-French three-factor model, this study mixes and stacks artificial neural networks recently used for forecasting and adds the CNN to them. This study adds a total of five models—namely, the stacked LSTM, stacked GRU, CNN, CNN-LSTM, and CNN-GRU.
According to Table 5, except for 2016 and 2019, LSTM has the minimum error, and the error increases after the CNN is added. Therefore, it is concluded in this study that the models with mixed or stacked networks are not helpful in explaining the daily price.
All models and variables are utilized in this study for forecasting, as shown in Table 6. The results indicate that the variables of the Fama-French three-factor model perform well across different models, including the ANN, LSTM, stacked LSTM, GRU, and stacked GRU. When combined with the short-term reversal factor, the performance is similar to using the Fama-French three-factor model alone. Therefore, the Fama-French three-factor model’s variables are identified as the most effective in explaining daily returns. Although the combination of stacked GRU with these variables performs well, its error rate is not lower than that of LSTM, and the model is more complex. Thus, this study concludes that the combination of LSTM with the Fama-French three-factor model is the most effective approach.

4. Conclusions

Artificial intelligence has advanced significantly, and models constructed using historical data have made financial markets with large datasets prime targets for forecasting research. This study applies artificial neural networks (ANNs) to the CAPM, Fama-French three-factor model, and Fama-French five-factor model, using historical data to verify if improved forecasting results can be achieved. In this paper, data ranging from 1 January 2010 to 31 December 2019 were utilized to forecast daily returns using various ANN models, including LSTM, GRUs, and CNNs. Key variables, such as market risk premium, book-to-market ratio, size factor, short-term and long-term reversal factors, profitability, investment, and momentum factors, were used to compare forecasting outcomes.
The performance of non-linear ANN regression was compared to linear ANNs, indicating that non-linear ANNs offer better predictive accuracy, as supported by studies on Taiwan’s ETFs and corroborated by other research [10] (pp. 41–73); Ref. [11] (pp. 53–60). Furthermore, the ANN was compared with other artificial neural network models, such as LSTM, GRUs, and CNNs, revealing that LSTM provided the best results. Therefore, this study suggests incorporating LSTM into future research models to improve forecasting accuracy.
While this study applies established deep learning models to Taiwan ETFs, its academic contribution lies in validating and comparing the effectiveness of different neural network models (ANN, LSTM, GRU, and CNN) within a specific market context. Notably, the results demonstrate that the Fama-French three-factor model, when paired with LSTM, provides superior forecasting accuracy. This study also investigates the limitations of mixing and stacking models for daily return predictions, finding that such techniques may not enhance accuracy, which contrasts with prior research. This nuanced insight contributes to the literature by offering a targeted application of deep learning models in Taiwan’s financial market and highlights the specific conditions under which these models perform best. Moreover, the findings suggest that investors and researchers could focus on using single-layer LSTM with specific variables to streamline model selection and improve forecasting accuracy in similar markets.
Adding additional variables to the neural network models showed that the Fama-French three-factor model produced the most substantial results with market risk premium, book-to-market ratio, and size factor, and no further improvements were observed by including other factors. Additionally, when models were mixed and stacked (e.g., stacked LSTM, stacked GRU, CNN-LSTM, and CNN-GRU), the results did not reduce errors, contrary to findings in previous studies [39] (pp. 1507–1515); Ref. [67] (pp. 1837–1844). This indicates that model mixing and stacking may not be suitable for studies focused on daily returns.
The results suggest that artificial neural networks can effectively reduce model errors, with single-layer LSTM combined with market risk premium, book-to-market ratio, and size factor proving to be the most effective for forecasting. This approach could help streamline the number of models required for financial market research and introduce new variables for future forecasting efforts. Investors can implement these models using Python 3.10 or commercial software for more accurate results.

Research Limitations and Future Directions

This study acknowledges several limitations that offer avenues for future research. One notable limitation is the potential risk of overfitting when using complex neural network models, particularly given the specific dataset used (Taiwanese ETFs from 2010 to 2020). This dataset choice may limit the generalizability of the findings to other markets or more recent conditions. Additionally, the study primarily relied on the Mean Absolute Error (MAE) as the error metric, which, while beneficial for its straightforward interpretation, might not capture all dimensions of prediction performance. Future research could incorporate additional metrics, such as mean squared error (MSE) or root mean squared error (RMSE), to provide a more comprehensive evaluation.
Furthermore, a correlation analysis between ETF assets would be valuable for understanding portfolio formation and enhancing the practical applications of the forecasting models. Including statistical tests to assess the significance of performance differences between models could also offer more robust insights, providing clearer validation of the results. These directions will help refine the models and strengthen their applicability in diverse market environments.

Author Contributions

Conceptualization, K.-H.S.; Methodology, Y.-H.W.; Software, I.-C.K.; Formal analysis, F.-M.L.; Investigation, I.-C.K. and F.-M.L.; Resources, I.-C.K.; Data curation, I.-C.K.; Supervision, Y.-H.W.; Project administration, K.-H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sharpe, W.F. Capital asset prices: A theory of market equilibrium under conditions of risk. J. Financ. 1964, 19, 425–442. [Google Scholar]
  2. Markowitz, H. Portfolio selection. J. Financ. 1952, 7, 77–91. [Google Scholar]
  3. Ross, S.A. Options and efficiency. Q. J. Econ. 1976, 90, 75–89. [Google Scholar] [CrossRef]
  4. Banz, R.W. The relationship between return and market value of common stocks. J. Financ. Econ. 1981, 9, 3–18. [Google Scholar] [CrossRef]
  5. Ball, R. Anomalies in relationships between securities’ yields and yield-surrogates. J. Financ. Econ. 1978, 6, 103–126. [Google Scholar] [CrossRef]
  6. Fama, E.F.; French, K.R. The cross-section of expected stock returns. J. Financ. 1992, 47, 427–465. [Google Scholar]
  7. Fama, E.F.; French, K.R. Dissecting anomalies. J. Financ. 2008, 63, 1653–1678. [Google Scholar] [CrossRef]
  8. Fama, E.F.; French, K.R. A five-factor asset pricing model. J. Financ. Econ. 2015, 116, 1–22. [Google Scholar] [CrossRef]
  9. Foye, J. A comprehensive test of the Fama-French five-factor model in emerging markets. Emerg. Mark. Rev. 2018, 37, 199–222. [Google Scholar] [CrossRef]
  10. Londoño, C.; Lopera, M.; Restrepo, S. Teoría de precios de arbitraje. Evidencia empírica para colombia a través de redes neuronales. Rev. Econ. Rosario 2010, 13, 41–73. [Google Scholar]
  11. Jafari, S.M.; Misaghi, F.J.; Ahmadvand, M. A comparison between CAPM, Fama and French models and artificial neural networks in predicting the iranian stock market. J. Econ. Bus. Res. 2013, 4, 53–60. [Google Scholar]
  12. Londoño, C.; Cuan, Y. Modelos de precios de los activos: Un ejercicio comparativo basado en redes neuronales aplicado al mercado de valores colombiano. Lect. Econ. 2011, 75, 59–87. [Google Scholar] [CrossRef]
  13. Diallo, B.; Bagudu, A.; Zhang, Q. A Machine Learning Approach to the Fama-French Three- and Five-Factor Models 2019. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3440840 (accessed on 1 September 2024).
  14. Cao, Q.; Leggio, K.B.; Schniederjans, M.J. A comparison between Fama and French’s model and artificial neural networks in predicting the Chinese stock market. Comput. Oper. Res. 2005, 32, 2499–2512. [Google Scholar] [CrossRef]
  15. Ayub, U.; Jan, M.N.; Afridi, M.A.; Jadoon, I.A. Capital asset pricing model and artificial neural networks: A case of Pakistan’s equity market. Fed. Hist. 2020, 40, 673–688. [Google Scholar]
  16. Jan, M.N.; Ayub, U. Do the FAMA and French Five-Factor model forecast well using ANN? J. Bus. Econ. Manag. 2019, 20, 168–191. [Google Scholar]
  17. Ap Gwilym, O. Forecasting volatility for options pricing for the UK stock market. J. Financ. Manag. Anal. 2001, 14, 55. [Google Scholar]
  18. Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  19. Fama, E.F.; French, K.R. Multifactor explanations of asset pricing anomalies. J. Financ. 1996, 51, 55–84. [Google Scholar] [CrossRef]
  20. Versace, M.; Bhatt, R.; Hinds, O.; Shiffer, M. Predicting the exchange traded fund DIA with a combination of genetic algorithms and neural networks. Expert Syst. Appl. 2004, 27, 417–425. [Google Scholar] [CrossRef]
  21. Chalvatzis, C.; Hristu-Varsakelis, D. High-performance stock index trading via neural networks and trees. Appl. Soft Comput. 2020, 96, 106567. [Google Scholar] [CrossRef]
  22. Xie, Z.; Lin, X.; Zhong, Y.; Chen, Q. Research on Gold ETF Forecasting Based on LSTM. In Proceedings of the 2019 IEEE Intl Conf on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking, Xiamen, China, 16–18 December 2019; pp. 1346–1351. [Google Scholar]
  23. Day, M.Y.; Lin, J.T. Artificial intelligence for ETF market prediction and portfolio optimization. In Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), Vancouver, BC, Canada, 27–30 August 2019; pp. 1026–1033. [Google Scholar]
  24. McCarthy, J.; Minsky, M.L.; Rochester, N.; Shannon, C.E. A proposal for the dartmouth summer research project on artificial intelligence. AI Mag. 2006, 27, 12–14. [Google Scholar]
  25. LeBaron, B.; Arthur, W.B.; Palmer, R. Time series properties of an artificial stock market. J. Econ. Dyn. Control 1999, 23, 1487–1516. [Google Scholar] [CrossRef]
  26. Holland, J.H.; Reitman, J.S. Cognitive systems based on adaptive algorithms. ACM SIGART Bull. 1977, 63, 48–49. [Google Scholar] [CrossRef]
  27. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386. [Google Scholar] [CrossRef] [PubMed]
  28. Chatzis, S.P.; Siakoulis, V.; Petropoulos, A.; Stavroulakis, E.; Vlachogiannakis, N. Forecasting stock market crisis events using deep and statistical machine learning techniques. Expert Syst. Appl. 2018, 112, 353–371. [Google Scholar] [CrossRef]
  29. Becerra-Vicario, R.; Alaminos, D.; Aranda, E.; Fernández-Gámez, M.A. Deep recurrent convolutional neural network for bankruptcy prediction: A case of the restaurant industry. Sustainability 2020, 12, 5180. [Google Scholar] [CrossRef]
  30. Du Jardin, P. Forecasting bankruptcy using biclustering and neural network-based ensembles. Ann. Oper. Res. 2019, 299, 531–566. [Google Scholar] [CrossRef]
  31. Ballestra, L.V.; Guizzardi, A.; Palladini, F. Forecasting and trading on the VIX futures market: A neural network approach based on open to close returns and coincident indicators. Int. J. Forecast. 2019, 35, 1250–1262. [Google Scholar] [CrossRef]
  32. Edet, S. Recurrent Neural Networks in Forecasting S&P 500 Index. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3001046 (accessed on 12 July 2017).
  33. Lee, J.; Kang, J. Effectively training neural networks for stock index prediction: Predicting the S&P 500 index without using its index data. PLoS ONE 2020, 15, e0230635. [Google Scholar]
  34. Arévalo, A.; Niño, J.; Hernández, G.; Sandoval, J. High-frequency trading strategy based on deep neural networks. In International Conference on Intelligent Computing; Springer: Cham, Switzerland, 2016; pp. 424–436. [Google Scholar]
  35. Sang, C.; Di Pierro, M. Improving trading technical analysis with tensorflow long short-term memory (lstm) neural network. J. Financ. Data Sci. 2019, 5, 1–11. [Google Scholar] [CrossRef]
  36. Fister, D.; Mun, J.C.; Jagrič, V.; Jagrič, T. Deep learning for stock market trading: A superior trading strategy. Neural Netw. World 2019, 29, 151–171. [Google Scholar] [CrossRef]
  37. Widiputra, H.; Mailangkay, A.; Gautama, E. Multivariate CNN-LSTM Model for Multiple Parallel Financial Time-Series Prediction. Complexity 2021, 2021, 9903518. [Google Scholar] [CrossRef]
  38. Nelson, D.M.; Pereira, A.C.; de Oliveira, R.A. Stock market’s price movement prediction with LSTM neural networks. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1419–1426. [Google Scholar]
  39. Chen, S.; Ge, L. Exploring the attention mechanism in LSTM-based Hong Kong stock price movement prediction. Quant. Financ. 2019, 19, 1507–1515. [Google Scholar] [CrossRef]
  40. Faustryjak, D.; Jackowska-Strumiłło, L.; Majchrowicz, M. Forward forecast of stock prices using LSTM neural networks with statistical analysis of published messages. In Proceedings of the 2018 International Interdisciplinary PhD Workshop (IIPhDW), Świnouście, Poland, 9–12 May 2018; pp. 288–292. [Google Scholar]
  41. Roondiwala, M.; Patel, H.; Varma, S. Predicting stock prices using LSTM. Int. J. Sci. Res. 2017, 6, 1754–1756. [Google Scholar]
  42. Shao, X.; Ma, D.; Liu, Y.; Yin, Q. Short-term forecast of stock price of multi-branch LSTM based on K-means. In Proceedings of the 2017 4th International Conference on Systems and Informatics (ICSAI), Hangzhou, China, 11–13 November 2017; pp. 1546–1551. [Google Scholar]
  43. Kim, H.Y.; Won, C.H. Forecasting the volatility of stock price index: A hybrid model integrating LSTM with multiple GARCH-type models. Expert Syst. Appl. 2018, 103, 25–37. [Google Scholar] [CrossRef]
  44. Damrongsakmethee, T.; Neagoe, V. Stock Market Prediction Using a Deep Learning Approach. In Proceedings of the 2020 12th International Conference on Electronics, Computers and Artificial Intelligence (ECAI), Bucharest, Romania, 25–27 June 2020; pp. 1–6. [Google Scholar]
  45. Althelaya, K.; Mohammed, S.; El-Alfy, E. Combining Deep Learning and Multiresolution Analysis for Stock Market Forecasting. IEEE Access 2021, 9, 13099–13111. [Google Scholar] [CrossRef]
  46. Hochreiter, S. Untersuchungen zu dynamischen neuronalen netzen. Diploma Tech. Univ. München 1991, 91, 1–70. [Google Scholar]
  47. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef]
  48. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  49. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
  50. Samarawickrama, A.J.P.; Fernando, T.G.I. A recurrent neural network approach in predicting daily stock prices an application to the Sri Lankan stock market. In Proceedings of the 2017 IEEE International Conference on Industrial and Information Systems (ICIIS), Peradeniya, Sri Lanka, 15–16 December 2017; pp. 1–6. [Google Scholar]
  51. Shahi, T.B.; Shrestha, A.; Neupane, A.; Guo, W. Stock price forecasting with deep learning: A comparative study. Mathematics 2020, 8, 1441. [Google Scholar] [CrossRef]
  52. Jiang, M.; Liu, J.; Zhang, L.; Liu, C. An improved Stacking framework for stock index prediction by leveraging tree-based ensemble models and deep learning algorithms. Phys. A Stat. Mech. Its Appl. 2020, 541, 1–16. [Google Scholar] [CrossRef]
  53. LeCun, Y.; Bengio, Y. Convolutional networks for images, speech, and time series. Handb. Brain Theory Neural Netw. 1995, 3361, 1441. [Google Scholar]
  54. Hosaka, T. Bankruptcy prediction using imaged financial ratios and convolutional neural networks. Expert Syst. Appl. 2019, 117, 287–299. [Google Scholar] [CrossRef]
  55. Neagoe, V.E.; Ciotec, A.D.; Cucu, G.S. Deep convolutional neural networks versus multilayer perceptron for financial prediction. In Proceedings of the 2018 International Conference on Communications (COMM), Bucharest, Romania, 14–16 June 2018; pp. 201–206. [Google Scholar]
  56. Sulistio, B.; Warnars, H.L.H.S.; Gaol, F.L.; Soewito, B. Energy sector stock price prediction using the CNN, GRU & LSTM hybrid algorithm. In Proceedings of the 2023 International Conference on Computer Science, Information Technology and Engineering, Jakarta, Indonesia, 16 February 2023; pp. 178–182. [Google Scholar]
  57. Gunduz, H.; Yaslan, Y.; Cataltepe, Z. Intraday prediction of Borsa Istanbul using convolutional neural networks and feature correlations. Knowl.-Based Syst. 2017, 137, 138–148. [Google Scholar] [CrossRef]
  58. Gudelek, M.U.; Boluk, S.A.; Ozbayoglu, A.M. A deep learning based stock trading model with 2-D CNN trend detection. In Proceedings of the 2017 IEEE Symposium Series on Computational Intelligence (SSCI), Honolulu, HI, USA, 27 November–1 December 2017; pp. 1–8. [Google Scholar]
  59. Hung, C.C.; Chen, Y.J.; Guo, S.J.; Hsu, F.C. Predicting the price movement from candlestick charts: A CNN-based approach. Int. J. Ad Hoc Ubiquitous Comput. 2020, 34, 111–120. [Google Scholar] [CrossRef]
  60. Guo, S.J.; Hsu, F.C.; Hung, C.C. Deep candlestick predictor: A framework toward forecasting the price movement from candlestick charts. In Proceedings of the 2018 9th International Symposium on Parallel Architectures, Algorithms and Programming (PAAP), Taipei, Taiwan, 26–28 December 2018; pp. 219–226. [Google Scholar]
  61. Song, H.; Choi, H. Forecasting stock market indices using the recurrent neural network based hybrid models: CNN-LSTM, GRU-CNN, and ensemble models. Appl. Sci. 2023, 13, 4644. [Google Scholar] [CrossRef]
  62. Kao, I.C. The Forecast Performance of ETF: Comparison of Deep Learning and Fama-French Three Factor Model. Master’s Thesis, Chinese Culture University, Taiwan, China, 2021. [Google Scholar]
  63. Jearanaitanakij, K.; Passaya, B. Predicting short trend of stocks by using convolutional neural network and candlestick patterns. In Proceedings of the 2019 4th International Conference on Information Technology (InCIT), Bangkok, Thailand, 24-25 October 2019; pp. 159–162. [Google Scholar]
  64. Dow, C.H. Scientific Stock Speculation; Magazine of Wall Street, Windsor Books: New York, NY, USA, 1920. [Google Scholar]
  65. Cochrane, J.H. Where is the market going? Uncertain facts and novel theories. Econ. Perspect. 1997, 21, 3–37. [Google Scholar]
  66. Fama, E.F.; French, K.R. The value premium and the CAPM. J. Financ. 2006, 61, 2163–2185. [Google Scholar] [CrossRef]
  67. Hossain, M.A.; Karim, R.; THulasiram, R.; Bruce, N.D.; Wang, Y. Hybrid deep learning model for stock price prediction. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI), Bangalore, India, 18–21 November 2018; pp. 1837–1844. [Google Scholar]
Table 1. Basic descriptive statistics of research samples.
Table 1. Basic descriptive statistics of research samples.
Code005000510053005400550056
Sample Size246524652465246524652465
Mean0.0003890.0002690.0003690.0002380.0003380.000312
Standard Deviation0.0095180.0114560.0104780.0112350.0108090.007821
Minimum−0.07027−0.085498−0.069711−0.07629−0.062885−0.065148
Maximum0.048510.0999970.0444080.069310.0695990.054395
Table 2. Comparison of linear and non-linear methods.
Table 2. Comparison of linear and non-linear methods.
VariableCAPM VariablesThree-Factor VariablesFive-Factor Variables
MethodLinear (Regression)Non-Linear (ANN)Linear (Regression)Non-Linear (ANN)Linear (Regression)Non-Linear (ANN)
20150.0137940.005101 0.0137640.004862 0.0137510.004921
20160.0113800.004753 0.0113320.004620 0.0112770.004670
20170.0104150.004199 0.0104520.004113 0.0104430.004120
20180.0107440.004771 0.0107310.004577 0.0107290.004613
20190.0102870.003863 0.0102660.003715 0.0102830.003775
Balance0.0113240.004537 0.0113090.004378 0.0112970.004420
Note: represents the lowest MAE of that year.
Table 3. Comparison of ANN, LSTM, and GRU.
Table 3. Comparison of ANN, LSTM, and GRU.
VariableCAPM VariablesThree-Factor VariablesFive-Factor Variables
ModelANNLSTMGRUANNLSTMGRUANNLSTMGRU
20150.0051010.0050240.0050610.0048620.004794 0.0049260.0049210.0049020.005260
20160.0047530.0047340.0046790.0046200.0045910.004563 0.0046700.0046470.004654
20170.0041990.0041280.0041340.0041130.004083 0.0040970.0041200.0041340.004151
20180.0047710.0047430.0048070.004577 0.0046030.0046880.0046130.0046700.004832
20190.0038630.0038400.0037320.0037150.0036600.003565 0.0037750.0037280.003651
Balance0.0045370.0044940.0044820.0043780.004346 0.0043680.0044200.0044160.004510
Note: represents the lowest MAE of that year.
Table 4. Comparison of variables.
Table 4. Comparison of variables.
Variables20152016201720182019Balance
3 Factor Variables0.004794 0.0045910.0040830.004603 0.0036600.004346
3 Factor Variables + Momentum Factor0.0048080.0046050.004047 0.0046430.0036860.004358
3 Factor Variables + Investment Factor0.0048580.0045980.0040570.0046360.0037310.004376
3 Factor Variables + Profitability Factor0.0048730.0046490.0041060.0046270.0036820.004388
3 Factor Variables + Dividend Yield Factor0.0048400.0046110.0040910.0047910.0037120.004409
3 Factor Variables + Long-Term Reversal0.0048720.0046270.0040720.0046110.0036630.004369
3 Factor Variables + Short-Term Reversal0.0048250.004582 0.0040600.0046350.003633 0.004347
Note: represents the lowest MAE of that year.
Table 5. Comparison of other neural network models.
Table 5. Comparison of other neural network models.
LSTMStacked LSTMGRUStacked GRUCNNCNN-LSTMCNN-GRU
20150.004794 0.0048540.0049260.0048370.0077200.0077270.007751
20160.0045910.004560 0.0045630.0045660.0075710.0075550.007777
20170.004083 0.0041220.0040970.0041040.0058350.0054310.005582
20180.004603 0.0046980.0046880.0046420.0074970.0074380.007515
20190.0036600.0035860.003565 0.0036200.0059470.0060520.008148
Balance0.004346 0.0043640.0043680.0043530.0069140.0068410.007355
Note: represents the lowest MAE of that year.
Table 6. Results of various model and variable combinations.
Table 6. Results of various model and variable combinations.
ANNLSTMStacked LSTMGRUStacked GRUCNNCNN-LSTMCNN-GRU
CAPM Variables0.0045370.0044940.0044820.0044820.0044800.0070440.0068830.006859 1
3 Factor Variables0.004378 30.004346 10.004364 10.004368 10.004353 10.0069140.006841 10.007355
5 Factor Variables0.0044200.0044160.0044630.0045100.0044410.006870 20.0070200.007203
3 Factor Variables + Momentum Factor0.004372 20.00435830.0044180.0044640.0043820.0071070.006869 20.006914 3
3 Factor Variables + Investment Factor0.0043800.0043760.0044170.0044480.0043900.0069360.006886 30.007162
3 Factor Variables + Profitability Factor0.0043930.0043870.0044100.0044580.0043830.006907 30.0068940.006943
3 Factor Variables + Dividend Factor0.0044320.0044090.0044640.0044680.0044200.0070250.006869 20.006878 2
3 Factor Variables + Long-Term Reversal0.0043940.0043690.004367 20.004395 20.004365 20.006859 10.0069100.007056
3 Factor Variables + Short-Term Reversal0.004365 10.004347 20.004386 30.004410 30.004373 30.0069200.0070160.006968
Note: 1 represents the lowest MAE. 2 represents the second lowest MAE. 3 represents the third lowest MAE.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shih, K.-H.; Wang, Y.-H.; Kao, I.-C.; Lai, F.-M. Forecasting ETF Performance: A Comparative Study of Deep Learning Models and the Fama-French Three-Factor Model. Mathematics 2024, 12, 3158. https://doi.org/10.3390/math12193158

AMA Style

Shih K-H, Wang Y-H, Kao I-C, Lai F-M. Forecasting ETF Performance: A Comparative Study of Deep Learning Models and the Fama-French Three-Factor Model. Mathematics. 2024; 12(19):3158. https://doi.org/10.3390/math12193158

Chicago/Turabian Style

Shih, Kuang-Hsun, Yi-Hsien Wang, I-Chen Kao, and Fu-Ming Lai. 2024. "Forecasting ETF Performance: A Comparative Study of Deep Learning Models and the Fama-French Three-Factor Model" Mathematics 12, no. 19: 3158. https://doi.org/10.3390/math12193158

APA Style

Shih, K.-H., Wang, Y.-H., Kao, I.-C., & Lai, F.-M. (2024). Forecasting ETF Performance: A Comparative Study of Deep Learning Models and the Fama-French Three-Factor Model. Mathematics, 12(19), 3158. https://doi.org/10.3390/math12193158

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop