2.2. Technical Analysis
Technical analysis-based methods posit that all information pertaining to the stock market will be reflected in price fluctuations. It is sufficient to use historical stock data directly to predict future stock trends. Specifically, assuming there is a time series
, the objective is to forecast future values
using the gathered historical data [
40]. There are numerous methods currently employed for stock market forecasting, primarily based on machine learning techniques, which can be divided into the following more specific categories of methods: traditional machine learning and deep learning [
41,
42].
In the field of stock forecasting, many conventional machine learning models have been widely utilized [
43,
44,
45]. Traditional machine learning methods include support vector machines [
43,
46], decision trees [
47], Naive Bayes [
48,
49], and so on. However, the inability to extract deep features from stock data to bridge the gap in predictive ability between machines and humans is one of the primary limitations of traditional machine learning methods. Beyond that, the performance of machine learning methods largely depends on the quality of features. Unsuitable features are likely to result in poor model performance.
Recently, there has been active research on deep learning-based methods for stock prediction, which have achieved remarkable performance [
45,
50,
51,
52]. The primary distinction between traditional machine learning methods and deep learning is that the network structure can independently learn effective features according to the target task. Selvin et al. [
53] proposed a hybrid deep model consisting of three modules: RNN, LSTM, and CNN, to predict stock prices. In [
54], RNN architecture is used to forecast stock market trends. Using real data from ten Nikkei companies as examples, the validity of the method is verified. In [
55], the authors used gold price, gold volatility index, crude oil price, and the crude oil price volatility index. This combination has demonstrated strong performance in predicting future stock price trends. Zhao et al. [
50] proposed a novel deep model that consists of an emotion-enhanced convolutional neural network, denoising autoencoder models, and LSTM for stock price movement forecasting. Long et al. [
56] introduced a multi-filter neural network to extract effective features from financial time series samples. In [
41], Lee et al. developed an end-to-end architecture for stock forecasting, which includes two feature extractor networks designed to learn high-level features from time series stock data. Qin et al. [
57] proposed a temporal attention mechanism model for predicting stock prices.
Graph Neural Networks (GNNs) fundamentally operate on the principle that the state of any given node depends on the states of its neighboring nodes. This assumption allows GNNs to effectively capture the spatial interdependencies among nodes while preserving the temporal dynamics of the data. GNNs have proven effective in capturing spatial dependencies across a wide range of domains. However, their application in stock market prediction has been somewhat limited compared to other areas. The proposed MG-Conv model integrates one-dimensional and multi-graph convolutional networks within a three-layer framework [
58]. The one-dimensional convolutional layer normalizes the data and extracts features, while the graph convolutional layer constructs both static and dynamic graphs to perform convolution. Performance evaluations conducted on 42 Chinese indices demonstrate that the MG-Conv model reduces the average prediction error by 5.11% compared to alternative methods. However, overfitting remains a concern, and future research may focus on enhancing generalization and improving the fusion of index trends. Wang et al. [
59] tackle the challenge of stock index prediction by addressing the high noise and dynamic nature of stock data. The inadequacy of existing methods to capture local spatial–temporal correlations prompted the development of the LoGCN model, which integrates graph construction, convolution, and pooling mechanisms. Experimental results on 42 Chinese stock indices demonstrate that the LoGCN model outperforms traditional methods such as MLP and LSTM across various evaluation metrics, including regression, classification, and market back-testing. Future research directions will focus on enhancing classification accuracy and devising innovative investment strategies. Ma et al. [
60] aim to address the challenges of stock price prediction. Existing methods often rely on predetermined graphs or overlook certain correlations. The proposed VGC-GAN model generates multiple correlation graphs from historical data, utilizing techniques such as Pearson correlation, Spearman rank correlation, and FastDTW. Additionally, it employs a Variational Mode Decomposition (VMD) algorithm with optimized parameters obtained through a Genetic Algorithm (GA) to manage data non-stationarity. The model’s framework includes a generative adversarial module comprising a generator (Multi-GCN and GRU) and a discriminator, along with components for relationship graph generation and stock sequence decomposition. Experiments conducted on ETF, SSE, and DJIA datasets demonstrate that VGC-GAN outperforms methods such as GRU and GCGRU in terms of prediction performance and computational efficiency. Liu et al. [
61] propose a novel model, ECHO-GL, for predicting stock movements. Existing models often exhibit insufficient or static stock relationships. ECHO-GL addresses these limitations by leveraging earnings calls. It constructs a heterogeneous graph (E-Graph) with various types of nodes and edges, employing mechanisms such as time assignment and sliding windows to ensure its dynamism. Qian et al. [
62] focuses on stock investment prediction. The stock market is complex, with prices affected by various factors. Traditional sequential and graph-based methods have limitations in capturing multifaceted and temporal influences. The proposed MDGNN framework uses a discrete dynamic graph to capture stock relations and their evolution. It includes an intra-day graph snapshot with a multi-relational graph construction and a hierarchical graph embedding layer. The inter-day temporal extraction layer uses a Transformer structure to handle temporal evolution. The prediction layer estimates the probability of a stock’s positive return.
The Transformer is another deep learning method used in stock market prediction, following CNN and LSTM. Its ability to process time series information has been proven in many other domains as well. Li et al. [
63] proposed an LSTM and attention-based model, which consists of a multi-input LSTM and an attention layer, followed by several ReLU layers to obtain the final prediction. Results show significant improvements over several LSTM or CNN-based methods on the CSI-300 index dataset with a time horizon of 4 years. Li et al. [
64] designed a hybrid deep learning model consisting of a Transformer, LSTM, Gate Recurrent Unit (GRU), and a high-frequency data adaptive decomposition architecture. The model is named Frequency Decomposition and includes a GRU Transformer (FDG-trans). Wang et al. [
3] utilize the Deep Transformer (DT) model to predict stock market indices, and the prediction performance was significantly better than that of other classic methods. Muhammad et al. [
31] developed a Transformer-based network with two input layers, three Transformer layers, one pooling layer, followed by two dropout layers, and two dense layers to make predictions on the Dhaka Stock Exchange, including daily, weekly, and monthly share prices. The results are promising. Sridhar and Sanagavarapu [
65] combine the Transformer with CNN to predict stock trend movement. The network structure comprises two multi-head layers, one Transformer, and two convolutional layers. The experiment showed that the model outperforms several models based on CNN or RNN on the S&P 500 dataset. Financial stock prices are complex to predict due to a multitude of influencing factors. Current methodologies, such as Recurrent Neural Networks (RNNs), Long Short-Term Memory networks (LSTM), Gated Recurrent Units (GRUs), and Transformers, each have their own limitations. The HPMG-Transformer model [
66] integrates the Hodrick–Prescott (HP) filter with a multi-scale Gaussian transformer. The HP filter decomposes stock time series into long-term and short-term fluctuations, while the multi-scale Gaussian transformer enhances the extraction of local features. Additionally, a multi-scale Gaussian prior is incorporated into the self-attention mechanism to improve the extraction of local contextual information. Li et al. [
67] focus on stock price forecasting, a challenging task due to market volatility. Existing methods have limitations in modeling stock correlations and accounting for market variations. The proposed MASTER model consists of five steps: market-guided gating, intra-stock aggregation, inter-stock aggregation, temporal aggregation, and prediction. It models both momentary and cross-time stock correlations and utilizes market information for automatic feature selection. MASTER offers a novel perspective on stock correlation modeling and effectively leverages market information for stock price prediction. Future research could explore improved methods for stock correlation mining and additional applications of market information. The efficient market hypothesis has inspired various methods for stock prediction; however, challenges remain in effectively integrating numerical and textual data. To address this issue, Zhang et al. [
68] proposed the CoATSMP model, which includes text and price feature extraction, a Transformer-based soft fusion method, and joint feature processing utilizing LSTM and a temporal attention mechanism. This model employs datasets from diverse sources and evaluates performance using metrics such as accuracy (ACC), Matthews correlation coefficient (MCC), F1 score, and area under the curve (AUC). Li et al. [
69] propose an innovative approach to enhancing stock price prediction by integrating Generative Adversarial Networks (GANs) with Transformer-based attention mechanisms. This study underscores the importance of accurate stock price forecasting in the financial sector, enabling investors to make informed decisions. The authors acknowledge the limitations of traditional statistical methods and position machine learning and deep learning as promising alternatives. The methodology includes data preprocessing to address imbalances in the distribution of news data and to enrich the dataset with technical indicators. The model architecture features a Variational Autoencoder (VAE) for feature extraction, GANs for generating synthetic data, and a Transformer-based attention mechanism for focused data analysis.
Compared to the other deep learning-based methodologies previously discussed, FDG-trans [
64], Transformer Network (TN) [
31], and DT [
3] are most closely related to the Deep Convolutional Transformer (DCT). In [
64], FDG-trans is built by integrating GRU, LSTM, and multi-head attention Transformers. Both TN and DT are stock prediction models based on the standard Transformer architecture. In contrast, our DCT model incorporates the separable fully connected layer and inception convolutional token embedding into the standard Transformer structure to enhance information richness and learn fine-grained features.
Due to the robust capabilities of deep learning models in autonomously extracting features from unprocessed data, contemporary methodologies for forecasting stock price trends favor these models.
Table 1 presents a comprehensive overview of research papers that utilize deep learning, categorizing them according to the types of features and model architectures employed. Nevertheless, there remains significant potential for enhancement in existing deep learning architectures aimed at predicting stock movements based on textual data and historical stock information. Forecasting stock price trends necessitates the analysis of time-series data, which is inherently reliant on stock-related information over time. Consequently, the primary challenge is to enhance the accuracy of stock movement analysis and prediction utilizing financial data, such as historical stock prices, while effectively addressing the issue of temporal dependence. The impetus for our research is to bridge this gap and investigate the efficacy of stock time series engineering through the application of the Transformer architecture.