Time-series data are among the most frequently gathered forms of information. This particular data type has the potential to be employed in the construction of a model that can accurately depict the actions of a dynamic system [
1]. Many approaches exist to accomplish this, each of which has undergone comprehensive research and demonstrated practical utility [
2,
3]. A primary challenge encountered in time-series analysis, prevalent in numerous fields, including computing, statistics, and banking, is selecting the most accurate forecasting model from historical data.
Many methodologies and strategies have been devised to construct prediction models from time-series data to approximate their future motions as stated in the introductory paragraph (see, for instance, references [
4,
5]. Deep learning is among the most efficient techniques for predicting time series. Artificial neural networks are helpful in forecasting time series because the models can autonomously identify detailed patterns within the data and circumvent conventional statistical techniques. When predicting time series, RNNs and variations on them, such as LSTM networks, demonstrate excellent performance. Recurrent neural networks, or RNNs, are algorithms specifically built to analyze sequential data, including time series. These models do this by preserving a hidden state that contains all the observations made up to this point [
6,
7]. In training deep networks using sequential data, vanishing gradients may manifest. In response, LSTM networks, derivatives of RNNs, were devised [
8]. Time-series data can be enhanced with the temporal feature extraction capabilities of CNNs, while long-range correlations between time steps can be discovered with transformer networks. One can use either of these methods to improve their time-series data. However, while most people agree that data attributes should guide model selection for time-lapse forecasting in deep learning, there are exceptions; the extensive evaluation of various models has sparked debates regarding the most suitable architecture or model for such data [
9]. Although transformer models have shown great success in natural language processing (NLP) and other domains, they are not always the best choice for predicting stock exchange closing prices. Stock price data are time-series based, meaning they have a natural sequential structure, where past values strongly influence future ones [
10,
11].
Traditional models like ARIMA, LSTM, or GRU are explicitly designed to handle this sequential dependence. While transformers can process sequences, their parallelized approach could be more intuitive for handling long-term temporal dependencies than models specifically designed for time-series data. Further, market volatility makes stock price data noisy and often unpredictable. With their complex architecture and large number of parameters, transformers may be more prone to overfitting on these noisy data, especially if the dataset is not large or diverse enough [
12,
13]. It is crucial that time series exhibit periodicity. Seasonal decomposition and ARIMA are two techniques that may prove beneficial when the time series demonstrates a clear and uninterrupted periodic pattern. Complex models, such as those comprising multiple deep networks, may improve performance when the time series displays irregular patterns or cycles. Aspects that include the complexity, accessibility, and necessary interoperability of the data may affect the final determination [
14]. Consequently, the model should align with the research’s data and objectives. The fundamental goal of this endeavor is to improve the accuracy of financial time-series forecasts by applying deep learning models. Finding out which deep learning model excels in predicting financial time-series observations is the driving force for this study. This study will employ a multivariate hybrid deep learning model capable of handling multivariable input and predicting the close price of the stock index. Unlike univariate models, multivariate models yield higher accuracy predictions for dynamic systems that comprise multiple interconnected variables, such as financial systems [
15] and capital market systems [
16]. This is the primary rationale for selecting this methodology for the present study.
Furthermore, predictions will be made simultaneously on all variables for the deep learning models that will be assessed. One potential approach to improve the prediction accuracy is to employ a trained deep learning network that simultaneously generates predictions and records the relationship between variables [
17]. A “multivariate” approach integrates the above-mentioned strategies into the deep learning model. Forecast accuracy may be affected by both the features of time-lapse data and the design of a deep neural network model; this is another research objective. A literature summary sets the stage for the study’s description of the research methodologies used to compare and contrast the observed deep neural network models’ performance.
To determine the most appropriate neural network technique for predicting the closing price of a stock market index from an OHLCAV (Open-High-Low-Close-Average-Volume) time-series dataset, which includes data points such as the stock’s opening price, highest price, lowest price, closing price, adjusted closing price, and the volume of stock traded in a single day,
Section 3 establishes the tests and analyses, and examines the evaluation outcomes. The study will be concluded in
Section 4 through a comprehensive summary of its main findings, some concluding remarks, and a delineation of its long-term objectives and approaches for future research.
1. Literature Review
This study will discuss a substantial issue related to the research topic. The problem is examining the methodology employed to forecast the economy’s condition. The primary challenge in time-series forecasting is to employ practical approaches that align with the characteristics of the data. Employing conventional statistical methods like regression analysis and ARIMA models may pose challenges because of the non-linear characteristics of financial data. Given the possibility of inaccurate results, it may be unwise to depend extensively on these approaches for forecasting the economy’s future performance. Numerous scholars have dedicated significant time and effort over the past several years to employing sophisticated methods to predict the stock markets of foreign countries. Global stock exchange market datasets are the basis for numerous academic research endeavors.
Several advanced deep learning architectures and models have been developed in recent years as feasible solutions for forecasting time-series data and other applications. These models and architectures have been produced [
18,
19]. A frequently chosen architecture for deep learning applications that involve predicting time-series data is the Convolutional RNN and Convolutional LSTM systems. According to studies by [
20], using LSTM to forecast the flow of financial data produces highly accurate results. The CNN is an important deep learning architecture that, along with the LSTM, can predict the future development of time-series data [
21]. The main impetus behind the development of CNN was to enhance the efficiency of data feature extraction to facilitate classification [
22]. In contrast, several studies have shown that CNNs can effectively make precise predictions on time-series data [
23].
Several other deep learning models, such as ConvRNN [
24], ConvLSTM [
25], ConvGRU [
26], Bi-directional ConvRNN [
27], Bi-directional ConvLSTM [
28], and Bi-directional ConvGRU [
29] integrate LSTM and CNN characteristics. LSTM is designed to store data motion patterns, and CNN is designed to extract features. Integrating the two models is to take advantage of each model’s qualities. As a consequence of this integration, the precision of time-series forecasts will subsequently improve. This study aligns with previous research carried out in the field of time-series prediction and uses the integrated deep learning approach. Ref. [
30] also confirmed the effectiveness of this strategy in predicting air quality. Furthermore, previous studies have indicated that CNN and LSTM can accurately predict the onset of COVID-19 [
31]. The researchers at [
32] conducted a study to validate the effectiveness of various models for complex wind speed forecasting and specifically explored using mixed deep learning algorithms applied to wind speed data. A further study that used electrical load forecasting data [
33] demonstrated similar findings, as it evaluated and empirically analyzed the most efficient deep learning models suitable for predicting short-term load.
This article evaluates six prominent deep learning models: the Convolutional RNN, Convolutional LSTM, Convolutional GRU, Bi-directional ConvRNN, Bi-directional ConvLSTM, and Bi-directional ConvGRU. These models can predict time series and may be applied to datasets with multiple variables. References to earlier research and the characteristics of the models served as inspiration for this choice.
Figure 1 presents a comparative examination of the six deep learning models’ input/output flow topologies.
1.1. Deep Neural Network
This part of the study summarizes the frameworks employed for deep learning, specifically Convolutional RNNs, Convolutional LSTM, and Convolutional GRUs, listed in that sequence.
1.1.1. Convolutional Neural Network
A certain kind of neural network, termed Convolutional CNNs, is distinguished by its grid-based topology and the application of mathematical convolution to analyze input data [
34]. Multiple-layered Convolutional CNN is utilized in natural language processing applications to extract local features. These networks process the input features using linear filters through the convolution process [
35].
1.1.2. Recurrent Neural Network (RNN)
An RNN, short for recurrent neural network, is a specialized neural network explicitly designed for sequence modeling. A directed graph is a representation that can be used to characterize the connections that exist between neurons in an RNN [
36]. The RNN is particularly suitable for Natural Language Processing (NLP) applications because it can analyze input sequences using its internal state. The outputs of the RNN are generated by iteratively applying the identical function for every occurrence using the identical approach. Calculations that have been used before are employed to determine the performance. Within the RNN design, the input length is utilized to ascertain the size of each separate time step. The data given to the architecture at time step
t are denoted as
, whereas the hidden state at time step
t is denoted as
:
Utilizing the hidden state from the previous time phase and the current input, as referenced in the study [
37], it is possible to compute the current hidden state
using Equation (
1) given in the previous sentence. The input and recurrent of the cell at time
t are represented by the symbols
and
, respectively. The weights that are shared over time are denoted by the symbols
U and
W, while the bias is represented by
b. The pointwise non-linear activation function is represented by the function
f. Regrettably, the RNN cannot identify the relationship between pertinent information when the time interval is extensive. Long Short-Term Memory is suggested by the study [
38] as a solution to the “long-term dependency” issue.
1.1.3. Long Short-Term Memory (LSTM)
An LSTM is an architecture for deep neural networks that uses forget gates to solve the problems of vanishing gradients or bursting. Ref. [
39] states that it is essential to note that, in contrast to standard recurrent neural network topologies, LSTM facilitates loss backpropagation over a restricted amount of time steps in numbers. A standard LSTM unit comprises three gates: an entry gate, an exit gate, and a gate that can be forgotten. The cell itself is also included in this unit. Through monitoring the gates’ operation, the cell can determine which data should be stored and when information should be granted to other units. The LSTM transition is built on the foundation established by the subsequent equations [
38]:
The input vector to the LSTM unit is denoted as , the activation vector for the forget gate is denoted as , the activation vector for the input gate is denoted as , and the activation vector for the output gate is denoted as .
The cell state vector is denoted as
, whereas the hidden state vector is represented as
. In this model, the letter
W represents the weight matrices, whereas the letter
b represents the parameters of the bias vector [
40]. Bi-LSTM employs bidirectional hidden layers to encompass both historical and future information. As a result, there is an enhanced acquisition of knowledge inside the network and a more robust and reliable exchange of information in both directions over time.
1.1.4. Gated Recurrent Unit (BGRU)
Nevertheless, an alternative artificial neural network (ANN), Gated Recurrent Unit (GRU), is a type of architecture based on recurrent neural networks (RNNs) as proposed by [
41] in 2014. The LSTM architecture effectively maintains the weight of long-term memory, resolving the disappearing gradient problem in RNN. Nevertheless, LSTM requires extensive computations due to its intricate architecture. In most GRU architectures, an update gate is utilized instead of an LSTM. The modification is incorporated into the GRU design based on the subsequent equations [
42]:
Let represent the input vector, represent the output vector, represent the reset gate vector, represent the update gate vector, and W, U, and b represent the parameter matrices and vectors, respectively.
2. Proposed Methodology
An extensive literature review is conducted to identify several effective deep learning architectures commonly employed for time-series data modeling and prediction. The subsequent phase involves collecting time-series data from actual-world situations after a diverse array of architectures has been acquired. The third stage of this study consists of pre-processing the collected dataset, which involves normalizing the values based on the various ranges of current stock indices values. This article proposes a hybrid architecture to predict the close price of the Asian Stock Exchange. The architecture incorporates the random and temporal fluctuations of the exchange rate and the vulnerability to external market dynamics. A two-stage feature extraction technique is employed on the dataset to streamline the architecture’s creation process and improve its capacity to accurately predict outcomes beyond the data it was trained on. This is necessary because of the abundance of influential factors and the unpredictable nature of exchange rates. One application of Auto-Encode (AE) is to simplify complex elements involved in influencing. Due to the volatility of stock exchange rates, the Self-Organize Map (SOM) is employed to cluster related data from the training set into one category. The Convolutional Neural Network (CNN) addresses traditional neural networks’ limitations in effectively extracting features from time-series data. Due to its capacity to accurately predict based on time-series data, it is a highly potent tool. Bi-directional Gated Recurrent Unit is recognized as a very accurate and effective technique for non-linear integration, as it excels in accurately considering the weight of each propagation to predict the close price of Asian stock indices as shown in
Figure 2.
For this study, Yahoo Finance
https://finance.yahoo.com/ (accessed on 1 April 2024) is the source of the time-series data obtained for each of the six Asian stock markets with corresponding tickers BSE (BSESN), Hongkong (HSI), Taiwan Index (TWII), Japan Index (JPXGY), Indonesia Index (JKSE), and Korea Index (KS11). The reason for choosing Yahoo Finance over other sources is its availability of an open-source library named “YFinance”. Compared to many paid services, YFinance is free, making it an appealing option for researchers with limited resources. Further, YFinance seamlessly integrates with Python version 3.6 a popular language among data scientists and researchers. This allows users to easily combine it with other Python libraries, such as Pandas, NumPy, and Matplotlib, for data analysis and visualization.
The dataset contains six features: open price, close price, high price, low price, adjusted close price, and volume. During this investigation’s third and final part, the dataset obtained is modified to correspond to the various percentage ranges of stock market index values. Pre-processing is the term that is most generally used to describe this method. In the following step, the dataset passes through the feature normalization and scaling, followed by the processed data being separated into two unique sets: training and testing. Before applying the models to time-series data that contain several variables, the subsequent stage is to construct and refine the deep learning structures that have been chosen. The training dataset is now fed into the series of deep learning models named the convolutional recurrent neural network (ConvRNN), Convolutional LSTM model, Convolutional GRU model, Convolutional Bi-directional RNN model, Convolutional Bi-directional LSTM model, and Convolutional Bi-directional GRU model. After completing the training phase of all six deep learning models, followed by a performance evaluation, the trained models are tested on the stock exchange’s test data to predict the exchange’s index value. The conclusion stage of this process involves identifying deep learning models and determining their structure to achieve the best level of prediction accuracy and lowest Root Mean Squared Error (RMSE), Mean Absolute Percentage Error (MAPE), and Mean Absolute Error (MAE).
The approach outlined in this study differs from the non-parallel prediction strategy, which involves creating individual forecasts for each variable of interest. Predictions created in parallel utilizing data significantly diverge from those generated in a non-parallel manner due to the intricate nature of the procedure. A multivariate parallel approach incorporates a complete model that receives multivariable inputs and generates multivariable outputs, in contrast to non-parallel prediction, which predicts each relevant variable individually. The prediction procedure is executed in parallel for all variables. Parallel prediction involves the execution of the prediction process for all variables simultaneously, while non-parallel prediction involves the execution of the process for each variable of concern individually. By integrating a well-trained model that establishes a correlation between variables into the prediction process, the prediction accuracy can be improved, thereby making prediction advantageous.
3. Dataset
As mentioned earlier, this study utilizes time-series data from the financial field. Based on results from the prior research [
15,
29] that showed the interdependence of these stock indices’ movements, this study integrates data from six stock indices into a single multivariate dataset. Utilizing a multivariate modeling technique considers the impact and interrelation of various stock markets, leading to more accurate outcomes.
Figure 3 shows the daily fluctuations of stock market indexes in India, while
Figure 4 displays similar data from Hong Kong.
Figure 5 presents the data from Taiwan, and
Figure 6 illustrates the data from Indonesia.
Figure 7 presents the data from Japan, and
Figure 8 illustrates the data from Korea.
Information was gathered between 1 April 2020 and 31 March 2024. As a result of the COVID-19 pandemic in 2020, the financial market indexes went through a period of highly unpredictable volatility throughout the year, prompting us to select our era. The authors can employ comprehensive data from the whole duration of the COVID-19 pandemic, which had a profound influence on the global economy, to assess the effectiveness of the statistical deep neural network model.
Figure 3,
Figure 4,
Figure 5,
Figure 6,
Figure 7 and
Figure 8 illustrate the financial data used in this experiment. They comprise roughly 770–800 data points, corresponding to the total number of trades from April 2020 to 31 March 2024 as mentioned in
Table 1.
3.1. Training–Testing
For training purposes, 80% data points are used out of roughly 770–800 data points for different exchange markets that represent the daily progression of stock market index values from 1 April 2020 to 31 March 2024. Using this information and up to nineteen data points daily, forecasts regarding the stock market index value fluctuations during March 2024 are created. This study utilizes the Python artificial intelligence libraries such as PyTorch, Scikit-Learn, TensorFlow, and Keras to create six deep learning models: ConvRNN, ConvLSTM, ConvGRU, ConvBRNN, ConvBLSTM, and ConvBGRU. The experimental methods are conducted as an ancillary inquiry utilizing a personal computer with an Intel Core i7 CPU, 32 gigabytes of random access memory (RAM), and the Windows 11 operating system. At the same time, the exhaustive parameter settings for every single deep learning model utilized in this investigation are presented in
Table 2. For Convolutional RNN, Convolutional LSTM, and Convolutional GRU, we use a single 1-D layer of 64 distinct convolutional filters with a maximum pool size of 2 followed by another 1-D CNN layer of 128 distinct convolutional filters with a pool size of 2 and another 1-D CNN layer of 64 distinct convolutional filters with a pool size of 2. For Convolutional Bi-directional RNN, Convolutional Bi-directional LSTM, and Convolutional Bi-directional GRU, we use a single 1-D layer of 64 distinct convolutional filters with a maximum pool size of 2 followed by another 1-D CNN layer of 128 distinct convolutional filters with a pool size of 2 and another 2-D CNN layer of 64 distinct convolutional filters with a pool size of 2. We use the same stride of 3 for every model to preserve more information in the pattern. We use a minimal learning rate of 1e-5 followed by the RELU activation function to minimize the error rate at every weight updation for each model. Further, every model has processed the training samples hundreds of times and updated their weights based on each sample’s error. The training samples, forty in number, are used in the forward pass for Conv. RNN, Conv. LSTM, and Conv. GRU, whereas forty training samples are used in forward and backward passes for Conv. BRNN, Conv. BLSTM, and Conv. BGRU. The choices of parameter values in all the six deep learning models for predicting BSE stock prices [
43] and in the studies [
41,
44] on how people moved around during the COVID-19 pandemic and how deep learning models were used to predict Asian stock market indices show how parameters are set up and structured.
Each of the six models is trained using a subset of the collected financial time-series dataset. The subsequent stage involves comparing the current values of each stock market index with their respective forecasts. It is determined that the Root Mean Squared Error (RMSE), Mean Absolute Percentage Error (MAPE), and Mean Absolute Error (MAE) are the most effective methods for determining the accuracy of each model’s predictions in this investigation [
45]. This study evaluates the deep learning models’ performance using one-step prediction accuracy instead of multistep predictions. Pragmatic considerations drive the decision to use a one-step prediction technique. Due to the inclination of stock market investments to prioritize short-term goals, it is crucial to have more accurate projections for this specific time frame.
3.2. Results and Discussion
The comparative analysis of training loss and value loss for all six Asian indices shown in
Figure 9a–f is achieved by the Bi-Directional Gated Recurrent Unit (BGRU) and evaluated during the training phase. The training phase for all indices is executed with 100 epochs in BGRU, and it is observed that
Figure 9a,c,f show almost the same loss at the 100th epoch. The decrease in loss during training is depicted for the BSE, HSI, TWII, JKSE, JPXCY, and KS11 exchange, respectively. The study employs data gathered between 1 April 2020 and 31 March 2024 to build the BGRU model. Upon further scrutiny of
Figure 9b, it becomes apparent that the loss value tends to stabilize after the 80th iteration (epoch) when all deep learning models have an epoch set to 100 to complete the training as indicated in
Table 2. In addition,
Figure 9a,f demonstrate that the BGRU model achieves the fastest stability in reducing the loss and value for HSI and KS11 exchange, while the BGRU model achieves the slowest stability in the case of JKSE.
From
Figure 10a–f, the predictions made by the BGRU model for all six Asian exchange markets stock markets on 31 March 2024 are shown. To summarize, the assessment of the six deep neural network models demonstrates the ability to accurately replicate the observed patterns of index value fluctuations in the six Asian stock exchanges. On the other hand, the expected yield curve shows a significant difference compared to the original value curve. Compared to the different exchange markets, the BGRU model produces the TWII index curve, which exhibits the most significant deviation from the original value curve in
Figure 10c. Conversely, the BGRU model produces the BSE prediction curve that resembles the original value curve in
Figure 10a.
Figure 10c displays a variety of possibilities depicted on the Taiwan stock market chart. The prediction curve of TWII has the highest level of deviation from the original value on this graph compared to the curves of the other exchange markets. The BGRU model accurately creates the prediction curve for the KS11 as shown in
Figure 10f. In addition, BGRU achieves the highest accuracy in predicting the curve for JKSE, while the BSE exchange remains an excellent performer for BGRU. These data show that none of the six Asian exchange markets under study have the highest level of prediction accuracy.
Further, the authors employ all six deep learning models to evaluate comparative loss and value on the BSE exchange market throughout the training phase.
Figure 11 illustrates the comparative loss and value decrease accomplished in these models. The relative decreases in training loss are shown in
Figure 11a–f for the Conv. RNN, Conv. LSTM, Conv. GRU, Conv. BRNN, Conv. BLSTM, and Conv. BGRU models that are used. The BSE exchange is created using data collected from 1 April 2020 to 31 March 2024, as stated in the paper.
Table 2 demonstrates that all deep learning models have an epoch value of 100. Moreover, upon careful examination of
Figure 11, it is evident that the loss value tends to reach a state of stability around the 91st iteration (epoch). Also,
Figure 11b and f show that the Conv. LSTM and Conv. BGRU models have the fastest loss and value reduction, which shows that they are stable, especially when compared to the Conv. RNN and Conv. LSTM models. Conversely, Conv. RNN exhibits the slowest rate of achieving stability.
Further, the six assessed deep learning models successfully replicate the observed patterns of index value fluctuations in the six stock markets shown in
Figure 12a–f. On the other hand, when compared to the initial value curve, the predicted yield curve demonstrates a considerable divergence. The Convolutional RNN model generates the BSE index curve that displays the most substantial deviation from the initial value curve compared to the additional models. The Convolutional BGRU model, on the other hand, generates the BSE prediction curve, which is comparable to the value curve that is first determined in
Figure 12a.
Figure 12 displays various possibilities on the BSE stock market chart. The prediction curve of Convolutional LSTM shows the most departure from the beginning value on this graph compared to the curves of the other models. The Convolutional Bi-directional RNN model accurately creates the prediction curve for the BSE as shown in
Figure 12d. In addition, the Convolutional BGRU model has the highest level of accuracy in predicting the curve for BSE. In contrast, the Convolutional BLSTM model remains the top performer in performance for BSE. The statistics presented here demonstrate that none of the six investigated deep learning models achieve the highest accuracy regarding the generation of predictions.
Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and Mean Absolute Percentage Error (MAPE) are the metrics that are utilized to measure the disparities that exist between the values that were anticipated and those that were observed.
Table 3 and
Table 4 illustrate these distinctions in a graphical format. The fact that the error rates for RMSE and MAPE are comparable is illustrated by this, which shows that standalone models can deliver findings that are superior to those produced by combining models.
Table 3, which can be located elsewhere, presents a comparison of the error rates that are observed during the training and testing stages of the process. During the training phase of the calculation, it is not unusual for the error rates to increase when dealing with time-series prediction. This is because of the nature of the problem. The fact that this particular case exists proves that this pattern does exist. The Convolutional LSTM algorithm has a value of 0.1404 in terms of the Root Mean Squared Error (RMSE), the Convolutional GRU algorithm has a value of 0.1582, the Convolutional RNN algorithm has a value of 0.1450, the Convolutional BLSTM algorithm has a value of 0.1494, the Convolutional BGRU algorithm has a value of 0.1205, and the Convolutional RNN algorithm has a value of 0.1545. The Convolutional BGRU and the Convolutional LSTM combinations are two models with the lowest total Root Mean Squared Error (RMSE) estimates. When simultaneously forecasting financial time-series data with many distinct components, the preliminary findings indicate that these two models perform much better than their competitors.
Further,
Table 5 presents the comparative analysis of outcomes of the suggested hybrid model, where layers of each of the six models are included and tested on the BSE stock exchange and other recent frameworks on RMSE, MAPE, and MAE evaluation parameters. The proposed hybrid model achieves an RMSE of 0.0122 while tested on the BSE index, and that is the least in comparison with other indices due to the large corpus available in BSE. Further, the hybrid model achieves a MAPE of 1.31% and an MAE of 0.0082 on the BSE index’s test data.
Based on the study conducted by other researchers, it has been consistently shown that hybrid models tend to outperform individual models across various types of datasets. While the difference between hybrid models may not be statistically significant, it has been discovered that using them produces better results in predicting the stock exchange close price than the traditional approach of deep learning models. Despite no significant and statistically meaningful difference between the hybrid models, this phenomenon continues.
4. Conclusions
Empirical studies and research findings indicate that no deep learning model can provide precise predictions for all potential states within the framework of time-series data. When using several models, such as Convolutional RNN, Convolutional LSTM, Convolutional GRU, Convolutional BRNN, Convolutional BLSTM, and Convolutional BGRU, to predict the movement of time-series data, it is evident that the performance varies for each dataset. The study also found a strong association between the level of instability in time-lapse information and the ideal architecture of the deep neural networks model for precise forecasting. A single deep learning model with consistent time-series data produces precise predictions compared to many models that work with unorganized data. Even though the framework suggested in this article can accurately forecast stock indices in a close process, numerous obstacles still need to be overcome in stock forecasting. Initially, the stock data utilized in this paper are exclusively from Asian markets and do not accurately predict stock indices in the Middle East, Africa, or Europe. Secondly, enhancing the parameters for tuning CNN and BGRU is necessary. The framework is posited to achieve superior outcomes among the stock indices’ intensive parameter adjustments. Furthermore, we believe that numerous hybrid models still need to be addressed in this article, which can potentially improve the error in predicting the close price of stock indices, even though the Convolutional BGRU hybrid approach is employed in this paper. Nevertheless, the Convolutional BGRU-equipped stock indices prediction approach proposed in this article can assist investors in making accurate judgments in the stock exchange to a certain extent.
Additional research is underway to create an independent deep learning system that can identify the most appropriate model depending on the unpredictable characteristics of the input data. To validate the conclusions of this research and create a deep learning model for predicting financial time series, it is crucial to acquire extensive and varied financial data from other businesses.