1. Introduction
The volatility of financial asset returns plays a key role in financial practice, which forms one of the core subjects of modern financial theory. Among them, as a barometer of the financial market, the predicable volatilities of tickets’ prices are crucial for risk management and financial supervision, portfolio optimization, and financial derivative pricing, which has always been a research hotspot.
Since 2020, COVID-19 has spread all over the world. The pandemic has had a severe impact on the global economy. The CSI300 Index, which is considered the ‘Blue Chip’ index for the Mainland China stock exchange, has experienced a maximum drawdown of 33.52% since 2021, marking the largest drawdown since 2015. In February and March 2020, the S&P 500 Index plummeted five times, triggering a market meltdown. Many investors incurred losses due to the abnormal volatility in the financial market. Consequently, the tail risk of asset returns under extreme volatility has become the focus of scholarly research. Volatility plays a crucial role in various areas of finance, such as derivative pricing, portfolio risk management, hedging strategies, and systemic risk. Therefore, it is valuable for investors to utilize volatility information effectively in constructing their trading strategies.
Modeling the economic mechanism and pricing of assets has been a crucial task in economics, with various methods employed to estimate the mean and variance of prices. One such approach is the autoregressive conditional heteroskedastic (ARCH) model, introduced by Engle [
1]. This model was further extended by Andersen and Bollerslev [
2] through the development of the generalized ARCH (GARCH) model, featuring a more rational lag structure. Later, Nelson [
3] proposed the exponential generalized ARCH (EGARCH) model, which incorporated an exponential component to better capture extreme events. Despite their contributions, these traditional time series models rely on simplified assumptions that may not always hold true in practice. As such, there exists a need for alternative or supplementary methodologies that can address these limitations and provide more accurate inflation rate estimates.
The application of artificial neural networks (ANNs) in finance has gained significant traction in recent years, particularly in the areas of volatility prediction and stock market forecasting. Barunik and Krehlik [
4] pioneered the use of ANNs in energy market volatility prediction, demonstrating improved accuracy with high-frequency data. Notably, Hochreiter and Schmidhuber [
5] introduced long short-term memory (LSTM) algorithms, a type of recurrent neural network (RNN), which has since become a widely used tool for tackling complex tasks with long time lags. Chen et al. [
6] successfully applied LSTM to predict Chinese stock returns, showcasing its potential in stock market prediction, which also indicated a possible way to predict volatility than the strategies currently used. Kim and Won [
7] developed a hybrid model combining LSTM with multiple GARCH-type models to improve realized volatility forecasts for the KOSPI 200 index. Their findings indicated that the integrated model outperformed individual GARCH-type models. The increasing availability of high-frequency financial data has fueled research in this area, driving the development of novel techniques and architectures to harness the power of advanced machine learning methods. As data science continues to evolve, the intersection of AI and finance holds great promise for unlocking new insights and improving decision-making processes in the investment industry.
According to Hornik et al. [
8], artificial neural network (ANN) models possess the ability to approximate continuous functions without imposing restrictions on the underlying data generation process, as demonstrated in D’Amato et al. [
9]. Numerous studies have shown that ANN models excel over traditional GARCH-type models in volatility prediction due to their capacity to capture nonlinearity and their lack of requirement for stationarity in the series (Tapia and Kristjanpoller [
10]; Bahareh Amirshahi, Salim Lahmiri [
11]). Notably, hybrid models combining deep learning and GARCH-type models exhibit superior performance compared to single deep learning or time series models (Kristjanpoller and Hernández [
12]; Vidal and Kris-tjanpoller [
13]). This paper contributes to the field by highlighting the significance of intelligent algorithms and economic connections in volatility prediction, offering a unique perspective on the interplay between these factors.
Realized volatility, a concept introduced by Andersen and Bollerslev [
2], has revolutionized the way we measure and understand volatility in financial markets. By utilizing high-frequency sample data, realized volatility captures ex post volatility and provides a more comprehensive picture of market fluctuations compared to traditional measures. Building upon this concept, Shao and Yin [
14] developed a realized volatility model and a realized range model, which were used to compute value at risk (VaR) using intraday high-frequency data. Their work demonstrated that models based on intraday data significantly outperform those relying on daily returns, highlighting the importance of high-frequency data in volatility modeling. Furthermore, Kuster et al. [
15] emphasized the critical role of accurate volatility predictions in estimating VaR, underscoring the significance of developing sophisticated models capable of capturing the complexity of modern financial markets.
This paper offers several significant contributions to the field of financial risk management. First, it employs multiple models to study realized volatility (RV), thereby enhancing the accuracy and robustness of predictions. Second, it performs out-of-sample forecasts to evaluate the performance of the developed models. Third, it utilizes estimated value at risk (VaR) to conduct risk analysis. Fourth, it combines artificial intelligence algorithms and traditional volatility models, not only improving model performance but also highlighting the relevance of each variable. Lastly, it provides a reference model for investment and risk management that can contribute to market pricing efficiency and stability.
Previous studies have demonstrated the superiority of hybrid models combining deep learning and GARCH-type models in volatility forecasting for cryptocurrencies (Bahareh Amirshahi and Salim Lahmiri [
11]; Kristjanpoller and Minutolo [
16]). In contrast, our study applies this approach to the stock market, which has a larger market value and greater practical significance. While Ramos-Pérez et al. [
17] and Liu [
18] utilized hybrid models to predict volatility in the S&P500 and Kim and Won [
7] examined the volatility of the Korean stock price index (KOSPI 200), these studies neglected the underlying economic mechanisms driving volatility. Our research addresses this gap by incorporating economic insights into hybrid artificial intelligence algorithms, rendering it the first study to bridge this divide. By doing so, we expand upon existing research and underscore the significance of economic variables and econometric models in volatility analysis.
The organization of this paper is as follows: In
Section 2, we conduct a literature review of relevant studies on realized volatility, GARCH-type models, and LSTM. We then propose a hybrid model that combines these approaches to better capture volatility and predict it. Next, we outline the basic models used in our study.
Section 3 presents the empirical results of our models and compares them with traditional models. In
Section 4, we discuss the potential applications of our models, including systemic risk prediction and portfolio management, as well as robustness tests. Finally, we conclude with a summary of our findings and implications for future research in
Section 5.
3. Results
This section, structured by subheadings, offers a condensed yet comprehensive account of the experimental outcomes, their meaningful interpretations, and the subsequent conclusions derived from the data. Through a meticulous examination of the results, we unveil new insights into the phenomenon under investigation, furnishing the field with valuable knowledge and paving the way for further investigations and practical applications.
3.1. Data
The historical 5 min trading data of the CSI300 Index and macro variables employed in this study were sourced from JoinQuant. Specifically, the CSI300 Index is designed to mirror the performance of the top 300 stocks listed on the Shanghai Stock Exchange and Shenzhen Stock Exchange. Our dataset encompasses 68,976 5 min data points and 1437 daily data points spanning the period of 23 August 2016 to 22 July 2022. To train the LSTM model, we utilized 90% of the data in the training set as a holdout set for model fitting and 10% as a validation set for hyperparameter tuning. Notably, the within-sample period ranges from 23 August 2016 to 18 January 2021, while the outside-sample period covers 21 January 2021 to 22 July 2022.
The following table shows the description statistics of the return and RV (realized volatility) of the CSI300 Index. As shown in
Table 2, the mean daily realized volatility (RV) of the CSI300 Index is 0.0082, and the standard deviation is 0.0035; the minimum value of RV is 0.0025, and the maximum value of RV is 0.0296; the mean daily return of the CSI300 Index is 0.0002, and the standard deviation is 0.0120; and the minimum daily return is −0.0821, and the maximum daily return is 0.0578. The ADF test result of return is −19, which means the return is a stationary series at the 1% level.
3.2. Volatility Prediction
This study leverages the rolling-time-window technique for volatility forecasting. Our approach involves first training three individual GARCH-type models, including GARCH (1,1), EGARCH (1,1), and GED-EGARCH (1,1). We then integrate these models with LSTM and macro variables to create a hybrid model. Additionally, we incorporate inputs such as the interbank offered rate, consumer price index (CPI), industrial growth, and consumer sentiment. To accommodate data limitations, we apply a one-period lag for macro variables. Finally, we evaluate the out-of-sample predictive performance using three loss functions: mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE).
According to
Table 3, the GARCH model exhibits the strongest performance among the GARCH-type models, with a mean absolute error (MAE) of 0.0035, root mean square error (RMSE) of 0.0043, and mean absolute percentage error (MAPE) of 0.4483. In comparison, the EGARCH model has a slightly higher MAE of 0.0041, while the GED-EGARCH model has an MAE of 0.0040, both of which are inferior to the GARCH model’s performance. Similarly, the RMSE and MAPE measurements also indicate that the GARCH model outperforms the other two models.
Our hybrid model, which integrates LSTM and GARCH-type models with macro variables, exhibits superior performance compared to the standalone GARCH model, as demonstrated in
Table 3. Specifically, the hybrid model achieves a mean absolute error (MAE) of 0.0020, root mean square error (RMSE) of 0.0027, and mean absolute percentage error (MAPE) of 0.2233. These values represent improvements of 43%, 37%, and 50% over the GARCH model, respectively. The table clearly shows that the inclusion of macro variables in the hybrid model leads to the most accurate predictions.
According to Kim and Won [
7], the mean absolute error (MAE) of the growth-based exponential smoothing–long short-term memory (GEW-LSTM) model is 0.0107, which represents a 37.2% reduction compared to the ensemble–dual factor nested (E-DFN) model (0.017). Additionally, the GEW-LSTM model exhibits inferior performance in terms of mean square error (MSE), half-life autoregressive moving average (HMAE), and half-life moving average (HMSE), with reductions of 57.3%, 24.7%, and 48%, respectively. Our models, which integrate macro variables, demonstrate even lower MAEs than those reported by Kim and Won [
7]. Moreover, our hybrid models outperform GARCH-type models, suggesting their superiority in predicting stock market volatility.
Figure 1 depicts the comparison between predicted and realized volatility for both GARCH-type models (part a) and the hybrid model (part b). The GARCH-type models’ predictions are contrasted with the actual volatility, which serves as the target value in this study. On the whole, the hybrid model’s forecasts are likewise compared to the realized volatility.
5. Conclusions
This study introduces a novel hybrid model that seamlessly integrates multiple GARCH-type models with long short-term memory (LSTM) networks to capture a wide range of economic characteristics. The GARCH (1,1) model and EGARCH model are employed to reflect the magnitude of volatility shocks, persistence of volatility, and leverage effects, respectively. These features are then fed into an LSTM network, which exhibits remarkable capabilities in identifying high-level temporal patterns in time series data. Furthermore, the incorporation of macroeconomic variables, such as interbank lending rates and consumer price indices, provides valuable information for long-term risk assessment. Our comprehensive evaluation of the hybrid model’s performance, conducted using three distinct loss functions and CSI300 Index data, demonstrates its superiority over single GARCH-type models in predicting realized volatility. The hybrid model’s ability to learn from multiple sources of information enhances its predictive accuracy, making it a promising tool for financial risk management.
The hybrid model, which synergistically combines GARCH-type models and LSTM, yields a substantial improvement in prediction performance compared to single GARCH-type models. Additionally, incorporating macro variables, such as interbank lending rates, as inputs to the LSTM model further enhances its predictive accuracy. Statistical comparisons reveal that the hybrid model with the optimal macro variable selection achieves improvements of 43%, 37%, and 50% in mean absolute error (MAE), root mean squared error (RMSE), and mean absolute percentage error (MAPE), respectively, relative to the best-performing single GARCH model. Consequently, the out-of-sample prediction error of the hybrid model is demonstrated to be the lowest across all evaluation metrics, underscoring its superior forecasting capability.
The present study’s empirical findings demonstrate that the proposed hybrid model significantly enhances the value-at-risk (VaR) prediction performance for the CSI300 Index. By implementing a basic trading strategy based on the predicted VaR values, the cumulative return is found to increase significantly under 90% and 99% confidence levels. These results suggest that the developed hybrid model offers considerable potential for practical applications in finance, contributing to the advancement of risk management and investment decision-making. The methodology and conclusions presented in this study pave the way for future research endeavors to build upon and expand the scope of this innovative approach.
The limitations of our study arise from the fact that our models need to be estimated separately for each asset, precluding a universal application. However, our findings offer valuable insights for traders and market participants, who can utilize our framework to evaluate the volatility of their portfolio holdings and determine the optimal critical level for adjusting their positions, thereby effectively managing risks. While our analysis has focused on two prominent Chinese stock indexes, the applicability of our models extends to other assets, including those in the US stock market, providing fertile ground for future research endeavors.