Next Article in Journal
Estimation of Citarum Watershed Boundary’s Length Based on Fractal’s Power Law by the Modified Box-Counting Dimension Algorithm
Previous Article in Journal
On Control Polygons of Planar Sextic Pythagorean Hodograph Curves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Network-Based Analysis for Evaluating Conditional Covariance Estimates

Facoltà di Economia, Università degli Studi Niccolò Cusano Roma, 00166 Roma, Italy
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(2), 382; https://doi.org/10.3390/math11020382
Submission received: 13 December 2022 / Revised: 28 December 2022 / Accepted: 4 January 2023 / Published: 11 January 2023
(This article belongs to the Section Financial Mathematics)

Abstract

:
The modeling and forecasting of dynamically varying covariances has received a great deal of attention in the literature. The two most widely used conditional covariances and correlations models are BEKK and the DCC. In this paper, we advance a new method based on network analysis and a new targeting approach for both the above models with the aim of better estimating covariance matrices associated with financial time series. Our approach is based on specific groups of highly correlated assets in a financial market and assuming that those relationships remain unaltered at least in the long run. Based on the estimated parameters, we evaluate our targeting method on simulated series by referring to two well-known loss functions introduced in the literature. Furthermore, we find and analyze all the maximal cliques in correlation graphs to evaluate the effectiveness of our method. Results from an empirical case study are encouraging, mainly when the number of assets is not large.

1. Introduction

Modeling financial volatility is a crucial issue in empirical finance. Heteroscedasticity is a peculiar feature of a financial time series, as it is characterized by periods more volatile than others due to unpredictable outer events. The existence of heteroscedasticity is a major concern in a univariate approach aiming to estimate and forecast a given (financial) phenomenon. Estimating financial volatility becomes a more difficult task if we consider N time series (multivariate approach) and we are interested in understanding the co-movements of financial returns. In this regard, different multivariate general auto-regressive conditional heteroscedasticity (MGARCH) models have been presented in the literature. In finance, the main aim of MGARCH models is to predict the future values of the variance and covariance matrix of asset returns. Readers may refer to the very recent paper [1] and the references therein for a comprehensive review of the existing literature on MGARCH approaches. These methods play an important role in different fields of finance such as, for instance, in portfolio optimization [2,3,4], option pricing [5,6], energy markets [7], and analysis of contagion and volatility spillovers [8,9,10]. In the literature, MGARCH models are classified into different categories (see, e.g., [11]): (1) direct generalizations of the univariate models; (2) linear combinations of univariate models, such as, generalized orthogonal models and latent factor models; (3) nonlinear combinations of univariate models; and (4) nonparametric and semi-parametric models that constitute an alternative to the parametric estimation of the financial volatility and do not impose a particular structure on the data. In this paper, we will focus on categories (1) and (3).
In general, estimating a time-varying covariance matrix is difficult due to the problem’s dimension and the fact that a covariance matrix must be positive definite. The most renowned MGARCH model of category (1) is BEKK, which has the attractive property that the conditional covariance matrices are positive definite by construction. However, estimation of a BEKK model involves somewhat heavy computations due to the fact that it contains a large number of parameters. To overcome this problem, one approach was to decompose the conditional covariance matrix into conditional standard deviations and a conditional correlation matrix (models of category (3)). In [12,13], the authors introduced the dynamic conditional correlation model (DCC-GARCH), for which the conditional correlation matrix is designed to vary over time. This approach has one important advantage, that is, the number of parameters to be simultaneously estimated is reduced, as a complex optimization problem is disaggregated into simpler ones.
Several authors have provided a number of different modifications of the MGARCH models, some of which are mentioned in [1]. Other meaningful approaches explicitly incorporate the effect of measurement errors and time-varying attenuation biases into the covariance estimations and forecasts (see, e.g., [14]). Engle and Kelly [15] studied a special case of the DCC-GARCH model, the dynamic equicorrelation model, in which all correlations are identical. This can be a useful model in situations where the number of parameters to estimate is large.
A specific modification of the MGARCH models is provided in [16]. The authors directly modified the structure of the BEKK and DCC models by introducing a “target” matrix and compared their modified models to the original ones. The authors considered the long-run covariance matrix as a target matrix, which can be consistently estimated by the corresponding sample estimator and considered to be different versions of the BEKK and DCC models. They point out that by imposing positive definiteness and covariance stationarity for the different versions of BEKK is extremely complicated. For the DCC models, targeting can be useful to reduce the number of parameters to estimate. However, in [17] the authors proved that the estimation of DCC models with targeting may be inconsistent. From a theoretical viewpoint, the paper [16] is worthy of interest because the authors study the availability of the analytical forms for the sufficient conditions for consistency, asymptotic normality of the appropriate estimators, and computational tractability for large-dimension problems. However, in [16], no empirical tests were presented, and the authors claimed that it is not possible to provide an appropriate evaluation of which of the different MGARCH models is preferred.
In this paper, we advance a novel approach based on network analysis for evaluating the estimates of the time-varying correlation matrices in financial markets. We refer to the BEKK and DCC models and propose a variant of these two models by suitably modifying the log-likelihood function to maximize. We call the two resulting models the modified BEKK and modified DCC models, respectively. This modification consists of introducing a term in the function incorporating a loss measure based on the difference between the time-varying covariance matrices and the covariance matrix estimated with respect to the whole in-sample period. In actuality, as observed in [18], a financial market is characterized by some stylized facts. More precisely, there are often specific groups of assets that are highly correlated in such a way that positive price changes of one asset in the group determine positive price changes of all the other assets in the group, and these relationships remain unaltered over time. Hence, the idea behind our modification is that (extremely) high/low values of correlations observed between pairs of assets do not change too much during time. In particular, any pair of highly positive or negative correlated assets in the market remains highly (positive/negative) correlated during the given observed time period.
In a financial market, the correlation between assets can be represented via a correlation graph where, given a time period, assets are identified by vertices of a complete graph G and distances (weights) assigned to pairs of assets (edges) incorporate the dependency structure of returns. Mantegna [19] was one of the first who constructed asset graphs based on stock price correlations in order to detect the hierarchical organization inside a stock market. In order to highlight only groups of assets whose correlation is above or below a given threshold δ , one can extract a (sub)graph G ( δ ) of G that contains only a subset of the connections in G between pairs of assets. In this new graph G ( δ ) , two stocks are linked if and only if they exhibit high/low values of correlation. In other words, the observed subgraph G ( δ ) reveals the strong relationships between groups of assets in a financial market during a given time period T. Examples of such interconnected groups are, for instance, firms belonging to a specific industrial sector. We introduce an additional term in the log-likelihood function that takes into account the fact that the time-varying covariance or correlation matrices must not alter the clusters formed by highly correlated assets observed during T. Hence, with our method we estimate a modified MGARCH model, simulate T realizations of returns of each of the N assets, compute the correlation matrix with respect to the whole simulated in-sample period and, given the threshold δ , we obtain the corresponding simulated graph G S ( δ ) . Then, we compare the observed and simulated graphs G ( δ ) and G S ( δ ) , respectively. In particular, we compare all the maximal cliques of the two graphs. Given a general graph G, a clique is a set of interconnected stocks that forms a complete subgraph of G. Hence, maximal cliques in G ( δ ) and G S ( δ ) represent highly (positive/negative) correlated assets, that is, a stock that belongs to a clique is highly correlated with all other stocks in the clique. The comparisons allow us to verify if a modified MGARCH model has been able to correctly reproduce the volatility and all the strong relationships among assets of the given financial market and if a modified model is able to outperform the corresponding original MGARCH model.
In light of the above discussion, the contribution of the present paper is threefold. First, we propose a new alternative targeting method that does not alter the models’ structure. In fact, we modify only the log-likelihood objective function by introducing a suitable loss or distance measure between the long-run sample covariance or correlation matrix and the corresponding conditional covariance or conditional correlation matrices, respectively. Hence, we do not have to impose additional constraints for covariance stationarity or to guarantee that the matrices are definite positive. Second, for the first time, we provide an empirical analysis for evaluating modified and original models. We do this with the caveat that it is not our goal to definitely decide which of the different models is the best; rather, we seek to understand whether our modified models allow to better capture the strong relationships between assets in a stock market. We hope that the modified models can be more accurate than the original ones. Third, we advance a new method for evaluating the effectiveness of the estimated models through some specific tools commonly used in network analysis.
Our findings enable future analysis on volatility. In actuality, the importance of identifying the volatility effect and correlation in a dynamic way is crucial for efficiently managing the investment portfolios and executing optimal diversification of assets. This will reduce the risk and controls the changes that may occur in financial markets due to the economic situation in general. Our approach can build a precise mathematical model for managing financial data, which can be also incorporated in artificial intelligence algorithms to avoid the problems associated with the problems’ dimensions.
The paper is organized as follows: Section 2 reports notation and some basic definitions. Section 2.1 and Section 2.2 summarize the two MGARCH models considered in the paper along with the new log-likelihood functions used in the optimization/estimation phase. Section 3 reports the structure of the empirical financial dataset we used in the experimental phase. Finally, some conclusions and further research are depicted in Section 4.

2. Notation and Definitions

Consider a financial market formed by N assets. Let P j t be the daily closing price of asset j at time t, t = 0 , , T and r j t = log ( P j t P j t 1 ) be the corresponding log-return of asset j. In the following, Pearson correlation coefficients are used to detect dependencies between assets returns. A general MGARCH model is defined:
r t = μ t + ϵ t ϵ t = H t 1 2 η t ,
where r t is the N × 1 vector of log-returns at time t, ϵ t is the N × 1 vector of mean-corrected returns of N assets at time t, E[ ϵ t ] = 0, and Cov[ ϵ t ] = H t . The vector μ t represents the expected value of r t . Observe that μ t may be modeled as a constant vector or as a time series model. In this paper, we assume μ t constant.
H t is the N × N matrix of conditional variances and covariances of the unpredictable component ϵ t at time t. Finally, η t is the N × 1 vector of i.i.d errors such that E[ η t ] = 0 and E[ η t η t ] = I.
In a general MGARCH model we have to estimate the conditional covariance matrix H t , which, in addition, has to be positive definite for all t. Depending on the possible specifications for H t , there are different MGARCH models each belonging to one of the four categories mentioned in Section 1.
In order to estimate the conditional covariance matrix H t , a common issue is to resort to the maximization of an appropriate log-likelihood function L ( θ ) , where θ denotes the vector of all the parameters to estimate. Depending on the MGARCH model, the function L ( θ ) assumes a different appearance. It is well-known that the quality of the maximum likelihood (ML) estimation relies also on the assumed data distribution. In general, when dealing with models with conditional heteroscedasticity, the estimates are known to be asymptotically normal [20]. In our approach, we consider a multivariate Gaussian distribution for the standardized error η t , even if, in principle, our method can be applied assuming different distributions.
Given a stock market, consider H t the conditional variance and covariance matrix of the returns of the N assets at time t, and let R = [ ρ i j ] , with ρ i i = 1 , i = 1 , , N , be the global correlation matrix with respect to the sample period. The idea is that if asset i is highly positively/negatively correlated with asset j, then the (high positive/negative) correlation does not change during the observed period. Furthermore, we assume that the relationship based on the correlation between two assets i and j does not change during time whereas the value of the correlation coefficient ρ i j , can (obviously) vary.
Let δ > 0 be a threshold. We construct the correlation graph G ( δ ) = ( V , E ) where V is the set of vertices each representing an asset, and E is the set of edges of G ( δ ) , that is, the set of connections between pairs of assets. We assume that there is an edge ( i , j ) between assets i and j if and only if the (global) correlation coefficient | ρ i j | > δ . Let A ( δ ) be the adjacency matrix of G ( δ ) . Observe that the generic element a i j ( δ ) can be 0 or 1 if we assume an unweighted graph, otherwise a i j ( δ ) = w i j if we assume that the graph G ( δ ) is weighted with weights w i j assigned to each edge ( i , j ) and obtained as a function of the correlation coefficients ρ i j , that is, w i j = f ( ρ i j ) . Given a general graph G = ( V , E ) , the following definitions hold [21]:
Definition 1. 
A subset C V is called a clique of G if any two vertices in C are connected by an edge. The order q of a clique is the cardinality of C.
Definition 2. 
A subset C V is called a maximal clique of order q of G if C is not included in any other clique C of order q + 1 .
Other concepts not defined in the paper can be found in the book [21].
In a modified MGARCH model, the modified likelihood function to maximize is:
L ( θ ) t = 1 T [ D i s t ( A ( δ ) , A t ( δ ) ] ,
where D i s t ( · ) is (any) distance or divergence measure between the adjacency matrix A ( δ ) referred to the global correlation matrix R , and the adjacency matrix A t ( δ ) related to the conditional correlation matrix R t . Because A ( δ ) is the correlation matrix computed with respect to the sample period, it can be considered to be a target matrix.
In our framework, as a measure we can consider any distance or divergence measure between two N × N positive definite matrices P and Q. In this paper, we consider the well-known Kullback–Leibler (KL) distance (or divergence) [22] between P and Q, namely K L ( P , Q ) . As a statistical measure, it is a measure of the distance between two probability densities P and Q. In the case of Gaussian multivariate distributions, this distance is completely defined by the correlation matrices of the whole system. Thus, it can be interpreted as how a multivariate probability distribution represented by the matrix Q is different from a reference multivariate probability distribution represented by the matrix P. It is also well-known that it is not a metric, as this measure is not symmetric and does not satisfy the triangle inequality. Given matrices P and Q, the K L ( P , Q ) measure is:
K L ( P , Q ) = 1 2 log ( | Q | | P | ) + T r ( Q 1 P ) N ,
where the operator | · | computes the determinant of a matrix and T r ( · ) is the trace of a square matrix, which is the sum of its diagonal elements. In our approach, D i s t ( A ( δ ) , A t ( δ ) ) = K L ( A ( δ ) , A t ( δ ) ) .
Note that assuming the identity function as the function of the graph’s weights, i.e., w i j = ρ i j , i j , the matrix A ( δ ) corresponds, in fact, to the unconditional correlation matrix where, given δ , a i j ( δ ) = ρ i j 0 , i j , if and only if | ρ i j | > δ , a i j ( δ ) = 0 , i j , otherwise (i.e., | ρ i j | δ ), and further imposing a i j ( δ ) = 1 , i = 1 , , N . Denote by Z ^ the matrix A ( δ ) defined so far. Under the above hypothesis on the graph’s weights, A t ( δ ) can also be considered to be a conditional correlation matrix at time t. Hence, in this special case, in order to preserve the strong relationships among specific groups of assets, in the maximization of the modified likelihood function (2) we are, in fact, requiring that a i j ( δ ) and a i j , t ( δ ) be as close as possible, that is, the corresponding correlation values | ρ i j | and | ρ i j , t | be as close as possible. In the rest of the paper, we assume w i j = ρ i j , i j .

2.1. The BEKK Model

Ding and Engle [23] introduced the d i a g o n a l BEKK model, which is:
H t = C · C + k = 1 K i = 1 p A k i · ( ϵ t i ϵ t i ) · A k i + k = 1 K j = 1 q B k j · H t j · B k j ,
where A k i and B k j are parameter matrices, C is a lower triangular matrix, p and q represent the lagged error term and the number of conditional covariance lags, respectively. K determines the generality of the process. We assume that p = q = 1 and K = 1 , so that the diagonal BEKK model can be written in a compact form as:
H t = C · C + A · ( ϵ t 1 ϵ t 1 ) · A + B · H t 1 · B ,
with A and B diagonal matrices. Positive definiteness of conditional covariance matrices is guaranteed by construction (see [23]). The procedure used in estimating the parameters of the model is the maximization of a likelihood function constructed under the assumption of an i.i.d. of the errors η t . Under the further assumption of conditional normality, the set of all parameters θ of the multivariate diagonal BEKK model can be estimated by maximizing the following sample log-likelihood function:
L ( θ ) = T N 2 log ( 2 π ) 1 2 t = 1 T log | H t | + ϵ t H t 1 ϵ t ,
with T the number of returns observations. Note that in the BEKK model, Equation (5) refers to the conditional covariance matrices. Thus, in the following modified log-likelihood function we have to use covariance matrices in place of the correlation matrices. Hence, under the assumption w i j = ρ i j , i j , given δ , we first compute matrix Z ^ and then the corresponding unconditional covariance matrix Σ ^ with respect to the whole sample period T, that is, Σ ^ = Γ Z ^ Γ , with Γ the diagonal matrix of standard deviations with respect to the whole sample period. Finally, in the Kullback–Leibler divergence measure of Formula (7) we compute the difference between matrices Σ ^ and H t . The modified log-likelihood function is:
L ( θ ) = T N 2 log ( 2 π ) 1 2 t = 1 T log | H t | + ϵ t H t 1 ϵ t K L ( Σ ^ , H t ) = T N 2 log ( 2 π ) 1 2 t = 1 T log | H t | + ϵ t H t 1 ϵ t 1 2 t = 1 T log ( | H t | | Σ ^ | ) + T r ( H t 1 Σ ^ ) N = 1 2 N ( T log ( 2 π ) ) + t = 1 T 2 log ( | H t | ) + ϵ t H t 1 ϵ t log ( | Σ ^ | ) + T r ( H t 1 Σ ^ ) .
Observe that the minimization of the distance between the target unconditional covariance matrix Σ ^ and H t forces the values σ i j , t to be as close as possible to the values σ i j , that is, forces the values ρ i j , t and i j to be as close as possible to the values ρ i j .

2.2. The DCC Model

The idea of this model, introduced and analyzed in [12], is that the conditional covariance matrix H t can be decomposed into the conditional standard deviations D t of each of the N series and a conditional correlation matrix of the returns R t . The dynamic of the model is described by Equation (1) and:
H t = D t R t D t .
The matrix D t is a diagonal matrix and consists of the N univariate GARCH models. Because it is a diagonal non-negative matrix with all diagonal elements positive, D t is positive definite. To ensure that H t is positive definite, it is necessary that the matrix R t is positive definite with the additional constraint that all its elements must be equal to or less than 1 by definition. The dynamic of correlation matrix R t is derived from another matrix Q t of the form:
R t = Q ¯ t 1 Q t Q ¯ t 1 ,
where Q ¯ t = d i a g ( Q t ) with d i a g ( · ) is the diagonal of a square matrix. The form of Q t determines the dynamic of the model and its complexity (see, e.g., [13,24]). For example, following [13] we have:
Q t = ( 1 θ 1 θ 2 ) Q ^ t + θ 1 ( ϵ t 1 ϵ t 1 ) + θ 2 Q t 1 ,
with Q ^ t = C o v [ ϵ t ϵ t ] = E [ ϵ t ϵ t ] . In order to ensure that R t is positive definite, the parameters θ 1 and θ 2 must satisfy:
θ 1 0 , θ 2 0 , θ 1 + θ 2 < 1 .
The parameter estimation phase is rather difficult, and hence for the DCC model a two-stage estimation procedure is provided. In the first stage, the parameters of the univariate GARCH models are estimated for each asset series. In the second stage, a second set of parameters are estimated given the parameters found in the previous phase. Referring to the dynamics described in (10) and assuming multivariate Gaussian distributed errors, after the first step only the parameters θ 1 and θ 2 are unknown so they are estimated in the second stage. In this second phase, the log-likelihood function is:
L ( θ ) = 1 2 t = 1 T log ( | R t | ) + ϵ t R t 1 ϵ t .
In our approach, we are interested in the second stage of the process where the log-likelihood function takes into account the correlation matrix of the assets returns at time t. Hence, under the assumption w i j = ρ i j , i j , given δ , we consider the matrix Z ^ and the matrix R t . The corresponding modified log-likelihood function is:
L ( θ ) = 1 2 t = 1 T log ( | R t | ) + ϵ t R t 1 ϵ t K L ( Z ^ , R t ) = 1 2 t = 1 T 2 log ( | R t | ) + ϵ t R t 1 ϵ t log ( | Z ^ | ) + T r ( R t 1 Z ^ ) N .
As for the modified BEKK model, to highlight the strong relationships among groups of assets in the market, the minimization of the distance between the (target) matrix Z ^ and R t forces the values ρ i j , t and i j to be as close as possible to the values ρ i j .
To conclude this section, we observe that if we set δ = 0 , and under the assumption that w i j = ρ i j , i j , the correlation matrix Z ^ and the unconditional covariance matrix Σ ^ correspond exactly to the long-run correlation and covariance matrices used in the model provided in [16]. Hence, on the one hand, our method can be precisely considered as an alternative targeting method with respect to the one proposed in [16]. On the other hand, assuming δ > 0 , with our method we are, in fact, “forcing” our modified MGARCH models to find clusters of highly correlated assets in the simulated series.

3. Data Analysis and Test

In order to demonstrate the effectiveness of our new estimation approach, which aims at correctly identifying highly (positive or negative) correlated groups of assets given a financial dataset, we first performed a preprocessing phase by a method of clustering financial time series (see [25]). Time series clustering creates groups based on similarities between time series without using pre-existing categorization labels [26]. Various approaches were proposed in time series clustering, and various authors attempted to synthesize the literature on the field (see, e.g., [27,28,29]). According to [29], there is a wide range of applications for time series clustering that can be found in a variety of domains (e.g., economic or medical time series analysis, visual processing, and anomalous activity detection).
Time series clustering is particularly relevant in finance. There is a particular interest in discovering typical dynamics of financial markets and the impact of different shocks on time series and portfolio allocation groups. In this respect, according to [30] clustering of financial time series is essential to determine how much wealth should be allocated to financial assets and opportunities. Therefore, financial time series need to be clustered to select an appropriate portfolio and analyze an economic system. When studying highly complicated phenomena such as financial time series, one has to deal with substantial heterogeneity and peculiar characteristics and features (see [31,32]). Therefore, robustness measures should address this challenge [33]. Moreover, studying their behavior over time in a dynamic framework is also essential, as these systems are associated with uncertainty and for this reason, other approaches representing this uncertainty were proposed in [34].
In this section, we provide a statistical framework for classifying time series and an example of applying the proposed method to a group of time series. Thus, our whole approach is based on two stages: In the first stage, we explicitly define the different groups of time series using a clustering procedure, which is helpful for identification of the different groups. In the second stage, we apply our new estimation method to correctly recognize the groups obtained so far. In our example, we perform the first stage on the following set of stocks: Facebook, Apple, Google, Boeing, Microsoft, Amazon, General Motors, Goldman Sachs, JPMorgan Chase, Intel, Verizon Communications, Visa, Cisco, Coca-Cola Company, and Salesforce. They were selected without any particular pattern from the stocks listed on the NYSE, the New York Stock Exchange. The period considered is from 1 January 2020 to 1 January 2021, and the resulting 253 observations refer to the daily closing prices of each financial time series.
Following [29], we have to clarify what is the claim in what two data are similar, and in what way we can obtain a good similarity and dissimilarity measure between two different observations (in this case, two different financial time series). These are the central questions of cluster analysis. In this regard, we have explicitly considered a hierarchical clustering algorithm that uses a Pearson correlation-based distance and the complete method (see [35]) as a helpful approach to distinguish the different groups that we can identify. Concerning financial time series, Pearson’s correlation coefficient is widely used in the related financial literature to quantify the degree of similarity or dissimilarity between two time series [30]. If we denote y t and x t as two different time series, we can consider their correlations [29]:
COR x t , y t = 1 t = 1 T y t x ¯ t y t y ¯ t t = 1 t x t x ¯ t 2 t = 1 t y t y ¯ t 2 ,
Then, we can compute the following distance, and we have:
d C o r x t , y t = 1 C O R x t , y t
We apply the distance and cluster approach to develop the dendrogram and interpret the different clusters (see Figure 1). Then, we follow an exploratory approach for identification of the groups of different stocks that are included in each cluster. We use the dendrogram to collect information about the variability of a cluster structure and the overall group’s structure.
No prior knowledge is required for this approach, and it allows for the understanding of patterns in financial time series data without the need for additional information [36]. This approach can be summarized as follows: use similarity measures to identify subgroups (see [37]) and determine which generic grouping each individual belongs to. Through the above clustering procedure, we can identify four general clusters (see Figure 1). First, we can identify Microsoft’s most significant joint movement (MSFT) and Amazon (AMZN). At the same time, Facebook (FB), Apple (AAPL), and finally, Salesforce (CRM) belong to the same cluster. This cluster is clearly based on the technological innovations that can have an overarching impact on different sectors and companies (see also [34]). Finally, at the same time, Coca Cola (KO) and, more importantly, Boeing (BA) and JPMorgan Chase (JPM) are in a different cluster. Furthermore, Intel (INTC) and Cisco (CSCO) can also be considered part of the same cluster. We can further observe the critical role of economic similarities for these two stocks in this case. In this respect, it is possible to suppose relevant economic shocks that affect both financial time series. Finally, in the third cluster, we can observe a slightly different behavior for Verizon Communications (VZ), but also daily solid joint movements for Google (GOOG) and Visa (V) as well as for General Motors (GM) and Goldman Sachs (GS). Overall, through observation and visual analysis of the dendrogram, the distinct clusters obtained will be used as a reference for the evaluation of the effectiveness of our new estimation approach (second stage).
In the following, we present some scenarios characterized by a different number N of time series considered with the aim of evaluating the differences between the simulated series obtained by the BEKK and DCC models and their modified versions. All MGARCH models have been implemented in Matlab R2018 version using the MFE Toolbox code repository by Kevin Sheppard [38]. In the resulting correlation graphs, we detected all the maximal cliques using the Bron–Kerbosch algorithm [39]. The experiments were conducted on a PC equipped with an Intel Core i7-3632MQ processor with 2.20 Ghz.
  • Scenario 1
In this first scenario, we consider only the group of technological innovation assets formed by MSFT, AMZN, CRM, FB, and AAPL (see Figure 1). We compute the 252 log-return values and the corresponding unconditional correlation matrix, which is (see Table 1):
Let us now consider the (sub)graph G ( δ ) of the complete correlation graph G whose vertex set contains only the subset of pairs of assets such that | ρ i j | > δ , with δ = 0.5 (see Figure 2). In the figure, the numbers associated with the vertices correspond to the following labeling: 1, MSFT; 2, AMZN; 3, CRM; 4, FB; and 5, AAPL.
Note that the graph G ( δ ) corresponds to the complete graph G, because for all pairs of assets i and j we have ρ i j > 0.5 .
Considering the whole original dataset consisting of T = 252 return observations, we first estimate the parameters of the BEKK and DCC models and then the parameters of each of the modified versions of the two models using the log-likelihood functions (7) and (12), respectively. Then, using these sets of parameters we simulate (new) T = 252 log-return observations of the five assets. Our aim is to verify whether our modified models allow to better capture the strong relationships between assets in the stock market. In other words, we want to compare the graph G ( δ ) with the simulated graphs G S ( δ ) (see Figure 3).
Comparing the five graphs in Figure 2 and Figure 3, only the original BEKK model was not able to correctly reproduce the original series. In this case, vertex 5, corresponding to asset AAPL, is not connected to all the other vertices despite its correlation value with other assets is greater that 0.5 . Observe that for the DCC approach the modification brings no gain. This can occur when the series’ dimension is small, and, in this specific case, the modification has no effect. In addition to the graphs G S ( δ ) , in Table 2 we report the Frobenius distance between the fitted covariance matrices H t with respect to the unconditional covariance matrix Σ ^ , and the Kullback–Leibler divergence between the matrix Σ ^ and the unconditional covariance matrix of the new simulated series of log-returns, denoted by Σ ^ S , with respect to all the considered models. The Frobenius norm is:
F = t = 1 T T r [ ( H t Σ ^ ) ( H t Σ ^ ) ] ,
and the Kullback–Leibler divergence is:
K L = 1 2 log ( | Σ ^ S | | Σ ^ | ) + T r ( Σ ^ S 1 Σ ^ ) N .
Both functions F and K L measure loss, so that lower values are preferable. We note that the values of the modified MGARCH models are better than the corresponding original models. Here, we are not interested in comparing the values of the two loss functions between all the models, as, again, it is not our goal to find a single winner.
Assume now that we are interested in detecting groups of assets with a higher value of correlation. On the basis of the values in Table 1, we set δ = 0.71 . The corresponding graph G ( δ ) is (see Figure 4):
We observe that with this value of δ , only assets { 1 , 2 , 3 } form a clique. The goal is to find simulated graphs G S ( δ ) , with δ = 0.71 , as similar as possible to the subgraph G ( δ ) of Figure 4. The results are reported in Figure 5 along with the corresponding table (see Table 3) reporting the values of the two loss functions considered.
The simulated graph of the modified BEKK model exactly replicates G ( δ ) . The two graphs associated with the DCC and modified DCC models correctly find the clique C = { 1 , 2 , 3 } , but the DCC model determines a new clique C = { 1 , 2 , 3 , 4 } that contains C, whereas the modified DCC model introduces the additional edge ( 4 , 2 ) . Thus, these two models “overestimate” the correlation between assets pointing out correlations that do not really exist. The original BEKK model has the lowest ability to correctly identify groups of highly correlated assets. On the other hand, the values of the loss functions reported in Table 3 are in line with the values in Table 2.
  • Scenario 2
In this second scenario, we consider N = 8 assets along with the corresponding correlation matrix reported in Table 4. The set of stocks is formed by the five previously considered assets and the group formed by KO, BA, and JPM (see Figure 1). In the corresponding graphs, the numbers associated with the vertices now correspond to the following labeling: 1, MSFT; 2, AMZN; 3, CRM; 4, FB; 5, AAPL; 6, KO; 7, BA; 8, JPM. We choose this new group of three stocks because the two clusters are adequately separated from each other, so that we expect finding two well-separated groups of assets in the following correlation graphs. The unconditional correlation matrix is:
Consider δ = 0.5 , the graph G ( δ ) is (see Figure 6):
In G ( δ ) , we clearly detect the three maximal cliques C 1 = { 1 , 2 , 3 , 4 , 5 } ; C 2 = { 6 , 7 , 8 } ; and C 3 = { 2 , 3 , 7 , 8 } . In particular, cliques C 1 and C 2 refer to the two clusters of assets in Figure 1. The simulated graphs G S ( δ ) are presented in Figure 7.
In Table 5, we report the values of the loss functions F and K L and the additional information on the number of maximal cliques in each of the simulated graphs G S ( δ ) .
The graphs G S ( δ ) related to the two modified MGARCH models are exactly the same as the graph G ( δ ) . It is worth noting that this does not mean that the correlation matrices of the simulated returns series with the modified BEKK and DCC models are equal to the correlation matrix of the original series, but that both the modified models are able to detect the same (original) clusters of highly correlated assets. In actuality, assuming highly (positive) correlated assets remain unaltered over time, by forecasting future values of the series using the results provided by the modified BEKK and/or DCC models, one might expect to simulate returns series that do not alter too much the correlation structure among assets as well as the strong relationships among stocks. This can be useful in a portfolio selection problem from a diversification viewpoint. It is well-known that correlation represents the degree of relationship between the price movements of different assets included in the portfolio for diversified portfolios. Thus, choosing pairs of assets less correlated decreases the portfolio’s overall risk. Consider, for example, the graph G S ( δ ) in Figure 7c referred to the original DCC model, which is very similar to the graph G ( δ ) (in fact there is just one additional edge in G S ( δ ) , that is, edge ( 2 , 6 ) ). It includes the maximal clique C = { 2 , 6 , 7 , 8 } , demonstrating that asset 2 now shows a correlation value with asset 6 greater than 0.5 (precisely, equal to 0.536 ), whereas from Table 4 we have ρ 26 = 0.4810 . Hence, the original DCC model inserts a strong correlation between these two assets, which is not truthful, and, without reason, this might prevent from choosing assets 2 and 6 in an optimal portfolio. Finally, note that these considerations are difficult to obtain by examination of only the values of the loss functions, which demonstrates that the modified models behave better than the original models.
Finally, observe that the graph G S ( δ ) , related to the original BEKK model (see Figure 7a), reports a very distorted relationship between the assets in the market. Hence, on the basis of the two above scenarios, we can state that it seems that the BEKK model benefits more effectively from the introduction of the modified log-likelihood function.
  • Scenario 3
In this third scenario, we consider all N = 15 assets described in Section 3. In the graphs, the new labels associated with each asset are (see Figure 1): 1, MSFT; 2, AMZN; 3, CRM; 4, FB; 5, AAPL; 6, VZ; 7, GOOG; 8, V; 9, GM; 10, GS; 11, KO; 12, BA; 13, JPM; 14, INTC; 15, CSCO. The corresponding correlation matrix is in Table 6:
The correlation graph for δ = 0.5 is depicted in Figure 8.
The maximal cliques in G ( δ ) with δ = 0.5 are: C 1 = { 1 , 2 , 3 , 4 , 5 , 6 , 14 } ; C 2 = { 1 , 2 , 3 , 5 , 6 , 8 , 9 , 14 , 15 } ; C 3 = { 2 , 3 , 6 , 8 , 9 , 10 , 12 , 13 , 14 , 15 } ; and C 4 = { 6 , 7 , 8 , 9 , 11 , 12 , 13 , 15 } . These cliques show the most interconnected groups of assets at a correlation level greater than 0.5 . In the following, we report a table of the values of the loss functions F and K L and the information on the maximal cliques in each of the simulated graphs G S ( δ ) . More precisely, in Table 7 we report in bold the cliques of the simulated graphs that coincide to the ones of the original graph G ( δ ) . The modified BEKK model seems to be the best performing both with respect to the loss function values and the number of the maximal cliques of G ( δ ) . We also note that the first clique in the graph, which refers to the modified BEKK model, has the same number of nodes of clique C 1 except with vertex 15 in place of vertex 14. We observe that the correlation values of the first six assets in Table 6 with respect to vertices 14 and 15 are very close to each other. Thus, the difference in those two cliques might be due to numerical reasons in the estimation and simulation phase of the modified BEKK model. In any case, all models introduce correlation values, and a larger number of quite different cliques, which do not appear in the graph G ( δ ) .
To evaluate the effectiveness of our method in finding groups of strongly correlated assets we consider a higher value for the threshold, that is, δ = 0.65 . The corresponding correlation graph G ( δ ) is in Figure 9:
As expected, the graph G ( δ ) , with δ = 0.65 , is sparser than the graph G ( δ ) , with δ = 0.5 . In the new graph, we identify a larger number of cliques, but of cardinality less than the cardinality of the maximal cliques when δ = 0.5 .
In Table 8, we report in bold the cliques of the simulated graphs that coincide to those in G ( δ ) with δ = 0.65 . It is evident that when N grows, the results are less clear-cut with respect to the two previous examples. On the one hand, this can be due to the fact that the log-likelihood functions to optimize are highly nonlinear functions, and it is difficult to find provably optimal solutions. Local solutions may be problematic, and this creates difficulties in the estimation of the models. Additionally, because in the optimization phase matrices have to be inverted in each iteration, it makes the overall computation demanding unless N is small. In any case, Table 7 and Table 8 still highlight that the modified MGARCH models performs slightly better than the original ones, in particular, when δ increases. For example, for δ = 0.65 the simulated series obtained with the modified BEKK and DCC models are still able to better capture the strong relationships among assets in the market than the corresponding original models.
From the above results, we can state that the application of our method to empirical case studies is encouraging, particularly when the number of assets is not large and we are able to find (global) optimal solutions in the maximization of the log-likelihood objective functions. In fact, in these cases, we observed a significant improvement in the ability of the modified MGARCH models to replicate the correlation structure of the assets in the market compared with the original models. When the number of assets increases, the estimation of the models involves somewhat heavy computations because they contain a large number of parameters, and there is no guarantee of finding provably global optimal solutions but only local optimal solutions. This undermines the ability of the estimated models to replicate the correlation structure of the assets correctly. Despite this, even when the number of assets is large, the modified MGARCH models seem to perform better than the original ones. Overall, our experiments demonstrate that the BEKK model benefits most from the modification of the log-likelihood function.

4. Conclusions and Further Research

In this paper, we advance a method for improving the estimation phase of financial time series with the aim of improving the analysis and evaluation of portfolios of financial assets whose performance strictly depends on the correlation among assets. Several different estimation models have been proposed in the literature; among others, the family of models known as multivariate general auto regressive conditional heteroscedasticity (MGARCH) models are the most widely used. These models consider that a financial time series suffers from heteroscedasticity. This paper considers two such models, namely the BEKK and DCC models, where we modified the log-likelihood objective function assuming that there are specific groups of assets that are highly correlated in a financial market, and these relationships remain unaltered over time. Hence, in the log-likelihood function we introduced a term referring to a loss measure computed on the difference between the time-varying covariance/correlation matrices and the covariance/correlation matrix estimated with respect to the whole in-sample period. Given the set of the estimated parameters, we use them for simulating new time series to evaluate the effectiveness of our modified estimation phase. We also propose a new approach for the evaluation of the results based on network analysis and, more precisely, on detecting maximal cliques in correlation graphs.
On the basis of the results reported in Section 3, we cannot state whether the modified models always outperform the original ones, and much more experiments are needed. Therefore, this leaves plenty of opportunity for further research. On the one hand, one can experiment with other loss functions to introduce in the log-likelihood objective functions to improve the estimates. In fact, we observe that our approach is extremely flexible. Different loss functions can be considered without imposing additional constraints on covariance or correlation matrices. On the other hand, much attention should be paid in order to improve the optimization phase. This involves developing ad hoc strategies for finding global optimal solutions or solutions close to the optimal ones. In this regard, the development of metaheuristic procedures could be a further line of research to improve the optimization phase. Finally, finding new network indexes that better highlight other peculiar aspects of a financial market is worth investigating.

Author Contributions

Methodology, C.D. and A.S.; Software, A.S.; Data curation, C.D.; Writing—original draft, A.S.; Funding acquisition, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

The research of the second author was partially supported by the research program supported by Junta de Andalucia, Consejeria de Economia y Conocimiento n. US-1256951. The authors have no financial or proprietary interests in any material discussed in this article.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Silvennoinen, A.; Terasvirta, T. Consistency and asymptotic normality of maximum likelihood estimators of a multiplicative time-varying smooth transition correlation GARCH model. Econom. Stat. 2021, in press. [CrossRef]
  2. Engle, R.; Colacito, R. Testing and valuing dynamic correlations for asset allocation. J. Bus. Econ. Stat. 2006, 24, 238–253. [Google Scholar] [CrossRef]
  3. Hsiao, C.W.; Chan, Y.C.; Lee, M.Y.; Lu, H.P. Heteroscedasticity and Precise Estimation Model Approach for Complex Financial Time-Series Data: An Example of Taiwan Stock Index Futures before and during COVID-19. Mathematics 2021, 9, 2719. [Google Scholar] [CrossRef]
  4. Lei, B.; Zhang, B.; Song, Y. Volatility forecasting for high-frequency financial data based on web search index and deep learning model. Mathematics 2021, 9, 320. [Google Scholar] [CrossRef]
  5. Posedel Simovic, P.; Tafro, A. Pricing the Volatility Risk Premium with a Discrete Stochastic Volatility Model. Mathematics 2021, 9, 2038. [Google Scholar] [CrossRef]
  6. Rombouts, J.; Stentoft, L.; Violante, F. The value of multivariate model sophistication: An application to pricing Dow Jones Industrial Average options. Int. J. Forecast. 2014, 30, 78–98. [Google Scholar] [CrossRef] [Green Version]
  7. Gargallo, P.; Lample, L.; Miguel, J.A.; Salvador, M. Co-movements between EU ETS and the energy markets: A VAR-DCC-GARCH approach. Mathematics 2021, 9, 1787. [Google Scholar] [CrossRef]
  8. Billio, M.; Caporin, M. Market linkages, variance spillovers, and correlation stability: Empirical evidence of financial contagion. Comput. Stat. Data Anal. 2010, 54, 2443–2458. [Google Scholar] [CrossRef]
  9. Chang, C.L.; Liu, C.P.; McAleer, M. Volatility spillovers for spot, futures, and ETF prices in agriculture and energy. Energy Econ. 2019, 81, 779–792. [Google Scholar] [CrossRef]
  10. Mata, M.N.; Razali, M.N.; Bentes, S.R.; Vieira, I. Volatility Spillovers Effect of Pan Asia’s Property Portfolio Markets. Mathematics 2021, 9, 1418. [Google Scholar] [CrossRef]
  11. Bauwens, L.; Laurent, S.; Rombouts, J.V. Multivariate GARCH models: A survey. J. Appl. Econom. 2006, 21, 79–109. [Google Scholar] [CrossRef] [Green Version]
  12. Engle, R.F.; Sheppard, K. Theoretical and Empirical Properties of Dynamic Conditional Correlation Multivariate GARCH; NBER Working Papers, No. 8554; 2001, National Bureau of Economic Research (NBER.org). Available online: https://www.nber.org/papers/w8554 (accessed on 4 February 2022).
  13. Engle, R.F. Dynamic conditional correlation: A simple class of multivariate generalized autoregressive conditional heteroskedasticity models. J. Bus. Econ. Stat. 2002, 20, 339–350. [Google Scholar] [CrossRef]
  14. Bollerslev, T.; Patton, A.J.; Quaedvlieg, R. Modeling and forecasting (un)reliable realized covariances for more reliable financial decisions. J. Econom. 2018, 207, 71–91. [Google Scholar] [CrossRef] [Green Version]
  15. Engle, R.; Kelly, B. Dynamic equicorrelation. J. Bus. Econ. Stat. 2012, 30, 212–228. [Google Scholar] [CrossRef]
  16. Caporin, M.; McAleer, M. Do we really need both BEKK and DCC? A tale of two multivariate GARCH models. J. Econ. Surv. 2012, 26, 736–751. [Google Scholar] [CrossRef] [Green Version]
  17. Aielli, G.P. Dynamic Conditional Correlations: On Properties and Estimation. 2009. Available online: http://ssrn.com/abstract=1507743 (accessed on 2 February 2022).
  18. Stoyanov, S.V.; Rachev, S.T.; Racheva-Yotova, B.; Fabozzi, F.J. Fat-tailed models for risk estimation. J. Portf. Manag. 2011, 37, 107–117. [Google Scholar] [CrossRef] [Green Version]
  19. Mantegna, R.N. Hierarchical structure in financial markets. Eur. Phys. J. B-Condens. Matter Complex Syst. 1999, 11, 193–197. [Google Scholar] [CrossRef] [Green Version]
  20. Bollerslev, T.; Wooldridge, J.M. Quasi-maximum likelihood estimation and inference in dynamic models with time-varying covariances. Econom. Rev. 1992, 11, 143–172. [Google Scholar] [CrossRef]
  21. Bondy, A.; Murty, U.S.R. Graph Theory; Springer: New York, NY, USA, 2008. [Google Scholar]
  22. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  23. Ding, Z.; Engle, R. Large scale conditional covariance modelling, estimation and testing. Academia Econ. Pap. 2001, 29, 157–184. [Google Scholar]
  24. Tse, Y.K.; Tsui, A.K.C. A multivariate generalized autoregressive conditional heteroscedasticity model with time-varying correlations. J. Bus. Econ. Stat. 2002, 20, 351–362. [Google Scholar] [CrossRef]
  25. Sarda-Espinosa, A. Comparing time-series clustering algorithms in r using the dtwclust package. R Package Vignette 2017, 12, 41. [Google Scholar]
  26. Radhimeenakshi, D.S.; Latha, K. Similarity Measure Selection for Clustering Stock Market Time Series Databases. Int. J. Eng. Sci. Comput. 2017, 7, 13879–13882. [Google Scholar]
  27. Aghabozorgi, S.; Shirkhorshidi, A.S.; Wah, T.Y. Time-series clustering—A decade review. Inf. Syst. 2015, 53, 16–38. [Google Scholar] [CrossRef]
  28. Liao, T.W. Clustering of time series data—A survey. Pattern Recognit. 2005, 38, 1857–1874. [Google Scholar] [CrossRef]
  29. Montero, P.; Vilar, J.A. TSclust: An R Package for Time Series Clustering. J. Stat. Softw. 2014, 62, 1–43. [Google Scholar] [CrossRef] [Green Version]
  30. Piccardi, C.; Calatroni, L.; Bertoni, F. Clustering financial time series by network community analysis. Int. J. Mod. Phys. C 2011, 22, 35–50. [Google Scholar] [CrossRef]
  31. Sewell, M. Characterization of Financial Time Series; Rn, 11(01), 01. Working Paper; 2011, UCL Department of Computer Science, University College London. Available online: http://www.cs.ucl.ac.uk/fileadmin/UCL-CS/images/Research_Student_Information/RN_11_01.pdf (accessed on 2 December 2021).
  32. Shi, Y.; Li, B.; Du, G.; Dai, W. Clustering framework based on multi-scale analysis of intraday financial time series. Phys. A Stat. Mech. Appl. 2021, 567, 125728. [Google Scholar] [CrossRef]
  33. D’Urso, P.; De Giovanni, L.; Massari, R. GARCH-based robust clustering of time series. Fuzzy Sets Syst. 2016, 305, 1–28. [Google Scholar] [CrossRef] [Green Version]
  34. Drago, C.; Lauro, C.; Scepi, G. Visualization and Analysis of Large Datasets by Beanplot PCA. In Advances in Latent Variables; Carpita, M., Brentari, E., Qannari, E.M., Eds.; Vita e Pensiero; Università Cattolica del S. Cuore: Milano, Italy, 2013; Available online: https://www.researchgate.net/profile/Carlo-Drago (accessed on 30 June 2013).
  35. Everitt, B.S.; Landau, S.; Leese, M. Cluster Analysis; Arnold: London, UK, 2001; ISBN 0-340-76119-9. [Google Scholar]
  36. Ahn, J.; Lee, J.H. Clustering Method for Financial Time Series with Co-Movement Relationship. In Proceedings of the 2018 3rd International Conference on Computational Intelligence and Applications (ICCIA), Hong Kong, China, 28–30 July 2018; pp. 260–264. [Google Scholar]
  37. Wang, X.; Smith, K.; Hyndman, R. Characteristic-based clustering for time series data. Data Min. Knowl. Discov. 2006, 13, 335–364. [Google Scholar] [CrossRef]
  38. Sheppard, K. MFE Toolbox. 2013. Available online: https://www.kevinsheppard.com/MFE_Toolbox (accessed on 16 August 2018).
  39. Bron, C.; Kerbosch, J. Algorithm 457: Finding all cliques of an undirected graph. Commun. ACM 1973, 16, 575–577. [Google Scholar] [CrossRef]
Figure 1. Clusters obtained in the first phase.
Figure 1. Clusters obtained in the first phase.
Mathematics 11 00382 g001
Figure 2. The graph G ( δ ) with δ = 0.5 . Vertex number refers to the associated asset.
Figure 2. The graph G ( δ ) with δ = 0.5 . Vertex number refers to the associated asset.
Mathematics 11 00382 g002
Figure 3. The simulated graphs G S ( δ ) with δ = 0.5 . (a) The graph G S ( δ ) resulting from the BEKK model; (b) the graph G S ( δ ) resulting from the modified BEKK model; (c) the graph G S ( δ ) resulting from the DCC model; (d) the graph G S ( δ ) resulting from the modified DCC model. Vertex number refers to the associated asset.
Figure 3. The simulated graphs G S ( δ ) with δ = 0.5 . (a) The graph G S ( δ ) resulting from the BEKK model; (b) the graph G S ( δ ) resulting from the modified BEKK model; (c) the graph G S ( δ ) resulting from the DCC model; (d) the graph G S ( δ ) resulting from the modified DCC model. Vertex number refers to the associated asset.
Mathematics 11 00382 g003
Figure 4. The graph G ( δ ) with δ = 0.71 . Vertex number refers to the associated asset.
Figure 4. The graph G ( δ ) with δ = 0.71 . Vertex number refers to the associated asset.
Mathematics 11 00382 g004
Figure 5. The simulated graphs G S ( δ ) with δ = 0.71 . (a) The graph G S ( δ ) resulting from the BEKK model; (b) the graph G S ( δ ) resulting from the modified BEKK model; (c) the graph G S ( δ ) resulting from the DCC model; (d) the graph G S ( δ ) resulting from the modified DCC model. Vertex number refers to the associated asset.
Figure 5. The simulated graphs G S ( δ ) with δ = 0.71 . (a) The graph G S ( δ ) resulting from the BEKK model; (b) the graph G S ( δ ) resulting from the modified BEKK model; (c) the graph G S ( δ ) resulting from the DCC model; (d) the graph G S ( δ ) resulting from the modified DCC model. Vertex number refers to the associated asset.
Mathematics 11 00382 g005
Figure 6. The graph G ( δ ) with δ = 0.5 for N = 8 assets. Vertex number refers to the associated asset.
Figure 6. The graph G ( δ ) with δ = 0.5 for N = 8 assets. Vertex number refers to the associated asset.
Mathematics 11 00382 g006
Figure 7. The simulated graphs G S ( δ ) with δ = 0.5 for N = 8 assets. (a) The graph G S ( δ ) resulting from the BEKK model; (b) the graph G S ( δ ) resulting from the modified BEKK model; (c) the graph G S ( δ ) resulting from the DCC model; (d) the graph G S ( δ ) resulting from the modified DCC model. Vertex number refers to the associated asset.
Figure 7. The simulated graphs G S ( δ ) with δ = 0.5 for N = 8 assets. (a) The graph G S ( δ ) resulting from the BEKK model; (b) the graph G S ( δ ) resulting from the modified BEKK model; (c) the graph G S ( δ ) resulting from the DCC model; (d) the graph G S ( δ ) resulting from the modified DCC model. Vertex number refers to the associated asset.
Mathematics 11 00382 g007
Figure 8. The graph G ( δ ) with δ = 0.5 for N = 15 assets. Vertex number refers to the associated asset.
Figure 8. The graph G ( δ ) with δ = 0.5 for N = 15 assets. Vertex number refers to the associated asset.
Mathematics 11 00382 g008
Figure 9. The graph G ( δ ) with δ = 0.65 for N = 15 assets. Vertex number refers to the associated asset.
Figure 9. The graph G ( δ ) with δ = 0.65 for N = 15 assets. Vertex number refers to the associated asset.
Mathematics 11 00382 g009
Table 1. The correlation matrix for N = 5 assets.
Table 1. The correlation matrix for N = 5 assets.
MSFTAMZNCRMFBAAPL
MSFT10.77120.76900.68550.6914
AMZN0.771210.84350.70280.6458
CRM0.76900.843510.74130.7508
FB0.68550.70280.741310.6121
AAPL0.69140.64580.75080.61211
Table 2. Values of the Frobenius and the Kullback–Leibler loss functions: case N = 5 and δ = 0.5 .
Table 2. Values of the Frobenius and the Kullback–Leibler loss functions: case N = 5 and δ = 0.5 .
BEKKModified BEKKDCCModified DCC
F0.06690.05550.08810.0870
KL0.08550.02290.01620.0154
Table 3. Values of the Frobenius and the Kullback–Leibler loss functions: case N = 5 and δ = 0.71 .
Table 3. Values of the Frobenius and the Kullback–Leibler loss functions: case N = 5 and δ = 0.71 .
BEKKModified BEKKDCCModified DCC
F0.06690.05750.08810.0840
KL0.09760.02690.03060.0262
Table 4. The correlation matrix for N = 8 assets.
Table 4. The correlation matrix for N = 8 assets.
MSFTAMZNCRMFBAAPLKOBAJPM
MSFT10.77120.76900.68550.69140.40380.48340.4387
AMZN0.771210.84350.70280.64580.48100.54120.5580
CRM0.76900.843510.74130.75080.47980.59120.5838
FB0.68550.70280.741310.61210.24510.26970.3162
AAPL0.69140.64580.75080.612110.39950.45620.4472
KO0.40380.48100.47980.24510.399510.72110.6461
BA0.48340.54120.59120.26970.45620.721110.7304
JPM0.43870.55800.58380.31620.44720.64610.73041
Table 5. Values of the Frobenius and the Kullback–Leibler loss functions and the maximal cliques: case N = 8 and δ = 0.5 .
Table 5. Values of the Frobenius and the Kullback–Leibler loss functions and the maximal cliques: case N = 8 and δ = 0.5 .
BEKKModified BEKKDCCModified DCC
F0.06690.05750.08810.0840
KL0.09760.02690.03060.0262
Maximal Cliques{1,2,3,4}, {1,3,5}, {6,7}, {7,8}{1,2,3,4,5}, {2,3,7,8}, {6,7,8}{1,2,3,4,5}, {2,3,7,8}, {2,6,7,8}{1,2,3,4,5}, {2,3,7,8}, {6,7,8}
Table 6. The correlation matrix for N = 15 assets.
Table 6. The correlation matrix for N = 15 assets.
MSFTAMZNCRMFBAAPLVZGOOGVGMGSKOBAJPMINTCCSCO
MSFT10.77120.76900.68550.69140.80670.46710.56190.66560.45660.40380.48340.43870.54930.5777
AMZN0.771210.84350.70280.64580.75810.45390.62250.71230.55230.48100.54120.55800.63330.6596
CRM0.76900.843510.74130.75080.85580.48490.64840.78300.60800.47980.59120.58380.71870.7171
FB0.68550.70280.741310.61210.68370.22880.38250.45190.39550.24510.26970.31620.50650.4991
AAPL0.69140.64580.75080.612110.68460.41910.51180.63530.41430.39950.45620.44720.51700.5358
VZ0.80670.75810.85580.68370.684610.53790.66440.79030.58770.51630.60200.61720.65350.6838
GOOG0.46710.45390.48490.22880.41910.537910.74880.70140.41350.69720.75900.58810.49130.5131
V0.56190.62250.64840.38250.51180.66440.748810.76040.58990.69630.89070.70010.59810.6526
GM0.66560.71230.78300.45190.63530.79030.70140.760410.62800.65410.77730.73630.64040.7295
GS0.45660.55230.60800.39550.41430.58770.41350.58990.628010.43660.59750.71330.56980.6544
KO0.40380.48100.47980.24510.39950.51630.69720.69630.65410.436610.72110.64610.49080.5061
BA0.48340.54120.59120.26970.45620.60200.75900.89070.77730.59750.721110.73040.56630.6145
JPM0.43870.55800.58380.31620.44720.61720.58810.70010.73630.71330.64610.730410.55420.5984
INTC0.54930.63330.71870.50650.51700.65350.49130.59810.64040.56980.49080.56630.554210.6700
CSCO0.57770.65960.71710.49910.53580.68380.51310.65260.72950.65440.50610.61450.59840.67001
Table 7. Values of the Frobenius and the Kullback–Leibler loss functions and the maximal cliques: case N = 15 and δ = 0.5 . In bold the cliques of the simulated graphs that coincide to the ones of the original graph G ( δ ) .
Table 7. Values of the Frobenius and the Kullback–Leibler loss functions and the maximal cliques: case N = 15 and δ = 0.5 . In bold the cliques of the simulated graphs that coincide to the ones of the original graph G ( δ ) .
BEKKModified BEKKDCCModified DCC
F0.17080.16820.19950.2028
KL0.40350.30020.33090.3091
Maximal Cliques{1,2,3,4,5,6},
{7,8,9,11,12,13},
{1,2,3,5,6,8,9,14,15},
{1,2,3,6,8,9,10,14,15},
{2,3,6,8,9,10,12,13,14,15},
{7,8,9,12,13,14,15}
{1,2,3,4,5,6,15},
{1,2,3,5,6,8,9,14,15},
{2,3,6,8,9,10,12,13,14,15},
{6,7,8,9,11,12,13,15},
{6,8,9,11,12,13,14,15}
{1,2,3,4,5,6}, {1,2,3,5,6,8,9},
{2,3,6,8,9,10,15},
{3,6,8,9,10,13,15},
{3,8,9,10,12,13,15},
{3,8,9,10,12,14,15},
{7,8,9,11,12,13},
{2,8,9,11,14,15},
{8,9,11,12,13,15},
{8,9,11,12,14,15}
{1,2,3,4,6},
{1,3,4,5,6}, {1,3,5,6,9},
{1,2,3,6,8,9,10,13,14,15},
{1,3,6,8,9,10,12,13,14,15},
{1,6,7,8,9,12,13},
{7,8,9,11,12,13}
Table 8. Values of the Frobenius and Kullback–Leibler loss functions and the maximal cliques: case N = 15 and δ = 0.65 . In bold the cliques of the simulated graphs that coincide to the ones of the original graph G ( δ ) .
Table 8. Values of the Frobenius and Kullback–Leibler loss functions and the maximal cliques: case N = 15 and δ = 0.65 . In bold the cliques of the simulated graphs that coincide to the ones of the original graph G ( δ ) .
BEKKModified BEKKDCCModified DCC
F0.17080.16590.19950.2038
KL0.37210.29820.26270.2433
Maximal Cliques{1,2,3,4,6}, {1,2,3,5,6},
{2,3,5,6,9}, {2,3,6,9,14},
{3,6,9,14,15}, {7,8,9,11,12},
{8,9,11,12,13}, {6,8,9},
{9,10,13}
{1,2,3,4,6}, {1,2,3,6,9},
{1,2,3,5,6}, {2,3,6,9,14,15},
{3,6,8,9,14,15},
{7,8,9,11,12},
{8,9,11,12,13}, {8,9,12,14},
{9,10,12,13}, {9,10,15}
{1,2,3,4,6}, {2,3,6,9},
{3,5}, {3,14}, {7,8,9,12},
{7,8,11,12}, {8,9,12,13},
{10,13}
{1,2,3,4,6}, {1,2,3,5,6},
{1,2,3,6,9}, {7,8,9,12},
{8,9,12,13}, {10,13},
{7,8,11,12}, {3,15}
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Drago, C.; Scozzari, A. A Network-Based Analysis for Evaluating Conditional Covariance Estimates. Mathematics 2023, 11, 382. https://doi.org/10.3390/math11020382

AMA Style

Drago C, Scozzari A. A Network-Based Analysis for Evaluating Conditional Covariance Estimates. Mathematics. 2023; 11(2):382. https://doi.org/10.3390/math11020382

Chicago/Turabian Style

Drago, Carlo, and Andrea Scozzari. 2023. "A Network-Based Analysis for Evaluating Conditional Covariance Estimates" Mathematics 11, no. 2: 382. https://doi.org/10.3390/math11020382

APA Style

Drago, C., & Scozzari, A. (2023). A Network-Based Analysis for Evaluating Conditional Covariance Estimates. Mathematics, 11(2), 382. https://doi.org/10.3390/math11020382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop