Next Article in Journal
Invariance of the Mathematical Expectation of a Random Quantity and Its Consequences
Previous Article in Journal
A Hybrid Model for Forecasting Realized Volatility Based on Heterogeneous Autoregressive Model and Support Vector Regression
Previous Article in Special Issue
Equity Price Dynamics under Shocks: In Distress or Short Squeeze
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multivariate Spectral Backtests of Forecast Distributions under Unknown Dependencies

by
Janine Balter
1,† and
Alexander J. McNeil
2,*
1
Deutsche Bundesbank, 40212 Düsseldorf, Germany
2
School for Business and Society, University of York, York YO10 5DD, UK
*
Author to whom correspondence should be addressed.
The opinions expressed here are our own and do not reflect the views of the Deutsche Bundesbank or its staff.
Risks 2024, 12(1), 13; https://doi.org/10.3390/risks12010013
Submission received: 19 December 2023 / Revised: 5 January 2024 / Accepted: 12 January 2024 / Published: 17 January 2024
(This article belongs to the Special Issue Applied Financial and Actuarial Risk Analytics)

Abstract

:
Under the revised market risk framework of the Basel Committee on Banking Supervision, the model validation regime for internal models now requires that models capture the tail risk in profit-and-loss (P&L) distributions at the trading desk level. We develop multi-desk backtests, which simultaneously test all trading desk models and which exploit all the information available in the presence of an unknown correlation structure between desks. We propose a multi-desk extension of the spectral test of Gordy and McNeil, which allows the evaluation of a model at more than one confidence level and contains a multi-desk value-at-risk (VaR) backtest as a special case. The spectral tests make use of realised probability integral transform values based on estimated P&L distributions for each desk and are more informative and more powerful than simpler tests based on VaR violation indicators. The new backtests are easy to implement with a reasonable running time; in a series of simulation studies, we show that they have good size and power properties.

1. Introduction

In this paper, we propose a multi-desk extension of the spectral backtest of Gordy and McNeil (2020). Our test is designed to address the requirement that banks should implement backtests of their risk models at the desk level. Although banks are required to report the results of single-desk value-at-risk backtests at the desk level, our proposal is for a test that simultaneously assesses the quality of all desk-level risk models, with respect to outcomes in the tail.
There are a number of potential advantages of our test. First, it may detect deficiencies in desk models that are missed when trading desk data are aggregated into a single portfolio. Aggregation will result in the loss of information and the netting effect across desks will potentially mask situations where an under-estimation of risk in one trading desk model is compensated for by the over-estimation of risk in another trading desk model. Moreover, when desks are tested jointly, we effectively increase the amount of data with respect to a single test at the level of the trading book, which improves the testing power. Furthermore, by embedding our test in the spectral framework, we are able to test the performance of risk models at a range of confidence levels, and not simply at a single level, as in standard VaR backtesting. To place our proposal in context, we give a brief review of the development of backtesting for the trading book in the following paragraphs.
Over the last three decades, a method of testing value-at-risk (VaR) predictions using historical data, known as backtesting (Jorion 2007), has become the industry standard for the validation of internal models of market risk. Under the previous regulatory frameworks of Basel II and Basel II.5 (see BCBS 2006, 2011), backtesting required the comparison of one-day-ahead VaR forecasts at the 99% probability level with the ex-post realised losses and was based on so-called VaR violation indicators. Most proposed backtests for the accuracy of VaR forecasts are of a univariate nature and are based on univariate time series of VaR violations for a single portfolio, such as the entire trading book of a bank.
Christoffersen (1998) defined criteria, known as the unconditional coverage and independence hypotheses, that should be satisfied by credible VaR forecasts. The former is the requirement that the expected number of violations of VaR forecasts at probability level α over n time periods should be n ( 1 α ) , while the latter is the requirement that such violations should occur independently in time. The criteria can also be combined to obtain the conditional coverage hypothesis. An early approach to verifying the unconditional coverage hypothesis can be found in Kupiec (1995), while an approach to testing for the independence of the sequence of VaR violation indicators was presented by Christoffersen (1998). Tests of the independence hypothesis generally require an assumption about the dependence structure under the alternative hypothesis and can have very limited power to detect departures from independence caused by other forms of serial dependence. Moreover, as pointed out in Campbell (2007), joint tests of the unconditional coverage and independence hypothesis are not automatically preferable to separate testing; poorly performing VaR models, which violate only one of the two hypotheses, are less likely to be detected by a joint test than by two separate tests.
Since a reasonable model of a profit-and-loss (P&L) distribution should deliver acceptable VaR estimates at a range of probability levels, the revised regulatory requirements for the measurement of market risk (BCBS 2019)1 emphasise that backtesting should be carried out at multiple probability levels beyond the usual 99% level. In these new regulations, the expected shortfall risk measure plays an important role in the formula used to determine the required capital for market risk in the trading book. Since the expected shortfall is defined as the mean of the losses that are greater than the VaR at level α , a good estimate of the expected shortfall requires a  P&L  forecast that is accurate for VaR estimation at a range of α values in the tail of the distribution. Kratz et al. (2018) proposed a simultaneous multinomial test of VaR estimates at different α levels and suggested that such a test could be viewed as an implicit backtest of the expected shortfall. Alternative multilevel VaR tests were also developed by Campbell (2007), who proposed a Pearson’s chi-squared test for goodness of fit, and Pérignon and Smith (2008), who developed a likelihood-ratio test generalising the unconditional coverage test of Kupiec (1995).
Testing the accuracy of VaR forecasts at more than one α level implies that we move away from a simple assessment of the validity of VaR to a more thorough assessment of the forecast of the  P&L  distribution from which the VaR forecast is calculated. If the  P&L  distribution is adequately estimated, the resulting VaR estimates must be accurate for every α level in [ 0 , 1 ] . In the extreme case that we test at every α level, we backtest the complete  P&L  distribution. One advantage of backtesting the forecast of the  P&L  distribution (or a region theoreof) is that we exploit much more information in comparison to using the series of VaR violation indicators at a single α level. The latter may take only the values one or zero, depending on whether a VaR violation has occurred or not, and, for typical α levels in the region of 99%, violations are rare and the resulting indicator data are sparse.
Backtests of the forecast of the P&L distribution can be based on realised probability-integral transform (PIT) values. These are transformations of the realised value of P&L by the cumulative distribution function of the model used to forecast P&L at the previous time point. An ideal forecaster, i.e., a forecaster in the sense of Gneiting et al. (2007) who has knowledge of the correct model, would produce independent and uniformly distributed PIT values. Berkowitz (2001) proposed the first backtest of this kind based on the transformation of realised PIT values to a normal distribution under the null hypothesis.
The spectral tests of Gordy and McNeil (2020) are constructed by transforming PIT values with a weighting function, referred to as a kernel, and are available in both unconditional coverage and conditional coverage variants. Their form is very flexible and they include many of the previously proposed approaches to backtesting as special cases. Under the spectral backtesting philosophy, the risk modelling group or banking regulator can choose the kernel to apply weight to the area of the forecast distribution that is of primary interest—for example, a region in the area around the 99% quantile. Among the tests subsumed in this framework are the test of Kupiec (1995), which corresponds to a kernel equivalent to the Dirac measure at a specific probability level α , and the test of Berkowitz (2001). The spectral risk measure test of Costanzino and Curran (2015) and the expected shortfall test of Du and Escanciano (2017) are further special cases obtained by choosing a kernel truncated to tail probabilities.
While the test of Gordy and McNeil (2020) can be viewed as an absolute test of model adequacy for a single candidate model, a weighted approach to forecast comparison in specific areas of a forecast distribution was proposed by Amisano and Giacomini (2007). Their weighted likelihood ratio test is a relative test that compares the performance of two competing density forecasts and is based on the weighted averages of the logarithmic scoring rule, where the weights are chosen according to the preferences of the risk modelling group. Diks et al. (2011) proposed similar tests but based them on conditional likelihood and censored likelihood scoring rules, while Gneiting and Ranjan (2011) applied the (quantile-weighted) continuous ranked probability score instead of the logarithmic score.
In contrast to the rich variety of tests available for the backtesting of a univariate series of density forecasts, or comparing competing sets of forecasts, the literature on multivariate backtesting is much more sparse. Multivariate extensions of univariate VaR backtests have not been widely developed, although Berkowitz et al. (2011) already sketched a number of ideas. We are aware of two papers by Danciulescu (2016) and Wied et al. (2016) that deal with the multivariate backtesting of VaR, in the sense of backtesting VaR forecasts for several portfolios (or desks) simultaneously.
Danciulescu (2016) suggests a test for the unconditional coverage hypothesis and a test for the independence hypothesis using multivariate portmanteau test statistics of the Ljung–Box type, which are applied to the multivariate time series of VaR violation indicator variables. The test for the independence hypothesis simultaneously tests for the absence of cross- and autocorrelations in the multivariate time series of VaR violation indicator variables up to some finite lag K. In Wied et al. (2016), multivariate tests are proposed for the detection of clustered VaR violations, which would violate the conditional coverage hypothesis. They argue that their test can detect the clustering of VaR exceedances for a single desk, which would indicate that the probability of VaR violations is varying over time, as well as the clustering of VaR exceedances across desks at different lags, which would cast doubt on the assumption of independent VaR violations for different desks at different time points.
Unfortunately, due to the sensitivity of the data concerned, there are very few empirical studies of bank-wide  P&L backtesting and even fewer studies of desk-level data. One exception is Berkowitz et al. (2011), who analysed daily realisations from the  P&L  distribution and daily forecasts of VaR (calculated by the widely used historical simulation method) for each of four separate business lines at a large international commercial bank. In this study, various univariate tests, including the Markov test of Christoffersen (1998) for the conditional coverage hypothesis, the CaViaR test for autocorrelation of Engle and Manganelli (2004), and the unconditional coverage hypothesis test of Kupiec (1995), were used to backtest the VaR for each trading desk separately. Moreover, the tests were assessed based on their finite sample size and power properties in a Monte Carlo study. The authors found that the VaR models for two out of the four business lines were rejected due to volatility clustering and that the model of a third business line was rejected by the unconditional coverage hypothesis test of Kupiec (1995).
The remainder of our paper is organised as follows. In Section 2, we explain the testing approach based on PIT values and recapitulate the main details of the spectral test of Gordy and McNeil (2020). We then show how this may be extended to obtain multivariate spectral tests based on a single spectrum or multiple spectra in Section 3. In Section 4, we carry out a simulation study to analyse the size and power of the proposed tests for different amounts of data, different numbers of desks, different choices of kernels, and different deviations from the null hypothesis. Concluding remarks are found in Section 5.

2. Spectral Backtests of Forecast Distributions

2.1. Realised PIT Values

Let Ω , F , P denote a probability space and let ( L t ) t N be a time series defined on this space representing losses (positive numbers) and profits (negative numbers) on a portfolio of risky assets. For every t N , define the conditional loss distribution function F t by
F t ( x t x t 1 , , x 1 ) = P L t x t L t 1 = x t 1 , , L 1 = x 1
and assume that F t is an absolutely continuous function of x t . If we now define the process ( U t ) t N by U 1 = F 1 ( L 1 ) and U t = F t L t L t 1 , , L 1 for t 2 , then the result of Rosenblatt (1952) implies that U t t N is a sequence of independent and identically distributed (iid) standard uniform variables. Note that this holds in general without any assumption of stationarity for ( L t ) t N . This is important because the time-varying nature of the functions F t may come both from the time-varying nature of the underlying risk factors affecting the portfolio value (for example, equity prices and interest rates) as well as changes in the composition of the portfolio caused by rebalancing and capital inflows and outflows.
We assume that, at each time t, a risk modelling group at a bank or other financial institution makes a forecast F ^ t of F t based on the information available to it up to time t 1 , which includes { L 1 , , L t 1 } and possibly additional information. The corresponding random variables P t obtained by setting P t = F ^ t ( L t L t 1 , , L 1 ) for t N are referred to as the realised PIT values. For an omniscient or ideal forecaster in the sense of Gneiting et al. (2007), which is a forecaster who possesses extra information about the exact form of the function F t at every time point, it would follow that F ^ t = F t and the realised PIT values would be iid uniform. For an ordinary mortal forecaster, we effectively hold them to this ideal standard and test the extent to which the realised PIT values satisfy the iid uniform assumption.
We assume that model validation is carried out by a regulator who only has access to the time series of realised PIT values ( P t ) and has no knowledge of the forecasting models ( F ^ t ) that were used.

2.2. Spectral Tests

The spectral tests of Gordy and McNeil (2020) are based on transformations of the realised PIT values through the level exceedance indicator function Y t ( u ) = I { P t u } according to
W t = [ 0 , 1 ] I { P t u } d ν ( u ) , t = 1 , , n
where ν is a probability measure on [ 0 , 1 ] , which is referred to as the kernel measure.
The risk modelling group or model validation group can choose this measure to weight a set of quantile levels according to their preferences for forecast model performance. For example, if the preference is for model performance around the 99th percentile, the kernel measure would typically be chosen to place weight in a sub-interval of [ 0 , 1 ] containing the value 0.99. In the most extreme case, by choosing the Dirac measure at the value α = 0.99 , the spectral transformation of the realised PIT value in (1) would simply satisfy
W t = Y t ( α ) = I { F ^ t ( L t ) α } = I { L t F ^ t 1 ( α ) } .
In other words, it would yield the indicator variable for a VaR exceedance at level α . The paper of Gordy and McNeil (2020) contains examples of discrete and continuous probability measures ν . In the continuous case, we have W t = 0 1 g ( u ) I { P t u } d u , where g is the probability density corresponding to the measure ν ; in this case, we usually assume that ν is supported on a strict sub-interval [ α 1 , α 2 ] of [ 0 , 1 ] , which is referred to as the kernel window.
Monospectral tests are based on the time series ( W t ) t N . Gordy and McNeil (2020) also propose bispectral and, more generally, m-spectral tests in which a set of distinct kernel measures ν 1 , , ν m is considered. In this case, the data consist of vectors ( W t ) , where W t = ( W t , 1 , , W t , m ) and each component W t , j is calculated using the probability measure ν j and a spectral transformation of the form given in (1).
We can consider hypotheses about the series ( W t ) that generalise the notions of unconditional coverage, independence, and conditional coverage promoted by Christoffersen (1998): a test of the unconditional coverage hypothesis is a test that the distribution of the W t values is the one implied by the uniformity of the realised PIT values; an independence test is a test of the hypothesis that the W t values are independent of lagged realised PIT values P 1 , , P t 1 (which implies that the series ( W t ) is iid). A test of the conditional coverage hypothesis is a combined test of both of these hypotheses.
In this paper, our focus is on unconditional tests. The generalisation of unconditional spectral tests to conditional spectral tests is extensively discussed in Gordy and McNeil (2020) and it is clear from the analysis in that paper that conditional extensions are more powerful at detecting the presence of serial dependence in PIT values—for example, serial dependence resulting from unmodelled volatility effects. In this paper, we concentrate on tests that address the hypothesis of the uniformity of PIT values and how these may be generalised to the multi-desk case.
Gordy and McNeil (2020) consider different styles of test based on data W 1 , , W n for the unconditional coverage hypothesis, including likelihood ratio tests and Z-tests. We will extend the latter to the multi-desk situation due to their tractability, ease of implementation, and good performance in the single portfolio setting. A Z-test appeals to the asymptotic normality of the sample average W ¯ n = n 1 t = 1 n W t and tests whether
T n = n W ¯ n μ W Σ W 1 W ¯ n μ W χ m 2
where μ W and Σ W are the mean vector and covariance matrix of W t when the underlying realised PIT value P t is standard uniform. In the case of a monospectral test, we consider the simpler test of whether the following holds:
T ˜ n = n ( W ¯ n μ W ) σ W N ( 0 , 1 ) .
In the case where the probability measures ν j have absolutely continuous distribution functions G j with densities g j , we obtain simple formulas for the moments of W t , j = 0 1 g j ( u ) I { P t u } d u taking the form
μ W j = E ( W t , j ) = 0 1 g j ( u ) ( 1 u ) d u
Σ W j , k = E ( W t , j W t , k ) = 0 1 g j ( u ) G k ( u ) + g k ( u ) G j ( u ) ( 1 u ) d u
and when m = 1 , these reduce to μ W = μ W 1 and σ W 2 = σ W 1 2 = Σ W 1 , 1 . This is the case that we will extend in this paper. The main practical consideration is that the density functions g j should permit the moments in (4) and (5) to be easily evaluated. We will consider parametric densities g j supported on an interval [ α 1 , α 2 ] in the neighbourhood of a standard VaR confidence level ( α = 0.99 ) such that g j is either constant in the interval or increasing, to give greater weight to more extreme quantiles; the exact forms used may be found in Section 4.1. This approach is shown to work well in Gordy and McNeil (2020).

2.3. Monospectral versus Bispectral Tests

Monospectral tests can be useful in carrying out one-sided tests of the null hypothesis that E ( W t ) μ W (where μ W is the value obtained when PITs are uniform) against the alternative that E ( W t ) > μ W ; in the case of a VaR exception test, this amounts to testing that E ( Y t ( α ) ) 1 α versus E ( Y t ( α ) ) > 1 α or, in other words, the null hypothesis that the VaR exceedance probability is no larger than the desired value ( 1 α ) .
Bispectral or multispectral tests are two-sided tests and, in our view, should generally be preferred to the one-sided monospectral tests. In assessing the quality of forecasts, the systematic overestimation of risk is as suggestive of a deficiency in the forecast model as is systematic underestimation. Moreover, at the desk level, if the net position in some risky asset class changes from long to short, then a different tail of the P&L distribution becomes relevant and an overestimate of risk may become an underestimate.
One means of understanding the difference between a monospectral and a bispectral test is that the former only tests for the consistency of weighted PIT values with the first moment of W t under the null hypothesis, while the latter tests for the consistency of two different statistical functionals and thus effectively captures the misspecification of two moments. This can be particularly beneficial when one of the chosen kernels fails to detect deviations of the distribution of the realised PIT values from a uniform distribution. Spectral tests are weighted measures of the discrepancy between the empirical distribution of PIT values and the uniform distribution within the kernel window. It can emerge that this discrepancy results in the crossing of the distribution functions and that this crossing is undetected by a single weighting scheme but is revealed by using two such schemes. Further discussion of this phenomenon is found in Gordy and McNeil (2020).

3. Multi-Desk Spectral Backtests

3.1. Framework for Multi-Desk Backtests

We generalise the notation to consider d desks or sub-portfolios. For i = 1 , , d , let L t , i represent the loss attributed to sub-portfolio or trading desk i at time t, let F ^ t , i denote the forecast model for L t , i constructed using information up to time t 1 , and let P t , i = F ^ t , i L t , i L t 1 , i , , L 1 , i be the corresponding realised PIT value.
If the risk management function of desk or sub-portfolio i is modelling its risks well, we expect that the univariate PIT time series ( P t , i ) will behave like an iid uniform series. We expect cross-correlation between the series at lag zero, since portfolio or desk models are usually built using the same underlying methodology and the same correlated risk factor data, but we do not expect cross-correlation at other lags. This leads us to adopt the hypothesis that the PIT vectors ( P t ) given by P t = ( P t , 1 , P t , d ) are iid random vectors with uniform marginal distributions.
We first note that a simple but crude method of extending the single-portfolio spectral testing methodology to a multi-desk situation is through the Bonferroni correction (see, for example, Dunn 1959). Suppose that we calculate either monospectral test statistics T ˜ n , i based on (3) or multispectral test statistics T n , i based on (2) for each desk i. Let the p-values for the d desks be p 1 , , p d . The null hypothesis that the desks are collectively delivering acceptable forecasts of P&L would be rejected in a test of size no greater than β if min { p 1 , , p d } β / d . However, it is well known that the Bonferroni correction leads to a loss of statistical power and so this is not a recommended procedure in the backtesting context.
We now consider how the spectral methodology can be adapted to construct a joint test over all desks. In the following sections, we use the notation
W t , i , j = [ 0 , 1 ] I { P t , i u } d ν j ( u )
to refer to the spectral-transformed realised PIT value for desk i at time t using kernel ν j .

3.2. The Monospectral Case

We first consider a single spectral transformation of the PIT values of the form W t , i , 1 = [ 0 , 1 ] I { P t , i u } d ν ( u ) using the spectrum ν , which leads to a multi-desk generalisation of the test in (3). Under the assumption of uniform PIT values, each of the W t , i , 1 satisfies E ( W t , i , 1 ) = μ W and var ( W t , i , 1 ) = σ W 2 , where these moments take the same values as in (3).
Let W ¯ n = n 1 t = 1 n Z t , where Z t = d 1 i = 1 d W t , i , 1 is the average of the spectrally transformed, realised PIT values over all desks at time t. We note that this is a generalisation of the notation used in (3) since, when d = 1 , we have W ¯ n = n 1 t = 1 n W t , 1 , 1 , which is the average weighted realised PIT value over time for a single desk and a single spectrum. Under the uniform assumption, E ( W ¯ n ) = E ( Z t ) = E ( W t , i , 1 ) = μ W . However, the variance of W ¯ n satisfies var ( W ¯ n ) = n 2 σ Z 2 , where σ Z 2 = var ( Z t ) and σ Z σ W unless d = 1 ; in general, the unknown dependencies between desks must be considered in evaluating σ Z . Thus, the monospectral multi-desk test relies on generalising (3) to test whether
T ˜ n = n ( W ¯ n μ W ) σ Z N ( 0 , 1 )
where σ Z is a parameter that is not fully identified under the null hypothesis and must be replaced by a suitable estimate.
A natural first approach is to use the unbiased and consistent estimator given by σ ^ Z 2 = 1 n 1 t = 1 n ( Z t W ¯ n ) 2 , but this does not work well in practice. Occasional large values of Z t have a tendency to inflate σ ^ Z 2 and reduce the size of the test statistic in (6), leading to the undersizing of the test. We have explored some alternative estimators that are better at controlling the size, and the best-performing estimator is based on a method that will be referred to as the correlation estimation (CE) method in the results. The CE method is based on the observation that σ Z 2 may be written as
σ Z 2 = var ( Z t ) = 1 d 2 i = 1 d j = 1 d cov ( W t , i , 1 , W t , j , 1 ) = σ W 2 d 2 i = 1 d j = 1 d ρ i j
where ρ i j denotes the correlation between the spectrally transformed realised PIT values for desks i and j. If R ^ W denotes the correlation matrix of the data { ( W t , 1 , , W t , d ) , t = 1 , , n } , we can use the consistent estimator
σ ^ Z 2 = σ W 2 d 2 1 R ^ W 1 .
The intuition is that by estimating a bounded quantity like correlation empirically while using the values of σ W 2 derived under the null hypothesis, we obtain an estimator that is relatively robust to outlying observations of ( W t , 1 , , W t , d ) , or, in other words, observations where multiple desks show extreme PIT values. In contrast, if we estimate empirical variances and covariances, we observe the tendency for inflated estimates to dilute the size and power of the test.
In general, we expect positive dependence between desks. The above estimator can occasionally give realised values that point to negative dependence, even in simulation experiments where we know that there is a positive correlation between desks. We find that better results are obtained by setting a floor for the estimator at the value that would be obtained when the desks are independent, i.e.,  σ Z 2 = σ W 2 / d . Thus, we work, in practice, with the estimator σ ˜ Z 2 = max ( σ W 2 / d , σ ^ Z 2 ) , where σ ^ Z 2 is given by (7).

3.3. The Bispectral Case

Recall from Section 2.2 that a multispectral test is based on m probability measures ν 1 , , ν m . In this section, we will consider m = 2 continuous measures with densities g 1 and g 2 so that, for each desk i and time t, we have a pair of spectrally transformed PIT values given by ( W t , i , 1 , W t , i , 2 ) . Let Z t be given by Z t = ( Z t , 1 , Z t , 2 ) , where Z t , j = 1 d i = 1 d W t , i , j is the desk average for measure j and j = 1 , 2 .
Consider the statistic W ¯ n = n 1 t = 1 n Z t , which generalises the statistic used in (2) to the case of more than one desk. The mean of W ¯ n under the null hypothesis is given by
μ W = E ( W ¯ n ) = E ( Z t ) = ( μ W 1 , μ W 2 )
where the μ W j are exactly as calculated in (4). The covariance matrix of W ¯ n is given by var ( W ¯ n ) = n 1 Σ Z , where
Σ Z = var ( Z t , 1 ) cov ( Z t , 1 , Z t , 2 ) cov ( Z t , 1 , Z t , 1 ) var ( Z t , 2 ) .
The application of the multivariate Central Limit Theorem Van der Vaart (2000) leads to a test of the form
T n = n W ¯ n μ W Σ Z 1 W ¯ n μ W χ 2 2 .
Note that in the case of a single desk ( d = 1 ), we have Σ Z = Σ W and we recover the test statistic T n given in (2) in Section 2.2.
As in the case of the monospectral test, we have to consider how to estimate Σ Z under the null hypothesis when we have no information about dependencies across desks. We use a method that is the natural extension of CE. Specifically, we decompose the variances and covariances in (8) into a variance part, which is known under the null hypothesis, and a part depending on the unknown correlation across desks.
Let V t , j = ( W t , 1 , j , W t , d , j ) denote the vector of W-values across all d desks at time t under weighting function g j , for j = 1 , 2 . Since Z t , j = d 1 1 V t , j , we have
var ( Z t , j ) = 1 d 2 1 var ( V t , j ) 1 = σ W j 2 d 2 1 R V j 1 , j = 1 , 2 , cov ( Z t , 1 , Z t , 2 ) = 1 d 2 1 cov ( V t , 1 , V t , 2 ) 1 = σ W 1 2 σ W 2 2 d 2 1 R V 1 V 2 1
where σ W j 2 = Σ W j , j with Σ W j , j as defined in (5) and where R V j and R V 1 V 2 denote correlation matrices. More specifically, R V j is the d × d correlation matrix for the W-values under measure ν j across all d desks; this is estimated by R ^ V j , the empirical correlation matrix of the vectors { V t , j , t = 1 , , n } . The matrix R V 1 V 2 is the d × d correlation matrix whose element ( j , k ) is the correlation between the W-value for desk j under measure 1 and desk k under measure 2; this is estimated by R ^ V 1 V 2 , the matrix whose element ( j , k ) is the sample correlation of the pairs ( W t , j , 1 , W t , k , 2 ) , t = 1 , , n .

4. Simulation Study

4.1. Design of the Study

We vary the following variables in the simulation study.
  • Sample size n. This corresponds to the number of days used in the bank’s backtesting exercise. The length of the backtesting period is typically small (one or two years of daily data corresponding to days on which markets are open) and so we consider n = 250 and n = 500 .
  • Number of desks d. This can be quite large in a bank with extensive trading operations and we consider the values d = 50 and d = 100 .
  • Copula C of PIT values across desks. We assume different dependence structures across desks by sampling PIT values with different copulas. In particular, we use the Gauss copula and the copula of a multivariate t distribution with 4 degrees of freedom. The latter case allows us to see how the properties of the spectral tests are affected by tail dependencies in the PIT data.
  • Level of dependence ρ across desks. For simplicity, we assume that all desks are equi-dependent by setting the correlation matrix R of the Gauss and t copulas to be an equicorrelation matrix with common parameter ρ , which takes the values ρ = 0 or ρ = 0.5 ; more details are given below.
  • Fraction of misspecified desks. In examining size and power, we vary the fraction of desks that use misspecified P&L  distributions in their risk models.
  • Specification of spectral test. We use a number of different monospectral and bispectral tests as detailed below.
We generate n independent realisations of the vector P t = ( P t , 1 , , P t , d ) representing the PIT values at time t for desks i = 1 , , d . The vectors P t are drawn from the distribution
P t C G 1 ( u 1 ) , , G d ( u d ) ,
where C denotes the copula (Gauss or Student) and G 1 , , G g are continuous univariate distribution functions, which are designed to capture the effects of correctly and incorrectly specified desk-level P&L models. The correlation matrix R, which is used to parameterise the Gauss copula and the Student t4 copula, is an equicorrelation matrix with parameter ρ = 0 or ρ = 0.5 . Note that, in the former case, the Gauss copula yields a model where the simulated PIT values are independent across desks, while the t4 copula yields a model with dependencies. This is because, even with a correlation matrix R equal to the identity, the Student t4 copula still has tail dependence, which will tend to lead to very large or very small PIT values occurring together across a number of desks; see, for example, McNeil et al. (2015)2.
For the marginal distributions G i , we use a construction first developed in Kratz et al. (2018) and also used in Gordy and McNeil (2020). We set G i ( u ) = Φ ( F i 1 ( u ) ) , where Φ is the standard normal distribution function and where F i is either the standard normal distribution or the distribution function of a univariate Student t4 distribution scaled to have variance one. If we wish to mimic a desk that is using a correctly specified desk model, we choose the normal and thus obtain G i ( u ) = u , the distribution function of standard uniform. If we wish to mimic a desk that is using an incorrectly specified model, we choose the scaled t4 and obtain a distribution function G i , which is supported on the unit interval [ 0 , 1 ] but is not uniform; on the contrary, it is the type of PIT value distribution that would be obtained if the desk were using a model (represented by Φ ) that was lighter-tailed than the true P&L distribution (represented by F i ) and thus underestimated the potential for large losses (and large gains). It is important to note that the use of these two distributions is simply a device to generate PIT data that either satisfy or violate the null hypothesis; we do not claim that these distributions are in any sense the true distributions. Recall that the null hypothesis is that the PIT vectors ( P t ) are iid random vectors with uniform marginal distributions but an unknown dependence structure.
For the spectral tests, we selected kernels corresponding to the following three continuous weighting functions g defined on [ α 1 , α 2 ] :
(1)
The uniform weighting function g ( u ) = 1 ;
(2)
The linear weighting g ( u ) = u ;
(3)
The exponential weighting function g ( u ) = exp ( k ( u α 1 ) ) 1 with k = 1 .
The values α 1 and α 2 determine the kernel window. We choose α 1 = 0.9805 and α 2 = 0.9995 , which gives a symmetric interval around 0.99 . Note that the functions are non-decreasing, placing more weight on more extreme outcomes (in the right tail).
For the multivariate monospectral Z-tests, the kernel functions listed above lead to three different tests, which we denote, respectively, by SP.U, SP.L, and SP.E. For the multivariate bispectral Z-tests, we will look at two different Z-tests, which combine the continuous kernel functions listed above. The test denoted SP.UL combines the uniform and linear weighting functions. The test denoted SP.UE combines the uniform and exponential weighting functions.
The simulation experiments described in the following sections were all performed using the R package simsalapar, which is a very flexible tool for conducting large-scale studies with a number of different dimensions (see Hofert and Mächler 2016). The tables that we provide show observed rejection rates for the null hypothesis in 1000 replications of the simulation experiment.
We use a colouring convention to help with the interpretation of these rejection rates and this is applied differently according to whether the results address the size or power of a test. Simulation results relating to size are colour-coded as follows: good results (observed size smaller or equal 6%) are coloured green; poor results (observed size in range [9–12%]) are coloured pink; very poor results (size above 12) are coloured red; all other values are uncoloured. Simulation results relating to power are colour-coded as follows: good results (observed power above 70%) are coloured green; poor results (observed power in range [30–10%]) are coloured pink; very poor results (power below 10%) are coloured red; all other values are uncoloured.

4.2. Evaluating the CE Method of Correcting for Inter-Desk Dependence

We first investigate the crucial CE method of correcting for the unknown dependence structure across desks. We consider two extreme situations—one in which all desks are correctly specified and one in which all desks are incorrectly specified. The former situation allows us to evaluate the size of the test, i.e., the probability of a significant test result when the null hypothesis holds. The latter situation is one that should certainly be picked up by any backtest with reasonable power.
Results for the monospectral tests are found in Table 1. These relate to one-sided tests of the null hypothesis, where we are interested in being able to detect the systematic underestimation of tail risk in the right tail of the loss distribution. The nominal level β of the tests is 0.05. The table shows the actual test rejection rates over 1000 replications for different backtest lengths n, desk numbers d, copulas C, distribution functions F i , and correlation values ρ . The field CE shows whether the CE correction method has been used or not.
The rows in which F i is recorded as “N” address the question of size and we expect values close to the nominal level of the test 0.05. However, when the CE correction method is not implemented, it is clear that the spectral tests are oversized in all cases except where the desks are independent, which is the column corresponding to a Gauss copula and ρ = 0 ; in the case of a t4 copula with ρ = 0 , the desks are still dependent and the tests are oversized. In the absence of correction for correlation, there are a number of results coloured red, indicating a complete inability to control the size. In contrast, when the CE correction is implemented, the results are coloured green in all cases.
The rows in which F i is recorded as “t4” address the question of power, since all desks are misspecified. For each of the spectral tests, the power increases with both n and d, as we would expect. However, the power decreases with strengthening dependence across desks; the power for the case C = t4 and ρ = 0 is greater than for C = Gauss and ρ = 0.5 , which in turn is greater than for C = t4 and ρ = 0.5 . Increasing levels of dependence can be thought of as effectively reducing the number of independent desk results. Turning to the different spectral tests, the performance shows similar patterns but the SP.L kernel (linear weighting function) seems to give the highest power in this case.
In view of this first set of results, we will apply the CE correction method to the spectral tests in all further experiments.

4.3. Size and Power of Bispectral Tests

We now consider the two bispectral tests under the two scenarios of Table 1—all desks correctly specified and all desks misspecified. For the bispectral test, we also add results for n = 1000 backtests, corresponding to 4 years of data. Results are shown in Table 2. Note that the bispectral tests are two-sided tests.
While the power of these tests is perfect, the size properties are not as good as for the monospectral tests. This is particularly apparent in the case of a backtest of length n = 250 ; the situation improves for n = 500 and the size results are very good for n = 1000 ; if anything, there is evidence of undersizing. We interpret these results as showing that more data are required in order for the estimators of (10) to give an accurate correction for the unknown desk dependence structure in the bispectral case.

4.4. Evaluating the Effect of Desk Misspecification Rate

We now consider the more realistic situation where only a certain fraction of desks are incorrectly specified. We choose values of this fraction in the set of 25% and 50%. All the misspecified desks use the model in which F i is t4. Results for monospectral tests are shown in Table 3.
When 25% of models are misspecified models, the power is rather low, except in the case where the desks are independent (i.e., have a Gauss copula and correlation ρ = 0 ). We attribute this to the intuition that independent data increase the effective sample size, whereas correlated data decrease it. The power increases with both n and d. As the misspecification rate increases, the power increases, as we would expect. The spectral test with a linear weighting function (SP.L) gives the most power, while the other two kernels are comparable.
The results for the bispectral tests SP.UL and SP.UE are reported in Table 4. In this case, we add results for n = 1000 , since we have observed that bispectral tests typically require more data for the correlation correction (CE) method to give tests that are well sized. We also add results for a misspecification fraction of 10%.
The results are clearly better than for the monospectral tests and show the advantages of bispectral tests—by using two kernels, we can effectively test for the correct specification of more moments of the distribution of PIT values in the kernel window (see discussion in Section 2.3). The general observations are the same as for the monospectral tests; the power increases with both n and d and with weakening dependence across desks. While it would clearly be best to base backtests on 1000 observations, to obtain tests that are both well sized and powerful, reasonable results are obtained for n = 500 even when only 25% of the desks are using misspecified risk models. When only 10% of the desks use misspecified models, the power is clearly weaker, but the test does still have some ability to detect that a number of desks are delivering poor P&L estimates.
Ther is little to choose from between the SP.UL and SP.UE tests. While the latter seems to be slightly more powerful, it also tends to have slightly worse size properties, as seen in Table 2; we would tend to favour the former for samples of size n = 500 .

5. Conclusions

In this paper, we proposed multivariate spectral Z-tests, which can be used to simultaneously backtest trading desk models when the dependence structure of P&L across trading desks is unknown. Multivariate backtests are potentially more powerful than univariate backtests at the level of the trading book, since they exploit a greater amount of data; typical banks can have 50–100 trading desks. Moreover, a simultaneous backtest avoids the problem of aggregating and drawing inferences from a set of single-desk backtests in the presence of unknown dependencies between test results.
The tests that we have developed are a multivariate extension of the spectral tests proposed in Gordy and McNeil (2020). They take the form of a Z-test against a normal or chi-squared reference distribution and make use of realised PIT values as input variables. PIT values provide more data about the quality of desk models than indicator variables for VaR violations and their benefits are already being exploited in the USA. It is likely that other regulatory authorities will use this type of information more systematically in the future as well.
The multivariate spectral tests are designed in such a way that the risk modelling group can select a kernel or weighting scheme to emphasise the region of the estimated P&L distribution where model performance is most critical, typically a region in the tail representing large losses. The tests can also provide an indirect validation of the expected shortfall measure, which now has a prominent role in the revised regulation.
We suggested a method of controlling for the unknown dependencies between desks, based on estimating correlations. The resulting tests were generally well sized, although bispectral tests typically required samples of n = 500 PIT vectors (2 years of daily data), while monospectral tests only required n = 250 (one year of daily data). However, in realistic situations where only a minority of the used desk models underestimated the risks in their P&L distributions, bispectral tests gave much better power and should generally be preferred to their monospectral counterparts.
The performance of the proposed multivariate tests suggests that they are a valuable addition to a bank’s validation framework. Since they take the form of Z-tests, they are easy to implement and have quick run times. In the event of a significant test result, a bank would be able to implement a post-hoc testing scheme on individual desks to see where the main problems lie.
Finally, the new multivariate backtests could also have an interesting application to backtesting in the banking sector as a whole, based on data for the trading book provided to the regulator by individual banks. Significant test results would be an indication of the potential for spillover effects across banks caused by trading activities and would contribute to the understanding of systemic risk.
Clearly, it would be of interest to apply our tests to actual reported data from banks. However, multi-desk data are not currently available to researchers due to the sensitivities surrounding bank risk reporting both in the EU and the USA. An anonymised study of real data at the trading-book level for US banks is provided by Gordy and McNeil (2020) and shows the advantages of the (univariate) spectral backtesting framework over simple VaR backtests; we would certainly expect that these advantages will carry over to the multivariate setting. We hope that, by setting out a methodology for multi-desk backtesting in this paper, we can help to promote the use of desk-level PIT values for model validation and encourage banks and regulators to allow some datasets of realised PIT values to enter the public domain and stimulate further research into the refinement of the methodology.

Author Contributions

Conceptualisation, J.B. and A.J.M.; methodology, J.B. and A.J.M.; software, J.B.; validation, J.B. and A.J.M.; formal analysis, J.B. and A.J.M.; investigation, J.B. and A.J.M.; writing—original draft preparation, J.B. and A.J.M.; writing—review and editing, J.B. and A.J.M.; visualisation, J.B. and A.J.M.; supervision, A.J.M.; project administration, J.B. and A.J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The code to reproduce the simulation experiments is available from the authors.

Conflicts of Interest

Author Janine Balter was employed by the company Deutsche Bundesbank. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The Deutsche Bundesbank had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Notes

1
The regulatory framework, known as the Fundamental Review of the Trading Book (FRTB), requires that the backtest compares whether the number of observed VaR exceptions is consistent with a 99% confidence level with the comment that other α levels should be added to the backtesting exercise; see BCBS (2019) [M32.5] and [M32.13].
2
At the suggestion of a referee, we also tried repeating all analyses using a Gumbel copula to simulate the dependence structure across desks, since this is a typical copula for extreme values and also leads to PIT values with upper tail dependence. The results were comparable to the Gauss and t4 copulas, in terms of size and power properties, and we have not included them in the tables.

References

  1. Amisano, Gianni, and Raffaella Giacomini. 2007. Comparing density forecasts via weighted likelihood ratio tests. Journal of Business & Economic Statistics 25: 177–90. [Google Scholar]
  2. Basel Committee on Banking Supervision. 2006. International Convergence of Capital Measurement and Capital Standards. Available online: https://www.bis.org/publ/bcbs128.pdf (accessed on 30 March 2020).
  3. Basel Committee on Banking Supervision. 2011. Revisions to the Basel II Market Risk Framework. Available online: https://www.bis.org/publ/bcbs193.pdf (accessed on 13 January 2021).
  4. Basel Committee on Banking Supervision. 2019. Minimum Capital Requirements for Market Risk. Available online: https://www.bis.org/bcbs/publ/d457.pdf (accessed on 18 January 2021).
  5. Berkowitz, Jeremy. 2001. Testing density forecasts, with applications to risk management. Journal of Business & Economic Statistics 19: 465–74. [Google Scholar]
  6. Berkowitz, Jeremy, Peter Christoffersen, and Denis Pelletier. 2011. Evaluating value-at-risk models with desk-level data. Management Science 57: 2213–27. [Google Scholar] [CrossRef]
  7. Campbell, Sean D. 2007. A review of backtesting and backtesting procedures. Journal of Risk 9: 1–17. [Google Scholar] [CrossRef]
  8. Christoffersen, Peter F. 1998. Evaluating interval forecasts. International Economic Review 39: 841–62. [Google Scholar] [CrossRef]
  9. Costanzino, Nick, and Mike Curran. 2015. Backtesting general spectral risk measures with application to expected shortfall. Journal of Risk Model Validation 9: 21–31. [Google Scholar] [CrossRef]
  10. Danciulescu, Cristina. 2016. Backtesting aggregate risk. Journal of Forecasting 35: 285–307. [Google Scholar] [CrossRef]
  11. Diks, Cees, Valentyn Panchenko, and Dick van Dijk. 2011. Likelihood-based scoring rules for comparing density forecasts in tails. Journal of Econometrics 163: 215–30. [Google Scholar] [CrossRef]
  12. Du, Zaichao, and Juan Carlos Escanciano. 2017. Backtesting expected shortfall: Accounting for tail risk. Management Science 63: 940–85. [Google Scholar] [CrossRef]
  13. Dunn, Olive Jean. 1959. Estimation of the medians for dependent variables. The Annals of Mathematical Statistics 30: 192–97. [Google Scholar] [CrossRef]
  14. Engle, Robert F., and Simone Manganelli. 2004. Caviar: Conditional autoregressive value at risk by regression quantiles. Journal of Business & Economic Statistics 22: 367–81. [Google Scholar]
  15. Gneiting, Tilmann, and Roopesh Ranjan. 2011. Comparing density forecasts using threshold-and quantile-weighted scoring rules. Journal of Business & Economic Statistics 29: 411–22. [Google Scholar]
  16. Gneiting, Tilmann, Fadoua Balabdaoui, and Adrian E. Raftery. 2007. Probabilistic forecasts, calibration and sharpness. Journal of the Royal Statistical Society 69: 243–68. [Google Scholar] [CrossRef]
  17. Gordy, Michael B., and Alexander J. McNeil. 2020. Spectral backtests of forecast distributions with application to risk management. Journal of Banking & Finance 116: 105817. [Google Scholar]
  18. Hofert, Marius, and Martin Mächler. 2016. Parallel and other simulations in R made easy: An end-to-end study. Journal of Statistical Software 69: 1–44. [Google Scholar] [CrossRef]
  19. Jorion, Philippe. 2007. Value at Risk. The New Benchmark for Managing Financial Risk, 3rd ed. New York: McGraw-Hill. [Google Scholar]
  20. Kratz, Marie, Yen H. Lok, and Alexander J. McNeil. 2018. Multinomial VaR backtests: A simple implicit approach to backtesting expected shortfall. Journal of Banking & Finance 88: 393–407. [Google Scholar]
  21. Kupiec, Paul. 1995. Techniques for verifying the accuracy of risk measurement models. The Journal of Derivatives 3: 73–84. [Google Scholar] [CrossRef]
  22. McNeil, Alexander J., Rüdiger Frey, and Paul Embrechts. 2015. Quantitative Risk Management: Concepts, Techniques and Tools, 2nd ed. Princeton: Princeton University Press. [Google Scholar]
  23. Pérignon, Christophe, and Daniel R. Smith. 2008. A new approach to comparing VaR estimation methods. The Journal of Derivatives 16: 54–66. [Google Scholar] [CrossRef]
  24. Rosenblatt, Murray. 1952. Remarks on a multivariate transformation. The Annals of Mathematical Statistics 23: 470–72. [Google Scholar] [CrossRef]
  25. Van der Vaart, Aad W. 2000. Asymptotic Statistics. Cambridge: Cambridge University Press. [Google Scholar]
  26. Wied, Dominik, Gregor N. F. Weiß, and Daniel Ziggel. 2016. Evaluating value-at-risk forecasts: A new set of multivariate backtests. Journal of Banking & Finance 72: 121–32. [Google Scholar]
Table 1. Size and power of monospectral Z-tests when all desk models are correctly specified and when all desk models are misspecified.
Table 1. Size and power of monospectral Z-tests when all desk models are correctly specified and when all desk models are misspecified.
d50100
ρ 00.500.5
TestCEn F i | C Gausst4 Gausst4Gausst4Gausst4
SP.UNo250N5.023.626.229.25.129.630.532.9
t4100.093.689.879.0100.096.492.482.4
500N5.124.727.532.95.429.432.536.7
t4100.099.497.791.1100.099.798.693.4
Yes250N4.64.44.24.54.94.94.44.6
t4100.080.065.741.7100.080.067.242.9
500N5.04.15.04.55.44.04.54.6
t4100.095.987.365.8100.097.388.966.3
SP.LNo250N4.622.725.528.95.128.529.332.1
t4100.098.796.490.2100.099.297.892.2
500N5.125.526.634.25.530.432.437.7
t4100.0100.099.797.4100.0100.099.998.4
Yes250N4.34.64.25.44.85.14.35.2
t4100.094.388.962.8100.096.490.164.5
500N5.03.35.15.05.53.54.65.2
t4100.099.998.186.0100.099.898.786.3
SP.ENo250N5.023.626.229.25.129.630.532.9
t4100.093.690.079.0100.096.592.582.5
500N5.124.727.532.95.429.432.536.8
t4100.099.497.791.3100.099.798.693.5
Yes250N4.64.44.24.44.94.94.44.6
t4100.080.266.142.0100.080.367.843.0
500N5.04.15.04.55.44.04.54.6
t4100.096.187.466.1100.097.589.166.4
Table 2. Size and power of bispectral Z-tests when all desk models are correctly specified and when all desk models are misspecified.
Table 2. Size and power of bispectral Z-tests when all desk models are correctly specified and when all desk models are misspecified.
d50100
ρ 0 0.5 0 0.5
Testn F i | C Gausst4Gausst4Gausst4Gausst4
SP.UL250N5.89.411.618.64.112.312.221.0
t4100.0100.0100.0100.0100.0100.0100.0100.0
500N2.75.45.010.23.17.26.610.0
t4100.0100.0100.0100.0100.0100.0100.0100.0
1000N3.03.13.14.82.34.54.65.5
t4100.0100.0100.0100.0100.0100.0100.0100.0
SP.UE250N6.611.212.720.24.914.014.522.4
t4100.0100.0100.0100.0100.0100.0100.0100.0
500N3.06.25.410.63.67.76.810.7
t4100.0100.0100.0100.0100.0100.0100.0100.0
1000N3.13.03.35.22.54.74.86.0
t4100.0100.0100.0100.0100.0100.0100.0100.0
Table 3. Power of monospectral Z-tests when a certain proportion of desks is misspecified.
Table 3. Power of monospectral Z-tests when a certain proportion of desks is misspecified.
d50100
ρ 0 0.5 0 0.5
TestFraction n | C Gauss t4 Gauss t4 Gauss t4 Gauss t4
SP.U2525043.015.612.810.169.916.814.110.7
50066.722.218.313.490.424.019.613.1
5025089.036.529.618.599.339.629.319.9
50099.356.243.127.8100.058.943.527.7
SP.L2525067.223.119.113.089.725.920.314.5
50088.834.728.118.799.337.830.819.7
5025099.256.746.527.4100.058.747.427.2
500100.080.068.743.3100.082.170.244.5
SP.E2525043.115.612.910.370.116.914.110.7
50066.822.318.313.490.424.219.613.1
5025089.136.629.618.599.339.829.520.0
50099.356.643.427.9100.059.143.728.0
Table 4. Power of bispectral Z-tests SP.UL and SP.UE when a certain proportion of desks is misspecified.
Table 4. Power of bispectral Z-tests SP.UL and SP.UE when a certain proportion of desks is misspecified.
d50100
ρ 0 0.5 0 0.5
TestFraction n | C Gauss t4 Gauss t4 Gauss t4 Gauss t4
SP.UL1025023.116.321.217.538.519.223.318.4
50032.714.618.111.457.418.723.810.9
100056.524.328.69.287.227.740.09.8
2525076.948.659.028.397.469.583.431.7
50095.066.580.931.3100.086.296.137.6
1000100.092.497.449.3100.098.699.955.6
5025099.997.298.978.9100.099.9100.083.8
500100.099.9100.092.7100.0100.0100.094.9
1000100.0100.0100.099.5100.0100.0100.099.9
SP.UE1025025.317.522.518.542.419.924.719.2
50034.514.518.711.660.018.724.610.9
100057.524.428.78.587.727.840.19.3
2525079.951.161.628.898.070.785.231.7
50095.567.281.830.3100.086.796.236.8
1000100.092.597.548.7100.098.599.954.8
5025099.997.799.079.2100.099.9100.084.2
500100.099.9100.092.8100.0100.0100.094.6
1000100.0100.0100.099.5100.0100.0100.099.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Balter, J.; McNeil, A.J. Multivariate Spectral Backtests of Forecast Distributions under Unknown Dependencies. Risks 2024, 12, 13. https://doi.org/10.3390/risks12010013

AMA Style

Balter J, McNeil AJ. Multivariate Spectral Backtests of Forecast Distributions under Unknown Dependencies. Risks. 2024; 12(1):13. https://doi.org/10.3390/risks12010013

Chicago/Turabian Style

Balter, Janine, and Alexander J. McNeil. 2024. "Multivariate Spectral Backtests of Forecast Distributions under Unknown Dependencies" Risks 12, no. 1: 13. https://doi.org/10.3390/risks12010013

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop