Next Article in Journal
Pricing of Commodity Derivatives on Processes with Memory
Previous Article in Journal
Measuring Financial Contagion and Spillover Effects with a State-Dependent Sensitivity Value-at-Risk Model
Previous Article in Special Issue
Developing an Impairment Loss Given Default Model Using Weighted Logistic Regression Illustrated on a Secured Retail Bank Portfolio
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Markov Chain Monte Carlo Methods for Estimating Systemic Risk Allocations

Department of Statistics and Actuarial Science, University of Waterloo, 200 University Avenue West, Waterloo, ON N2L 3G1, Canada
*
Author to whom correspondence should be addressed.
Submission received: 24 September 2019 / Revised: 2 January 2020 / Accepted: 7 January 2020 / Published: 15 January 2020
(This article belongs to the Special Issue Computational Methods for Risk Management in Economics and Finance)

Abstract

:
In this paper, we propose a novel framework for estimating systemic risk measures and risk allocations based on Markov Chain Monte Carlo (MCMC) methods. We consider a class of allocations whose jth component can be written as some risk measure of the jth conditional marginal loss distribution given the so-called crisis event. By considering a crisis event as an intersection of linear constraints, this class of allocations covers, for example, conditional Value-at-Risk (CoVaR), conditional expected shortfall (CoES), VaR contributions, and range VaR (RVaR) contributions as special cases. For this class of allocations, analytical calculations are rarely available, and numerical computations based on Monte Carlo (MC) methods often provide inefficient estimates due to the rare-event character of the crisis events. We propose an MCMC estimator constructed from a sample path of a Markov chain whose stationary distribution is the conditional distribution given the crisis event. Efficient constructions of Markov chains, such as the Hamiltonian Monte Carlo and Gibbs sampler, are suggested and studied depending on the crisis event and the underlying loss distribution. The efficiency of the MCMC estimators is demonstrated in a series of numerical experiments.

1. Introduction

In portfolio risk management, risk allocation is an essential step to quantifying the risk of each unit of a portfolio by decomposing the total risk of the whole portfolio. One of the most prevalent rules to determine risk allocations is the Euler princple, proposed by Tasche (1995) and justified from various viewpoints, such as the RORAC compatibility (Tasche (1995) and Tasche (2008)) and cooperative game theory (Denault (2001)). For the popular risk measures, such as VaR, RVaR, and ES, Euler allocations take the form of conditional expectations of the underlying loss random vector given a certain rare event on the total loss of the portfolio; see Tasche (2001) for derivations. We call this rare event the crisis event.The decomposition of risks is also required in the context of systemic risk measurement. Systemic risk is the risk of financial distress of an entire economy as a result of the failure of individual components of the financial system. To quantify such risks, various systemic risk measures have been proposed in the literature, such as conditional VaR (CoVaR) (Adrian and Brunnermeier (2016)), conditional expected shortfall (CoES) (Mainik and Schaanning (2014)), and marginal expected shortfall (MES) (Acharya et al. (2017)). These three measures quantify the risk of individuals by taking the VaR, ES, and expectation of the individual loss, respectively, under some stressed scenario—that is, given the crisis event. Chen et al. (2013), Hoffmann et al. (2016), and Kromer et al. (2016) proposed an axiomatic characterization of systemic risk measures, where the risk of the aggregated loss in a financial system is first measured and then decomposed into the individual economic entities. Due to the similarity of risk allocations with the derivation of systemic risk measures, we refer to both of them as systemic risk allocations. In fact, MES coincides with the Euler allocation of ES, and other Euler allocations can be regarded as special cases of systemic risk measures considered in Gourieroux and Monfort (2013).
Calculating systemic risk allocations given an unconditional joint loss distribution is generally challenging, since analytical calculations often require knowledge of the joint distribution of the marginal and aggregated loss. Furthermore, MC estimation suffers from the rare-event character of the crisis event. For computing CoVaR, CoES, and MES, Mainik and Schaanning (2014), Bernardi et al. (2017), and Jaworski (2017) derived formulas based on the copula of the marginal and aggregated loss; Asimit and Li (2018) derived asymptotic formulas based on the extreme value theory; and Girardi and Ergün (2013) estimated CoVaR under a multivariate GARCH model. Vernic (2006), Chiragiev and Landsman (2007), Dhaene et al. (2008), and Furman and Landsman (2008) calculated Euler allocations for specific joint distributions. Asimit et al. (2011) derived asymptotic formulas for risk allocations. Furman and Zitikis (2009) and Furman et al. (2018) calculated weighted allocations, which include Euler allocations as special cases, under a Stein-type assumption. Concerning the numerical computation of Euler allocations, Glasserman (2005), Glasserman and Li (2005), and Kalkbrener et al. (2004) considered importance sampling methods, and Siller (2013) proposed the Fourier transform Monte Carlo method, all specifically for credit portfolios. For general copula-based dependence models, analytical calculations of systemic risk allocations are rarely available, and an estimation method is, to the best of our knowledge, only addressed in Targino et al. (2015), where sequential Monte Carlo (SMC) samplers are applied.
We address the problem of estimating systemic risk allocations under general copula-based dependent risks in the case where the copula between the marginal and aggregated losses are not necessarily available. We consider a general class of systemic risk allocations in the form of risk measures of a conditional loss distribution given a crisis event, which includes CoVaR, CoES, MES, and Euler allocations as special cases. In our proposed method, the conditional loss distribution, called the target distribution π , is simulated by a Markov chain whose stationary distribution is the desired distribution π by sequentially updating the sample path based on the available information from π . While this MCMC method resembles the SMC in Targino et al. (2015), the latter requires a more complicated implementation involving the choice of forward and backward kernels, resampling and move steps, and even MCMC in the move steps. Our suggested approach directly constructs a single sophisticated Markov chain depending on the target distribution of interest. Applications of MCMC to estimating risk allocations have been studied in Koike and Minami (2019), specifically for VaR contributions. Our paper explores and demonstrates the applicability of MCMC methods to a more general class of systemic risk allocations.
Almost all MCMC methods used in practice are of the Metropolis–Hastings (MH) type (Metropolis et al. (1953) and Hastings (1970)), where the so-called proposal distribution q generates a candidate of the next state based on the current state. This candidate is then accepted or rejected according to the so-called acceptance probability to adjust the stationary distribution to be the target distribution π . As explained in Section 3.1 below, the resulting Markov chain has serial correlation, which adversarially affects the efficiency of the estimator. An efficient MCMC of MH type is such that the proposal distribution generates a candidate which exhibits low correlation with the current state with sufficiently large acceptance probability. The main difficulty in constructing such an efficient MCMC estimator for systemic risk allocations is that the support of the target distribution π is subject to constraints determined by the crisis event. For such target distributions, simple MCMC methods, such as random walk MH, are not efficient since a candidate is immediately rejected if it violates the constraints; see Section 3.2 for details.
To tackle this problem, we consider two specific MCMC methods, Hamiltonian Monte Carlo (HMC) (Duane et al. (1987)) and the Gibbs sampler (GS) (Geman and Geman (1984) and Gelfand and Smith (1990)). In the HMC method, a candidate is generated according to the so-called Hamiltonian dynamics, which leads to a high acceptance probability and low correlation with the current state by accurately simulating the dynamics of sufficiently long length; see Neal et al. (2011) and Betancourt (2017) for an introduction to HMC. Moreover, the HMC candidates always belong to the crisis event by reflecting the dynamics when the chain hits the boundary of the constraints; see Ruján (1997), Pakman and Paninski (2014), Afshar and Domke (2015), Yi and Doshi-Velez (2017), and Chevallier et al. (2018) for this reflection property of the HMC method. An alternative method to handle the constraints is the GS, in which the chain is updated in each component. Since all the components except the updated one remain fixed, a componentwise update is typically subject to weaker constraints. As long as such componentwise updates are feasible, the GS candidates belong to the crisis event, and the acceptance probability is always 1; see Geweke (1991), Gelfand et al. (1992), and Rodriguez-Yam et al. (2004) for the application of the GS to constrained target distributions, and see Gudmundsson and Hult (2014) and Targino et al. (2015) for applications to estimating risk contributions.
Our findings include efficient MCMC estimators of systemic risk allocations achieved via HMC with reflection and GSs. We assume that the unconditional joint loss density is known, possibly through its marginal densities and copula density. Depending on the supports of the marginal loss distributions and the crisis event, different MCMC methods are applicable. We find that if the marginal loss distributions are one-sided, that is, the supports are bounded from the left, then the crisis event is typically a bounded set and HMC shows good performance. On the other hand, if the marginal losses are two-sided, that is, they have both right and left tails, the crisis event is often unbounded and the GSs perform better, provided that the random number generators of the conditional copulas are available. Based on the samples generated by the MC method, we propose heuristics to determine the parameters of the HMC and GS methods, for which no manual interaction is required. Since, in the MCMC method, the conditional loss distribution of interest is directly simulated, in contrast to MC where rejection is applied based on the unconditional loss distribution, the MCMC method generally outperforms the MC method in terms of the sample size, and thus the standard error. This advantage of MCMC becomes more pronounced as the probability of the crisis event becomes smaller. We demonstrate this efficiency of the MCMC estimators of systemic risk allocations by a series of numerical experiments.
This paper is organized as follows. The general framework of the estimation problem of systemic risk allocations is introduced in Section 2. Our class of systemic risk allocations is proposed in Section 2.1, and their estimation via the MC method is presented in Section 2.2. Section 3 is devoted to MCMC methods for estimating systemic risk allocations. After a brief review of MCMC methods in Section 3.1, we formulate our problem of estimating systemic risk allocations in terms of MCMC in Section 3.2. HMC and GS for constrained target distributions are then investigated in Section 3.3 and Section 3.4, respectively. In Section 4, numerical experiments are conducted, including simulation and empirical studies, and a detailed comparison of MC and our introduced MCMC methods is provided. Section 5 concludes with practical guidance and limitations of the presented MCMC methods. An R script reproducing the numerical experiments is available as Supplementary Material.

2. Systemic Risk Allocations and Their Estimation

In this section, we define a broad class of systemic risk allocations, including Euler allocations, CoVaR, and CoES as special cases. Then, the MC method is described to estimate systemic risk allocations.

2.1. A Class of Systemic Risk Allocations

Let ( Ω , F , P ) be an atomless probability space, and let X 1 , , X d , d 2 be random variables on this space. The random vector X = ( X 1 , , X d ) can be interpreted as losses of a portfolio of size d, or losses of d economic entities in an economy over a fixed time period. Throughout the paper, a positive value of a loss random variable represents a financial loss, and a negative loss is interpreted as a profit. Let F X denote the joint cumulative distribution function (cdf) of X with marginal distributions F 1 , , F d . Assume that F X admits a probability density function (pdf) f X with marginal densities f 1 , , f d . Sklar’s theorem (Nelsen (2006)) allows one to write
F X ( x ) = C ( F 1 ( x 1 ) , , F d ( x d ) ) , x = ( x 1 , , x d ) R d ,
where C : [ 0 , 1 ] d [ 0 , 1 ] is a copula of X . Assuming the density c of the copula C to exist, f X can be written as
f X ( x ) = c ( F 1 ( x 1 ) , , F d ( x d ) ) f 1 ( x 1 ) f d ( x d ) , x R d .
An allocation A = ( A 1 , , A d ) is a map from a random vector X to ( A 1 ( X ) , , A d ( X ) ) R d . The sum j = 1 d A j ( X ) can be understood as the capital required to cover the total loss of the portfolio or the economy. The jth component A j ( X ) , j = 1 , , d is then the contribution of the jth loss to the total capital j = 1 d A j ( X ) . In this paper, we consider the following class of allocations
A ϱ 1 , , ϱ d , C = ( A 1 ϱ 1 , C , , A d ϱ d , C ) , A j ϱ j , C ( X ) = ϱ j ( X j | X C ) ,
where ϱ j is a map from a random variable to R called the jth marginal risk measure for j = 1 , , d , and  C R d is a set called the crisis event. The conditioning set { X C } is simply written as C if there is no confusion. As we now explain, this class of allocations covers well-known allocations as special cases. For a random variable X F , we define the Value-at-Risk (VaR) of X at confidence level α ( 0 , 1 ]  by
VaR α ( X ) : = inf { x R : F ( x ) α } .
Range Value-at-Risk (RVaR) at confidence levels 0 < α 1 < α 2 1 is defined by
RVaR α 1 , α 2 ( X ) = 1 α 2 α 1 α 1 α 2 VaR γ ( X ) d γ ,
and, if it exists, expected shortfall (ES) at confidence level α ( 0 , 1 ) is defined by ES α ( X ) = RVaR α , 1 ( X ) . Note that ES is also known as C(onditional)VaR, T(ail)VaR, A(verage)VaR and C(onditional)T(ail)E(expectation). These risk measures are law-invariant in the sense that they depend only on the distribution of X. Therefore, we sometimes write ϱ ( F ) instead of ϱ ( X ) .
We now define various crisis events and marginal risk measures. A typical form of the crisis event is an intersection of a set of linear constraints
C = m = 1 M h m x v m , h m R d , v m R , m = 1 , , M , M N .
Several important special cases of the crisis event of Form (2) are provided in the following.
Definition 1
(VaR, RVaR, and ES crisis events). For S = j = 1 d X j , the VaR, RVaR and ES crisis events are defined by
C α VaR = { x R d | 1 d x = VaR α ( S ) } , α ( 0 , 1 ) , C α 1 , α 2 RVaR = { x R d | VaR α 1 ( S ) 1 d x VaR α 2 ( S ) } , 0 < α 1 < α 2 1 , C α ES = { x R d | VaR α ( S ) 1 d x } , 0 < α < 1 , α ( 0 , 1 ) ,
respectively, where 1 d is the d-dimensional vector of ones.
Definition 2
(Risk contributions and conditional risk measures). For j { 1 , , d } , we call A j ϱ j , C of
1. 
risk contribution-type if ϱ j = E ;
2. 
CoVaR type if ϱ j = VaR β j for β j ( 0 , 1 ) ;
3. 
CoRVaR type if ϱ j = RVaR β j , 1 , β j , 2 for 0 < β j , 1 < β j , 2 1 ; and
4. 
CoES-type if ϱ j = ES β j for β j ( 0 , 1 ) .
The following examples show that A j ϱ j , C coincides with popular allocations for some specific choices of marginal risk measure and crisis event.
Example 1
(Special cases of A ϱ 1 , , ϱ d , C ).
(1) 
Risk contributions. If the crisis event is chosen to be C α VaR , C α 1 , α 2 RVaR or C α ES , the allocations of the risk contribution type ϱ j = E reduce to the VaR, RVaR, or ES contributions defined by
VaR α ( X , S ) = E [ X | S = VaR α ( S ) ] , RVaR α 1 , α 2 ( X , S ) = E [ X | VaR α 1 ( S ) S VaR α 2 ( S ) ] , ES α ( X , S ) = E [ X | S VaR α ( S ) ] ,
respectively. These results are derived by allocating the total capital VaR α ( S ) , RVaR α 1 , α 2 ( S ) and ES α ( S ) according to the Euler principle; see Tasche (1995). The ES contribution is also called the MES and used as a systemic risk measure; see Acharya et al. (2017).
(2) 
Conditional risk measures. CoVaR and CoES are systemic risk measures defined by
CoVaR α , β = ( X j , S ) = VaR β ( X j | S = VaR α ( S ) ) , CoVaR α , β ( X j , S ) = VaR β ( X j | S VaR α ( S ) ) , CoES α , β = ( X j , S ) = ES β ( X j | S = VaR α ( S ) ) , CoES α , β ( X j , S ) = ES β ( X j | S VaR α ( S ) ) ,
for α , β ( 0 , 1 ) ; see Mainik and Schaanning (2014) and Bernardi et al. (2017). Our CoVaR and CoES-type allocations with crisis events C = C VaR α or C ES α coincide with those defined in the last displayed equations.
Remark 1
(Weighted allocations). For a measurable function w : R d R + : = [ 0 , ) , Furman and Zitikis (2008) proposed the weighted allocation ϱ w ( X ) with the weight function w being defined by ϱ w ( X ) = E [ X w ( X ) ] / E [ w ( X ) ] . By taking an indicator function as weight function w ( x ) = 1 [ x C ] and provided that P ( X C ) > 0 , the weighted allocation coincides with the risk contribution-type systemic allocation A E , , E , C .

2.2. Monte Carlo Estimation of Systemic Risk Allocations

Even if the joint distribution F X of the loss random vector X is known, the conditional distribution of X given X C , denoted by F X | C , is typically too complicated to analytically calculate the systemic risk allocations A ϱ 1 , , ϱ d , C . An alternative approach is to numerically estimate them by the MC method, as is done in Yamai and Yoshiba (2002) and Fan et al. (2012). To this end, assume that one can generate i.i.d. samples from F X . If P ( X C ) > 0 , the MC estimator of A j ϱ j , C , j = 1 , , d is constructed as follows:
(1)
Sample from X : For a sample size N N , generate X ( 1 ) , , X ( N ) ind . F X .
(2)
Estimate the crisis event: If the crisis event C contains unknown quantities, replace them with their estimates based on X ( 1 ) , , X ( N ) . Denote by C ^ the estimated crisis event.
(3)
Sample from the conditional distribution of X given C ^ : Among X ( 1 ) , , X ( N ) , determine X ˜ ( n ) such that X ˜ ( n ) C ^ for all n = 1 , , N .
(4)
Construct the MC estimator: The MC estimate of A j ϱ j , C is ϱ j ( F ^ X ˜ ) where F ^ X ˜ is the empirical cdf (ecdf) of the X ˜ ( n ) ’s.
For an example of (2), if the crisis event is C RVaR α 1 , α 2 = { x R d | VaR α 1 ( S ) 1 d x VaR α 2 ( S ) } , then VaR α 1 ( S ) and VaR α 2 ( S ) are unknown parameters, and thus they are replaced by VaR α 1 ( F ^ S ) and VaR α 2 ( F ^ S ) , where F ^ S is the ecdf of the total loss S ( n ) : = X 1 ( n ) + + X d ( n ) for n = 1 , , N . By the law of large numbers (LLN) and the central limit theorem (CLT), the MC estimator of A ϱ 1 , , ϱ d , C is consistent, and the approximate confidence interval of the true allocation can be constructed based on the asymptotic normality; see Glasserman (2005).
The MC cannot handle VaR crisis events if S admits a pdf, since P ( X C VaR α ) = P ( S = VaR α ( S ) ) = 0 , and thus, no subsample is picked in (3) above. A possible remedy (although the resulting estimator suffers from an inevitable bias) is to replace C α VaR with C α δ , α + δ RVaR for sufficiently small δ > 0 , so that P ( S C α δ , α + δ RVaR ) = 2 δ > 0 .
The main advantage of MC for estimating systemic risk allocations A ϱ 1 , , ϱ d , C is that only a random number generator for F X is required for implementing the method. Furthermore, MC is applicable for any choice of the crisis event C as long as P ( X C ) > 0 . Moreover, the main computational load is simulating F X in (1) above, which is typically not demanding. The disadvantage of the MC method is its inefficiency concerning the rare-event characteristics of ϱ 1 , , ϱ d and C . To see this, consider the case where C = C α 1 , α 2 RVaR and ϱ j = RVaR β 1 , β 2 for α 1 = β 1 = 0.95 and α 2 = β 2 = 0.975 . If the MC sample size is N = 10 5 , there are N × ( α 2 α 1 ) = 2500 subsamples resulting from (3). To estimate RVaR β 1 , β 2 in (4) based on this subsample, only 2500 × ( β 2 β 1 ) = 62.5 samples contribute to computing the estimate, which is generally not enough for statistical inference. This effect of sample size reduction is relaxed if ES and/or ES crisis events are considered, but is more problematic for the VaR crisis event since there is a trade-off concerning reducing bias and MC error when choosing δ ; see Koike and Minami (2019).

3. MCMC Estimation of Systemic Risk Allocations

To overcome the drawback of the MC method for estimating systemic risk allocations, we introduce MCMC methods, which simulate a given distribution by constructing a Markov chain whose stationary distribution is F X | C . In this section, we first briefly review MCMC methods, including the MH algorithm as a major subclass of MCMC methods, and then study how to construct an efficient MCMC estimator for the different choices of crisis events.

3.1. A Brief Review of MCMC

Let E R d be a set and E be a σ -algebra on E. A Markov chain is a sequence of E-valued random variables ( X ( n ) ) n N 0 satisfying the Markov property P ( X ( n + 1 ) A | X ( k ) = x ( k ) , k n ) = P ( X ( n + 1 ) A | X ( n ) = x ( n ) ) for all n 1 , A E , and  x ( 1 ) , , x ( n ) E . A Markov chain is characterized by its stochastic kernel K : E × E [ 0 , 1 ] given by x × A K ( x , A ) : = P ( X ( n + 1 ) A | X ( n ) = x ) . A probability distribution π satisfying π ( A ) = E π ( d x ) K ( x , A ) for any x E and A E is called stationary distribution. Assuming K ( x , · ) has a density k ( x , · ) , the detailed balance condition (also known as the reversibility) with respect to π is given by
π ( x ) k ( x , y ) = π ( y ) k ( y , x ) , x , y E ,
and is known as a sufficient condition for the corresponding kernel K to have the stationary distribution π ; see Chib and Greenberg (1995). MCMC methods simulate a distribution as a sample path of a Markov chain whose stationary distribution π is the desired one. For a given distribution π , also known as target distribution, and a functional ϱ , the quantity of interest ϱ ( π ) is estimated by the MCMC estimator ϱ ( π ^ ) where π ^ is the empirical distribution constructed from a sample path X ( 1 ) , , X ( N ) of the Markov chain whose stationary distribution is π . Under regularity conditions, the MCMC estimator is consistent and asymptotically normal; see Nummelin (2002), Nummelin (2004), and Meyn and Tweedie (2012). Its asymptotic variance can be estimated from ( X ( 1 ) , , X ( N ) ) by, for instance, the batch means estimator; see Jones et al. (2006), Geyer (2011) and Vats et al. (2015) for more details. Consequently, one can construct approximate confidence intervals for the true quantity ϱ ( π ) based on a sample path of the Markov chain.
Since the target distribution π is determined by the problem at hand, the problem is to find the stochastic kernel K having π as the stationary distribution such that the corresponding Markov chain can be easily simulated. One of the most prevalent stochastic kernels is the Metropolis–Hastings (MH) kernel, defined by K ( x , d y ) = k ( x , y ) d y + r ( x ) δ x ( d y ) , where δ x is the Dirac delta function; k ( x , y ) = q ( x , y ) α ( x , y ) ; q : E × E R + is a function called a proposal density such that x q ( x , y ) is measurable for any y E and y q ( x , y ) is a probability density for any x E ;
α ( x , y ) = min π ( y ) q ( y , x ) π ( x ) q ( x , y ) , 1 , if π ( x ) q ( x , y ) > 0 , 0 , otherwise ;
and r ( x ) = 1 E k ( x , y ) d y . It can be shown that the MH kernel has stationary distribution π ; see Tierney (1994). Simulation of the Markov chain with this MH kernel is conducted by the MH algorithm given in Algorithm 1.
Algorithm 1 Metropolis–Hastings (MH) algorithm.
Require: Random number generator of the proposal density q(x,·) for all xE, x(0) ∈ supp(π) and the ratio π(y)/π(x) for x, yE, where π is the density of the stationary distribution.
Input: Sample size N N , proposal density q, and initial value X ( 0 ) = x ( 0 ) .
Output: Sample path X ( 1 ) , , X ( N ) of the Markov chain.
for n : = 0 , , N 1 do
  (1) Generate X ˜ ( n ) q ( X ( n ) , · ) .
  (2) Calculate the acceptance probability
α n : = α ( X ( n ) , X ˜ ( n ) ) = min π ( X ˜ ( n ) ) q ( X ˜ ( n ) , X ( n ) ) π ( X ( n ) ) q ( X ( n ) , X ˜ ( n ) ) , 1 .
  (3) Generate U U ( 0 , 1 ) and set X ( n + 1 ) : = 1 [ U α n ] X ˜ ( n ) + 1 [ U > α n ] X ( n ) .
end for
An advantage of the MCMC method is that a wide variety of distributions can be simulated as a sample path of a Markov chain even if generating i.i.d. samples is not directly feasible. The price to pay is an additional computational cost to calculate the acceptance probability (4), and a possibly higher standard deviation of the estimator ϱ ( π ^ ) compared to the standard deviation of estimators constructed from i.i.d. samples. This attributes to the serial dependence among MCMC samples, which can be seen as follows. Suppose first that the candidate X ˜ ( n ) is rejected (so { U > α n } occurs). Then  X ( n + 1 ) = X ( n ) , and thus, the samples are perfectly dependent. The candidate X ˜ ( n ) is more likely to be accepted if the acceptance probability α n is close to 1. In this case, π ( X ( n ) ) and π ( X ˜ ( n ) ) are expected to be close to each other (otherwise, π ( X ˜ ( n ) ) / π ( X ˜ ( n ) ) and thus α n can be small). Under the continuity of π , X ˜ ( n ) and X ( n ) are expected to be close and thus dependent with each other. An efficient MCMC method is such that the candidate X ˜ ( n ) is sufficiently far from X ( n ) with the probability π ( X ˜ ( n ) ) being as close to π ( X ( n ) ) as possible. The efficiency of MCMC can indirectly be inspected through the acceptance rate (ACR) and the autocorrelation plot (ACP); ACR is the percentage of times a candidate X ˜ is accepted among the N iterations, and ACP is the plot of the autocorrelation function of the generated sample path. An efficient MCMC method shows high ACR and steady decline in ACP; see Chib and Greenberg (1995) and Rosenthal et al. (2011) for details. Ideally, the proposal density q is constructed only based on π , but typically, q is chosen among a parametric family of distributions. For such cases, simplicity of the choice of tuning parameters of q is also important.

3.2. MCMC Formulation for Estimating Systemic Risk Allocations

Numerous choices of proposal densities q are possible to construct an MH kernel. In this subsection, we consider how to construct an efficient MCMC method for estimating systemic risk allocations A ϱ 1 , , ϱ d , C depending on the choice of the crisis event C . Our goal is to directly simulate the conditional distribution X | C by constructing a Markov chain whose stationary distribution is
π ( x ) = f X | X C ( x ) = f X ( x ) P ( X C ) 1 [ x C ] , x E R d ,
provided P ( X C ) > 0 . Samples from this distribution can directly be used to estimate systemic risk allocations with crisis event C and arbitrary marginal risk measures ϱ 1 , , ϱ d . Other potential applications are outlined in Remark 2.
Remark 2
(Gini shortfall allocation). Samples from the conditional distribution F X | C α ES can be used to estimate, for example, the tail-Gini coefficient TGini α ( X j , S ) = 4 1 α Cov ( X j , F S ( S ) | S VaR α ( S ) ) for α ( 0 , 1 ) , and the Gini shortfall allocation (Furman et al. (2017)) GS α ( X j , S ) = E [ X j | S VaR α ( S ) ] + λ · TGini α ( X j , S ) , λ R + more efficiently than by applying the MC method. Another application is to estimate risk allocations derived by optimization, given a constant economic capital; see Laeven and Goovaerts (2004) and Dhaene et al. (2012).
We now construct an MH algorithm with target distribution (5). To this end, we assume that
  • the ratio f X ( y ) / f X ( x ) can be evaluated for any x , y C , and that
  • the support of f X is R d or R + d .
Regarding Assumption 1, the normalization constant of f X and the probability P ( X C ) are not necessary to be known, since they cancel out in the numerator and the denominator of π ( y ) / π ( x ) . In Assumption 2, the loss random vector X refers to the profit&loss (P&L) if supp ( f X ) = R d , and to pure losses if supp ( f X ) = R + d . Note that the case supp ( f X ) = [ c 1 , ] × × [ c d , ] , c 1 , , c d R is essentially included in the case of pure losses as long as the marginal risk measures ϱ 1 , , ϱ d are law invariant and translation invariant, and the crisis event is the set of linear constraints of Form (2). To see this, define X ˜ j = X j c j , j = 1 , , d , X ˜ = ( X ˜ 1 , , X ˜ d ) and c = ( c 1 , , c d ) . Then  supp ( f X ˜ ) = R + d and X | ( X C ) = d X ˜ | ( X ˜ C ˜ ) + c , where C ˜ is the set of linear constraints with parameters h ˜ m = h m and v ˜ m = v m h m c . By law invariance and translation invariance of ϱ 1 , , ϱ d ,
ϱ j ( X j | X C ) = c j + ϱ j ( X ˜ j | X ˜ C ˜ ) , j = 1 , , d .
Therefore, the problem of estimating A ϱ 1 , , ϱ d , C ( X ) reduces to that of estimating A ϱ 1 , , ϱ d , C ˜ ( X ˜ ) for the shifted loss random vector X ˜ (such that supp ( f X ˜ ) = R + d ) and the modified crisis event of the same form.
For the P&L case, the RVaR and ES crisis events are the set of linear constraints of Form (2) with the number of constraints M = 2 and 1, respectively. In the case of pure losses, additional d constraints e j , d x 0 , j = 1 , , d are imposed, where e j , d is the jth d-dimensional unit vector. Therefore, the RVaR and ES crisis events are of Form (2) with M = d + 2 and d + 1 , respectively. For the VaR crisis event, P ( X C ) = 0 , and thus, (5) cannot be properly defined. In this case, the allocation A ϱ 1 , , ϱ d , C VaR depends on the conditional joint distribution X | C α VaR , but is completely determined by its first d : = d 1 variables ( X 1 , , X d ) | C α VaR , since X d | C α VaR = d ( VaR α ( S ) j = 1 d X j ) | C α VaR = d VaR α ( S ) j = 1 d X j | C α VaR . Estimating systemic risk allocations under the VaR crisis event can thus be achieved by simulating the target distribution
π VaR α ( x ) = f X | S = VaR α ( S ) ( x ) = f ( X , S ) ( x , VaR α ( S ) ) f S ( VaR α ( S ) ) = f X ( x , VaR α ( S ) 1 d x ) f S ( VaR α ( S ) ) 1 [ VaR α ( S ) 1 d x supp ( f d ) ] , x R d ,
where X = ( X 1 , , X d ) and the last equation is derived from the linear transformation ( X , S ) X with unit Jacobian. Note that other transformations are also possible; see Betancourt (2012). Under Assumption 1, the ratio π VaR α ( y ) / π VaR α ( x ) can be evaluated and f S ( VaR α ( S ) ) is not required to be known. In the case of pure losses, the target distribution π VaR α is subject to d linear constraints e j , d x 0 , j = 1 , , d and 1 d x VaR α ( S ) , where the first d constraints come from the non-negativity of the losses and the last one is from the indicator in (6). Therefore, the crisis event C VaR for ( X 1 , , X d ) is of Form (2). In the case of P&L, supp ( f d ) = R and VaR α ( S ) 1 d x supp ( f d ) holds for any x R d . Therefore, the target distribution (6) is free from any constraints and the problem reduces to constructing an MCMC method with target distribution π ( x ) f X ( x , VaR α ( S ) 1 d x ) , x R d . In this paper, the P&L case with VaR crisis event is not investigated further, since our focus is the simulation of constrained target distributions; see Koike and Minami (2019) for an MCMC estimation in the P&L case.
MCMC methods to simulate constrained target distributions require careful design of the proposal density q. A simple MCMC method is Metropolis–Hastings with rejection in which the support of the proposal density q may not coincide with that of the target distribution, which is the crisis event C , and a candidate is immediately rejected when it violates the constraints. This construction of MCMC is often inefficient due to a low acceptance probability, especially around the boundary of C . In this case, an efficient MCMC method can be expected only when the probability mass of π is concentrated near the center of C . In the following sections, we introduce two alternative MCMC methods for the constrained target distributions F X | C of interest, the HMC method and the GS. Each of them is applicable and can be efficient for different choices of the crisis event and underlying loss distribution functions F X .

3.3. Estimation with Hamiltonian Monte Carlo

We find that if the HMC method is applicable, it is typically the most preferable method to simulate constrained target distributions because of its efficiency and ease of handling constraints. In Section 3.3.1, we briefly present the HMC method with a reflection for constructing a Markov chain supported on the constrained space. In Section 3.3.2, we propose a heuristic for determining the parameters of the HMC method based on the MC presamples.

3.3.1. Hamiltonian Monte Carlo with Reflection

For the possibly unnormalized target density π , consider the potential energy U ( x ) , kinetic energy K ( p ) , and the Hamiltonian H ( x , p ) defined by
U ( x ) = log π ( x ) , K ( p ) = log f K ( p ) and H ( x , p ) = U ( x ) + K ( p ) ,
with position variable x E , momentum variable p R d , and kinetic energy density f K ( p ) such that f K ( p ) = f K ( p ) . In this paper, the kinetic energy distribution F K is set to be the multivariate standard normal with K ( p ) = 1 2 p p and K ( p ) = p ; other choices of F K are discussed in Appendix B.2. In the HMC method, a Markov chain augmented on the state space E × R d with the stationary distribution π ( x ) f K ( p ) is constructed and the desired samples from π are obtained as the first | E | -dimensional margins. A process ( x ( t ) , p ( t ) ) , t R on E × R d is said to follow the Hamiltonian dynamics if it follows the ordinary differential equation (ODE)
d d t x ( t ) = K ( p ) , d d t p ( t ) = U ( x ) .
Through the Hamiltonian dynamics, the Hamiltonian H and the volume are conserved, that is, d H ( x ( t ) , p ( t ) ) / d t = 0 and the map ( x ( 0 ) , p ( 0 ) ) ( x ( t ) , p ( t ) ) has a unit Jacobian for any t R ; see Neal et al. (2011). Therefore, the value of the joint target density π · f K remains unchanged by the Hamiltonian dynamics, that is,
π ( x ( 0 ) ) f K ( p ( 0 ) ) = exp ( H ( x ( 0 ) , p ( 0 ) ) ) = exp ( H ( x ( t ) , p ( t ) ) ) = π ( x ( t ) ) f K ( p ( t ) ) , t 0 .
In practice, the dynamics (7) are discretized for simulation by, for example, the so-called leapfrog method summarized in Algorithm 2; see Leimkuhler and Reich (2004) for other discretization methods.
Algorithm 2 Leapfrog method for Hamiltonian dynamics.
Input: Current states ( x ( 0 ) , p ( 0 ) ) , stepsize ϵ > 0 , gradients U and K .
Output: Updated position ( x ( ϵ ) , p ( ϵ ) ) .
 (1) p ϵ 2 = p ( 0 ) ϵ 2 U ( x ( 0 ) ) .
 (2) x ( ϵ ) = x ( 0 ) + ϵ K ( p ϵ 2 ) .
 (3) p ( ϵ ) = p ( ϵ / 2 ) + ϵ 2 U ( x ( ϵ ) ) .
Note that the evaluation of U does not require the normalization constant of π to be known, since U = ( π ) / π . By repeating the leapfrog method T times with stepsize ϵ , the Hamiltonian dynamics are approximately simulated with length T ϵ . Due to the discretization error, the Hamiltonian is not exactly preserved, while it is expected to be almost preserved for ϵ which is small enough. The discretization error H ( x ( T ϵ ) , p ( T ϵ ) ) H ( x ( 0 ) , p ( 0 ) ) is called the Hamiltonian error.
All the steps of the HMC method are described in Algorithm 3. In Step (1), the momentum variable is first updated from p ( 0 ) to p , where p follows the kinetic energy distribution F K so that the value of the Hamiltonian H = log ( π · f K ) changes. In Step (3), the current state ( x ( 0 ) , p ) is moved along the level curve of H ( x ( 0 ) , p ) by simulating the Hamiltonian dynamics.
Algorithm 3 Hamiltonian Monte Carlo to simulate π .
Require: Random number generator of FK, x(0) ∈ supp(π), π(y)/π(x), x, yE and FK(p)/FK(p), p, p ∈ ℝd.
Input: Sample size N N , kinetic energy density f K , target density π , gradients of the potential and kinetic energies U and K , stepsize ϵ > 0 , integration time T N and initial position X ( 0 ) = x ( 0 ) .
Output: Sample path X ( 1 ) , , X ( N ) of the Markov chain.
for n : = 0 , , N 1 do
  (1) Generate p ( n ) F K .
  (2) Set ( X ˜ ( n ) , p ˜ ( n ) ) = ( X ( n ) , p ( n ) ) .
  (3) for t : = 1 , , T ,
( X ˜ ( n + t / T ) , p ˜ ( n + t / T ) ) = Leapfrog ( X ˜ ( n + ( t 1 ) / T ) , p ˜ ( n + ( t 1 ) / T ) , ϵ , U , K ) .
   end for
  (4) p ˜ ( n + 1 ) = p ( n + 1 ) .
  (5) Calculate α n = min π ( X ˜ ( n + 1 ) ) f K ( p ˜ ( n + 1 ) ) π ( X ( n ) ) f K ( p ( n ) ) , 1 .
  (6) Set X ( n + 1 ) : = 1 [ U α n ] X ˜ ( n + 1 ) + 1 [ U > α n ] X ( n ) for U U ( 0 , 1 ) .
end for
By flipping the momentum in Step (4), the HMC method is shown to be reversible w.r.t. π (c.f. (3)) and thus to have the stationary distribution π ; see Neal et al. (2011) for details. Furthermore, by the conservation property of the Hamiltonian dynamics, the acceptance probability in Step (5) is expected to be close to 1. Moreover, by taking T as sufficiently large, the candidate X ˜ ( n + 1 ) is expected to be sufficiently decorrelated from the current position X ( n ) . Consequently, the resulting Markov chain is expected to be efficient.
The remaining challenge for applying the HMC method to our problem of estimating systemic risk allocations is how to handle the constraint C . As we have seen in Section 2.1 and Section 3.2, C is assumed to be an intersection of linear constraints with parameters ( h m , v m ) , m = 1 , , M describing hyperplanes. Following the ordinary leapfrog method, a candidate is immediately rejected when the trajectory of the Hamiltonian dynamics penetrates one of these hyperplanes. To avoid it, we modify the leapfrog method according to the reflection technique introduced in Afshar and Domke (2015) and Chevallier et al. (2018). As a result, the trajectory is reflected when it hits a hyperplane and the Markov chain moves within the constrained space with probability one. Details of the HMC method with the reflection for our application are described in Appendix A.

3.3.2. Choice of Parameters for HMC

HMC requires as input two parameters, the stepsize ϵ , and the integration time T. As we now explain, neither of them should be chosen too large nor too small. Since the stepsize ϵ controls the accuracy of the simulation of the Hamiltonian dynamics, ϵ needs to be small enough to approximately conserve the Hamiltonian; otherwise, the acceptance probability can be much smaller than 1. On the other hand, an ϵ which is too small requires the integration time T to be large enough for the trajectory to reach a farther distance, which is computationally costly. Next, the integration time T needs to be large enough to decorrelate the candidate state with the current state. Meanwhile, the trajectory of the Hamiltonian dynamics may make a U-turn and come back to the starting point if the integration time T is too long; see Neal et al. (2011) for an illustration of this phenomenon.
A notable characteristic of our problem of estimating systemic risk allocations is that the MC sample from the target distribution π is available but its sample size may not be sufficient for statistical inference, and, in the case of the VaR crisis event, the samples only approximately follow the target distribution. We utilize the information of this MC presample to build a heuristic for determining the parameters ( ϵ , T ) ; see Algorithm 4.
In this heuristic, the initial stepsize is set to be ϵ = c ϵ d 1 / 4 for some constant c ϵ > 0 , say, c ϵ = 1 . This scale was derived in Beskos et al. (2010) and Beskos et al. (2013) under certain assumptions on the target distribution. We determine ϵ through the relationship with the acceptance probability. In Step (2-2-2-1) of Algorithm 4, multiple trajectories are simulated, starting from each MC presample with the current stepsize ϵ . In the next Step (2-2-2-2), we monitor the acceptance probability and the distance between the starting and ending points while extending the trajectories. Based on the asymptotic optimal acceptance probability 0.65 (c.f. Gupta et al. (1990) and Betancourt et al. (2014)) as d , we set the target acceptance probability as
α ̲ = 1 + ( d 1 ) × 0.65 d ( 0.65 , 1 ] .
The stepsize is gradually decreased in Step (2-1) of Algorithm 4 until the minimum acceptance probability calculated in Step (2-3) exceeds α ̲ . To prevent the trajectory from a U-turn, in Step (2-2-2-3), each trajectory is immediately stopped when the distance begins to decrease. The resulting integration time is set to be the average of these turning points, as seen in Step (3). Note that other termination conditions of extending trajectories are possible; see Hoffman and Gelman (2014) and Betancourt (2016).
Algorithm 4 Heuristic for determining the stepsize ϵ and integration time T.
Input: MC presample X 1 ( 0 ) , , X N 0 ( 0 ) , gradients U and K , target acceptance probability α ̲ , initial constant c ϵ > 0 and the maximum integration time T max ( c ϵ = 1 and T max = 1000 are set as default values).
Output: Stepsize ϵ and integration time T.
 (1) Set α min = 0 and ϵ = c ϵ d 1 / 4 .
 (2) while α min < α ̲
  (2-1) Set ϵ = ϵ / 2 .
  (2-2) for n : = 1 , , N 0
   (2-2-1) Generate p n ( 0 ) F K .
   (2-2-2) for t : = 1 , , T max
    (2-2-2-1) Set Z n ( t ) = Leapfrog ( Z n ( t 1 ) , ϵ , U , K ) for Z n ( t 1 ) = ( X n ( t 1 ) , p n ( t 1 ) ) .
    (2-2-2-2) Calculate
α n , t = α ( Z n ( t 1 ) , Z n ( t ) ) and Δ t = | | X n ( t ) X n ( 0 ) | | | | X n ( t 1 ) X n ( 0 ) | | .
    (2-2-2-3) if Δ t < 0 and Δ t 1 > 0 , break and set T n * = t 1 .
    end for
   end for
  (2-3) Compute α min = min ( α n , t | t = 1 , 2 , , T n * , n = 1 , , N 0 ) .
end while
(3) Set T = 1 N 0 n = 1 N 0 T n * .
At the end of this section, we briefly revisit the choice of the kinetic energy distribution F K , which is taken to be a multivariate standard normal throughout this work. As discussed in Neal et al. (2011), applying the HMC method with target distribution π and kinetic energy distribution N ( 0 , Σ 1 ) is equivalent to applying HMC with the standardized target distribution x π ( L x ) and F K = N ( 0 , I ) , where L is the Cholesky factor of Σ such that Σ = L L . By taking Σ to be the covariance matrix of π , the standardized target distribution becomes uncorrelated with unit variances. In our problem, the sample covariance matrix Σ ^ = L ^ L ^ calculated based on the MC presample is used alternatively. The new target distribution π ˜ ( y ) = π ( L ^ y ) | L ^ | where | L ^ | denotes the Jacobian of L ^ , is almost uncorrelated with unit variances, and thus the standard normal kinetic energy fits well; see Livingstone et al. (2019). If the crisis event consists of the set of linear constraints ( h m , v m ) , m = 1 , , M , then the standardized target density is also subject to the set of linear constraints ( L ^ h m , v m ) , m = 1 , , M . Since the ratio f X ( L ^ y ) / f X ( L ^ x ) can still be evaluated under Assumption 1, we conclude that the problem remains unchanged after standardization.
Theoretical results of the HMC method with normal kinetic energy are available only when C is bounded (Cances et al. (2007) and Chevallier et al. (2018)), or when C is unbounded and the tail of π is roughly as light as that of the normal distribution (Livingstone et al. (2016) and Durmus et al. (2017)). Boundedness of C holds for VaR and RVaR crisis events with pure losses; see Koike and Minami (2019). As is discussed in this paper, convergence results of MCMC estimators are accessible when the density of the underlying joint loss distribution is bounded from above on C , which is typically the case when the underlying copula does not admit lower tail dependence. For other cases where C is unbounded or the density explodes on C , no convergence results are available. Potential remedies for the HMC method to deal with heavy-tailed target distributions are discussed in Appendix B.2.

3.4. Estimation with Gibbs Sampler

As discussed in Section 3.3.2, applying HMC methods to heavy-tailed target distributions on unbounded crisis events is not theoretically supported. To deal with this case, we introduce the GS in this section.

3.4.1. True Gibbs Sampler for Estimating Systemic Risk Allocations

The GS is a special case of the MH method in which the proposal density q is completely determined by the target density π via
q G S ( x , y ) = i = ( i 1 , , i d ) I d p i π ( y i 1 | x i 1 ) π ( y i 2 | y i 1 , x ( i 1 , i 2 ) ) π ( y i d | y i d ) ,
where x ( j 1 , , j l ) is the ( d l ) -dimensional vector that excludes the components j 1 , , j l from x , π ( x j | x j ) = π j | j ( x j | x j ) is the conditional density of the jth variable of π given all the other components, I d { 1 , , d } d is the so-called index set, and ( p i [ 0 , 1 ] , i I d ) is the index probability distribution such that i I d p i = 1 . For this choice of q, the acceptance probability is always equal to 1; see Johnson (2009). The GS is called deterministic scan (DSGS) if I d = { ( 1 , , d ) } and p ( 1 , , d ) = 1 . When the index set is the set of permutations of ( 1 , , d ) , the GS is called random permulation (RPGS). Finally, the random scan GS (RSGS) has the proposal (8) with I d = { 1 , , d } d and p ( i 1 , , i d ) = p i 1 p i d with probabilities ( p 1 , , p d ) ( 0 , 1 ) d such that j = 1 d p j = 1 . These three GSs can be shown to have π as stationary distribution; see Johnson (2009).
Provided that the full conditional distributions π j | j , j = 1 , , d can be simulated, the proposal distribution (8) can be simulated by first selecting an index i I d with probability p i and then replacing the jth component of the current state with a sample from π j | j sequentially for j = i 1 , , i d . The main advantage of the GS is that the tails of π are naturally incorporated via full conditional distributions, and thus the MCMC method is expected to be efficient even if π is heavy-tailed. On the other hand, the applicability of the GS is limited to target distributions such that π j | j is available. Moreover, fast simulators of π j | j , j = 1 , , d , are required, since the computational time linearly increases w.r.t. the dimension d.
In our problem of estimating systemic risk allocations, we find that the GS is applicable when the crisis event is of the form
C = { x R d or R + d | v 1 h x v 2 } , v 1 , v 2 R { ± } , h = ( h 1 , , h d ) R d { 0 d } .
The RVaR crisis event is obviously a special case of (9), and the ES crisis event is included as a limiting case for v 2 . Furthermore, the full conditional copulas of the underlying joint loss distribution and their inverses are required to be known as we now explain. Consider the target density π = f X | v 1 h X v 2 . For its jth full conditional density π j | j ( x j | x j ) , notice that
{ v 1 h X v 2 , X j = x j } = v 1 h j x j h j X j v 2 h j x j h j , X j = x j
and thus, for v i , j ( x j ) = ( v i h j x j ) / h j , i = 1 , 2 , we obtain the cdf of π j | j as
F X j | ( v 1 h X v 2 , X j = x j ) ( x j ) = F X j | X j = x j ( x j ) F X j | X j = x j ( v 1 , j ( x j ) ) F X j | X j = x j ( v 2 , j ( x j ) ) F X j | X j = x j ( v 1 , j ( x j ) )
for v 1 , j ( x j ) x j v 2 , j ( x j ) . Denoting the denominator of (10) by Δ j ( x j ) , we obtain the quantile function
F X j | ( v 1 h X v 2 , X j = x j ) 1 ( u ) = F X j | X j = x j 1 Δ j ( x j ) · u + F X j | X j = x j ( v 1 , j ( x j ) ) .
Therefore, if F X j | X j = x j and its quantile function are available, one can simulate the full conditional target densities π j | j with the inversion method; see Devroye (1985). Availability of F X j | X j = x j and its inverse typically depends on the copula of X . By Sklar’s theorem (1), the jth full conditional distribution of F X can be written as
F X j | X j = x j ( x j ) = C j | j ( F j ( x j ) | F j ( x j ) ) ,
where F ( j 1 , , j l ) ( x ( j 1 , , j l ) ) = ( F j 1 ( x j 1 ) , , F j l ( x j l ) ) , ( j 1 , , j l ) = { 1 , , d } ( j 1 , , j l ) and C j | j is the jth full conditional copula defined by
C j | j ( u j | u j ) = P ( U j u j | U j = u j ) = D j C ( u ) D j C ( u 1 , , u j 1 , 1 , u j + 1 , , u d ) ,
where D denotes the operator of partial derivatives with respect to the components given as subscripts and U C . Assuming the full conditional copula C j | j and its inverse C j | j 1 are available, one can simulate X ˜ j π j | j via
U U ( 0 , 1 ) , U ˜ = U + ( 1 U ) C j | j ( F j ( v 1 ( x j ) | F j ( x j ) ) , X ˜ j = F j 1 C j | j 1 ( U ˜ | F j ( x j ) ) .
Examples of copulas for which the full conditional distributions and their inverses are available include normal, Student t, and Clayton copulas; see Cambou et al. (2017). In this case, the GS is also applicable to the corresponding survival ( π -rotated) copula C ^ , since
C ^ j | j ( u ) = 1 C j | j ( 1 u j | 1 d u j ) , C ^ j | j 1 ( u ) = 1 C j | j 1 ( 1 u j | 1 d u j ) , j = 1 , , d ,
by the relationship U ˜ = 1 U C ^ for U C . In a similar way, one can also obtain full conditional copulas and their inverses for other rotated copulas; see Hofert et al. (2018) Section 3.4.1 for rotated copulas.
In the end, we remark that even if the full conditional distributions and their inverses are not available, π j | j can be simulated by, for example, the acceptance and rejection method, or even the MH algorithm; see Appendix B.3.

3.4.2. Choice of Parameters for GS

As discussed in Section 3.3.2, we use information from the MC presamples to determine the parameters of the Gibbs kernel (8). Note that standardization of the variables as applied in the HMC method in Section 3.3.2 is not available for the GS, since the latter changes the underlying joint loss distribution, and since the copula after rotating variables is generally not accessible, except for in the elliptical case; see Christen et al. (2017). Among the presented variants of GSs, we adopt RSGS, since determining d probabilities ( p 1 , , p d ) is relatively easy, whereas RPGS requires d ! probabilities to be determined. To this end, we consider the RSGS with the parameters ( p 1 , , p d ) determined by a heuristic described in Algorithm 5.
Algorithm 5 Random scan Gibbs sampler (RSGS) with heuristic to determine ( p 1 , , p d ) .
Require: Random number generator of πj|−j and x(0) ∈ supp(π).
Input: MC presample X ˜ 1 ( 0 ) , , X ˜ N 0 ( 0 ) , sample size N N , initial state x ( 0 ) , sample size of the pre-run N pre and the target autocorrelation ρ ( N pre = 100 and ρ = 0.15 are set as default values).
Output:N sample path X ( 1 ) , , X ( N ) of the Markov chain.
 (1) Compute the sample covariance matrix Σ ^ based on X ˜ 1 ( 0 ) , , X ˜ N 0 ( 0 ) .
 (2) Set p j Σ ^ j , j Σ ^ j , j Σ ^ j , j 1 Σ ^ j , j and X ( 0 ) = X pre ( 0 ) = x ( 0 ) .
 (3) for n : = 1 , , N pre
  (3-1) Generate J = j with probability p j .
  (3-2) Update X pre , J ( n ) π J | J ( · | X pre ( n 1 ) ) and X pre , J ( n ) = X pre , J ( n 1 ) .
  end for
 (4) Set
T = argmin h N 0 estimated autocorrelations of X pre ( 1 ) , , X pre ( N pre ) with lag h ρ .
 (5) for n : = 1 , , N , t : = 1 , , T
  (5-1) Generate J = j with probability p j .
  (5-2) Update X J ( n 1 + t / T ) π J | J ( · | X ( n 1 + ( t 1 ) / T ) ) and X J ( n 1 + t / T ) = X J ( n 1 + ( t 1 ) / T ) .
  end for
The RSGS kernel is simulated in Steps (3) and (5) of Algorithm 5. To determine the selection probabilities p 1 , , p d , consider a one-step update of the RSGS from X ( n ) to X ( n + 1 ) with X ( n ) π and the one-step kernel
K RSGS ( x , y ) = j = 1 d p j π j | j ( y j | x j ) 1 [ y j = x j ] .
Liu et al. (1995, Lemma 3) implies that
Cov ( X j ( n ) , X j ( n + 1 ) ) = i = 1 d p i E [ E [ X j | X i ] ] = i = 1 d p i { m j ( 2 ) E [ Var ( X j | X i ) ] } i = 1 d p i E [ Var ( X j | X i ) ] ) ,
where m j ( k ) is the kth moment of π j .
For the objective function j = 1 d Cov ( X j ( n ) , X j ( n + 1 ) ) , its minimizer ( p 1 * , , p d * ) under the constraint j = 1 d p j = 1 satisfies
p j * E [ Var ( X j | X j ) ] .
While this optimizer can be computed based on the MC presamples, we observed that its stable estimation is as computationally demanding as estimating the risk allocations themselves. Alternatively, we calculate (11) under the assumption that π follows an elliptical distribution. Under this assumption, (11) is given by
p j Σ j , j Σ j , j Σ j , j 1 Σ j , j ,
where Σ is the covariance matrix of π and Σ J 1 , J 2 , J 1 , J 2 { 1 , , d } is the submatrix of Σ with indices in J 1 × J 2 . As seen in Step (2) of Algorithm 5, Σ is replaced by its estimate based on the MC presamples.
As shown in Christen et al. (2017), Gibbs samplers require a large number of iterations to lower the serial correlation when the target distribution has strong dependence. To reduce serial correlations, we take every Tth sample in Step (5-2), where T N is called the thinning interval of times. Note that we use the same notation T as that of the integration time in HMC, since they both represent a repetition time of some single step. Based on the preliminary run with length N pre in Step (3) in Algorithm 5, T is determined as the smallest lag h such that the marginal autocorrelations with lag h are all smaller than the target autocorrelation ρ ; see Step (4) in Algorithm 5.

4. Numerical Experiments

In this section, we demonstrate the performance of the MCMC methods for estimating systemic risk allocations by a series of numerical experiments. We first conduct a simulation study in which true allocations or their partial information are available. Then, we perform an empirical study to demonstrate that our MCMC methods are applicable to a more practical setup. Finally, we make more detailed comparisons between the MC and MCMC methods in various setups. All experiments were run on a MacBook Air with 1.4 GHz Intel Core i5 processor and 4 GB 1600 MHz of DDR3 RAM.

4.1. Simulation Study

In this simulation study, we compare the estimates and standard errors of the MC and MCMC methods under the low-dimensional risk models described in Section 4.1.1. The results and discussions are summarized in Section 4.1.2.

4.1.1. Model Description

We consider the following three-dimensional loss distributions:
(M1)
generalized Pareto distributions (GPDs) with parameters ( ξ j , β j ) = ( 0.3 , 1 ) and survival Clayton copula with parameter θ = 2 so that Kendall’s tau equals τ = θ / ( θ + 2 ) = 0.5 ;
(M2)
multivariate Student t distribution with ν = 5 degrees of freedom, location vector 0 , and dispersion matrix Σ = ( ρ i , j ) , where ρ j , j = 1 and ρ i , j = | i j | / d for i , j = 1 , , d , i j .
Since the marginals are homogeneous and the copula is exchangeable, the systemic risk allocations under the loss distribution (M1) are all equal, provided that the crisis event is invariant under the permutation of the variables. For the loss distribution (M2), by ellipticality of the joint distribution, analytical formulas of risk contribution-type systemic risk allocations are available; see McNeil et al. (2015) Corollary 8.43. The parameters of the distributions (M1) and (M2) take into account the stylized facts that the loss distribution is heavy-tailed and extreme losses are positively dependent.
We consider the VaR, RVaR, and ES crisis events with confidence levels α VaR = 0.99 , ( α 1 RVaR , α 2 RVaR ) = ( 0.975 , 0.99 ) and α ES = 0.99 , respectively. For each crisis event, the risk contribution, VaR, RVaR, and ES-type systemic risk allocations are estimated by the MC and MCMC methods, where the parameters of the marginal risk measures VaR, RVaR, and ES are set to be β VaR = 0.99 , ( β 1 RVaR , β 2 RVaR ) = ( 0.975 , 0.99 ) and β ES = 0.99 , respectively.
We first conduct the MC simulation for the distributions (M1) and (M2). For the VaR crisis event, the modified event C mod = { VaR α δ ( S ) 1 d x VaR α + δ ( S ) } with δ = 0.001 is used to ensure that P ( X C mod ) > 0 . Based on these MC presamples, the Markov chains are constructed as described in Section 3.3 and Section 3.4. For the MCMC method, (M1) is the case of pure losses and (M2) is the case of P&L. Therefore, the HMC method is applied to the distribution (M1) for the VaR and RVaR crisis events, the GS is applied to (M1) for the ES crisis event and the GS is applied to the distribution (M2) for the RVaR and ES crisis events. The target distribution of (M2) with VaR constraint is free from constraints and was already investigated in Koike and Minami (2019); we thus omit this case and consider the five remaining cases.
Note that 99.8% of the MC samples from the unconditional distribution are discarded for the VaR crisis event and a further 97.5% of them are wasted to estimate the RVaR contributions. Therefore, 1 / ( 0.002 × 0.025 ) = 10 5 / 5 = 20,000 MC samples are required to obtain one MC sample from the conditional distribution. Taking this into account, the sample size of the MC estimator is set to be N MC = 10 5 . The sample size of the MCMC estimators is free from such constraints and thus is chosen to be N MCMC = 10 4 . Initial values x 0 for the MCMC methods are taken as the mean vector calculated from the MC samples. Biases are computed only for the contribution-type allocations in the distribution (M2) since the true values are available in this case. For all the five cases, the MC and the MCMC standard errors are computed according to Glasserman (2013) Chapter 1, for MC, and Jones et al. (2006) for MCMC. Asymptotic variances of the MCMC estimators are estimated by the batch means estimator with batch length L N : = N 1 2 = 100 and batch size B N : = N / L N = 100 . The results are summarized in Table 1 and Table 2.

4.1.2. Results and Discussions

Since a fast random number generators are available for the joint loss distributions (M1) and (M2), the MC estimators are computed almost instantly. On the other hand, the MCMC methods cost around 1.5 min for simulating the N = 10 4 MCMC samples, as reported in Table 1 and Table 2. For the HMC method, the main computational cost consists of calculating gradients N × T times for the leapfrog method, and calculating the ratio of target densities N times in the acceptance/rejection step, where N is the length of the sample path and T is the integration time. For the GS, simulating an N-sample path requires N × T × d random numbers from the full conditional distributions, where T here is the thinning interval of times. Therefore, the computational time of the GS linearly increases w.r.t. the dimension d, which can become prohibitive for the GS in high dimensions. To save computational time, MCMC methods generally require careful implementations of calculating the gradients and the ratio of the target densities for HMC, and of simulating the full conditional distributions for GS.
Next, we inspect the performance of the HMC and GS methods. We observed that the autocorrelations of all sample paths steadily decreased below 0.1 if lags were larger than 15. Together with the high ACRs, we conclude that the Markov chains can be considered to be converged. According to the heuristic in Algorithm 4, the stepsize and integration time for the HMC method are selected to be ( ϵ , T ) = ( 0.210 , 12 ) in Case (I) and ( ϵ , T ) = ( 0.095 , 13 ) in Case (II). As indicated by the small Hamiltonian errors in Figure 1, the acceptance rates in both cases are quite close to 1.
For the GS, the thinning interval of times T and the selection probability p are determined as T = 12 and p = ( 0.221 , 0.362 , 0.416 ) in Case (III), T = 10 and p = ( 0.330 , 0.348 , 0.321 ) in Case (IV) and T = 4 and p = ( 0.241 , 0.503 , 0.255 ) in Case (V). For biases of the estimators, observe that in all cases ((I) to (V)), the estimates of the MC method and the MCMC method are close to each other. In Cases (I), (II), and (III), the true allocations are the homogeneous allocations, whereas their exact values are not known. From the estimates in Table 1 and Table 2, the MCMC estimates are, on average, more equally allocated compared to those of the MC method, especially in Case (III) where heavy-tailedness may lead to quite slow convergence rates of the MC method. Therefore, lower biases of the MCMC estimators are obtained, compared to those of the MC estimators. In the case of risk contributions in Case (IV) and (V), exact biases are computed based on ellipticality, and they show that the GS estimator has a smaller bias than the one of the MC estimator.
Although the MC sample size is 10 times larger than that of the MCMC method, the standard error of the latter is, in most cases, smaller than the MC standard error. This improvement becomes larger as the probability of the crisis event becomes smaller. The largest improvement is observed in Case (I) with the VaR crisis event, and the smallest one is in Cases (III) and (V) with the ES crisis event. MCMC estimates of the risk contribution-type allocations have consistently smaller standard errors than the MC ones. For the RVaR, VaR, and ES-type allocations, the improvement of standard error varies according to the loss models and the crisis event. A notable improvement is observed for ES-type allocation in Case (III), although a stable statistical inference is challenging due to the heavy-tailedness of the target distribution.
Overall, the simulation study shows that the MCMC estimators outperform the MC estimators due to the increased effective sample size and its insusceptibility to the probability of the crisis event. The MCMC estimators are especially recommended when the probability of the crisis event is too small for the MC method to sufficiently simulate many samples for a meaningful statistical analysis.
Remark 3
(Joint loss distributions with negative dependence in the tail). In the above simulation study, we only considered joint loss distributions with positive dependence. Under the existence of positive dependence, the target density f X | v α S v β puts more probability mass around its mean, and the probability decays as the point moves away from the mean, since positive dependence among X 1 , , X d prevents them from going in opposite directions (i.e., one component increases and another one decreases) under the sum constraint; see Koike and Minami (2019) for details. This phenomenon leads to the target distributions being more centered and elliptical, which in turn facilitates efficient moves of Markov chains. Although it may not be realistic, joint loss distributions with negative dependence in the tail are also possible. In this case, the target distribution has more variance, heavy tails, and is even multimodal, since two components can move in opposite directions under the sum constraint. For such cases, constructing efficient MCMC methods becomes more challenging; see Lan et al. (2014) for a remedy for multimodal target distributions with Riemannian manifold HMC.

4.2. Empirical Study

In this section, we illustrate our suggested MCMC methods for estimating risk allocations from insurance company indemnity claims. The dataset consists of 1500 liability claims provided by the Insurance Services Office. Each claim contains an indemnity payment X 1 and an allocated loss adjustment expense (ALAE) X 2 ; see Hogg and Klugman (2009) for a description. The joint distribution of losses and expenses is studied, for example, in Frees and Valdez (1998) and Klugman and Parsa (1999). Based on Frees and Valdez (1998), we adopt the following parametric model:
(M3) univariate marginals are X 1 Par ( λ 1 , θ 1 ) and X 2 Par ( λ 2 , θ 2 ) with ( λ 1 , θ 1 ) = ( 14 , 036 , 1.122 ) and ( λ 2 , θ 2 ) = ( 14 , 219 , 2.118 ) , and the copula is the survival Clayton copula with parameter θ = 0.512 (which corresponds to Spearman’s rho ρ S = 0.310 ).
Note that in the loss distribution (M3), the Gumbel copula used in Frees and Valdez (1998) is replaced by the survival Clayton copula, since both of them have the same type of tail dependence and the latter possesses more computationally tractable derivatives. The parameter of the survival Clayton copula is determined so that it reaches the same Spearman’s rho observed in Frees and Valdez (1998). Figure 2 illustrates the data and samples from the distribution (M3). Our goal is to calculate the VaR, RVaR, and ES-type allocations with VaR, RVaR, and ES crisis events for the same confidence levels as in Section 4.1.1. We apply the HMC method to all three crisis events since, due to the infinite and finite variances of X 1 and X 2 , respectively, the optimal selection probability of the second variable calculated in Step 2 of Algorithm 5 is quite close to 0, and thus the GS did not perform well. The simulated HMC samples are illustrated in Figure 2. The results of estimating the systemic risk allocations are summarized in Table 3.
The HMC samples shown in Figure 2 indicate that the conditional distributions of interest are successfully simulated from the desired regions. As displayed in Figure 3, the Hamiltonian errors of all three HMC methods are sufficiently small, which led to the high ACRs of 0.997 , 0.986 , and 0.995 , as listed in Table 3. We also observed that autocorrelations of all sample paths steadily decreased below 0.1 if lags were larger than 80. Together with the high ACRs, we conclude that the Markov chains can be considered to be converged. Due to the heavy-tailedness of the target distribution in the case of the ES crisis event, the stepsize is very small and the integration time is very large compared to the former two cases of the VaR and RVaR crisis events. As a result, the HMC algorithm in this case has a long run time.
The estimates of the MC and HMC methods are close in all cases, except Case (III). In Case (III), the HMC estimates are smaller than the MC ones in almost all cases. Based on the much smaller standard errors of HMC, one could infer that the MC estimates are likely overestimating the allocations due to a small number of extremely large losses, although the corresponding conditional distribution is extremely heavy-tailed, and thus no estimation method might be reliable. In terms of the standard error, the estimation of systemic risk allocations by the HMC method were improved in Cases (I) and (III) compared to that of the MC method; the MC standard errors are slightly smaller than those of HMC in Case (II). All results considered, we conclude from this empirical study that the MCMC estimators outperform the MC estimators in terms of standard error. On the other hand, as indicated by the theory of HMC with normal kinetic energy, the HMC method is not recommended for heavy-tailed target distributions due to the long computational time caused by a small stepsize and large integration time determined by Algorithm 5.

4.3. Detailed Comparison of MCMC with MC

In the previous numerical experiments, we fixed the dimensions of the portfolios and confidence levels of the crisis events. Comparing the MC and MCMC methods after balancing against computational time might be more reasonable, although one should keep in mind that run time depends on various external factors, such as the implementation, hardware, workload, programming language, or compiler options (and our implementation was not optimized for any of these factors). In this section, we compare the MC and MCMC methods with different dimensions, confidence levels, and parameters of the HMC methods in terms of bias, standard error, and the mean squared error (MSE), adjusted by run time.
In this experiment, we fix the sample size of the MC and MCMC methods as N MC = N MCMC = 10 4 . In addition, we assume X t ν ( 0 , P ) , that is, the joint loss follows the multivariate Student t distribution with ν = 6 degrees of freedom, location vector 0 , and dispersion matrix P, which is the correlation matrix with all off-diagonal entries equal to 1 / 12 . The dimension d of the loss portfolio will vary for comparison. We consider only risk contribution-type systemic risk allocations under VaR, RVaR, and ES crisis events, as true values of these allocations are available to compare against; see McNeil et al. (2015), Corollary 8.43. If b and σ denote the bias and standard deviation of the MC or MCMC estimator and S the run time, then (under the assumption that run time linearly increases by sample size) we define the time-adjusted MSEs by
MSE MC = b MC 2 + σ MC 2 S MCMC S MC × N MCMC and MSE MCMC = b MCMC 2 + σ MCMC 2 N MCMC .
We can then compare the MC and MCMC estimators in terms of bias, standard error, and time-adjusted MSE under the following three scenarios:
(A)
VaR 0.99 , RVaR 0.95 , 0.99 , and ES 0.99 contributions are estimated by the MC, HMC, and GS methods for dimensions d { 4 , 6 , 8 , 10 } . Note that the GS is applied only to RVaR and ES contributions, not to VaR contributions (same in the other scenarios).
(B)
For d = 5 , VaR α VaR , RVaR α 1 RVaR , α 2 RVaR and ES α ES contributions are estimated by the MC, HMC, and GS methods for confidence levels α VaR { 0.9 , 0.99 , 0.999 , 0.9999 } , ( α 1 RVaR , α 2 RVaR ) { ( 0.9 , 0.9999 ) , ( 0.9 , 0.99 ) , ( 0.99 , 0.999 ) , ( 0.999 , 0.9999 ) } and α ES { 0.9 , 0.99 , 0.999 , 0.9999 } .
(C)
For d = 5 , VaR 0.9 , RVaR 0.9 , 0.99 and ES 0.9 contributions are estimated by the MC and HMC methods with the parameters ( ϵ opt , T opt ) (determined by Algorithm 4) and ( ϵ , T ) { ( 10 ϵ opt , 2 T opt ) , ( 10 ϵ opt , T opt / 2 ) , ( ϵ opt / 10 , 2 T opt ) , ( ϵ opt / 10 , T opt / 2 ) } .
In the MC method, the modified VaR contribution E [ X | C α δ , α + δ RVaR ] with δ = 0.01 is computed. Moreover, if the size of the conditional sample for estimating RVaR and ES contributions is less than 100, then the lower confidence level of the crisis event is subtracted by 0.01, so that at least 100 MC presamples are guaranteed. For the sample paths of the MCMC methods, ACR, ACP, and Hamiltonian errors for the HMC methods were inspected and the convergences of the chains were checked, as in Section 4.1 and Section 4.2.
The results of the comparisons of (A), (B), and (C) are summarized in Figure 4, Figure 5 and Figure 6. In Figure 4, the performance of the MC, HMC, and GS estimators is roughly similar across dimensions from 4 to 10. For all crisis events, the HMC and GS estimators outperform MC in terms of bias, standard error, and time-adjusted MSE. From (A5) and (A8), standard errors of the GS estimators are slightly higher than those of the HMC ones, which result in slightly improved performance of the HMC estimator over the GS in terms of MSE. In Figure 5, bias, standard error, and MSE of the MC estimator tend to increase as the probability of the conditioning set decreases. This is simply because the size of the conditional samples in the MC method decreases proportionally to the probability of the crisis event. On the other hand, the HMC and GS estimators provide a stably better performance than MC since such sample size reduction does not occur. As seen in (B4) to (B9) in the cases of RVaR 0.999 , 0.9999 and ES 0.9999 , however, if the probability of the conditioning event is too small and/or the distribution of the MC presample is too different from the original conditional distribution of interest, then the parameters of the HMC method determined by Algorithm 4 can be entirely different from the optimal, which leads to a poor performance of the HMC method, as we will see in the next scenario (C). In Figure 6, the HMC method with optimally determined parameters from Algorithm 4 is compared to non-optimal parameter choices. First, the optimal HMC estimator outperforms MC in terms of bias, standard error, and time-adjusted MSE. On the other hand, from the plots in Figure 6, we see that some of the non-optimal HMC estimators are significantly worse than MC. Therefore, a careful choice of the parameters of the HMC method is required to obtain an improved performance of the HMC method compared to MC.

5. Conclusion, Limitations and Future Work

Efficient calculation of systemic risk allocations is a challenging task, especially when the crisis event has a small probability. To solve this problem for models where a joint loss density is available, we proposed MCMC estimators where a Markov chain is constructed with the conditional loss distribution, given the crisis event as the target distribution. By using HMC and GS, efficient simulation methods from the constrained target distribution were obtained and the resulting MCMC estimator was expected to have a smaller standard error compared to that of the MC estimator. Sample efficiency is significantly improved, since the MCMC estimator is computed from samples generated directly from the conditional distribution of interest. Another advantage of the MCMC method is that its performance is less sensitive to the probability of the crisis event, and thus to the confidence levels of the underlying risk measures. We also proposed a heuristic for determining the parameters of the HMC method based on the MC presamples. Numerical experiments demonstrated that our MCMC estimators are more efficient than MC in terms of bias, standard error, and time-adjusted MSE. Stability of the MCMC estimation with respect to the probability of the crisis event and efficiency of the optimal parameter choice of the HMC method were also investigated in the experiments.
Based on the results in this paper, our MCMC estimators can be recommended when the probability of the crisis event is too small for MC to sufficiently simulate many samples for a statistical analysis and/or when unbiased systemic risk allocations under the VaR crisis event are required. The MCMC methods are likely to perform well when the dimension of the portfolio is less than or around 10, losses are bounded from the left, and the crisis event is of VaR or RVaR type; otherwise, heavy-tailedness and computational time can become challenging. Firstly, a theoretical convergence result of the HMC method is typically not available when the target distribution is unbounded and heavy-tailed, which is the case when the losses are unbounded and/or the crisis event is of ES type; see the case of the ES crisis event in the empirical study in Section 4.2. Secondly, both the HMC and GS methods suffer from high-dimensional target distributions since the algorithms contain parts of steps where the computational cost linearly increases in dimension. We observed that, in this case, although the MCMC estimator typically improves bias and standard error compared to MC, the improvement vanishes in terms of time-adjusted MSE due to the long computational time of the MCMC method. Finally, multimodality of joint loss distributions and/or the target distribution is also an undesirable feature since full conditional distributions and their inverses (which are required to implement the GS) are typically unavailable in the former case, and the latter case prevents the HMC method from efficiently exploring the entire support of the target distribution. Potential remedies for heavy-tailed and/or high-dimensional target distributions are the HMC method with a non-normal kinetic energy distribution and roll-back HMC; see Appendix B for details. Further investigation of HMC methods and faster methods for determining the HMC parameters are left for future work.

Supplementary Materials

An R script for reproducing the numerical experiments conducted in this paper is available at https://www.mdpi.com/2227-9091/8/1/6/s1.

Author Contributions

Conceptualization, T.K.; methodology, T.K.; formal analysis, T.K.; investigation, T.K. and M.H.; resources, T.K.; data curation, T.K.; writing—original draft preparation, T.K.; writing—review and editing, M.H.; visualization, T.K. and M.H.; supervision, M.H.; project administration, T.K.; funding acquisition, M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSERC through Discovery Grant RGPIN-5010-2015.

Acknowledgments

We wish to thank to an associate editor and anonymous referees for their careful reading of the manuscript and their insightful comments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
i.i.d.Independent and identically distributed
pdfProbability distribution function
cdfCumulative distribution function
ecdfEmpirical cdf
GPDGeneralized Pareto distribution
MSEMean squared error
LLNLaw of large numbers
CLTCentral limit theorem
VaRValue-at-Risk
RVaRRange VaR
ESExpected shortfall
MESMarginal expected shortfall
CoVaRConditional VaR
CoESConditional ES
MCMonte Carlo
SMCSequential Monte Carlo
MCMCMarkov chain Monte Carlo
ACRAcceptance rate
ACPAutocorrelation plot
MHMetropolis–Hastings
GSGibbs sampler
MGSMetropolized Gibbs sampler
DSGSDeterministic scan GS
RPGSRandom permutation GS
RSGSRandom scan GS
HMCHamiltonian Monte Carlo
RBHMCRoll-back HMC
RMHMCRiemannian manifold HMC
ALAEAllocated loss adjustment expense
P&LProfit and loss

Appendix A. Hamiltonian Dynamics with Boundary Reflection

In this appendix, we describe details of the HMC method with boundary reflection, as mentioned in Section 3.3.1. Let ( h , v ) be the hyperplane which the trajectory of the Hamiltonian dynamics hit at ( x ( t ) , p ( t ) ) . At this time, ( x ( t ) , p ( t ) ) is immediately replaced by ( x ( t ) , p r ( t ) ) where p r ( t ) is the reflected momentum defined by
p r ( t ) = p ( t ) p ( t ) ,
where p ( t ) and p ( t ) are such that p ( t ) = p ( t ) + p ( t ) and p ( t ) and p ( t ) are parallel and perpendicular to the hyperplane ( h , v ) , respectively. Afshar and Domke (2015) and Chevallier et al. (2018) showed that the map ( x ( t ) , p ( t ) ) ( x ( t ) , p r ( t ) ) preserves the volume and the Hamiltonian, and that this modified HMC method has the stationary distribution π . As long as the initial position x ( 0 ) belongs to C , the trajectory of the HMC method never violates the constraint C . The algorithm of this HMC method with reflection is obtained by replacing the Leapfrog function call in Step (3) of Algorithm 3 by Algorithm A1. Accordingly, the parameters of the hyperplanes need to be passed as input to Algorithm 3.
In Step (3-1) of Algorithm A1 the time t m at which the trajectory hits the boundary ( h m , v m ) is computed. If 0 < t m < 1 for some m { 1 , , M } , then the chain hits the boundary during the dynamics with length ϵ . At the smallest time t m * among such hitting times, the chain reflects from ( x * , p ) to ( x r * , p r ) against the corresponding boundary ( h m * , v m * ) as described in Step (3-2-1) of Algorithm A1. The remaining length of the dynamics is ( 1 t m * ) ϵ temp and Step (3) is repeated until the remaining length becomes zero. Other techniques of reflecting the dynamics are discussed in Appendix B.1.
Algorithm A1 Leapfrog method with boundary reflection.
Input: Current state ( x ( 0 ) , p ( 0 ) ) , stepsize ϵ > 0 , gradients U and K , and constraints ( h m , v m ) , m = 1 , , M .
Output: Updated state ( x ( ϵ ) , p ( ϵ ) ) .
 (1) Update p ( ϵ / 2 ) = p ( 0 ) + ϵ / 2 U ( x ( 0 ) ) .
 (2) Set ( x , p ) = ( x ( 0 ) , p ( ϵ / 2 ) ) , ϵ temp = ϵ .
 (3) while ϵ temp > 0
  (3-1) Compute
x * = x + ϵ temp K ( p ) , t m = ( v m h m x ) / ( ϵ h m p ) , m = 1 , , M .
  (3-2) if t m [ 0 , 1 ] for any m = 1 , , M ,
   (3-2-1) Set
m * = argmin { t m | 0 t m 1 , m = 1 , , M } , x r * = x * 2 h m * x * v m * h m * h m * h m * , p r = x * x t m * ϵ p ϵ ( 1 t m * ) .
   (3-2-2) Set ( x , p ) = ( x r * , p r ) and ϵ temp = ( 1 t m * ) ϵ temp .
   else
   (3-2-3) Set ( x , p ) = ( x * , p ) and ϵ temp = 0 .
   end if
  end while
 (4) Set x ( ϵ ) = x and p ( ϵ ) = p + ϵ 2 U ( x ) .

Appendix B. Other MCMC Methods

In this appendix, we introduce some advanced MCMC techniques potentially applicable to the problem of estimating systemic risk allocations.

Appendix B.1. Roll-Back HMC

Yi and Doshi-Velez (2017) proposed roll-back HMC (RBHMC), in which the indicator function 1 [ x C ] in the target distribution (5) is replaced by a smooth sigmoid function so that the Hamilotonian dynamics naturally move back inwards when the trajectory violates the constraints. HMC with reflection presented in Section 3.3.1 requires to check M boundary conditions at every iteration of the Hamiltonian dynamics. In our problem the number M linearly increases with the dimension d in the case of pure losses, which leads to a linear increase in the computational cost. The RBHMC method avoids such explicit boundary checks, and thus can reduce the computational cost of the HMC method with constrained target distributions. Despite saving computational time, we observed that the RBHMC method requires a careful choice of the stepsize ϵ > 0 and the smoothness parameter of the sigmoid function involved, and we could not find any guidance on how to choose them to guarantee a stable performance.

Appendix B.2. Riemannian Manifold HMC

Livingstone et al. (2019) indicated that non-normal kinetic energy distributions can potentially deal with heavy-tailed target distributions. In fact, the kinetic energy distribution F K can even be dependent on the position variable x . For example, when F K ( · | x ) = N ( 0 , G ( x ) ) for a positive definite matrix G ( x ) > 0 and x E , the resulting HMC method is known as Riemannian manifold HMC (RMHMC) since this case is equivalent to applying HMC on the Riemannian manifold with metric G ( x ) ; see Girolami and Calderhead (2011). Difficulties in implementing RMHMC are in the choice of metric G and in the simulation of the Hamiltonian dynamics. Due to the complexity of the Hamiltonian dynamics, simple discretization schemes such as the leapfrog method are not applicable, and the trajectory is updated implicitly by solving some system of equations; see Girolami and Calderhead (2011). Various choices of the metric G are studied in Betancourt (2013), Lan et al. (2014) and Livingstone and Girolami (2014) for different purposes. Simulation of RMHMC is studied, for example, in Byrne and Girolami (2013).

Appendix B.3. Metropolized Gibbs Samplers

Müller (1992) introduced the Metropolized Gibbs sampler (MGS) in which the proposal density q in the MH kernel is set to be q = f Y | v 1 h Y v 2 where Y has the same marginal distributions as X but a different copula C q for which C j | j q and C j | j q , 1 are available so that the GS can be applied to simulate this proposal. This method can be used when the inversion method is not feasible since C j | j or C j | j 1 are not available. Following the MH algorithm, the candidate is accepted with the acceptance probability (4), which can be simply written as
α ( x , x ˜ ) = min c ( F ( x ˜ ) ) c q ( F ( x ) ) c ( F ( x ) ) c q ( F ( x ˜ ) ) , 1 .
As an example of the MGS, suppose C is the Gumbel copula, for which the full conditional distributions cannot be inverted analytically. One could then choose the survival Clayton copula as the proposal copula C q above. For this choice of copula, q j | j is available by the inversion method as discussed in Section 3.4.1. Furthermore, the acceptance probability is expected to be high especially on the upper tail part because the upper threshold copula of C defined as P ( U > v | U > u ) , v [ u , 1 ] , u [ 0 , 1 ] d , U C is known to converge to that of a survival Clayton copula when lim u j , j = 1 , , d ; see Juri and Wüthrich (2002), Juri and Wüthrich (2003), Charpentier and Segers (2007) and Larsson and Nešlehová (2011).

References

  1. Acharya, Viral V., Lasse H. Pedersen, Thomas Philippon, and Matthew Richardson. 2017. Measuring systemic risk. The Review of Financial Studies 30: 2–47. [Google Scholar] [CrossRef]
  2. Adrian, Tobias, and Markus K. Brunnermeier. 2016. Covar. The American Economic Review 106: 1705. [Google Scholar] [CrossRef]
  3. Afshar, Hadi Mohasel, and Justin Domke. 2015. Reflection, refraction, and hamiltonian monte carlo. In Advances in Neural Information Processing Systems. Cambridge: The MIT Press, pp. 3007–15. [Google Scholar]
  4. Asimit, Alexandru V., Edward Furman, Qihe Tang, and Raluca Vernic. 2011. Asymptotics for risk capital allocations based on conditional tail expectation. Insurance: Mathematics and Economics 49: 310–24. [Google Scholar] [CrossRef] [Green Version]
  5. Asimit, Alexandru V., and Jinzhu Li. 2018. Systemic risk: An asymptotic evaluation. ASTIN Bulletin: The Journal of The IAA 48: 673–98. [Google Scholar] [CrossRef]
  6. Bernardi, Mauro, Fabrizio Durante, and Piotr Jaworski. 2017. Covar of families of copulas. Statistics & Probability Letters 120: 8–17. [Google Scholar]
  7. Beskos, Alexandros, Natesh Pillai, Gareth Roberts, Jesus-Maria Sanz-Serna, and Andrew Stuart. 2010. The acceptance probability of the hybrid monte carlo method in high-dimensional problems. AIP Conference Proceedings 1281: 23–6. [Google Scholar]
  8. Beskos, Alexandros, Natesh Pillai, Gareth Roberts, Jesus-Maria Sanz-Serna, and Andrew Stuart. 2013. Optimal tuning of the hybrid monte carlo algorithm. Bernoulli 19: 1501–34. [Google Scholar] [CrossRef]
  9. Betancourt, Michael. 2012. Cruising the simplex: Hamiltonian monte carlo and the dirichlet distribution. AIP Conference Proceedings 1443: 157–64. [Google Scholar]
  10. Betancourt, Michael. 2013. A general metric for riemannian manifold hamiltonian monte carlo. In International Conference on Geometric Science of Information. New York: Springer, pp. 327–34. [Google Scholar]
  11. Betancourt, Michael. 2016. Identifying the optimal integration time in hamiltonian monte carlo. arXiv arXiv:1601.00225. [Google Scholar]
  12. Betancourt, Michael. 2017. A conceptual introduction to hamiltonian monte carlo. arXiv arXiv:1701.02434. [Google Scholar]
  13. Betancourt, Michael, Simon Byrne, and Mark Girolami. 2014. Optimizing the integrator step size for hamiltonian monte carlo. arXiv arXiv:1411.6669. [Google Scholar]
  14. Byrne, Simon, and Mark Girolami. 2013. Geodesic monte carlo on embedded manifolds. Scandinavian Journal of Statistics 40: 825–45. [Google Scholar] [CrossRef] [PubMed]
  15. Cambou, Mathieu, Marius Hofert, and Christiane Lemieux. 2017. Quasi-random numbers for copula models. Statistics and Computing 27: 1307–29. [Google Scholar] [CrossRef] [Green Version]
  16. Cances, Eric, Frédéric Legoll, and Gabriel Stoltz. 2007. Theoretical and numerical comparison of some sampling methods for molecular dynamics. ESAIM: Mathematical Modelling and Numerical Analysis 41: 351–89. [Google Scholar] [CrossRef]
  17. Charpentier, Arthur, and Johan Segers. 2007. Lower tail dependence for archimedean copulas: Characterizations and pitfalls. Insurance: Mathematics and Economics 40: 525–32. [Google Scholar] [CrossRef]
  18. Chen, Chen, Garud Iyengar, and Ciamac C. Moallemi. 2013. An axiomatic approach to systemic risk. Management Science 59: 1373–88. [Google Scholar] [CrossRef] [Green Version]
  19. Chevallier, Augustin, Sylvain Pion, and Frédéric Cazals. 2018. Hamiltonian Monte Carlo With Boundary Reflections, and Application To Polytope Volume Calculations. Available online: https://hal.archives-ouvertes.fr/hal-01919855/ (accessed on 8 January 2020).
  20. Chib, Siddhartha, and Edward Greenberg. 1995. Understanding the Metropolis–Hastings algorithm. The American Statistician 49: 327–35. [Google Scholar]
  21. Chiragiev, Arthur, and Zinoviy Landsman. 2007. Multivariate pareto portfolios: Tce-based capital allocation and divided differences. Scandinavian Actuarial Journal 2007: 261–80. [Google Scholar] [CrossRef]
  22. Christen, J. Andrés, Colin Fox, and Mario Santana-Cibrian. 2017. Optimal direction gibbs sampler for truncated multivariate normal distributions. Communications in Statistics-Simulation and Computation 46: 2587–600. [Google Scholar] [CrossRef]
  23. Denault, Michel. 2001. Coherent allocation of risk capital. Journal of Risk 4: 1–34. [Google Scholar] [CrossRef] [Green Version]
  24. Devroye, Luc. 1985. Non-Uniform Random Variate Generation. New York: Springer. [Google Scholar]
  25. Dhaene, Jan, Luc Henrard, Zinoviy Landsman, Antoine Vandendorpe, and Steven Vanduffel. 2008. Some results on the cte-based capital allocation rule. Insurance: Mathematics and Economics 42: 855–63. [Google Scholar] [CrossRef]
  26. Dhaene, Jan, Andreas Tsanakas, Emiliano A. Valdez, and Steven Vanduffel. 2012. Optimal capital allocation principles. Journal of Risk and Insurance 79: 1–28. [Google Scholar] [CrossRef] [Green Version]
  27. Duane, Simon, Anthony D. Kennedy, Brian J. Pendleton, and Duncan Roweth. 1987. Hybrid monte carlo. Physics Letters B 195: 216–22. [Google Scholar] [CrossRef]
  28. Durmus, Alain, Eric Moulines, and Eero Saksman. 2017. On the convergence of hamiltonian monte carlo. arXiv arXiv:1705.00166. [Google Scholar]
  29. Fan, Guobin, Yong Zeng, and Woon K. Wong. 2012. Decomposition of portfolio var and expected shortfall based on multivariate copula simulation. International Journal of Management Science and Engineering Management 7: 153–60. [Google Scholar] [CrossRef]
  30. Frees, Edward W., and Emiliano A. Valdez. 1998. Understanding relationships using copulas. North American Actuarial Journal 2: 1–25. [Google Scholar] [CrossRef]
  31. Furman, Edward, Alexey Kuznetsov, and Ričardas Zitikis. 2018. Weighted risk capital allocations in the presence of systematic risk. Insurance: Mathematics and Economics 79: 75–81. [Google Scholar] [CrossRef]
  32. Furman, Edward, and Zinoviy Landsman. 2008. Economic capital allocations for non-negative portfolios of dependent risks. ASTIN Bulletin: The Journal of the IAA 38: 601–19. [Google Scholar] [CrossRef] [Green Version]
  33. Furman, Edward, Ruodu Wang, and Ričardas Zitikis. 2017. Gini-type measures of risk and variability: Gini shortfall, capital allocations, and heavy-tailed risks. Journal of Banking & Finance 83: 70–84. [Google Scholar]
  34. Furman, Edward, and Ričardas Zitikis. 2008. Weighted risk capital allocations. Insurance: Mathematics and Economics 43: 263–9. [Google Scholar] [CrossRef]
  35. Furman, Edward, and Ričardas Zitikis. 2009. Weighted pricing functionals with applications to insurance: An overview. North American Actuarial Journal 13: 483–96. [Google Scholar] [CrossRef]
  36. Gelfand, Alan E., and Adrian F. M. Smith. 1990. Sampling-based approaches to calculating marginal densities. Journal of the American Statistical Association 85: 398–409. [Google Scholar] [CrossRef]
  37. Gelfand, Alan E., Adrian F. M. Smith, and Tai-Ming Lee. 1992. Bayesian analysis of constrained parameter and truncated data problems using gibbs sampling. Journal of the American Statistical Association 87: 523–32. [Google Scholar] [CrossRef]
  38. Geman, Stuart, and Donald Geman. 1984. Stochastic relaxation, gibbs distributions, and the bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence 6: 721–41. [Google Scholar] [CrossRef]
  39. Geweke, John. 1991. Efficient simulation from the multivariate normal and student-t distributions subject to linear constraints and the evaluation of constraint probabilities. In Computing Science and Statistics: Proceedings of the 23rd Symposium on the Interface. Fairfax: Interface Foundation of North America, Inc., pp. 571–8. [Google Scholar]
  40. Geyer, Charles. 2011. Introduction to markov chain monte carlo. In Handbook of Markov Chain Monte Carlo. New York: Springer, pp. 3–47. [Google Scholar]
  41. Girardi, Giulio, and A. Tolga Ergün. 2013. Systemic risk measurement: Multivariate garch estimation of covar. Journal of Banking & Finance 37: 3169–80. [Google Scholar]
  42. Girolami, Mark, and Ben Calderhead. 2011. Riemann manifold langevin and hamiltonian monte carlo methods. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 73: 123–214. [Google Scholar] [CrossRef]
  43. Glasserman, Paul. 2005. Measuring marginal risk contributions in credit portfolios. Journal of Computational Finance 9: 1–41. [Google Scholar] [CrossRef] [Green Version]
  44. Glasserman, Paul. 2013. Monte Carlo Methods in Financial Engineering. New York: Springer. [Google Scholar]
  45. Glasserman, Paul, and Jingyi Li. 2005. Importance sampling for portfolio credit risk. Management Science 51: 1643–56. [Google Scholar] [CrossRef] [Green Version]
  46. Gourieroux, Christian, and Alain Monfort. 2013. Allocating systemic risk in a regulatory perspective. International Journal of Theoretical and Applied Finance 16: 1350041. [Google Scholar] [CrossRef]
  47. Gudmundsson, Thorbjörn, and Henrik Hult. 2014. Markov chain monte carlo for computing rare-event probabilities for a heavy-tailed random walk. Journal of Applied Probability 51: 359–76. [Google Scholar] [CrossRef]
  48. Gupta, Sourendu, A. Irbäc, Frithjof Karsch, and Bengt Petersson. 1990. The acceptance probability in the hybrid monte carlo method. Physics Letters B 242: 437–43. [Google Scholar] [CrossRef] [Green Version]
  49. Hastings, W. Keith. 1970. Monte carlo sampling methods using markov chains and their applications. Biometrika 57: 97–109. [Google Scholar] [CrossRef]
  50. Hofert, Marius, Ivan Kojadinovic, Martin Mächler, and Jun Yan. 2018. Elements of Copula Modeling with R. New York: Springer Use R! Series. [Google Scholar] [CrossRef]
  51. Hoffman, Matthew D., and Andrew Gelman. 2014. The no-u-turn sampler: Adaptively setting path lengths in hamiltonian monte carlo. Journal of Machine Learning Research 15: 1593–623. [Google Scholar]
  52. Hoffmann, Hannes, Thilo Meyer-Brandis, and Gregor Svindland. 2016. Risk-consistent conditional systemic risk measures. Stochastic Processes and Their Applications 126: 2014–37. [Google Scholar] [CrossRef] [Green Version]
  53. Hogg, Robert V., and Stuart A. Klugman. 2009. Loss Distributions. Hoboken: John Wiley & Sons, Volume 249. [Google Scholar]
  54. Jaworski, Piotr. 2017. On conditional value at risk (covar) for tail-dependent copulas. Dependence Modeling 5: 1–19. [Google Scholar] [CrossRef] [Green Version]
  55. Johnson, Alicia A. 2009. Geometric Ergodicity of Gibbs Samplers. Available online: https://conservancy.umn.edu/handle/11299/53661 (accessed on 8 January 2020).
  56. Jones, Galin L., Murali Haran, Brian S. Caffo, and Ronald Neath. 2006. Fixed-width output analysis for markov chain monte carlo. Journal of the American Statistical Association 101: 1537–47. [Google Scholar] [CrossRef] [Green Version]
  57. Juri, Alessandro, and Mario V. Wüthrich. 2002. Copula convergence theorems for tail events. Insurance: Mathematics and Economics 30: 405–20. [Google Scholar] [CrossRef]
  58. Juri, Alessandro, and Mario V. Wüthrich. 2003. Tail dependence from a distributional point of view. Extremes 6: 213–46. [Google Scholar] [CrossRef]
  59. Kalkbrener, Michael, Hans Lotter, and Ludger Overbeck. 2004. Sensible and efficient capital allocation for credit portfolios. Risk 17: S19–S24. [Google Scholar]
  60. Klugman, Stuart A., and Rahul Parsa. 1999. Fitting bivariate loss distributions with copulas. Insurance: Mathematics and Economics 24: 139–48. [Google Scholar] [CrossRef]
  61. Koike, Takaaki, and Mihoko Minami. 2019. Estimation of risk contributions with mcmc. Quantitative Finance 19: 1579–97. [Google Scholar] [CrossRef]
  62. Kromer, Eduard, Ludger Overbeck, and Konrad Zilch. 2016. Systemic risk measures on general measurable spaces. Mathematical Methods of Operations Research 84: 323–57. [Google Scholar] [CrossRef]
  63. Laeven, Roger J. A., and Marc J. Goovaerts. 2004. An optimization approach to the dynamic allocation of economic capital. Insurance: Mathematics and Economics 35: 299–319. [Google Scholar] [CrossRef]
  64. Lan, Shiwei, Jeffrey Streets, and Babak Shahbaba. 2014. Wormhole hamiltonian monte carlo. In Twenty-Eighth AAAI Conference on Artificial Intelligence. Available online: https://www.aaai.org/ocs/index.php/AAAI/AAAI14/paper/viewPaper/8437 (accessed on 8 January 2020).
  65. Larsson, Martin, and Johanna Nešlehová. 2011. Extremal behavior of archimedean copulas. Advances in Applied Probability 43: 195–216. [Google Scholar] [CrossRef]
  66. Leimkuhler, Benedict, and Sebastian Reich. 2004. Simulating Hamiltonian Dynamics. Cambridge: Cambridge University Press, Volume 14. [Google Scholar]
  67. Liu, Jun S., Wing H. Wong, and Augustine Kong. 1995. Covariance structure and convergence rate of the gibbs sampler with various scans. Journal of the Royal Statistical Society: Series B (Methodological) 57: 157–69. [Google Scholar] [CrossRef]
  68. Livingstone, Samuel, Michael Betancourt, Simon Byrne, and Mark Girolami. 2016. On the geometric ergodicity of hamiltonian monte carlo. arXiv arXiv:1601.08057. [Google Scholar]
  69. Livingstone, Samuel, Michael F. Faulkner, and Gareth O. Roberts. 2019. Kinetic energy choice in hamiltonian/hybrid monte carlo. Biometrika 106: 303–19. [Google Scholar] [CrossRef]
  70. Livingstone, Samuel, and Mark Girolami. 2014. Information-geometric markov chain monte carlo methods using diffusions. Entropy 16: 3074–102. [Google Scholar] [CrossRef]
  71. Mainik, Georg, and Eric Schaanning. 2014. On dependence consistency of covar and some other systemic risk measures. Statistics & Risk Modeling 31: 49–77. [Google Scholar]
  72. McNeil, Alexander J., Rüdiger Frey, and Paul Embrechts. 2015. Quantitative Risk Management: Concepts, Techniques and Tools. Princeton: Princeton University Press. [Google Scholar]
  73. Metropolis, Nicholas, Arianna W. Rosenbluth, Marshall N. Rosenbluth, Augusta H. Teller, and Edward Teller. 1953. Equation of state calculations by fast computing machines. The Journal of Chemical Physics 21: 1087–92. [Google Scholar] [CrossRef] [Green Version]
  74. Meyn, Sean P., and Richard L. Tweedie. 2012. Markov Chains and Stochastic Stability. New York: Springer. [Google Scholar]
  75. Müller, Peter. 1992. Alternatives to the Gibbs Sampling Scheme. Available online: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.48.5613 (accessed on 8 January 2020).
  76. Neal, Radford M. 2011. Mcmc using hamiltonian dynamics. Handbook of Markov Chain Monte Carlo 2: 2. [Google Scholar]
  77. Nelsen, Roger B. 2006. An Introduction to Copulas. New York: Springer. [Google Scholar]
  78. Nummelin, Esa. 2002. Mc’s for mcmc’ists. International Statistical Review 70: 215–40. [Google Scholar] [CrossRef]
  79. Nummelin, Esa. 2004. General Irreducible Markov Chains and Non-Negative Operators. Cambridge: Cambridge University Press. [Google Scholar]
  80. Pakman, Ari, and Liam Paninski. 2014. Exact hamiltonian monte carlo for truncated multivariate gaussians. Journal of Computational and Graphical Statistics 23: 518–42. [Google Scholar] [CrossRef] [Green Version]
  81. Rodriguez-Yam, Gabriel, Richard A. Davis, and Louis L. Scharf. 2004. Efficient gibbs sampling of truncated multivariate normal with application to constrained linear regression. Unpublished manuscript. [Google Scholar]
  82. Rosenthal, Jeffrey S. 2011. Optimal proposal distributions and adaptive mcmc. In Handbook of Markov Chain Monte Carlo. Edited by Steve Brooks, Andrew Gelman, Galin Jones and Xiao-Li Meng. Boca Raton: CRC Press. [Google Scholar]
  83. Ruján, Pál. 1997. Playing billiards in version space. Neural Computation 9: 99–122. [Google Scholar] [CrossRef]
  84. Siller, Thomas. 2013. Measuring marginal risk contributions in credit portfolios. Quantitative Finance 13: 1915–23. [Google Scholar] [CrossRef]
  85. Targino, Rodrigo S., Gareth W. Peters, and Pavel V. Shevchenko. 2015. Sequential monte carlo samplers for capital allocation under copula-dependent risk models. Insurance: Mathematics and Economics 61: 206–26. [Google Scholar] [CrossRef] [Green Version]
  86. Tasche, Dirk. 1995. Risk Contributions and Performance Measurement. Working Paper. München: Techische Universität München. [Google Scholar]
  87. Tasche, Dirk. 2001. Conditional expectation as quantile derivative. arXiv arXiv:math/0104190. [Google Scholar]
  88. Tasche, Dirk. 2008. Capital allocation to business units and sub-portfolios: The euler principle. In Pillar II in the New Basel Accord: The Challenge of Economic Capital. Edited by Andrea Resti. London: Risk Books, pp. 423–53. [Google Scholar]
  89. Tierney, Luke. 1994. Markov chains for exploring posterior distributions. The Annals of Statistics 1994: 1701–28. [Google Scholar] [CrossRef]
  90. Vats, Dootika, James M. Flegal, and Galin L. Jones. 2015. Multivariate output analysis for markov chain monte carlo. arXiv arXiv:1512.07713. [Google Scholar]
  91. Vernic, Raluca. 2006. Multivariate skew-normal distributions with applications in insurance. Insurance: Mathematics and Economics 38: 413–26. [Google Scholar] [CrossRef]
  92. Yamai, Yasuhiro, and Toshinao Yoshiba. 2002. Comparative analyses of expected shortfall and value-at-risk: Their estimation error, decomposition, and optimization. Monetary and Economic Studies 20: 87–121. [Google Scholar]
  93. Yi, Kexin, and Finale Doshi-Velez. 2017. Roll-back hamiltonian monte carlo. arXiv arXiv:1709.02855. [Google Scholar]
Figure 1. Hamiltonian errors of the HMC methods for estimating systemic risk allocations with VaR (left) and RVaR (right) crisis events for the loss distribution (M1). The stepsize and integration time are set to be ( ϵ , T ) = ( 0.210 , 12 ) in Case (I) and ( ϵ , T ) = ( 0.095 , 13 ) in Case (II).
Figure 1. Hamiltonian errors of the HMC methods for estimating systemic risk allocations with VaR (left) and RVaR (right) crisis events for the loss distribution (M1). The stepsize and integration time are set to be ( ϵ , T ) = ( 0.210 , 12 ) in Case (I) and ( ϵ , T ) = ( 0.095 , 13 ) in Case (II).
Risks 08 00006 g001
Figure 2. Plots of N = 1500 MCMC samples (green) with VaR (left), RVaR (center), and ES (right) crisis events. All plots include the data and the MC samples with sample size N = 1500 in black and blue dots, respectively. The red lines represent x 1 + x 2 = VaR ^ α 1 ( S ) and x 1 + x 2 = VaR ^ α 2 ( S ) where VaR ^ α 1 ( S ) = 4.102 × 10 4 and VaR ^ α 2 ( S ) = 9.117 × 10 4 are the MC estimates of VaR α 1 ( S ) and VaR α 2 ( S ) , respectively, for α 1 = 0.975 and α 2 = 0.99 .
Figure 2. Plots of N = 1500 MCMC samples (green) with VaR (left), RVaR (center), and ES (right) crisis events. All plots include the data and the MC samples with sample size N = 1500 in black and blue dots, respectively. The red lines represent x 1 + x 2 = VaR ^ α 1 ( S ) and x 1 + x 2 = VaR ^ α 2 ( S ) where VaR ^ α 1 ( S ) = 4.102 × 10 4 and VaR ^ α 2 ( S ) = 9.117 × 10 4 are the MC estimates of VaR α 1 ( S ) and VaR α 2 ( S ) , respectively, for α 1 = 0.975 and α 2 = 0.99 .
Risks 08 00006 g002
Figure 3. Hamiltonian errors of the HMC methods for estimating systemic risk allocations with VaR, RVaR, and ES crisis events for the loss distribution (M3). The stepsize and the integration time are chosen as ( ϵ , T ) = ( 0.015 , 34 ) , ( ϵ , T ) = ( 0.026 , 39 ) and ( ϵ , T ) = ( 5.132 × 10 5 , 838 ) , respectively.
Figure 3. Hamiltonian errors of the HMC methods for estimating systemic risk allocations with VaR, RVaR, and ES crisis events for the loss distribution (M3). The stepsize and the integration time are chosen as ( ϵ , T ) = ( 0.015 , 34 ) , ( ϵ , T ) = ( 0.026 , 39 ) and ( ϵ , T ) = ( 5.132 × 10 5 , 838 ) , respectively.
Risks 08 00006 g003
Figure 4. Bias (left), standard error (middle) and time-adjusted mean squared error (right) of the MC, HMC, and GS estimators of risk contribution-type systemic risk allocations under VaR 0.99 (top), RVaR 0.95 , 0.99 (middle), and ES 0.99 (bottom) crisis events. The underlying loss distribution is t ν ( μ , P ) , where ν = 6 , μ = 0 and P = 1 / 12 · 1 d 1 d + diag d ( 11 / 12 ) for portfolio dimensions d { 4 , 6 , 8 , 10 } . Note that the GS method is applied only to RVaR and ES contributions.
Figure 4. Bias (left), standard error (middle) and time-adjusted mean squared error (right) of the MC, HMC, and GS estimators of risk contribution-type systemic risk allocations under VaR 0.99 (top), RVaR 0.95 , 0.99 (middle), and ES 0.99 (bottom) crisis events. The underlying loss distribution is t ν ( μ , P ) , where ν = 6 , μ = 0 and P = 1 / 12 · 1 d 1 d + diag d ( 11 / 12 ) for portfolio dimensions d { 4 , 6 , 8 , 10 } . Note that the GS method is applied only to RVaR and ES contributions.
Risks 08 00006 g004
Figure 5. Bias (left), standard error (middle), and time-adjusted mean squared error (right) of the MC, HMC, and GS estimators of risk contribution-type systemic risk allocations with the underlying loss distribution t ν ( μ , P ) , where ν = 6 , μ = 0 , P = 1 / 12 · 1 d 1 d + diag d ( 11 / 12 ) and d = 5 . The crisis event is taken differently, as VaR α VaR (top), RVaR α 1 RVaR , α 2 RVaR (middle) and ES α ES (bottom) for confidence levels α VaR { 0.9 , 0.99 , 0.999 , 0.9999 } , ( α 1 RVaR , α 2 RVaR ) { ( 0.9 , 0.9999 ) , ( 0.9 , 0.99 ) , ( 0.99 , 0.999 ) , ( 0.999 , 0.9999 ) } , and α ES { 0.9 , 0.99 , 0.999 , 0.9999 } . Note that the GS method is applied only to RVaR and ES contributions.
Figure 5. Bias (left), standard error (middle), and time-adjusted mean squared error (right) of the MC, HMC, and GS estimators of risk contribution-type systemic risk allocations with the underlying loss distribution t ν ( μ , P ) , where ν = 6 , μ = 0 , P = 1 / 12 · 1 d 1 d + diag d ( 11 / 12 ) and d = 5 . The crisis event is taken differently, as VaR α VaR (top), RVaR α 1 RVaR , α 2 RVaR (middle) and ES α ES (bottom) for confidence levels α VaR { 0.9 , 0.99 , 0.999 , 0.9999 } , ( α 1 RVaR , α 2 RVaR ) { ( 0.9 , 0.9999 ) , ( 0.9 , 0.99 ) , ( 0.99 , 0.999 ) , ( 0.999 , 0.9999 ) } , and α ES { 0.9 , 0.99 , 0.999 , 0.9999 } . Note that the GS method is applied only to RVaR and ES contributions.
Risks 08 00006 g005
Figure 6. Bias (left), standard error (middle), and time-adjusted mean squared error (right) of the MC and HMC estimators of risk contribution-type systemic risk allocations under VaR 0.9 , RVaR 0.9 , 0.99 , and ES 0.9 crisis events. The underlying loss distribution is t ν ( μ , P ) , where ν = 6 , μ = 0 , P = 1 / 12 · 1 d 1 d + diag d ( 11 / 12 ) and d = 5 . The parameters of the HMC method are taken as ( ϵ opt , ϵ opt ) determined by Algorithm 4 and ( ϵ , T ) { ( 10 ϵ opt , 2 T opt ) , ( 10 ϵ opt , T opt / 2 ) , ( ϵ opt / 10 , 2 T opt ) , ( ϵ opt / 10 , T opt / 2 ) } . In the labels of the x-axes, each of the five cases ( ϵ opt , ϵ opt ) , ( 10 ϵ opt , 2 T opt ) , ( 10 ϵ opt , T opt / 2 ) , ( ϵ opt / 10 , 2 T opt ) and ( ϵ opt / 10 , T opt / 2 ) is denoted by HMC.opt, HMC.mm, HMC.md, HMC.dm, and HMC.dd, respectively.
Figure 6. Bias (left), standard error (middle), and time-adjusted mean squared error (right) of the MC and HMC estimators of risk contribution-type systemic risk allocations under VaR 0.9 , RVaR 0.9 , 0.99 , and ES 0.9 crisis events. The underlying loss distribution is t ν ( μ , P ) , where ν = 6 , μ = 0 , P = 1 / 12 · 1 d 1 d + diag d ( 11 / 12 ) and d = 5 . The parameters of the HMC method are taken as ( ϵ opt , ϵ opt ) determined by Algorithm 4 and ( ϵ , T ) { ( 10 ϵ opt , 2 T opt ) , ( 10 ϵ opt , T opt / 2 ) , ( ϵ opt / 10 , 2 T opt ) , ( ϵ opt / 10 , T opt / 2 ) } . In the labels of the x-axes, each of the five cases ( ϵ opt , ϵ opt ) , ( 10 ϵ opt , 2 T opt ) , ( 10 ϵ opt , T opt / 2 ) , ( ϵ opt / 10 , 2 T opt ) and ( ϵ opt / 10 , T opt / 2 ) is denoted by HMC.opt, HMC.mm, HMC.md, HMC.dm, and HMC.dd, respectively.
Risks 08 00006 g006
Table 1. Estimates and standard errors for the MC and HMC estimators of risk contribution, RVaR, VaR, and ES-type systemic risk allocations under (I) the VaR crisis event, and (II) the RVaR crisis event for the loss distribution (M1). The sample size of the MC method is N MC = 10 5 , and that of the HMC method is N MCMC = 10 4 . The acceptance rate (ACR), stepsize ϵ , integration time T, and run time are ACR = 0.996 , ϵ = 0.210 , T = 12 , and run time = 1.277 mins in Case (I), and ACR = 0.984 , ϵ = 0.095 , T = 13 , and run time = 1.649 mins in Case (II).
Table 1. Estimates and standard errors for the MC and HMC estimators of risk contribution, RVaR, VaR, and ES-type systemic risk allocations under (I) the VaR crisis event, and (II) the RVaR crisis event for the loss distribution (M1). The sample size of the MC method is N MC = 10 5 , and that of the HMC method is N MCMC = 10 4 . The acceptance rate (ACR), stepsize ϵ , integration time T, and run time are ACR = 0.996 , ϵ = 0.210 , T = 12 , and run time = 1.277 mins in Case (I), and ACR = 0.984 , ϵ = 0.095 , T = 13 , and run time = 1.649 mins in Case (II).
MCHMC
Estimator A 1 ϱ , C ( X ) A 2 ϱ , C ( X ) A 3 ϱ , C ( X ) A 1 ϱ , C ( X ) A 2 ϱ , C ( X ) A 3 ϱ , C ( X )
(I) GPD + survival Clayton with VaR crisis event: { S = VaR 0.99 ( S ) }
E [ X | C VaR ] 9.5819.4009.8299.5939.5999.619
Standard error0.1260.1180.1200.0070.0090.009
RVaR 0.975 , 0.99 ( X | C VaR ) 12.98612.91913.63013.29813.20413.338
Standard error0.2290.1310.0860.0610.0490.060
VaR 0.99 ( X | C VaR ) 13.59213.23513.79613.74213.56513.768
Standard error0.6470.3330.2700.0880.0700.070
ES 0.99 ( X | C VaR ) 14.77513.95514.56814.46114.22714.427
Standard error0.6600.4980.6050.1920.1760.172
(II) GPD + survival Clayton with RVaR crisis event: { VaR 0.975 ( S ) S VaR 0.99 ( S ) }
E [ X | C RVaR ] 7.8737.7807.8167.8127.8027.780
Standard error0.0460.0460.0460.0120.0120.011
RVaR 0.975 , 0.99 ( X | C RVaR ) 11.79011.90811.68011.68611.69611.646
Standard error0.0470.0570.0430.0530.0550.058
RVaR 0.99 ( X | C VaR ) 12.20712.38212.08712.10212.05312.044
Standard error0.1830.1970.1820.0740.0690.069
ES 0.99 ( X | C RVaR ) 13.07913.10213.05912.85912.79112.713
Standard error0.1820.1730.1880.2310.2180.187
Table 2. Estimates and standard errors for the MC and the GS estimators of risk contribution, VaR, RVaR, and ES-type systemic risk allocations under (III) distribution (M1) and the ES crisis event, (IV) distribution (M2), and the RVaR crisis event, and (V) distribution (M2) and ES crisis event. The sample size of the MC method is N MC = 10 5 and that of the GS is N MCMC = 10 4 . The thinning interval of times T, selection probability p and run time are T = 12 , p = ( 0.221 , 0.362 , 0.416 ) and run time = 107.880 secs in Case (III), T = 10 , p = ( 0.330 , 0.348 , 0.321 ) and run time = 56.982 secs in Case (IV) and T = 4 , p = ( 0.241 , 0.503 , 0.255 ) and run time = 22.408 secs in Case (V).
Table 2. Estimates and standard errors for the MC and the GS estimators of risk contribution, VaR, RVaR, and ES-type systemic risk allocations under (III) distribution (M1) and the ES crisis event, (IV) distribution (M2), and the RVaR crisis event, and (V) distribution (M2) and ES crisis event. The sample size of the MC method is N MC = 10 5 and that of the GS is N MCMC = 10 4 . The thinning interval of times T, selection probability p and run time are T = 12 , p = ( 0.221 , 0.362 , 0.416 ) and run time = 107.880 secs in Case (III), T = 10 , p = ( 0.330 , 0.348 , 0.321 ) and run time = 56.982 secs in Case (IV) and T = 4 , p = ( 0.241 , 0.503 , 0.255 ) and run time = 22.408 secs in Case (V).
MCGS
Estimator A 1 ϱ , C ( X ) A 2 ϱ , C ( X ) A 3 ϱ , C ( X ) A 1 ϱ , C ( X ) A 2 ϱ , C ( X ) A 3 ϱ , C ( X )
(III) GPD + survival Clayton with ES crisis event: { VaR 0.99 ( S ) S }
E [ X | C ES ] 15.65715.80615.72115.20915.17515.190
Standard error0.4340.4750.3950.2570.2580.261
RVaR 0.975 , 0.99 ( X | C ES ) 41.62641.02645.93945.50645.00845.253
Standard error1.2111.0651.6151.0311.1331.256
VaR 0.99 ( X | C ES ) 49.68948.81857.48855.03354.74654.783
Standard error4.9014.3884.9738.0795.6303.803
ES 0.99 ( X | C ES ) 104.761109.83597.94471.87472.58870.420
Standard error23.00527.89517.9084.8324.5844.313
(IV) Multivariate t with RVaR crisis event: { VaR 0.975 ( S ) S VaR 0.99 ( S ) }
E [ X | C RVaR ] 2.4561.9342.4762.3942.0602.435
Bias0.019−0.0970.038−0.0430.029−0.002
Standard error0.0260.0360.0270.0140.0230.019
RVaR 0.975 , 0.99 ( X | C RVaR ) 4.6704.9984.8934.6025.1884.748
Standard error0.0370.0420.0310.0320.0700.048
RVaR 0.99 ( X | C VaR ) 5.2175.3975.2404.8785.7175.092
Standard error0.2380.1570.1450.0490.1740.100
ES 0.99 ( X | C RVaR ) 5.9295.9775.9465.4466.5176.063
Standard error0.2040.1790.1990.1560.2480.344
(V) Multivariate t with ES crisis event: { VaR 0.99 ( S ) S }
E [ X | C ES ] 3.7583.0993.7703.7353.1263.738
Bias0.017−0.0180.029-0.0050.009−0.003
Standard error0.0550.0720.0600.0310.0270.030
RVaR 0.975 , 0.99 ( X | C ES ) 8.5168.4899.0518.5868.3178.739
Standard error0.0890.1670.1610.1440.1560.158
VaR 0.99 ( X | C ES ) 9.2569.75410.3279.4549.5179.890
Standard error0.5170.6800.6980.2480.2930.327
ES 0.99 ( X | C ES ) 11.12912.52012.94611.85712.46912.375
Standard error0.5951.3210.8260.7850.9480.835
Table 3. Estimates and standard errors for the MC and HMC estimators of RVaR, VaR, and ES-type systemic risk allocations under the loss distribution (M3) with the (I) VaR crisis event, (II) RVaR crisis event, and (III) ES crisis event. The MC sample size is N MC = 10 5 , and that of the HMC method is N MCMC = 10 4 . The acceptance rate (ACR), stepsize ϵ , integration time T, and run time are ACR = 0.997 , ϵ = 0.015 , T = 34 and run time = 2.007 min in Case (I), ACR = 0.986 , ϵ = 0.026 , T = 39 and run time = 2.689 min in Case (II), ACR = 0.995 , ϵ = 5.132 × 10 5 , T = 838 and run time = 44.831 min in Case (III).
Table 3. Estimates and standard errors for the MC and HMC estimators of RVaR, VaR, and ES-type systemic risk allocations under the loss distribution (M3) with the (I) VaR crisis event, (II) RVaR crisis event, and (III) ES crisis event. The MC sample size is N MC = 10 5 , and that of the HMC method is N MCMC = 10 4 . The acceptance rate (ACR), stepsize ϵ , integration time T, and run time are ACR = 0.997 , ϵ = 0.015 , T = 34 and run time = 2.007 min in Case (I), ACR = 0.986 , ϵ = 0.026 , T = 39 and run time = 2.689 min in Case (II), ACR = 0.995 , ϵ = 5.132 × 10 5 , T = 838 and run time = 44.831 min in Case (III).
MCHMC
Estimator A 1 ϱ , C ( X ) A 2 ϱ , C ( X ) A 1 ϱ , C ( X ) A 2 ϱ , C ( X )
(I) VaR crisis event: { S = VaR 0.99 ( S ) }
E [ X | C VaR ] 842465.49773553.738844819.90171199.334
Standard error7994.5737254.5676306.8366306.836
RVaR 0.975 , 0.99 ( X | C VaR ) 989245.360443181.466915098.833428249.307
Standard error307.85824105.16372.56820482.914
VaR 0.99 ( X | C VaR ) 989765.514500663.072915534.362615801.118
Standard error4670.96654576.957669.85396600.963
ES 0.99 ( X | C VaR ) 990839.359590093.887915767.076761038.843
Standard error679.05575024.69247.74431211.908
(II) RVaR crisis event: { VaR 0.975 ( S ) S VaR 0.99 ( S ) }
E [ X | C RVaR ] 528455.72960441.368527612.75160211.561
Standard error3978.4772119.4614032.4752995.992
RVaR 0.975 , 0.99 ( X | C RVaR ) 846956.570349871.745854461.670370931.946
Standard error1866.1336285.5232570.9979766.697
VaR 0.99 ( X | C RVaR ) 865603.369413767.829871533.550437344.509
Standard error5995.34129105.05912780.74121142.135
ES 0.99 ( X | C RVaR ) 882464.968504962.099885406.811529034.580
Standard error3061.11017346.2073134.14423617.278
(III) ES crisis event: { VaR 0.99 ( S ) S }
E [ X | C ES ] 8663863.925137671.6532934205.458140035.782
Standard error3265049.59010120.557165794.77214601.958
RVaR 0.975 , 0.99 ( X | C ES ) 35238914.131907669.46217432351.450589309.196
Standard error2892208.68931983.660443288.6493471.641
VaR 0.99 ( X | C ES ) 56612082.9051131248.05520578728.307615572.940
Standard error1353975.612119460.4111364899.75212691.776
ES 0.99 ( X | C ES ) 503537848.1922331984.18125393466.446649486.810
Standard error268007317.199468491.1271138243.1377497.200

Share and Cite

MDPI and ACS Style

Koike, T.; Hofert, M. Markov Chain Monte Carlo Methods for Estimating Systemic Risk Allocations. Risks 2020, 8, 6. https://doi.org/10.3390/risks8010006

AMA Style

Koike T, Hofert M. Markov Chain Monte Carlo Methods for Estimating Systemic Risk Allocations. Risks. 2020; 8(1):6. https://doi.org/10.3390/risks8010006

Chicago/Turabian Style

Koike, Takaaki, and Marius Hofert. 2020. "Markov Chain Monte Carlo Methods for Estimating Systemic Risk Allocations" Risks 8, no. 1: 6. https://doi.org/10.3390/risks8010006

APA Style

Koike, T., & Hofert, M. (2020). Markov Chain Monte Carlo Methods for Estimating Systemic Risk Allocations. Risks, 8(1), 6. https://doi.org/10.3390/risks8010006

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop