Next Article in Journal
Modeling High Frequency Data with Long Memory and Structural Change: A-HYEGARCH Model
Previous Article in Journal
The Role of Inflation-Indexed Bond in Optimal Management of Defined Contribution Pension Plan During the Decumulation Phase
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Misspecification Tests for Log-Normal and Over-Dispersed Poisson Chain-Ladder Models

Department of Economics, University of Oxford & Oriel College, Oxford OX1 4EW, UK
Risks 2018, 6(2), 25; https://doi.org/10.3390/risks6020025
Submission received: 28 February 2018 / Revised: 20 March 2018 / Accepted: 21 March 2018 / Published: 23 March 2018

Abstract

:
Despite the widespread use of chain-ladder models, so far no theory was available to test for model specification. The popular over-dispersed Poisson model assumes that the over-dispersion is common across the data. A further assumption is that accident year effects do not vary across development years and vice versa. The log-normal chain-ladder model makes similar assumptions. We show that these assumptions can easily be tested and that similar tests can be used in both models. The tests can be implemented in a spreadsheet. We illustrate the implementation in several empirical applications. While the results for the log-normal model are valid in finite samples, those for the over-dispersed Poisson model are derived for large cell mean asymptotics which hold the number of cells fixed. We show in a simulation study that the finite sample performance is close to the asymptotic performance.

1. Introduction

“Can we trust chain-ladder models?” is a central question in non-life insurance claim reserving. It hinges on the model assumptions: if these are violated the answer would be “no”. For example, the popular over-dispersed Poisson chain-ladder model assumes a fixed variance to mean ratio across the run-off triangle. If this is false then distribution forecasts are bound to fail. Yet, there is no statistical theory available to test for a violation of this assumption.
We show that testing for a violation of central assumptions is straightforward in two popular chain-ladder models: over-dispersed Poisson and log-normal. While the over-dispersed Poisson model assumes a fixed variance to mean ratio, the log-normal model imposes a common variance of the log data. Further, both models assume a chain-ladder structure. That is, accident year effects do not vary by development year and vice versa. We show that these assumptions are not only testable, but testable with standard tools that can easily be implemented in a spreadsheet.
The over-dispersed Poisson model arguably owes its special status to the ubiquitous chain-ladder technique. Kremer (1985) showed that this deterministic technique so commonly used in claim reserving is replicated by maximum likelihood estimation in a Poisson model. However, integer support and the implicit assumption that the variance equals the mean cannot be reconciled with insurance claim data. This explains the need for the over-dispersed Poisson model which relaxes both of these assumptions. Unlike the Poisson model, the over-dispersed Poisson model is moment-based and does not come equipped with a distributional framework. Despite this shortcoming, distribution forecasts are needed and bootstrapping (England 2002; England and Verrall 1999) is in widespread use. Yet, so far we do not have a statistical theory for the bootstrap in this setting.
Recently, Harnau and Nielsen (2017) proposed a distributional framework that incorporates the moment assumptions of the over-dispersed Poisson model. This framework allows for a compelling asymptotic theory that does not require a large array but rather large cell means. The practical implication is that for a run-off triangle with a large, potentially unknown, number of payments, we can use a fixed sample size Gaussian distribution theory. They derive parameter distributions, tests for model reduction, such as the absence of calendar effects, and closed form distribution forecasts. Their assumptions accommodate, among others, many compound Poisson distributions. In insurance, these have the interpretation that each cell of aggregate incremental claims is the sum of a Poisson number of claims each with a random individual claim amount. The asymptotic theory then does not assume that we have many such cells, but rather that the mean of the Poisson number of claims is large. We stress that while Harnau and Nielsen (2017) largely use terminology from the age-period-cohort literature, the theory immediately applies to the reserving literature by renaming age, period, and cohort effects to development, calendar, and accident effects.
Modeling aggregate incremental claims as log-normal rather than over-dispersed Poisson is also common. Kremer (1982) introduced a log-normal model with multiplicative mean structure mirroring the over-dispersed Poisson chain-ladder model. While this model does not replicate the classic chain-ladder technique, it is easily estimated by least squares. Recently, Kuang et al. (2015) derived explicit expression for the estimators in the log-normal model. These have interpretation as a geometric, rather than the classic arithmetic chain ladder. Other contributions for the log-normal model are discussed in the excellent overview of stochastic reserving models by England and Verrall (2002).
We are of course not the first to question the validity of the assumptions in these models. Yet, so far the problem was dealt with by specifying more flexible models. For example, Hertig (1983) considers a log-normal model that allows the log data variance to vary by development year. The double-chain-ladder model by Martínez Miranda et al. (2012) has, conditional on the incurred counts, an approximate over-dispersed Poisson structure where the over-dispersion varies by accident year. The “distribution-free” model by Mack (1993) has separate variance parameters for each development year. We note that while this model also replicates the classical chain-ladder point forecasts, it differs from the over-dispersed Poisson model and so far lacks a distributional framework that would allow for a rigorous statistical theory. Thus, while it is a popular model, we do not consider it further in this paper.
While using more flexible models seems sensible when assumptions are violated, we should not be too quick to dispose of well-known simple models. Particularly for forecasting, such simpler models may be advantageous. A statistical framework for misspecification testing is thus needed. The tests may corroborate the initial modeling choice of the expert, draw attention to an issue, or confirm the suspicion that the model is not well suited for the task. Whichever scenario the expert encounters, the misspecification tests can help to make an informed choice.
The test statistics we propose in this paper are well known in an analysis of variance (ANOVA) context. There, the researcher is usually presented with several samples and wants to test for treatment effects. The data are often assumed to be independent Gaussian. The first step is to test for common variances across samples. This is done with a Bartlett test based on an easily computed likelihood ratio statistic. Then, given common variances, a standard F-test can be used to test for different means between the samples, indicating a treatment effect.
The difference to the ANOVA application is that we generally have data for only one sample, often a run-off triangle. We thus reverse engineer the ANOVA situation by splitting the data into several artificial sub-samples. This idea has a long history in the econometric literature. For instance, Chow (1960) proposed a test for structural breaks that involved splitting the sample at the known breakpoint. In the (weak) instrumental variable literature, Angrist and Krueger (1995) proposed a split-sample procedure with the objective to break the bias of the instrumental variable estimator towards the ordinary least squares estimator. Figure 1 shows examples of how we could split run-off triangles into sub-samples.
In Section 2, we give a precise definition for the conditions that both the data set as well as the artificial sub-samples must meet. We note that while we do not provide guidance on how to chose the sub-sample structure in this paper, the choice does not affect the size of the proposed tests under the null hypothesis.
In a log-normal model, taking logs yields Gaussian data such that we can directly apply the Bartlett and F-test from the ANOVA scenario. While the finite sample distribution of the Bartlett test statistic has no closed form, it does not have nuisance parameters and critical values could easily be simulated. However, Bartlett (1937) suggests a χ 2 approximation to the exact distribution that allows us to sidestep simulations. For a special case with just two sub-samples, we can also apply an F-test for the hypothesis for common variances of the log data ; while Bartlett and F-tests are not identical, simulations indicate that they give similar information. Next, we show that an F-test for common mean parameters is not only straightforward but also independent of the Bartlett test. These results are collected in Section 3.
In the over-dispersed Poisson model, the asymptotic framework by Harnau and Nielsen (2017) catapults us into a finite dimensional Gaussian world. Therefore, the results developed for the log-normal model carry over. We can now asymptotically use a Bartlett test as a test for common over-dispersion across sub-samples. Similarly, an F-test for common mean parameters across sub-samples is asymptotically F-distributed and asymptotically independent of the dispersion parameter tests. We stress again that the asymptotic theory does not require a large triangle but rather large means of the cells in the triangle. As for the log-normal model, we could simulate critical values for the Bartlett test; however a χ 2 approximation can still be justified. We show all this in Section 4.
The Bartlett test is easily implemented and makes an empirical application straightforward. The same is true for an F-test on the means. We illustrate the testing procedure, splitting the data, estimating the sub-models, Bartlett testing for common dispersion parameters, and F-testing for common mean parameters, in Section 5 with several empirical applications.
We clear up remaining questions about the power of the tests and the performance of approximations in a simulation study based on a run-off triangle. First, it would not take much to simulate critical values of the Bartlett test statistic under the null, rather than to use a χ 2 approximation. However, we show in a simulation study that this approximation works so well that simulating critical values seems superfluous. Second, we produce power curves under several alternatives for the test for common variances of the log data in a log-normal model. Third, we find that the asymptotic results for the over-dispersed Poisson model are well approximated in finite samples, at least in our simulations. The simulation study is in Section 6.
Finally, we discuss some open questions for future research such as how to choose the sub-sample structure and whether one can select between over-dispersed Poisson and log-normal model. With this, Section 7 concludes the paper.

2. Data and Sub-Samples

Our aim is to test model specification by using statistics that are usually employed to test for common parameters across separate samples. However, we are presented with just a single sample, such as a run-off triangle. Thus, we artificially construct separate samples by splitting the data at hand into sub-samples. Many intuitive splits can be accommodated by the theory, for example, all sub-samples in Figure 1. Here, we define precisely the permissible structures for data and sub-samples, illustrated on an example of a run-off triangle.
For the theory in this paper, we assume that data are a generalized trapezoid as defined by Kuang et al. (2008). This flexible format allows for different numbers of accident and development years, and can accommodate missing past and future calendar years. Run-off triangles are a special case with as many accident as development years and only future calendar years missing. For accident year i and development year j, we count calendar years k with an offset so k = i + j 1 . Generalized trapezoids are characterized by the index set
I = { ( i , j ) : I l i I u , J l j J u , K l k K u } ,
where I l and I u , J l and J u , and K l and K u are the smallest and largest accident, development and calendar year indices available, respectively. We denote the number of cells in I by n. The run-off triangle in Table 1, taken from Taylor and Ashe (1983), are a generalized trapezoid with I l = J l = K l = 1, I u = J u = K u = 10 and n = 55 .
We also assume that each sub-sample is a generalized trapezoid. We denote sub-samples by I 1 , I m . The sub-samples should be disjoint so I s I t = and their union should be the original sample so I = I . All sub-samples of the examples in Figure 1 are generalized trapezoids. For instance, the sub-sample I 2 in Figure 1c is specified by I 2 l = J 2 l = 2 , I 2 u = J 2 u = 5 , K 2 l = 6 , K 2 u = 9 and n 2 = 10 .
The purpose of the generalized trapezoid assumption is to ensure parameter identification later on. We note that this assumption is often more restrictive than needed. Examples for arrays that do not fall into the generalized trapezoid category are arrays with missing cells and disconnected arrays such as the combination of sub-samples I 1 and I 3 in Figure 1b. However, for many of these arrays identification may still be given and then the theory developed below will still be valid.

3. Log-Normal Model

Given data and sub-samples, we can specify a log-normal model, define estimators, and provide the theory for specification testing. The idea is to start with a model that allows parameters to vary across sub-samples and then to test for reductions to a model with common parameters. The latter, most restrictive, model is commonly used in claim reserving. If we reject a reduction to this model, it is likely misspecified. Estimation is done by least squares. The first hypothesis is that log data variances are common across sub-samples; we can test this with a Bartlett test. The second hypothesis is for common linear predictors and can be assessed with an independent F-test.

3.1. Model and Hypotheses

The unrestricted model allows both log data means and variances to vary across sub-samples. For this model, we assume that the aggregate incremental claims Y i j , for accident year i, development year j, and sub-sample are independent log-normal with
M L N : log ( Y i j , ) = D N ( μ i j , , σ 2 ) , μ i j , = α i , + β j , + δ ( i , j ) I , { 1 , , m } .
While we focus on linear predictors μ i j , with accident and development year effect, the theory in this paper allows for more general or restrictive linear predictors. For example, we could incorporate calendar year effects as in Zehnwirth (1994) or Kuang et al. (2011).
The first hypotheses restricts log data variances to be common across sub-samples I . The remaining assumptions are maintained; thus, linear predictors are still allowed to vary across sub-samples. We write the hypothesis as
H σ 2 : σ 2 = σ 2 { 1 , , m } .
The model that arises by imposing this restriction is
M σ 2 L N : log ( Y i j , ) = D N ( μ i j , , σ 2 ) .
The second hypothesis nests the first but also restricts linear predictors to be common across sub-samples. The hypothesis is
H μ , σ 2 : σ 2 = σ 2 and μ i j , = μ i j = α i + β j + δ { 1 , , m } .
Under this hypothesis, all parameters are common across sub-samples I . Thus, we can feasibly drop the sub-script and write the model under this hypothesis as
M μ , σ 2 L N : log ( Y i j ) = D N ( μ i j , σ 2 ) .
This is the log-normal geometric chain-ladder model.
We can also think about the hypotheses on the original scale. Mean and variance parameters on the log-scale map into median and coefficients of variations on the original scale. Taking the model M μ , σ 2 L N under H μ , σ 2 as an example,
log ( Y i j ) = D N ( μ i j , σ 2 ) Median ( Y i j ) = exp ( μ i j ) , S D ( Y i j ) E ( Y i j ) = exp ( σ 2 ) 1 .
Thus, the separation between mean and variance on the log-scale translates to separation between median and coefficient of variation on the original scale. Hence, we can alternatively think of H σ 2 as the hypothesis of common coefficients of variation and of H μ , σ 2 as further imposing common median parameters.

3.2. Estimation

We estimate on the log-scale with standard estimators, least squares for log data means and residual sum of squares for log data variances. Since the theory for testing developed below is adapted from a Gaussian framework, estimation on the log-scale is intuitive. Before specifying the estimators, we briefly discuss identification.
The identification problem is that
μ i j = α i + β j + δ = ( α i + a ) + ( β j + b ) + ( δ a b )
for any a , b . Thus, the levels of accident and development effects are not identified. However, the linear predictors μ are identified (Kuang et al. 2008). These are thus invariant to the identification constraints imposed on the individual effects. Therefore, it does not matter whether we impose ad-hoc constraints such as α I l = β J l = 0 or non ad-hoc constraints as suggested by Kuang et al. (2008). We choose to discuss estimation based on the latter, which has the advantage that it allows for straightforward counting of degrees of freedom. By way of example, we apply the identification by Kuang et al. (2008) to a run-off triangle with I = { ( i , j ) : 1 i , j , k I } . Defining the first difference operator as Δ , the idea is to re-write
μ i j = μ 11 + s = 2 I 1 ( i s ) Δ α s + t = s I 1 ( j s ) Δ β s .
Then, μ i j = x i j ξ where the design vector x i j = ( 1 , 1 ( i 2 ) , , 1 ( i I ) , 1 ( j 2 ) , , 1 ( j I ) , ) and the identified parameter vector is ξ = ( μ 11 , Δ α 2 , , Δ α I , Δ β 2 , , Δ β I ) . We denote the number of parameters as p = length ( ξ ) . The identification method can be extended to generalized trapezoids as well as to linear predictors with calendar year effects.

3.2.1. Estimation in Unrestricted Model M L N

For the unrestricted model M L N , we estimate linear predictors as
μ ^ i j , L N = x i j , ξ ^ L N where ξ ^ L N = i j I x i j , x i j , 1 i j I x i j , log ( Y i j , ) .
With degrees of freedom d f = n p , we estimate log data variances by
σ ^ 2 , L N = R S S d f where R S S = i j I { log ( Y i j , ) μ ^ i j , L N } 2 .

3.2.2. Estimation with Common Variances in M σ 2 L N

Imposing the restriction of common log data variances H σ 2 does not require re-estimation as the estimators from M L N can be re-used. The estimators for the linear predictors μ i j , are identical to those of M L N . The log data variance in M σ 2 L N is estimated by
σ ¯ 2 , L N = = 1 m d f d f . σ ^ 2 , L N = R S S . d f .
where d f . = = 1 m d f and R S S . = = 1 m R S S .

3.2.3. Estimation with Common Variances and Linear Predictors in M μ , σ 2 L N

Under the hypothesis H μ , σ 2 L N , which imposes common log data mean and variance parameters, both estimators change. We drop the -subscript indicating the sub-sample since estimation is done over the full sample I . With that, we write the estimators for the linear predictors in M μ , σ 2 L N as
μ ^ i j L N = x i j ξ ^ L N with ξ ^ L N = i j I x i j x i j 1 i j I x i j log ( Y i j ) .
We estimate the log data variance σ 2 under this hypothesis, defining d f = n p , by
σ ^ 2 , L N = R S S d f where R S S = i j I { log ( Y i j ) μ ^ i j L N } 2 .

3.2.4. Remarks

Least squares estimation for the identified parameter vector ξ is maximum likelihood estimation in the log-normal model. Kuang et al. (2015) derive a representation of the least squares estimators that is interpretable as a geometric chain-ladder, in contrast to the classic, arithmetic, chain-ladder.
For many regression models, there is little difference between scaling the residual sum of squares by the degrees of freedom or the number of observations; the former yields an unbiased estimator for σ 2 , the latter the maximum likelihood estimator. However, the scaling does matter here due to the large parameter to observation ratio. By way of example, the Taylor and Ashe (1983) data has n = 55 observations but only d f = 36 degrees of freedom so that σ ^ 2 , L N is some 50% larger then the rival estimator R S S / n . This is amplified for the sub-samples.

3.3. Testing for Common Variances

We show how to test for common log data variances, that is for H σ 2 in M L N using a Bartlett test. In a special case with two sub-samples, we can use an F-test instead of a Bartlett test.
The Bartlett test (Bartlett 1937) was designed to test for common variances across several Gaussian samples. Thus, it is directly applicable to the log sub-samples. We only give a rough overview of the theory; for a more detailed derivation in contemporary terminology see Jørgensen (1993, pp. 94–96). The test rests on the independent χ 2 -distribution of σ ^ 2 , L N in M L N . Rather than deriving a test in the Gaussian model for log ( Y i j , ) , Bartlett (1937) considers a joint χ 2 model for the variance estimators. In this χ 2 model, the log-likelihood ratio statistic for the hypothesis H σ 2 is
L R L N = d f . log ( σ ¯ 2 , L N ) = 1 m d f log ( σ ^ 2 , L N )
for σ ^ 2 and σ ¯ 2 as defined in (1) and (2), respectively. Define now the Bartlett distribution Ba ( · ) such that L R L N = D Ba ( d f 1 , , d f m ) under the hypothesis. Considering L R L N as a function of the estimators so L R L N = L R ( σ ^ 1 2 , L N , , σ ^ m 2 , L N ) , the Bartlett distribution is characterized by
P { Ba ( d f 1 , , d f m ) y } = A ( y ) = 1 m d G ( x )
where G ( · ) is the χ d f 2 cdf and A ( y ) = { ( x 1 , , x m ) : L R ( x 1 , , x m ) y } . Likelihood theory tells us that Ba ( d f 1 , , d f m ) and thus L R L N approaches a χ m 1 2 as min ( d f 1 , , d f m ) goes to infinity. However, Bartlett (1937) goes a step further and suggest to divide L R L N by
C = 1 + 1 3 ( m 1 ) = 1 m 1 d f 1 d f . .
Comparing L R L N / C rather than L R L N to a χ m 1 2 substantially improves the quality of the approximation and makes it useful even in rather small samples. That is, under H σ 2 ,
B L N = L R L N C D χ m 1 2 .
The Bartlett correction factor C improves the order of magnitude of the error term. This idea has been shown to apply generally to likelihood ratio tests; see, for instance, Lawley (1956) and Barndorff-Nielsen and Cox (1984).
While using an asymptotic approximation for the Bartlett test is appealing, we could also simulate critical values of the exact distribution. This is feasible because the exact distribution of L R L N , Ba , is free of nuisance parameters. However, if Ba / C is sufficiently close to χ m 1 2 , simulating the critical values may be unnecessary even for rather small degrees of freedom. Looking ahead, we confirm in a simulation study in Section 6.1 that the asymptotic approximation indeed works very well.
As an alternative to the Bartlett test, we can test the equality of dispersion parameters across two sub-samples with an F-test that is not equivalent to a Bartlett test. The F-test follows quickly given independence and distribution of the log data variance estimators σ 2 , L N in (1). Under H σ 2 ,
F σ 2 L N = σ 2 2 , L N σ 1 2 , L N = D F d f 2 , d f 1
so that we can use a (two-sided) F-test to test the hypothesis; see, for example, Snedecor and Cochran (1967, chp. 4.15). We can write L R L N as a function of F σ 2 L N . With r = d f 2 / d f 1 ,
L R L N = L R ( F σ 2 L N ) = d f 1 log 1 + r F σ 2 L N 1 + r + d f 2 log 1 + ( r F σ 2 L N ) 1 1 + r 1 .
This mapping is not monotone. Intuitively, the Bartlett test is one-sided compared to a two-sided F-test. Thus, we would expect L R to be increasing both for small and large F σ 2 L N . We can now find scenarios in which the F-test and the Bartlett test lead to different decisions: for example, with d f 1 = 1 and d f 2 = 2 an equal-tailed 5% F-test just about rejects the null for a draw F σ 2 L N = 0.025, while a 5% Bartlett test does not reject with L R ( 0.025 ) = 4.23 and a (simulated) exact critical value of 4.91. This leaves the question which test should be used; we investigate this in Section 6.2.
Usually, a drawback of both F and Bartlett test is their sensitivity to departures from Gaussianity of the log data log ( Y i j , ) . Box (1953) goes as far as comparing the Bartlett test to a test for Gaussianity and argues in favor of robust tests, prioritizing robustness over other qualities such as power. However, sensitivity to non-Gaussianity is not necessarily undesirable for an application to insurance claim-reserving since distribution forecasts of the log-normal model would also be invalid if the data is not log-normal. Besides, we find F-test and Bartlett test appealing for their simplicity and because they carry over to over-dispersed Poisson models as we will see later. Thus, we do not consider methods to improve robustness to departures from Gaussianity such as made by Shoemaker (2003) for F-tests.

3.4. Testing for Common Linear Predictors

Now that we know how to test for common variances, we turn to testing for common linear predictors. The idea is to test sequentially: first for common variances, then for common linear predictors. We show how to use an F-tests for the latter and prove that this test is independent of Bartlett and F-tests for common variances. Thus, size control is not an issue.
If we take the model with common variances M σ 2 L N as given, then testing for H μ , σ 2 amounts to testing for common linear predictors. Since standard Gaussian theory applies,
F μ L N = ( R S S R S S . ) / ( d f d f . ) R S S . / d f . = D F d f d f . , d f .
under the hypothesis. Thus, we can use a (one-sided) F-test to test for a reduction from M σ 2 L N to M μ , σ 2 L N . Unlike the dispersion Bartlett and F-tests, this F-test is equivalent to the corresponding exact Gaussian likelihood ratio test. However, a χ 2 approximation to the likelihood ratio test may not work well due to rather few degrees of freedom. Thus, we prefer the F-test since it is easier to implement.
A sequential test approach for common variance and common linear predictors is sensible. This is because we can show the tests are independent. We formulate the independence result in a theorem; all proofs are in Appendix A.
Theorem 1.
In model M μ , σ 2 L N , the test statistic F μ L N is independent of F σ 2 L N and L R L N .
In applications, we would first conduct a, say, 5% Bartlett test for H σ 2 . Conditional on non-rejection of the hypothesis, we can conduct an F-test for H μ , σ 2 at 5% critical values and be assured that it truly has a 5% size if the hypothesis is correct.

4. Over-Dispersed Poisson

The over-dispersed Poisson model is appealing because it naturally links to the classic chain-ladder technique, unlike the log-normal model. Harnau and Nielsen (2017) developed an asymptotically framework in which the over-dispersed Poisson model is asymptotically Gaussian. Using their results, we show that finite sample results from the log-normal model hold asymptotically in the over-dispersed Poisson model. The structure of this section reflects the similarities between the log-normal and over-dispersed Poisson model. After setting up the model, we specify the estimators; these are based on a Poisson quasi-likelihood, thus replicating the chain-ladder. Before we can proceed, the over-dispersed Poisson model needs another ingredient, a sampling scheme for the asymptotic theory that we take from Harnau and Nielsen (2017). Then, we show that we can use test for common over-dispersion with a Bartlett test. Finally, we can use an F-test to test for common mean parameters. We prove that this F-test is independent of the over-dispersion test.

4.1. Model and Hypotheses

We set up a model that allows over-dispersion and mean parameters to vary across sub-samples, and specify hypotheses for common over-dispersion, and common mean parameters. This mirrors the process from the log-normal model. The key assumption of the over-dispersed Poisson model involves infinitely divisible distributions: to justify it we provide an example that is appealing for insurance claim-reserving.
We adopt the assumptions for the over-dispersed Poisson model from Harnau and Nielsen (2017). One assumption is distributional and allows for an asymptotic theory, the other imposes the desired over-dispersed Poisson chain-ladder structure. Specifically, we assume that aggregate incremental claims Y i k , are independent across ( i , k ) I and = { 1 , , m } with non-degenerate infinitely divisible distribution, at least three moments, and non-negative support. The second assumption imposes a log-linear mean and common over-dispersion within the sub-sample:
M O D P : E ( Y i j , ) = exp ( μ i j , ) , μ i j , = α i , + β j , + δ , v a r ( Y i j , ) E ( Y i j , ) = σ 2
for all ( i , j ) I and { 1 , , m } .
The first hypotheses imposes common over-dispersion parameters across sub-samples. It matches the hypothesis from the log-normal model:
H σ 2 : σ 2 = σ 2 { 1 , , m } .
The remaining assumptions are maintained. We can write the model under this assumption as
M σ 2 O D P : E ( Y i j , ) = exp ( μ i j , ) , v a r ( Y i j , ) E ( Y i j , ) = σ 2 .
The second hypothesis again nests the first and imposes common linear predictors. The hypothesis is
H μ , σ 2 : σ 2 = σ 2 and μ i j , = μ i j = α i + β j + δ { 1 , , m } .
Dropping the superfluous subscript, we write the model under this hypothesis as the familiar
M μ , σ 2 O D P : E ( Y i j ) = exp ( μ i j ) , v a r ( Y i j ) E ( Y i j ) = σ 2 .
The model under this hypothesis in a run-off triangle replicates the chain-ladder. Thus, M μ , σ 2 O D P is the model we would ideally like to use.
We can motivate the assumption of an over-dispersed infinitely divisible distribution for the aggregate incremental claims by a compound Poisson story. We can think of the aggregate incremental claims Y as a random Poisson number of claims N each with an independent random claim amount X so the Y = s = 1 N X s are compound Poisson. Compound Poisson distributions are infinitely divisible. The over-dispersion σ 2 simplifies to E ( X 2 ) / E ( X ) . Thus, it is common across the data set if the same is true for the claim amount distribution. If the claim amount distribution varies across sub-samples, so does the over-dispersion.

4.2. Estimation

With the model and hypotheses in place, we move on to estimation. The estimators match those in Harnau and Nielsen (2017). Means are estimated by Poisson quasi-likelihood, over-dispersion parameters by Poisson log-likelihood ratios. By estimating means by Poisson quasi-likelihood, we match the classic arithmetic chain-ladder forecasts in run-off triangles as Kremer (1985) showed. Just as the results for the log-normal model, the theory in this section is invariant to the identification scheme since the statistics are functions of the identified linear predictors. We choose the same identification scheme as in the log-normal model, matching the notation.

4.2.1. Estimation in Unrestricted Model M O D P

We estimate linear predictors by Poisson quasi-likelihood
μ ^ i j , O D P = x i j , ξ ^ O D P where ξ ^ O D P = arg max ξ R p i j I { Y i j , ( x i j , ξ ) exp ( x i j , ξ ) } .
The over-dispersion parameter estimators are Poisson quasi log-likelihood ratios; looking ahead, this is justified by their asymptotic χ 2 distribution. Specifically, the estimator for σ 2 is the Poisson deviance divided by the degrees of freedom. The deviance is the log-likelihood ratio against a saturated model with as many parameters as observations and perfect fit. Specifically for deviance D , the estimator for σ 2 is
σ ^ 2 , O D P = D d f where D = 2 i j I Y i j , { log ( Y i j , ) μ ^ i j , O D P } .

4.2.2. Estimation with Common Variances in M σ 2 O D P

In the model with common variances we can, as in the log-normal model, compute estimators from those for the unrestricted model. Estimators for the linear predictors μ i j , are unchanged. The estimator for the over-dispersion parameters is the degree of freedom weighted average
σ ¯ 2 , O D P = = 1 m d f d f . σ ^ 2 , O D P = D . d f .
where D . = = 1 m D and, as before, d f . = = 1 m d f .

4.2.3. Estimation with Common Variances and Linear Predictors in M μ , σ 2 O D P

In the model with common linear predictors and over-dispersion parameters, we estimate over the full sample. Dropping the subscript,
μ ^ i j O D P = x i j ξ ^ O D P where ξ ^ O D P = arg max ξ R p i j I { Y i j ( x i j ξ ) exp ( x i j ξ ) }
and
σ ^ 2 , O D P = D d f where D = 2 i j I Y i j { log ( Y i j ) μ ^ i j O D P } .

4.3. Sampling Scheme

The asymptotic theory requires a sampling scheme. The challenge is that the number of observations n grows with the number of parameters: new accident or development years would demand their own parameters. Harnau and Nielsen (2017) circumvent this problem. They propose a sampling scheme that requires the means of the cells in the data set I to grow proportionally. This is reminiscent of multinomial sampling as used, for example, by Martínez Miranda et al. (2015) in a Poisson model. Crucially, the number of observations n, thus the number of parameters, remains fixed. We adopt their sampling scheme and motivate it by a compound Poisson example.
The sampling scheme stipulates that the aggregate mean E ( Y . . ) = E ( i j I Y i j ) over the array grows in such a way that the skewness s k e w ( Y i j , ) vanishes while keeping the frequencies E ( Y i j , ) / E ( Y . . ) fixed. The requirement on the skewness is somewhat unconventional and is motivated by a limit theorem proved by Harnau and Nielsen (2017, Theorem 1).
For intuitive appeal, the skewness in the compound Poisson example from Section 4.1 vanishes as the expected number of claims grows. More precisely, considering once again aggregate incremental claims Y = s = 1 N X s with N being the random Poisson number of claims and X s the random claim amounts, the skewness of Y vanishes if the mean of the number of claims N grows for a fixed claim amount distribution X s .

4.4. Asymptotic Testing for Common Over-Dispersion

Having set up the model and sampling scheme, we turn to the asymptotic theory. We show that the asymptotic distribution of the Bartlett test and the two-sample F-test for common over-dispersion match the finite sample distribution of the test for common log data variance in the log-normal model. We can justify a χ 2 approximation to the distribution of the Bartlett test through a sequential asymptotic argument.
To test for common over-dispersion across sub-samples in the over-dispersed Poisson model, we can proceed just as is the log-normal model. This is because the asymptotic distribution of σ ^ 2 , O D P matches the exact distribution of σ ^ 2 , L N in the log-normal model (Harnau and Nielsen 2017, Lemma 1):
σ ^ 2 , O D P D σ 2 d f χ d f 2 .
Therefore, to test H σ 2 , we merely replace the estimators from the log-normal model with the over-dispersion estimators and compute
L R O D P = d f . log ( σ ¯ 2 , O D P ) = 1 m d f log ( σ ^ 2 , O D P ) .
Since the theory for the variance tests in the log-normal model hinged on the distribution of the log data variance estimators alone, we can immediately jump to the main result of the paper.
Theorem 2.
In the over-dispersed Poisson model with common over-dispersion M σ 2 O D P of Section 4.1 and Section 4.3, L R O D P converges to the Bartlett distribution Ba ( d f 1 , , d f ) from (4). Further, the F-statistic F σ 2 O D P = σ ^ 2 2 , O D P / σ ^ 1 2 , O D P is asymptotically F d f 2 , d f 1 distributed.
In Section 6.3 below, we show that finite sample approximations to the asymptotic results in Theorem 2 work well. To make the χ 2 approximation for the Bartlett test work we can use a sequential asymptotic argument. In the log-normal model, the χ 2 approximation followed through large degree of freedom asymptotics. In the over-dispersed Poisson model, we first let the aggregate mean E ( Y . . ) grow such that L R O D P / C is distributed Ba . Then, we can increase the sub-sample dimension and thus the degrees of freedom so Ba becomes χ 2 . Then, under H σ 2 , we can expect
B O D P = L R O D P C D χ m 1 2 .
A simultaneous double asymptotic theory for large E ( Y . . ) and degrees of freedom would have to wrestle with the complication that the number of mean parameters grows with the dimension of the sub-samples. Hence, such a generalization is by no means trivial and the simulations in Section 6 make it seem unnecessary.

4.5. Asymptotic Testing for Common Linear Predictors

We show how to F-test for common mean parameters. We also prove asymptotic independence of this F-test and tests for common over-dispersion.
As in the log-normal model, we can use a sequential testing strategy, first testing for H σ 2 , then for H μ , σ 2 . Harnau and Nielsen (2017, Theorem 4) showed that under H μ , σ 2 and thus in M μ , σ 2 O D P , an F-statistic has an asymptotic F-distribution:
F μ O D P = ( D D . ) / ( d f d f . ) D . / d f . D F d f d f . , d f . .
Thus, we can use a (one-sided) F-test to test for a reduction from M σ 2 O D P to M μ , σ 2 O D P . If we compare to the test in the log-normal model, we simply replaced the residual sum of squares R S S with Poisson quasi-deviances D. The difference is that the F-distribution is now asymptotic and not exact.
To justify a sequential testing approach, it is useful to show that the test is independent of the Bartlett and F-test for common dispersion, just as it was for the log-normal model.
Lemma 1.
In the over-dispersed Poisson model M μ , σ 2 O D P of Section 4.1 and Section 4.3, F μ O D P is asymptotically independent of F σ 2 O D P and L R O D P .
Therefore, under H μ , σ 2 the distribution of F μ O D P is asymptotically unaffected by conditioning on non-rejection of tests for common over-dispersion. We confirm in simulations below that this result holds approximately in finite samples. Hence, size control is not an issue in sequential testing, just as for the log-normal model.

5. Empirical Applications

To illustrate implementation of the theory we take it to the data. A run-off triangle first analyzed by Verrall et al. (2010) is appealing for a log-normal application: Kuang et al. (2015) raised the question of misspecification for this model on this data. As an over-dispersed Poisson example, we chose the data set by Taylor and Ashe (1983) in Table 1 which has become a sort of benchmark data set for this model. Verrall (1991), England and Verrall (1999), and Pinheiro et al. (2003) all use this data, to name but a few. Finally, the data by Barnett and Zehnwirth (2000) seem to require a calendar effect for modeling; we take this opportunity to demonstrate that we can easily test for specification in a model with an extended chain-ladder structure that includes a calendar effect. We use the R (R Core Team 2016) package apc (Nielsen 2015) for the empirical applications and simulations below.

5.1. Log-Normal Chain-Ladder

Kuang et al. (2015) employ a log-normal chain-ladder model for data in a run-off triangle first analyzed by Verrall et al. (2010). They remark that the largest residuals congregate within the first five accident years, indicating a potential misspecification. Verrall et al. (2010) used the data to illustrate a model that makes use of the number of reported claims that is also available; we do not make use of this information. The data relate to a portfolio of motor policies from the insurer Royal & Sun Alliance. We show this triangle in Table A1.
We take the remarks about misspecification by Kuang et al. (2015) as an opportunity to apply the specification tests for common log data variance and mean parameters. To do so, we first specify the sub-samples. Then, we set up the unrestricted model and test the hypotheses. Figure 2 summarizes the results.
Figure 2a shows how we split the data I , a run-off triangle with ten accident and development years. We split into two sub-samples: I 1 contains the first five and I 2 the last five accident years. Choosing this specific structure seems intuitive given Kuang et al. (2015) remarks about the location of large residuals.
Given the sub-samples, we specify the unrestricted independent log-normal model
M L N : log ( Y i j , ) = D N ( α i , + β j , + δ , σ 2 ) .
We first consider the hypothesis H σ 2 : σ 1 2 = σ 2 2 for a reduction to
M σ 2 L N : log ( Y i j , ) = D N ( α i , + β j , + δ , σ 2 ) .
Figure 2b shows the relevant estimates and test results. Since we have just two sub-samples, we can test the hypothesis either with a Bartlett test or an F-test for common variances. The two test give a rather similar indication. The Bartlett statistic B L N has a χ 2 p-value of 0.09 and the F-statistic F σ 2 L N a two-sided F p-value of 0.12.
If we take the variance test results as an indication not to reject H σ 2 , we can take M σ 2 L N as our primary model and test for H μ , σ 2 . That is, we test for a reduction to
M μ , σ 2 L N : log ( Y i j ) = D N ( α i + β j + δ , σ 2 ) .
Based on the F-statistic F μ L N , we cannot reject this hypothesis with a p-value of 0.91. Thus, we do not find compelling evidence against a reduction to M μ , σ 2 L N .
Alternatively, we could make use of the information that there is not just a discrepancy between the sub-samples when it comes to residuals, but that those in I 1 are larger. With this information, we could alternatively have conducted a one-sided F-test for a one-sided hypothesis H σ 2 : σ 1 2 > σ 2 2 . This test yields a p-value of 0.06, a much closer call. Note that we cannot evaluate one-sided hypotheses with a Bartlett test.

5.2. Over-Dispersed Poisson Chain-Ladder

The Taylor and Ashe (1983) data in Table 1 has served many times as an empirical application for over-dispersed Poisson chain-ladder models. Thus, it seems only appropriate to investigate the model specification. We summarize results in Figure 3.
Figure 3a shows the chosen sub-sample structure. We split the sample after the fifth accident, development, and calendar year into four sub-samples. Unlike in the case of the Verrall et al. (2010) data above, we do not have information indicating a specific sub-sample structure. While arbitrary, we find the chosen structure appealing because all sub-samples are run-off triangles themselves and of relatively similar size. Further, we hope that splits after each of the three time-scales increases our chances to find breaks. We point out that the specific sub-sample structure has no effect on the size of the tests if the hypothesis is true.
Figure 3b shows estimates and test results. The unrestricted model is the over-dispersed Poisson model discussed in Section 4.1 so that
M σ 2 O D P : E ( Y i j , ) = exp ( μ i j , ) , v a r ( Y i j , ) E ( Y i j , ) = σ 2 .
Looking at evidence for varying over-dispersion, we test for H σ 2 with a Bartlett test. While we can see quite a bit of variation in the dispersion estimates, ranging from σ ^ 4 2 , O D P = 17,592 to σ ^ 2 2 , O D P = 168,293, the test does not convincingly reject the hypothesis with a p-value of 0.08. Even though relative deviations from the degree of freedom weighted average σ ¯ 2 , O D P = 68,038 are less stark, it seems to us that making a decision by eyeballing alone would be difficult in this case.
If the Bartlett test results convince us that a reduction to M σ 2 O D P is sensible, we can test for common linear predictors. Given an F-statistic of F μ O D P = 0.46, we cannot reject this simplification with a p-value of 0.93.
Overall, the target over-dispersed Poisson model for the Taylor and Ashe (1983) data survives both misspecification tests at a 5% level for this sub-sample structure. Thus, we may be more confident now to model it with an over-dispersed Poisson chain-ladder model.
We could also opt to repeat the test for other sub-sample structures, adjusting the size to take into account that tests for different sub-sample structures on the same data are generally not independent. For example, retesting for the split into two sub-samples consider above and shown in Figure 1a. For this structure, a Bartlett test statistic of B O D P = 2.89 yields a p-value of 0.09 and an F-test statistic of F μ O D P = 0.63 a p-value of 0.64. Further, we can test for a split into three sub-samples after calendar years four and seven, similar to the structure in Figure 1b. For this structure, we get B O D P = 1.27 with a p-value of 0.53 and F μ O D P = 1.84 with a p-value of 0.11. Controlling the overall size of the thrice repeated sequential tests with a Bonferroni correction, we would reject if any p-value was below 5%/3 ≈ 0.017. This is not the case so the model survives this battery of tests as well.

5.3. Log-Normal (Extended) Chain-Ladder

As a final empirical application, we look at a run-off triangle first considered by Barnett and Zehnwirth (2000). We show this data in Table A2. These data are known to be modeled best with a predictor with not just accident and development, but also calendar effects. We look at a model with and without calendar effects. Barnett and Zehnwirth (2000) and also Kuang et al. (2011) consider a log-normal model for this data and we follow them in this choice. As before, we split the data, specify the model, and test for the hypotheses. The results are summarized in Figure 4.
Figure 4a shows the sub-sample structure we choose. Given the apparent need for calendar effects, we aim to maximize power for varying dispersion parameters along the same time dimension and split the run-off triangle, this time with eleven accident and development years, after periods five and eight into three sub-samples.
The top of Figure 4b shows estimation and test results for a model without calendar effect. This model is given by
M L N : log ( Y i j , ) = D N ( α i , + β j , + δ , σ 2 ) .
A Bartlett test for the hypothesis H σ 2 of common log data variances has a χ 2 p-value of (just under) 0.05. We may consider this as evidence against H σ 2 . For comparison with the model with calendar effect considered next, we still compute an F-test for the hypothesis H μ , σ 2 . We point out that this test is not strictly a test for common linear-predictors if we are not comfortable to accept M σ 2 L N as a model. The statistic F μ L N = 11.20 has a 0.00 p-value so that we reject H μ , σ 2 . Thus, M μ , σ 2 L N is not well specified.
At the bottom of Figure 4b we show results for a model with calendar effects γ for calendar years k = i + j 1 . The model is
M L N e : log ( Y i j , ) = D N ( α i , + β j , + γ k , + δ , σ 2 ) .
The theory for specification tests is not affected by this change and thus still valid. A Bartlett test for H σ 2 in this model yields a χ 2 p-value of 0.36 so we may feel comfortable to impose common log data variances and take M σ 2 L N e as given. An F-test for common linear predictors leaves us with a p-value of 0.41. Thus, reducing the model to M μ , σ 2 L N e seems sensible. Therefore, we cannot reject the specification of the model with calendar effect.
If we directly compare the two models, we can see that the calendar effect has a substantial impact on the specification tests. While the model with calendar effect seems to be well specified, the model without this effect raises red flags for both a test for common variances and common linear predictors. The test for common linear predictors is much more strongly affected by dropping the calendar effect than the Bartlett test. This indicates that the shift in log data variances is smaller than that in linear predictors.
We look at the shift in linear predictor in two ways. First, we can directly test for for dropping the calendar effects from the well specified M μ , σ 2 L N e . A standard F-test for the hypothesis H γ : γ k = 0 k yields a p-value of 0.00, consistent with the rejection of the model without calendar effects M μ , σ 2 L N above.
Alternatively, we can test for a reduction from M σ 2 L N e : log ( Y i j , ) = D N ( α i , + β j , + γ k , + δ , σ 2 ) to M σ 2 L N , corresponding to the hypothesis H γ k , : γ k , = 0 k , . This reduction allows for breaks in linear predictors between sub-samples. Interestingly, an F-test cannot reject H γ k , (p-value 0.92). As an intuition, we recall that the chain-ladder predictor without calendar effects can accommodate a constant trend in calendar years, but not deviations from that trend. Thus, allowing for separate sets of linear predictors on the sub-samples implicitly allows for three different calendar trends. While still less flexible than the model with an effect for each calendar year, this seems to be good enough. Note, however, that the Bartlett flags the reduction from M L N to M σ 2 L N (but not from M L N e to M σ 2 L N e ).
Overall, the analysis suggests that calendar effects are needed in this data set for two reasons for this sub-sample structure: to capture the structure of the linear predictors themselves, and, to a lesser extent, to achieve homogeneous variance across the log data.
We note that for this data, repeating the tests for different sub-samples structures does affect the results. Indeed, considering sub-samples similar to before, the specification of the log-normal extended chain-ladder model is rejected. Specifically, splitting the data into two sub-samples after the fifth accident year, a Bartlett test yields a p-value of 0.017 and an F-test a p-value of 0.004. Considering four sub-samples with splits after the fifth calendar year, the fifth development year and the sixth accident year, the p-value of the Bartlett test is 0.03 and that of the F-test 0.05. Again controlling the size of the repeated tests with a Bonferroni correction, we would reject the null hypothesis if we can find a p-value below about 0.017. This is the case for the F-test and a knife-edge decision for the Bartlett test in the two sub-sample scenario. Thus, for this data we may want to consider a different model or at least be somewhat more skeptical of its results.

6. Simulations

The developed theory begs several questions that we answer in a simulation study. First, we argued that we can sidestep simulating critical values of the Bartlett distribution Ba and instead approximate these by a Bartlett corrected χ 2 critical value. We show that this works very well. Second, we compute power curves of Bartlett and F-test for common log data variances under several alternatives in a log-normal model to get a better understanding for the tests’ behavior. Third, we show that an asymptotic approximation in an over-dispersed Poisson model resembles the asymptotic distribution closely, both under the null and the considered alternatives. Finally, we derived above that F-tests for common linear predictors in the over-dispersed Poisson model are asymptotically independent of tests for common over-dispersion. We confirm that the size of the former test seems unaffected by conditioning on the results of the latter, even in finite samples.

6.1. Performance of Bartlett test χ 2 Approximation

The theory tells us that the distribution of Ba / C , which is the exact distribution of the Bartlett statistic B L N in the log-normal model, is close to a χ 2 for large degrees of freedom. We show that the approximation works very well for a range of degrees of freedom.
We draw realizations from the adjusted Bartlett distribution Ba ( d f 1 , , d f m ) / C as follows. For = 1 , , m , we draw independent χ 2 distributed V with d f degrees of freedom and compute s = V / d f and s ¯ = = 1 m d f / d f . s . Then, { d f . log ( s ¯ ) = 1 m d f log ( s ) } / C is Ba / C distributed.
Figure 5a shows the upper 10% probability spectrum of a pp-plot for the adjusted Bartlett distribution Ba ( d f 1 , , d f m ) / C against a χ m 1 2 . We show plots for the tuples ( 26 , 6 ) , ( 3 , 5 , 8 ) , ( 6 , 6 , 9 ) , and ( 6 , 3 , 6 , 6 ) encountered in the empirical applications above. The plots are based on 10 7 draws for each tuple. The plots seem indistinguishable from the 45-degree line, even though we zoomed in to the upper 10% of the spectrum.
Figure 5b is constructed in the same way as Figure 5a, except the degrees of freedom are halved and rounded down. Now, we can see some deviations from the 45-degree line. As expected, we can see convergence to the 45-degree line as the degrees of freedom increase.
In Table 2, we take a closer look at the approximation at α = 1%, 5%, 10% critical values c α of a χ m 1 2 specifically. The table shows P ( Ba / C > c α ) , corresponding to the true size of a Bartlett test in a log-normal model if we use the χ 2 approximation rather than simulated critical values.
While we can see some differences for some of the halved critical values, we would argue that the approximation for the degree of freedom tuples from the empirical applications is so good that using it is reasonable and should not affect the modeling decision.

6.2. Rejection Frequencies of Tests for Common Variance in Log-Normal Model

As a supplement to the behavior of the tests for common log data variance under the null hypothesis in the log-normal model given in Section 3.3, we now also take a look at power. We simulate the three sub-sample structures from the empirical applications and consider rejection frequencies of the tests used in the corresponding applications. We find that the Bartlett and F-test for common variance have very similar power, at least in this simulation. Further, we see that the power does not necessarily decrease with the number of sub-samples.
For the sub-sample structures from the empirical applications (see Figure 1), we simulate M L N : log ( Y i j , ) = D N ( μ i j , , σ 2 ) . Thus, we simulate for m = 2 , 3 , 4 sub-samples. Before specifying the parameter values, we point out that the distribution of tests in this model depends only on ratios σ s 2 / σ t 2 , the degrees of freedom d f , and the number of sub-samples m. To see this, we first re-write
L R L N = d f . log ( σ ¯ 2 , L N ) = 1 m d f log ( σ ^ 2 , L N ) = = 1 m d f log d f d f . s = 1 R S S s R S S .
Now, under M L N , R S S = D σ 2 χ d f 2 independently. Thus, the distribution of L R L N is invariant to common changes in levels of σ 2 as well as to μ i j , . Therefore, we can normalize the smallest σ 2 to unity and set μ i j , = 0 without loss of generality. The distribution of the F-statistic F σ 2 L N shares these properties.
For each sub-sample scenario, we consider a range of values for the log data variance ratios σ s 2 / σ t 2 . For m > 2 sub-samples there is more than one ratio such that we cannot effectively visualize all combinations. We thus consider the following special case. For each sub-sample structure, we the compute the spacing of the estimates from the corresponding empirical application. That is, we order the empirical estimates σ ^ ( 1 ) 2 < < σ ^ ( m ) 2 and compute the m spacing-coefficients x = ( σ ^ 2 σ ^ ( 1 ) 2 ) / ( σ ^ ( m ) 2 σ ^ ( 1 ) 2 ) . We note that x ( 1 ) = 0 and x ( m ) = 1 . The spacings ( x 1 , , x m ) in the empirical examples are ( 1 , 0 ) (Verrall et al. 2010), ( 0 , 0.76 , 1 ) (Barnett and Zehnwirth 2000), and ( 0.09 , 1 , 0.58 , 0 ) (Taylor and Ashe 1983). The log data variance for the -th subset is then
σ 2 = σ ( 1 ) 2 + x ( σ ( m ) 2 σ ( 1 ) 2 ) .
To trace out power curves, we vary the largest ratio σ ( m ) 2 / σ ( 1 ) 2 from one, corresponding to H σ 2 : σ 2 = σ 2 , to twenty in 0.5 increments. As noted above, we can set σ ( 1 ) 2 = 1 without loss of generality. For each degree of freedom scenario and for each ratio σ ( m ) 2 / σ ( 1 ) 2 , we draw 10 6 sub-samples.
For each draw, we compute the test statistics used in the corresponding empirical application, F σ 2 L N as in (6) and B L N as in (5). We note that for m = 3 , we compute only the Bartlett test statistic for the model with calendar effect B L N e to make the plot less cluttered. Thus, the degrees of freedom for σ ^ 2 , L N in the three scenarios are ( 26 , 6 ) , ( 3 , 5 , 8 ) and ( 6 , 3 , 6 , 6 ) . We use χ 2 critical values for the Bartlett tests.
Figure 6 shows rejection frequencies at 5% critical values.
We can see that all tests have the right size under H σ 2 , that is for σ ( m ) 2 / σ ( 1 ) 2 = 1 . The power of two-sided F-test and Bartlett test in the two sub-sample scenario is very similar with a slight advantage for the Bartlett test. Thus, the choice between the two test may mostly depend on taste. Comparing Bartlett tests across scenarios, we see that the power for m = 4 sub-samples is larger than that for m = 3 sub-samples. Thus, fewer sub-samples do not necessarily imply higher power. Intuition comes from the degree of freedom weighting. For m = 3 sub-samples, if we drop the variance with the smallest degree of freedom the larger two variances are relatively homogeneous. Meanwhile, for m = 4 sub-samples there is still plenty of variation left among the largest three variances. Thus, since the test attributes more weight to the better estimates with higher degrees of freedom, the scenario with m = 3 sub-samples is a rather tough case.
We indicated the σ ( m ) 2 / σ ( 1 ) 2 ratios we found in the individual empirical applications by vertical lines. We recall that the spacing of intermediate variances is taken from the empirical applications. Therefore, suppose that the empirical estimates are the truth such that H σ 2 is violated. Then we can read of the power against this scenario directly from the plot. For example, in the application to the Verrall et al. (2010) data, the F-test would have a power of about 35% while the Bartlett test power would be closer to 40%.

6.3. Performance of Over-Dispersed Poisson Model Asymptotics

The theoretical results for the over-dispersed Poisson model are asymptotic, rather than exact as in the log-normal model. We show that an asymptotic approximation works well. Tests for common over-dispersion have the right size under the null. The power under the alternative in finite samples is close to the asymptotic power. Further, F-tests for common linear predictors conditional on non-rejection of over-dispersion tests are very close to F distributed in finite samples.

6.3.1. Rejection Frequencies of Tests for Common Over-Dispersion

We can use the rejection frequencies from the log-normal simulations as a benchmark for those in the over-dispersed Poisson model. To see this we recall that as the overall mean E ( Y . . ) , the over-dispersion estimator σ ^ 2 , O D P D σ 2 χ d f 2 / d f in the over-dispersed Poisson model M O D P . This matches the exact distribution of σ ^ 2 , L N in M L N . Thus, asymptotically, the distribution of L R O D P in M O D P and L R L N in M L N are identical for identical ratios σ s 2 / σ t 2 . The same holds for F σ 2 O D P and F σ 2 L N .
We simulate for the same three sub-sample structures as in the log-normal simulations. For the simulation design, we set-up an unrestricted model M O D P that satisfies the assumptions in Section 4.1. For the distribution of the cells we choose compound Poisson-gamma so Y i j , = s = 1 C i j , X s , where C i j , = D Poisson { exp ( μ i j , ) } independent of the i.i.d. gamma distributed X s , with scale σ 2 1 and shape ( σ 2 1 ) 1 . We note that the parametrization for the linear predictors μ i j , and the level of the over-dispersion σ 2 matters in finite samples. This is in contrast to the log-normal model. The reason is that the finite sample distribution of σ ^ 2 , O D P in M O D P is generally not σ 2 χ d f 2 / d f . Thus, for each considered scenario, we set the linear predictors μ i j , to the estimates μ ^ i j , O D P from the data in the corresponding empirical application. Similarly, we set the smallest over-dispersion σ ( 1 ) 2 = σ ^ ( 1 ) 2 , O D P .
We again vary the ratios σ ( m ) 2 / σ ( 1 ) 2 from one to twenty, using the exact same spacing x from (13) in the log-normal simulations so σ 2 = σ ( 1 ) 2 + x ( σ 2 σ ( 1 ) 2 ) . The only difference is that now σ ( 1 ) 2 = σ ^ ( 1 ) 2 , O D P . Therefore, asymptotically, the power for a common σ ( m ) 2 / σ ( 1 ) 2 is identical in over-dispersed Poisson and log-normal models. We draw 10 6 sub-samples for each over-dispersion ratio σ ( m ) 2 / σ ( 1 ) 2 and sub-sample structure.
Figure 7a shows the rejection frequencies at 5% critical values for the four test statistics from the empirical applications as in the log-normal model but now computed based on σ ^ 2 , O D P .
For σ ( m ) 2 / σ ( 1 ) 2 = 1 we are under the null; we can see that the rejection frequencies are very close to 5% so the tests have the correct size. Under the alternative where σ ( m ) 2 / σ ( 1 ) 2 > 1 , the ordering of the rejection frequencies matches that in the log-normal simulations (Figure 6). Generally, the plot is reassuringly reminiscent of its equivalent in the log-normal simulations.
Figure 7b shows the gap to the asymptotic rejection frequencies that arises in the finite sample simulations. The simulation set-up implies that this is the difference between rejection frequencies in log-normal and over-dispersed Poisson simulations. Thus, the plots shows the impact of the asymptotic approximation in the over-dispersed model. Since the difference under the alternative is positive throughout, the power in the over-dispersed model is lower than in the log-normal model. We next interpret the plots under the alternative in turn for the three sub-sample scenarios.
For m = 2 , the power gap of Bartlett and F-test initially increases with σ ( m ) 2 / σ ( 1 ) 2 , hitting 10 p p (percentage points)for the F-test at σ ( 2 m ) / σ ( 2 1 ) 7.5 , before it decreases. The initial increase relates to the asymptotic theory by Harnau and Nielsen (2017) which assumes fixed dispersion parameters. Since we keep σ ( 1 ) 2 constant, the remaining dispersion parameters grow with the ratio. Thus, we would require larger cell-means to achieve the same asymptotic approximation quality. The later decrease reflects the upper bound of one for the power: even as the asymptotic approximation becomes worse, the difference between dispersion parameters becomes so large that it is easily caught. For m = 4 , the power gap is increasing throughout the considered range for σ ( m ) 2 / σ ( 1 ) 2 . The intuition for the increase again comes from the asymptotic theory. We do not see a decrease since the power is still quite far from unity, staying below 80% even for the largest maximum to minimum ratio. Meanwhile, for m = 3 , the power gap is essentially zero so that the finite sample power matches the asymptotic power. The intuition for this follows because the dispersion to mean ratio is small. As a rough indication, dividing the largest considered dispersion 20 · σ ^ ( 1 ) 2 , O D P by the mean over all cells n 1 i j Y i j yields 0.8% for the Barnett and Zehnwirth (2000) simulations compared with 70% and 56% for the Verrall et al. (2010) and Taylor and Ashe (1983) simulations, respectively.
We again indicate the power at the particular alternative generated by taking the estimates in the empirical applications as true values by vertical lines. Figure 7b shows that for these alternatives, the power for all asymptotic approximations is within 5 p p of their asymptotic power.

6.3.2. Independence of Test for Common Linear Predictors

We move on to evaluate the quality of a finite sample approximation to the asymptotic independence in Lemma 1. Specifically, we consider the finite sample distribution of F μ O D P as in (11) given that a tests for common over-dispersion did not reject. Arguably, this is the most interesting case since it matches the natural order of the two specification tests.
We simulate under the null H μ , σ 2 , that is for a model with common linear predictors and over-dispersion M μ , σ 2 O D P . As before, cells Y i j are compound Poisson-gamma. We consider three scenarios, setting the parameters to the estimates for M μ , σ 2 O D P in the three empirical examples. We draws 106 triangles per scenario.
For each draw, we compute tests based on the sub-sample structure of the corresponding empirical application. We first conduct a Bartlett test for H σ 2 at 5% critical values. If we do not reject H σ 2 based on this test, we keep the triangle, otherwise we throw it out. Since we simulate under the null hypothesis, we thus keep about 95% of the draws. Only for the draws we keep do we compute the F-statistic for common linear predictors F μ O D P .
Figure 8 shows a pp-plot for the F μ O D P against F d f d f . , d f . for the triangles that survived Bartlett testing.
To be able to tell a difference from the 45-degree line, we limit our attention to the upper 10% of the probability spectrum since. This is also the most interesting range for testing. Even in this spectrum, each plot is very close to the 45-degree line. Therefore, under H μ , σ 2 , we can be reassured that an F-test for common linear predictors has the correct size in finite samples even if we apply it only conditionally on non-rejection of a test for common over-dispersion.

6.4. Remark

We note that all simulations are for tests that consider the correct sub-sample structure under the alternative. Of course, this does not seem realistic in applications. However, for tests computed on a given sub-sample structure, it appears we would generally be able to choose a true, different, sub-sample structure against which the tests would at best have limited power. For example, say we compute the tests on the two sub-samples with a split after the fifth accident year in Figure 1a while really there are three sub-samples with an additional split after the fifth development year. Then, we could choose parameterizations for the three true true sub-samples to balance out the variation between the two incorrectly chosen sub-samples, thus minimizing power. Therefore, it seems to us that such simulation results would be almost entirely driven by our chosen parametrization and provide little insight beyond that. We believe the real answer to this problem must come from a theory that is agnostic to the sub-sample structure as discussed below. However, we stress again that the size of the tests under the null hypothesis is not affected by the chosen sub-sample structure.

7. Discussion

Some questions are left open for future research. For example, it is not clear how to best choose the sub-sample structure and the number of sub-samples. Further, the question arises whether we can somehow select between the over-dispersed Poisson and log-normal model. Finally, a misspecification test for independence of the cells would be a useful addition to the modeling toolkit.
So far, we chose the sub-sample structures somewhat arbitrarily if potentially informed by prior knowledge of the data. While the size of the tests under the null is not affected by the sub-sample structure, the power of the tests under the alternative is affected both by the chosen number of sub-samples and their structure. In applications, the expert may consider choosing a range of sub-samples structures and conducting tests for each, adjusting the size based on the number of tests to account for multiple testing as shown in the empirical applications. For future research, it would be useful to derive a theory that is agnostic to the number of sub-samples and their structure while still directly controlling size. It might be fruitful to look for ideas in time-series econometrics which has been concerned with tests for parameter breaks for a long time. In this literature, Chow (1960) had proposed a test for parameter breaks that required knowledge of the breakpoint. By now, there are several test available that are agnostic with respect to the number of breaks, related to the number of sub-samples in our problem, and the position of breaks, akin to the sub-sample structure. Examples include Andrews’ test (Andrews 1993), generalizations of Chow tests (Nielsen and Whitby 2015), and indicator saturation (Hendry 1999). However, these tests are designed for data with a single time-scale and results are generally based on long time-series. In contrast, we are confronted with data with three interlinked time-scales and the arrays are often small with a large number of parameters that is growing with the array size. Thus, the known results do not carry over and a it appears that a new theory is needed.
Since we have seen two models in this paper, log-normal and over-dispersed Poisson, a natural question is when we should choose which model. As we have seen, the log-normal model assumes a fixed standard deviation to mean ratio while the over-dispersed Poisson model considers the variance to mean ratio to be fixed. Making use of recent results for generalized log-normal models by Kuang and Nielsen (2018), a class of models that includes the log-normal but is more general, Harnau (2018) proposes a test to distinguish between (generalized) log-normal and over-dispersed Poisson models based on this discrepancy.
Finally, a misspecification test for the assumption that the cells in the array are independent would be useful. This is an assumption that both the log-normal and the over-dispersed Poisson model impose. In contrast, the “distribution free” model by Mack (1993) relaxes this somewhat, assuming independence only across accident years.

Acknowledgments

The author thanks two anonymous referees for their helpful comments. Discussions with Bent Nielsen (Department of Economics, University of Oxford & Nuffield College, UK) are gratefully acknowledged. The author was supported by the Economic and Social Research Council, grant number ES/J500112/1, and the European Research Council, grant AdG 694262.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

Appendix A.1. Proof of Theorem 1

The proof relies on two properties. First, R S S R S S . , the numerator of F μ L N , reduces to a comparison of least squares fitted log-means = 1 m i j I ( μ ^ i j L N μ ^ i j , L N ) 2 , and is therefore, in the Gaussian framework at hand, independent of the residual sum of squares R S S 1 , , R S S m .
Second, the denominator of F μ L N , the aggregated residual sum of squares R S S . and the relative contributions π 1 , , π m for π = R S S / R S S . are mutually independent. To see this, we first recall that under the hypothesis R S S 1 , , R S S m are independent σ 2 χ 2 . The proof is unaffected by setting σ 2 = 1 . Thus, let X 1 , X m be independent χ 2 . Recall that the sum of independent χ 2 ’s is χ 2 . Let S = s = 1 X s and V = S 1 / S . We note that we can map V 2 , , V m to the frequencies X 1 / S m , , X m / S m . For the special case with m = 2 , Johnson et al. (1995, p. 212) note the independence of S 2 and V 2 . The general case for independence of S m and V 2 , , V m can be proved by induction. Here, we partially replicate the (originally Danish) argument from Andersson and Jensen (1987, p. 180). We show the induction step from m 1 to m. Suppose V 2 , , V m 1 , S m 1 and X m are independent. Then S m = S m 1 + X m and V m = S m 1 / S m . S m and V m match the setting for the special case with m = 2 from above and are thus independent. Hence, V 2 , , V m and S m are independent, completing the induction step. Independence of V 2 , , V m and S m implies independence of π 1 , , π m and R S S . ,
Taken together, R S S R S S . , R S S . , and π 1 , , π m , are mutually independent. Now, we can write the test statistics for the dispersion parameters as functions of the relative contributions π :
L R L N = L R ( π 1 , , π ) = = 1 m d f log d f d f . log ( π ) , F σ 2 L N = d f 1 d f 2 π 2 π 1 .
Thus, F μ L N is independent of F σ 2 L N and L R L N .

Appendix A.2. Proof of Theorem 2

This follows from (8), independence of d across due to disjoint sub-samples made up of independent Y i j , the continuous mapping theorem, and the results discussed in Section 3.

Appendix A.3. Proof of Lemma 1

Once we show that D D . , D . and D 1 / D . , , D m / D . are asymptotically mutually independent, the result follows from the proof of Theorem 1 since the asymptotic distribution of the deviances in the over-dispersed Poisson model matches the exact distribution of the residual sum of squares in the log-normal model.
We can set σ 2 = 1 without loss of generality. Then, to prove mutual independence, we build on the insight of Harnau and Nielsen (2017) that asymptotics for the over-dispersed Poisson model match standard exponential family asymptotics. Thus, D D . and D 1 , , D m are asymptotically equivalent to quadratic forms (Johansen 1979, Theorem 7.8) of asymptotically Gaussian projections on orthogonal subspaces (Johansen 1979, Theorem 7.6). Thus, independence and hence Lemma 1 follows.
Table A1. Insurance run-off triangle taken from Verrall et al. (2010, Table 1) as used in the empirical application in Section 5.1 and the simulations in Section 6.
Table A1. Insurance run-off triangle taken from Verrall et al. (2010, Table 1) as used in the empirical application in Section 5.1 and the simulations in Section 6.
i, j12345678910
1451,288339,519333,371144,988093,243045,51125,21720,40631,4821729
2448,627512,882168,467130,674056,044033,39756,07126,52214,346-
3693,574497,737202,272120,753125,046037,15427,60817,864--
4652,043546,406244,474200,896106,802106,75363,688---
5566,082503,970217,838145,181165,519091,313----
6606,606562,543227,374153,551132,743-----
7536,976472,525154,205150,564------
8554,833590,880300,964-------
9537,238701,111--------
10684,944---------
Table A2. Insurance run-off triangle taken from Barnett and Zehnwirth (2000, Table 3.5) as used in the empirical application in Section 5.1 and the simulations in Section 6.
Table A2. Insurance run-off triangle taken from Barnett and Zehnwirth (2000, Table 3.5) as used in the empirical application in Section 5.1 and the simulations in Section 6.
i, j1234567891011
1153,638188,412134,534087,456060,34842,40431,23821,25216,62214,44012,200
2178,536226,412158,894104,686071,44847,99035,57624,81822,66218,000-
3210,172259,168188,388123,074083,38056,08638,49633,76827,400--
4211,448253,482183,370131,040078,99460,23245,56838,000---
5219,810266,304194,650120,098087,58262,75051,000----
6205,654252,746177,506129,522096,78682,400-----
7197,716255,408194,648142,328105,600------
8239,784329,242264,802190,400-------
9326,304471,744375,400--------
10420,778590,400---------
11496,200----------
References

References

  1. Andersson, Steen A., and Søren T. Jensen. 1987. Forelæsningsnoter i Sandsynlighedsregning, 3rd ed. Copenhagen: Institute of Mathematical Statistics, University of Copenhagen. [Google Scholar]
  2. Andrews, Donald W. K. 1993. Tests for Parameter Instability and Structural Change with Unknown Change Point. Econometrica 61: 821–56. [Google Scholar] [CrossRef]
  3. Angrist, Joshua D., and Alan B. Krueger. 1995. Split-sample instrumental variables estimates off the return to schooling. Journal of Business and Economic Statistics 13: 225–35. [Google Scholar]
  4. Barndorff-Nielsen, Ole E., and David R. Cox. 1984. Bartlett Adjustments to the Likelihood Ratio Statistic and the Distribution of the Maximum Likelihood Estimator. Journal of the Royal Statistical Society. Series B (Methodological) 46: 483–95. [Google Scholar]
  5. Bartlett, Maurice S. 1937. Properties of Sufficiency and Statistical Tests. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 160: 268–82. [Google Scholar] [CrossRef]
  6. Barnett, Glen, and Ben Zehnwirth. 2000. Best estimates for reserves. Proceedings of the Casualty Actuarial Society 87: 245–321. [Google Scholar]
  7. Box, George E. P. 1953. Non-Normality and Tests on Variances. Biometrika 40: 318–35. [Google Scholar] [CrossRef]
  8. Chow, Gregory C. 1960. Tests of Equality Between Sets of Coefficients in Two Linear Regressions. Econometrica 28: 591. [Google Scholar] [CrossRef]
  9. England, Peter D. 2002. Addendum to “Analytic and bootstrap estimates of prediction errors in claims reserving”. Insurance: Mathematics and Economics 31: 461–66. [Google Scholar] [CrossRef]
  10. England, Peter D., and Richard J. Verrall. 1999. Analytic and bootstrap estimates of prediction errors in claims reserving. Insurance: Mathematics and Economics 25: 281–93. [Google Scholar] [CrossRef]
  11. England, Peter D., and Richard J. Verrall. 2002. Stochastic Claims Reserving in General Insurance. British Actuarial Journal 8: 443–518. [Google Scholar] [CrossRef]
  12. Harnau, Jonas. 2018. Log-Normal or Over-Dispersed Poisson? Unpublished manuscript. UK: University of Oxford. [Google Scholar]
  13. Harnau, Jonas, and Bent Nielsen. 2017. Over-dispersed age-period-cohort models. Journal of the American Statistical Association. [Google Scholar] [CrossRef]
  14. Hendry, David F. 1999. An Econometric Analysis of US Food Expenditure. In Methodology and Tacit Knowledge: Two Experiments in Econometrics. Edited by Jan R. Magnus and Mary S. Morgan. New York: Wiley, pp. 341–61. [Google Scholar]
  15. Hertig, Joakim. 1983. A Statistical Approach To Ibnr-Reserves in Marine Reinsurance. ASTIN Bulletin 15: 171–83. [Google Scholar] [CrossRef]
  16. Johansen, Søren. 1979. Introduction to the Theory of Regular Exponential Families. Copenhagen: Institute of Mathematical Statistics, University of Copenhagen. [Google Scholar]
  17. Johnson, Norman L., Samuel Kotz, and Narayanaswamy Balakrishnan. 1995. Continuous Univariate Distributions Volume 2, 2nd ed. Chichester: Wiley. [Google Scholar]
  18. Jørgensen, Bent. 1993. The Theory of Linear Models. New York: Chapman & Hall. [Google Scholar]
  19. Kremer, Erhard. 1985. Einführung in die Versicherungsmathematik, 7th ed. Göttingen: Vandenhoeck & Ruprecht, pp. 130–36. [Google Scholar]
  20. Kremer, Erhard. 1982. IBNR-claims and the two-way model of ANOVA. Scandinavian Actuarial Journal 1982: 47–55. [Google Scholar] [CrossRef]
  21. Kuang, Di, and Bent Nielsen. 2018. Generalized Log Normal Chain-Ladder. Unpublished manuscript. UK: University of Oxford. [Google Scholar]
  22. Kuang, Di, Bent Nielsen, and Jens P. Nielsen. 2011. Forecasting in an Extended Chain-Ladder-Type Model. Journal of Risk and Insurance 78: 345–59. [Google Scholar] [CrossRef]
  23. Kuang, Di, Bent Nielsen, and Jens P. Nielsen. 2008. Identification of the age-period-cohort model and the extended chain-ladder model. Biometrika 95: 979–86. [Google Scholar] [CrossRef]
  24. Kuang, Di, Bent Nielsen, and Jens P. Nielsen. 2015. The geometric chain-ladder. Scandinavian Actuarial Journal 2015: 278–300. [Google Scholar] [CrossRef]
  25. Lawley, Derrick N. 1956. A general method for approximating to the distribution of likelihood ratio criteria. Biometrika 43: 295–303. [Google Scholar] [CrossRef]
  26. Mack, Thomas. 1993. Distribution free calculation of the standard error of chain ladder reserve estimates. ASTIN Bulletin 23: 213–25. [Google Scholar] [CrossRef]
  27. Martínez Miranda, Maria D., Bent Nielsen, and Jens P. Nielsen. 2015. Inference and forecasting in the age-period-cohort model with unknown exposure with an application to mesothelioma mortality. Journal of the Royal Statistical Society: Series A (Statistics in Society) 178: 29–55. [Google Scholar]
  28. Martínez Miranda, Maria D., Jens P. Nielsen, and Richard J. Verrall. 2012. Double Chain Ladder. ASTIN Bulletin 42: 59–76. [Google Scholar]
  29. Nielsen, Bent. 2015. apc: An R Package for Age-Period-Cohort Analysis. The R Journal 7: 52–64. [Google Scholar]
  30. Nielsen, Bent, and Andrew Whitby. 2015. A Joint Chow Test for Structural Instability. Econometrics 3: 156–86. [Google Scholar] [CrossRef]
  31. Pinheiro, Paulo J. R., João M. Andrade e Silva, and Maria de Lourdes Centeno. 2003. Bootstrap methodology in claim reserving. Journal of Risk and Insurance 70: 701–14. [Google Scholar] [CrossRef]
  32. R Core Team. 2016. R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing. [Google Scholar]
  33. Shoemaker, Lewis H. 2003. Fixing the F Test for Equal Variances. The American Statistician 57: 105–14. [Google Scholar] [CrossRef]
  34. Snedecor, George W., and William G. Cochran. 1967. Statistical Methods, 6th ed. Ames: The Iowa State University. [Google Scholar]
  35. Taylor, Greg C., and Frank R. Ashe. 1983. Second moments of estimates of outstanding claims. Journal of Econometrics 23: 37–61. [Google Scholar] [CrossRef]
  36. Verrall, Richard J. 1991. On the estimation of reserves from loglinear models. Insurance: Mathematics and Economics 10: 75–80. [Google Scholar] [CrossRef]
  37. Verrall, Richard J., Jens P. Nielsen, and Anders H. Jessen. 2010. Prediction of RBNS and IBNR claims using claim amounts and claim counts. ASTIN Bulletin 40: 871–87. [Google Scholar]
  38. Zehnwirth, Ben. 1994. Probabilistic development factor models with applications to loss reserve variability, prediction intervals and risk based capital. Insurance: Mathematics and Economics 15: 82. [Google Scholar]
Figure 1. Examples for splits of run-off triangles into two (a), three (b) and four (c) sub-samples. Sub-samples are denoted by I . Accident years i are in the rows, development years j in the columns.
Figure 1. Examples for splits of run-off triangles into two (a), three (b) and four (c) sub-samples. Sub-samples are denoted by I . Accident years i are in the rows, development years j in the columns.
Risks 06 00025 g001
Figure 2. Log-normal chain-ladder model for Verrall et al. (2010) data. Sub-sample structure shown in (a), estimation and test results in (b).
Figure 2. Log-normal chain-ladder model for Verrall et al. (2010) data. Sub-sample structure shown in (a), estimation and test results in (b).
Risks 06 00025 g002
Figure 3. Over-dispersed Poisson chain-ladder model for Taylor and Ashe (1983) data. Sub-sample structure shown in (a), estimation and test results in (b).
Figure 3. Over-dispersed Poisson chain-ladder model for Taylor and Ashe (1983) data. Sub-sample structure shown in (a), estimation and test results in (b).
Risks 06 00025 g003
Figure 4. Log-normal chain-ladder ( L N ) and extended chain-ladder ( L N e ) model for Barnett and Zehnwirth (2000) data. Sub-sample structure shown in (a), estimation and test results in (b).
Figure 4. Log-normal chain-ladder ( L N ) and extended chain-ladder ( L N e ) model for Barnett and Zehnwirth (2000) data. Sub-sample structure shown in (a), estimation and test results in (b).
Risks 06 00025 g004
Figure 5. pp-plots for the adjusted Bartlett distribution Ba / C against χ 2 for varying degrees of freedom. (a) and (b) show results for degrees of freedom corresponding to the empirical applications and half those degrees of freedom, respectively.
Figure 5. pp-plots for the adjusted Bartlett distribution Ba / C against χ 2 for varying degrees of freedom. (a) and (b) show results for degrees of freedom corresponding to the empirical applications and half those degrees of freedom, respectively.
Risks 06 00025 g005
Figure 6. Power curves for log-normal dispersion tests based on sub-sample structures from empirical applications. Empirical maximum to minimum ratios indicated by horizontal lines. BZ is short for Barnett and Zehnwirth (2000), VNJ for Verrall et al. (2010), and TA for Taylor and Ashe (1983).
Figure 6. Power curves for log-normal dispersion tests based on sub-sample structures from empirical applications. Empirical maximum to minimum ratios indicated by horizontal lines. BZ is short for Barnett and Zehnwirth (2000), VNJ for Verrall et al. (2010), and TA for Taylor and Ashe (1983).
Risks 06 00025 g006
Figure 7. Power gap for log-normal dispersion tests based on sub-sample structures from empirical applications. Empirical maximum to minimum ratios indicated by horizontal lines. Rejection frequencies shown in (a), gap to asymptotic rejection frequencies in (b).
Figure 7. Power gap for log-normal dispersion tests based on sub-sample structures from empirical applications. Empirical maximum to minimum ratios indicated by horizontal lines. Rejection frequencies shown in (a), gap to asymptotic rejection frequencies in (b).
Risks 06 00025 g007
Figure 8. Distribution of F μ O D P conditional on non-rejection of a 5% Bartlett test.
Figure 8. Distribution of F μ O D P conditional on non-rejection of a 5% Bartlett test.
Risks 06 00025 g008
Table 1. Insurance run-off triangle taken from Taylor and Ashe (1983) as an example for a generalized trapezoid. Entries are aggregate incremental paid amounts for claims of accident year i and development year j. Calendar years k = i + j 1 are on the diagonals increasing from the top left.
Table 1. Insurance run-off triangle taken from Taylor and Ashe (1983) as an example for a generalized trapezoid. Entries are aggregate incremental paid amounts for claims of accident year i and development year j. Calendar years k = i + j 1 are on the diagonals increasing from the top left.
i, j12345678910
1357,8480766,9400610,5420482,940527,326574,398146,342139,950227,22967,948
2352,1180884,0210933,8941,183,289445,745320,996527,804266,172425,046-
3290,5071,001,7990926,2191,016,654750,816146,923495,992280,405--
4310,6081,108,2500776,1891,562,400272,482352,053206,286---
5443,1600693,1900991,9830769,488504,851470,639----
6396,1320937,0850847,4980805,037705,960-----
7440,8320847,6311,131,3981,063,269------
8359,4801,061,6481,443,370-------
9376,6860986,608--------
10344,014---------
Table 2. P ( Ba / C > c α ) where c α is the χ 2 α critical value. Results are in %. Degrees of freedom shown as ( d f 1 , , d f m ) .
Table 2. P ( Ba / C > c α ) where c α is the χ 2 α critical value. Results are in %. Degrees of freedom shown as ( d f 1 , , d f m ) .
(13, 3)(1, 2, 4)(3, 3, 4)(3, 1, 3, 3)(26, 6)(3, 5, 8)(6, 6, 9)(6, 3, 6, 6)
α = 10%9.949.059.869.209.989.939.979.92
α = 5%4.934.224.854.374.984.924.974.92
α = 1%0.950.690.920.760.990.950.980.96

Share and Cite

MDPI and ACS Style

Harnau, J. Misspecification Tests for Log-Normal and Over-Dispersed Poisson Chain-Ladder Models. Risks 2018, 6, 25. https://doi.org/10.3390/risks6020025

AMA Style

Harnau J. Misspecification Tests for Log-Normal and Over-Dispersed Poisson Chain-Ladder Models. Risks. 2018; 6(2):25. https://doi.org/10.3390/risks6020025

Chicago/Turabian Style

Harnau, Jonas. 2018. "Misspecification Tests for Log-Normal and Over-Dispersed Poisson Chain-Ladder Models" Risks 6, no. 2: 25. https://doi.org/10.3390/risks6020025

APA Style

Harnau, J. (2018). Misspecification Tests for Log-Normal and Over-Dispersed Poisson Chain-Ladder Models. Risks, 6(2), 25. https://doi.org/10.3390/risks6020025

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop