Next Article in Journal
Transition Risk in Climate Change: A Literature Review
Previous Article in Journal
Do Board Characteristics Matter with Greenwashing? An Investigation in the Financial Sector with the Integration of Entropy Weight and TOPSIS Multicriteria Decision-Making Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Exponential Dispersion Family (EDF) Chain Ladder and Data Granularity

School of Risk and Actuarial Studies, University of New South Wales, Randwick, NSW 2052, Australia
Risks 2025, 13(4), 65; https://doi.org/10.3390/risks13040065
Submission received: 29 November 2024 / Revised: 18 March 2025 / Accepted: 20 March 2025 / Published: 27 March 2025

Abstract

:
This paper is concerned with the choice of data granularity for application of the EDF (Exponential Dispersion Family) chain ladder model to forecast a loss reserve. As the duration of individual accident and development periods is decreased, the number of data points increases, but the volatility of each point increases. This leads to a question as to whether a decrease in time unit leads to an increase or decrease in the variance of the loss reserve estimate. Is there an optimal granularity with respect to the variance of the loss reserve? A preliminary question is that of whether an EDF chain ladder that is valid for one duration (here called mesh size) remains so for another. The conditions under which this is so are established. There are various ways in which the mesh size of a data triangle may be varied. The paper identifies two of particular interest. For each of these two types of variation, the effect on variance of loss reserve is studied. Subject to some technical qualifications, the conclusion is that an increase in mesh size always increases the variance. It follows that one should choose a very high degree of granularity in order to maximize efficiency of loss reserve forecasting.

1. Introduction

1.1. Background

The chain ladder is a widely used model for insurance loss reserving (Mack 1993; Taylor 1985, 2000; Wüthrich and Merz 2008). The data set to which it is applied is typically triangular, with rows labelled by accident period and columns by development period.
Accident and development periods are commonly years, but other units of time are possible, e.g., quarters, months, weeks, etc. As the time unit is decreased, the number of data points increases, but the variance of each point increases. This leads to a question as to whether a decrease in time unit leads to an increase or decrease in the variance of the loss reserve estimate. A typical question faced by a practitioner might be the following one.
Example 1. 
Past chain ladder analyses of data triangles have been performed on quarterly data. The next valuation of liability could be conducted in the same way. However, monthly data are also available, and so application of the chain ladder to a monthly data triangle is also feasible. Is this likely to lead to a more or less reliable (as measured by predictive variance) estimate of liability?
A natural companion to this question is whether there is an optimal time unit at which the variance of the forecast loss reserve is minimized.
This question arises in numerous scientific contexts, e.g., statistics, energy, demand forecasting, finance, medicine, traffic flow and air quality, to name some.
Cases that closely mirror the question investigated in this paper include Kim et al. (2022), Z. Wang et al. (2023), Das et al. (2022), Grolinger et al. (2016), Li et al. (2019) and Jain et al. (2014). All of these point out that the calibration of statistical models can depend on the choice of data granularity, and propose methods for making this choice.
Other papers are not specifically concerned with optimization of granularity, but still point out that it can affect accuracy and efficiency of modelling. Examples are Elton et al. (2010), Huang and Zhu (2016), Lusis et al. (2017) and J. Wang et al. (2020).
While some of the actuarial literature discusses the choice of granularity of data for claim modelling in general terms, it remains silent on the specific questions just stated.
Even practitioner textbooks offer little. For example, an American text (Peterson 1981, p. 122) makes the following statement:
Annual exposure periods [accident periods] and development intervals are used in this example, but shorter periods can also be used; quarterly semi-annual or monthly periods are the most common, other than annual.”
Similarly, an Australian text (Hart et al. 1996, p. 226) includes the statement:
The difference between the payment year and the accident year is referred to as the development year. Other time periods can also be used particularly for short-tail classes.”
Both of these volumes refer to the possibility of different data granularities, but with no advice on the selection of a specific one. Associated general comment implies that the length of development periods might be chosen to be broadly proportional to the average delay from occurrence to closure of claims, but no reasoning for this is given.

1.2. Purpose of the Paper

The present paper aims to address those questions. It is necessary that a discussion of the chain ladder recognize two quite distinct formulations of that model. These are:
These two models differ in a number of ways, but the main difference relevant to the present discussion is that, in the former, a distribution (EDF, in fact) is placed on each data point, whereas in the latter observations are distribution-free (though subject to an assumed variance).
The questions raised in Section 1.1 can be addressed for each type of model, but the difference just identified necessitates different mathematical approaches. As the title of the present paper indicates, it deals with the EDF chain ladder model, and especially the sub-models Tweedie chain ladder and Poisson chain ladder. A sequel paper deals with the Mack chain ladder.
The present paper thus discusses the influence of data granularity on the variance of the loss reserve forecast by the EDF chain ladder. Under suitable conditions, it is possible to identify the granularity that minimizes this variance.

1.3. Layout of the Paper

The paper considers changes of mesh size on the triangular data set, which is to say changes in the units of time used for accident and development periods. The effect of an increase in mesh size (i.e., an increase in the length of time spanned by one of these periods) on the variance of the estimated loss reserve is investigated.
There are various ways in which the mesh size of a data set can be varied. Some are more sensible than others, and two fundamental types of variation are considered here. Both, though subject to limitation, are extremely versatile. One is of a type usually found in practice; the other, though not unknown in practice, is rarer, but possessed of one or two useful properties. These types of variation are described in Section 2.
Since, under the EDF chain ladder, observations come equipped with distributions, sufficient statistics can be sought and put to use. These types of statistics are also discussed in Section 2.
Section 3 largely recites the background literature, introducing the EDF, Tweedie and Poisson chain ladders, their parameter estimation algorithms, and discusses the sufficient statistics associated with each. Section 4 and Section 5 then contain this paper’s own further development, examining the effect of changed mesh size on the variance of estimated loss reserve. These two sections deal with the two types of mesh variation just described, and how they influence the variance of the EDF chain ladder loss reserve. A numerical example of the findings of Section 5 is given in Section 6. Finally, Section 7 summarizes and discusses the findings of these sections.

2. Notation and Mathematical Preliminaries

2.1. Fundamentals

Let i = 1 , 2 , , I denote accident period and j = 0 , 1 , , I 1 development period. Let Y i j , with mean μ i j , denote the random variable representing the amount of claim payments made during development period j of accident period i . These are usually referred to as incremental claim payments. All accident and development periods are of equal duration, though this will change later.
A calendar period consists of all those combinations of i and j such that i + j is constant. For definiteness, label the calendar t if i + j = t .
It will be assumed that one stands at the end of calendar period I , in possession of certain data from calendar periods 1 , 2 , , I . Specifically, these data comprise the claim triangle  D U = Y i j : i = 1 , , I , j = 0 , , I i . This will be referred to as the upper triangle. One wishes to use these data to forecast the lower triangle  D L = Y i j : i = 2 , , I , j = I i + 1 , , I 1 .
In a loss reserving context, a realization of the upper triangle will have been observed. This will be denoted d U , which is the same as D U with each random variable Y i j replaced by its realization y i j .
It will also be convenient to define cumulative claim payments. Thus define
C i j = k = 0 I j Y i k ,
which is the random variable representing the amount of cumulative claim payments up to the end of development period j of accident period i . Further, let c i j be defined in the same way in terms of the y i k , so that c i j is a realization of C i j .
It is assumed that there is no claim activity beyond development period I 1 , in which case C i , I 1 is the ultimate claim cost of accident period i .
The objective of the loss reserving exercise is to forecast the unobserved lower triangle. This will be denoted D ^ L , which is the same as D L with each random variable Y i j replaced by its forecast Y ^ i j . Here and throughout, a hatted symbol will denote an estimate or forecast of the unhatted quantity.
The upper and lower triangles have been defined in terms of incremental claim payments. It will sometimes be convenient to define them in terms of cumulative claim payments, e.g., D U = C i j : i = 1 , , I , j = 0 , , I i . As there is an equivalence relation between the two forms of triangle, the terms upper and lower triangle will be used, with a slight abuse of terminology, to refer to either form, provided that the context removes any ambiguity.
Let R i denote row i of the incremental d U , i.e., R i = y i j : j = 1 , , I i , and let r i denote the sum of this row:
r i = R i y i j = c i , I i   .
Further, let C j denote column j of the cumulative d U , i.e., C j = c i j : i = 1 , , I j , and let s j denote the sum of this column:
s j = C j c i j = i = 1 I j c i j   .
As noted above, C i , I 1 is the ultimate claim cost of accident period i , and so C ^ i , I 1 is an estimate of this quantity. The amount of outstanding losses (the loss reserve) for this accident period is equal to
L i = C i , I 1 C i , I i ,
and is estimated by
L ^ i = C ^ i , I 1 c i , I i .
The total reserve for all accident years of interest is
L = i = 2 I L i ,
which is estimated by
L ^ = i = 2 I L ^ i .
The following vector and matrix quantities are also introduced. Let u k denote the k -dimensional column vector with all components equal to unity, and let I k denote the k × k identity matrix.

2.2. Mesh Size

2.2.1. Preservation of Calendar Periods

Section 2.1 is phrased in terms of accident, development and calendar periods, without any specification of the meaning of “period”. Often, in the literature, this unit is a year. But it need not be; it might be a quarter, month, week, or any other convenient length of time. This will be referred to as the mesh size of the data triangle.
Consider how the mesh size might be changed. Suppose that, in D U , I = N q for some strictly positive integers N , q . The mesh size can be changed from one unit of time to q units. There will then be N accident and development periods, instead of the original I .
An example would be the case in which the units of time in D U are quarters and I = 40 , N = 10 , q = 4 . Here, the mesh size is changed from quarters to years, and D U contains 10 accident and development years.
In the case of general N and q , the change of mesh will induce a new upper triangle D U * in which row i * will be obtained from the merger of rows q i * 1 + 1 , q i * 1 + 2 , , q i * from D U .
Now suppose, in addition, that the change of mesh is required to preserve calendar periods. By this is meant that calendar period t * in D U * will comprise calendar periods q t * 1 + 1 , q t * 1 + 2 , , q t * from D U .
For given i * and t * , development period j * = t * i * . Then the i * , j * cell of D U * will consist of all pairs of q i * 1 + r , q t * 1 + s q i * 1 + r , r , s = 1 , , q from D U such that the second member of the pair is non-negative. Equivalently, it consists of all pairs q i * 1 + r , m a x 0 , q j * + s r , r , s = 1 , , q .
This is the form of aggregation from unit to q -unit periods commonly used in commercial practice. It is illustrated in Figure 1, in which a 40 × 40 quarterly triangle is collapsed to a 10 × 10 yearly triangle. The upper triangle is shaded yellow, the lower green. Calendar years are delineated by the red diagonals. Development years 5 and 6 are shaded blue and purple, respectively. Development year 0 is also shown orange to illustrate its exceptional nature.

2.2.2. Preservation of Development Periods

It is seen in Figure 1 that the quarterly development periods that contribute to a specific yearly development period differ by accident quarter. Although this is standard commercial practice, one might regard this as undesirable. Indeed, Section 5.1 gives some justification for this view.
An alternative is as follows. Define merged accident periods exactly as in Section 2.2.1, but define merged development periods in such a way that development period j * always consists of the same development periods from D U .
In this case, there is no requirement that accident and development periods be subject to the mesh size. There is not even a requirement that all development periods be of the same duration. So partition the ordered set 0 , , I by the insertion of cut-points c k , k = 0 , , K with c 0 = 0 , c K = I so that 0 , , I = k = 1 K S k , where S k = c k 1 , , c k 1 . Then S k , k = 1 , , K become the merged development periods, where S k spans c k c k 1 units of time.
The situation is illustrated in Figure 2, where development quarters 16 to 19 have been merged into a single development year, shaded blue; and similarly development quarters 20 to 23, shaded purple. The other development quarters have been left intact. Note that, for each of these development years, there is a triangle of data relating to the latest accident periods. These have not been shaded as they are incomplete and not comparable with the merged development years for earlier accident periods.

2.3. Sufficient Statistics

The present paper makes use of sufficient statistics in its argument. These and their properties are therefore briefly reviewed here. The review is general rather than specific to the context of the present paper.
Suppose that one makes observations y k , k = 1 , , K on some random variable Y that depends on some generic parameter θ , and let T y 1 , , y K be an estimator of θ . The general idea of a sufficient statistic for a parameter θ is that it contains all information relevant to the estimation of θ . Thus, for example, if T is sufficient for θ , and is a dimension reducing mapping, i.e., dim T < K , then there is no gain in information to be made by augmenting T with any further detail from y 1 , , y K .
The following few sub-sections elaborate on the technical aspects of sufficient statistics, and how they relate to Generalized Linear Models (“GLMs”).

2.3.1. Sufficiency

Continue with the notation immediately above. Further, let the vector y denote the sample of observations y 1 , , y K , and suppose that this vector is subject to the pdf p y ; θ .
Definition 1.
Let  T : S R M , M K  for some  S R K . Then  T y  is a sufficient statistic (Cox and Hinkley 1974) for  θ  if the conditional density  P r o b y | T y  is independent of  θ .
Note that  T  reduces the data set of dimension  K to an M -dimensional random vector with M K . In the case M < K , the sufficient statistic T will be said to be dimension-reducing.
A useful identification of sufficient statistics is given by the following result (Cox and Hinkley 1974).
Proposition 1.
(Fisher-Neyman factorization theorem). Let  T : S R M , M K  for some  S R K . Then  T y  is a sufficient statistic for  θ  if the density of  y  can be decomposed in the following manner:
p y ; θ = f y g T y , θ
 for some non-negative functions  f .
Some important estimation properties follow from sufficiency. The derivation of the first of these is elementary. Simply take logs throughout (8) and differentiate with respect to θ , to obtain
l n   p y ; θ θ = l n   g T y , θ θ   ,
from which the following proposition is immediate.
Proposition 2.
Let  T y  be a sufficient statistic for  θ  under the conditions of Definition 1. Assume further that  p y ; θ  is differentiable with respect to  θ . Then the maximum likelihood estimator (“MLE”) of  θ  is a function of only  T y .
Proof. 
The maximum likelihood (“ML”) equation sets (9) equal to zero. Since the right side of (9) involves only  T y , θ , the MLE of  θ  must be a function of just T y . □
Sufficient statistics also lead to estimators of maximal efficiency (Lehmann and Scheffé 1950, 1955). This result requires the concept of completeness of a family of distributions.
Definition 2.
Consider the family of densities  p y ; θ : θ Θ  for some parameter space  Θ , and let  T y  be a sufficient statistic for  θ . T y  is said to be complete if (Cox and Hinkley 1974),  E q T | θ = 0  for every  q : R M R  only if  T y = 0  a.s.
Proposition 3.
(Lehmann-Scheffé). Let  p y ; θ : θ Θ  be a family of densities, and let  T y  be a sufficient statistic for  θ  under the conditions of Definition 2.1. Let  T y  be complete within the family, and suppose that  ψ  is a mapping of  T y  to  Θ , with the unbiased property  E ψ T = θ , where the expectation is taken over  Y 1 , , Y K . Then  ψ T  is the unique minimum-variance unbiased estimator (“MVUE”) of  θ .

2.3.2. Generalized Linear Models

The standard GLM (Nelder and Wedderburn 1972) takes the form
Y i ~ E D F μ i , φ , i = 1 , , n
with
μ = E Y = h 1 X β
where
  • the Y i are stochastically independent real-valued random variables, each with range D ;
  • Y is the column n -vector with components Y i ;
  • E D F μ i , φ denotes a distribution from the exponential dispersion family (“EDF”) (Nelder and Wedderburn 1972) with mean μ i and dispersion parameter φ ;
  • μ is the column n -vector with components μ i ;
  • β is a column p -parameter;
  • X is an n × p design matrix;
  • h : D R is an invertible function, called the link function, with R a subset of the real line, and h operating component-wise in (11).
The notations here are generic to GLMs and will take on different forms in the specific GLM of Section 3, Section 4 and Section 5. Observations Y i j in that specific model are doubly suffixed (instead of just Y i as in the generic description) and so the vector Y there becomes the vector of all Y i j . Similarly for other quantities such as μ .
A distribution is a member of the EDF if its density takes the form (McCullagh and Nelder 1989)
p y ; θ , φ = e x p y θ b θ a φ + c y , φ
for location and scale parameters θ , φ and some functions a . , b . , c . .
A member of the EDF is defined by its cumulant function  b , a real-valued function of the real canonical parameter  θ i which is related to the mean μ i by the relation μ i = b θ i . The link function is said to be canonical if θ = X β , where θ is the vector of values θ i .
McCullagh and Nelder (1989, p. 32) point out an interesting and useful sufficient statistic of a GLM.
Proposition 4.
In the case of the GLM defined in (10) and (11), and provided that the link is canonical,  X T Y  is a sufficient statistic for  β .
Note that X T Y is a p -vector, so of the same dimension as β .
A special case of the above GLM is obtained by restricting the EDF to its Tweedie sub-family (Tweedie 1984), for which the cumulant function is
b θ = 1 2 p 1 p θ 2 p 1 p
for some parameter p   ϵ   , 0 [ 1 , ) .
A further special case is that in which p = 1 , and b θ = e x p   θ , a φ = φ , obtained by allowing p 1 in (12). This is the case of an over-dispersed Poisson (“ODP”) distribution. Note that, in this case,
e x p   X β = b θ = e x p θ
where all functions operate on vectors component-wise. By definition, then, the log link function is canonical in this case. Specialization even further, with φ = 1 , gives the Poisson case.

2.4. Notational Summary

For ease of reading the remainder of this paper, Table 1 summarizes both the notation introduced above and the main items of notation introduced in later sections. This necessarily involves forward references and these are noted in the table.
In addition, there are other notations (e.g., y , μ , θ ), occurring in Section 2.3.1 and Section 2.3.2, that relate to generic background results, and are not specific to the models used in Section 3 and subsequently. These have not been included in Table 1.

3. EDF Chain Ladder

3.1. Model Assumptions

The EDF chain ladder was discussed by Wüthrich and Merz (2008, Chapters 5 and 6). Taylor (2011) discussed the special case of the Tweedie chain ladder. It is the GLM defined by (10) to (12) and with the specific link function, observation vector, parameter vector and design matrix as follows:
h = l n
Y = Y 10 , , Y 1 , I 1 , Y 20 , , Y 2 , I 2 , , Y I 1 , 0 , Y I 1 , 1 , Y I 0 T
β = α 1 , , α I , β 0 , , β I 1 T
X = u I       I I   u I 1     I I 1 *             u 1 I 1 *
Here the Y i j in (14) are the observations defined in Section 2.1, the α i are the chain ladder row effects, the β j the column effects, and I i * denotes the identity matrix I I with all but its first i diagonal entries set to zero. As is well known, this model contains one degree of parameter redundancy, and so one linear constraint must be imposed on the 2 I parameters if parameter estimation is to be carried out.
The vector μ of cell means μ i j is given by μ = e x p   X β and, by substitution of (16) and (17),
μ i j = e x p α i   e x p β j
For p = 1 , the EDF chain ladder becomes the Poisson chain ladder.

3.2. Parameter Estimation

A GLM is typically calibrated by ML estimation, using GLM-specific software. However, Taylor (2011) shows that the ML equations for the Tweedie chain ladder take the simple form
j     R i μ i j 1 p y i j μ i j , i = 1 , , I ,
i     C j μ i j 1 p y i j μ i j , j = 0 , , I 1 .
In the Poisson chain ladder case, these ML equations reduce, by virtue of (18), to
e x p α i = r i / j     R i e x p β j , i = 1 , , I ,
e x p β j = s j /   i     C j e x p α i , j = 0 , , I 1 .

3.3. Sufficient Statistics

Taylor (2011, Theorem 5.2) derives the following result.
Proposition 5.
For the Tweedie chain ladder defined in Section 3.1,
(a)
For  p 1 ,  the only sufficient statistic for any  α i  or  β j  is the full data set  D U ; there is no sufficient statistic that is a proper subset of  D U .
(b)
For  p = 1  (Poisson chain ladder), the set  r 1 , , r I , s 0 , , s I 1  is a sufficient statistic for  α 1 , , α I , β 0 , , β I 1 .
Although Taylor proves this result by appeal to the Fisher-Neyman factorization theorem, it is interesting to note in passing that Proposition 5(b) can be obtained by substitution of (15) and (17) into Proposition 4 after taking note of the fact (Section 2.3.2) that the log link is canonical for the Poisson distribution. Thus,
X T Y = u I T         u I 1 T                 u 1 T I I I I 1 * I 1 * Y 1 Y 2 Y I ,
where Y i is the sub-vector of Y that relates to row i of the data set. Simple evaluation of the sufficient statistic X T Y then reveals it to be as claimed by Proposition 5(b).
A further result that will be found useful is the following, from Cox and Hinkley (1974).
Proposition 6.
Let  p y ; θ : θ Θ  be a family of EDF densities for specific cumulant function, and let  T y  be a sufficient statistic for  θ  under the conditions of Definition 1. A necessary and sufficient condition that  T y  be complete is that  d i m   T y = d i m   θ .

3.4. Forecast Bias

3.4.1. Poisson Case

As noted in Section 3.1, the parameters in Section 3.2 must be subject to a linear constraint. There are many possibilities but, for convenience, set
j = 0 I 1 e x p β j = 1 ,
in which case (21) may be re-expressed as
e x p α i = r i / 1 j = I i + 1 I 1 e x p β j , i = 1 , , I .
In the case of the Poisson chain ladder, (22) and (25) provide an estimation algorithm if taken in the order i = 1 , j = I 1 , i = 2 , j = I 2 , , j = 0 .
For clarity of subsequent reasoning, it is desirable to re-state (22) and (25) in their random forms, as follows:
e x p α ^ i = R i / 1 j = I i + 1 I 1 e x p β ^ j , i = 1 , , I .
e x p β ^ j = S j /   i     C j e x p α ^ i , j = 0 , , I 1 .
where R i , S j are the random variables of which r i , s j are realizations, and α ^ i , β ^ j are the estimates of α i , β j .
The above recursive computation of the estimates indicates that they are explicit functions (albeit of a complicated nature) of the random variables R i , S j , which are Poisson variates with means that depend explicitly on the set of parameters α i , β j . By (18), future values of C i j are forecast by C ^ i j = e x p α ^ i   e x p β ^ j , and so, by (5) and (7), L ^ i and L ^ are also explicit functions of these Poisson variates. Write L ^ i = λ i ( α ^ , β ^ | α , β ) , L ^ = λ ( α ^ , β ^ | α , β ) , where the true parameters α , β are unknown.
For given parameters α , β , the expectation of the estimated loss reserves is
E L ^ i = E D U λ i α ^ , β ^ | α , β .
The chain ladder forecast is conventionally taken as just L ^ i = λ i α ^ , β ^ | α , β , and so the bias in this estimator might be taken as the ratio of E L ^ i to E L i . Unfortunately, the latter is not calculable because α , β in (28) are unknown. An alternative estimate of bias can be obtained by comparison between L ^ i and E L ^ ^ i = E D U λ i α ^ ^ , β ^ ^ | α ^ , β ^ , where α ^ ^ , β ^ ^ are the estimates of the chain ladder parameters from a data triangle D U randomly sampled from parameters α ^ , β ^ .
The (estimated) chain ladder bias factor is then B i = L ^ i / E L ^ ^ i . A similar factor B can be calculated as the bias associated with L ^ .

3.4.2. General EDF Case

Bias in the general EDF case can be estimated in a similar way, but with one variation. The need for this arises from the fact that the ML Equations (19) and (20) do not yield explicit solutions for α , β .
An iterative solution is suggested by Taylor (2009), in which (19) and (20) are replaced by
j     R i μ i j m 1 p y i j μ i j m + 1 , i = 1 , , I ,
i     C j μ i j m 1 p y i j μ i j m + 1 , j = 0 , , I 1 ,
for m = 0 , 1 , , where μ i j m = α i m β j m is the m -th iterate on the estimate of μ i j .
In this schema, at each m , (29) and (30) are solved for α i m + 1 , β j m + 1 with values of μ i j m fixed by the preceding iteration. The procedure is initiated by some convenient selection of α i 0 , β j 0 , possibly the Poisson chain ladder solution (21) and (22), and continued until successive iterates of the α i m , β j m are sufficiently close that convergence may be assumed.
Convergence may or may not occur. It might occur, but there might be multiple solutions. Taylor (2017) gives an example of a triple solution for an inverse Gaussian distribution (Tweedie, p = 3 ). In such cases, bias correction cannot be obtained by these means.
Bias factors may be derived just as in Section 3.4.1.

3.5. Forecast Variance

Consider the estimator L i = L ^ i / B i , where B i is the bias factor derived in Section 3.4.1 (and which is similarly applicable to Section 3.4.2). It is evident that L i is an unbiased estimator of L i .

3.5.1. Poisson Case

Proposition 7.
Consider the Poisson chain ladder model of Section 3.1. The bias-corrected reserve estimators  L , L i  defined in Section 3.4.1 are unique MVUEs.
Proof. 
It is evident from (21) and (22) that the estimates of the parameters  α , β  are functions of the summary statistics  r 1 , , r I , s 0 , , s I 1  which, by Proposition 5, form a sufficient statistic  T D U . Hence  L ^ , L ^ i  are functions of this sufficient statistic. □
From Section 3.4.1, the bias factors B , B i are functions of the parameter estimates α ^ , β ^ , and therefore functions of the sufficient statistic T D U . It then follows that the bias-corrected loss reserve estimates L i , L = i L i are functions of T D U .
It remains to check that T D U is complete. This can be demonstrated with the help of Proposition 6. Note that, as written, T D U has dimension 2 I , and this is a sufficient statistic for α T , β T T , which also has dimension 2 I .
However, it has been noted in Section 3.1 that this parameter vector has in fact only 2 I 1 independent components. Moreover, it may also be noted that the components of T D U are also not independent. This can be seen from the fact that
r 1 + + r I = s 0 + + s I 1 = D U Y i j
the sum of all observations.
Thus, it is more useful to regard r 1 , , r I 1 , s 0 , , s I 1 T , of dimension 2 I 1 , as a sufficient statistic for α 1 , , α I 1 , β 0 , , β I 1 T , also of dimension 2 I 1 . The equality of these two dimensions, together with Proposition 6, then guarantees that T D U , or its abbreviated form, is a complete sufficient statistic.
All conditions of Proposition 3 are now satisfied, justifying the stated result.
The discussion hitherto has been confined to the Poisson case for the sake of simplicity. However, Taylor’s result (Taylor 2011, Theorem 5.2), cited earlier as relating to the Poisson case, actually applies to the more general ODP case. Here, for known scale parameter φ , the above sufficient statistic for the Poisson case remains sufficient for the ODP case. Hence the following remark.
Remark 1.
Proposition 7 applies equally to the ODP chain ladder with known scale parameter  φ .
Remark 2.
Further, it is well known that the magnitude of the scale parameter has no effect on the GLM’s parameter estimates, but all variance estimates are proportional to it.

3.5.2. General EDF Case

Now consider the general EDF chain ladder model of Section 3.1. In this case, as noted by Proposition 5(a), for p 1 , there is no sufficient statistic for the parameter set that is a proper subset of D U . One can, of course, choose T D U = D U as the sufficient statistic.
The bias-corrected reserve estimates L , L i are evidently functions of this sufficient statistic. However, T D U is not complete, since d i m   T D U = ½ I I + 1 , which is not equal to the dimension of the parameter vector α 1 , , α I 1 , β 0 , , β I 1 T . The sufficient statistic therefore fails the completeness test, according to Proposition 6.
As a result, Proposition 3 cannot be invoked. This does not prove that the reserve estimates L , L i are not MVUE, but it does leave the question open.

4. Effects of Change of Mesh Size Under Preservation of Calendar Periods

4.1. Model Assumptions

4.1.1. EDF Chain Ladder

Consider a given claim data set D U and the imposition on it of a larger mesh size, as described in Section 2.2.1. It is commonly assumed that one can make such changes at will while retaining an EDF chain ladder model. This assumption, however, warrants consideration.
Cells under the enlarged mesh are obtained by simple addition of the cells of smaller mesh, but note that the EDF is not closed under addition of random variables except under restrictive conditions. These are given for the Tweedie sub-family by Taylor (2021, Theorem 2.5), which leads to the following result.
Proposition 8.
Consider a claim data set  D U  for which all observations are Tweedie distributed with Tweedie parameter  p . Then observations under increased mesh size remain Tweedie distributed if and only if one of the following two conditions holds:
(a)
All observations are normally distributed ( p = 0 ); or
(b)
All observations have common ratio of mean to variance.
If these conditions are not satisfied, one is not strictly entitled to apply the EDF (or, more particularly, Tweedie) chain ladder to the data set with increased mesh.

4.1.2. Poisson Chain Ladder

Proposition 9.
(special case of Proposition 8). A data set of Poisson observations remains so under enlargement of mesh size.
Proof. 
The Poisson family of distributions satisfies condition (b) of Proposition 8. Specifically, the ratio of mean to variance is unity. The stated result is therefore a special case of Proposition 8. □

4.1.3. Structure of Cell Means

The Poisson chain ladder model of Section 3.1 requires not only Poisson distributions for all observations, but also that cell means assume the form (18). It is necessary to check whether this structure remains after mesh enlargement. The situation is, in fact, as set out in Proposition 10 below.
For the purpose of just this section, adopt the notation
a i = e x p   α i , b j = e x p   β j
so the required cell structure (18) can be expressed in the form
μ i j = a i b j
Further, for given integer q > 1 , define
ξ q j * k = l = 1 q b q j * + l k
and
ρ q j * k = ξ q j * + 1 k ξ q j * k   .
Then introduce the matrix and vector notation
w i * = a q i * 1 + 1 , , a q i * 1 + q T , v j * = ξ q j * 1 , , ξ q j * q T , D j * = d i a g ρ q j * 1 , , ρ q j * q , M j * = D j * v j * v j * T v j * v j * T D j *
Proposition 10.
Let  D U  be a data triangle satisfying the conditions for which a Poisson chain ladder model is valid, i.e.,  Y i j ~ P o i s s o n μ i j , with  μ i j  taking the form (18). Now increase the mesh size of the triangle by a factor  q , preserving calendar periods (as illustrated in  Section 2.2.1). A necessary and sufficient condition for the Poisson chain ladder to be a valid model for the triangle with enlarged mesh is that, for all choices of  i * , i * * ,
w i * T M j * w i * * = 0 , j * 1 ,
 in the above notation, with a similar but slightly different requirement for the case  j * = 0 .
Proof. 
A different characterization of this structure, which follows from and implies (18), is
μ i , j + 1 μ i j = b j + 1 b j   ,   i n d e p e n d e n t   o f   i .
Now consider this requirement under mesh enlargement by a factor of q as set out in Section 2.2.1. In the notation established there, the left side of (38) is replaced, in the case , j * 1 , by
μ i * , j * + 1 μ i * , j * = k = 1 q a q i * 1 + k l = 1 q b q j * + 1 + l k k = 1 q a q i * 1 + k l = 1 q b q j * + l k = k = 1 q a q i * 1 + k ρ q j * k ξ q j * k k = 1 q a q i * 1 + k ξ q j * k   .
This may be abbreviated by use of the matrix and vector notation above, thus
μ i * , j * + 1 μ i * , j * = w i * T D j * v j * w i * T v j *   .
In parallel with (38), it is required that, for each j * 1 , the ratio immediately above be independent of i * . Equivalently, for i * , i * * = 1 , , I ,
w i * T D j * v j * w i * T v j * = w i * * T D j * v j * w i * * T v j *   , j * 1 .
For conciseness, it will be convenient to drop the subscript j * , cross-multiply in the last equation, and rearrange slightly, to obtain
w i * T D v v T w i * * = w i * T v v T D w i * *
which, by (36), is the same as (37). □
Remark 3.
It is straightforward to check that  M j *  is a  q × q  matrix with  k , l  element  ρ q j * k ρ q j * l ξ q j * k ξ q j * l , k , l = 1 , , q . It is thus anti-symmetric. It immediately follows that  w i * T M j * w i * = 0  in (37), as it must.
The condition (37) is very restrictive, but there are some simple conditions under which it is satisfied.
Corollary 1.
Each of the following conditions is sufficient for the satisfaction of (37):
(a)
D j *  is a scalar matrix, i.e.,  ρ q j * 1 = ρ q j * 2 = ρ q j * q ;
(b)
w i * * = C w i *  for some constant  C > 0 .
Proof of Corollary 1.
Combine condition (a) with Remark 3 to obtain  M j * = 0 , whence (37) is satisfied.
Under condition (b),  w i * T M j * w i * * T = C w i * T M j * w i * T = 0 , by Remark 3. □
Remark 4.
The meaning of condition (a) of the corollary is that, within each development period under the enlarged mesh, the ratio of expected claim payments in one of the component unenlarged development periods to the expected payments in the corresponding component of the previous enlarged development period is independent of the component chosen.
The meaning of condition (b) is that each accident period under the enlarged mesh is distributed in proportions over its component accident periods from the unenlarged mesh in the same way as every other accident period.

4.2. Forecast Variance

Consider a model with enlarged mesh that jumps all the hurdles of Section 4.1, i.e., it is a Poisson chain ladder model that satisfies (37) under an enlargement of mesh size. Then, by Proposition 10, it remains a Poisson chain ladder model under the enlarged mesh. A question arises as to which mesh size produces the most efficient forecast of loss reserve.
By Proposition 7, application of the chain ladder algorithm to D U (unenlarged mesh) produces a unique MVUE estimate of loss reserve for that data set. As seen in Section 2.2.1, each cell of the upper triangle D U * induced by the change of mesh size is obtained by summation of cells of D U .
The row and column sums of D U * are a sufficient statistic for D U * when the available data comprise the latter. However, they are not a sufficient statistic for D U when those more granular data are available. This result is summarized as follows.
Proposition 11.
Consider a Poisson chain ladder model on data triangle D U , and subject this triangle to the increase in mesh size described in Section 2.2.1. Suppose that condition (37) in Proposition 10 is satisfied, in which case the induced data triangle D U * is also described by a Poisson chain ladder model.
Then
(a)
V a r L ^ | D U is the unique MVUE of all unbiased estimators of L , so conditioned;
(b)
V a r L ^ * | D U * is the unique MVUE of all unbiased estimators of L , so conditioned;
(c)
V a r L ^ | D U < V a r L ^ * | D U * .
Proof. 
Proposition 5(b) defines the (different) sufficient statistics for the respective model parameters in the two cases of data triangle. The sufficient statistic defines an MVUE forecast of loss reserve in each of the cases of D U and D U * . Denote these forecasts by L ^ and L ^ * , respectively. Then the proposition follows. □
In other words, subject to the sustained validity of a Poisson chain ladder model, the bias-corrected chain ladder forecast of a loss reserve is always improved by increased granularity, subject to preservation of diagonals.

5. Effects of Change of Mesh Size Under Preservation of Development Periods

5.1. Model Assumptions

5.1.1. EDF and Poisson Chain Ladder

Consider a given claim data set D U and the imposition on it of a larger mesh size, as described in Section 2.2.2. Propositions 8 and 9 continue to hold. The supporting arguments are as in Section 4.1.1 and Section 4.1.2.

5.1.2. Structure of Cell Means

It was seen in Section 4.1.3 that the validity of a Poisson chain ladder model continued to hold under mesh enlargement with preservation of calendar periods only under restrictive conditions on the chain ladder parameters of the smaller-mesh model. It is of interest to consider the conditions under which a Poisson chain ladder model remains valid when the mesh is enlarged with preservation of development periods, as in Section 2.2.2.
Consider the Poisson chain ladder model of Section 3.1 and, again, temporarily adopt the notation of (32) and (33). Now, in the notation of Section 2.2.2, define development period j * under a mesh enlargement to be the merged period S j * .
Accident periods 1 , , I can be partitioned in any way desired, so let a new accident period i * under mesh enlargement comprise accident periods h i * , h i * + 1 , , h i * + d i * 1 under the unenlarged mesh. Here accident period i * spans d i * of the original time units. The question arises as to the extent to which cells remain Poisson distributed under this re-definition of development periods. The next result answers this question.
Proposition 12.
Let  D U  be a data triangle satisfying the conditions for which a Poisson chain ladder model is valid, i.e.,  Y i j ~ P o i s s o n μ i j  , with  μ i j  taking the form (18). Now increase the mesh size of the triangle, preserving development periods (as illustrated in Section 2.2.2). The Poisson chain ladder continues to be a valid model for all cells of the triangle with enlarged mesh except the edge cells defined in the proof below.
Proof. 
The expected value of payments in cell i * , j * under the new mesh is
μ i * , j * = k = 0 d i * 1 l = c j * 1 c j * 1 a h i * + k b l = k = 0 d i * 1 a h i * + k l = c j * 1 c j * 1 b l ,
provided that none of the cells h i * + k , c j * 1 lies outside the triangle D U , i.e., provided that h i * + d i * 1 + c j * 1 I .
Relation (43) is of the form (18). Just as in Proposition 9, the amount of claim payments in cell i * , j * , as the sum of Poisson variates, is Poisson.
If the constituent cells do not all lie within D U , then the factorization in (43) occurs only if d i * = 1 (no aggregation of accident periods) or the trivial case c j * = c j * 1 (no aggregation of development periods). Call these cells edge cells. They are exemplified in Figure 2, where they are bordered by the I -th diagonal in D U , the most recent past diagonal. □
Note that the edge cells are incomplete in the sense that they include fewer of the smaller-mesh cells than is the case for earlier accident periods, and so are not comparable with any larger-mesh development period from earlier accident periods.
It is supposed here that one makes a decision as to the granularity of collected data, after which data of greater granularity are unavailable. There are two alternative ways forward. Either:
  • formulate a model of the incomplete larger-mesh development periods (which will not be informed by more granular data); or
  • exclude the edge cells from the data set as uninformative.
The first alternative is viable. For example, one might assume that claim payments occur uniformly over the larger-mesh development periods, or according to some other convenient distribution. However, the resulting model, while perfectly respectable, would no longer be chain ladder, and so is not pursued here.
In the second case, the edge cells are deleted from the upper triangle, and regarded as belonging to the future, therefore requiring forecast. The situation is illustrated by Figure 3, which is a replica of Figure 2 except that edge cells have been marked in green as belonging to the future.
In this figure, development quarters 16 to 19 have been merged into a single development year, as in Figure 2, and similarly for development quarters 20 to 23. The figure also indicates that accident quarters 17 to 24 are to be merged into four accident half-years.
The general situation is summarized by the following statement.
Proposition 13.
Let  D U  be a data triangle satisfying the conditions for which a Poisson chain ladder model is valid, i.e.,  Y i j ~ P o i s s o n μ i j , with  μ i j  taking the form (18). Now increase the mesh size of the triangle for certain accident and development periods, preserving development periods (as illustrated in Section 2.2.2). The mesh-enlarged development periods need not be of equal lengths, nor need the mesh-enlarged accident periods. There need be no relation between the lengths of mesh-enlarged accident and development periods. Define  D U *  to be the data set induced on  D U  by the mesh enlargement, and let  D U * ( )  be the data set obtained by the deletion of edge cells from  D U * . Then a Poisson chain ladder model remains valid on  D U * ( ) .
Proof. 
A Poisson chain ladder model can be applied to the data set D U * ( ) (in the above example, this consists of the part of Figure 3 not shaded green and not forming the incomplete accident half-years comprising accident quarters 17 and 21). The model is simply
Y i * j *   ~ P o i s s o n μ i * j *
μ i * j * = e x p α i *   e x p β j *
(compare (18)) where the cells of D U * ( ) are labelled i * , j * with i * , j * labelling the accident and development periods under the enlarged mesh.
This chain ladder model will forecast the ultimate cost of each mesh-enlarged accident period. Total payments up to the valuation date are known for each of these accident periods, and so the outstanding liability for each accident period can be inferred. □

5.2. Forecast Variance

Consider the model and mesh enlargement described in Proposition 12. As in Section 4.2, a question arises as to which mesh size produces the most efficient forecast of loss reserve. This question is answered by logic parallel to that applied in Section 4.2, leading to the following result.
Proposition 14.
Consider a Poisson chain ladder model on data triangle  D U , and subject this triangle to an increase in mesh size described in Section 2.2.2. Then the induced data triangle  D U * ( )  is also described by a Poisson chain ladder model.
Proposition 5(b) defines the (different) sufficient statistics for the respective model parameters in the two cases of data triangle. The sufficient statistic defines a unique MVUE forecast of loss reserve in each of the cases of D U and D U * ( ) . Denote these forecasts by L ^ and L ^ * ( ) , respectively.
Then
(a)
V a r L ^ | D U is the unique minimum variance of all unbiased estimators of L , so conditioned;
(b)
V a r L ^ * ( ) | D U * is the unique minimum variance of all unbiased estimators of L , so conditioned;
(c)
V a r L ^ | D U < V a r L ^ * ( ) | D U * ( ) .

6. Numerical Example

A simple numerical example is given to illustrate the result in Proposition 14. It is based on the synthetic data triangle set out in Table 2, which is deliberately constructed as well behaved and consistent with the chain ladder structure set out in (18).
Initially in the example, development years 3 and 4 are merged, just as two consecutive development periods are merged in the theoretical workings of Section 5. This is reflected in the blue shading in the table. In addition, cell (7,3) is shaded brown to indicate that there is no corresponding value in cell (7,4) with which it can be merged, whereas a combined “development years 3 and 4” cell can be formed for each of accident years 1 to 6, and used in chain ladder modelling of this combination, cell (7,3) is unusable in the modelling.
Call the triangle in the Table 2 the unmerged triangle, and apply the name “merged triangle” to the new version in which development years 3 and 4 are merged into a single “3 and 4” development period. The Poisson chain ladder is applied to each, on the basis of Section 3 in the unmerged case and Section 5 in the merged case. The latter of these is achieved by eliminating development year 3 and cell (7,3) from the design matrix X in (17).
In each case, the outstanding losses are forecast and the variance of this forecast calculated from the relevant GLM. The calculation follows (Wüthrich and Merz 2008, Section 6.4, especially (6.47) and Lemma 6.13). The results are as set out in Table 3.
The standard deviations here include only those components of forecast error referred to in the actuarial literature as parameter error, i.e., process error and model error are specifically excluded (Taylor and McGuire 2023). Nevertheless, the standard deviations are quite low by comparison with empirical works on this sort of error. In practice, one would usually apply an ODP rather than simple Poisson chain ladder, using a guessed scale parameter φ > 1 , which would produce larger standard deviations in both merged and unmerged cases. According to Remark 2, all of these standard deviations would simply increase by a factor equal to φ .
The following observations can be made on Table 3:
  • As expected, the results for accident years 2 to 6 are not affected at all by the merger of development years 3 and 4, because their forecasts do not depend in any way on these development years;
  • The results for accident years 8 to 10 are slightly affected by the omission of cell (7,3) from the modelling;
  • The result for accident year 7 is substantially affected, with a noticeable change in forecast and a 44% increase in the associated standard deviation. This is the result of the loss of information of cell (7,3).
The last result here is as predicted by Proposition 14. It shows that a single merger of a pair of development periods can produce a substantial increase in forecast variance for a single accident period. If multiple mergers are carried out over a number of development periods, e.g., changed in mesh from quarterly to yearly, all accident periods may be affected, the total effect may be very substantial indeed.
This is further illustrated by more extensive merger of development periods. In the next case, pairs of development periods, 2&3, 4&5, 6&7 and 8&9, respectively, are merged. In this case, cells (2,8), (4,6), (6,4), and (8,2) become edge cells (as defined in Proposition 12) in the same way as did (7,4) in the above example.
Development periods 1 and 3 have not been merged. If they had been, no complete cell would have been available for accident year 10, and the chain ladder would be unable to generate a forecast.
The results of application of the chain ladder to this further merged triangle are set out in Table 4, where comparison with the unmerged triangle is again made.
Once again, the restructure of the data set makes little difference to the forecast. However, standard deviations of the forecasts are substantially affected. This is less so for accident years 9 and 10, which are not affected by edge cells.
However, accident years 2, 4, 6 and 8 are noted above to be subject to incomplete cells, and it is seen in the table that their standard deviations of forecasts increase sharply. Accident years 3, 5 and 7 do not contain any incomplete cells, but their standard deviations of forecasts still increase as a result of the general loss of information induced by the merger of development years. Indeed, the increases of these standard deviations, taking account of dependencies between accident years, are as set out in Table 5.
These results are all consistent with Proposition 14.

7. Discussion and Conclusions

This paper is concerned with the identification of MVUEs of the EDF chain ladder estimated loss reserve. It is natural, therefore, to draw on sufficient statistics where possible (see Proposition 6).
However, it is known that the EDF chain ladder produces biased estimates. A bias correction factor is defined in Section 3.4. The bias-corrected chain ladder loss reserve is MVUE when the EDF is restricted to the Poisson family (Proposition 7). This result is obtained by reference to sufficient statistics.
In the case of a more general member of the EDF, the same argument cannot be used, because the sufficient statistic for the chain ladder parameters is not complete. This leaves open the question of whether the bias-corrected EDF chain ladder loss reserve is MVUE.
Two types of increase of the mesh size of the data triangle are considered:
  • preserving calendar periods; and
  • preserving development periods.
The first of these cases is considered in Section 4. This type of mesh variation is the one most commonly found in practice, where accident and development periods are of identical duration. Here, the initial question does not concern the variance of the EDF chain ladder loss reserve, but rather whether that chain ladder model remains valid under the change of mesh.
This involves two sub-questions:
  • Do the cells of the data triangle remain EDF under mesh enlargement?
  • Do the cell expectations retain, under mesh enlargement, the multiplicative parameter structure required by the chain ladder model?
These are questions which, though fundamental in nature, are rarely considered by practitioners. It is found that the preservation of the pre-conditions for the chain ladder is by no means a foregone conclusion in the event of a change in mesh size.
It is shown (Proposition 8) that the only practical case under which an EDF distribution is maintained is that in which all observations have a common ratio of mean to variance. In other cases, one should not be applying the EDF chain ladder under the increased mesh size. The required condition is very restrictive, but it is satisfied by the case in which each cell of the data set is Poisson distributed.
As to the second sub-question, it is found (Proposition 10) necessary for maintenance of the required multiplicative structure that an apparently obscure algebraic condition holds. However, there are a couple of special cases of this condition with clear physical meaning (Remark 4).
First, it is sufficient that, for each development period under the enlarged mesh, claim payments are uniformly distributed over the development periods under the unenlarged mesh. Second, it is sufficient that each accident period under the enlarged mesh is distributed in proportions over its component accident periods from the unenlarged mesh in the same way as every other accident period.
Under these conditions for a Poisson chain ladder, enlargement of mesh always increases the variance of the forecast loss reserve, and so a smaller mesh is always preferable to a larger one (Proposition 11). In other words, the strictly optimal mesh size is that which is small enough that each cell contains at most one transaction (a single payment in respect of a single claim). Of course, practicalities might militate against this choice, but this strict optimality provides a guide.
Section 5 considers the case of mesh enlargement under preservation of development periods. As in Section 4, the ratio of mean to variance over all cells of the data set is required to be constant if the EDF is to be retained under mesh enlargement. Once again, this places focus on the Poisson chain ladder model.
Beyond this requirement, the mesh enlargement under preservation of development periods is considerably less demanding, in terms of assumptions, than under preservation of calendar periods. For example, a Poisson chain ladder model under the original mesh remains so under the enlarged mesh without the imposition of further conditions (Proposition 12), except in relation to the so-called “edge cells”.
One may also select the durations of accident and development periods more or less arbitrarily without invalidating this result. Different accident periods need not have the same duration; nor development periods. There need be no relation between the durations of accident periods and those of development periods (Proposition 13). The chosen durations do, however, affect the definition of edge cells.
If one wishes to continue to work with a Poisson chain ladder model, one must exclude the edge cells from analysis. It is then found once again that enlargement of mesh always increases the variance of the forecast loss reserve, and so a smaller mesh is always preferable to a larger one (Proposition 14).
This paper began with a question as to whether there existed an optimal granularity for the EDF chain ladder with respect to variance of the loss reserve estimate. It has been seen that whether this question is even meaningful, in relation to variation of mesh size, depends on a number of details of the EDF model in use.
However, if the EDF chain ladder is the special case of a Poisson chain ladder, then, under certain realistic conditions, it is indeed meaningful to compare forecast variances for different mesh sizes. It is found that a smaller mesh is always to be preferred. On this basis, it is optimal to reduce the mesh size as far as possible, even using time units of days.
In short, practitioners should not be shy of reducing mesh size. The results here suggest that this action will be beneficial.
This is reminiscent of the Kaplan-Meier estimate, sometimes called the product limit estimate, of a mortality rate (Kaplan and Meier 1958). Those authors show that the maximum likelihood estimate of a mortality rate from exposure data and dates of all deaths in the data set consists of combining the estimates obtained by considering separately each period between successive deaths; in other words, a mortality rate estimate relating to each death.
It is often pointed out that the chain ladder model structure is rarely used in practice without some form of adjustment, destroying theoretical results such as derived here. Very often, the early diagonals of the data triangle will be down-weighted or even totally discarded. In other cases, an actuary may hand-pick parameter estimates with only partial regard to the formal estimates produced by the GLM.
It is difficult to deal with such objections in any organized manner. Naturally, no rigorous theoretical results are available in any case in which a formal definition of model is unavailable. Perhaps the most useful statement is that the formal results for the “pure” chain ladder are indicative of other closely related models.
They show that maximum granularity is beneficial to the pure chain ladder, and intuition then suggests that it would also be beneficial to many other “adjusted forms of chain ladder”. It is to be emphasized, nonetheless, that rigorous formulation of a model, followed by formal derivation of its statistical properties is preferable to intuitive extrapolation of properties from one model to another.
In similar vein, one might speculate on the optimal granularity of data provided as input to machine learning models, such as gradient boosting machines and neural networks. Here, there is no formal statistical model, and so no theoretical statistical results to be had. Again, however, the results obtained here may, in certain cases, be suggestive of maximum granularity.

Funding

This research received no external funding.

Data Availability Statement

The data is contained within the article.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Cox, David R., and David V. Hinkley. 1974. Theoretical Statistics. London: Chapman and Hall. [Google Scholar]
  2. Das, Rituparna, Asif Iqbal Middya, and Sarbani Roy. 2022. High granular and short term time series forecasting of PM2.5 air pollutant—A comparative review. Artificial Intelligence Review 55: 1253–87. [Google Scholar]
  3. Elton, Edwin J., Martin J. Gruber, Christopher R. Blake, Yoel Krasny, and Sadi O. Ozelge. 2010. The effect of holdings data frequency on conclusions about mutual fund behavior. Journal of Banking & Finance 34: 912–22. [Google Scholar]
  4. Grolinger, Katarina, Alexandra L’Heureux, Miriam A. M. Capretz, and Luke Seewald. 2016. Energy forecasting for event venues: Big data and prediction accuracy. Energy and Buildings 112: 222–33. [Google Scholar] [CrossRef]
  5. Hart, David G., Robert A. Buchanan, and Bruce A. Howe. 1996. The Actuarial Practice of General Insurance, 5th ed. Sydney: Institute of Actuaries of Australia. [Google Scholar]
  6. Huang, Zhichuan, and Ting Zhu. 2016. Leveraging multi-granularity energy data for accurate energy demand forecast in smart grids. Paper presented at the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, December 5–8; Piscatway: IEEE, pp. 1182–91. [Google Scholar]
  7. Jain, Rishi K., Kevin M. Smith, Patricia J. Culligan, and John E. Taylor. 2014. Forecasting energy consumption of multi-family residential buildings using support vector regression: Investigating the impact of temporal and spatial monitoring granularity on performance accuracy. Applied Energy 123: 168–78. [Google Scholar] [CrossRef]
  8. Kaplan, Edward L., and Paul Meier. 1958. Nonparametric estimation from incomplete observations. Journal of the American Statistical Association 53: 457–81. [Google Scholar] [CrossRef]
  9. Kim, Mingyung, Eric T. Bradlow, and Raghuram Iyengar. 2022. Selecting data granularity and model specification using the scaled power likelihood with multiple weights. Marketing Science 41: 848–66. [Google Scholar] [CrossRef]
  10. Lehmann, Erich L., and Henry Scheffé. 1950. Completeness, similar regions, and unbiased estimation-Part I. Sankhyā 10: 305–40. [Google Scholar]
  11. Lehmann, Erich L., and Henry Scheffé. 1955. Completeness, similar regions, and unbiased estimation. Part II. Sankhyā 15: 219–36. [Google Scholar]
  12. Li, Peikun, Chaoqun Ma, Jing Ning, Yun Wang, and Caihua Zhu. 2019. Analysis of prediction accuracy under the selection of optimum time granularity in different metro stations. Sustainability 11: 5281. [Google Scholar] [CrossRef]
  13. Lusis, Peter, Kaveh Rajab Khalilpour, Lachlan Andrew, and Ariel Liebman. 2017. Short-term residential load forecasting: Impact of calendar effects and forecast granularity. Applied Energy 205: 654–69. [Google Scholar]
  14. Mack, Thomas. 1993. Distribution-free calculation of the standard error of chain ladder reserve estimates. ASTIN Bulletin 23: 213–25. [Google Scholar]
  15. McCullagh, Peter, and John A. Nelder. 1989. Generalized Linear Models, 2nd ed. Boca Raton: Chapman & Hall. [Google Scholar]
  16. Nelder, John A., and Robert W. M. Wedderburn. 1972. Generalised linear models. Journal of the Royal Statistical Society. Series A 135: 370–84. [Google Scholar]
  17. Peterson, Timothy M. 1981. Loss Reserving: Property/Casualty Insurance. New York: Ernst & Whinney. [Google Scholar]
  18. Taylor, Greg C. 1985. Claim Reserving in Non-Life Insurance. Amsterdam: North-Holland. [Google Scholar]
  19. Taylor, Greg. 2000. Loss Reserving: An Actuarial Perspective. Boston: Kluwer Academic Publishers. [Google Scholar]
  20. Taylor, Greg. 2009. The chain ladder and Tweedie distributed claims data. Variance 3: 96–104. [Google Scholar]
  21. Taylor, Greg. 2011. Maximum likelihood and estimation efficiency of the chain ladder. ASTIN Bulletin 41: 131–55. [Google Scholar]
  22. Taylor, Greg. 2017. Existence and uniqueness of chain ladder solutions. ASTIN Bulletin 47: 1–41. [Google Scholar] [CrossRef]
  23. Taylor, Greg. 2021. A special Tweedie sub-family with application to loss reserving prediction error. Insurance: Mathematics and Economics 101B: 262–88. [Google Scholar] [CrossRef]
  24. Taylor, Greg, and Gráinne McGuire. 2023. Model error (or ambiguity) and its estimation, with particular application to loss reserving. Risks 11: 185. [Google Scholar] [CrossRef]
  25. Tweedie, Maurice C. K. 1984. An index which distinguishes between some important exponential families. In Statistics: Applications and New Directions, Proceedings of the Indian Statistical Golden Jubilee International Conference. Edited by Jayanta Kumar Ghosh and Jogabrata Roy. Kolkata: Indian Statistical Institute, pp. 579–604. [Google Scholar]
  26. Wang, Jue, Hao Zhou, Tao Hong, Xiang Li, and Shouyang Wang. 2020. A multi-granularity heterogeneous combination approach to crude oil price forecasting. Energy Economics 91: 104790. [Google Scholar] [CrossRef]
  27. Wang, Zongqiang, Yan Xian, Guoyin Wang, and Hong Yu. 2023. A Multi-granularity Network for Time Series Forecasting on Multivariate Time Series Data. Paper presented at the International Joint Conference on Rough Sets, Krakow, Poland, October 5–8; Cham: Springer Nature Switzerland, pp. 324–38. [Google Scholar]
  28. Wüthrich, Mario V., and Michael Merz. 2008. Stochastic Claims Reserving Methods in Insurance. Chichester: John Wiley & Sons Ltd. [Google Scholar]
Figure 1. Change of mesh size from quarterly to yearly.
Figure 1. Change of mesh size from quarterly to yearly.
Risks 13 00065 g001
Figure 2. Change of mesh size with preservation of development periods.
Figure 2. Change of mesh size with preservation of development periods.
Risks 13 00065 g002
Figure 3. Change of mesh size with preservation of development periods and edge cells deleted from data set.
Figure 3. Change of mesh size with preservation of development periods and edge cells deleted from data set.
Risks 13 00065 g003
Table 1. Summary of notation.
Table 1. Summary of notation.
SymbolIntroduced in
Section
Interpretation
i Section 2.1Accident period
j Section 2.1Development period
t Section 2.1Calendar period
Y i j Section 2.1Amount of claim payments (random variable) made during development period j of accident period i
y i j Section 2.1Realization of the random variable Y i j
μ i j Section 2.1Mean of Y i j
Y Section 3.1Vector of the Y i j
D U Section 2.1Upper triangle of observations (random variables) Y i j
D L Section 2.1Lower triangle of observations (random variables) Y i j
d U Section 2.1Realization of D U
C i j Section 2.1Cumulative claim payments (random variable) corresponding to (non-cumulative) Y i j
c i j Section 2.1Realization of C i j
R i Section 2.1Row i of upper triangle d U
C j Section 2.1Column j of upper triangle d U
r i Section 2.1Row sum of d U (over R i )
s j Section 2.1Column sum of d U (over C j )
L i Section 2.1Amount of outstanding losses (the loss reserve) for accident period i
L Section 2.1Amount of outstanding losses (the loss reserve) for all accident periods
i * , j * Section 2.2Accident and development periods under changed mesh size
D U * Section 2.2Upper triangle under changed mesh size
c 0 , , c K Section 2.2.2Partition of the interval 0 , I , defining development periods under changed mesh
α 1 , , α I , β 0 , , β I 1 Section 3.1EDF parameters for row and column effects in D U
B i Section 3.4.1Loss reserve bias factor
Table 2. Data Triangle For Numerical Example.
Table 2. Data Triangle For Numerical Example.
AccidentClaim Payments in Development Year
Year0123456789
1584015,16910,338751859544451296721351123593
2620515,31110,682755260804608299021881185
3625415,91511,10780356217470531302172
4634916,18011,2727999637549083193
5701117,55912,396896771085272
6675517,46412,18087877061
7727718,48112,5119212
8752518,71413,134
9750819,019
107834
Table 3. Forecasts And Standard Deviations Of Forecasts For Merged And Unmerged Data Triangles.
Table 3. Forecasts And Standard Deviations Of Forecasts For Merged And Unmerged Data Triangles.
AccidentForecast Outstanding LossesStandard Deviation of Forecast
Year
UnmergedMergedUnmergedMerged
26076072525
3183518353737
4413741375151
5803680367171
613,14513,1459494
721,06520,997133192
831,09231,081194195
944,58944,578311313
1066,36766,356807807
Total190,871190,77110621087
Table 4. Forecasts And Standard Deviations Of Forecasts For Merged And Unmerged Data Triangles.
Table 4. Forecasts And Standard Deviations Of Forecasts For Merged And Unmerged Data Triangles.
AccidentForecast Outstanding LossesStandard Deviation of Forecast
Year
UnmergedMergedUnmergedMerged
26075702544
3183518163745
4413740985173
5803680097179
613,14512,92494135
721,06521,000133140
831,09230,884194314
944,58944,502311316
1066,36766,277807808
Total190,871190,07910621178
Table 5. Forecasts And Standard Deviations Of Forecasts For Aggregated Accident Years In Merged And Unmerged Data Triangles.
Table 5. Forecasts And Standard Deviations Of Forecasts For Aggregated Accident Years In Merged And Unmerged Data Triangles.
AccidentStandard Deviation of ForecastChange
Due to
Years
UnmergedMergedMerger
%
2, 4, 6, 822235258
2 to 827139044
All1062117811
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Taylor, G. The Exponential Dispersion Family (EDF) Chain Ladder and Data Granularity. Risks 2025, 13, 65. https://doi.org/10.3390/risks13040065

AMA Style

Taylor G. The Exponential Dispersion Family (EDF) Chain Ladder and Data Granularity. Risks. 2025; 13(4):65. https://doi.org/10.3390/risks13040065

Chicago/Turabian Style

Taylor, Greg. 2025. "The Exponential Dispersion Family (EDF) Chain Ladder and Data Granularity" Risks 13, no. 4: 65. https://doi.org/10.3390/risks13040065

APA Style

Taylor, G. (2025). The Exponential Dispersion Family (EDF) Chain Ladder and Data Granularity. Risks, 13(4), 65. https://doi.org/10.3390/risks13040065

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop