1. The variability problem
1.1. The challenge of reserving
The property/casualty business model relies on the accurate measurement of risk. Of relevance to this paper is the measurement of reserving risk. Accurate actuarial loss reserving is one of the regulatory requirements in measuring solvency. Consequently, one of the most important tasks for an actuary is to estimate the proper amount of reserves to be set aside to meet future liabilities of current in-force business.
Because the stated reserve is an estimate and not the true number, there is error and it is important to measure this error. Quantifying the potential error allows for setting ranges around a best estimate, allows for measures of risk, and can assist in the setting of risk-based capital requirements. Recent papers such as Hayne (2004) and Shapland (2007) make the importance of this issue clear.
In this paper we propose a method for measuring the total risk involved in reserve estimates. It is simple to apply and uses data that is almost always available from the reserve-setting process. A key feature is that our measure of reserve variability does not depend on the method used to determine the reserves.
1.2. Literature review
There are many papers regarding different loss reserving techniques, some deterministic, some stochastic. For a comprehensive review of existing deterministic methods, the reader is referred to Wiser (2001) and Brown and Gottlieb (2001). For an excellent overview of a wide range of stochastic reserving methods in general insurance, the reader is referred to England and Verrall (2002). Other references of interest include Bornhuetter and Ferguson (1972), Finger (1976), and Taylor (2000).
While the traditional chain ladder technique provides only a point estimate of the total reserve, it has become evident recently that actuaries also need a measure of variability in loss reserving estimation. Sound methodologies that quantify risks related to the balance sheet are important for attracting and retaining capital in the firm. Standard and Poor’s ratings as well as investors are very interested in the value at risk (VaR) measurements. The NAIC and state regulators are interested in monitoring the reserve variability in the form of reserve ranges. Variability and sometimes the entire distribution of the loss reserve estimate are important for risk management purposes. Questions such as what is the 95th percentile of losses or the cost of loss portfolio transfer are important in managing and assessing risk. These questions can be properly addressed with a thorough analysis of the variability of the reserve estimate.
Over the last 30 years many researchers have made significant contributions to the study of the variability of reserving methods. The CAS working party paper (2005) presents a comprehensive review that brings all of the important historical research together. A representative, but not exhaustive, list follows. Taylor and Ashe (1983) introduced the second moment of estimates of outstanding claims. Hayne (1985) provided an estimate of statistical variation in development factor methods when a lognormal distribution is assumed for these factors. Verrall (1991) derived unbiased estimates of total outstanding claims as well as the standard errors of these estimates. Mack (1993) used a distribution-free formula to calculate the standard errors of chain ladder reserve estimates. England and Verrall (2002) presented analytic and bootstrap estimates of prediction errors in claims reserving. De Alba (2002) gave a Bayesian approach to obtain a predictive distribution of the total reserves. Taylor and McGuire (2004) applied general linear model techniques to obtain an alternative method in cases where the chain ladder method performs poorly. Verrall (2004) implemented Bayesian models within the framework of generalized linear models that led to posterior predictive distributions of quantities of interest.
The CAS working party (2005) ultimately concluded, “there is no clear preferred method within the actuarial community.” Actuaries need to select one or several methods that are considered appropriate for the specific situation. When it comes to the final decision, judgment still overrules.
1.3. The risk measurement problem
The CAS working party report (2005) notes that the sources of uncertainty in the reserve estimate come from three types of risk: process, parameter, and model. Model risk is the uncertainty in the choice of model. Parameter risk is the uncertainty in the estimates of the parameters. Process risk is the uncertainty in the observations given the model and its parameters.
Shapland (2007, 124) highlights the importance of all the risks:
Returning to the earlier definition of loss liabilities. . . all three types or risk. . . should be included as part of the calculated expected value. Alternatively, some or all of these types of risk could be included in a “risk margin” as defined under ASOP No. 36.
These three risks are intertwined and thus hard to separate. In particular, process and parameter risk are often calculated as if the model were correct.
In this paper, the three risks will not be separately measured nor directly treated.[1] Rather, we capture the total risk from all three sources underlying the reserve estimate. The reserving model (or ultimate loss selection if no specific method is used) will be considered fixed and any errors measured will be a consequence of that choice.
From a risk management perspective, this is appropriate. Now that the reserves have been established, what is the potential error that may result, given the current reserve review?
1.4. A summary of our approach
Most approaches to risk measurement rely on the statistical properties of the data as reflected by the model selected for calculating the reserve. Those approaches attempt to capture the underlying distribution of losses. In this setting, parameter estimation error can be estimated using statistical measures such as Fisher information. We choose to look at the reserves (as reflected in the estimated ultimate losses) themselves as they evolve over time. This provides a way to reflect all the sources of error. Each reserve set in the past is an estimate of its distribution and thus its errors can be estimated from the historical errors made in the estimations. Because the ultimates will converge to the true value, the errors made along the way reflect all sources of error.
Our methodology will be introduced within a stable context. In particular, initially we assume that the distribution underlying the loss development process is constant over time and that we are working with indicated reserves developed under a consistent methodology. However, it is important to note that, under these assumptions, the ranges produced do not include the possibility of a change in the distribution, such as inflation moving to a new level not in the available history or sharp increases in the deductibles or retentions written over or ceded. After working through this situation we will discuss modifications that can allow our model to be applied in more general settings.
Note that we model indicated reserves instead of held reserves. This is because indicated reserves do not have management or other ad hoc adjustments. Therefore, indicated reserves are more stable and thus easier to model. What is interesting about held reserves is where they fall within the probability distribution of the unknown true reserve. This can provide an indication of the degree to which held reserves are conservative or aggressive.
Because the method presented here is free of the choice of the reserving method used, it is not necessary to even have a specific method. The only requirement is a history of ultimate loss selections. Thus we rely on the actual error history of the reserving department.
We make a theoretical and pragmatic case for the lognormal distribution for the errors in aggregate reserves, line or total. The focus on the aggregate distribution also removes the need to choose individual size of loss distributions.
Each of the following sections will take one step through the development of the risk measure. An example will be followed throughout to illustrate the formulas.
2. Reserving process and data
The reserve review process generates reserves based on raw data analysis. The indicated ultimate losses are selected by line using perhaps several methods as well as judgment. These line ultimates are then added to yield the total indicated reserve.
Management adjustments called margins may be applied to the total reserve. These margins may then be allocated by line and by accident year to the indicated ultimate losses and reported in Schedule P Part 2. As noted earlier, our method does not work with these reserves.
There are three issues of interest relating to measuring reserve *variability:
-
The distribution of the true (but unknown) ultimate losses by line and in aggregate. This shows the volatility and the bias in the actuarial selections.
-
The held reserves that are reported in Schedule P Part 2 as a percentile on the distribution of reserves.
-
A procedure to allocate the margin by line and accident year such that ultimate losses for all years are at a constant percentile.
It is instructive to understand several aspects of indicated reserves.
-
Indicated reserves are the actuary’s best estimates based on the data and exclude management adjustments. Therefore, errors are due only to actuarial selections, methods, or randomness in the data.
-
Data triangles used for reserve reviews can be quarterly or annual. Generally, most companies like to have consistency in the indicated reviews because it is easier to update spreadsheets for each review if they are the same size. Also, most companies like to track development to ultimate and prefer complete triangles if data is available.
-
Reserve reviews are done on a net and or gross basis and the underlying triangles are based on the relevant data. They are definitely conducted annually to report Annual Statement reserves but many companies do them quarterly.
-
Reserving actuaries refer to lines of business as “segments,” as they can be custom-defined by the actuary for reserve reviews. These can be different than the usual lines of business such as those defined in Schedule P. For consistency of notation in this paper, we will refer to reserve segments as lines of business. The methodology will be the same in both cases.
Generally the DCC (Defense and Cost Containment) reserve review (indicated reserves) is done using data where they are either a part of the loss triangle (loss+DCC) or are treated separately (DCC only). In the first case we treat DCC as part of the losses, and in the latter case as another data segment (if broken out separately). In this paper, loss shall mean whatever appears in the analysis being evaluated.
The variability for ULAE reserves is outside the scope of the paper and will not be discussed. For most companies ULAE reserves are a relatively small part of the total reserves.
2.1. Scope of model
The approach presented in this paper is generic and applies to any type of triangle, such as
-
Paid or incurred
-
Count or severity
-
Accident year, policy year, or report year
-
Quarterly or annual.
Each of these triangles eventually leads to the appropriate ultimate losses. Depending on the choices made above and the selected reserve methodology, the distribution of errors will differ. For example, if only the paid loss development method is used and we track its history of ultimate losses over time, then the resulting reserve distribution will pertain to the paid loss development method. If several methods are used and the actuary finally selects reserves based on several indications (typically this is the case), then the triangle history of final selected ultimate losses will provide the distribution of the selected reserves.
An interesting point is that the method can even be applied to the raw data. This is equivalent to treating the data as the selected ultimate loss. In this case, the resulting distribution will pertain to the data itself (paid or incurred losses, etc.). This essentially creates a new reserving method. However, the purpose of this paper is not to promote a new way of calculating reserves, but to develop a method for determining the distribution of the indicated reserves resulting from actuarial selections.
2.2. Reasonable estimates
A standard assumption that is in line with Actuarial Standards of Practice (ASOP) on reserves is that the reserve ranges are set around reasonable estimates. This is partly achieved by using indicated ultimate reserves instead of held reserves, as these do not have management adjustments.
Cases where the indicated ultimate losses are unreasonable are outside the scope of this paper. This does not necessarily mean that the model will not apply, but rather that the authors have not given consideration for such cases in this paper.
3. A model for errors
3.1. Measuring the error from the data
The notation will be illustrated with an example that will be carried through the paper. Suppose for a particular line we believe that losses are fully developed after 10 years. There have been 12 reserve reviews completed and in each year an ultimate loss has been estimated. For notation, let
- —the estimated ultimate loss as of calendar year k for accident year i.
The results for an example block of business are in Table 1. Note that the available data has i and k range from 1 through 12 but, for example, the accident year 1 losses were fully developed by calendar year 10.
For example, the value 42,894 is the indicated ultimate loss estimated at the end of calendar year 10 for losses incurred in accident year 6. We are interested in the errors made in the estimates of the ultimate losses. Some of those errors can be determined from Table 1. We will use the logarithm of the ratio for the errors, for reasons to be explained later.
For example, the ultimate value for accident year 3 was estimated at the end of calendar year 11 (development year 8) to be 85,626 . A year later, the actual value of 85,650 was known. The error is
We cannot calculate the error for later accident years because they have not yet been fully developed. The errors we care about are where The numerator is the fully developed ultimate loss and the denominator is the estimate as of calendar year For our example we are currently at and so are concerned with the errors made in estimates from accident year 4 onward.In order to gather more information about errors, begin with an error that is not immediately useful. Consider
where available. This represents the error realized as the estimated ultimate value is updated one calendar year later. Here is one pragmatic reason for using logarithms-the errors are additive. In fact,\[ \begin{aligned} e_{i}^{k} & =\ln \left(U_{i}^{i+9} / U_{i}^{k}\right) \\ & =\ln \left(\frac{U_{i}^{k+1}}{U_{i}^{k}} \frac{U_{i}^{k+2}}{U_{i}^{k+1}} \cdots \frac{U_{i}^{i+9}}{U_{i}^{i+8}}\right) \\ & =\sum_{g=0}^{i+8-k} \ln \left(U_{i}^{k+g+1} / U_{i}^{k+g}\right) \\ & =\sum_{j=k}^{i+8} e_{i}^{j *}. \end{aligned} \]
In this notation, the development year of the denominator value is d = j − i + 1. The advantage of this approach is that the available data in our example provides many estimated values, as presented in Table 2.
There is another interpretation of these errors. Regardless of the reserving method, a factor representing the ultimate development can be inferred. Suppose we are looking at accident year
and calendar year The factor is where is the paid loss for that accident year at the end of year One year later the factor is These represent age to ultimate for two different development years. The ratio represents how losses were expected to develop, while is how they actually developed. The ratio of actual to expected is which is the error measurement we have been using.Now that the key data values have been calculated, it is time to construct a model.
The model
We propose that the error random variables have the normal distribution, and in particular,
\[ e_{i}^{j *} \sim N\left(\mu_{j-i+1}^{*}, \sigma_{j-i+1}^{* 2}\right) . \]
Note that the mean and variance are constant for a given development year. The rest of this section is devoted to justifying these assumptions. It should be noted that justifying the assumptions through data analysis is not sufficient. If this method is to be useful regardless of the loss reserving method used, the justification must be based on our beliefs about the loss development and reserving processes and not any particular models or empirical evidence. No independence assumptions are made and the covariance measurement will be presented formally as part of the model.
3.2.1. Normal distribution
After the ultimate loss is estimated at a particular time, what factors will cause it to change when it is re-estimated one year later? Day by day during that year a variety of events may take place:[2]
-
Economic forces such as inflation and changes in the legal environment will alter the amounts paid on open claims or those newly reported.
-
The rate at which claims develop may change.
-
Purely random events may affect individual open claims.
-
The actuary’s opinion on IBNR may change, depending on the adequacy of case reserves.
These factors will tend to act proportionally on the current estimate of the ultimate loss. Because there are many such factors happening many times in the course of a year, it is reasonable to assume we are looking at the product of a variety of random variables, most with values near 1. The Central Limit Theorem indicates that the result is a lognormal distribution and thus measuring the error in the logarithm will produce values with a normal distribution. We recognize that this reasoning is not accepted by all actuaries but we feel it is a reasonable starting point.
3.2.2. Normal approximation
The above arguments relating to the central limit theorem apply to a fictitious accident year with infinite claims. This accident year need not be evaluated infinitely many times but the changes at each valuation should be driven by infinite reasons underlying infinite claims.
In practice, a finite subset of this hypothetical accident year is available. The reasons causing the change in the ultimate loss are finite and not infinite and thus the distribution is approximately normal.[3] The approximation can be improved by increasing the number of reasons driving the change in ultimate losses. This can be done in the following ways:
-
Increase the valuation time (for example, once a year rather than four times a year).
-
Increase the claim count (large volume line).
The claim count required for a good normal approximation in turn depends on the skewness of the underlying size of loss distribution. As a practical matter, the model will work well even for a modest claim count.
Reserve valuations conducted less frequently (for example, once a year rather than four times a year) will allow greater “reasons” for changes and thus help with the normal approximation. The tradeoff is the loss of accuracy in ultimate loss estimation and consequential bias. Also note that both the parameters of the distribution and its shape (to which extent it is lognormal) will change depending on the frequency of reserve review.
A subtle point concerns the later development intervals. In these intervals, the changes are usually driven by a handful of claims (few reasons) and thus violate the normality assumption. However, at those valuations the aggregate errors are often close to zero with small volatility (constants) and thus departure from normality has little impact on the total distribution for the entire accident year.
3.2.3. Mean
At first it may seem that the mean should be zero. There are two reasons that is not so. First, consider the expected revised ultimate loss given the current estimate:
\[ E\left(U_{i}^{j+1} / U_{i}^{j}\right)=e^{\mu_{j-i+1}^{*}+\sigma_{j-i+1}^{* 2} / 2}. \]
For the reserve estimate to be unbiased, it is necessary that
and not zero.In addition, it is possible that the reserving method is biased. This may be a property of the method selected and may even be a deliberate attempt by the reserving actuary to adjust the reserves based on knowledge that is outside the data.
Having the mean be constant from one accident year to the next is a consequence of the assumptions that were made at the beginning. That is, there is no distributional change in the underlying development process or reserving method.
Having the mean depend on the development year seems reasonable. As the development year increases, any systematic bias is likely to be reduced in the expectation that the ultimate value will not be much different from the current value. In addition, we expect the variance of the errors to decrease, and if the ultimate loss estimates are unbiased, the means will then also decrease (in absolute value).
3.2.4. Complete triangle of errors
For at least one accident year we need the losses to be fully developed. Even better would be to have a few fully developed years so there would be more data available for computing covariances and variances. There is a tradeoff: older fully developed accident years that are not part of the in-force book may be less predictive and may not reflect the current prevailing business environment.
The opposite concern is when there are not enough accident years available to obtain fully developed losses. This can happen for relatively new companies or lines of business. Note that tail errors generally result from estimation errors of pending court cases or simply absence of data due to a new line. Since we are dealing with a year in aggregate, these errors are usually close to zero and therefore contribute less to the variance than earlier development intervals.
4. Parameter estimation
Return to the continuing example and recall Table 2. The numbers in this table are percentage changes (errors) of the estimated ultimate losses and have been assumed to have been drawn from six normal distributions. The upper half of the triangle is fixed and known and our goal is to obtain the normal means and variances for the yet-to-be observed errors in the lower right part of the table.
4.1. Estimation of the mean
We choose to calculate an initial estimate of the mean by calculating the sample mean,
\[ \hat{\mu}_{d}^{*}=\frac{\sum_{i=1}^{12-d} e_{i}^{(i+d-1) *}}{12-d}, \]
where d is the development year, the sum is taken over all accident years for which errors are available.[4] The sample mean is also the maximum likelihood estimate of the true mean and works best if the data is stable. If the dataset is not stable, then other estimates such as those involving time series analysis can be used.
The results for our example are in Table 3.
However, there are reasons why the sample mean may not be the appropriate choice. The results may be biased due to model risk. The estimate of the mean affects the estimate of the expected ultimate loss.
For example, if μ < −σ2/2 then the estimated ultimate loss will be less than the indicated ultimate showing redundancy in resulting reserves. The opposite is true if the inequality is reversed. Thus any value of mean that is different than −σ2/2 should be justified.
The estimation of the mean can effectively result in taking a position that the actuarial estimate is biased. This is not trivial and there should be careful analysis before making that assertion. We discuss the mean selection in greater detail under distribution reviews later in the paper.
4.2. Estimation of the variance
The variance parameter measures the risk historically faced by the book in force. The variance estimate is the usual unbiased estimate. The equation is
\[ \hat{\sigma}_{d}^{* 2}=\frac{\sum_{i=1}^{12-d}\left[e_{i}^{(i+d-1) *}-\hat{\mu}_{d}\right]^{2}}{11-d} . \]
Note that the estimated mean must be the sample mean and the above is not the maximum likelihood estimate.
For our data the estimated standard deviations by development year are given in Table 4. As expected, for the most part, the standard deviations decrease by lag.
4.3. Correlations
For a given accident year, the errors for one development year may be correlated with those from other development years.
Any such correlations can be estimated. A formula for the covariance is
\[ \hat{\sigma}_{d, d^{\prime}}^{*}=\frac{\sum_{i}^{12-d}\left[e_{i}^{(i+d-1) *}-\hat{\mu}_{d}\right]\left[e_{i}^{\left(i+d^{\prime}-1\right) *}-\hat{\mu}_{d^{\prime}}\right]}{11-d} \]
where d > d′ represent two different development years and each sample mean is based only on the first 10 − d observations.
The matrix of covariances is given in Table 5.
The total variance for any given accident year can now be calculated from Table 5. For example, the variance for accident year 12 will be the sum of the values in the table (shown as 0.46129 in Table 6).
The correlation matrix underlying the above covariances should be tested for statistical significance. This is outside the scope of this paper.
5. The error distribution for a given accident year
Now that we have a model for the errors from one calendar year to the next, we need to return to the total error. Recall that
\[ e_{i}^{k}=\sum_{j=k}^{i+8} e_{i}^{j *}. \]
From the assumptions made previously, this error has a normal distribution and its mean depends only on the latest development year, k − i + 1.
\[ \begin{aligned} e_{i}^{k} & \sim N\left(\mu_{k-i+1}, \sigma_{k-i+1}^{2}\right) \\ \mu_{k-i+1} & =\sum_{j=k}^{i+8} \mu_{j-i+1}^{*} \\ \sigma_{k-i+1}^{2} & =\sum_{j=k}^{i+8} \sigma_{j-i+1}^{* 2}+2 \sum_{j=k+1}^{i+8} \sum_{j^{\prime}=k}^{j} \sigma_{j-i+1, j^{\prime}-i+1}^{*} . \end{aligned} \]
These quantities can be estimated by adding the respective mean, variance, and covariance estimates.
For our example, we now have estimates of the distribution of the error in the ultimate loss as estimated from the data available at the end of calendar year 12. They are given in Table 6.
6. The error distribution for all accident years combined
While we have been using logarithms to measure the error, when all is done we are interested in the ultimate losses themselves. To make the formulas easier to follow, rather than allow arbitrary values, we will follow the example and assume we are at the end of calendar year 12 and losses are fully developed after 10 years. In particular, we care about
\[ U=U_{4}^{13}+\cdots+U_{12}^{21} . \]
Recall from the notation that the terms on the right-hand side represent the fully developed losses at a time in the future when the ultimate results are known. Rewrite this expression as
\[ \begin{aligned} U & =U_{4}^{12} \exp \left(e_{4}^{12}\right)+\cdots+U_{12}^{12} \exp \left(e_{12}^{12}\right) \\ & =V \sum_{i=4}^{12} r_{i} \exp \left(e_{i}^{12}\right) \\ r_{i} & =\frac{U_{i}^{12}}{U_{4}^{12}+\cdots+U_{12}^{12}} \\ V & =U_{4}^{12}+\cdots+U_{12}^{12} . \end{aligned} \]
Here V is the estimated ultimate loss as of calendar year 12 for all open years and the weights are the relative proportions in each accident year. This indicates that the ultimate loss is a weighted average of lognormal random variables. This can be painful to work with (though not hard to simulate). Because the error random variables are usually close to zero and will vary about their mean, consider the following Taylor series approximations.
\[ \begin{aligned} X & =\ln \frac{U}{V}=\ln \left[\sum_{i=4}^{12} r_{i} \exp \left(e_{i}^{12}\right)\right] \\ & \approx \ln \left[\sum_{i=4}^{12} r_{i}\left(1+e_{i}^{12}\right)\right] \\ & =\ln \left(1+\sum_{i=4}^{12} r_{i} e_{i}^{12}\right) \\ & \approx \sum_{i=4}^{12} r_{i} e_{i}^{12} . \end{aligned} \]
The approximate log-ratio has a normal distribution and thus U has an approximate lognormal distribution. The moments of the normal distribution are:
\[ \begin{aligned} E\left(\ln \frac{U}{V}\right) & =E\left(\sum_{i=4}^{12} r_{i} e_{i}^{12}\right) \\ & =\sum_{i=4}^{12} r_{i} \mu_{13-i} \\ & =\mu \\ \operatorname{Var}\left(\ln \frac{U}{V}\right) & \approx \operatorname{Var}\left(\sum_{i=4}^{12} r_{i} e_{i}^{12}\right) \\ & =\sum_{i=4}^{12} r_{i}^{2} \sigma_{13-i}^{2} \\ & =\sigma^{2}. \end{aligned} \]
For the example, V = 760,808, μ = 0.01927, σ2 = 0.01123, and then the expected ultimate loss is 779,978 and the standard deviation is 82,892.
7. Extension to multiple lines of business
Suppose there are two lines of business. Each can be analyzed separately using the method previously outlined. However, it is likely that the results for the lines are not independent. One approach would be to model the correlation structure between error values from the two lines. The problem with that approach is that there may not be corresponding cells to match and also the number of parameters may become prohibitive.[5]
An alternative is to combine the data from the two lines into a single triangle and analyze it using the methods of this paper. When finished there will be a distribution for each line separately and for the combined lines. Usually, only the total reserve is of importance and thus only the combined results are needed. The individual line results will be useful for internal analyses and also if it is desired to allocate the reserves back to the lines. This method assumes that the relative sizes of the lines has been consistent over time. Adjustments to the historical triangles can be made if this is not the case, as discussed later.
To illustrate this idea, we add a second line of business. The same analysis gives: V = 244,537, μ = −0.30759, σ2 = 0.008933, and then the expected ultimate loss is 180,593 and the standard deviation is 17,107.
Combining the two lines creates a single table with the totals from each. An analysis of these tables produces V = 1,005,376, μ = −0.02674, σ2 = 0.009582, and then the expected ultimate loss is 983,520 and the standard deviation is 96,506.
8. Allocation of ultimate losses
An illustrative allocation model is presented here but other models are also possible. From the previous examples there is an interesting situation. If the two lines were independent, the standard deviation of the total would be
\[ \sqrt{82,892^{2}+17,107^{2}}=84,638 \]
which is less than the standard deviation from the combined lines, which was 96,506. There is thus an implied positive correlation between the lines and when setting reserves and then allocating them to the lines, something needs to be done.
Suppose we set reserves with a margin for conservatism based on the results of the previous section. For example, suppose we set ultimate losses to be at the 95th percentile. That is,
\[ \begin{array}{l} 1,005,376 \exp (-0.02674+1.645 \sqrt{0.009582}) \\ \quad=1,149,833 \end{array} \]
where the mean and standard deviation are for the corresponding normal distribution.
We also want to set ultimate losses for each line of business such that they add to 1,149,833. However, this cannot be achieved by setting each line at its 95th percentile. One possible approach is to set each line at the same percentile, using the percentile that makes the sum work out. It turns out that if each line is set at the 96.28th percentile, the ultimate estimates will be 937,025 and 212,808.
9. Model extension to practical settings
To this point we have assumed constant distributions and reserving methods. In practice these are not always present, so some extensions are discussed here.
9.1. Distribution reviews
Once the reserving actuary has completed the reserve review the distribution review should follow. The actuary will remember the considerations for selecting the ultimate losses. These include data considerations, coverage changes, mix changes, etc. All of these can now be factored in the bias (mean parameter) estimation of resulting reserve distribution.
The distribution review allows the actuary to consider the possibility of estimation bias in the current reserve estimate. For example, if the historical errors of the selected ultimate losses for a line are positive and the actuary has not changed the selection approach, then the current estimate will likely have a positive error (too low an estimate).
The reserve diagnostics are particularly important in evaluating such errors, as the past is not always indicative of the future. For example, a sign of case reserve weakening should lead to a higher IBNR, all else being equal.
These mechanisms of monitoring errors allow early warning signs for future potential reserve deficiencies.
Finally, note that changes in conditions do not invalidate the lognormal property. However, changes in conditions will change the lognormal parameters, thereby increasing the degree of difficulty for projection accuracy.
9.2. Changing internal/external conditions
While the distribution reviews and the resulting mean parameter will account for changes in reserve adequacy, the volatility parameter reflects historical data and is therefore slow to respond to changes in the errors. Thus, if the reserving methods change abruptly or the reserving actuary is replaced, the volatility will change slowly as the data emerges. Such cases can be handled in the following two ways:
-
In many cases it is wise to let data lead the way as it may be preemptive to conclude about the volatility of the new process. This is especially relevant in light of the fact that processes themselves usually change slowly over time (such as experience of reserving actuary, etc.).
-
In some cases, the change is driven by an abrupt decision, such as outsourcing the reserve review to a consultant or change in treaty limits. In that case, the historical error data will have to be restated with the new process. This involves recalculation of the past reserve estimates adjusted for the new data and using the new methodology.
To be explicit, the volatility measurement in the model relies on two main assumptions:
-
The data is consistent in terms of reserve model/process and retentions.
-
Future volatility in ultimate loss selections can be measured from historical data. Note that we are asking the future volatility in loss selections to be similar to the past and not the consistency in interest rates, etc.
9.3. Booked reserve
The model was used on indicated reserve data for consistency reasons since management adjustments can often lead to consistency issues. If the held reserve data is relatively consistent, then the model will perform well with that data (Schedule P Part 2), as well.
9.4. Lognormal assumption
We present here two discussions:
Discussion 1. “The ultimate losses are hypothesized to be lognormally distributed and therefore the sums of such lognormal distributions are not exactly lognormal.”
We first note that at an accident year level we do not add lognormal random variables. In other words, we study the given data that has been obtained from an underlying reserve review process and not artificially created by an addition process. These two things are not the same, as the belief in the underlying process is the key to describing the correct modeling distribution.
In instances where we add lognormal random variables (such as Section 6) we show mathematically that if the cumulative accident year error
is relatively small, then the sums of lognormal random variables will be approximately lognormal. Note also that the weighted average is approximately lognormal, so adding a nonlognormal accident year ultimate loss random variable with a small weight with a lognormal random variable with a large weight will still produce a lognormal random variable. The weight is based on selected ultimate loss.Discussion 2. Why not use any other distribution besides the lognormal?
Beside the theoretical argument given for the use of lognormal, we suggest performing an empirical test on the data. In the end, the model does not depend on the lognormal assumption and any other distribution can be used, but we believe that it is the best distribution that describes the underlying reserve process. Additionally, the lognormal has the advantage that the mathematical calculations are tractable. With other distributions the calculations are much harder.
The Q-Q plot is a simple way to test for normality, the error triangle can be standardized using the mean and variance of each column of errors and the graph inspected for N(0, 1).
9.5. Changes in volume
In Section 7 we added the ultimate loss data from different lines and applied the model to the total dataset. This approach assumes relative consistency in volume by line.
9.5.1. Analytical approach
An alternative is to measure the total distribution from the parts. Suppose we know the individual line lognormal distributions. Let
- Vm = Total (for all accident years) selected ultimate loss for line m.
We showed earlier that for a given line,
\[ e_{m}=\ln \frac{U_{m}}{V_{m}} \sim N\left(\mu_{m}, \sigma_{m}^{2}\right) . \]
We now combine individual lines,
\[ \begin{aligned} U & =\sum_{m=1}^{i=M} U_{m}=\sum_{m=1}^{m=M} V_{m} \exp e_{m}=V \sum_{m=1}^{m=M} w_{m} \exp e_{m} \\ w_{m} & =\frac{V_{m}}{\sum_{m=1}^{m=M} V_{m}} ; \quad V=\sum_{m=1}^{m=M} V_{m} \\ \ln \frac{U}{V} & =\ln \sum_{m=1}^{m=M} w_{m} \exp e_{m} \approx \ln \sum_{m=1}^{m=M} w_{m}\left(1+e_{m}\right) \\ & =\ln \left(1+\sum_{m=1}^{m=M} w_{m} e_{m}\right) \approx \sum_{m=1}^{m=M} w_{m} e_{m} \\ \ln \frac{U}{V} & \sim N\left(\sum_{m=1}^{m=M} w_{m} \mu_{m}, \sum_{m=1}^{m=M} \sum_{k=1}^{k=M} w_{m} w_{k} \operatorname{cov}\left(e_{m}, e_{k}\right)\right) . \end{aligned} \]
The above used Taylor series approximation of the “exp” and “ln” operator. To measure the covariance term in the above expression, we need additional notation. Let
\[ j=\text { Development interval. } \]
Due to the lognormal assumption of the total ultimate loss distribution, we can view the line error as the sum of errors by development interval,
\[ e_{m}=\sum j=1^{j=J} e_{m j}. \]
Thus,
\[ \begin{aligned} \operatorname{cov}\left(e_{m}, e_{k}\right) & =\operatorname{cov}\left(\sum_{j=1}^{j=J} e_{m j}, \sum_{l=1}^{l=J} e_{k l}\right) \\ & =\sum_{j=1}^{j=J} \sum_{l=1}^{l=J} \operatorname{cov}\left(e_{m j}, e_{k l}\right). \end{aligned} \]
9.5.2. Simulation approach
The analytical approach uses the first-order Taylor series approximation to ensure that the total reserve distribution for all lines combined is lognormal. This means that we have used the Taylor approximation twice—at a line level and then at the total reserve level. To avoid approximating twice, we can simulate the line distributions using the covariance shown below:
\[ \begin{aligned} \operatorname{cov} & \left(U_{m}, U_{n}\right) \\ & =\operatorname{cov}\left(\sum_{m=1}^{m=M} V_{m} \exp e_{m}, \sum_{k=1}^{k=M} V_{k} \exp e_{k}\right) \\ & =\operatorname{cov}\left(\sum_{m=1}^{m=M} V_{m}\left(1+e_{m}\right), \sum_{k=1}^{k=M} V_{k}\left(1+e_{k}\right)\right) \\ & =\operatorname{cov}\left(V+\sum_{m=1}^{m=M} V_{m} e_{m}, V+\sum_{k=1}^{k=M} V_{k} e_{k}\right) \\ & =\sum_{m=1}^{m=M} \sum_{k=1}^{k=M} V_{m} V_{k} \operatorname{cov}\left(e_{m}, e_{k}\right). \end{aligned} \]
The above used Taylor series approximation of the “exp” operator.
10. Applications
We now present a few applications of the previous results.
10.1. Loss estimation method
The chain ladder estimate is not biased under loss development factor independence assumption of Mack (1993). If the development factors are correlated, then the model provides a way to incorporate covariance of loss development factors into the chain ladder estimate. Using the technique presented in this paper (including the variance covariance matrix of a triangle) the expected ultimate loss for an accident year is
\[ E(\mathrm{U})=\mathrm{U}_{\mathrm{o}} \exp \left(\mu+\frac{\sigma^2}{2}\right), \]
where U = Ultimate loss, Uo = Incurred or paid loss, σ2 = sum of variance covariance matrix (the number of terms depend on the age of the accident year).
10.2. Claim commutations
Claim commutations involve transfer of reserves from a ceding carrier to an acquiring carrier. The risk to the acquiring company is that the indicated reserves may not be sufficient to pay all claims. One complication is that the agreed-to reserve value may not be the same as that used in the determination of the reserve distribution. Let H be this arbitrary reserve value. For pricing it would be useful to have C, the expected reserve cost above H (given that the actual loss is above H). The formula for C is
\[ \mathrm{C}=\int_{\mathrm{H}}^{\infty}(x-\mathrm{H}) f_{\mathrm{R}}(x) d x \]
where R is the random true reserve. Let P be the amount paid and, as before, U is the random, true, ultimate loss. Then R = U − P and so
\[ \begin{aligned} \mathrm{C}= & \int_{\mathrm{H}}^{\infty}(x-\mathrm{H}) f_{\mathrm{R}}(x) d x \\ = & \int_{\mathrm{H}}^{\infty}(x-\mathrm{H}) f_{U}(x+\mathrm{P}) d x \\ = & \int_{\mathrm{P}+\mathrm{H}}^{\infty}(y-\mathrm{P}-\mathrm{H}) f_{U}(y) d y \\ = & E(U)-E(U \wedge \mathrm{P}+\mathrm{H}) \\ = & \exp \left[\mu+\ln (V)+\sigma^{2} / 2\right] \\ & \times\left[1-\Phi\left(\frac{\ln (\mathrm{H}+\mathrm{P})-\mu-\ln (V)-\sigma^{2}}{\sigma}\right)\right] \\ & -(\mathrm{H}+\mathrm{P})\left[1-\Phi\left(\frac{\ln (\mathrm{H}+\mathrm{P})-\mu-\ln (V)}{\sigma}\right)\right] . \end{aligned} \]
The second line is a change of variable using R= U − P.
The same concept applies in situations where one company acquires or merges with another company or a reinsurer acquires the reserve of the ceding carrier (loss portfolio transfer).
Another way to determine the charge is to calculate the stress reserve value for a given percentile. This can be done easily since we know the reserve distribution. The difference between this and the held reserve is the reserving capital or the commutation charge.
In these cases, C or the full capital is likely conservative and a certain fraction may be more appropriate as a charge.
10.3. Insurance market segmentation
The expected excess cost formula gives an insight into reinsurance economics. The assuming carrier will pool the assumed reserves into existing homogenous reserve segments. If the pooling decreases the total risk, captured in the estimated σ2 for the line, the assuming carrier’s expected excess cost will be lower than ceding carrier’s. This assumes that the reserve adequacy of the reserve segment is estimated identically by both parties.
The above argument explains insurance market segmentation. Companies grow their business in a given line and continue to acquire business from smaller companies in the same line because they face different total risks. Unlike other businesses, the volatility of losses underlying estimates drives decisions to acquire and grow an existing business segment. This is also the reason why low-risk (short-tailed, fast-developing) lines are seldom ceded to other companies.
10.4. Net and gross reserve distributions
It is possible to quantify the reserve distribution net of reinsurance and/or recoveries using the method explained in the paper because we are following the net reserve reviews and simply measuring the uncertainty in the estimates.
One caveat when dealing with net distributions is that the true mean can be harder to estimate, especially if treaties have changed recently. As stated earlier, quarterly error triangles are more helpful in such cases because they are more responsive to changes. In other cases, the current actuary’s estimate can be taken to be unbiased until further history is developed under the new treaty.
If reinsurance has a relatively small impact on total reserves, a change in treaty provisions will not change the resulting net error distribution significantly. For example, quota share reinsurance on a loss occurring basis on a line with large claim count per accident year will not necessarily lead to a lower variability of the net aggregate error distribution. The reserving actuary will see proportionally lower losses for new accident years but this may not impact the net error distribution.
Another example is casualty excess of loss coverage that attaches at a high layer. If the ground up claim count is large enough, the change of reinsurance treaty many have very little impact on the net error distribution as few losses out of the total will be ceded in that layer.
In some cases, especially for low-frequency and high-severity lines such as personal umbrella excess of loss coverage, the impact of a change in treaty can be significant and the error history will not be relevant to the current treaty. This will require the actuary to conduct net historical reserve reviews based on the new treaty. This can be time consuming but should be done if the line forms a significant part of the total reserve.
Net distributions result in net reserve ranges and net reserving capital. These are relevant from a solvency, regulatory, and company rating standpoint.
10.5. Regulatory and rating agency applications
Regulators and rating agencies are interested in quantifying reserve ranges, percentiles, and reserving capital in order to monitor the solvency of the company. The regulator in particular is interested in measuring the performance of the held reserve, shown in Schedule P.
Once the distribution of the reserves is known, we can state the percentile of the held reserves. Note that the management bias will now make a difference, as it will position the held reserves to the right or left of the mean.
We can also calculate the reserving risk capital as the difference between a selected cutoff value (say 95th percentile) of the indicated reserves and the held reserves. The resulting reserving risk capital can be used to modify the RBC reserving risk charge. Note that the NAIC tests of reserve development to surplus ratios suggest comparing dollar ultimate loss errors to company surplus. This is very similar to our approach of measuring reserve uncertainty. Thus, the reserving risk charge measured using the method outlined in this paper would be consistent with the current annual statement and NAIC practices.
11. Conclusion
This paper presents a shift in paradigm from loss distributions of the underlying losses to distributions of the company’s estimates. This represents a new way of measuring reserving risk. By being able to quantify risk both by line and for the company, effective management of capital, reinsurance, and other company functions becomes feasible.
We provide a framework for assessing reserve review accuracy as well as measuring the distribution of the current reserve review. This is done using a “distribution review” immediately after a reserve review using the same segments and models currently used by the reserving department.
Given the current complex reserving environment where reserving is both an art and a science, there is no statistical formula to set reserves or its distribution. Rather we present a rigorous framework that involves the same considerations and process as the underlying reserve review itself.
Acknowledgments
We are indebted to Zhongzian Han (Statistics and Actuarial Science Department of the University of Central Florida) for his assistance with the literature review. The authors would also like to thank William J. Gerhardt (Nationwide Insurance) for being a great sounding board in the formative stages of the paper.
The authors are also grateful to an anonymous reviewer for extensive data testing and feedback.
The model presented will capture the parameter and model risk in the actuary’s estimate but will not measure the parameter uncertainty due to its own estimate.
At this point in the paper we focus on lognormality as a consequence of unchanging development processes. We relax this assumption later in the paper.
Technically, the distribution remains approximately normal even for the infinite claim accident year, but the distribution is closer to normal than in the finite case.
An alternative is to use a weighted average where the weights are the indicated ultimate values for that accident year. This allows more weight to be placed on those accident years in which there is more data.
Combining distributions also requires the multivariate normality assumption. This is hard to test but is often made in practice.