1. Introduction
Over the past several decades, generalized linear models (GLMs) have become ubiquitous in actuarial science. Within the insurance industry, GLMs were first utilized in the property and casualty industry, due to its inherent need for custom modeling. However, GLMs and predictive modeling have recently found application in the life insurance industry as well. GLMs have enjoyed this popularity largely due to their ability to overcome the limitations of ordinary least squares regression (OLS) when applied to data which exhibit non-linear relationships, and non-normal distributions. In their traditional form, GLMs require that the dependent variable have a distribution from the exponential family of distributions. This requirement primarily ensures the existence of a desirable relationship between the variance and mean of dependent variable. In most situations this restriction is at most a minor inconvenience; however, there are circumstances in which data can be best described by distributions outside of the exponential family of distributions. In such cases, it can be beneficial to have access to an even more general class of models. In response to this need, Parsa and Klugman (2011) proposed a regression modeling framework which exploits the flexibility of copula techniques.
Not long after the introduction of copula regression, however, it was noticed that the formulation of copula regression as a conditional expected value could be approximated by an OLS regression on appropriately transformed data. Such an approximation would have the advantage of ease of implementation, using almost any statistical software or even Microsoft Excel. Moreover, upon investigating this linear approximation to copula regression, Parsa and Klugman (2011) noticed that the approximation often produced estimates which were quite close to, yet consistently underestimated those from exact copula regression. Further, there seemed to be a systematic deviation between the estimates from each method. These observations motivated the current research. This paper will proceed as follows: In section 2 we briefly describe copulas, and copula regression. In section 3 we describe the linear approximation to copula regression, which is the main subject of this research. In section 4 we investigate sufficient conditions which ensure a predictable pattern of bias in the estimates from the linear approximation to copula regression. In section 5 we introduce transmutation mappings, which play an integral role within the linear approximation to copula regression, and also within the theory of copulas in general. In section 6, we investigate the most commonly used loss distributions to see if they satisfy the sufficient conditions from section 4, and hence lead to predictably biased estimates (with respect to full copula regression) from the linear approximation to copula regression.
2. Copula regression
Copula methods are essentially just a particular way to construct multivariate distributions. Copula functions themselves are simply multivariate distributions whose domain has uniform marginal distributions. As described in Parsa and Klugman (2011), a copula model can be constructed for an n-dimensional set of data by first fitting n marginal distributions to the data Then the joint-distribution, can be created by applying a copula function to specifically The copula function induces a correlation structure between the variables By construction, this correlation structure will be independent of the marginal distributions of the data. In order to ensure that the resulting multivariate function is in fact a distribution function, only certain functions can be used as copula functions. However, Sklar’s theorem guarantees that, for any set of random variables with associated marginal distributions a copula function, exists s.t.; Moreover, if the marginal distributions for are continuous, then there exists a unique copula function, defined on such that Further, in the case of continuous marginal distributions, the copula function allows the multivariate dependence structure to be separated from the marginal distributions of the data. At this point, it is important to point out that, although the mathematical theory of copulas is very rich, and many powerful results have been proven, not every multivariate distribution can be naturally, or “usefully”, modeled using copulas. To quote Paul Embrechts: “copulas form a most useful concept for a lot of applied modeling, they do not yield, however, a panacea for the construction of useful and well-understood multivariate density functions” Embrechts (2009). While there are many copula models, most are only bivariate, and hence do not allow for fully independent correlation between the response variable and each of the covariates. For this reason, Parsa and Klugman assert that the multivariate normal and -copula are the most useful for the purposes of copula regression. In basic OLS regression, the distribution of given the covariates is assumed to be normal, and the predicted values are specified by The predicted values, under copula regression, are defined analogously.
Definition
Given N observations of the RVs Parsa and Klugman (2011), is the process of estimating the observed values of the RV based on the corresponding values of the RVs namely
copula regression, as defined by\[ \hat{y}_{i}=E\left[Y \mid X_{1}=x_{1 ;}, \ldots, X_{n-1}=x_{n-1 ;}\right], \tag{1} \]
where the conditional expected value is computed with respect to the conditional density of the RVs
induced from a multivariate copula function i.e.,\[ \begin{array}{l} F_{x_{1}, \ldots, x_{n}}\left(X_{1}, X_{2}, \ldots, X_{n-1}, Y\right) \\ \quad=C\left[F_{1}\left(X_{1}\right), F_{2}\left(X_{2}\right), \ldots, F_{n}\left(X_{n-1}\right), F_{Y}(Y)\right] \end{array} \tag{2} \]
In this paper, we concentrate on the case where the copula function C is the multivariate normal copula, and reserve investigation of the t-copula for subsequent research. We now review the basics of copula regression, under the multivariate normal copula. For a more complete treatment, the interested reader is referred to Parsa and Klugman (2011). The distribution function induced by the multivariate normal copula has the form:
\[ \begin{array}{l} F_{x_{1}, \ldots, x_{n-1},}\left(X_{1}, \ldots, X_{n-1}, Y\right) \\ =G\left(\Phi^{-1}\left[F_{1}\left(X_{1}\right)\right], \ldots, \Phi^{-1}\left[F_{n-1}\left(X_{n-1}\right)\right], \Phi^{-1}\left[F_{y}(Y)\right]\right) \end{array} \tag{3} \]
where
are the marginal distributions of the RV’s and is the multivariate normal distribution. In what follows we assume that all marginal distributions are continuous. Under this assumption, the density function induced by the multivariate normal copula can be written as follows:\[ \begin{array}{r} f\left(X_{1}, X_{2}, \ldots, X_{n-1}, Y\right)=f_{1}\left(X_{1}\right) \ldots f_{n}\left(X_{n-1}\right) f_{y}(Y) \\ \cdot \exp \left\{-\frac{\vec{v}^{T}\left(R^{-1}-\mathrm{I}\right) \vec{v}}{2}\right\} \times|R|^{-\frac{1}{2}} \end{array} \tag{4} \]
where:
- I is the Identity matrix.
- is the copula correlation matrix.
where
- and
- is the correlation between and and is the correlation between and for
Correspondingly, it is shown in Clement and Reilly (1999) that the conditional density of given has the following form:
\[ \begin{array}{l} f\left(Y \mid X_{1}, X_{2}, \ldots, X_{n-1}\right)=\frac{f_{y}(Y)}{\sqrt{1-\vec{r}^{T} \cdot R_{n-1}^{-1} \cdot \vec{r}}} \times \\ \exp \left\{-\frac{1}{2}\left[\frac{\left(\Phi^{-1}\left[F_{y}(Y)\right]-\vec{r}^{T} \cdot R_{n-1}^{-1} \cdot v^{*}\right)^{2}}{1-\vec{r}^{T} \cdot R_{n-1}^{-1} \cdot \vec{r}}\right]\right\} \times \\ \quad \exp \left\{\frac{1}{2}\left(\Phi^{-1}\left[F_{y}(Y)\right]\right)^{2}\right\} \end{array} \tag{5} \]
where
with and and are defined as above. Under the multivariate normal copula, specifies the predicted values up to specification of the marginal distributionsIdeally, to parameterize the copula regression model based on empirical data, maximum likelihood estimation is performed using the density function Parsa and Klugman (2011), this alternative will produce suboptimal results. The relative complexity of fitting the parameters of a full multivariate normal copula regression model with marginals, each with parameters, is one of the motivations for the linear approximation to copula regression, which is the topic this paper. For more details on the parametrization of copula regression models, the reader is referred to Parsa and Klugman (2011).
above. If and each marginal distribution has two parameters, it is necessary to estimate 44 parameters. Alternatively, one can use the marginal data to fit each of the marginal distributions and then a second optimization can be performed to estimate the correlations, within the MVN copula. As pointed out in3. A linear approximation to copula regression
Now that copula regression has been described, we introduce what the authors have dubbed the linear approximation to copula regression. The possibility of such an approximation to copulal regression arose organically in response to a question posed by a practicing property and casualty (P&C) actuary at the 2011 CAS Spring meeting. During the question and answer section of their presentation on copula regression, Parsa and Klugman were asked the following question: "For a given set of
since the distribution induced by the multivariate normal copula is essentially the multivariate normal distribution applied to the transformed RVs and since a fully probabilistic version of multivariate linear regression assumes that the RVs follow a multivariate normal distribution, how does copula regression, under a multivariate normal copula, differ from simply applying OLS regression to the transformed RVs:\[ \begin{array}{l} \Phi^{-1}\left[F_{1}\left(X_{1}\right)\right], \Phi^{-1}\left[F_{2}\left(X_{2}\right)\right], \ldots, \\ \Phi^{-1}\left[F_{n-1}\left(X_{n-1}\right)\right], \Phi^{-1}\left[F_{Y}(Y)\right] " \end{array} \]
In other words, if Y is considered the response variable, what is the difference in the estimates from the following models?
Perform OLS regression in the U, V→ space, and then transform back to the Y, X→ space:
- Transform each of the variables \[ \begin{aligned} U=\Phi^{-1}\left[F_{y}(y)\right] \text { and } V_{i}= & \Phi^{-1}\left[F_{i}\left(x_{i}\right)\right] \\ & \text { for } i=\{1,2, \ldots, n-1\} \end{aligned} \]
- Perform an ordinary OLS of on the to obtain : \[ U=\beta_{0}+\beta_{1} \cdot V_{1}+\cdots+\beta_{n-1} \cdot V_{n-1}+\varepsilon where \varepsilon \propto N(0,1) \]
- Then transform the back to the original scale: to obtain the estimates \[ \hat{y}_{i}=F_{y}^{-1}\left(\Phi\left(\hat{u}_{i} \mid V_{1}=v_{1, i}, \ldots, V_{n-1}=v_{n-1, i}\right)\right) \]
Model 2
Perform copula regression in the Y, X→ space:
\[\begin{align} \hat{y}_{i}&=E\left[Y \mid X_{1}=x_{1 i}, \ldots, X_{n-1}=x_{n-1 i}\right], \text{where} \\ &f\left(Y \mid X_{1}, X_{2}, \ldots, X_{n-1}\right) \text{is defined as in equation (5)}.\end{align} \]
Parsa and Klugman were immediately struck by the simplistic beauty of this reductive argument, and intrigued by the possibility that this approach could actually produce a close approximation to copula regression. Moreover, they soon realized that this question transcended copula regression, and actually addressed the larger issue of the effect of transformations of variables within regression modeling.
4. Sufficient conditions for bias
It seems likely that an understanding of the limitations of variable transformations, as a means to coerce data to fit the restrictive normal distribution assumption of linear regression, informed the development of GLMs. A simple example illustrates the difference between these two approaches: Suppose that it is known that the dependent variable Y follows a lognormal distribution. In introductory applied statistic classes, students are taught that in this case they should apply a log transformation to Y, and then perform OLS. The underlying model is that log(Y) = aX + b + where ∼ N(0, σ2). In other words, the assumption is that log(Y)|X has a normal distribution with mean aX + b, and constant variance. Then, one simply applies the inverse of the log transform to obtain the predicted values Ŷ = E(Y|X) = exp(aX + b). However, if Y in fact follows a lognormal distribution, then E(Y|X) = exp (aX + b + 1/2σ2). Hence, by applying the log transformation and assuming a normal distribution, the resulting estimate differs from the theoretically correct estimate by a factor of exp(1/2σ2). The disparity between these two estimates of E(Y|X) can be viewed as representing a bias in the estimates from the model involving transformations. Although it may escape consideration, such a disparity arises whenever transformations of variables are used within OLS. In light of this observation, it can be seen that the question at hand is fundamentally a question of the bias that is induced by applying the specific transformation Φ−1[Fi (·)] within OLS regression. Further, due to the general nature of the transformations Φ−1[Fi(·)], an investigation of the bias between copula regression and the linear approximation to copula regression may shed light on the bias induced by the use of transformations within OLS in general. As a first step towards investigating this bias, the linear approximation to copula regression was computed for the examples provided in Parsa and Klugman (2011). This resulted in several interesting observations. As anticipated by the motivating question, it was noticed that the estimates from the linear approximation to copula regression were in fact often very close to the estimates based on copula regression. Moreover, a consistent relationship between the two sets of estimates was observed. Specifically, the estimates from the linear approximation to copula regression were noticed to be consistently slightly lower than those from copula regression. This piqued the interest of the authors, making further investigation irresistible. The following results pave the way toward a more quantitative understanding of the relationship between the estimates from copula regression and the linear approximation to copula regression. Lemma 4.1 provides a connection between the copula regression of Y on X→, and the transformed variables U, V→. In what follows, the authors sometimes use the slightly more suggestive notation (Fy−1 ∘ Φ)(x), in place of Fy−1(Φ(x)).
Lemma 4.1
If
and for are the continuous CDF’s, corresponding to the RV’s and respectively, where then:\[ E(Y \mid \vec{X})=E\left[\left(F_{y}^{-1} \circ \Phi\right)(U \mid \vec{V})\right] \tag{6} \]
where U = Φ−1[Fy(Y)] and Vi = Φ−1[Fi(Xi)] for i {1, 2, . . . , n − 1}
Proof. Let U and V→ be defined as U = (Φ−1 ∘ Fy)(Y) and V→ = (Φ−1 ∘ Fi)(X→), where Φ−1 ∘ Fi is applied component-wise to X→. Let f(U|V→) be the conditional multivariate normal distribution of U given V→, then:
\[ \begin{aligned} E & {\left[\left(F_{y}^{-1} \circ \Phi\right)(U \mid \vec{V}=\vec{v})\right] } \\ & =\int_{u=-\infty}^{u=\infty}\left(F_{y}^{-1} \circ \Phi\right)(U) \cdot f(U \mid \vec{V}=\vec{v}) d U \\ & =\int_{u=-\infty}^{u=\infty} Y \cdot f(U \mid \vec{V}=\vec{v}) d U \\ & =\int_{y=h^{-1}(-\infty)}^{y=h^{-1}(\infty)} Y \cdot f(Y \mid \vec{V}=\vec{v})\left|\frac{d Y}{d U}\right| d U \\ & =\int_{y=0}^{y=\infty} Y \cdot f(Y \mid \vec{X}=\vec{x}) d Y=E(Y \mid \vec{X}) \end{aligned} \]
where h(y) = (Φ−1 ∘ Fy)(y). The
to last equality holds since Vi = vi → Φ−1[Fi(Xi)] = Φ−1[Fi(xi)] → Xi = xi, and since h′(y) > 0 for all y by the inverse function theorem, and the last equality holds since h(0) = −∞, and h(∞) = ∞.Next, Theorem 4.2 quantifies when the estimates from the linear approximation to copula regression will underestimate those from copula regression, by providing a sufficient condition for the under-estimation.
Theorem 4.2
If the CDFFy(Y) is continuous, and the mapping g(U) = (Fy−1 ∘ Φ)(U) is convex, then:
\[ E(Y \mid \vec{X}) \geq F_{y}^{-1}(\Phi(\hat{U})) \tag{7} \]
where E[Y|X→] is the conditional expectation used to define copula regression, and Fy−1(Φ(Û)) defines the linear approximation to copula regression.
Proof. Let
and for be the continuous CDF’s corresponding to the RV’s and respectively, where Let and be defined as 。 and where is applied component-wise to If is the conditional expected value with respect to the density induced by the multivariate copula given in equation (5), then by Lemma 4.1 we have that: and since is convex, Jensen’s inequality gives that:\[ E\left[\left(F_{y}^{-1} \circ \Phi\right)(U \mid \vec{V})\right] \geq\left(F_{y}^{-1} \circ \Phi\right)(E[U \mid \vec{V}]) \tag{8} \]
Further, since Fy−1(Φ(Û)) is the linear approximation to copula regression, Û is obtained from the OLS regression of U on the Vi. So, by definition Û = E[U|V→], we have that:
\[ F_{y}^{-1}(\Phi(\hat{U}))=\left(F_{y}^{-1} \circ \Phi\right)(E[U \mid \vec{V}]) \tag{9} \]
Hence, in conclusion, E[Y|X→] ≥ Fy−1(Φ(Û)).
So, the question is, for what CDFs Fy is the mapping (Fy−1 ∘ Φ)(·) convex? Armed with Theorem 4.2, we now know that convexity of the mapping (Fy−1 ∘ Φ)(·) is sufficient to ensure that the estimates from the linear approximation to copula regression will be bounded above, by the corresponding estimates from exact copula regression. We now present a set of criteria which ensure the convexity of the mappings (Fy−1 ∘ Φ)(·), and hence help quantify the bias in the estimates from the linear approximation to copula regression.
Lemma 4.3
IfFy(Y) is a continuous CDF, and y(x) = (Fy−1 ∘ Φ)(x), then y(x) is convex, for all x IFF:
\[ \frac{\phi^{\prime}(x)}{\phi^{2}(x)} \geq \frac{f_{y}^{\prime}(y(x))}{f_{y}^{2}(y(x))} \tag{10} \]
for all x where fy′(y(x)) denotes the derivative of fy(y(x)) WRT y, evaluated y(x), and Φ(x) is the standard normal CDF, and correspondingly, φ(x) is the standard normal density.
Proof. Since, y(x) = (Fy−1 ∘ Φ)(x), we have that Fy(y(x)) = Φ(x). Taking the derivative of both sides WRT x, we have that: f(y(x)) ·
= φ(x), which implies:\[ \frac{d y(x)}{d x}=\frac{\phi(x)}{f(y(x))} \tag{11} \]
Once again, taking the derivative of both sides WRT x, re-arranging terms, and substituting in the equality from equation (11) for
we have: Hence, we have:\[ \frac{d^{2} y(x)}{d x^{2}} \geq 0 \Leftrightarrow \frac{\phi^{\prime}(x)}{(\phi(x))^{2}} \geq \frac{f^{\prime}(y(x))}{(f(y(x)))^{2}} \tag{12} \]
The next corollary gives an equivalent condition for convexity of the mappings (Fy−1 ∘ Φ)(·), and is an immediate consequence of Lemma 4.3.
Corollary 4.4
If Fy(Y) is a continuous CDF, and y(x) = (Fy−1 ∘ Φ)(x), then y(x) is convex, for all x IFF:
\[ \frac{d}{d x} \log [\phi(x)] \geq \frac{d}{d x} \log [f(y(x))] \tag{13} \]
for all x where fy′(y(x)) denotes the derivative of fy(y) WRT y, evaluated at y(x), and Φ(x) is the standard normal CDF, and correspondingly, φ(x) is the standard normal density.
Proof. Since (x) 0 we have that:
where the last inequality follows, since:\[ \begin{aligned} \frac{f_{y}^{\prime}(y(x))}{f_{y}(y(x))} \frac{\phi(x)}{f_{y}(y(x))} & =\frac{d \log [f(y(x))]}{d y(x)} \cdot \frac{d y(x)}{d x} \\ & =\frac{d}{d x} \log [f(y(x))] \end{aligned} \]
which follows from equation (12). So, in summary, if y(x) (Fy−1)(x), where Fy(Y) is a continuous CDF, and (x) is the standard normal CDF, then either of the following equivalent conditions imply the convexity of y(x), for all x:
Before investigating the convexity of the mappings y(x) = (Fy−1 ∘ Φ)(x), when F is a common loss distribution, we pause to consider the interpretation of the preceding results. We focus on the
condition, since it appears to be more amenable to interpretation. By integrating from x = 0 to x = x′, and recalling that y(x) sends the percentiles of φ(x) to the corresponding percentiles of f(y), and in particular, that y(0) is the median of f(y), the condition implies:\[ f_{y}\left(y\left(x^{\prime}\right)\right) \leq f_{y}\left(y_{m}^{f}\right) e^{-\frac{x^{\prime 2}}{2}} \quad \text { for all } x^{\prime}>0 \tag{14} \]
Similarly,
\[ f_{y}\left(y\left(x^{\prime}\right)\right) \geq f_{y}\left(y_{m}^{f}\right) e^{-\frac{x^{\prime 2}}{2}} \quad \text { for all } x^{\prime}<0 \tag{15} \]
where yfm denotes the median of f(y). The first condition shows that fy(y(x)), as a function of x, decays, at least, proportionally to
as x grows large. Regarding the condition, corresponding to the left-hand tail, we consider the case when F has positive support. The condition given in equation (15) is surely satisfied if f(y) is monotonically decreasing, as is the case for the exponential distribution. However, even if f(y) is uni-modal, with f(0) = 0, as is the case for the Gamma distribution with scale parameter α > 1, the condition requires that, if y(x) = c 0 (c > 0), then the corresponding x = y−1(c) 0 is sufficiently small that fy(yfm) is, nonetheless, less than fy(c).5. Transmutation mappings
In the previous section it was seen that convexity of the mappings;
ensures that the linear approximation to copula regression will underestimate exact copula regression. In fact, the investigation of the mappings dates back at least to 1937 when Cornish-Fisher (C-F) formed polynomial expansions to approximate the quantiles of a given non-normal distribution function, in terms of the quantiles of the standard normal distribution. More specifically, if is a with distribution and is the quantile of the standard normal distribution, then the quantile of can be approximated by:\[ \begin{aligned} Y_{\alpha}= & m+\sigma\left(X_{\alpha}+\frac{1}{6} \frac{\kappa_{3}}{\sigma_{3}}\left(X_{\alpha}^{2}-1\right)+\frac{1}{24} \frac{\kappa_{4}}{\sigma_{4}}\left(X_{\alpha}^{3}-3 X_{\alpha}\right)\right. \\ & \left.-\frac{1}{36}\left(\frac{\kappa_{3}}{\sigma_{3}}\right)^{2}\left(2 X_{\alpha}^{3}-5 X_{\alpha}\right)+\cdots\right) \end{aligned} \tag{16} \]
where (2007), and later by Steinbrecher and Shaw (2008), Shaw and Brickman (2010), and Munir and Shaw (2012). Shaw and Buckley (2007) dubbed the mappings transmutation mappings, and noted that these mappings, essentially, turn samples from one distribution into samples from another Hence, a practical and precise representation of transmutation mappings has the potential to produce very efficient sampling algorithms, via the leveraging of existing samples to create samples from the desired distribution(s). In addition to having obvious application to sampling theory, especially copula-based sampling, Shaw and Buckley also note that transmutation mappings have utility within hypercubefilling quasi-Monte-Carlo (QMC) methods.
and denote the mean, standard deviation, and order cumulant of the It is important to note that C-F expansions are only approximations. Moreover, there are problems with C-F expansions, including the fact that introduction of additional terms does not always lead to a more accurate approximation. In fact, the likelihood of negative density values increases as higher-order terms are added to the series. As a result, C-F expansions are not useful for investigating the convexity of However, during the 2013 Actuarial Research Conference it was pointed out by Vytaras Brazauskas, from the University of Wisconsin at Milwaukee, that the mappings were recently studied by Shaw and BuckleyMore relevant to our purposes, Shaw and Buckley claim that, prior to their research, there had been few, or no, analytical results published on transmutation maps, outside the asymptotic domain. Shaw and Buckley ended this drought through an investigation of the differential equations that transmutation mappings obey. They then used these results to form new transmutation map expansions, based on the power series solutions to these non-linear ODEs. Though the results of Shaw and Buckley represent a step forward in the analytical analysis of transmutation mappings, and are based on differential equations, their results are not, however, sufficiently strong enough to quantify properties of transmutation mappings, such as convexity. In fact, to the authors knowledge, there has been no analytical investigation into the higher-order mathematical properties, such as convexity, of transmutation mapping. This statement is consistent with those of Shaw and Buckley, who go on to, quite rightly, point out that the dearth of such research is likely due to the fact that many common statistical distributions, and their inverses, including the ubiquitous normal distribution, involve rather intractable mathematical special functions. As a result, an attempt to investigate the convexity of transmutation mappings is tantamount, in many cases, to proving results about the composition of two mathematical special functions, both of which may be very intractable, on their own. This is especially the case when one of the distributions is the Gamma distribution, or the regularized incomplete gamma, as it is known within mathematical physics. Regardless, the final section of this paper contains a proof of the convexity of this particular transmutation mapping.
6. Results under specific loss distributions
We now investigate the convexity of transmutation mappings when the distribution F is among those commonly used for severity, or size-of-loss, modeling. The authors suspect that the results of this section will be useful, even outside the statistical community. In particular, we feel it is likely that researchers in finance and even applied mathematics will find interest in these results. We first present an example where convexity of the transmutation mapping, almost trivially, follows. Since each distribution F under consideration is a loss, or severity, distribution, each distribution F has support on the positive reals.
Proposition 6.1
(lognormal distribution) If Fy(y) is a lognormal distribution, with parameters (μ,σ), and Φ(x) (φ(x)) is the CDF (density) of the standard normal distribution, then y(x) = (Fy−1 ∘ Φ)(x) is convex, for all x.
Proof. Since (Fy−1 ∘ Φ)(x), and its inverse (Φ−1 ∘ Fy)(y) are both increasing functions, the convexity of (Fy−1 ∘ Φ)(x) is equivalent to the concavity of (Φ−1 ∘ Fy)(y). Hence, it suffices to show that (Φ−1 ∘ Fy)(y) is concave. Since Y ∼ LN(μ, σ), we have that:
\[ F_{y}(y)=\Phi\left(\frac{\ln (y)-\mu}{\sigma}\right) \]
Hence: (Φ−1 ∘ Fy)(y) =
and so, for all y:\[ \frac{d^{2}}{d y^{2}}\left(\Phi^{-1} \circ F_{y}\right)(y)=-\frac{1}{\sigma y^{2}} \leq 0 \tag{17} \]
Next, we consider another distribution that is commonly used for severity, or loss-size, modeling: the Pareto distribution. In particular, we consider the Pareto Type II distribution, or Lomax distribution, to which it is sometimes referred, with CDF parameterized as follows, for y > 0, and α, θ > 0:
\[ F(y)=1-\left(\frac{\theta}{y+\theta}\right)^{\alpha} \tag{18} \]
Unlike the lognormal case, convexity of the transmutation mapping is not as easily verified when F is a Pareto distribution.
Proposition 6.2
(Pareto distribution) If
is a Pareto distribution, with shape parameter and scale parameter and is the CDF (density) of the standard normal distribution, then is convex, for allProof. First, consider the case where x 0. Note that, since φ(x) is the normal density, the sufficient condition for convexity in Corollary 4.4, namely, equation (13), can be written: −x ≥
log[f(y(x))], or:\[ -x-\frac{d}{d x} \log [f(y(x))] \geq 0 \tag{19} \]
However, since f(y) is monotonically decreasing, for all y, and y(x), and log(x) are both increasing, we have that:
log[f(y(x))] 0, for all y, and hence equation (19) is satisfied, for x 0. Now, consider the case where x > 0. Since Fy−1(τ) = we have that:\[ y(x)+\theta=F_{y}^{-1}(\Phi(x))+\theta=\theta(1-\Phi(x))^{-\frac{1}{\alpha}} \tag{20} \]
so we have that
Therefore, after some routine computation and simplification, and using equation (20) for the last equality, we have:\[ \begin{aligned} -\frac{d}{d x} \log \left[f_{y}(y(x))\right] & =\frac{\alpha+1}{y(x)+\theta} \cdot \frac{\phi(x)}{f_{y}(y(x))} \\ & =\frac{\alpha+1}{\alpha} \phi(x)[1-\Phi(x)]^{-1} \end{aligned} \tag{21} \]
Again, by equation (19), we only need to show that: x ≤ −
log[f(y(x))], for x > 0, which by equation (21) is equivalent to: x ≤ φ(x)[1 − Φ(x)]−1, or for x > 0:\[ \frac{\alpha+1}{\alpha} \cdot \frac{1}{x} \geq \frac{1-\Phi(x)}{\phi(x)} \tag{22} \]
But, we note that the quantity m(x) = (1 − Φ(x))/φ(x) is the well-known Mill’s Ratio for the normal distribution, which has been the focus of a good amount of research. In particular, Baricz (2008, 2010, 2012) points out that in 1941 R. D. Gordon proved that the Mill’s ratio, for the normal distribution obeys the following inequalities, for x > 0:
\[ \frac{x}{x^{2}+1}<\frac{1-\Phi(x)}{\phi(x)}<\frac{1}{x} \tag{23} \]
Hence, since
> 1 we have, for x > 0, and for all α > 0, that:\[ \frac{\alpha+1}{\alpha} \cdot \frac{1}{x} \geq \frac{1}{x} \geq \frac{1-\Phi(x)}{\phi(x)} \tag{24} \]
We now turn our attention to the Gamma distribution, or the regularized incomplete gamma function, as it is known within the applied mathematics and physics communities. The incomplete gamma function appears ubiquitously within applied mathematics, and is related to many other mathematical special functions, including the Confluent Hypergeometric functions, Bessel functions Tricomi (1950a, 1950b), the Legendre and Laguerre polynomials, Kummer’s functions, the Gaussian error function, as well as the exponential integrals. For this reason, the theory of the incomplete gamma function, has been (and continues to be) of keen interest to many mathematicians. The Italian mathematician Tricomi is one such example. In fact, in the paper The Incomplete Gamma Function Since Tricomi, Gautschi (1998) remarks that “the incomplete gamma function held a special fascination for him (Tricomi), as he was fond of calling it, affectionately, the Cinderella of special functions.” More relevantly, Gautshi goes on to state that “Monotonicity, convexity, and higher monotonicity results abound for the gamma function, but seem to be scarce for the incomplete gamma function.” Moreover, despite the author’s assiduous attempt to find inequality results which are tight enough across a sufficiently broad range of the domain, to facilitate a direct proof of the convexity of the incomplete gamma function composed with the inverse error function, the breadth of published inequalities was, once again, found to be lacking. Some relevant investigations of the regularized incomplete gamma, and other special functions, include Alm (2003), Alzer (1997, 2005), Berg and Pedersen (2008), Carlitz (1963), Cerone and Dragomir (2008), Gautschi (1998), Short (2013), and Strecok (1968).
The authors did have some success using various approaches, including the use of approximations to (Φ−1 ∘ Fy)(y) involving inverse hyperbolic trigonometric functions, but this necessitated the breaking-up of the domain, and additional work would have been necessary to make the result rigorous over the full domain, Y. More importantly, this approach required a constructive argument, which was quite protracted. As a result, the focus of the authors returned to obtaining the result through contradiction. Finally, due to an éclair de génie, experienced by the third author, a surprisingly concise and elegant version of the proof was made possible. To the author’s knowledge, this analytical result is unique within the study of special functions.
Proposition 6.3
(Gamma distribution) If
is a Gamma distribution, with shape parameter and scale parameter and resp. is the CDF (resp. density) of the standard normal distribution, then is convex, for all and for all values of the shape parameterDue to the complexity of the proof of Proposition 6.3, especially when α > 1, the proof is relegated to the appendix, and is proceeded by several supporting lemmas.
7. Conclusion
Parsa and Klugman (2011) proposed a generalization of ordinary least squares regression which is better suited to the modeling of actuarial data sets, which often possess heavy-tailed marginal distributions and non-linear relationships between variables. However, not long after the introduction of copula regression, a surprisingly simple approximation was suggested. In this paper we have presented and investigated this approximation, which takes the form of an OLS regression, under particular transformations of the variables. Next, we described how this linear approximation to copula regression can produce estimates which are close to those from exact copula regression. However, we present a set of criteria which guarantee that the estimates from the linear approximation will underestimate those from exact copula regression. Further, it is described how the main driver of the discrepancy, or bias, between copula regression and its linear approximation is a consequence of the use of transformations of the variables. Moreover, we explain how this discrepancy is not due to the form of the transformations used within the linear approximation to copula regression, but rather such a discrepancy will likely arise whenever transformations of the variables are used within OLS regression. This realization has consequences well beyond the specific models investigated in this paper, and serves as a salient reminder of the dangers of using transformations of variables within OLS regression. Finally, armed with the aforementioned sufficient conditions for underestimation, we continued by investigating which of the common loss distributions satisfy these criteria. In particular, we were able to prove that the lognormal, Pareto, and gamma distributions all satisfy these criteria and in particular, satisfy these criteria for all parameter values. Hence, these results allow the partitioner to determine when the estimates from the linear approximation to copula regression will underestimate the true values. Further, if OLS regression involving transformations is used within the reserving, capital modeling, or even pricing processes of a firm, these results can aid the practitioner in avoiding the understatement of reserves and even insolvency.
Acknowledgments
This work was sponsored by the Casualty Actuarial Society (CAS), the Actuarial Foundation’s research committee, and the Committee on Knowledge Extension Research (CKER) of the Society of Actuaries (SOA). In addition to the CAS, the Actuarial Foundation, and the CKER, the authors wish to extend thanks to Alice Underwood, Vice-President, Research and Development, CAS; David Core, Director of Professional Education and Research, CAS; Curtis Huntington, FSA, MAAA, FCA, MSPA (1942–2013) Former Chair, the Actuarial Foundation’s Research Committee, the CEKR of the SOA, and the Actuarial Foundation of Canada; Cynthia MacDonald, Senior Experience Studies Actuary, SOA; and Eileen Streu, Executive Director, The Actuarial Foundation.
Also, the authors wish to extend special thanks to the following individuals who provided guidance, feedback, and motivations for this research: Stuart Klugman, Staff Fellow, Education, SOA; and Thomas Struppeck, Longhorn Analytics LLC, and The University of Texas at Austin.