Chapter 4 Transferring Multiple Risks including Reinsurance
Chapter Preview. This chapter begins our study of multivariate risks. As emphasized in Chapter 1, any risk owner, be it a person, firm, or insurer, faces a multitude of interrelated risks related to the risk owner itself, its property, as well as its obligations and responsibilities.
In the historical development of risk retention, the focus has largely been on either individual consumers or insurers as the risk owners. When an insurer transfers risk to another insurer, it is referred to as reinsurance. Reinsurance not only holds historical significance but also serves as a natural context to introduce significant risk retention agreements. As demonstrated in Section 1.1, reinsurance continues to be an active commercial activity and so Section 4.1 delves deeper into its specific agreements.
With several risks, one cannot disregard their interrelationships. Section 4.2 begins our investigations into dependent risks; dependency impacts diversification, a critical feature when constructing portfolios. This section introduces copulas as tools to account for interdependencies among risks. The subsequent Section 4.3 illustrates their application in various reinsurance agreements.
As modeling becomes more complex, the special case of independence, while not always realistic, serves as a valuable benchmark for comparison. Similarly, in this chapter, the variance operator is used to gauge the uncertainty of risks in a multivariate context, offering interpretability and serving as a benchmark. However, we have already seen limitations of the variance operator and in subsequent chapters will employ quantile-based risk measures. This means that calculations will become more complex and so Section 4.4 introduces numerical tools that can be utilized to solve constrained optimization problems.
Another intriguing aspect of reinsurance is that it naturally suggests the introduction of more than two parties taking responsibility for a risk. Even in a simple setting, a firm might have a contract with an insurer to take on a portion of its risks, and this insurer has a separate contract with another insurer (a reinsurer) that may include ceding a portion of these risks. Thus, there is a sharing of the risks among three parties. In more complex settings, it is common to have more parties involved. Section 4.5 concludes the chapter with a comprehensive discussion of risk sharing among multiple parties known as a risk exchange.
4.1 Reinsurance
As noted above, reinsurance is simply insurance purchased by an insurer. Insurance purchased by non-insurers is sometimes known as primary insurance to distinguish it from reinsurance. Reinsurance differs from personal insurance purchased by individuals, such as auto and homeowners insurance, in contract flexibility. Like insurance purchased by major corporations, reinsurance programs are generally tailored closely to the needs of the buyer. In contrast, in personal insurance buyers typically cannot negotiate on the contract terms although they may have a variety of different options (coverages) from which to choose.
The two broad types of agreements are proportional and non-proportional reinsurance. A proportional reinsurance contract is an agreement between a reinsurer and a ceding company (also known as the reinsured) in which the reinsurer assumes a given percent of losses and premiums. Non-proportional agreements are simply everything else. As examples of non-proportional agreements, this chapter looks first to stop-loss and excess of loss contracts and then to more complex arrangements.
A reinsurance contract can also be categorized as facultative or as a treaty. Facultative reinsurance contracts typically cover a single, large complex risk that cannot be readily grouped with other risks. In contrast, a treaty is a reinsurance contract that covers more than a specific risk, typically a specified share of all the insurance policies issued by the ceding company which come within the scope of the contract.
4.1.1 Proportional Reinsurance
The simplest example of a proportional treaty is called quota share. In a quota share reinsurance treaty, the reinsurer receives a flat percent, say 50%, of the premium for the book of business reinsured. In exchange, the reinsurer pays the same flat percent of losses, including allocated loss adjustment expenses.
For loss sharing, this is exactly the same as the risk retention agreement with a coinsurance parameter \(c\) introduced in Section 2.2 where the insurer cedes a proportion \(1-c\) of losses and retains the fraction \(c\). However, in a quota share treaty, the premium is also specified as part of the treaty. Moreover, the insurer, or ceding company, has already incurred expenses for acquiring the business. Therefore, it is customary for the reinsurer to also pay the ceding company a ceding commission which is designed to reflect the differences in underwriting expenses incurred.
Quota share reinsurance treaties are particularly suitable for growing insurers or established companies which are new to a certain class of business. Quota share also makes sense for primary insurers who are seeking capital relief in light of solvency considerations or protection against random fluctuations across an entire portfolio. Because of their automatic nature, quota share treaties work well with high frequency lines of business. From the point of view of the reinsurer, quota share is desirable because the interests of the insurer and reinsurer are aligned and so there is little opportunity for moral hazard issues (such as sloppy claims handling).
From a theoretical viewpoint, we have already seen in Section 3.1 (Proposition 3.3) the desirability of the quota share contract for the reinsurer. There, we learned that the quota share reinsurance treaty minimizes the reinsurer’s uncertainty, measured by the variance, when subject to some mild constraints on the insurer’s variance of retained losses. Intuitively, quota share insurance is appealing because the insurer shares the responsibility for large losses in the tail of the distribution. As will be seen, this is in contrast to non-proportional agreements where reinsurers take responsibility for large losses.
4.1.2 Non-Proportional Reinsurance
The simplest example of a non-proportional treaty is called a stop-loss treaty. This acts as the same as the upper limit risk retention agreement introduced in Section 2.2. That is, the insurer sets a retention level \(u\) and pays in full total losses for which \(X \le u\). Further, for total losses for which \(X > u\), the direct insurer pays \(u\) and the reinsurer has responsibility for the remaining amount \(X-u\). Thus, the insurer retains an amount \(u\) of the risk. Summarizing, the amounts paid by the direct insurer and the reinsurer are \(Y_{retained} = X \wedge u\) and \(Y_{reinsurer} = \max(0,X-u) = (X-u)_+\). As before, note that \(Y_{retained}+Y_{reinsurer}=X\). From a reinsurer’s perspective, note that \(u\) could be considered a deductible. However, this book adopts the risk owner’s perspective and so \(u\) represents the upper limit of its obligation.
Stop-loss, in its pure form, is not employed extensively in practice. This is likely due to its potential moral hazard weakness. With a pure stop-loss treaty, insurers have few incentives to carefully underwrite and monitor policies, as well as settle associated losses that contribute to experience above the stop-loss limit. Recall from basic economics that the phrase moral hazard is used to describe the situation where a contract alters the behavior of parties to the agreement. This would certainly be a concern to reinsurers that offer a stop-loss coverage. An exception to this is coverage between connected entities, such as a firm and its captive insurer, where incentives are aligned (see Section 6.2.2).
Nonetheless, at least conceptually, the stop-loss type of contract is particularly desirable for the insurer. As we learned in Section 3.1 (Proposition 3.4), the stop-loss reinsurance treaty minimizes the insurer’s uncertainty as measured by the variance subject to a mild budget constraint. Intuitively, with stop-loss insurance, the reinsurer takes responsibility for very large losses in the tail of the distribution, not the insurer.
Excess of Loss
A closely related form of non-proportional reinsurance is the excess of loss coverage. Under this contract, the total risk can be thought of as composed as \(p\) separate risks \(X_1, \ldots, X_p\) where each risk is subject to an upper limit, say, \(u_i\). So the insurer retains \[ Y_{retained} = \sum_{i=1}^p X_i \wedge u_i = S(u_1, \ldots, u_p), \] and the reinsurer is responsible for the excess, \(Y_{reinsurer}= S - Y_{retained}\). (Recall that the portfolio sum is \(S= X_1 + \cdots + X_p\).) The retention limits may vary by risk or may be the same for all risks, that is, \(u_i =u\) for all \(i\). The development in this book focuses on this contract and so I introduce the notation \(S(u_1, \ldots, u_p)\) to represent the sum of losses that are limited by the retention parameters \(u_1, \ldots, u_p\).
In addition to arguments from theory, there are important practical considerations that recommend the use of non-proportional insurance to an insurer. Non-proportional insurance can limit an insurer’s liability with upper limits that reflect its willingness and capacity to bear risk. Smaller losses are the sole responsibility of the insurer and are not shared with the reinsurer, suggesting an insurer may focus its expertise on this portion of the portfolio and thus enhancing its earnings potential. In particular, administration costs can be lower for both parties as the insurer no longer needs to calculate the proportion of a loss for each risk as under surplus share reinsurance.
Optimal Choice for Excess of Loss Retention Limits
What are the best choices of the excess of loss retention limits \(u_i\)? One way to formalize this question is to find those values of \(u_i\) that minimize \(\mathrm{Var}(Y_{retained})\) subject to the constraint that \(\mathrm{E}(Y_{reinsurer}) = RTC_{max}\). Subject to this budget constraint, the insurer wishes to minimize the uncertainty of the retained risks (as measured by the variance).
As before, only equality constraints are invoked and so a solution can be derived using the Lagrange method. Specifically, define the Lagrangian \[ \begin{array}{ll} LA &= \mathrm{Var}(Y_{retained}) + LME \{\mathrm{E}(Y_{reinsurer}) - RTC_{max} \} \\ &= \sum_{i=1}^p ~\mathrm{Var}(X_i \wedge u_i) + LME \{ [\sum_{i=1}^p ~\mathrm{E}(X_i)-\mathrm{E}(X_i \wedge u_i)]- RTC_{max}\}. \end{array} \]
\(Under~the~Hood.\) Show the Optimal Excess of Loss Limits
Optimizing the function \(LA\) over choices of upper limits provides insights into properties of the optimal retention limits \(u_i^*\). In particular, it turns out that the optimal retention limit less the insurer’s expected retained losses, \(u_i^* - E[X_i \wedge u_i^*]\), is the same for all risks. Gray and Pitts (2012) (Section 5.8) attribute the discovery of this result to Bruno de Finetti (1906-1985). Section 11.3.1 will provide extensions to additional contracts.
Example 4.1.3. Excess of Loss for Three Pareto Risks. Consider three risks that have a Pareto distribution, each having a different set of parameters (so they are independent but non-identical). Use the same set of parameters as in Example 4.1.1.
By minimizing the Lagrangian, one can numerically determine the optimal retention limits \(u_1^*\), \(u_2^*\), and \(u_3^*\). This is done in the associated R
code using the package alabama
. Results from the numerical optimization show that each optimal retention limit minus expected insurer’s losses, \(u_i^* - E[X_i \wedge u_i^*]\), is the same for all risks, as derived theoretically. In addition, one can graphically compare the distribution of total risks to that retained by the insurer and by the reinsurer, as displayed in Figure 4.3. This graph summarizes the distributions of possible outcomes that result from the optimal retention policy.
Show Example 4.1.3 R Code
4.1.3 Layers of Coverage
Typically excess of loss contracts limit the exposure for a reinsurer, as illustrated in the following example.
Example 4.1.4. Consider a specific risk of an insurer. For this risk, the reinsurer is responsible for losses in excess of 500 up to a limit of 100,000, that is, \(Y_{reinsurer} =\min[(X-500)_+, 100000]\). The insurer retains the remaining risk, \(Y_{retained} = X - Y_{reinsurer}\).
From the reinsurer’s point of view, in the language of Section 2.2 one would think of 500 as a “deductible” and 100,000 as its “upper limit” of coverage.
In reinsurance, it is common to partition loss amounts into multiple layers. This partitioning of losses is easiest to think about when there are more than two parties responsible for a risk, a concept further developed in Section 4.5. For example, instead of simply an insurer and reinsurer or an insurer and a policyholder, think about the situation with all three parties, a policyholder, insurer, and reinsurer, who agree on how to share a risk. More generally, consider \(k\) parties. If \(k=3\), it could be an insurer and two different reinsurers.
Example 4.1.5. Layers of Coverage for Three Parties.
- Suppose that there are \(k=3\) parties. The first party is responsible for the first 100 of losses, the second responsible for losses from 100 to 3000, and the third responsible for losses above 3000.
- If there are four losses in the amounts 50, 600, 1800 and 4000, then they would be allocated to the parties as follows: \[ {\small \begin{array}{l|rrrrrrr} \hline \text{Layer}& \text{Claim 1}& \text{Claim 2}& \text{Claim 3}& \text{Claim 4}& \text{Total} \\ \hline (0, 100]& 50& 100& 100 & 100 & 350\\ (100, 3000]& 0& 500 & 1700 & 2900& 5100 \\ (3000, \infty)& 0 & 0 & 0 & 1000 & 1000\\ \hline \text{Total} & 50 & 600& 1800& 4000& 6450 \\\hline \end{array} } \]
To handle the general situation with \(k\) groups, partition the positive real line into \(k\) intervals using the cut-points \[ 0 = u_0 < u_1 < \cdots < u_{k-1} < u_k = \infty. \] Note that the \(i\)th interval is \((u_{i-1}, u_i]\). Now let \(Y_i\) be the amount of risk shared by the \(i\)th party. To illustrate, if a loss \(X\) is such that \(u_{i-1} <X \le u_i\), then \[ \left(\begin{array}{c} Y_1\\ Y_2 \\ \vdots \\ Y_i \\Y_{i+1} \\ \vdots \\Y_k \end{array}\right) =\left(\begin{array}{c} u_1-u_0 \\ u_2-u_1 \\ \vdots \\ X-u_{i-1} \\ 0 \\ \vdots \\0 \end{array}\right) . \] More succinctly, we have \[ Y_i = \min(X,u_i) - \min(X,u_{i-1}) . \] With this expression, one can see that the \(i\)th party is responsible for losses in the interval \((u_{i-1}, u_i].\) With this, it is easy to check that \(X = Y_1 + Y_2 + \cdots + Y_k.\) To illustrate, applying this structure to Example 4.1.4, we see that the primary insurer has responsibility for the first and third layers (\(Y_1\) and \(Y_3\)) whereas the reinsurer has responsibility for the second layer (\(Y_2\)). Thus, the parties taking on different layers need not be different.
For an example used in practice, see the Wisconsin Property Fund site for an illustration of layers of reinsurance.
See Appendix Section 4.6.4 for an overview of a mathematical framework that provides conditions under which a layered coverage scheme is optimal.
4.1.4 Complexities of Reinsurance Coverage
Historically, reinsurance agreements were contracts negotiated based on individual risks. These contracts, done on a case-by-case basis, today are known as facultative reinsurance and continue to represent about 15% of the marketplace (in terms of premium volume), cf. Albrecher, Beirlant, and Teugels (2017). The remainder of the market has an obligatory treaty form, where both parties are bound by an obligation to cede and assume a contractually agreed share of a risk portfolio.
Because of their experience in risk modeling, reinsurers are key drivers in the adoption of better risk management practices by many organizations. Given their technical expertise and the extensive marketplace for unique contract arrangements, it is not surprising that reinsurance contracts can easily become quite complicated. This section describes a few of the more popular contracts.
To illustrate, some excess of loss contracts have in addition an (annual) aggregate deductible \(AAD\) so that the reinsurer’s obligation may be written as \[ Y_{reinsurer} =\left[\left( \sum_{i=1}^p \min[(X_i-u_{1,i})_+, u_{2,i}]\right)- AAD\right]_+ . \] Here, \(u_{1,i}\) represents a reinsurer’s deductible and \(u_{2,i}\) represents is upper limit for the \(i\)th risk. Deductibles and upper limits may vary by risk or be the same. In the same way, excess of loss contracts may have an (annual) aggregate limit \(AAL\), so that \[ Y_{reinsurer} = \min \left( \sum_{i=1}^p \min[(X_i-u_{1,i})_+, u_{2,i}], AAL \right) . \] Naturally, a contract may have both an overall deductible and an upper limit. Albrecher, Beirlant, and Teugels (2017) refer to these as types of global protections.
As another variation, a reinsurance cover may be defined on a per event, or per occurrence of a claim, basis in lieu of the per risk basis that has been described so far. This approach is very handy to protect against, for example, natural catastrophes such as earthquakes or hurricanes where there may be many losses due to a common event. For example, a reinsurer’s obligation may be of the form \[ Y_{reinsurer} = \sum_{j \in \text{event}} \min[(X_j-u_{1,j})_+, u_{2,j}] . \] When combining several reinsurance features into a single contract, proportional arrangements such as quota share arrangements are easy to include. As an example, it is common for an insurer to have simultaneous quota share contracts on the same portfolio with different reinsurers. Many other variations are routinely used in practice.
Video: Section Summary
4.2 Modeling Dependence with Copulas
4.2.1 Copulas Defined
Copulas are widely used in insurance and many other fields to model the dependence among multivariate outcomes. A copula is a multivariate distribution function with uniform marginals. Specifically, let \(\{V_1, \ldots, V_p\}\) be \(p\) uniform random variables on \((0,1)\). Their joint distribution function \[ C(v_1, \ldots, v_p) = \Pr(V_1 \leq v_1, \ldots, V_p \leq v_p) \] is a copula. Of course, applications are based on more than just uniformly distributed data. Thus, consider arbitrary marginal distribution functions \(F_1(\cdot)\),…,\(F_p(\cdot)\). Then, one can define a multivariate distribution function \(F\) using the copula such that \[\begin{equation} F(y_1, \ldots, y_p)= C[F_1(y_1), \ldots, F_p(y_p)]. \tag{4.3} \end{equation}\] Sklar (1959) showed the converse, that is, any multivariate distribution function can be written in the form of equation (4.3), that is, using a copula representation. Note on Notation: We use \(V\) and \(v\) for uniform random variables and their realizations in order to distinguish them from other terms, including utility functions \(U\) and upper limits \(u\).
As an example with \(p=2\) variables, let us look to the copula due to Frank (1979), \[\begin{equation} C_F(v_1,v_2) = \frac{1}{\gamma} \log \left( 1+ \frac{ (\exp(\gamma v_1) -1)(\exp(\gamma v_2) -1)} {\exp(\gamma) -1} \right). \tag{4.4} \end{equation}\] This is a bivariate distribution function with its domain on the unit square \([0,1]^2.\) It is easy to see that \(C_F(v_1,1) = v_1\). From this, one can argue that the marginals have a uniform (on \([0,1]\)) distribution and so \(C_F(\cdot,\cdot)\) is a copula. Further, one can interpret \(\gamma\) to be a dependence parameter in that the extent of dependence is controlled by the parameter \(\gamma\).
The copula due to Frank (1979) has a simple expression in equation (4.4) (at least simple relative to other copulas) and enjoys several desirable properties. However, it does not readily extend to multiple dimensions (a point of emphasis in this book). So, as an alternative, let us also introduce the normal, or Gaussian, copula. In two dimensions, this is given as
\[\begin{equation}
C_N(v_1,v_2) = \Phi_2\left[\Phi^{-1}(v_1), \Phi^{-1}(v_2) \right].
\tag{4.5}
\end{equation}\]
The expression in equation (4.5) appears more forbidding than Frank’s copula but is just as easy to evaluate. The symbol \(\Phi(\cdot)\) represents the cumulative standard normal distribution and \(\Phi^{-1}(\cdot)\) is the corresponding inverse function, or quantile. In R
, these can be evaluated using the functions pnorm
and qnorm
, respectively. The symbol \(\Phi_2(\cdot)\) represents a bivariate normal distribution function that, in R
, can be evaluated using the function mvtnorm
.
Because both the Frank and normal copulas are used extensively in applications, it is not surprising that they are both available as R
functions (e.g., in the package copula
, see examples that follow). So, users do not really need the mathematical expressions in equations (4.4) and (4.5). Nonetheless, it is helpful to have some familiarity with the mathematical underpinnings of the technique.
One important strength of the copula approach is evident from equation (4.3). We will be able to work with multiple risks, perhaps the first one being gamma distribution, the second having a Pareto distribution, the third having a normal distribution. Even though each marginal distribution may be very different from one another, one can link (or couple) them using a copula.
4.2.2 Copulas, Risk Retention, and Simulation
Copulas are well known to work in many arenas of scientific applications but how do they work in risk retention problems? To get some insights, consider the following example.
Example 4.2.1. Standard Deviation of an Excess of Loss Policy with Two Risks. Consider two risks:
- That \(X_1\) has a gamma distribution with mean 10000 and scale 5,000. The 95th percentile of \(X_1\) is 23,719.
- That \(X_2\) has a Pareto distribution with mean 1000 and scale 2,000. The 95th percentile of \(X_2\) is 3,429.
Their relationship is governed by a normal copula with dependence parameter \(\rho_S = 0.30\).
The risk owner retains \(S(u_1,u_2) = X_1 \wedge u_1 + X_2 \wedge u_2\) and calibrates uncertainty through
\[
\begin{array}{ll}
\mathrm{Var}[S(u_1,u_2)] & = \mathrm{E}[S(u_1,u_2)]^2 - [\mathrm{E} S(u_1,u_2)]^2\\
& = \mathrm{E}[(X_1 \wedge u_1)^2]+\mathrm{E}[(X_2 \wedge u_2)^2]+2\mathrm{E}[(X_1 \wedge u_1)(X_2 \wedge u_2)]\\
& \ \ \ \ \ - [\mathrm{E}(X_1 \wedge u_1)+\mathrm{E}(X_2 \wedge u_2)]^2 .\\
\end{array}
\]
As in Section 4.1.2, we utilize the relationships
\[
\mathrm{E}(X \wedge u) = \int_0^u ~[1- F(x)]dx
~~~~ \text{and} ~~~~
\mathrm{E}(X \wedge u)^2 = 2\int_0^u ~x[1- F(x)]dx.
\]
These functions are easy to compute and, for example, are readily available in R
through the package actuar
. One can evaluate the expected cross-product of limited variables using
\[\begin{equation}
\begin{array}{rl}
\mathrm{E} [(X_1 \wedge u_1)(X_2 \wedge u_2)]
& = \int_0^{u_2} \int_0^{u_1} ~\Pr(X_1 >x_1, X_2> x_2) ~dx_1 dx_2 \\
& = \int_0^{u_2} \int_0^{u_1} ~\left[1 -F_1(x_1) - F_2(x_2) +F(x_1,x_2) \right]~dx_1 dx_2 .\\
\end{array}
\tag{4.6}
\end{equation}\]
\(Under~the~Hood.\) Check the Limited Cross Product Expectation Equation (4.6)
Although the double integral in equation (4.6) can be readily evaluated for a single point \((u_1,u_2)\), this becomes burdensome as we go through many iterations seeking optimal values through iterative constrained optimization procedures beginning in the next section. (Chapter 5 will demonstrate this complexity.) A simpler approach is the following:
- Using simulation, create a random realization of the bivariate uniform random variates \((V_{1r}, V_{2r})\). In the
R
packagecopula
, one can do this with the functionrCopula
. - Use the inverse of the distribution function to create random realizations of the random variables, \(X_{1r} = F_1^{-1}(V_{1r})\) and \(X_{2r} = F_2^{-1}(V_{2r})\).
- For \(F_1^{-1}\), in
R
use the functionqgamma
. - For \(F_2^{-1}\), in
R
use the functionqpareto
.
- For \(F_1^{-1}\), in
- Create a random realization of the limited sum, \(S_r(u_1,u_2) = X_{1r} \wedge u_1 + X_{2r} \wedge u_2\).
- Repeat this for a large number, \(R\), of replications. Summarize using the approximation \[ \widehat{Var}[S_R(u_1,u_2)] = \frac{1}{R} \sum_{r=1}^R \left[S_r(u_1,u_2)-\bar{S}_R(u_1,u_2)\right]^2, \] where \(\bar{S}_R\) is the average value of the replicates \(S_r(u_1,u_2)\).
R Code to Evaluate Variance
Figure 4.4 summarizes results. As anticipated, the uncertainty of retained risk increases as the upper limits increase.
R Code to Show the Figures
Video: Section Summary
4.3 Incorporating Dependence
Now return to the basic reinsurance treaties and let us see how they fare in the face of dependence. Let us start with a simple example to show how dependence affects risk measures. After this, one can then address the question as to how these optimal allocations change as the dependence changes. We develop this strategy first for quota share, then for excess of loss contracts.
4.3.1 Effects of Dependence on Risk Measures
It is not surprising that dependence affects our measures of uncertainty. In general, positive dependence leads to greater uncertainty, and hence larger values of risk measures. To be concrete, let us see how these general ideas work in a specific example.
Example 4.3.1. Effects of Dependence on Portfolio Summary Measures. Consider a portfolio management situation outlined in Exercise 4.1.1 that assumed that the property and liability risks are independent. Now, let us incorporate dependence by allowing the four lines of business to depend on one another through a Gaussian copula. For simplicity, assume a constant parameter equal 0.5 among the four lines.
Table 4.1 summarizes the results for the dependent portfolio in the lower three rows. For comparison, the top three rows provide results for the independence case. As anticipated, the mean losses are virtually identical when comparing the independent to dependent portfolios. The standard deviations (square root of our variance measure) for the dependent portfolio are larger for the reinsurer and in total. This larger uncertainty is also reflected in the values at risk at different levels of confidence.
Mean | Std Dev | \(VaR_{0.90}\) | \(VaR_{0.95}\) | \(VaR_{0.99}\) | |
---|---|---|---|---|---|
Retained (Ind) | 269 | 48 | 300 | 300 | 300 |
Reinsurer (Ind) | 2333 | 3907 | 4487 | 6269 | 12644 |
Total (Ind) | 2602 | 3909 | 4760 | 6541 | 12907 |
Retained (Dep) | 269 | 48 | 300 | 300 | 300 |
Reinsurer (Dep) | 2336 | 4067 | 4498 | 6283 | 12772 |
Total (Dep) | 2605 | 4068 | 4770 | 6556 | 13052 |
R Code for the Table
4.3.3 Excess of Loss
As in Section 4.1.2, for the excess of loss setting we consider \(p\) risks \(X_1, \ldots, X_p\) where each risk is subject to an upper limit, say, \(u_i\). With this, the insurer retains \(Y_{retained} =\) \(\sum_{i=1}^p X_i \wedge u_i\) and the reinsurer is responsible for the excess, \(Y_{reinsurer}=S - Y_{retained}\). The retention limits may vary by risk or may be the same for all risks, \(u_i =u\), for all \(i\).
What is the best choice of the excess of loss retention limits \(u_i\)? To formalize this question, we seek to find those values of \(u_i\) that minimize \(\mathrm{Var}(Y_{retained})\) subject to the constraint that \(\mathrm{E}(Y_{reinsurer}) = RTC_{max}\). That is, subject to this budget constraint, the insurer wishes to minimize the uncertainty of the retained risks (as measured by the variance).
Section 4.4 will outline several approaches for solving this problem numerically. Let us see how one might do this using the root finding methods that will be described in Section 4.4.4. In part, we explore this now simply to demonstrate the difficulties associated with this approach. To implement this approach, assume that the budget constraint is binding. Further, one needs derivatives. So, start with the Lagrangian \[ LA(u_1, \ldots, u_p, LME) = \mathrm{Var}(Y_{retained}) + LME ~ [\mathrm{E}(Y_{reinsurer}) - RTC_{max}]. \] It is convenient to introduce an integrated tail function \[ \begin{array}{ll} IT_{ij}(a,b) &=\int_0^{b} ~\left\{1 -F_i(a) - F_j(z) +F_{ij}(a,z) \right\}~ dz \\ &=\int_0^{b} ~\Pr(X_i > a ,X_j > z)~ dz ,\\ \end{array} \] where \(F_{ij}\) is the joint distribution function of the random variables \(X_i\) and \(X_j\). After some pleasant calculus, one can check that the partial derivatives are \[\begin{equation} \begin{array}{rl} \partial_{u_i} ~LA(u_1, \ldots, u_p, LME) &=2 \left[u_i - \sum_{j=1}^p ~\mathrm{E}(X_j \wedge u_j) - \frac{LME}{2} \right][1-F_i(u_i)] \\ &\ \ \ \ \ +2 \sum_{j \ne i} \left\{ IT_{ij}(u_i,u_j) \right\} , \\ \end{array} \tag{4.7} \end{equation}\] for \(i=1, \ldots, p\), and \[\begin{equation} \partial_{LME} ~ LA(u_1, \ldots, u_p, LME) = \mathrm{E}(S) - \sum_{i=1}^p ~\mathrm{E}(X_i \wedge u_i) - RTC_{max} . \tag{4.8} \end{equation}\]
\(Under~the~Hood.\) Verification of Equation (4.7)
Setting these derivatives to zero, there are \(p+1\) equations and \(p+1\) unknowns which can be solved numerically. Numerical solutions require starting values; solutions can sometimes be sensitive to the choice of starting values. Fortunately, for this problem, one can look to the case of independence developed in Section 4.1.2 for starting values.
Special Case - Independent Risks. It is convenient to first define an integrated distribution function \(H_{1i}(x) = \int_0^{x} F_i(z) dz\). With this notation, it can be shown we may determine \(LME^*\) as the solution of the equation \[ \mathrm{E}(S) - RTC_{max} = \sum_{i=1}^p \left\{ H_{1i}^{-1}\left(\frac{LME}{2}\right) - \frac{LME}{2} \right\} .\\ \] With a value of \(LME^*\), one can calculate the optimal retention limits using \(u_i^* = H_{1i}^{-1}(\frac{LME^*}{2})\).
\(Under~the~Hood.\) Independence Case Details
Example 4.3.3. Optimal Upper Limits with Excess of Loss for Three Dependent Risks. Now return to the same set of marginal distributions as in the quota share Example 4.3.2. Also as in this example, let \(RTC_{max} = 4,200\), which is 20% of the sum of expected losses. For the Gaussian copula parameters, we use \[ \boldsymbol \rho = \left( \begin{array}{ccc} 1 & \rho_{12} & \rho_{13}\\ \rho_{12} & 1& \rho_{23}\\ \rho_{13} & \rho_{23}& 1\\ \end{array} \right) \] with \(\rho_{12} =\) 0.95, \(\rho_{13} =\) 0, and \(\rho_{23} =\) 0. Note that the rho parameters are based on the copula, not (Pearson) correlations among the risks.
R Code For Excess of Loss Dependent Risks Calculations
Mean | Optimal Ind \(u^*\) | Ind \(u^*\) Prop | Optimal Dep \(u^*\) | Dep \(u^*\) Prop | |
---|---|---|---|---|---|
Gamma1 | 10000 | 11806 | 0.683 | 11938 | 0.689 |
Pareto | 1000 | 4775 | 0.974 | 1214 | 0.759 |
Gamma2 | 10000 | 11806 | 0.683 | 12673 | 0.72 |
LME | 7725 | 8942 |
Table 4.5 shows the optimal upper limits under both models of independence and dependence. Comparing these results, it appears that there can be important differences between the optimal limits based on the dependence.
In addition, another way to view these is to think about their position in the marginal distribution. Specifically, the optimal retention parameters can also be expressed as proportions for the corresponding marginal distribution functions. So, for example, for the best choice of \(u_1\) under the model of independence on the original scale is 11,806; this is the 68.3 percentile of this gamma distribution. Extending this logic to the second row, the upper limits are 97.4 percentile for the independence model versus 75.9 percentile for the dependence model. Thus, when viewed as a percentile or on the original scale, 4,775 versus 1,214, it appears that the dependence assumption can influence the optimal upper limits.
Video: Section Summary
4.4 Numerical Optimization
Recall the general constrained optimization problem in Display (3.9) from Section 3.3.4, \[ \boxed{ \begin{array}{cccr} {\small \text{minimize}_{{\bf z}}} & f_0({\bf z}) & \\ {\small \text{subject to}} & f_{con,j}({\bf z}) \le 0 & j \in CON_{in} & \ \ \ \ \ \ \ \ \ \ \\\ & f_{con,j}({\bf z}) = 0 & j \in CON_{eq} . \end{array} } \] Here, \(f_0(\cdot)\) is the objective function that we wish to minimize subject to conditions on the constraints \(f_{con,j}(\cdot)\) in the inequality (\(CON_{in}\)) and equality (\(CON_{eq}\)) sets.
4.4.1 Convex Optimization
In today’s computing environment, the optimization problem described in Display (3.9) is relatively straightforward if we can identify the problem as convex. There are some minor variations in the definition of convex optimization in the literature; I use the conditions put forward in Boyd and Vandenberghe (2004). Specifically, let us require that the objective function \(f_0({\bf z})\) and inequality constraints \(f_{con,j}({\bf z}), j \in CON_{in}\) are convex and that the equality constraints \(f_{con,j}({\bf z}), j \in CON_{eq}\) are affine. By affine (or linear), I mean that one can write \(f_{con,j}({\bf z}) = {\bf A} {\bf z} - {\bf b}\) for some matrix \({\bf A}\) and vector \({\bf b}\). (For a refresher, refer to the Appendix Section 3.6.3 for definitions of convexity and other terms.)
Classical problems based on proportional (e.g., quota sharing) rules with the variance as a measure of risk are typically convex. As an example, consider the quota sharing problem introduced in Section 4.1.1.1. The quota share coefficients are the parameters of interest and so \({\bf z} = \bf c\). The variance of the portfolio risk, \(f_0({\bf z}) = \mathrm{Var}({\bf z}^{\prime} \mathbf{X})\) \(= {\bf z}^{\prime} \boldsymbol \Sigma{\bf z}\), is the objective function to be minimized. This is quadratic in the parameters and hence typically convex. In the same way, the risk transfer condition can be seen to be linear in the parameters. Moreover, it will be straightforward to include parameter constraints of the form \(c_i = z_i \ge 0\) and \(c_i = z_i \le 1\) as these are linear and hence also convex. Recall that Section 4.3.2 described this problem in more detail.
For convex problems, there are many algorithms that are computationally efficient and that fare remarkably well even when the number of decision variables (the dimension of \({\bf z}\)) is large. To illustrate this in our demonstrations, we use the R
package CVXR
; see Exercise 4.5.5. Moreover, as emphasized by Boyd and Vandenberghe (2004), local solutions to convex problems are also global solutions, a property that is not generally available for constrained optimization problems. Because of this property, many convex algorithms do not even require starting values.
Special Case: Asset Allocations
Asset allocations, summarized in Section 3.4, can be cast as a convex optimization problem. In addition, mathematically the asset allocation problem is similar to the quota share problem described in Section 4.3.2. Naturally, the practices are very different; the former considers allocation concerns of investors whereas the latter is about diversifying liabilities that is typically only done by insurers and major companies. A common simplifying feature of both problems is that the decision variables, \(c_1, \ldots, c_p\), are linear in the risks. This means that the problem is convex in many situations. In this section, I expand the discussion and use the expected shortfall \(ES\) as our risk measure.
To see the convexity of the problem, the approach is to define an expanded version of the expected shortfall function, \[ \begin{array}{ll} ES1(z_0,\mathbf{c}) &= z_0 + \frac{1}{1-\alpha} \left\{\mathrm{E}[\mathbf{c}'\mathbf{X}] -\mathrm{E}[\mathbf{c}'\mathbf{X} \wedge z_0]\right\}\\ &= z_0 + \frac{1}{1-\alpha} \mathrm{E}[\mathbf{c}'\mathbf{X} - z_0]_+ .\\ \end{array} \tag{4.9} \] For fixed values of \(\mathbf{c}\), it is easy to check that the solution of the unconstrained minimization problem is the quantile. Moreover, the minimum value is the expected shortfall. This point will be developed in more detail in Section 7.2.2.
Moreover, we verify below that the function \(ES1(z_0, \mathbf{c})\) is convex in the decision variables \(z_0\) and \(\mathbf{c}\). Because the minimum of a convex function is also convex (see, for example, Boyd and Vandenberghe (2004), Section 3.2.5), this means that the \(ES\) function is convex.
\(Under~the~Hood.\) Show the Verification of the Convexity
Earlier in Display (3.13), we wrote the asset allocation problem as minimizing the portfolio variance subject to constraints on the mean return. Now, we extend this by using the expected shortfall as our risk measure. Specifically, we consider \[\begin{equation} \boxed{ \begin{array}{lcc} {\small \text{minimize}}_{z_0, \mathbf{c}} & ES1(z_0, \mathbf{c}) & \\ {\small \text{subject to}} & \mathbf{c}^{\prime} ~\mathrm{E}(\mathbf{R}) \ge Req_0 .\\ \end{array} } \tag{4.10} \end{equation}\]
Example 4.4.1. Optimizing a Portfolio of Insurance Stock Returns. This is a continuation of Example 1.3.1 and Example 3.4.1. We are now ready to compute the optimal portfolio specified in Display (4.10) with the CVXR
package.
As in Example 3.4.1, we determine optimal allocations using the training data. With these allocations, various summary measures are computed. Then, the optimal allocations are used on the test data and several summary measures are determined.
Table 4.6 summarizes performance for both the training and test samples. For comparison purposes, the performance measures summarized earlier in Example 3.4.1 are also included.
R Code to ES Portfolios
Train Ann Return | Train Std Dev | Train Sharpe (Rf=0) | Test Ann Return | Test Std Dev | Test Sharpe (Rf=0) | |
---|---|---|---|---|---|---|
MinVar | 0.164 | 0.119 | 1.378 | 0.168 | 0.143 | 1.178 |
Markowitz | 0.203 | 0.126 | 1.614 | 0.166 | 0.143 | 1.160 |
Equal | 0.173 | 0.130 | 1.329 | 0.155 | 0.142 | 1.093 |
ES-\(\alpha = 0.95\) | 0.173 | 0.120 | 1.436 | 0.158 | 0.144 | 1.099 |
ES-\(\alpha = 0.99\) | 0.181 | 0.128 | 1.412 | 0.155 | 0.152 | 1.020 |
Another way of presenting the performance of the differing portfolios is to track their performance over time. In Figure 4.5, we can see that all portfolios behave similarly over the test period.
As emphasized in Example 3.4.1, one cannot conclude too much from these results. Practicing analysts will use this approach while conducting more extensive testing by varying the training period, the test period, and the sample of stocks (risks) selected. In addition, many of the constraints on allocations introduced in Section 3.4.2
will come into play (note that these allocation constraints are linear and so can readily be incorporated into the convex optimization problem). Readers are invited to experiment with the code to determine their own optimal portfolios.
4.4.2 Constrained Optimization Without Coding Derivatives
Many risk retention problems cannot be classified as convex and so we need a more general approach. For some problems, one only needs to code the objective and constraint functions and use these specifications in a numerical optimization program. Many modern programs now have the ability to calculate approximate derivatives using finite difference methods and use these approximations in lieu of analytical expressions for the derivatives. For many problems, this approach is much simpler for the user and often works well particularly when there are small numbers of decision variables and constraints. Nonetheless, for some problems, they can take much longer for convergence to optimal solutions when compared to methods using analytical derivatives. Further, they can be sensitive to algorithmic parameters such as starting values. I show how to use this approach in R
using the function auglag
from the package alabama
.
However, in applied risk management problems even the evaluation of objective and constraint functions can be difficult. As seen in Example 4.2.1, this difficulty can be addressed using simulation as a method of approximating these functions. Virtually all numerical methods exhibit deviations from true values when evaluating a solution and simulation is certainly no exception. A strength of the simulation approach is that the approximation error can be reduced simply by increasing the number of simulated replicates. So, if the analyst is satisfied with a result based on the analysis of a simulated trial, then accuracy can be increased by replicating the simulated trial with a larger number of replicates. Let us see how the simulation approach is valuable in the following example.
Example 4.4.2. Constrained Optimization for Excess of Loss with Two Risks. Let us continue from Example 4.2.1. The risk owner retains \(S(u_1,u_2) = X_1 \wedge u_1 + X_2 \wedge u_2\) and desires to minimize \(f_0(u_1,u_2) = \mathrm{Var}[S(u_1,u_2)]\). This is constrained by the budget requirement \(f_{con,1}(u_1,u_2) = RTC(u_1,u_2) - RTC_{max}\le 0\) and constraints on the parameters \(f_{con,2}(u_1,u_2) =-u_1 \le 0\) and \(f_{con,3}(u_1,u_2) =-u_2 \le 0\). For simplicity, we use a fair risk transfer cost \(RTC(u_1,u_2) = \mathrm{E} [S(\infty,\infty)] - \mathrm{E}[S(u_1,u_2)]\). We use simulation to evaluate the objective function but not the risk transfer cost which is easy to evaluate.
R Code to Evaluate Constraint Functions
Recall from the earlier Example 4.2.1, Figure 4.4 showed the standard deviation of retained risks as a function of each upper limit. Figure 4.6 supplements this by showing the risk transfer cost \(RTC\) as a function of each upper limit. As anticipated, the risk transfer cost \(RTC\) decreases as each upper limit increases. The right-hand panel of Figure 4.6 demonstrates the trade-off between the risk transfer cost and the uncertainty of retained risk.
R Code to Show the Figures
With these objective and constraint functions, one can then solve the constrained optimization problem that is demonstrated in the associated code. This is done for a specific value of \(RTC_{max}\), which is then varied. This shows how the solution changes as the maximal risk transfer cost changes. Figure 4.7 summarizes results. Here, we see that as \(RTC_{max}\) increases, the optimal upper limits decrease, as does the uncertainty of retained risk.
R Code for Optimization without Derivatives
4.4.3 Constrained Optimization with Coding Derivatives
Another approach is to use a numerical optimization program that allows one to include derivatives of objective and constraint functions. This approach requires more upfront effort for the analyst, since you have to derive and code the derivatives, but often provides more stable and robust results. However, as seen in the following example, even developing the derivatives can be difficult.
Example 4.4.2. Excess of Loss with Two Risks – Continued. The derivatives of the constraint functions are relatively straightforward. We have \[ \begin{array}{ll} \partial_{u_j} f_{con,1}(u_1,u_2) & = \partial_{u_j} \left\{ \mathrm{E} (S(\infty,\infty)) - \mathrm{E}(S(u_1,u_2)) -RTC_{max} \right\} \\ & = -\partial_{u_j} \left\{ \mathrm{E}(X_1\wedge u_1 )+ \mathrm{E}(X_2\wedge u_2)\right\} \\ & = -\left\{ 1- F_j(u_j)\right\} , \end{array} \] as well as \[ \begin{array}{ll} \partial_{u_1} f_{con,2}(u_1,u_2) =\partial_{u_1} (-u_1) = -1, & \partial_{u_2} f_{con,2}(u_1,u_2) =\partial_{u_2} (-u_1) = ~~~0, \\ \partial_{u_1} f_{con,3}(u_1,u_2) =\partial_{u_1} (-u_2) = ~~~0, & \partial_{u_2} f_{con,3}(u_1,u_2) =\partial_{u_2} (-u_2) = -1 . \\ \end{array} \] For the objective function, a partial derivative can be shown to be \[\begin{equation} \begin{array}{ll} \partial_{u_1} f_0(u_1,u_2) & = \partial_{u_1} \mathrm{Var}[S(u_1,u_2)] \\ & = 2 \left\{ (u_1+u_2-\mathrm{E} [S(u_1,u_2)])[1-F_1(u_1)] \right. \\ & \left. \ \ \ \ -\int_0^{u_2} ~\left[ F_2(x_2) -F(u_1,x_2) \right]dx_2 \right\} .\\ \end{array} \tag{4.11} \end{equation}\] The result for the other gradient, \(\partial_{u_2} f_0(u_1,u_2)\), follows in the same way, simply switch the indices.
\(Under~the~Hood.\) Show the Verification of the Derivative of the Objective Function
The associated R
code for constrained maximization with derivatives uses the auglag
function. This code provides essentially the same results as those without derivatives and so the output is suppressed (the exception is when \(RTC_{max}\) is very small and there are numerical difficulties). For this example, it takes longer than the results without derivatives. This is because the numerical evaluation of the integral above required for each step of the minimization process. The code numerically evaluates the partial derivatives of the constraints and objective function yet uses a simulation approximation for the objective function. This is because the calculation of the variance requires double integration which can be computationally intensive when done as part of an iterative optimization procedure.
R Code to Optimization with Derivatives
4.4.4 Root-Finding Methods (with Derivatives)
With derivatives developed and coded, we can also employ a more basic approach. Simply take the derivatives of the Lagrangian, set these equal to zero, and solve a system of equations using a root-finding algorithm. That is, the beauty of the Lagrangian approach is that it turns a constrained optimization problem into an unconstrained one. And, we know from calculus, for many smooth problems, you simply take a derivative of the function and solve for zeros to get its critical values. The limitation of this approach, mentioned earlier, is the working assumption that the constraints are active, or binding, at the optimum.
This approach can be implemented numerically in R
using the function multiroot
from the rootSolve
package, see the associated R
code. Like any root-finding algorithm, this function only finds critical values (of which there may be many) and it is up to the analyst to ensure that they minimize the objective function over the entire feasible region. Further, the function requires starting values which, when varied, can lead to different critical values.
Example 4.4.2. Excess of Loss with Two Risks – Continued. The associated R
code shows how to solve the problem using root finding methods. As before, because the results are similar to those from other approaches, the findings are not displayed.
R Code for Optimization Using Root-Finding Methods Example 4.4.2
As another example of root-finding methods, you may wish to revisit the code in Example 4.3.3.
For analysts who wish to take a deeper dive into the intricacies of optimization problems in R
, visit
CRAN Task View: Optimization and Mathematical Programming. Among many other packages and approaches, this page has further links to three packages that focus on variations of the Lagrange method:
- The package
alabama
provides an implementation of the augmented Lagrange multiplier method for solving nonlinear optimization problems with nonlinear equality and inequality constraints. The acronym stands for Augmented Lagrangian Adaptive Barrier Minimization Algorithm. This is the primary package used in this book. - The package
nloptr
provides access to various nonlinear optimization algorithms, including the augmented Lagrangian approach for nonlinear constraints. - The package
Rsolnp
also provides an implementation of the augmented Lagrange multiplier method for solving problems with equality and inequality constraints.
Video: Section Summary
4.5 Risk Exchanges
We have focused on a risk transfer defined to be the change of possession of an uncertain obligation, a risk, from one party to another in exchange for (certain) monetary compensation. This definition includes commercial insurance as well as many other mechanisms. For example, you might be the firm risk manager seeking to transfer risks due to unforeseen events that could damage the company’s property, financial health due to cyber events, and the well-being of the workforce due to accidents at the work facility. Corporate insurance is one way to transfer these potential liabilities but there are other possibilities, including self-insurance, utilizing a captive insurer, participation in a pool or group, insurance from a governmental entity, and other alternative risk transfer mechanisms. Section 6.2 will provide more detail on these alternatives. Although our primary focus is on insurance as a motivation for risk transfer, we aim to develop frameworks that are sufficiently general to handle other mechanisms as well.
Risk owners seek ways to offload some or all of their risk. Instead of simply ceding a portion of the risk, it is helpful to think about a general framework of exchanging risks. Specifically, we consider \(n\) exchange participants known as agents. These agents may be risk owners, such as a firm or an individual, or they may be another party, such as an insurer or a pool. For example, an agent may be a reinsurer that brings no risk to the exchange but agrees to take on risks in exchange for a fee. Associated with each agent, say the ith one, is a set of p risks \(\mathbf{X}_i = (X_{i1}, \ldots, X_{ip})\). A risk may be a (degenerate) random variable that has no uncertainty, such as cash.
When terms of a risk exchange have been agreed upon, a set of rules or functions \(g_i\) is established, such that \[\begin{equation} \mathbf{Y}_i = g_i(\mathbf{X}_1, \ldots, \mathbf{X}_n), \ \ \ i=1, \ldots, n , \tag{4.12} \end{equation}\] defines the collection of risks of the ith agent after the exchange has taken place (but before the random variables have been realized). The goal of the risk exchange, also known as the risk-sharing, literature is to establish guidance for suitable choices of the rules \(g_i\). For example, one basic principle in selecting the rules is to ensure that \(\sum_{i=1}^n \mathbf{Y}_i = \sum_{i=1}^n \mathbf{X}_i\). This ensures that the sum of risks from all risk owners is equal before and after the exchange. In financial transactions, this principle is known as “clearing the market” and so clearing the exchange seems appropriate here.
Another fundamental goal is to achieve an improvement in the risk position for all parties. That is, the measure of uncertainty for the ith party after the exchange should not be larger than prior to the exchange. The reasoning is that if a party’s uncertainty increases due to the exchange, that party will not enter into the exchange.
In addition to these basic goals, there is another desirable principle known as Pareto optimality. An exchange is said to be Pareto-optimal if there is no other exchange that is strictly better. This principle, as with the risk improvement principle, depends on the type of risk measure used to quantify uncertainty, as well as the application area. Section 4.5.2 provides additional discussion on Pareto optimality. In the following three subsections, we provide some guidance based on separate strands of the actuarial literature. In the appendix Section 4.6.4, we describe an alternative approach for determining risk exchanges based on quantile risk measures such as \(RVaR\).
4.5.1 Linear Exchanges
We now know that the risk exchange literature likely began with the work of De Finetti (1940); Seal (1969) provides a detailed description of its early history up to its publication date. Based on this long history, we adopt the structure put forth in Feng, Liu, and Taylor (2023).
Similar to most approaches, let us focus on only a single type of risk, so \(p=1\). If cash is required to equate a risk-sharing agreement, one regards it as a “side payment.” In this section, we use variance as a measure of risk and, consistent with the early literature, emphasize proportional risk-sharing agreements.
Consider \(n\) agents who have risks \(X_1, \ldots, X_n\) to be shared among them. In the special case where all risks have the same distribution, it is desirable to share risks equally. That is, the share for agent i after the exchange is \(Y_i = (X_1 + \cdots + X_n)/n\), the average. When the risks are independent, the uncertainty after the exchange \(\text{Var}(Y_i)= \text{Var}(X_i)/n\) is much lower than prior to the exchange.
In general, risks do not have the same distribution nor are they independent. To handle this more complex situation, this subsection considers linear exchanges, or sharing rules, where risks are combined in an additive fashion and the proportion of each risk is set as part of the sharing rule. Specifically, define \(c_{ij}\) to be the proportion of the \(j\)th risk that agent \(i\) takes on after a risk-sharing agreement. In sum, the \(i\)th agent has responsibility for \(Y_i = c_{i1}X_1 + \cdots + c_{in} X_n\), a random linear combination of losses. This can be summarized for all agents as: \[ \mathbf{Y} = \left( \begin{array}{c}Y_1 \\ \vdots \\Y_n \end{array} \right)= \left( \begin{array}{ccc}c_{11} &\cdots& c_{1n} \\ \vdots &\ddots &\vdots \\c_{n1} &\cdots &c_{nn} \end{array} \right) \left( \begin{array}{c}X_1 \\ \vdots \\X_n \end{array} \right) = \mathbf{C}\mathbf{X} . \] Several conditions potentially apply: \[\begin{equation} \boxed{ \begin{array}{cc} \mathbf{C}' {\bf 1}= {\bf 1} & \text{clearing the exchange} \\ \mathbf{C} \boldsymbol \mu = \boldsymbol \mu & \text{no profits}\\ 0 \le c_{ij} \le 1 & \text{no short/long selling} \\ \mathrm{Var} (Y_i) \le \mathrm{Var} (X_i) & \text{risk improvement} \end{array} } . \tag{4.13} \end{equation}\] The first condition, “clearing the exchange,” ensures that \({\bf 1' X} = {\bf 1}' {\bf C X} = {\bf 1' Y}\), so that the sum of risks before the exchange equals the sum of risks after the exchange, for any realization of the random vector \(\mathbf{X}\). The second condition on the means (\(\mathrm{E} ({\bf X}) = \boldsymbol \mu\)) implies that this is a non-profit arrangement; that is, no agent can expect to make profit (so that \(\mathrm{E} ({\bf Y}) = \mathrm{E} ({\bf X})\)). The third condition excludes “short selling” (and “long” positions) as is common in financial investment transactions. An agent may not take a negative position nor more than 100% of a risk. The fourth condition is about improving a risk position; no agent will be worse off by participating in the risk-sharing agreement.
There are many choices of the matrix \(\mathbf{C}\) that satisfy the conditions in Display (4.13). The identity matrix is certainly one, ensuring that the set of feasible parameters is not empty. To obtain a unique set of parameters, we focus on minimizing \(\sum_i \mathrm{Var} (Y_i)\) over choices of \(\mathbf{C}\) subject to the conditions in Display (4.13). Here, one can interpret \(\sum_{i=1}^n \mathrm{Var}(Y_i)\) to be a “system variance.”
As is common for optimization problems, with only equality constraints one can obtain closed-form expressions for the solutions. This approach is outlined in Exercise 4.5.2 and Exercise 4.5.3. More generally (including inequality constraints), this is a convex programming problem that can be readily solved numerically. Specifically, it is a quadratic program, and so fast solvers are available (see, for example, Boyd and Vandenberghe (2004), Section 4.4). Fast solvers mean that it can be easy to optimize large dimensional problems. As an illustration, Exercise 4.5.3 outlines a strategy for \(n=50\) agents (that works in about 4 minutes). Nonetheless, to start, first consider the following example.
Example 4.5.1. Linear Sharing Among Three Agents. Assume that \(n=3\) and the mean vector and variance-covariance matrix are given as
\[
\boldsymbol \mu =
\left(\begin{array}{rrr}20 \\ 2.5 \\ 10
\end{array}\right)
\ \ \ \ \ \ \text{and} \ \ \ \ \ \ \
\boldsymbol \Sigma =
\left(\begin{array}{rrr}
10 & -4 & -1 \\
-4 & 8 & 1 \\
-1 & 1 & 1\\
\end{array}\right) .
\]
Prior to the exchange, the sum of all variances is \(\sum_{i=1}^n \mathrm{Var}(X_i) =\) \(10 + 8 + 1 = 19\). We then determine values of the coefficient matrix \(\bf C\) that minimizes the system variance after the exchange, \(\sum_{i=1}^n \mathrm{Var}(Y_i)\), subject to a sequence of constraints. The results are summarized in Tables 4.7 and Table 4.8 based on the associated R
code.
R Code for Example 4.5.1
Table 4.7 shows the impact of the exchange on the uncertainty measure for each agent. To illustrate, for Agent 1, the original (prior to the exchange) variance is \(\mathrm{Var}(X_1)=10\). After imposing on the clearing exchange condition, Agent 1’s variance is \(\mathrm{Var}(Y_1)=1.2222\). If instead one does the optimization problem with the clear exchange and the no profits condition, Agent 1’s variance is \(\mathrm{Var}(Y_1)=2.6281\). Repeating, with the constraints clear exchange, no profits, and short sell, Agent 1’s variance is \(\mathrm{Var}(Y_1)=2.8881\). Finally, with all four constraints, Agent 1’s variance is \(\mathrm{Var}(Y_1)=3.3164\). The bottom row provides the sum of variances for all three agents. After the exchange, we see that the system variance increases as more constraints are imposed as one would expect. Nonetheless, all four options provide a smaller system variance than the original set-up. This illustrates the benefits of diversification.
It is also interesting to observe the uncertainty measure for Agent 3. From the third row of the table, we see that this agent’s variance actually increases under the first three exchanges. It is only when we specifically impose the risk improvement condition is this agent at least no worse off.
Original | Clear Exchange | No Profits | Short Sell | Risk Improvement | |
---|---|---|---|---|---|
Agent 1 | 10 | 1.2222 | 2.6281 | 2.8881 | 3.3164 |
Agent 2 | 8 | 1.2222 | 0.6695 | 0.3415 | 0.4148 |
Agent 3 | 1 | 1.2222 | 1.1359 | 1.3959 | 1.0000 |
Total | 19 | 3.6667 | 4.4335 | 4.6256 | 4.7312 |
Table 4.8 provides more details about the optimal coefficients. Each optimization procedure yields nine coefficients, three for each risk. The table shows only those for Risk 3. First note that each column sums to one, as required under the clearing the exchange condition. Also observe the negative coefficient in the No Profits
column. This corresponds to an optimization problem before we explicitly introduce the short sell condition, indicating the importance of this condition. In fact, from the Short Sell
column we see that the lower bound is a binding constraint for some coefficients.
Clear Exchange | No Profits | Short Sell | Risk Improvement | |
---|---|---|---|---|
To Agent 1 | 0.3333 | 0.9286 | 0.8247 | 0.7119 |
To Agent 2 | 0.3333 | -0.2078 | 0.0000 | 0.0000 |
To Agent 3 | 0.3333 | 0.2792 | 0.1753 | 0.2881 |
4.5.2 Using Utility Functions to Determine Risk Exchanges
Following the work of De Finetti (1940), intellectual development of risk exchanges was subsequently reignited by contributions due to Karl Borch, summarized in Borch (1974). Borch focused on economic modeling using utility theory. To keep this treatment self-contained, a brief overview of utility theory can be found in Section 4.6.3.
In this context, the \(i\)th firm or agent has an initial wealth \(w_i\), faces potential losses \(X_i\), and values money using the utility function \(U_i(\cdot)\). Before the risk exchange, its expected utility is \(\mathrm{E}[U_i(w_i - X_i)]\). As in equation (4.12), a set of risk exchange rules is denoted by \(g_i\), so that after the exchange, one is responsible for the loss \(Y_i = g_i(X_1, \ldots, X_n)\). After the risk exchange, the expected utility is \(\mathrm{E}[U_i(w_i - Y_i)]\). Recall that a risk exchange is said to be Pareto optimal if it is not possible to improve the situation of one company without worsening the situation of at least one other company.
Theorem (Borch). An exchange \(\{g_1, \ldots, g_n\}\) is Pareto-optimal if and only if there exist non-negative constants \(k_1, \ldots , k_n\) such that, with probability one, \(k_i~ U_i'[w_i - g_i(X_1, \ldots, X_n)]\) is the same for all \(i\).
Before discussing the implications of this result, let us first see how it works in a special case.
Example 4.5.2. Exponential Utility Functions. Suppose that each agent has an exponential utility function of the form \(U_i(x) = \alpha_i[1-\exp(-x/\alpha_i)]\). In this case, the solution of Borch’s Theorem is a familiar quota share treaty with agent \(i\)’s responsibility being \(Y_i = c_i~ S + P_i\), where \[ c_i = \frac{\alpha_i}{\alpha_1 + \cdots + \alpha_p} \ \ \ \text{and} \ \ \ P_i = w_i - c_i\sum_{j=1}^n \left[w_j + \alpha_j \ln \frac{k_i}{k_j} \right] . \] Thus, each company takes a proportional share \(c_i\) of the overall risk pool, \(S =X_1 + \cdots + X_n\). To equalize the transactions, cash side-payments in the form of \(P_i\) are made.
As summarized by Lemaire (1991):
- This kind of exchange is common in reinsurance practice. Each agent will pay a share \(c_i\) of each claim, inversely proportional to risk aversion \(\alpha_i\).
- In order to compensate for the fact that the least risk-averse companies will pay greater claim amounts, zero-sum (\(\sum_i P_i = 0\)) side-payments or “fees” \(P_i\) between the participants occur.
- Note that the quota proportions are determined by the risk-aversion parameters only: they are non-negotiable. Hence the bargaining process will involve only the monetary compensations. Stated another way, agents will negotiate about amounts of money, not about abstract bargaining constants \(k_i\).
- It turns out that quota-share treaties with side-payments also occur when all exchange participants use quadratic or logarithmic utilities, and only in those cases. In these cases, both the quotas and the monetary compensations will be subject to negotiation.
\(Under~the~Hood.\) Development of the Exchange for Exponential Utility Functions
Example 4.5.2 helps to underscore some additional features of Borch’s theorem from, for example, Lemaire (1991):
First, note that the condition of the theorem does not depend on the loss amount distributions. There is no mention of the shape of the distribution, the number of moments required, or any other restrictions.
Second, in the exponential utility example, we saw that the exchange rules only depend on the sum of losses, \(S=X_1 + \cdots + X_n\) although we allowed for more general functions of the form \(g_i(X_1, \ldots, X_n)\). This is true more broadly under mild regularity conditions such as, for example, requiring differentiability of \(g_i\).
From Borch (1962), a Pareto-optimal allocation can be achieved by choosing positive constants \(k_1, \ldots, k_n\) and trying to maximize \(\sum_{i=1}^n k_i~ \mathrm{E}[U_i(X_i)]\). We may interpret the positive constants as weighting factors to equalize contributions of each company.
It is important to remark that the whole model relies on two important implicit assumptions:
- Shared beliefs: All companies use the same joint distribution of claim amounts.
- Truthfulness: All companies honestly report the utility function that represents their attitude toward risk.
4.5.3 Conditional Mean Risk-Sharing Rule
Another type of risk exchange assumes that all risks go into a central pool. Now denote the pool risk as \(S_n = X_1 + \cdots + X_n\), the total amount to be exchanged with conditional mean \(g_{i,n} (s) = \mathrm{E}[X_i|S_n = s]\). With the conditional mean risk-sharing rule, the \(i\)th participant’s share of the total risk is \(Y_i = g_{i,n} (S_n)\). Denuit and Robert (2021a) refers to this setting as a self-governing, or no insurance company, model.
Using conditional expectations, it is straightforward to check that this rule satisfies the market clearing and fairness conditions, that is, that \(\sum_i X_i = S_n = \sum_i Y_i\) and that \(\mathrm{E}(X_i) = \mathrm{E}(Y_i)\). Moreover, from Jensen’s inequality for conditional expectations, we have that \[ \mathrm{E} [U_i(w_i - X_i)] \le \mathrm{E} [U_i(w_i -g_{i,n}(S_{n}))], \ \ \ i=1, \ldots, n, \] for concave utility functions \(U_i\) and initial wealth levels \(w_i\). Thus, the conditional mean risk allocation benefits all risk‐averse economic agents, thereby improving their risk position. Moreover, under some mild regularity conditions (when \(g_{i,n}(s)\) is non-decreasing in \(s\) for all \(i\)), the conditional mean risk-sharing rule is Pareto-optimal.
As an example of other results, Denuit and Robert (2021a) also shows \[ \mathrm{E} [U_i(w_i - g_{i,n}(S_{n}))] \le \mathrm{E} [U_i(w_i - g_{i,n+1}(S_{n+1})], \ \ \ i=1, \ldots, n . \] We may interpret this to mean the risk position of the \(i\)th participant improves as the exchange pool becomes larger. They also cite asymptotic results about the pool (as it grows larger), such as law of large numbers and central limit theorem type results, that give qualitative information about the behavior of large pools.
Example 4.5.3. Conditional Mean Risk-Sharing Among Three Agents. This is a continuation of Example 4.5.1 with the same means and variances for \(n=3\) agents. In Example 4.5.1, we only needed means and variances but now supplement this by assuming that the first and third agents follow a gamma distribution while the second agent follows a Pareto distribution. Further, the correlations of Example 4.5.1 are now treated as association parameters for a Gaussian copula to allow for dependencies.
With knowledge of the entire distribution of the losses \(\{X_1, X_2, X_3\}\), one can calculate the conditional mean risk-sharing rules via simulation. With simulated observations, I use nonparametric regression techniques as described in, for example, Härdle (1990) (Chapter 3). Specifically:
- For \(r=1 , \ldots, R\), generate simulated losses \(\{X_1^r, X_2^r, X_3^r\}\) and sum them to get \(S^r\).
- From simulated outcomes \(\{S^r\}\), build a nonparametric density estimator \(\hat{f}_{S}(s)\). This appears in the left-hand panel of Figure 4.8.
- For different values of \(s\), calculate the estimator \[ \hat{g}_j(s) = \frac{\sum_{r=1}^R X_j^r \times I( s - bw \le S^r \le s+bw)}{\sum_{r=1}^R I( s - bw \le S^r \le s+bw)}, \ \ \ j=1,2,3, \] where \(bw\) is known as a bandwidth for the estimator (we use \(bw=1\) in our example). In this literature, this is sometimes called a Nadaraya-Watson estimator. These appear in the middle panel of Figure 4.8. It is also of interest to examine the proportion of the realized risk (\(S=s\)) taken on by each agent, \(\hat{g}_j(s)/s\). These appear in the right-hand panel.
R Code for Example 4.5.3
From Example 4.5.1, recall that the sum of the means was \(20+ 2.5+ 10 = 32.5\); this is consistent with the density plot in the left-hand panel of Figure 4.8. The third risk had a mean of 10 and a variance of 1; so when we examine its risk-sharing behavior as the solid black line, the middle panel shows it is relatively level (about the mean of 10) whereas the right-hand panel shows that its proportion decreases as the sum of (realized) risks, \(S=s\), increases. In contrast, the second risk is the more volatile Pareto risk, depicted as the red dashed line. This risk’s conditional mean proportion increases as \(s\) increases. The first risk, the other gamma distributed random variable, serves as a balancing item for the other two (balancing because the sum of the three risks must add to \(s\)). The conditional mean for this risk increases as \(s\) increases but decreases when viewed as a proportion of the total risk \(s\).
One can also compare the results from the analysis in Example 4.5.1 for the Section 4.5.1 proportional risk-sharing rules to those from the conditional mean sharing rules. Table 4.9 provides this comparison, extending the results reported in Table 4.7 to include those from the conditional mean sharing rules. As in Table 4.7, the first column of Table 4.9 gives the original (prior to the exchange) variances for each agent and columns two through five gives optimal variances imposing more conditions on the proportional risk-sharing rules. The right-hand column shows variances of each agent assuming that a conditional mean sharing rule is in place.
Not surprisingly, these compared favorably to other rules considered. Recall that the Section 4.5.1 rules use only information from the first two moments and restrict consideration to proportional sharing rules. In contrast, the conditional mean sharing rule requires knowledge of the entire joint distribution to compute the conditional means.
Original | Clear Exchange | No Profits | Short Sell | Risk Improve | Conditional Mean | |
---|---|---|---|---|---|---|
Agent 1 | 10 | 1.2222 | 2.6281 | 2.8881 | 3.3164 | 2.6257 |
Agent 2 | 8 | 1.2222 | 0.6695 | 0.3415 | 0.4148 | 0.5706 |
Agent 3 | 1 | 1.2222 | 1.1359 | 1.3959 | 1.0000 | 0.0289 |
Total | 19 | 3.6667 | 4.4335 | 4.6256 | 4.7312 | 3.2251 |
Video: Section Summary
4.6 Supplemental Materials
4.6.1 Further Resources and Readings
There are many superb treatments of reinsurance in the literature. An outstanding book-long introduction is Albrecher, Beirlant, and Teugels (2017); see also Verlaak and Beirlant (2003).
Lourdes Centeno and Simões (2009) identify the pioneering work by Bruno de Finetti in 1940, De Finetti (1940), as the first treatment of risk sharing in the academic literature. Especially with the variance as a measure of uncertainty, Beard, Pentikäinen, and Pesonen (1984) summarizes work that influenced generations of actuaries.
The Section 4.4.1 asset allocation problem has already been discussed in Sections 1.3 and 3.4. Additional details on asset allocations can be found in Ingersoll (1987) and Panjer et al. (1998) (Chapter 8).
In addition to the references already cited in Section 4.5.2 on utility based risk exchanges, see H. U. Gerber (1979) and Albrecher, Beirlant, and Teugels (2017) for further discussions and references to this fascinating literature. An accessible proof of Borch’s Theorem in Section 4.5.2 may be found in H. Gerber and Pafumi (1998). The Section 4.5 risk exchanges are motivated by pooling of risks that are generally viewed as beneficial for both insurance firms, the suppliers of insurance, as well consumers, who demand insurance. For example, Albrecht and Huggenberger (2017) provide general conditions under which consumers with exchangeable risks will improve their risk positions by pooling (exchangeable risks have the same distribution but may exhibit some positive dependence, a generalization of the idea of “i.i.d.” - independent and identically distributed risks). However, as demonstrated in Exercise 4.5.1, pooling can fail when risks have different distributions, the so-called heterogeneous case.
The conditional mean risk-sharing rule is due to Denuit and Dhaene (2012). The peer-to-peer risk exchange development in Section 4.5.3 follows the work of Denuit and Robert (2021a). Further exploration of this topic appears in Section 6.5.
4.6.2 Exercises
Section 4.1 Exercise
Exercise 4.1.1. Portfolio Management. Let us consider the same structure as in Exercise 2.2.2, with risks:
- \(X_1\) - buildings, modeled using a gamma distribution with mean 200 and scale parameter 100.
- \(X_2\) - motor vehicles, modeled using a gamma distribution with mean 400 and scale parameter 200.
- \(X_3\) - directors and executive officers risk, modeled using a Pareto distribution with mean 1000 and scale parameter 1000.
- \(X_4\) - cyber risks, modeled using a Pareto distribution with mean 1000 and scale parameter 2000.
Denote the total risk as \(S = X_1 + X_2 + X_3 + X_4\). For simplicity, in this exercise we assume that these risks are independent.
To manage the risk, you seek some reinsurance protection. You wish to manage internally small and intermediate building and motor vehicles amounts, up to \(u_1\) and \(u_2\), respectively. That is, your retained risk is \(Y_{retained}= \min(X_1,u_1) + \min(X_2,u_2)\). You seek reinsurance to cover all other risks. Specifically, the reinsurer’s portion is \[ Y_{reinsurer} = S- Y_{retained}=(X_1 - u_1)_+ + (X_2 - u_2)_+ + X_3 + X_4 . \] Using upper limits \(u_1=\) 100 and \(u_2=\) 200, this example will:
a. Determine the expected claim amount of (i) that retained, (ii) that accepted by the reinsurer, and (iii) the total overall amount.
b. Determine the 80th, 90th, 95th, and 99th percentiles for (i) that retained, (ii) that accepted by the reinsurer, and (iii) the total overall amount.
c. Compare the distributions by plotting the densities for (i) that retained, (ii) that accepted by the reinsurer, and (iii) the total overall amount.
Show Exercise 4.1.1 Solution with R Code
Section 4.3 Exercise
Exercise 4.3.1. Sensitivity of Optimal Quota Share Proportions. Consider the quota share problem with dependent risks described in Section 4.3.2. How sensitive are these optimal proportions to small changes in \(\boldsymbol \Sigma\)? One can derive explicit results in the simpler case where we do not impose non-negative constraints on the parameters. The more general problem, with non-negativity constraints, will be taken up in Chapter 10.
Specifically, if we let \(\sigma\) be a generic element of \(\boldsymbol \Sigma\), show that where
Solution to Exercise 4.3.1
Section 4.4 Exercise
Exercise 4.4.1. Markowitz Investment Optimal Allocations. Recall the notation from Section 3.4 on the asset allocation problem. In this exercise, we do not require \(c_i \ge 0\), so that “short sales” are permitted, yet do require \(\sum_i c_i =1\). With this assumption, show that the optimal allocations can be expressed as \[\begin{equation} \mathbf{c} = \mathbf{c}(LME_1, LME_2) = LME_1 ~\boldsymbol \Sigma^{-1} [\mathrm{E}(\mathbf{R})] +LME_2 ~\boldsymbol \Sigma^{-1} \mathbf{1} , \tag{4.16} \end{equation}\] where \[\begin{equation} {\small LME_1 = \frac{A_1A_3-A_2^2}{Req_0 ~A_3-A_2} ~~~~~~~ LME_2 = \frac{A_2^2 - A_1A_3}{Req_0 ~A_2-A_1} } \tag{4.17} \end{equation}\]
and \(A_1=\mathrm{E}(\mathbf{R})^{\prime}\boldsymbol \Sigma^{-1} \mathrm{E}(\mathbf{R})\), \(A_2=\mathbf{1}^{\prime}\boldsymbol \Sigma^{-1} \mathrm{E}(\mathbf{R})\), and \(A_3=\mathbf{1}^{\prime}\boldsymbol \Sigma^{-1} \mathbf{1}\).
Solution to Exercise 4.4.1
Section 4.5 Exercises
Exercise 4.5.1. Risk Sharing Does not Always work with Heterogeneous Risks. Consider a set of independent risks \(X_j\) that have the same mean (say \(\mu\)) with variances \(\sigma_j^2\) that are possibly different.
a. Suppose that two risks are shared equally so that \(Y_1 = (X_1 + X_2)/2 = Y_2\) but that the first risk has a higher variance such that \(\sigma_1^2 = 4\) and \(\sigma_2^2 = 1\). Show that this risk sharing arrangement improves the risk position for the first risk but not the second. That is, show that
\[
\mathrm{Var}(Y_1) \le \mathrm{Var}(X_1) \ \ \ \ \text{and} \ \ \ \mathrm{Var}(Y_2) \ge \mathrm{Var}(X_2) .
\]
b. Suppose that \(n\) risks are shared equally so that \(Y_j = (X_1 + \cdots +X_n)/n\) but that the first risk has a higher variance \(\sigma_1^2\) and other risks have equal variance, say \(\sigma_2^2\). Show that the risk position for all other risks improves only if \(\sigma_1^2\) is not too large relative to the others. Specifically, show that \(\mathrm{Var}(Y_j) \le \mathrm{Var}(X_j)\) if \(\sigma_1^2 \le (n^2-n+1) \sigma_2^2\).
c. For two risks, suppose that the first risk takes \(Y_1 =c X_1 + (1-c) X_2\) and the second takes \(Y_2 = (X_1+X_2) - Y_1\) \(= (1-c) X_1 + c X_2\). Show that the variability of each risk must be relatively similar for risk improvement for both risks. Specifically, for \(\mathrm{Var}(Y_j) \le \mathrm{Var}(X_j)\), \(j=1,2\), we require \(\sigma_2^2 k_c \le \sigma_1^2 \le \sigma_2^2/k_c\), with the constant \(k_c = (1-c)/(1+c)\).
Solution to Exercise 4.5.1
Exercise 4.5.2. Optimal Linear Exchange Based on a Clearing Exchange Condition. Based on the risk sharing structure described in Section 4.5.1, now use only the exchange clearing condition for this problem. The goal is to determine the set of risk sharing/allocation parameters \(c_{ij}\) that \[ \boxed{ \begin{array}{lc} {\small \text{minimize}}_{\bf C} & \sum_i \mathrm{Var} (Y_i) \\ {\small \text{subject to}} & \mathbf{C}' {\bf 1}= {\bf 1} . \end{array} } \]
a. Show that the solution is given as \[\begin{equation} \begin{array}{ll} \mathbf{C}_{opt} &= \frac{1}{n}{\bf 1} {\bf 1}' .\\ \end{array} \tag{4.18} \end{equation}\]
b. Show that the variance is \[ \begin{array}{ll} \mathbf{C}_{opt} \boldsymbol \Sigma \mathbf{C}_{opt}' &= \frac{k_{\Sigma}}{n^2}{\bf 1} {\bf 1}' .\\ \end{array} \]
with the constant \(k_{\Sigma}= {\bf 1}'\boldsymbol \Sigma {\bf 1}\).
Solution to Exercise 4.5.2
Exercise 4.5.3. Optimal Linear Exchange with Clearing Exchange and No Profits Conditions. Continuing from Exercise 4.5.2, now add to the exchange clearing the no profits condition. As before, the goal is to find a set of risk sharing/allocation parameters \(c_{ij}\) that minimize, for example, the system variance \(\sum_{i=1}^n \mathrm{Var} (Y_i)\). In summary, \[ \boxed{ \begin{array}{lc} {\small \text{minimize}}_{\bf C} & \sum_i \mathrm{Var} (Y_i) \\ {\small \text{subject to}} & \mathbf{C} \boldsymbol \mu = \boldsymbol \mu \ \ \ \text{ and } \mathbf{C}' {\bf 1}= {\bf 1} . \end{array} } \] a. Show that the solution is given as \[\begin{equation} \begin{array}{ll} \mathbf{C}_{opt} &= \frac{1}{n} {\bf 1}{\bf 1}' + \frac{1}{k_{\mu}} \boldsymbol \mu^*~\boldsymbol \mu'\boldsymbol \Sigma ^{-1} . \end{array} \tag{4.19} \end{equation}\] with the constant \(k_{\mu}= \boldsymbol \mu'\boldsymbol \Sigma^{-1} \boldsymbol \mu\) and the deviations from the means vector \(\boldsymbol \mu^* = \boldsymbol \mu- {\bar{\mu}} {\bf 1}\). Note that this vector is zero when the means are constant.
b. Show that the variance is \[\begin{equation} \begin{array}{ll} \mathbf{C}_{opt} \boldsymbol \Sigma \mathbf{C}_{opt}' &=\frac{1}{n^2} k_{\bf 1} {\bf 1}{\bf 1}' + \frac{1}{k_{\mu}} \bar{\mu} \left[ \boldsymbol \mu^* {\bf 1}' + {\bf 1} {\boldsymbol \mu^*}' \right] + \frac{1}{k_{\mu}} {\boldsymbol \mu^*} {\boldsymbol \mu^*}' . \\ \end{array} \tag{4.20} \end{equation}\] where we recall the constant \(k_{\Sigma}= {\bf 1}'\boldsymbol \Sigma {\bf 1}\).
Solution to Exercise 4.5.3
Exercise 4.5.4. Optimal Linear Exchange - Special Case. Based on the risk sharing structure described in Section 4.5.1, now consider a risk exchange where all participants share a proportion of the total risk \(\sum_{j=1}^n X_j = {\bf 1' X}\). That is, the allocation matrix is expressed as \({\bf C} = {\bf c 1}'\) and \(\bf c\) is an \(n \times 1\) vector of coefficients. Check that:
a. The implication of clearing the exchange condition is that \({\bf c' 1} = 1\), that is, the coefficients sum to one.
b. The implication of the no profits condition is that \({\bf c} = \frac{1}{n \bar{\mu}} \boldsymbol \mu\), that is, the coefficient vector equals the mean vector, rescaled by the sum of means. Here, the average mean is denoted as \(\bar{\mu} = \frac{1}{n} \bf 1'\boldsymbol \mu\).
c. The implication of the risk improvement condition is that \(k_{\Sigma} ~c_i^2 \le \sigma_i^2\) with the constant \(k_{\Sigma}= {\bf 1}'\boldsymbol \Sigma {\bf 1}\).
d. Use the parameter values from the Section 4.5.1 example to determine optimal risk exchange variances and coefficients under the
- clear exchange,
- clear exchange and no short sell, as well as the
- clear exchange, no short sell, and risk improvement conditions.
Solutions are provided in Tables 4.10 and 4.11, respectively.
Original | Clear Exchange | Short Sell | Risk Improvement | |
---|---|---|---|---|
Agent 1 | 10 | 1.2222 | 1.2222 | 1.3417 |
Agent 2 | 8 | 1.2222 | 1.2222 | 1.3417 |
Agent 3 | 1 | 1.2222 | 1.2222 | 1.0000 |
Total | 19 | 3.6667 | 3.6667 | 3.6834 |
Clear Exchange | Short Sell | Risk Improvement | |
---|---|---|---|
Agent 1 | 0.3333 | 0.3333 | 0.3492 |
Agent 2 | 0.3333 | 0.3333 | 0.3492 |
Agent 3 | 0.3333 | 0.3333 | 0.3015 |
Solution to Exercise 4.5.4
R Code For Exercise 4.5.4
Exercise 4.5.5. Following up on Example 4.5.3, modify the code to allow for a larger number of risks so that we can demonstrate the quick computational ability of the convex solvers. As the purpose is to demonstrate the utility of numerical convex solvers, use a random number generator to generate generic mean and variance-covariance matrix (although make sure the variance-covariance matrix remains positive definite). Figure 4.9 demonstrates risk sharing variances under different conditions.
Solution to Exercise 4.5.5
Exercise 4.5.6. Conditional Mean Risk-Sharing Rules. From the notation introduced in Section 4.5.3, we have \(w_{i,n} = g_{i,n}(w_n)\) and \(B_{i,n} = (w_{i,n} - g_{i,n}(S_n))_+\). Assume that \(g_{i,n}(s) = E(X_i |S_n=s)\) is non-decreasing in \(s\). Establish that \((w_n - S_n)_+ = \sum_i ~B_{i,n}\).
Solution to Exercise 4.5.6
Exercise 4.5.7. Comparing Proportional to Conditional Mean Risk-Sharing Rules. Following up on Example 4.5.1 and Example 4.5.3, use the statistical code to compare the proportional sharing rules to the conditional mean sharing rules. Compare these two rules by varying the level of dependence. Also examine the effects of changing the marginal distributions to more short-tail (e.g. exponential) and long-tail (e.g., Pareto). What other assumptions do you think might be important? (e.g., how about the number of risks?)
4.6.3 Appendix. Utility Theory
This is a brief review of utility functions, drawn primarily from H. Gerber and Pafumi (1998).
To explain certain phenomena, the usefulness of money can be measured on a new scale. Thus, the usefulness of $x is \(U(x)\), the utility (or moral value) of $x. Typically, \(x\) is the wealth or a gain of a decision-maker.
Suppose that a utility function \(U(x)\) has the following two basic properties:
- \(U(x)\) is an increasing function of \(x\)
- \(U(x)\) is a concave function of \(x\).
Usually one assumes that the function \(U(x)\) is twice differentiable; then (1) and (2) state that \(U'(x) > 0\) and \(U''(x) < 0\).
For a given utility function \(U(x)\), one can define \[ r(x) = -\frac{U''(x)}{U'(x)} = -\frac{d}{dx} \ln U'(x), \] known as a risk aversion function. Note that properties (1) and (2) imply that \(r(x) > 0\). One can check that \[ U(x) = \int^x_{\xi} \exp \left(-\int^z_{\xi} r(y)dy \right)dz \] for some \(\xi\), so knowing \(U(\cdot)\) and \(r(\cdot)\) are equivalent.
Two Applications of Utility Functions
Preference Ordering of Random Gains. Consider a decision-maker with initial wealth \(w\) who has the choice between a certain number of random gains. By using a utility function, two random gains can be directly compared: he or she prefers \(G_1\) to \(G_2\), if \[ \mathrm{E}[U(w + G_1)] > \mathrm{E}[U(w + G_2)]. \] Premium Calculation. We consider a company with initial wealth \(w\). The company is to insure a risk and has to pay the total claims \(S\) (a random variable) at the end of the period. The expected utility of wealth with the contract should be equal to the utility without the contract: \[ \mathrm{E}[U(w + P - S)] = U(w). \] This is known as the principle of equivalent utility.
Optimality of a Stop-Loss Contract
We consider a company that has to pay the total amount \(S\) (a random variable) to its policyholders at the end of the year. We compare two reinsurance agreements:
A stop-loss contract with deductible \(d\). Here the reinsurer will pay \[ (S-d)_+ = \left\{ \begin{array}{cl} 0 & \text{if } S<d \\ S-d & \text{if } S > d . \end{array} \right. \]
A general reinsurance contract, given by a function \(h(\cdot)\), where the reinsurer pays \(h(S)\) at the end of the year. The only restriction on the function \(h(\cdot)\) is that \(0 \le h(x) \le x\).
We assume that the two contracts are comparable, in the sense that the expected payments of the reinsurer are the same, that is, that \(\mathrm{E}[(S - d)_+ ] = \mathrm{E}[h(S)].\) Furthermore, we make the convenient assumption that the two reinsurance premiums are the same. Then, in terms of utility, the stop-loss contract is preferable: \[ \mathrm{E}[U(w - S + h(S))] \le \mathrm{E}[U(w - S + (S - d)_+ )]. \] In this context, \(w\) represents the wealth after receipt of the premiums and payment of the reinsurance premiums. For a proof of this result, see H. Gerber and Pafumi (1998) or Proposition 3.2 in Section 3.1.