5.1. Estimation Results
The logic of estimating the parameters using the benchmark model is straightforward, so that one can directly maximize the joint log-likelihood with a numerical routine, for example, the
optim function in
R.
Table 4 shows the estimated values for the binomial thinning model and the benchmark, independent frequency model. The point estimates of
, and
under the two models are similar. Additionally, the standard errors of the mean parameters are smaller compared with the standard errors of dispersion parameters,
r,
,
, and
. It is observed that there is a significant level of improvement in the log-likelihood by incorporating dependence via the common factor. We note that the improvements in AIC and BIC are even greater with the dependence modelling as the binomial thinning model is more parsimonious than the independent model.
As mentioned in the previous section, the joint distribution of a copula-based model combines marginal distributions with a copula function. Here, we use the inference by margin (IFM) method, so that the marginal distributions in the independent model are considered as given, while only the copula part is additionally estimated.
Table 5 shows the estimated copula parameters and the log-likelihood values of each of the copula models. We note that the parameter estimated with the Gaussian copula model implies positive relationships among the three lines. By comparing the log-likelihood, the Gaussian copula outperforms the others.
For the severity components, we use several composite models. For the body part (modelled with a light-tailed distribution), we consider the gamma and exponential distributions. For the tail part (modelled with a heavy-tailed distribution), we use inverse-gamma, Pareto, and lognormal distributions. In the following
Table 6,
Table 7 and
Table 8, the model selection criteria for various composite models are demonstrated, fitted with the building/content/profit severity data, respectively.
In the case of building losses, the gamma and lognormal (
and
) and gamma and Pareto (
and
) distributions are shown to have the best goodness-of-fit. Likewise, we find that gamma and lognormal (
and
) is the best for modelling contents losses, and gamma and Pareto (
and
) fits the profits losses well.
Table 9 shows the point estimates of three composite distributions’ parameters, given the best combinations for each coverage, along with the splicing points distribution and the corresponding weight parameter values based on the parameter estimates. We note that some transformation is required to make the weight parameter meaningful. For example, in the case of building losses, we can interpret
as the proportion of
Y that is from the body part, of the gamma distribution. On the other hand,
of
Y is from the tail part, of the lognormal distribution. Specifically, a larger weight parameter value indicates that the composite model is more heavy-tailed, and vice versa. The splicing point parameter,
u, indicates the change of distribution components. For the building coverage, the splicing parameter is 2.08943, which means that the building losses greater than 2.08943 million Danish Krone are modelled by a lognormal distribution. Additionally, we observe that the losses from the profit line are more heavy-tailed compared with the losses from the other two lines.
5.2. Empirical Findings for Risk Management
In the insurance industry, estimating the risk level for a product or portfolio is critical for determining appropriate levels of the premium and reserve. We recall that
and
mean that the random variable stands for the aggregate loss amount from the
jth line and the aggregate loss for all lines of insurance, as defined by (
1) and (
2). It is also straightforward to see that
due to the additivity of expectation. However, such a property generally does not hold for other types of risk measures, so it is important to properly analyze the risk level of total claims
, rather than summing up the risk level of
, and
. For our risk analysis, we use the following well-known risk measures:
Value at Risk (VaR)—;
Tail Value at Risk (TVaR)—;
Proportional Hazard (PH) Risk Measure (
Wang 1995)—
,
.
Dual Power (DP) Risk Measure—.
While TVaR is not a coherent risk measure unless the underlying distribution is continuous, it is innocuous to assume that the TVaR of
, and
are coherent, as we mainly focus on the tail part, where the claim amounts are for sure strictly positive and the underlying distributions are continuous. We also note that (
Wang 1994) showed that the integration of the transformed distribution is coherent when the transformation is a concave function, so that PH and DP are both coherent.
For comparison of the calculated risk measures under each of the models, we apply a Monte Carlo simulation to numerically evaluate the values of risk measures. More specifically, we simulated 100,000 data points, number of accidents
M, the claim numbers for three business lines
and
, and, subsequently, the claim amounts
and
under each of the model specifications with the estimated parameters shown in
Section 5.1.
The simulation for the independent model is straightforward. Because of the independence, we apply random generations for the negative binomial distributions to simulate claim numbers and for all business lines. Unlike the independent one, the binomial thinning model requires the simulation of the reported number of accidents M. After that, a binomial random generation with size parameters corresponding to the reported claim numbers is applied to get the claim numbers and for three lines of business. The logic for the copula models is similar. We first generate trivariate uniform random numbers from the copula functions. With the generated uniform random numbers, we get the claim numbers and using the inverse of the marginal distributions. Once the claim frequencies and were generated, the severity components are generated, subsequently. For example, if is given, then uniform random numbers are generated times and they are converted to the individual severities via the inverse distribution (or quantile) function of the composite distribution function for building losses. Lastly, these values are summed up as .
Figure 2,
Figure 3 and
Figure 4 show the scatterplots of the combinations of building, content, and profit claims for observed and simulated frequency data. Based on the plots of observed data, there are apparent positive relationships among the marginal frequencies. As we expected, however, the independent model cannot capture such dependent behaviours. In the case of the other models, the binomial thinning model shows a substantial linear relationship between the building and contents claim numbers, which is the most similar to the observed. In the case of
Figure 3, however, the Joe copula best captures the relationship between the building and profit frequencies.
Lastly,
Table 10 shows the approximated risk measures under different models. The independent model reproduces relatively smaller values of VaR and TVaR for the aggregated claims
, whereas the calculated risk measure values for each coverage,
and
are more or less the same, regardless of the chosen model. This is quite natural as, regardless of the (assumed) dependence structure, the marginal distributions for
and
(and, subsequently,
and
) are the same. As a result, the TVaR for
under the independent model is severely underestimated compared to the observed (or empirical) TVaR, while the other dependent models are able to reproduce the empirical TVaR for
with less deviations. It implies that it is required to consider possible dependence among different types of insurance coverage for an effective enterprise risk management purpose.