Next Article in Journal
Demand Subsidy versus Production Regulation: Development of New Energy Vehicles in a Competitive Environment
Next Article in Special Issue
Contributions from Spatial Models to Non-Life Insurance Pricing: An Empirical Application to Water Damage Risk
Previous Article in Journal
The Influence of Transport Link Density on Conductivity If Junctions and/or Links Are Blocked
Previous Article in Special Issue
A Review of Software for Spatial Econometrics in R
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fire Risk Sub-Module Assessment under Solvency II. Calculating the Highest Risk Exposure

by
Elena Badal-Valero
1,*,
Vicente Coll-Serrano
2,* and
Jorge Segura-Gisbert
2
1
Department of Business Finances, School of Economics, University of Valencia, 46022 Valencia, Spain
2
Department of Applied Economics, School of Economics, University of Valencia, 46022 Valencia, Spain
*
Authors to whom correspondence should be addressed.
Mathematics 2021, 9(11), 1279; https://doi.org/10.3390/math9111279
Submission received: 8 May 2021 / Revised: 28 May 2021 / Accepted: 28 May 2021 / Published: 2 June 2021
(This article belongs to the Special Issue Spatial Statistics with Its Application)

Abstract

:
The European Directive 2009/138 of Solvency II requires adopting a new approach based on risk, applying a standard formula as a market proxy in which the risk profile of insurers is fundamental. This study focuses on the fire risk sub-module, framed within the man-made catastrophe risk module, for which the regulations require the calculation of the highest concentration of risks that make up the portfolio of an insurance company within a radius of 200 m. However, the regulations do not indicate a specific methodology. This study proposes a procedure consisting of calculating the cluster with the highest risk and identifying this on a map. The results can be applied immediately by any insurance company, covered under the Solvency II regulations, to determine their maximum exposure to the catastrophic man-made risk of fire, instantly providing them with the necessary input for calibration of the solvency capital requirement.

Graphical Abstract

1. Introduction

The management of an insurance company is based on its capacity, present, and future, to meet its commitments. The randomness of the phenomena to which an insurance company is exposed implies the development of probabilistic models of a stochastic nature for the correct valuation of the phenomena insured. However, traditional rating procedures do not work for catastrophic risks due to the usually very low frequency of loss events. These risks are the ones that the current legislation and the new European regulatory framework [1] known as Solvency II, whose principles revolve around the concept of solvency, try to capture. The Directive is structured around three pillars closely linked to each other and very similar to those harmonized by the banking sector in Basel II/III: risk-based capital requirements (Pillar I), enhanced governance (Pillar II), and increased transparency (Pillar III).
The primary objective of solvency capital requirements is to ensure an optimal capital level that allows insurance companies to face unforeseen losses while ensuring policyholders’ protection. This paper proposes a methodology that allows the calculation of solvency capital related to fires’ concentration risk, framed in man-made risk catastrophe of non-life underwritten insurance.
This risk requires a geospatial analysis of the set of risks based on the highest sum insured in the portfolio and where coverage includes fire or explosion and terrorist attacks. Thus, the regulations establish that every insurance company must geolocate the risks that it has insured in its portfolio to identify those partially or fully located within a 200-m radius of the highest concentration of risk [2] (pp. 85–86). However, this regulation, which is mandatory at the European level, does not specify how the cluster of sums insured are calculated. In the absence of a reference methodology, it is common for insurance companies to outsource the calculation of this sub-module, trying in some way to avoid possible future liabilities derived from a poor calibration of a risk of this type.
The paper is structured as follows. Section 2 relates the capital requirements called for by Solvency II with the capital requirements of the non-life catastrophe risk and the man-made catastrophe risk sub-module. Section 3 presents the proposed methodology for determining the highest concentration of risk for an insurance company. Section 4 demonstrates the methodology´s implementation on a portfolio whose risks have been simulated, and then, the outcomes are discussed. The paper ends with the conclusions.

2. Solvency II

The approval of the European Solvency II Directive 2009/138 was a step forward in revising the previous rules for the valuation of the financial situation of insurance companies, adopting a risk-based approach. Under this new paradigm, insurance companies’ valuation rules are thoroughly reviewed, reflecting a better measurement and management of all the risks to which they are exposed.
Solvency II’s fundamental essence is the establishment of the risk-based capital requirements, including an appropriate capital assessment as a guarantor of policyholder protection. The quantitative requirements applicable to insurance and reinsurance companies, which constitute Pillar I of the Solvency II regime, are based on an economic approach centered on the overall balance sheet. The available financial resources must cover the overall financial requirements. Thus, the solvency capital requirements’ fundamental objective is to ensure an optimum level of capital to enable insurance companies to cope with unforeseen losses while guaranteeing the protection of policyholders. Two types of solvency levels are established: an upper limit, called SCR (Solvency Capital Requirement), and a lower limit called MCR (Minimum Capital Requirement), which determines the minimum level of security or absolute minimum amount of capital below which an entity could not continue to operate [1,2].
For its calculation, a general method is established based on a standard formula that adopts a modular approach to risk aggregation. However, the Directive also establishes that under certain prerequisites and authorization by the supervisory body, the SCR can be obtained by means of an internal model or calibration of specific parameters.
Solvency II establishes that the SCR is calculated considering the effects of diversification, with the regulator assuming that VaR (Value At Risk) is a measure that complies with the subadditivity axiom. It could be thought that the model proposed in the Directive, the standard model, is based on multivariate normality assumptions regarding the random variables representing risks.
The capital requirement for non-life underwriting risk is calculated taking into account the possible combinations and correlations of the following risk groups [2] (pp. 72–87): non-life premium and reserve, non-life catastrophe, and non-life lapse. More specifically, the non-life catastrophe risk sub-module (SCRnlCAT) consists of all of the following sub-modules: the natural catastrophe risk sub-module (SCRnatCAT), non-proportional property reinsurance catastrophe risk (SCRnppropperty) sub-module, the man-made catastrophe risk sub-module (SCRmmCAT), and the sub-module for another non-life catastrophe risk (SCRCATother), as shown in the following expression [2] (pp. 75–76):
S C R nlCAT = ( S C R natCAT + S C R nppropperty ) 2 + S C R 2 mmCAT + S C R 2 CATother

2.1. Man-Made Catastrophe Risk

There are risks that, although they can be classified as catastrophic, are not due to natural phenomena, but are caused by man or by anthropogenic effects. In contrast to the former, these types of risks affect a very specific location, with a significant degree of severity in the case of an extreme event. Figure 1 shows the evolution of catastrophic events that occurred between 1970 and 2018. There were 304 catastrophe events in 2018, and 123 were man-made disasters.
Catastrophic risks arise from extreme or irregular events that are not sufficiently reflected by the mandatory capital of the risk of premiums and reserves. These include risks of natural catastrophes, catastrophic risks in non-proportional reinsurance of property damage, risk of catastrophes caused by man, as well as other catastrophic risks in non-life insurance.
In Spain’s case, given the particularities due to the Insurance Consortium’s protection, an insurance company’s capital requirements will be lower than that of other European countries’ capital requirements. In this sense, the Spanish Insurance Consortium is a unique entity in Europe. In relation to catastrophic risks, its purpose is to indemnify the losses arising from extraordinary events occurring in Spain and affect risks located therein. For these purposes, direct damage to persons and property will be considered, as well as the loss of profits as a consequence thereof. For the risk in question, only those caused violently due to terrorism, rebellion, sedition, riot, and civil commotion will be considered.
However, the Consortium’s consideration as a risk mitigator has not been taken into account when considering the impact in terms of solvency capital requirements in this work. The Spanish Insurance Consortium would only cover a part of the fire risks (related to terrorism), leaving aside other risks subject to this sub-module and on the other hand, considering the standard formula and simplified calculations emanate from a European Directive of obligatory compliance for the different member states.
This paper focuses on the risk of man-made disasters where capital requirement must be calibrated at 99.5% of VaR. Unlike natural catastrophes, where the gross sum insured would be divided among competitors in the insurance market, man-made catastrophic events can only harm one insurer, since the impact is felt in a single location (or by a single policyholder) and its surroundings, with the damage being very concentrated and of a greater severity.
Thus, for the man-made catastrophic risk module, the following risk sub-modules described in the Solvency II Directive regulations must be considered, as well as the corresponding transposition into the legal system of each country: (a) the motor vehicle liability risk sub-module, (b) the marine risk sub-module, (c) the aviation risk sub-module, (d) the fire risk sub-module. (e) the liability risk sub-module, and (f) the credit and suretyship risk sub-module. The capital requirement for the man-made catastrophe risk is equal to the following [2] (pp. 83–84):
S C R m m C A T = i S C R i 2
where the sum includes all sub-modules set out and SCRi denotes the capital requirements for sub-module i.

Fire Risk Sub-Module

Until the advent of the Solvency II regulatory framework, the non-life fire sub-module was not considered, and there was limited knowledge of the protection of each policy by insurers. In this regard, to ensure that all insurance and reinsurance companies are on an equal footing when calculating the solvency capital requirement, a standard formula was determined with the sub-modules covering the different risks and the methods, assumptions, and general parameters to be used for its calculation. Within the underwriting risks, the weight of the catastrophe burden is very significant in the non-life module and has required the creation of specific working groups for its calibration.
Thus, the European legislator’s final approach is the unification of property damage and terrorism coverages through the more significant accumulation of risks reduced by the reinsurance program presented by each insurance company [2].
Fire risk solvency capital requires identifying the set of buildings with the highest concentration sum insured, concerning damages on each building due to fire or explosion, including as a result of terrorist attack. All buildings located, totally or partially, within a radius of 200 m [2] must be determined (pp. 85–86). This is a legal obligation and obligatory compliance, but there are no guidelines for calculating the accumulation of sums insured. Solvency II establishes that this risk must be calculated, but it leaves insurance companies free to carry out the calculation by adopting methodologies that better reflect the underlying risk.
The standard formula does not consider the real risk for the insurance companies according to the type of risks they contract for the calculation of the exposure based on the capital at risk. Any considered risk accumulation will be based on the aggregation of the total sums insured of all the policies subscribed within a radius of 200 m, regardless of the inherent activity they present.
This can be somewhat complicated since, although an accumulation of risks within a radius of 200 m may imply an expected loss, this risk may vary depending on the probability of the extreme event, the radius of impact and its severity, and the type of risks underwritten. The risk typology basis is vital because a severe claim in a location linked to an industrial location does not have the same impact as a home or business claim. The factors that lead to significant exposure to risk are different.
Likewise, under the Solvency II prism, insurance and reinsurance entities have the possibility of using a simplified calculation [1] in a standard formula environment if the nature, volume, and complexity of the risks justify it, under the principle of proportionality of the risks underlying the European Directive itself. In this regard, consideration (15) of the Commission Delegated Regulation (EU) 2019/981 [4] provides for the possibility for an insurance entity to restrict its process of identifying the highest concentration of fire risk to the environment of its most significant fire risk exposures, provided that such an approach is proportionate to the nature, scale, and complexity of the fire risk exposure of the insurance or reinsurance undertakings.
This alternative and simplified calculation approach does not invalidate the methodology proposed in this paper. Thus, the use of the simplified calculation should not lead to an incorrect calculation of the solvency capital requirement that could have a relevant influence on the decision making (as stated in the European Commission Delegated Regulation (EU) 2009/981 of 8 March 2009) after obtaining the highest sum insured per insured risk for industrial, commercial, and residential fire after mitigation of the reinsurance amounts. The calculation of the highest risk cluster within a radius of 200 m implies that it has to be calculated with the proposed methodology. Its calculation may or may not be limited only to the surroundings of the largest fire risk exposure, in which case, the calculation process would be significantly expedited since the number of insured risks included as input in the proposed model would be reduced, resulting in a reduction in the number of possible clusters of the highest risk and of the computation time.
From a simulated base of 1000 sums insured, this paper shows how to calculate the cluster containing the highest risk by geolocation of risks within a radius of 200 m of this point. The application is immediate and requires these calculations to be completed by all European insurance companies that emanate from the Solvency II regulatory framework. The principle of proportionality between entities based on their size and structure must prevail at all times. In this way, it is possible to establish specific reinsurance programs to mitigate solvency capital requirements and determine a risk management mechanism for possible adverse events of catastrophic nature.

3. Determination of the Highest Concentration of Risk for an Insurance Company

3.1. Methodology

For some years, the risk cluster concept has been used in the actuarial field, focusing predominantly on life and accident insurance [5,6,7,8]. In particular, it measures the sum of insured values liable to be affected by the same claim event. Thus, Solvency II requires calculating the capital requirement for fire risk to determine the set of buildings with the largest sum insured partly or fully located within a radius of 200 m [2] (pp. 85–86). However, the regulations do not indicate how to calculate this estimate. In this paper, we address this problem through the following steps:
Step 1. We calculate the distance between each pair of sums insured. This may seem straightforward, but calculating the distance between two geographical points located on the earth’s sphere (longitude and latitude) and not on a plane (Cartesian coordinates) is not so simple. One of the approximations most used in practical applications [9,10,11,12,13,14,15,16] to calculate the distance between two geographical points is the Haversine distance [17], which is given by:
d H a v e r s i n e = 2 · r · a r c s i n ( s i n 2 ( φ 2 φ 1 2 ) + c o s ( φ 1 ) c o s ( φ 2 ) s i n 2 ( λ 2 λ 1 2 ) ) .
where λ 1 and λ 2 are the lengths of points 1 and 2 expressed in radians, respectively; φ 1 and φ 2 are the latitudes of points 1 and 2 expressed in radians, respectively, and r is the radius of the earth (the equatorial radius is 6378.137 km and the mean radius is 6371 km).
The Haversine distance assumes that the earth is spherical when it is slightly ellipsoidal. In this case, Vicenty distance [18] can be used. From the following notation:
a , b = m a j o r   a n d   m i n o r   s e m i a x e s   o f   t h e   e l l i p s o i d f ,   f l a t t e n i n g = ( a b ) / a φ 1   a n d   φ 2 = g e o d e t i c   l a t i t u d e L = d i f f e r e n c e   i n   l o n g i t u d e U 1 = atan ( 1 f ) tan ( φ 1 ) U 2 = atan ( 1 f ) · tan ( φ 2 ) L = λ λ = 2 π ,
the algorithm to calculate the Vincenty distance iteratively evaluate equations from 4 to 12, while abs ( λ λ ) > 10 12 .
sin ( σ ) = ( cos ( U 2 ) sin ( λ ) ) 2 + ( cos ( U 1 ) sin ( U 2 ) sin ( U 1 ) cos ( U 2 ) cos ( λ ) ) 2
cos ( σ ) = sin ( U 1 ) sin ( U 2 ) + cos ( U 1 ) cos ( U 2 ) cos ( λ )
σ = a t a n 2 ( sin ( σ ) , cos ( σ ) )
sin ( α ) = cos ( U 1 ) cos ( U 2 ) sin ( λ ) sin ( α )
c o s 2 ( α ) = 1 s i n 2 ( α )
cos ( 2 σ m ) = cos ( σ ) 2 sin ( U 1 ) sin ( U 2 ) c o s 2 ( α )
C = f 16 c o s 2 ( α ) [ 4 + f ( 4 3 c o s 2 ( α ) ) ]
λ = λ
λ = L + ( 1 C ) f   s i n ( α ) { σ + C   s i n ( α ) [ cos ( 2 σ m ) + C   c o s ( σ ) ( 1 + 2 cos 2 ( 2 σ m ) ) ] } .
Then, when λ converges, the algoritm evaluates:
u 2 = c o s 2 ( α ) a 2 b 2 a 2
A = 1 + u 2 16,384 { 4096 + u 2 [ 768 + u 2 ( 320 175 u 2 ) ] }
B = u 2 1024 { 256 + u 2 [ 128 + u 2 ( 74 47 u 2 ) ] }
Δ σ = B s i n ( σ ) { cos ( 2 σ m ) + B 4 [ cos ( σ ) ( 1 + 2 cos 2 ( 2 σ m ) B 6 cos ( 2 σ m ) ( 3 + 4 s i n 2 ( σ ) ) ( 3 + 4 cos 2 ( 2 σ m ) ) ) ] } .
Thus, the Vincenty distance between two points will be given by Equation (17):
d V i n c e n t y = b A ( σ Δ σ ) .
Although Vincenty distance is quoted as being highly accurate, it is an iterative process and computationally time consuming.
Alternatively, the longitudes and latitudes in decimal degrees can be converted to UTM (Universal Transverse Mercator) coordinate system. From cartesian coordinates, the Euclidean distance between two points can be calculated.
Step 2. We create two matrices (empty), one to record the distance between each pair of geographic points (distance matrix) and the other to record the sum insured by the insurance company if the distance between two policyholders is less than 200 m (risk matrix).
Step 3. The total sum insured is calculated from the risk matrix and considering each policyholder as the centroid of a risk cluster for each cluster. Next, the cluster (policyholders) that represents the highest concentration of risks is identified as well as the set of risks that are part of that cluster.
To work out the highest risk for the insurance company, we have written a function in R (cumulus.R) (a simplified and limited version of this function is found in Appendix C).
The cumulus.R function has 6 arguments. The first argument is the database of the insurance company´s policyholders, and it must be a dataframe or a tibble. The data must necessarily have four columns, in this order: (1) identification of the policyholder, (2) longitude, (3) latitude (in decimal degrees), and (4) total sum insured. If the Euclidean distance is used, it is necessary to calculate the Cartesian coordinates previously. In the dataset, the Cartesian coordinates are located in columns 7 and 8. The second argument (cumulus) refers to the number of clusters to be obtained. By default, the cumulus argument is prefixed to the value 1. The third argument refers to the distance within which the policyholders must be located to calculate the highest risk, i.e., radius. In accordance with the Solvency II Directive, the radius is pre-set at 200 m. The fourth argument (radius) is the additional margin of distance that we want to incorporate to counteract possible inaccuracies in distances’ estimates. Thus, an insurance company may consider it appropriate to extend the radius of the calculation of the highest risk, e.g., incorporating a margin of 20 m, meaning the calculation radius would actually reach 220 m. The fifth argument is extended. This argument takes the value 0 by default, but it must be set to the value 1 if a margin is indicated and if the highest accumulation of risk is maintained at a radius of 200 m or to the value 2 for a new cluster of greater risk when locating those policyholders in the new radius (radius + margin). Lastly, the sixth argument is the type of distance used (by default, the argument is set to the Haversine distance).
With respect to distances, although several R packages allow calculation of the Haversine distance between two geographic points (e.g., the package geosphere (2019) [19]), cumulus refers to the function as hav_distance (see Appendix A). hav_distance is a function that we have written to calculate the Haversine distance to reduce the calculation time. Specifically, with the geosphere distHaversine function, it takes an average of 4.39 min to obtain the matrix of distances between each pair of points. This time is reduced to 38.36 s with the hav_distance function. These average times have been obtained using an Intel Core i5-10210U computer, 1.60 GHz, and 8 GB of RAM. On the other hand, to calculate the Vincenty distance, we call the corresponding function in the geosphere package. This is because although Vincenty distance is quoted as being highly accurate, it is an iterative process and computationally time consuming. We only calculate the Vincenty distance for comparison results. In fact, since the distances that we really need to calculate are relatively small, the ellipsoidal effect of the earth is minimal, so the haversine and Euclidean distances (see Appendix B) are preferred.
In our study, the geographic area established for a cluster is centered on an insurance policy, whereas many clusters can be created as an entity insures policies. The geographical representation of the policies insured by points based on their geolocation coordinates allows evaluating risks over a short distance.

3.2. Dataset

To determine the highest risk cluster that an insurance company has to assume as a previous step for calculating the solvency capital requirement (SCR), we use a dataset of 1000 risks. The sums insured (column 4, in euros) in this dataset come from a random sample of policyholders from a real insurance company. However, their geographic locations (columns 2 and 3, longitude and latitude in decimal degrees) have been simulated in another city (specifically in Valencia, Spain) in order to preserve privacy. Data file (xlsx file) is supplied as the Supplementary Material with this article. The sum insured refers to the risk coverage of both the buildings and content. The distribution of the policyholders can be seen in Figure 2.
The main descriptive statistics of the sum insured are shown in Table 1.

4. Case Study Results and Discussion

To obtain the highest risk for the insurance company, we have to execute the cumulus function, which returns a list of 6 elements as a result of the analysis:
  • data: original dataframe to which a first column (ref) was added to list the policyholders consecutively.
  • distance.matrix: matrix with the selected distance (Haversine by default) for each pair of geographic points.
  • cumulus.matrix: risk matrix, in which each policyholder constitutes the centroid of a risk cluster.
  • maximum.cumulus: amount of the highest risk for the insurance company.
  • identification.cumulus: policyholder representing the centroid of the highest risk cluster.
  • cumulus.data: dataframe made up of the policyholders who form the highest risk cluster.
We begin by analyzing the case in line with regulations: determination of the highest risk of the policyholders located within a radius of 200 m. To do this, we first load the function, source (“cumulus.R”) and then execute the instruction shown below. Note that Harversine distance is calculated by default.
result1 ← cumulus(data).
Depending on the computer, the estimated average execution time is 41.39 s. the main results, it should be noted that the highest risk for the insurance company amounts to 41,431,645 euros. Table 2 shows the results that cumulus provides for the highest risk cluster using the Haversine distance. Column 6 (highest.sum.insured), which is ordered in descending order, represents the highest risk considering each policyholder (id) as the centroid of a cluster. Policyholder 2266 represents the centroid of this risk cluster, which is made up of the remaining 28 policyholders listed in column 2. Column 7 (insured.cluster) represents the number of policyholders that make up each cluster.
The cluster with the highest risk (Table 2) was also identified on the map to ease interpreting the results by managing the insurance company [20] (see Figure 3). The policyholders that form the highest risk cluster appear as blue dots, while the rest of the policyholders appear as black dots. For each point, the Id of the policyholder and their sum insured are shown. A red marker identifies the centroid of the cluster.
With both the Vincenty distance and the Euclidean distance, we obtain the same results, which are slightly different from those obtained with the Haversine distance (see Table 2). Now, policyholder 2266 remains as the centroid of the highest risk cluster (Table 2), but policyholder 1960 is included and the highest risk amounts to 49,925,568 euros. However, the time consumed to obtain the results based on the Vincenty distance was 3.83 min and the Euclidean distance was 38.54 s.
Given that both the policyholders’ geolocation and the calculation of the distance between them are not accurate, the insurance company may decide to incorporate a margin of error to estimate the highest risk. Instead of considering all the policyholders located within a radius of 200 m, the insurer may consider the policyholders located within a radius of, e.g., 220 m. In this case, the margin of error assumed by the insurance company is 20 m. We have to consider the essential role of reinsurance programs in reducing capital charges in hypothetical adverse scenarios of high severity. The migration of reinsurance acts as a guarantor optimizing capital structure and the increase in risk transfers.
Under this assumption, the insurance company can consider two scenarios:
  • Scenario 1. Maintain policyholder 2266 as the centroid of the highest risk cluster and extend the cluster radius by 20 m to identify the policyholders who are within the extended radius. The results of this scenario 1 using the Haversine distance are obtained by executing the following instruction:
result2 ← cumulus(data, margin = 20, extended = 1).
Figure 4 shows the original cluster (in gray) and the new risk cluster (in red); in both cases, the centroid of the cluster is policyholder 2266. By expanding the radius from 200 to 220 m, a total of 4 further policyholders are incorporated into the cluster: 1690, 2538, 2551, and 2593. Under this scenario 1, the highest risk for the insurance company amounts to 44,695,192 euros. The results obtained are the same when the Vincenty or Euclidean distances are used, while the biggest difference is in the time consumed in executing the cumulus function: Haversine (42.75 s), Euclidean (38.82 s), and Vincenty (3.85 min).
  • Scenario 2. Determine a new risk cluster considering all the policyholders who are located in the new radius (200 m plus the margin of error). To obtain the results of this scenario, we execute the following instruction:
    result3 ← cumulus(data, margin = 20, extended = 2).
The results obtained are the same as that of the results obtained using all three distances and are shown in Table 3.
In this new situation, the centroid of the risk cluster is policyholder 667. The risk cluster consists of a total of 34 policyholders. As we can see in Table 3, the highest risk for the insurance company is quantified at 45,090,147 euros. In Figure 5, the new risk cluster (the red circle) and the risk cluster under Scenario 1 (the gray circle) have been identified.

5. Conclusions

This paper presents a methodological proposal to calibrate the catastrophic risk of fire caused by human activity to obtain the highest risk cluster for an insurance company. Although the Solvency II regulations indicate in detail and regularly what risks this sub-module must include, it does not offer any market proxy or frame of reference that serves as a guide for obtaining the necessary input [2] (pp. 85–86). In addition, there is no reference as to how to approach this in the scientific or professional field. Beirut´s recent events should serve as a reminder of the importance of a correct assessment of this risk sub-module.
A simplified version of the R function written to implement the procedure can be found in Appendix C. To illustrate this procedure for obtaining the highest risk cluster, a dataset of 1000 policyholders is used. The calculation proposal presented here constitutes a reference method for all European insurance companies covered by the Solvency II regulatory framework, among which the principle of proportionality must prevail at all times.
According to Solvency II, insurance companies must calculate the capital requirement for fire risk to determine the set of buildings with the largest sum insured partly or fully located within a radius of 200 m.
In the spirit of Solvency II, policyholder 2266 has been identified as the centroid of the highest risk cluster, with the highest sum insured at 41,431,645 euros or 49,925,568 euros, depending on the distance used. In this sense, it is essential to highlight that precision of the highest risk clusters’ estimation depends fundamentally on the correct geolocation of the insured risks and the methodological approach used to calculate the distances. In this paper, three different distances were used: Harversine, Vincenty, and Euclidean distances. However, we need to take into account that the Haversine distance calculates the great circle distance on a sphere, while the Vincenty distance calculates the shortest geodesic distance on a slightly ellipsoidal surface. Consequently, the Haversine distance can be in error up to 0.5%. Since the results obtained in our study with the different distances are quite similar, they are preferable because the Vincenty distance is time consuming.
As a suggestion for future research, one of the lines of improvement in the calibration of this sub-risk should be directed towards making changes in the determination of risk exposure to better adjust the occurrence of the extreme event. The introduction of a corrective factor to the sum insured may be pertinent given that in contrast to calculating hypothetical damages produced by the loss, the actual loss would be more significant at the center of a high risk than at the margins of the 200 m considered.
Finally, in the model proposed in this paper, the origin of the catastrophe has been considered to be located within a radius of 200 m of a policyholder, which is the criterion most commonly accepted by professionals in the sector. However, it is possible that the loss may occur at any other point but affect the company’s policyholders. We are currently working on the issue, but it must be borne in mind that the solution to this problem involves exceptionally high computational requirements, even more so given the large volume of most portfolios of European insurers.

Supplementary Materials

Data file (Excel file) is supplied as supplementary material with this article. Data can also be downloaded from: https://go.uv.es/vcoll/cumulus_data (accessed on 31 May 2021).

Author Contributions

Investigation E.B.-V., V.C.-S., and J.S.-G. All authors contributed equally to this research paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

R code: Haversine distance
hav.distance ← function(lon1, lat1, lon2, lat2){
radius ← 6378137 # ecuatorial radius in meters
rad.lat1 ← lat1 * pi/180 # lat in radians
rad.lat2 ← lat2 * pi/180
dif.lat ← (lat2-lat1) * pi/180 # lon in radians
dif.lon ← (lon2-lon1) * pi/180
a ← sin(dif.lat/2) * sin(dif.lat/2) + cos(rad.lat1) * cos(rad.lat2) * sin(dif.lon/2) * sin(dif.lon/2)
c ← 2 * atan2(sqrt(a), sqrt(1-a))
distance ← radius * c
return(distance)
}

Appendix B

R code: Euclidean distance
euclidean ← function(x1, y1, x2, y2){
distance ← sqrt((x1 - x2)^2 + (y1 - y2)^2)
return(distance)
}

Appendix C

R code: Cumulus function to calculate the highest risk
cumulus ← function(x, cumulus = 1, radius = 200, margin = 0, extended = 0, distance = “Haversine”){
 source(“hav.distance.R”)
 source(“euclidean.R”)
 ref ← 1:nrow(x)
 data ← cbind(ref,x)
 cumulus.matrix ← matrix(0,nrow= nrow(data),ncol=nrow(data))
 distance.matrix ← matrix(0,nrow= nrow(data),ncol=nrow(data))
 dimnames(cumulus.matrix) ← list(data[,2],data[,2])
 dimnames(distance.matrix) ← list(data[,2],data[,2])
 radius2 ← radius + margin
 for(i in 1:nrow(data)){
  
  for(j in 1:nrow(data)){
   
   if(extended == 0 | extended == 1) {
    
    if(distance == “Haversine”){
    calculated.distance ← hav.distance(data[i,3],data[i,4],data[j,3],data[j,4])
    
    if(calculated.distance <= radius){
     
     cumulus.matrix[i,j] ← data[5][j,]
     distance.matrix[i,j] ← calculated.distance
     
    } else{
     
     cumulus.matrix[i,j] ← 0
     distance.matrix[i,j] ← calculated.distance
     
    }
    }
    
    if(distance == "Vincenty"){
     
     calculated.distance ← geosphere::distVincentyEllipsoid(c(data[i,3],data[i,4]),c(data[j,3],data[j,4]))
     
     if(calculated.distance <= radius){
      
      cumulus.matrix[i,j] ← data[5][j,]
      distance.matrix[i,j] ← calculated.distance
      
     } else{
      
      cumulus.matrix[i,j] ← 0
      distance.matrix[i,j] ← calculated.distance
      
     }
    }
    
    
    if(distance == “Euclidean”){
     
     calculated.distance ← euclidean(data[i,8],data[i,9],data[j,8],data[j,9])
     
     if(calculated.distance <= radius){
      
      cumulus.matrix[i,j] ← data[5][j,]
      distance.matrix[i,j] ← calculated.distance
      
     } else{
      
      cumulus.matrix[i,j] ← 0
      distance.matrix[i,j] ← calculated.distance
      
     } 
     
    }
     
   }
   
   if(extended == 2) {
    
    if(distance == “Haversine”){
     
     calculated.distance ← hav.distance(data[i,3],data[i,4],data[j,3],data[j,4])
    
    if(calculated.distance <= radius2){
     
     cumulus.matrix[i,j] ← data[5][j,]
     distance.matrix[i,j] ← calculated.distance
     
    } else{
     
     cumulus.matrix[i,j] ← 0
     distance.matrix[i,j] ← calculated.distance
     
    }
    }
    
    if(distance == “Vincenty”){
     
     calculated.distance ← geosphere::distVincentyEllipsoid(c(data[i,3],data[i,4]),c(data[j,3],data[j,4]))
     
     if(calculated.distance <= radius2){
      
      cumulus.matrix[i,j] ← data[5][j,]
      distance.matrix[i,j] ← calculated.distance
      
     } else{
      
      cumulus.matrix[i,j] ← 0
      distance.matrix[i,j] ← calculated.distance
     }
    }
    
    if(distance == “Euclidean”){
     
     calculated.distance ← euclidean(data[i,8],data[i,9],data[j,8],data[j,9])
     
     if(calculated.distance <= radius2){
      
      cumulus.matrix[i,j] ← data[5][j,]
      distance.matrix[i,j] ← calculated.distance
      
     } else{
      
      cumulus.matrix[i,j] ← 0
      distance.matrix[i,j] ← calculated.distance
      
     } 
     
    }
    
   }
  }
 }
 df.cumul ← as.data.frame(cumulus.matrix)
 names(df.cumul) ← data[,1]
 df.cumul$cumul.sum ← apply(df.cumul,1,sum)
 df.cumul$insured ← nrow(df.cumul) - apply(df.cumul==0,1,sum) # numero de asegurados en un cúmulo
 highest.risk ← max(df.cumul$cumul.sum)
 # añadimos esta información al data
 data$highest.sum.insured ← df.cumul$cumul.sum
 data$insured.cluster ← df.cumul$insured
 # selección de número de cúmulos
 select.main.cumulus ← cumulus
 list.cumulus ← sort(df.cumul$cumul.sum, index.return=TRUE, decreasing=TRUE)
 selected.list.cumulus ← lapply(list.cumulus, ‘[‘, list.cumulus$x %in% head(list.cumulus$x,select.main.cumulus))
 if(length(selected.list.cumulus$x) > select.main.cumulus){ # si hay cúmulo que coinciden
  
  selected.cumulus ← length(selected.list.cumulus$x)
  
 } else {
  
  selected.cumulus ← select.main.cumulus
  
 }
 identification.cumulus ← head(selected.list.cumulus$ix,selected.cumulus)
 names.identification.cumulus ← data[identification.cumulus,2]
 cumulus.values ← df.cumul[identification.cumulus,]
 cumulus.insured ← simplify2array(
  apply(cumulus.values[1:nrow(data)], 1,
     function(x) paste(names(cumulus.values[1:nrow(data)])[x != 0], collapse = " "))
 ) %>%
  as.data.frame()
 names(cumulus.insured) ← “insured”
 cumulus.insured ← cumulus.insured %>%
  separate(“insured”,paste(1:max(cumulus.values$insured),sep=“”),sep = “ ”)
 cumulus.insured ← cumulus.insured[!is.na(cumulus.insured)] %>%
  as.numeric() %>%
  sort() %>%
  unique()
 # identificación del origen del cúmulo
 cumulus.data ← data[cumulus.insured,] %>%
  arrange(desc(highest.sum.insured))
 ### extended cluster
 if(extended == 1 ){
  
  identification.cumulus ← identification.cumulus
  extended.cumulus ← matrix(0,nrow= length(identification.cumulus),ncol=nrow(data))
  
  for(i in 1:length(identification.cumulus)){
   
   for(j in 1:nrow(data)){
    
    if(distance == "Haversine"){
     
    if(hav.distance(data[identification.cumulus,3],data[identification.cumulus,4],data[j,3],data[j,4]) <= radius2){
     
     extended.cumulus[i,j] ← data[j,5]
     
    } else{
     
     extended.cumulus[i,j] ← 0
     
    }
    }
    
    if(distance == “Vincenty”){
     
     if(geosphere::distVincentyEllipsoid(c(data[identification.cumulus,3],data[identification.cumulus,4]),c(data[j,3],data[j,4])) <= radius2){
  
      extended.cumulus[i,j] ← data[j,5]
      
     } else{
      
      extended.cumulus[i,j] ← 0
      
     }
    }
    
    
    if(distance == "Euclidean"){
     
     
if(euclidean(data[identification.cumulus,8],data[identification.cumulus,9],data[j,8],data[j,9]) <= radius2){
      
      extended.cumulus[i,j] ← data[j,5]
      
     } else{
      
      extended.cumulus[i,j] ← 0
      
     }
     
    }
   }
  }
  
  extended.cumulus ← as.data.frame(extended.cumulus)
  names(extended.cumulus) ← data[,2]
  row.names(extended.cumulus) ← identification.cumulus
  insured.extended.cumulus ← which(extended.cumulus !=0)
  
  highest.risk ← apply(extended.cumulus,1,sum)
  cumulus.data ← data[insured.extended.cumulus,] %>%
   arrange(desc(highest.sum.insured))
 }  
 cumulusOutput ← list(data = data,
            distance.matrix = distance.matrix,
            cumulus.matrix = cumulus.matrix,
            highest.risk = highest.risk,
            identificaction.cumulus = names.identification.cumulus,
            cumulus.data = cumulus.data)
 return(cumulusOutput)
}

References

  1. European Parliament of the Council. Directive 2009/138/EC of the European Parliament and of the Council of 25 November 2009 on the Taking-Up and Pursuit of the Business of Insurance and Reinsurance (Solvency II). 2009. Available online: https://eur-lex.europa.eu/legal-content/en/TXT/?uri=CELEX:32009L0138 (accessed on 1 June 2021).
  2. European Parliament of the Council. Commission Delegated Regulation (EU) 2015/35 of 10 October 2014 Supplementing Directive 2009/138/EC of the European Parliament and of the Council on the Taking-Up and Pursuit of the Business of Insurance and Reinsurance (Solvency II). 2015. Available online: https://eur-lex.europa.eu/eli/reg_del/2015/35/oj (accessed on 1 June 2021).
  3. SWISS RE. Sigma Nº2/2019: Natural Catastrophes and Man-Made Disasters in 2018: “Secondary” Perils on the Frontline. 2019. Available online: https://www.swissre.com/dam/jcr:c37eb0e4-c0b9-4a9f-9954-3d0bb4339bfd/sigma2_2019_en.pdf (accessed on 1 June 2021).
  4. European Parliament of the Council. Commission Delegated Regulation (EU) 2019/981 of 8 March 2019 Amending Delegated Regulation (EU) 2015/35 Supplementing Directive 2009/138/EC on the Taking-Up and Pursuit of the Business of Insurance and Reinsurance (Solvency II). 2019. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32019R0981 (accessed on 1 June 2021).
  5. Yeo, A.C.; Smith, K.A.; Willis, R.J.; Brooks, M. Clustering technique for risk classification and prediction of claim costs in the automobile insurance industry. Intell. Syst. Acc. Financ. Manag. 2011, 10, 39–50. [Google Scholar] [CrossRef]
  6. Abolmakarem, S.; Abdi, F.; Khalili-Damghani, K. Insurance customer segmentation using clustering approach. Int. J. Knowl. Eng. Data Min. 2016, 4, 18–39. [Google Scholar] [CrossRef]
  7. Zhuang, K.; Wu, S.; Gao, X. Auto Insurance Business Analytics Approach for Customer Segmentation Using Multiple Mixed-Type Data Clustering Algorithms. Teh. Vjesn. 2011, 25, 1783–1791. [Google Scholar]
  8. Hidalgo, H.; Chipulu, M.; Ojiako, U. Risk segmentation in Chilean social health insurance. Int. J. Health Care Qual. Assur. 2013, 26, 666–681. [Google Scholar] [CrossRef] [PubMed]
  9. Alkan, H.; Celebi, H. The Implementation of Positioning System with Trilateration of Haversine Distance. In Proceedings of the 30th Annual International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC), Istanbul, Turkey, 8–11 September 2019; pp. 1–6. Available online: https://ieeexplore.ieee.org/abstract/document/8904289 (accessed on 1 June 2021).
  10. Alam, C.N.; Manaf, K.; Atmadja, A.R.; Aurum, D.K. Implementation of Haversine Formula for Counting Event Visitor in the Radius Based on Android Application. In Proceedings of the 2016 4th International Conference on Cyber and IT Service Management IEEE, Bandung, Indonesia, 26–27 April 2016; pp. 1–6. [Google Scholar]
  11. Basyir, M.; Nasir, M.; Suryati, S.; Mellyssa, W. Determination of nearest emergency service office using haversine formula based on android platform. Emit. Int. J. Eng. Technol. 2017, 5, 270–278. [Google Scholar] [CrossRef]
  12. Cho, J.H.; Baik, E.Y. Geo-spatial Analysis of the Seoul Subway Station Areas Using the Haversine Distance and the Azimuth Angle Formulas. J. Inf. Technol. Serv. 2018, 12, 139–150. [Google Scholar]
  13. Dauni, P.; Firdaus, M.D.; Asfariani, R.; Saputra, M.I.N.; Hidayat, A.A.; Zulfikar, W.B. Implementation of Haversine Formula for School Location Tracking. J. Phys. Conf. Ser. 2017, 1402, 077028. [Google Scholar] [CrossRef]
  14. Hartanto, A.D.; Susanto, M.R.; Ilham, H.D.; Retnaningsih, R.; Nurdiyanto, H. Mobile Technologies of Formulation Haversine Application and Location Based Service. In Proceedings of the Joint Workshop KO2PI and the 1st International Conference on Advance & Scientific Innovation, Medan, Indonesia, 23–24 April 2018. [Google Scholar]
  15. Huang, J.; Huang, H. Looking around in the Neighbourhood: Location Estimation of Outdoor Urban Images. IET Image Process. 2021, 1–13. [Google Scholar] [CrossRef]
  16. Van Kampen, J.; Knapen, L.; Pauwels, E.; Van der Mei, R.; Dugundji, E.R. Bicycle Parking in Station Areas in the Netherlands. Procedia Comput. Sci. 2021, 184, 338–345. [Google Scholar] [CrossRef]
  17. Sinnott, R.W. Virtues of the Haversine. Sky Telesc. 1984, 68, 159. [Google Scholar]
  18. Vincenty, T. Direct and inverse solutions of geodesics on the ellipsoid with application of nested equations. Surv. Rev. 1975, 23, 88–93. Available online: https://www.ngs.noaa.gov/PUBS_LIB/inverse.pdf (accessed on 1 June 2021). [CrossRef]
  19. Hijmans, R.J. Geosphere: Spherical Trigonometry. R Package Version 1.5-10. 2019. Available online: https://CRAN.R-project.org/package=geosphere (accessed on 1 June 2021).
  20. Che, J.; Karambelkar, B.; Xie, Y. Leaflet: Create Interactive Web Maps with the JavaScript ’Leaflet’. R package version 2.0.3. 2019. Available online: https://CRAN.R-project.org/package=leaflet (accessed on 1 June 2021).
Figure 1. Number of catastrophe events, 1970–2018; Swiss RE (2019) [3].
Figure 1. Number of catastrophe events, 1970–2018; Swiss RE (2019) [3].
Mathematics 09 01279 g001
Figure 2. Simulated data: geolocation of the policyholders of the insurance company. Interactive map available at: https://go.uv.es/vcoll/cumulus_1 (accessed on 31 May 2021).
Figure 2. Simulated data: geolocation of the policyholders of the insurance company. Interactive map available at: https://go.uv.es/vcoll/cumulus_1 (accessed on 31 May 2021).
Mathematics 09 01279 g002
Figure 3. Higher risk cluster. The policyholders located within a 200 m radius (Haversine distance is used). Interactive map available at: https://go.uv.es/vcoll/cumulus_2 (accessed on 31 May 2021).
Figure 3. Higher risk cluster. The policyholders located within a 200 m radius (Haversine distance is used). Interactive map available at: https://go.uv.es/vcoll/cumulus_2 (accessed on 31 May 2021).
Mathematics 09 01279 g003
Figure 4. The highest risk cluster considering the original risk centroid (gray circle, radius = 200 m) and expanding the radius by 20 m (red circle, margin = 20 m). Haversine distance is used. Interactive map available at: https://go.uv.es/vcoll/cumulus_3 (accessed on 31 May 2021).
Figure 4. The highest risk cluster considering the original risk centroid (gray circle, radius = 200 m) and expanding the radius by 20 m (red circle, margin = 20 m). Haversine distance is used. Interactive map available at: https://go.uv.es/vcoll/cumulus_3 (accessed on 31 May 2021).
Mathematics 09 01279 g004
Figure 5. The highest risk cluster for the insurance company considering a radius of 220 m (radius = 200 m, margin = 20 m). Haversine distance is used. Interactive map available at: https://go.uv.es/vcoll/cumulus_4 (accessed on 31 May 2021).
Figure 5. The highest risk cluster for the insurance company considering a radius of 220 m (radius = 200 m, margin = 20 m). Haversine distance is used. Interactive map available at: https://go.uv.es/vcoll/cumulus_4 (accessed on 31 May 2021).
Mathematics 09 01279 g005
Table 1. Descriptives of the sum insured.
Table 1. Descriptives of the sum insured.
StatisticSum Insured (€)
Minimum34,099.73
First quartile194,856.66
Median461,285.27
Third quartile1,120,281.02
Mean1,249,490.66
Maximum12,388,953.79
Standard deviation1,639,387.156
Source: compiled by the authors.
Table 2. The highest risk cluster for the insurance entity.
Table 2. The highest risk cluster for the insurance entity.
RefidLongitudeLatitudeSum.InsuredHighest.Sum.InsuredInsured.Cluster
2312266−0.374540339.47245323,603,53941,431,64529
8992940−0.374550939.47236781,467,11641,242,62129
405667−0.374396339.4721817295,88840,118,95227
9291552−0.374892639.4713922255,45532,632,38424
7171394−0.374331939.47183961,979,42832,484,72122
5101517−0.375196339.4735817565,81731,562,82624
376494−0.374739939.4709996117,54330,211,50922
635685−0.374899839.4724242134,52729,676,49427
762878−0.373171339.4720993372,65229,629,74924
2283067−0.372371539.47254793,608,69828,545,75117
723111−0.374759139.4725809552,03028,406,18126
8902927−0.372926239.47125789,610,40927,250,42820
2131899−0.374440339.4732532154,92326,992,85222
9842453−0.376129439.472035696,18026,919,38626
4132998−0.376035239.4727656442,31226,690,70826
752642−0.372271339.4722993130,27926,005,06917
365524−0.376434239.47211136,467,36326,004,92426
3342656−0.373905939.4738459250,59125,825,80622
6962640−0.373905939.47384594,211,25425,825,80622
22511−0.376736239.4720473999,98325,294,01227
5311831−0.375051339.47361091,980,90925,095,46323
101480−0.375598439.4711045104,03223,984,46125
8552638−0.373192739.4734223829,23823,877,07521
5372725−0.374218739.47081881,265,50723,729,71819
1202926−0.374093339.471072691,24423,719,72419
7902910−0.37306439.4737415123,68623,446,66019
422985−0.375801439.4715086246,71822,861,34824
179712−0.374240339.4738532191,37722,659,42022
6411352−0.37260639.4733964682,94719,410,81316
Source: compiled by the authors.
Table 3. The highest risk cluster for the insurance entity. Scenario 2 (Haversine distance).
Table 3. The highest risk cluster for the insurance entity. Scenario 2 (Haversine distance).
RefidLongitudeLatitudeSum.InsuredHighest.Sum.InsuredInsured.Cluster
405667−0.374396339.4721817295,88845,090,14734
723111−0.374759139.4725809552,03044,710,17734
2312266−0.374540339.47245323,603,53944,695,19233
8992940−0.374550939.47236781,467,11644,695,19233
635685−0.374899839.4724242134,52744,214,27933
7171394−0.374331939.47183961,979,42840,521,60630
5101517−0.375196339.4735817565,81737,391,76529
2131899−0.374440339.4732532154,92336,989,62930
5311831−0.375051339.47361091,980,90934,662,86527
762878−0.373171339.4720993372,65234,480,80128
376494−0.374739939.4709996117,54333,791,78726
2283067−0.372371539.47254793,608,69833,223,48826
9291552−0.374892639.4713922255,45533,064,17627
4132998−0.376035239.4727656442,31232,482,61631
5022593−0.372078339.472529497,24831,950,59922
365524−0.376434239.47211136,467,36330,998,18235
9842453−0.376129439.472035696,18030,107,16932
22511−0.376736239.4720473999,98329,944,75235
179712−0.374240339.4738532191,37729,548,61327
392538−0.37685339.472110,13329,378,93534
7952551−0.37685339.4722,162,24329,378,93534
752642−0.372271339.4722993130,27928,919,79822
8902927−0.372926239.47125789,610,40928,467,01023
101480−0.375598439.4711045104,03227,151,36427
3342656−0.373905939.4738459250,59127,138,01825
6962640−0.373905939.47384594,211,25427,138,01825
8552638−0.373192739.4734223829,23825,856,50322
5372725−0.374218739.47081881,265,50725,856,38023
422985−0.375801439.4715086246,71825,804,87126
1202926−0.374093339.471072691,24425,108,20722
2421690−0.373845839.4707353493,92324,698,91521
7902910−0.37306439.4737415123,68623,877,07521
6411352−0.37260639.4733964682,94723,706,75620
1081075−0.373015539.4706526394,95521,838,16518
Source: compiled by the authors.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Badal-Valero, E.; Coll-Serrano, V.; Segura-Gisbert, J. Fire Risk Sub-Module Assessment under Solvency II. Calculating the Highest Risk Exposure. Mathematics 2021, 9, 1279. https://doi.org/10.3390/math9111279

AMA Style

Badal-Valero E, Coll-Serrano V, Segura-Gisbert J. Fire Risk Sub-Module Assessment under Solvency II. Calculating the Highest Risk Exposure. Mathematics. 2021; 9(11):1279. https://doi.org/10.3390/math9111279

Chicago/Turabian Style

Badal-Valero, Elena, Vicente Coll-Serrano, and Jorge Segura-Gisbert. 2021. "Fire Risk Sub-Module Assessment under Solvency II. Calculating the Highest Risk Exposure" Mathematics 9, no. 11: 1279. https://doi.org/10.3390/math9111279

APA Style

Badal-Valero, E., Coll-Serrano, V., & Segura-Gisbert, J. (2021). Fire Risk Sub-Module Assessment under Solvency II. Calculating the Highest Risk Exposure. Mathematics, 9(11), 1279. https://doi.org/10.3390/math9111279

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop