Modeling Cyber Loss Severity Using a Spliced Regression Distribution with Mixture Components
Meng Sun
Simon Fraser University, Burnaby, Canada.
DOI: 10.4236/ojs.2023.134021   PDF    HTML   XML   143 Downloads   673 Views  

Abstract

Cyber losses in terms of number of records breached under cyber incidents commonly feature a significant portion of zeros, specific characteristics of mid-range losses and large losses, which make it hard to model the whole range of the losses using a standard loss distribution. We tackle this modeling problem by proposing a three-component spliced regression model that can simultaneously model zeros, moderate and large losses and consider heterogeneous effects in mixture components. To apply our proposed model to Privacy Right Clearinghouse (PRC) data breach chronology, we segment geographical groups using unsupervised cluster analysis, and utilize a covariate-dependent probability to model zero losses, finite mixture distributions for moderate body and an extreme value distribution for large losses capturing the heavy-tailed nature of the loss data. Parameters and coefficients are estimated using the Expectation-Maximization (EM) algorithm. Combining with our frequency model (generalized linear mixed model) for data breaches, aggregate loss distributions are investigated and applications on cyber insurance pricing and risk management are discussed.

Share and Cite:

Sun, M. (2023) Modeling Cyber Loss Severity Using a Spliced Regression Distribution with Mixture Components. Open Journal of Statistics, 13, 425-452. doi: 10.4236/ojs.2023.134021.

1. Introduction

According to Gartner [1] , the global market for information security would reach $170.4 billion in 2022. Given the potential economic impact of a successful large-scale cyber attack, cybersecurity risks remain the second-most important emerging issue risk experts highlight [2] . Beyond traditional technologies, a larger set of risks at the intersection of technology and society is rapidly emerging. AI and big data have led emerging technologies, transforming economic and social structures. The full-scale implications of cyber threats are yet to be experienced, especially since technology is rapidly evolving. COVID-19 has compelled businesses to establish remote workforces and utilize cloud-based platforms. Due to the pandemic, remote work and digital transformation increased the average total cost of a data breach. The FBI reports a 300% increase in reported cybercrimes since the pandemic began. According to IBM [3] , data breach costs increased from $3.86 million to $4.24 million in 2021, the highest average total cost in the report’s history. Attackers stole $121,000 in bitcoin through nearly 300 transactions due to a Twitter breach that affected 130 accounts [4] resulting in attackers swindling. A security breach disclosed by Marriott compromised the data of more than 5.2 million hotel guests [5] . The Equifax Data Breach, which cost over $4 billion in total, affected 147.9 million consumers [6] . The increasing number of large-scale, widely publicized security breaches suggests that the number of security breaches and their severities is increasing. In 2021, the average cost of a breach will increase by 10%, necessitating quantifying its effects. Regulations and best practices in cyber security hygiene and risk management are changing due to the frequency and severity of cyberattacks. Prominent legislations like the European Union’s 2018 General Data Protection Regulation (GDPR)1, California’s 2020 California Consumer Privacy Act (CCPA)2 and Illinois’s 2018 Biometric Information Privacy Act (BIPA)3 have been passed to enforce severe consequences. To collect, store, process and transfer consumer data, these regulations all have one thing in common: they require organizations to adhere to specific standards.

In addition to reducing vulnerable exposure and increasing technology defence investment, cyber insurance is a fundamental and wildly applicable tool for organizations to maintain their enterprise solvency in light of the rise in cybersecurity threats. Cyber insurance is a type of insurance intended to product against the financial costs associated with the failure or compromise of an organization’s information system [7] . Cyber events include a hacking attack by an external party or malware infection, fraud involving debit and credit cards, and the unintentional disclosure of electronic records due to human error. Most cyber insurance providers offer a core set of coverages and various supplement coverages. With the cyber risk insurance market is at an inflection point, it provides an opportunity to embrace a paradigm shift. To safeguard its profitability, the cyber insurance market took four deliberate measures to combat rising loss ratios [8] : cyber premiums increased across the board, regardless of the industry sector or organization size; many carriers imposed sub-limits and coinsurance provisions specific to ransomware claims; carriers wanted to limit their exposure by limiting capacity; and almost all carries requested more information regarding data security control efforts. An insurance company cannot only rely solely on standard actuarial methods when modeling cyber risk and pricing cyber insurance products. Given how difficult it is to quantify this emerging and evolving risk, new methods and techniques must be developed. In order to identify risk characteristics and assess cyber risk severity, this paper takes an innovated rather than an empirical approach.

A growing number of disciplines are researching Cyber risk. However, acturial risk management is hindered by the need for mature predictive analysis approaches for quantifying and predicting risk severity. We review quantitative research works in actuarial science and describe several research works that focus on loss severity modeling and predictive analysis. [9] combines regression models based on the class of Generalized Additive Models for Location Shape and Scale (GLMLSS), which permits parameters in both the severity and frequency distributions, and a class of ordinal regressions. [10] model hacking data breaches frequency using a hurdle Poisson model and severity using a non-parametric generalized Pareto distribution (GPD). [11] particularly focus mainly on severe claims by combining a generalized Pareto modelling and a regression tree approach for severity analysis. Most of these methods pay special attention to large claims with heavy tail distributions. Traditional actuarial modelling techniques for heavy-tailed insurance loss data concentrate on simple models based on a single parametric distribution that adapts the tail well, such as generalized linear models (GLMs), regression models and quantile regression [12] . Because these techniques are based on a single distribution, they may not be applicable when the behaviour of the tail is inconsistent with the behaviour of the entire loss distribution. It is well known that the actuarial loss distribution is strongly skewed with heavy tails and consists of small, medium and large claims that are difficult to fit with a single parametric distribution. The Extreme Value Theory (EVT) approach, which employs GPD to model excesses over a high threshold ( [13] and [14] ), gained popularity when dealing with heavy-tailed and high losses data. However, they cannot capture the characteristics across the entire loss distribution range making them unsuitable for use as a global fit distribution [15] . In order to model the complete loss distribution, it is frequently necessary to obtain a global fit for the distribution of losses by splicing [16] several distributions in order to model the complete loss distribution. Several actuarial works proposed splicing models for the application of risk measures. For financial risk analysis, [17] suggest a splicing model with a mixed Erlang (ME) distribution for the body and a Pareto distribution for the tail. [18] suggests a three-component spliced regression model for fitting insurance loss data and demonstrate that spliced results outperform the Tweedie loss model regarding tail fitting and prediction accuracy.

The risk portfolio typically contains unobserved heterogeneity in terms of claim severity, such as workers’ compensation and cyber risk data. Given this reality, researchers typically employ a mixture approach to capture the multimodality of the observed loss distribution. [19] designs an optimal Bonus-Malus system in automobile insurance using finite mixture models. [20] models an actual motor insurance claim data set using a mixture Lognormal distribution. [21] proposes a finite mixture of skew-normal distributions that better describes insurance data. [22] suggests a different method for modelling mixture data with heavy tails and skewness in insurance loss distribution that exhibit multimodality. [23] proposes an Erlang loss model using a generalized expectation-maximization (GEM) and clustered method of moments (CMM) algorithm to fit insurance loss data and calculate quantities of interest for insurance risk mixture management. Finally, [24] propose a class of logit-weighted reduced mixtures of experts (LRMoE) models for multivariate claim frequencies or severity distributions and perform the estimation and application to correlated claim frequencies [25] .

Upon the above literature review, all the studies on cyber loss severity do not consider excess of zeros loss, spliced composites and mixture models under a global distribution with corresponding sets of covariates. Motivated by cyber risk specific nature, our study aims to fill these gaps using a finite mixture model (FMM) under a non-linear regression framework and a three-component splicing model with a zero-inflated component. We provide four significant contributions overall.

First, we adapt the methodology by providing a flexible mixture distribution model. Instead of modelling the univariate distributions of all mixture components within the same parametric distribution family, we propose a different method that combines different types of distributions such as Gamma, Log-Normal, Weibull, Burr, Inverse Gaussian and Pareto within a single FMM frame. The aim is to determine the degree to which wide cyber loss range of severity distribution and the exhibition of heavy-tailed and skewed. The model’s compositional distributions make it a valuable, understandable modelling tool for various risks with heterogeneous performance.

Second, in addition to developing components from parametric non-Gaussian families of distributions, we incorporate FMM into a generalized linear model (GLM) to fully utilize the risk characteristics by treating them as covariates within the regression framework. Traditional actuarial loss distribution analysis examines the loss distribution but rarely considers the impact of individual risk characteristics. Moreover, a generalized linear regression mixture model rarely relates a dependent variable to a set of explanatory variables. Our model divides the unobserved mixture probabilities of observations into subgroups and simultaneously estimates a GLM model for each subgroup.

Third, we built an FMM model together with a zero-inflated regression component for our continuous data type. This adaptable strategy enables using covariates to model both the non-zero mixture distribution and the rate of point mass zero. We demonstrate the adaptability of our method by applying continuous mixture model components to the zero-inflation component, in contrast to zero-inflated discrete models such as Poisson and Hurdle. Then we create a zero-inflated mixture model with a complete cumulative distribution function by adjusting the mixture proportions following logistic odds.

Last but not least, our work enables cyber risks to be completely quantified under one distribution taking into consider its extreme loss heavy tail. Cyber risk loss exposures permeate every facet of an organization’s operations, making the consequences of a data breach potentially catastrophic. Unlike other kinds of property and casualty risk that capping incurred loss at a 95% level could effectively rule out extreme value, cyber risk has the nature that, even upon logarithm, loss distribution is too heavily skewed to be capped at a bell shape distribution. Traditional insurance pricing set up a policy limit and doesn’t consider extreme loss when training the model. However, this technique can’t be applied to cyber risk as there is no such limit can be set so that cyber loss could be modeled via one single distribution. We brought up a more statistically rigorous attempt to incorporate excess zero, mixture components and heavy tail of cyber risk in a single, statistically consistent step where other estimation process, such as covariates dependence, is also going on.

This paper presents a series of finite mixture regression (FMR) models and discusses their application in cyber risk estimation. In Section 2, we introduce our PRC dataset and conduct cluster analysis on geographical information. Section 3 reviews the definition and composition of FMR models and propose our unique FMR model adjusted by zero-inflated component based on dataset. Next, we introduce the expectation-maximization (EM) algorithm used to estimate coefficients and model parameters in Section 4. Followed by details on how to fit and choose from among these models as well as information about how to assess the goodness of fit of a model in Section 5. Then, we combine the proposed severity model together with our previous frequency model to simulate aggregate cyber loss over a future time interval in Section 6. Finally, we discuss a model application from the insurers’ perspective and suggest rate filing and future discussions in Section 7.

2. Chronology of Data Breaches from PRC Dataset

In this section, we perform an empirical data analysis which supports and motivates our data-driven modeling approach and further analysis and application. Several necessary initialization procedures must be investigated. Starting with the explanatory data analysis, we investigate unique features of the dataset through an empirical data analysis in the Section 2.1, followed by a cluster analysis to study a multidimensional location feature of the dataset in Section 2.2.

2.1. Empirical Data Analysis

The dataset serves as a resource for researchers examining the effect of data breaches on the performance of insurance companies. It encourages research on the loss prediction and premium determination. Our previous work discussed several frequently considered databases from nonprofit corporations. It used a generalized linear mixed model (GLMM) [26] to examine a quarterly frequency modelling approach.

In this paper, we consider a publicly available cyber security dataset, Privacy Rights Clearinghouse (PRC)4 dataset. Most of the PRC data comes from state attorneys general and the U.S. Department of Health and Human Services. This dataset contains the data breach incidents as well as the number of records breached due to breach incidents. As a sample of the choronology shown in Figure 1, the type of cyber event and the victim’s information, such as the company’s name, type of business, and location, were gathered from breach incidents. Because it contains risk-related characteristics that can be utilized as rating factors, this information is essential for filing insurance rates. In [26] , a generalized linear mixed model (GLMM) is proposed to study the quarterly frequency (number of incidents) of the data breaches recorded in this same PRC database and its application to the cyber insurance is discussed. In this study, we are interested in the number of records breached affected by each recorded data breach incident, which is considered as the severity of the breach caused by cyber breach incidents and collected in PRC database. We late convert the breached data record to dollar amount loss in order to get a dollar amount magnitude.

We illustrate in Section 6 an application of our proposed severity methodology in examining aggregate cyber losses by combining the frequency modeling approach proposed in [26] based on the same dataset.

As discussed in the previous section, fitting an adequate loss distribution to the cyber breach data set is difficult due to its nature. Here, we conduct an empirical data analysis of related target and explanatory variables on the PRC data set to demonstrate the necessity of addressing/accounting for several risk features. Our work is based on the most recent download of the PRC data breach chronology, including 9012 data breach incidents observed in the United States. After removing incomplete and inconsistent observations, 8095 incidents from 2001 to 2019 are investigated and modeled. A summary statistics of this data set

Table 1. Sample of PRC Chronology.

is provided in Table 1.

The first row of Table 2 provides summary statistics for the target variable, the “number of records” breached for a total number of 8095 data breach incidents, rounded to the nearest 100 units, where qα denotes the empirical α-quantile. We observe from these summary statistics that the number of records has a 32.9% excess of zeros and a very heavy right tail, given that the sample mean is significantly larger than the sample median. This can be revealed by the fact that some types of loss such as competitive advantage and reputation damage can not be measured in digital record units. The breach incident occurred without any reported loss or expenses if there were no records lost in that incident. Loss records range from 0 to 500 billion which is difficult to quantify under one distribution. In this regard, the remainder of the analysis in this paper is based on the logarithm of severity in order to maintain complete low and high loss amounts information.

The PRC data set contains three explanatory variables that can be used as regressors: breach type, organization type, and company location. The first two variables are documented to have 7 subcategories each, while the location is listed in 50 geographical states. We modify on their levels, based on their nature and characteristics to reduce factor dimensions and increase predictive power. Table 3 summarizes the combined model inputs of business and breach types. The level combination of the geographic locations is discussed and introduced in Section 2.2. After obtaining 6 combined levels of information, we investigate their performance on the target variable and find those medical and non-medical organizations behave differently concerning the number of breached records. It can be observed from last two rows of Table 2 the significant differences between the medical data and the non-medical data; the latter refers to the business and non-business types of organizations in Table 3. Although they all process a similar point mass of zero, their zero proportions differ notably to the extent of the heavy tail and maximum amount measured on the non-zero claim amount. In addition, the medical losses are more compact compared to non-medical losses. We hence postulate that the underlying severity distribution features multimodality; in this sense, a multimodal distribution or mixture distribution could be candidates for modeling the overall losses.

The above-mentioned fact can also be observed from Figure 1, where both histograms of logarithmic records breached and incurred by the medical and the non-medical organizations are displayed. The non-zero severity body part of density of medical organization has a peak of around 600 records, and the probability

Table 2. Summary statistics of target variable.

Table 3. Type of organizations and breaches.

Figure 1. Histograms between medical and non-medical organizations.

for losses being smaller than the mode value is relatively low. Meanwhile, the body part density for non-medical organizations shows a relatively smooth and flat distributional pattern before and after its mode point and relatively a heavier tail. Compared to financial services industry, which has spent the last 20 years focusing on cyber security and protection [27] , healthcare organization is not as frequently attacked by cyber related incidents. Medical organizations form traditionally risk retention group to mitigate huge liability losses caused by cyber breaches, making them reluctant to understand, track, report, and manage threats via open market cyber insurance coverage. Besides, mature incident and vulnerability risk management processes are lacking in most medical organizations [28] . Thus, daily threats are not even reported or managed effectively, which explains the low occurrence of cyber-severity loss of less than 600 incident records. Even though some of the data are already appropriate to model losses with heavy tails, they do not account for this type of multimodality case resulting from data variations observed between medical and non-medical organizations. In this regard, estimating the moderate loss density component with a fixed number of mixed components is advantageous.

2.2. Cluster Analysis

Because the PRC data also contains the geographical location of the victims of cyber attacks, a list of 51 states of U.S. with their latitudes and longitudes serves as the raw data information. It is a common practice that the number of levels in the geographical rating factor are to be reduced in order to provide effective risk measurement for insurance rate-making. For this purpose, we use one of the initialization strategies, cluster analysis [29] , to do the analysis. Clustering analysis is a newly developed computer-oriented data analysis which utilize unsupervised machine learning algorithm. Clustering is segmenting a data set based on similarities between the data points. We conduct cluster analysis for three reasons. First, it avoids diluting predictive power caused by the geographical location factor with 51 levels. Second, when states with similar characteristics are grouped, implementing rate-making is simpler. Third, it reduces the likelihood that the rate for one area is drastically differ from that for its neighbouring areas. Cluster analysis divides observations into distinct groups so that the observations within each group are quite similar to one another, as opposed to grouping 51 states into some official government regions, such as those used by the U.S. Census Bureau and the Standard Federal Regions. Cluster analysis divides observations into distinct groups such that the observations within each group are similar to each other. Before clustering, we conduct a cluster analysis using the means of latitude and longitude in each state as representatives. In this regard, we smooth the regression coefficients to make them more reasonable and interpretable, given that clustered groups are based on state average level. Now we have a set of 8095 observations, each with two features, longitude and latitude, that can be used to identify subgroups. We are attempting to discover geographical heterogeneity structures based on the PRC data set, which is an unsupervised problem.

Figure 2 represents the geographical heat map information in a two-dimensional space of longitude and latitude. These are the first two principal components of the data, which summarize the 8095 investigated incidents for location information

Figure 2. Average severity level among states.

in terms of two geographical dimensions. Each small, closed area corresponds to one of the 51 states, allowing for a visual examination of the average severity level for signs of clustering. There appear to be multiple groups of clusters with similar colour patterns. Two commonly used clustering techniques are k-means [30] and hierarchical [31] , which have been widely applied in territory studies for finding patterns and investigating the underlying geographical structure of the data. This study uses the k-means method with elbow [32] to improve an efficient and effective k-means performance. The elbow method is the default method for determining the optimal number of clusters for a characteristic process. The k-means clustering algorithm formalizes finding the best similarity grouping where the observations within each cluster are as small as possible, and the variation between clusters is significant. The similarity is measured by the sum of the squared Euclidean distance (SSE) [33] , one of the most widely used cluster distance criteria:

SSE = k = 1 K [ x i = C k ( x i μ k ) 2 ] ,

where μ k is the cluster centroid/mean, and C k represents one of the K clusters. We manually conduct a k-means cluster analysis with one to six clusters and calculate the ratio between cluster sum of squares and the total sum of squares for each round. We take this ratio as the y-axis and create an elbow plot as inllustrated in Figure 3(a). The plot demonstrates the elbow at k = 5 , beyond which the gains in the between cluster of sum of squares appear to be minimal as the increase in total sum of squares after k = 5 is greatly shrinking down; therefore 5 is the best cluster cut-off point. Figure 3(b) depicts the relative

Figure 3. Cluster selection. (a) Elbow plot for clusters; (b) Five geographical clusters.

geographical location of five clusters, while Appendix A provides context-specific information about cluster partitioning by state. This way, we can identify the geographical segments of cyber severity and classify them according to similar risk factor factors.

3. Modeling Framework

One of the professional ethics of actuaries is to study loss distributions. Suggested by the empirical data analysis in Section 2.1, the severity distribution of cyber loss records possesses point masses of zero, over-dispersion and heavy tail on extreme values, and various severity performance types on medical and non-medical organizations, which hardly can be fitted by a single analytic and parametric distribution. Our data set also allows us to examine individual risk characteristics via regression predictors, such as breach type, business type, and location. Based on these characteristics, we propose an FMM with three components integrated into a GLM framework to analyze the severity of a cyber loss.

3.1. Splicing Models

Many risk and loss variables, such as bodily injury costs and cyber losses, have long tails. Therefore, when modelling claim size to set premiums, calculating loss measures, and determining capital requirements for solvency regulations, it is frequently necessary for the actuarial analytic domain to obtain a global fit for risk distributions. In the literature, a splicing model is also called a composite model, in which multiple light-tailed distributions for the body and a heavy-tailed distribution for the tail are combined. The general density form of an m-component spliced distribution is as follows:

f ( y ) = ( p 1 f 1 ( y ) y C 1 , p 2 f 2 ( y ) y C 2 , p m f m ( y ) y C m , (3.1)

where f i is a legitimate density function with all probability on the mutually exclusive and sequentially ordered interval C i , and positive weights p 1 , , p m that add up to one, i.e., i = 1 m p i = 1 . In this regard, the density function (3.1) and its corresponding cumulative distribution function can be written as a compact form and as

f ( y ) = i = 1 m I C i ( y ) p i f i ( y ) , F ( y ) = i = 1 m I C i ( y ) ( j = 1 i 1 p j + p i F i ( y ) ) ,

where I is an indicator function with I C i ( y ) = 1 , if y C i , otherwise 0, F i is the corresponding cumulative distribution function of f i in the interval C i .

Based on the empirical analysis results shown in Table 2, we consider a spliced distribution with three components: the first component contains zeros, the second component models the middle segment of the amount of lost data, and the third component models the tail segment. Let Y j denote the random variable that represents the jth loss amount, c is the non-zero loss threshold, and then the pdf of Y j can be expressed as

f ( y j | α ) = ( p 1 ( y j ; α ) y j = 0 , p 2 ( y j ; α ) f 1 ( y j ; α 1 ) F 1 ( c ; α 1 ) F 1 ( 0 + ; α 1 ) y j ( 0 , c ] , [ 1 p 1 ( y j ; α ) p 2 ( y j ; α ) ] f 2 ( y j ; α 2 ) 1 F 2 ( c ; α 2 ) y j ( c , ) , (3.2)

where f 1 and f 2 are two density functions with cdf F 1 and F 2 , respectively, defined on ( 0, ) , α = ( α 1 T , α 2 T ) T is a set of parameter vector associated with the distributions for the components. The threshold c is a parameter to be estimated from the data which is investigated in Section 4.3. The remaining unknown parameters p 1 , p 2 and α can be estimated using the maximum likelihood estimation (MLE) method by maximizing the log-likelihood function based on observations y 1 , y 2 , , y n , which is

log L ( α ) = log ( p 1 ( y j ; α ) ) j = 1 n I { 0 } ( y j ) + j = 1 n I ( 0 + , c ] ( y j ) [ log ( p 2 ( y j ; α ) ) + log f 1 ( y j ; α 1 ) log ( F 1 ( c ; α 1 ) F 1 ( 0 + ; α 1 ) ) ] + j = 1 n I ( c , ) ( y j ) [ log ( 1 p 1 ( y j ; α ) p 2 ( y j ; α ) ) + log f 2 ( y j ; α 2 ) log ( 1 F 2 ( c ; α 2 ) ) ] . (3.3)

3.2. Finite Mixture Models

Due to the adaptability in utilizing high-dimensional features, coping with population heterogeneity, and achieving multiple interrelated goals, mixture distributions have gained popularity in recent years. [34] provides a thorough discussion of using the EM algorithm to find maximizers of MLE and the selection of the number of components in finite mixture models. Let Y 1 , , Y n denote a random sample of size n, and y = ( y 1 , , y n ) T is the observed value of random vector. Suppose that Y j follows a finite mixture distribution with density function f on , which can be written in the form5

f M ( y j ) = i = 1 g π i f i ( y j ) , (3.4)

where for i = 1,2, , g , f i is a density function and π i is a non-negative quantity such that 0 π i 1 and i = 1 g π i = 1 . The quantities π 1 , , π g are the mixing proportions or weights, and f 1 , , f g are called the component densities of the mixture. We call density (3.4) as a g-component finite mixture density function and its corresponding distribution F M as a g-component finite mixture distribution function.

In order to well interpret the mixture models, let Z j be a g-dimensional component label vector where the ith element Z i j = ( Z j ) i is defined to be one or zero according to whether the component of Y j in the mixture is equal to i or not ( i = 1, , g ). Thus this categorical random vector Z j can be viewed as following a multinomial distribution with probabilities π 1 , , π g ; that is,

{ Z j = z j } = π 1 z 1 j π 2 z 2 j π g z g j (3.5)

according to a multinomial distribution consisting of one draw on g categories with probabilities π 1 , , π g We write

Z j ~ Mult g ( 1, π ) , (3.6)

where π = ( π 1 , , π g ) T . In the interpretation of a mixture model, Y j is drawn from a population with G with g groups, G 1 , , G g in proportions π 1 , , π g , where the density of Y j in group G i is f i ( y j ) The component-indicator variables z i j will be used in finding optimizers under ML estimation via the EM algorithem in Section 4.

Generally, the components can be any exponential family distribution [35] ; observations are available from a population known to be a mixture of K sub-populations. In our study, each subpopulation will not necessarily be assumed to have the same type of distribution, which is one of the most significant departures from previous research. For a single observation, the probability density of the exponential family can be expressed as follows:

f ( y j ; θ j , ϕ ) = exp { y j θ j b ( θ j ) ϕ + c ( y j , ϕ ) } (3.7)

where θ j is a natural parameter or canonical parameter, ϕ is the dispersion parameter or scale parameter6, b ( θ j ) is a known function called cumulant function and c ( y j , ϕ ) is a normalizing function, thus ensuring that (3.7) is a probability function. The mean and variance of exponential family distributions can be expressed by b ( θ j ) as follows:

μ = E ( Y j ) = μ j = b ( θ j ) and V a r ( Y j ) = ϕ b ( θ j ) . (3.8)

Considerations are given to the family of mixtures of generalized linear models (GLMs) because many applications of non-normal mixtures involve components from the exponential family. GLMs are a statistical framework for unifying several significant exponential family models [37] . In this framework, it is permissible for the mixing proportions and the component distributions are allowed to depend on some associated covariates. We have independent sample data { x j , y j } , j = 1, , n , in which y j is the response/target variable, n is the sample size, and x j = ( x j 1 , , x j p ) T is a vector of p explanatory variables, GLMs thus fit the following:

g ( μ j ) = η j = β T x j = β 1 x j 1 + + β p x j p (3.9)

where η j is the linear predictor, and μ j , the mean of an exponential family distribution f ( y j ; θ j , ϕ ) , is a known function of the canonical parameter θ j described in (8), g ( ) is a known link function that connects distribution mean and linear combination together.

From the above methodologies, we propose a finite mixture regression model where mixture components can be from same or different types of parametric family. Our model employs candidate distributions like Gamma, Log Normal, Inverse Gaussian, and Weibull because loss or severity is typically modelled as continuous random variables.

3.3. Zero Inflated Mixture and Composite Regression Models

In this subsection, we introduce a order to provide a clear understanding of our combined finite mixtures and splicing model, we first perform our model under a general splicing framework with three parts spliced densities jointing with weighting probabilities, followed by a detailed finite mixture section expression of the moderate spliced density part.

Let Y + be the claim severity random variable, and let x P be the vector of covariate information. The density of the zero-inflated mixture composite regression model written in a spliced form with zero and two densities f M and f T and their corresponding cumulative distribution functions (CDFs) F M and F T is given by

f Y ( y j ; α , W , B , ϕ , γ , κ , x j ) = p 1 ( y j ; α , x j ) 1 { y j = 0 } + p 2 ( y j ; α , x j , c ) f M ( y j ; W , B , ϕ , x j ) F M ( c ; W , B , ϕ , x j ) F M ( 0 + ; W , B , ϕ , x j ) 1 { 0 < y j c } + [ 1 p 1 ( y j ; x j , α ) p 2 ( y j ; α , x j , c ) ] f T ( y ; γ , κ , x j ) 1 F T ( c ; γ , κ , x j ) 1 { y j > c } , (3.10)

where { p 1 , p 2 } ( 0,1 ) are the splicing weights, c is the splicing point which is the threshold separating the moderate and extreme loss values, α is covariate coefficients of zero-inflated weight, W , B and ϕ are parameter vectors of the density of body f M which is a finite mixture model, and γ and κ are coefficients of the density of tail f T .

In this study, the finite mixture distribution f M is the density of positively defined continuous distributions with upper truncation at the threshold loss level c.

f M ( y j ; W , B , ϕ , x j ) = i = 1 g π i j ( x j ; W ) f i ( y j ; exp ( β i T x j ) , ϕ i ) (3.11)

where B contains the elements of β 1 , , β g known a priori to be distinct, ϕ = ( ϕ 1 , , ϕ g ) T is a vector of fixed dispersion parameters of g distribution components from the exponential family. The parameter π i j is the mixing proportion of the ith function and jth observation which is a function of x j and commonly modeled by logistic distributions

π i j = π i ( x j ; W ) = exp ( ω i T x j ) 1 + h = 1 g 1 exp ( ω h T x j ) , (3.12)

where W = ( ω 1 T , , ω g 1 T , ω g T ) T , with ω g = 0 , contains the logistic regression coefficients. Lastly, f T is the tail density function from the exponential family with heavy-tailed performance, given by

f T ( y j ; γ , κ , x j ) = f T ( y j ; exp ( γ T x j ) , κ ) (3.13)

where exp ( γ j T x j ) = θ j , the canonical parameter in (3.7), and κ is the dispersion parameter.

4. Parametric Estimation

Let Ψ represent the set of vectors representing all the unknown parameters in (3.10) that must be estimated, namely,

Ψ = ( α T , W T , B T , γ T ) T . (4.1)

The density of our spliced mixture regression model (3.10) of the jth response variable Y j can then be written as follows:

f ( y j ; Ψ ) = p 1 j 1 { y j = 0 } + p 2 j i = 1 g π i j f i ( y j ; β i j , ϕ i ) F M ( c ; Ψ ) F M ( 0 + ; Ψ ) 1 { 0 < y j c } + ( 1 p 1 j p 2 j ) f T ( y j ; γ j , k ) 1 F T ( c ; Ψ ) 1 { y j > c } . (4.2)

where p 1 j = p 1 ( x j ; α ) and p 2 j = p 2 ( x j ; α , c ) . In this way, the log likelihood for Ψ can be formed as

log L ( Ψ ) j = 1 n log ( p 1 j ) 1 { y j = 0 } + j = 1 n { log ( p 2 j ) + log ( i = 1 g π i j f i ( y j ; β i j , ϕ i ) ) log [ F M ( c ; Ψ ) F M ( 0 + ; Ψ ) ] } 1 { 0 < y j c } + j = 1 n { log ( 1 p 1 j p 2 j ) + log f T ( y j ; γ j , k ) log [ 1 F T ( c ; Ψ ) ] } 1 { y j > c } . (4.3)

The EM algorithm [38] can be applied to obtain the MLE of Ψ in this spliced mixture regression model. The complete-data log likelihood is given by the following

log L c ( Ψ ) j = 1 n log ( p 1 j ) 1 { y j = 0 } + j = 1 n { log ( p 2 j ) + i = 1 g z i j [ log ( π i j ) + log f i ( y j ; β i j , ϕ i ) ] log [ F M ( c ; Ψ ) F M ( 0 + ; Ψ ) ] } 1 { 0 < y j c } + j = 1 n [ log ( 1 p 1 j p 2 j ) + log f T ( y j ; γ j , k ) log ( 1 F T ( c ; Ψ ) ) ] 1 { y j > c } (4.4)

where z i j denotes the component-indicator variables as defined in (3.5).

4.1. E-Step

The EM algorithm is applied to this problem by treating the z i j as missing data. E (for expectation) and M (for maximization) are the two iterative steps. Given an observed data y , we take the conditional expectation of the compelete-data log likelihood (4.4) using the current fit for Ψ . We consider Ψ ( 0 ) as an initial value of the iterative computation. The E-step computes the conditional expectation of log L c ( Ψ ) given y using Ψ ( 0 ) for Ψ on the first EM algorithm iteration, that is

Q ( Ψ ; Ψ ( 0 ) ) = E Ψ ( 0 ) [ log L c ( Ψ ) | y ] (4.5)

where the subscript Ψ ( 0 ) means that using Ψ ( 0 ) for Ψ effect the expectation. In this manner, the E-step calculates Q ( Ψ ; Ψ ( k ) ) on the ( k + 1 ) th iteration, where Ψ ( k ) ) is the value of Ψ ) after kth iteration. The E-step requires the calculation of the current conditional expectation of Z i j given the observed data y , respectively

E Ψ ( k ) ( Z i j | y ) = Ψ ( k ) { Z i j = 1 | y } = τ i j ( y j ; Y ( k ) ) , (4.6)

where

τ i j ( y j ; Ψ ( k ) ) = π i j ( k ) f i ( y j ; β i j ( k ) , ϕ i ) f M ( y j ; Ψ ( k ) ) = π i j ( k ) f i ( y j ; β i j ( k ) , ϕ i ) h = 1 g π h j ( k ) f h ( y j ; β h j ( k ) , ϕ h ) (4.7)

for i = 1 , , g ; j = 1 , , n . The quantity τ i j ( y j ; Ψ ( k ) ) is the posterior probability that the jth member of the sample with observed value y j belongs to the ith component of the mixture. Taking the conditional expectation of (4.4) using (4.6) that

Q ( Ψ ; Ψ ( k ) ) = j = 1 n log ( p 1 j ) 1 { y j = 0 } + j = 1 n { log ( p 2 j ) + i = 1 g τ i j ( y j ; Ψ ( k ) ) [ log ( π i j ) + log f i ( y j ; β i j , ϕ i ) ] log [ F M ( c ; Ψ ) F M ( 0 + ; Ψ ) ] } 1 { 0 < y j c } + j = 1 n [ log ( 1 p 1 j p 2 j ) + log f T ( y j ; γ j , k ) log ( 1 F T ( c ; Ψ ) ) ] 1 { y j > c } (4.8)

We assume F M ( 0 + ; Ψ ) = 0 in the following derivations, which is generally the case.

4.2. M-Step

The M-step on the ( k + 1 ) th iteration entails solving the following system of four equations:

j = 1 n α [ log ( p 1 j ) 1 { y j = 0 } + log ( p 2 j ) 1 { 0 < y j c } + log ( 1 p 1 j p 2 j ) 1 { y j > c } ] = 0 , (4.9)

j = 1 n i = 1 g τ i ( y j ; Ψ ( k ) ) W log ( π i j ) 1 { 0 < y j c } = 0 , (4.10)

j = 1 n i = 1 g τ i ( y j ; Ψ ( k ) ) B [ log f i ( y j ; β i j , ϕ i ) log [ F M ( c ; Ψ ) F M ( 0 + ; Ψ ) ] ] 1 { 0 < y j c } = 0 (4.11)

j = 1 n γ [ log f T ( y j ; γ , k ) log ( 1 F T ( c ; Ψ ) ) ] 1 { y j > c } = 0 . (4.12)

Equations (4.9) and (4.10) can be solved using a standard algorithm for logistic regression to produce updated estimate α ( k + 1 ) and W ( k + 1 ) for the logistic regression coefficients as they both represent the probabilities between 0 and 1. Concerning the computation of B and γ and applying the chain rule of [39] , the likelihood equation for γ (4.12) can be expressed as

j = 1 n w ( μ j ) ( y j μ j ) η ( μ j ) x j = 0 , (4.13)

where μ j = exp ( γ T x j ) , η ( μ j ) = d η j / d μ j and w ( μ j ) is the weight function defined by

w ( μ j ) = 1 / [ η j ( μ j ) ] 2 V ( μ j ) . (4.14)

The likelihood Equation (4.13) can be solved using iteratively reweighted least squares (IRLS) [37] . The adjusted response variable y ˜ j for the ( k + 1 ) th iteration is

y ˜ j ( k ) = η ( μ j ( k ) ) + ( y j μ j ( k ) ) η ( μ j ( k ) ) . (4.15)

These n adjusted responses are then regressed on the covariates x 1 , , x n using weights w ( μ 1 ( k ) ) , , w ( μ n ( k ) ) . This produces an updated estimate γ ( k + 1 ) for γ and, consequently, the updated estimates μ j ( k + 1 ) for the μ j for use in the right-hand side of (4.15) to update the adjusted responses, and so on. This procedure is repeated until the variations in the estimates are small enough. Same as (4.13), the likelihood for B in (4.11) can be written as

i = 1 g j = 1 n τ i j ( y j ; Ψ ( k ) ) w ( μ i j ) ( y j μ i j ) η i ( μ i j ) [ B η i ( μ i j ) ] = 0 (4.16)

where μ i j is the mean of Y j for the ith component. Given that

β h η i ( μ i j ) = { x j , if h = i 0, otherwise , (4.17)

Equation (4.16) reduces to solving

j = 1 n τ i j ( y j ; Ψ ( k ) ) w ( μ i j ) ( y j μ i j ) η i ( μ i j ) x j = 0 (4.18)

Separately for each β i to produce β i ( k + 1 ) , i = 1, , g . As (4.15) responses y 1 , , y n are fitted with weights τ i 1 ( y 1 ; Ψ ( k ) ) , , τ i n ( y n ; Ψ ( k ) ) and fixed dispersion parameter k i . (4.18) then can be solved using the IRLS approach for a single GLM. The double summation over i and j in (4.16) can be handled by expanding the response vector to have dimension g × n by replicating each original observation ( y j ; x j T ) T g times, with weights τ i 1 ( y 1 ; Ψ ( k ) ) , , τ i n ( y n ; Ψ ( k ) ) , fixed dispersion parameters k 1 , , k g , and linear predictors x j T β 1 , , x j T β g .

4.3. Starting Values

To allow the overall distribution to disjoint at the splicing point, we maintain a density equation at c by setting

p 2 f M ( c ; x , B , Φ ) F M ( c ; x , B , Φ ) F M ( 0 + ; x , B , Φ ) = [ 1 p 1 ( y ; x , α ) p 2 ] f T ( c ; x , γ , θ ) 1 F T ( c ; x , γ , θ ) . (4.19)

After simplifying 4.19, we derive p 2 as a function of any given p 1

p 2 = [ 1 p 1 ( y ; x , α ) ] f T ( c ; x , γ , θ ) 1 F T ( c ; x , γ , θ ) / { f M ( c ; x , B , Φ ) F M ( c ; x , B , Φ ) F M ( 0 + ; x , B , Φ ) + f T ( c ; x , γ , θ ) 1 F T ( c ; x , γ , θ ) }

Until now, we have considered the fitting of a Zi-MCR for a given value of severity threshold c in the composite model. Typically, where our model is being used to handle overdispersion, c is predetermined from data using extreme value analysis or expert opinion via performing cyber insurance policy limit and similar matters. This is primarily motivated by estimation stability, which is adopted by [17] . Furthermore, conducting formal tests at any stage of this sequential process is chanllenging because regularity conditions for the likelihood ratio test statistic’s typical asymptotic null distribution do not hold [34] . Observing the trend in the log-likelihood as c is increased from a sequence of severity levels (1000, 5000, 10,000, 50,000, 100,000) can provide us with a heuristic for determining the optimal value for c. When dealing with a data-driven model, this method for selecting a splicing point makes more sense and is widely used [40] .

5. Analysis of Severity of Data Breaches

In this Section, we illustrate the efficiency of the EM algorithm on estimation by fitting a Zi-MCR model, as proposed in Section 3.3, to the PRC dataset. Furthermore, general model form and covariates are discussed, and several distribution combinations are tested to select the best performance.

This study is based upon the PRC cyber breach incident data by stratifying the residuals. The training set fine-tunes all candidate models, and their performance and out-of-sample validation are checked upon the test set. We conduct 5-fold cross-validation and set 80% as the training data to fit the models. Based on a set of breach observations, the problem is to estimate whether the unknown parameters can be contained in the vector Ψ = ( α T , W T , B T , γ T ) T , as in 4.1. We represent the logarithm rescaled number of loss records of data breach incidents explained in Section 2 as a target or dependent random variable Y j , and 9 covariates, including intercept, 2 business levels, 2 breach levels, and 4 location area levels, as described in the Sections 2.1 and 2.2 as vector coefficients x j .

Table 4 displays the summaries of three categorical variables; the proportion of zeros and differences between the mean values of the categories numerically illustrate their distribution patterns. These results demonstrate the importance of letting splicing weights depend on covariates and separately modelling body and tail parts.

In the PRC dataset, a source of the heterogeneity is mainly from businesses that have or do not have high prevention defence systems and active cyber risk managing activities, such as healthcare and financial service organizations. This is explained in Section 2.1 by comparing their kernel plots and enterprise features. The body part component may be viewed as two groups corresponding to whether those incidents happened within medical organizations. The problem is to estimate the medical and non-medical organization mixture rate, that is, the mixing proportion π 1 . Given g = 2 , α ( k + 1 ) and W ( k + 1 ) can be calculated using binomial error structure with the canonical logit transformation as the link. For illustrative purposes, we fit several popular distribution combinations on a mixture of body and heavy tail parts. To measure the overall goodness of fit of those fitted distributions, we calculate the Akaike information criterion (AIC) statistics. Table 5 reports the global fit distributions overall AIC values on a given c = 5000 threshold. The fit from Lognormal and Weibull body mixture and Pareto tail outperforms with the lowest AIC. We conduct a simulation study

Table 4. Summary of categorical variables.

Table 5. Overall goodness-fit.

based on the entire data set to comprehend the chosen model’s adaptability further. The procedure is repeated 200 times to ensure a thorough analysis of the chosen distribution combination. Finally, the estimated parameters are summarized in Table 6 using the above distribution combined with the lowest AIC.

Since the metric is based on data-driven analysis, we can draw a conclusion

Table 6. Parameter estimations.

that the selected combination, Lognormal-Weibull and Pareto, has the most explanation power of PRC dataset. If other dataset is given, methodology and algorithm in generating the metric should not change, selected distribution combination would depend on a case by case basis.

6. Cyber Loss Aggregation and Application

The premium calculation algorithm, known as rate order calculation, is applied to categorize segmentation to derive final premium rates. In order to set competitive premiums and develop sustainable underwriting plans, insurers extensively use historical loss data to seek economies of scale and premium balancing. Statistical algorithms and mathematical modelling arguments are used to structure aggregate cyber risk. The purpose of this Section is to describe an aggregate loss model based on the total amount of cyber loss that occurs in a quarter concerning a group of homogeneous risk characteristics and apply this model to determining increased limit factors (ILFs) based on the underlying data in order to balance statistical and economic constraints. A case study is established for two locations and business types (non-business and business, respectively). Their cyber risks are aggregated under all possible policy limits and deductible combinations. One method for addressing this issue is to separately model the individual severity levels and the quarterly incidence rate. These components can then derive a distribution for the total in a period. This approach has several advantages: changes over time can be monitored and attributed to frequency or severity as cyber risk shows a strong seasonal pattern. The other is that methodology can be quickly adapted to find suitable models for the components, as in the previous model selection. According to risk theory [41] , a collective risk model with aggregate loss S, which represents the total amount for a quarterly cyber risk, can be defined as follows

S = X 1 + X 2 + + X N ,

where loss count N and non-negative severities X 1 , , X N are random variables with independence assumptions that N does not rely on the severity of loss and X i s are i.i.d. independently with respect to a given count N. Under this assumption, the aggregate loss distribution (ALD) F S is a compound cdf of the following form [16] :

F S ( s ) = n = 0 p N ( n ) P r ( S s | N = n ) = E N F X ( n ) ( s ) ,

where p N ( n ) = P r ( N = n ) and F X ( n ) is the n-fold convolution of F X , which is defined by:

F X ( n ) ( x ) = ( 0 x f X ( y ) d F X ( n 1 ) ( x y ) d y n = 2 , 3 , F X ( x ) n = 1.

Because of the complexity of this claim amount distribution in practical applications, the following approximations for the distribution of the total loss are typically taken into account:

· Pure Risk Premium

P = E ( N ) E ( X ) ,

· Premium with Safety Loading Factor θ

P S L ( θ ) = ( 1 + θ ) E ( N ) E ( X ) θ 0,

· Premium with Variance Loading Factor a

P V ( a ) = E ( N ) E ( X ) + a [ E ( N ) V a r ( X ) + E ( X ) 2 V a r ( N ) ] a 0,

· Premium with Standard Deviation Loading factor b

P S D ( b ) = E ( N ) E ( X ) + b E ( N ) V a r ( X ) + E ( X ) 2 V a r ( N ) b 0,

In our case study, where loading factors are not specified, pure premium creates an example, which can be modified once loading information is obtained from businesses. Table 7 displays the total quarterly dollar loss caused by cyber risk. While frequency distribution is determined by our previous investigation of the generalized linear mixed model [26] of the same dataset, severity distribution is the distribution combined with the lowest AIC in Section 5. With dollar amount transferred from units of loss record by applying following rule [42] :

ln ( dollaramountloss ) = 7.68 + 0.76 × ln ( lossrecordsbreached )

Table 7. Quarterly aggregate loss in dollar amount.

For the collective risk model, the expected value and variance of aggregate claims S are as follows:

E ( S ) = E ( N ) E ( X ) ,

V a r ( S ) = E ( N ) V a r ( X ) + E ( X ) 2 V a r ( N ) .

Estimates of quantities such as VaR, TVaR, and ILFs can be analyzed using aggregated results at a given risk tolerance level. Utilizing risk characteristics, our model divides homogeneous risks into segments. Then, product designers can decide whether to implement policy limits or seek reinsurance. It can be broken down into four steps: establishing the base rate, applying risk factors, multiplying discount and surcharge, and factoring in expense retention. The first and second steps lay the groundwork for the entire pricing process. They are frequently carried out using an experienced-based pricing technique with preliminary data for analysis. Our cyber risk loss aggregation results are generated within a Bayesian framework, which proves to be a useful prediction tool for estimating future loss among segmentation with confidence.

7. Conclusions and Discussions

Once an insurance loss model has been constructed, in addition to applying techniques to data sets, we must consider numerous modelling-related factors, such as risk management and pricing decisions (for insurers) or the impact on capital requirements (for enterprises). Our model and findings provide meaningful insights to risk mitigation and risk transfer techniques, which benefit not only the individual organization, but also the overall economy.

Cyber risk exists because computer data is valuable to individuals. Business, and governments; therefore, the data must be protected by organizations that store privileged information. Financial firms receive, maintain, and store large amounts of personally identifiable information. Recent security research [43] indicates that most businesses have unprotected data and inadequate cybersecurity practices, making them susceptible to data loss. As more executives and decision-makers recognize the value and significance of security investments, cybersecurity budgeting has steadily risen to successfully combat potential digital property loss. A systemic cyber event could cost multiple times the current risk retention estimate. As a result of regulatory scrutiny and the need for improved portfolio management, businesses conduct scenario modelling and sensitivity tests regularly based on their changing risk appetite. To reduce cyber risk, organization can adopt threshold limits by monitoring risk with preset limits based on established risk criteria, trigger will be placed in threshold that has been breached. The objective is to achieve and maintain an acceptable level of risk at a reasonable cost. Under the leadership of the Chief Risk Officer, companies must revise their strategies, including changes to their risk appetite and the composition of their hedge products. Due to some businesses’ nature or responsibilities, increasing risk appetite or security investments may not be sufficient to achieve the risk management objective. Such limited reserved retention can have disastrous financial consequences if a data breach occurs, forcing the organization to absorb the costs associated with internal remediation and its liability to third parties. In this perspective, cyber insurance has become an effective alternative or backup tool for managing cyber risk.

Our investigation of large claims and an excess of zeros raises the issue of the risk’s insurability under various feature characteristics. To eliminate the variance caused by the heavy tail, the non-catastrophe loss will then be used to train a predictive model. Reinsurance will kick in if the loss exceeds the company’s tolerance level to ensure that insurers are not severely impacted. Our model offers additional perspective on coping with extreme loss value, such as cyber risk. Risk selection is one of the most crucial processes when designing an insurance product. Since not all customers are equally attracted to an insurance product, segmenting the risks into distinct groups is advantageous to prevent adverse selection. To ensure that cyber insurance products are priced appropriately, we use the results from the Section 6 to divide risks into categorizable segments. In addition, our model can be utilized to perform a preliminary pre-screening of a prospective client to facilitate rate discrimination and the creation of customized contracts. This security audit enables the insurer to capitalize on the profit opportunity presented by the interdependence of cyber risk.

Combining the loss frequency and severity distribution through convolution is the conventional method for estimating the aggregate loss distribution. Given the proposed mixture and composite severity model, aggregate losses can be estimated through simulation since our previous frequency model was semi-parametric with a simulated posterior distribution without a closed form of distribution.

The financial sector faces cybersecurity risks in their daily operations while insuring product providers. Insurers receive personal health and financial information from both policyholders and claimants. The cost of cyber insurance increased by an average of 96% in the third quarter of 2021 as organizations faced a daily onslaught of cyberattacks [44] . To mitigate the premium price increase, policyholders increased their retention. As a result, insurers must improve predictive analysis and cyber risk models to maintain market share and company solvency. Our model provides a method for measuring cyber risk severity, and there are multiple ways to extend this method. As previously stated, all of our results are based on the assumption of equal exposure, whereas exposure is the most crucial factor in determining the pure premium. Cyber risk loss exposures is any condition that presents the possibility of financial loss to an organization from property, net income and liability as a consequence of advanced technology transmissions, operations, maintenance, development and support. Training the predictive model under the assumption of equal exposure in a defined time period would be an important direction for future research once prior experience data with exposure information is obtained.

Appendix

A. Geographical Location Clusters

NOTES

1General Data Protection Regulation https://gdpr.eu/.

2California Consumer Privacy Act https://oag.ca.gov/privacy/ccpa.

3Illinois Biometric Information Privacy Act https://www.ilga.gov/legislation/.

4Privacy Rights Clearinghouse (PRC) database is available for public download at https://privacyrights.org/data-breaches.

5In this formulation of the mixture model, the number of components g is considered fixed. In many applications, the value of g is unknown and inferred from the available data, along with the mixing proportions and the parameters in the specified forms of the component densities.

6When ϕ is known, the distribution of Y j is one-parameter canonical exponential family member. When ϕ is unknown, it is often a nuisance parameter and then it is estimated by the method of moments. In most of GLM theory, the role of ϕ is often treated as an unknown constant but not as a parameter [36] .

Conflicts of Interest

The authors declare no conflicts of interest.

References

[1] Gartner Research (2018) Forecast Analysis: Information Security, Worldwide, 2Q18 Update. Gartner Research.
[2] AXA (2019) AXA & Eurasia Group Future Risks Report. AXA.
[3] (2020) IBM: Cost of a Data Breach Report. Computer Fraud & Security, 2021.
https://doi.org/10.1016/S1361-3723(21)00082-8
[4] Leswing, K. (2020) Twitter Hackers Who Targeted Elon Musk and Others Received $121,000 in Bitcoin, Analysis Shows. CNBC TECH.
https://www.cnbc.com/2020/07/16/twitter-hackers-made-121000-in-bitcoin-analysis-shows.html
[5] Marriott (2020) Marriott International Notifies Guests of Property System Incident. Marriott International News Center.
https://news.marriott.com/news/2020/03/31/marriott-international-notifies-guests-of-property-system-incident
[6] Equifax (2017) Equifax acquires data-crédito.
https://investor.equifax.com/news-events/press-releases/detail/1221/equifax-acquires-data-crdito
[7] Michael, A.B. (2020) Exposure Measures for Pricing and Analyzing the Risks in Cyber Insurance. Casualty Actuarial Society and Society of Actuaries.
[8] Farley, J. (2022) The Cyber Insurance Market Struggles with Continued Hardening Market Conditions. Gallagher.
[9] Malavasi, M., Gareth, P., Shevchenko, P.V., Trück, S., Jang, J. and Sofronov, G. (2021) Cyber Risk Frequency, Severity and Insurance Viability. (Preprint)
https://doi.org/10.2139/ssrn.3940329
[10] Sun, H., Xu, M. and Zhao, P. (2021) Modeling Malicious Hacking Data Breach Risks. North American Actuarial Journal, 25, 484-502.
https://doi.org/10.1080/10920277.2020.1752255
[11] Farkas, S., Lopez, O. and Thomas, M. (2021) Cyber Claim Analysis Using Generalized Pareto Regression Trees with Applications to Insurance. Insurance: Mathematics and Economics, 98, 92-105.
https://doi.org/10.1016/j.insmatheco.2021.02.009
[12] McNeil, A.J. (1997) Estimating the Tails of Loss Severity Distributions Using Extreme Value Theory. ASTIN Bulletin: The Journal of the IAA, 27, 117-137.
https://doi.org/10.2143/AST.27.1.563210
[13] Allen, D.E., Singh, A.K. and Powell, R.J. (2013) EVT and Tail-Risk Modelling: Evidence from Market Indices and Volatility Series. The North American Journal of Economics and Finance, 26, 355-369.
https://doi.org/10.1016/j.najef.2013.02.010
[14] Park, M.H. and Kim, J.H.T. (2016) Estimating Extreme Tail Risk Measures with Generalized Pareto Distribution. Computational Statistics & Data Analysis, 98, 91-104.
https://doi.org/10.1016/j.csda.2015.12.008
[15] Beirlant, J., Goegebeur, Y., Segers, J. and Teugels, J.L. (2004) Statistics of Extremes: Theory and Applications. John Wiley & Sons, Hoboken.
https://doi.org/10.1002/0470012382
[16] Klugman, S.A., Panjer, H.H. and Willmot, G.E. (2012) Loss Models: Further Topics. John Wiley & Sons, Hoboken.
https://doi.org/10.1002/9781118787106
[17] Reynkens, T., Verbelen, R., Beirlant, J. and Antonio, K. (2017) Modelling Censored Losses Using Splicing: A Global Fit Strategy with Mixed Erlang and Extreme Value Distributions. Insurance: Mathematics and Economics, 77, 65-77.
https://doi.org/10.1016/j.insmatheco.2017.08.005
[18] Gan, G. and Valdez, E.A. (2018) Fat-Tailed Regression Modeling with Spliced Distributions. North American Actuarial Journal, 22, 554-573.
https://doi.org/10.1080/10920277.2018.1462718
[19] Tzougas, G., Vrontos, S. and Frangos, N. (2014) Optimal Bonus-Malus Systems Using Finite Mixture Models. ASTIN Bulletin: The Journal of the IAA, 44, 417-444.
https://doi.org/10.1017/asb.2013.31
[20] Sattayatham, P. and Talangtam, T. (2012) Fitting of Finite Mixture Distributions to Motor Insurance Claims. Journal of Mathematics and Statistics, 8, 49-56.
https://doi.org/10.3844/jmssp.2012.49.56
[21] Bernardi, M., Maruotti, A. and Petrella, L. (2012) Skew Mixture Models for Loss Distributions: A Bayesian Approach. Insurance: Mathematics and Economics, 51, 617-623.
https://doi.org/10.1016/j.insmatheco.2012.08.002
[22] Miljkovic, T. and Grün, B. (2016) Modeling Loss Data Using Mixtures of Distributions. Insurance: Mathematics and Economics, 70, 387-396.
https://doi.org/10.1016/j.insmatheco.2016.06.019
[23] Gui, W., Huang, R. and Lin, X.S. (2018) Fitting the Erlang Mixture Model to Data via a GEM-CMM Algorithm. Journal of Computational and Applied Mathematics, 343, 189-205.
https://doi.org/10.1016/j.cam.2018.04.032
[24] Fung, T.C., Badescu, A.L. and Lin, X.S. (2019) A Class of Mixture of Experts Models for General Insurance: Theoretical Developments. Insurance: Mathematics and Economics, 89, 111-127.
https://doi.org/10.1016/j.insmatheco.2019.09.007
[25] Fung, T.C., Badescu, A.L. and Lin, X.S. (2019) A Class of Mixture of Experts Models for General Insurance: Application to Correlated Claim Frequencies. ASTIN Bulletin: The Journal of the IAA, 49, 647-688.
https://doi.org/10.1017/asb.2019.25
[26] Sun, M. and Lu, Y. (2022) A Generalized Linear Mixed Model for Data Breaches and Its Application in Cyber Insurance. Risks, 10, Article 224.
https://doi.org/10.3390/risks10120224
[27] Bell, G. and Ebert, M. (2015) Health Care and Cyber Security: Increasing Threats Require Increased Capabilities. KPMG.
[28] Williams, P.A.H. and Woodward, A.J. (2015) Cybersecurity Vulnerabilities in Medical Devices: A Complex Environment and Multifaceted Problem. Medical Devices: Evidence and Research, 8, 305-316.
https://doi.org/10.2147/MDER.S50048
[29] Roberts, S.J. (1997) Parametric and Non-Parametric Unsupervised Cluster Analysis. Pattern Recognition, 30, 261-272.
https://doi.org/10.1016/S0031-3203(96)00079-9
[30] Likas, A., Vlassis, N. and Verbeek, J.J. (2003) The Global k-Means Clustering Algorithm. Pattern Recognition, 36, 451-461.
https://doi.org/10.1016/S0031-3203(02)00060-2
[31] Johnson, S.C. (1967) Hierarchical Clustering Schemes. Psychometrika, 32, 241-254.
https://doi.org/10.1007/BF02289588
[32] Bholowalia, P. and Kumar, A. (2014) EBK-Means: A Clustering Technique Based on Elbow Method and K-Means in WSN. International Journal of Computer Applications, 105, 17-24.
[33] Agrawal, R., Faloutsos, C. and Swami, A. (1993) Efficient Similarity Search in Sequence Databases. In: Lomet, D.B., Eds., FODO 1993: Foundations of Data Organization and Algorithms, Lecture Notes in Computer Science, Vol. 730, Springer, Berlin, 69-84.
https://doi.org/10.1007/3-540-57301-1_5
[34] Peel, D. and MacLahlan, G. (2000) Finite Mixture Models. John & Sons, Hoboken.
https://doi.org/10.1002/0471721182
[35] Hasselblad, V. (1969) Estimation of Finite Mixtures of Distributions from the Exponential Family. Journal of the American Statistical Association, 64, 1459-1471.
https://doi.org/10.1080/01621459.1969.10501071
[36] Yee, T.W. (2015) Vector Generalized Linear and Additive Models: With an Implementation in R. Springer, Berlin.
[37] Nelder, J.A. and Wedderburn, R.W.M. (1972) Generalized Linear Models. Journal of the Royal Statistical Society: Series A (General), 135, 370-384.
https://doi.org/10.2307/2344614
[38] Dempster, A.P., Laird, N.M. and Rubin, D.B. (1977) Maximum Likelihood from Incomplete Data via the EM Algorithm. Journal of the Royal Statistical Society: Series B (Methodological), 39, 1-22.
https://doi.org/10.1111/j.2517-6161.1977.tb01600.x
[39] McCullagh, P. and Nelder, J.A. (2019) Generalized Linear Models. Routledge, New York.
https://doi.org/10.1201/9780203753736
[40] Gan, G. and Valdez, E.A. (2018) Regression Modeling for the Valuation of Large Variable Annuity Portfolios. North American Actuarial Journal, 22, 40-54.
https://doi.org/10.1080/10920277.2017.1366863
[41] Bühlmann, H. (2007) Mathematical Methods in Risk Theory. Springer Science & Business Media, Berlin.
[42] Jacobs, J. (2014) Analyzing Ponemon Cost of Data Breach. Data Driven Security, 11, 5.
[43] Financial Services Varonis (2021) 2021 Financial Services Data Risk Report.
[44] Marsh McLennan (2021) Cyber Insurance Market Overview: Fourth Quarter 2021. Marsh Cyber Risk Report.
https://www.marsh.com/us/services/cyber-risk/insights/cyber-insurance-market-overview-q4-2021.html

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.