1. Introduction
Default forecasting is crucial for financial institutions and investors. Prior to investing in or extending credit to a company, investors and creditors must assess the company’s financial distress risk in order to avoid incurring a significant loss. In the literature on financial distress, default risk modelling can be grouped into two main categories: structural and reduced-form approaches. This paper uses the reduced-form method of correlated default timing. The interested readers may refer to
Nguyen and Zhou (
2023) for a general view of the literature on reduced-form models of correlated default timing.
Accounting-based measures are the first generation of reduced-form models for predicting the failure of a company. The earliest works predicting this type of financial distress are univariate analyses (
Beaver 1966,
1968), which employ financial ratios independently and adopt the cut-off point for each financial ratio in order to improve the precision of classifications for a distinct sample.
Altman (
1968) conducted a multivariate analysis of business failure based on multiple discriminant analyses by combining the data from numerous financial ratios from the financial statement into a singular weighted index. The second generation of default literature is the logistic model (
Ohlson 1980). This method was developed to deal with the shortcomings of the Altman Z-score method.
Shumway (
2001) attempts to predict defaults and shows that half of the accounting ratios utilized by
Altman (
1968) and
Zmijewski (
1984) have poor prediction on the default models, while a large number of market-driven independent variables are significantly associated with default probability. The recent expansion of reduced-form default risk models has centred on duration analysis.
Jarrow and Turnbull (
1995) and
Jarrow et al. (
1997) are the pioneers of term structure and credit spread modelling.
With regard to duration analysis, recent research indicates that observable macroeconomic and firm-specific factors may not be sufficient to characterize the variation in default risk, as corporate default rates are strongly correlated with latent factors. The need for and importance of the hidden factor in a default model are discussed in several recent studies, such as
Koopman and Lucas (
2008),
Duffie et al. (
2009),
Chava et al. (
2011),
Koopman et al. (
2011,
2012),
Creal et al. (
2014),
Azizpour et al. (
2018), and
Nguyen (
2023).
To improve the prediction accuracy of default models, the utilization of expert judgement in the decision-making process is common in practice, as there may not be enough statistically significant empirical evidence to reliably estimate the parameters of complicated models. This problem is considered to be of central interest in simulating a number of debates in the empirical literature regarding the issue of Bayesian inference. In the process of inference, however, the majority of Bayesian analyses utilize non-informative priors formed by formal principles. The theoretical foundation utilized by the majority of Bayesians is that of
Savage (
1971,
1972) and
De Finetti (
2017). Despite the fact that non-informative prior distribution plays a crucial role in defining the model for certain problems, it appears that there is an unavoidable drawback, as it is sometimes impossible to specify only non-informative priors and disregard the informative priors. It is observed that Bayes factors are sensitive to the selection of unknown parameters of informative prior distributions, which has a greater likelihood of influencing the posterior distribution. As a consequence, it generates debates regarding the selection of priors. Moreover, real prior information is beneficial for specific applications, whereas non-informative priors do not take advantage of this; consequently, such circumstances require informative priors. In other words, this is where subjective views and expert opinion are combined. Assuming we have a complex, high-dimensional posterior distribution, it is uncertain whether we have exhaustively summarized it. This should likely be completed by an experienced statistician. Choosing informative priors and establishing a connection with expert opinion are still the subject of debate in academic research, and interesting stories about them are still being continued. Recently, there has been research on the default prediction combined with expert opinion using machine learning techniques, such as by
Lin and McClean (
2001),
Kim and Han (
2003),
Zhou et al. (
2015), and
Gepp et al. (
2018). However, these studies adopt machine learning techniques with single classifiers.
Motivated by these findings, this paper aims to answer the research question of whether adding expert opinions to the frailty-correlated default risk model can give us better prediction results. To do so, we combine prior distributions to the frailty-correlated default model in
Duffie et al. (
2009) and adopt the Particle MCMC approach in
Nguyen (
2023) to estimate the unknown parameters and predict the default risk in the model using the dataset of U.S. public non-financial firms spanning 1980–2019. Our findings show that the 1-year prediction for frailty-correlated default models with different prior distributions is relatively good, whereas the prediction accuracy of models decrease significantly as the prediction horizons increase. The results also indicate that prediction accuracy ratios for frailty-correlated default models with non-informative and subjective prior distributions over various prediction horizons are not significantly different. Specifically, the out-of-sample prediction accuracy for the frailty-correlated default models with subjective prior distribution is slightly higher than that of the frailty-correlated default models with uniform prior distribution (95.00% for 1-year prediction, 85.23% for 2-year prediction, and 83.18% for 3-year prediction of the frailty default model with uniform prior distribution; and 96.05% for 1-year prediction, 86.32% for 2-year prediction, and 84.71% for 3-year prediction of the frailty default model with subjective prior distribution).
To obtain the research objective, the remainder of the paper is organized as follows:
Section 2 presents the econometric model and the estimation methodology for the frailty-correlated default models with the different prior distributions.
Section 3 reports major results. Data and the choice of covariates are also presented in this section.
Section 4 provides the model performance evaluation.
Section 5 presents the concluding remarks and limitations of the research.
2. Econometric Model
This part outlines the econometric model used by
Duffie et al. (
2009) and our extension to the model and improvement of the method to examine and forecast default risk at the firm level. We first provide an introduction to the notations used in this study. We consider a complete filtered probability space
, where the filtration
describes the flow of information over time and
P is a real-world probability measure. Further on, we use the standard convention where capital letters denote random variables, whereas lower case letters are used for their values.
The complete Markov state vector is described as , where we let be a Markov state vector of firm-specific and macroeconomic covariates, a vector of observable firm-specific covariates for firm i at the first observation time until the last observation time , be an unobservable firm-specific covariate, be a vector of observable macroeconomic variables at all times, and be an unobservable frailty (latent macroeconomic factor) variable; denote a vector of observable covariates for firms i at time t, where 1 is a constant.
On the event of
of survival to
t, given the information set
, the conditional probability of survival to time
is
and the conditional default probability at time
is of the form:
The information filtration of
includes the information set of the observed macroeconomic/firm-specific variables:
The complete information filtration contains the variables in the information filtration of and the frailty process .
The assumptions are imposed as follows:
Assumption 1. All firms’ default intensities at time t depend on a Markov state vector which is only partially observable.
Assumption 2. Conditional on the path of the Markov state process W determining the default intensities, the firm default times are the first event times of an independent Poisson process with time-varying intensities determined by the path of W. This is referred to as a doubly stochastic assumption.
Assumption 3. Set the level of mean-reversion of H, , the unobserved frailty process H is a mean-reverting Ornstein–Uhlenbeck (OU) process which is given by the stochastic differential equation below:where are parameters; is a standard Brownian motion with respect to ; η is a nonnegative constant, the speed of mean-reversion of H; σ is the volatility of the Brownian motion. In the general case, without Assumption
3, we would need extremely numerically intensive Monte Carlo integration in a high dimensional space due to our large dataset from 1980 to 2019. Thus, we assume process H is an OU process, as in
Duffie et al. (
2009).
The default intensity of a firm
i at the time
t is:
, where
is the component of the state vector at time
t and
is a parameter vector to be estimated;
is a parameter vector of the observable covariates
Z;
is a parameter of the frailty variable
,
is the speed of mean-reversion of
; and
is a Brownian motion parameter of
. The parameters
and
need to be estimated through a mean-reverting OU process, which we assume the unobserved frailty process
H will follow. The proportional hazards form is expressed by
is the default indicators of
m firms. Default indicator
of the firm
i at the time
t is defined as:
Now we will start with the conditional probability of the
m company. As mentioned above, we let
be the first observation time for firm
i and
is the last observation time for firm
i. For each firm
i and fixed time
t, we have
and then, in our case, the conditional probability of the individual firm is given by
Thus, the conditional probability of the
m firm is expressed as:
Applying Bayes’ theorem:
We have two cases for the prior distribution
: (i) Uniform prior and (ii) Subjective prior.
Prior distribution is uniform
where
(non-informative prior distribution). This case is exactly studied by
Duffie et al. (
2009). Our extension to the model by combining with priors is given below.
Prior distribution is subjective
where
is the multivariate normal prior with a mean vector
and a covariance matrix
.
If the observable covariate process
Z is independent of the frailty process
H, the likelihood function of intensity parameter vector
is given by
where
is the unconditional probability density of the unobservable frailty process
H.
Now we show how to transform the model with the frailty-correlated defaults to the one combined with the subjective prior distribution. We have found the posterior probability density earlier as
Taking the logarithm of Equation (
11)
Recall that the log-likelihood of parameter value
given the observable and hidden variables is given by
We proceed to take the logarithm for the second term of the Equation (
12)
In the second term, the central interest is the covariance matrix. For notational simplicity, set
. It is then rewritten as
Then, the second term can be rewritten as
We combine terms of Equation (
11) to obtain an overall likelihood function given the filtration
Now the central interest is to estimate Equation (
17). We used a Bayesian approach coupled with the Particle MCMC algorithm to estimate and forecast the frailty-correlated default models with uniform and subjective prior distributions. Particle filters can be understood as sequential Monte Carlo (SMC) methods, as introduced by
Handschin and Mayne (
1969) and
Handschin (
1970). Particles are a set of points in the sample space, and particle filters provide approximations to the posterior densities via these points. Each particle has an assigned weight, and then the posterior distribution can be approximated by a discrete distribution. Several algorithms about particle filters have been proposed in the literature review, and it can be said that the difference between algorithms consist in the way that a set of the particles evolves and adapts to inputs data. Algorithm 1 presents the Sequential Monte Carlo process we applied in our method.
Algorithm 1: Sequential Monte Carlo algorithm |
- (1)
Sample - (2)
Calculate and normalize the weights
At time :
- (1)
Resample the particles, i.e., sample the indices , - (2)
Sample and set
- (3)
Calculate and normalize the weights
|
One disadvantage of this approach is that the SMC approximation to
deteriorates when
is too large.
Andrieu et al. (
2010) have proposed the Particle PIMH method to overcome this difficulty. This is a class of MCMC using the SMC algorithm as its component to design its multi-dimensional proposal distributions. The advantage of this method is that the PIMH sampler does not call for the SMC algorithm to generate all samples which approximate
but only to choose a sample which can be approximated for
(see
Andrieu et al. 2010). Algorithm 2 presents the PIMH method applied in our model.
Algorithm 2: PIMH algorithm |
Set Sample by SMC Algorithm 1, Draw from Set For
- (1)
Sample by SMC Algorithm 1 Draw - (2)
Draw U with the uniform distribution (0, 1) If Set Set Else Set Set
|
In our method, we combine Particle MCMC with the maximum likelihood method to estimate the intensity parameter vector
for the frailty-correlated model. We present the implementation steps in Algorithm 3. See
Nguyen (
2023) for further discussions about the methods.
Algorithm 3: Particle MCMC Expectation-Maximization algorithm |
Initialize Set i := 0 Set , where is an estimate of κ in the model without the hidden factors Loop Set i := i + 1 Sample from by PIMH Algorithm 2 Employ the maximum likelihood method to estimate parameters from Equation (17) using generated samples Exit when achieving reasonable numerical convergence of the likelihood .
|
4. Out-of-Sample Performance and Robustness Check
To evaluate the model performance, we use the cumulative accuracy profile (CAP) and the accuracy ratio (AR). The companies are divided into two equal groups: estimation and evaluation. We estimate the parameters based on the estimation group and then evaluate the prediction accuracy using the evaluation group. The implementation steps are shown as follows: Firstly, we estimate parameters in the frailty-correlated default model with subjective prior distribution using the historical default rates in the period from 1981 to 2011. Secondly, using the estimation results obtained from Step 1, we forecast the data for the period from 2012 to 2018 based on the covariates time series model for observable firm-specific/macroeconomic covariates. Thirdly, we forecast the data of the frailty variable for the period (2012-2018) using the PIMH Algorithm 2. Fourthly, after obtaining the estimates from Step 1 and the data obtained from Steps 2 and 3, we compute the default probability based on Equation (
2). Lastly, we can determine a CAP and its associated AR. The CAPs and ARs for the out-of-sample prediction horizons are displayed in
Figure 1 and
Figure 2.
Table 5 reports the results of out-of-sample predictions of frailty-correlated models with uniform and subjective prior distributions. From two default models, it can be seen that the prediction ratios of the frailty-correlated default model with subjective prior distribution are higher than those of the model with uniform prior distribution. The out-of-sample prediction accuracy for 1-year prediction on average is good. Specifically, 95 percent for the frailty-correlated model with a non-informative prior distribution and 96.05 percent for the model with a subjective prior distribution. When the time horizon for predictions is extended to three years, the AR of the models suffers a significant decline, falling to 85.23 percent for frailty-correlated models with uniform prior distributions and 86.32 percent for those with subjective prior distributions. We also perform out-of-sample default predictions using the logistic regression method
1 to compare the accuracy with our proposed method in
Table 6. The results show that our method has better prediction power compared with the logistic regression method.
Overall, two notable conclusions can be drawn from these parameter estimation results: (i) The 1-year prediction for both models is good and when the prediction horizons increase, the prediction accuracy of the models decreases significantly. (ii) It can be seen that there has not been much difference about prediction accuracy ratios for frailty-correlated default models with non-informative and subjective prior distributions over three out-of-sample prediction horizons, including 2012–2018 for 1-year default distribution, 2013–2018 for 2-year default prediction, and 2014–2018 for 3-year default prediction.
To check the robustness of the estimation results for the frailty-correlated default model with subjective prior distribution, we estimated a subperiod from 1980 to 2011 as a sensitivity test. The outcomes correspond with the signs and magnitude of the entire sample. On the other hand, the value of log-likelihood in the frailty-correlated default model with subjective prior distribution (−2202.45) is larger than that in the frailty-correlated default model with non-informative prior distribution (−2379.61), which confirms that the frailty-correlated default model should incorporate the expert opinion.
5. Concluding Remarks and Limitations
Risk assessment is part of the decision-making process in many fields of discipline including finance. In the financial distress literature, the credit risk evaluation entails the evaluation of the hazard of potential future exposure or probable loss to lenders in the context of lending activities. The effective management of credit risk is a crucial aspect of risk management and crucial to the long-term survival of any bank. Credit risk management’s objective is to maximize the bank’s risk-adjusted return by keeping credit risk exposure within acceptable limits. The ability to accurately forecast a company’s financial distress is a major concern for many stakeholders. This practical relevance has motivated numerous studies on the topic of predicting corporate financial distress. To improve the prediction accuracy of default models, the utilization of expert judgement in the decision-making process is common in practice as there may not be enough a statistically significant amount of empirical evidence to reliably estimate parameters of complicated models. This problems is considered to be of central interest of simulating a number of debates in the empirical literature regarding the issue of Bayesian inference.
This paper proposes a method to add expert judgement to the frailty-correlated default risk model in
Duffie et al. (
2009) by incorporating subjective prior distributions into the model. Then we employ the Bayesian method coupled with a Particle MCMC approach in
Nguyen (
2023) in order to evaluate the unknown parameters and predict the default risk models on a historical defaults dataset of 424,601 firm-month observations from January 1980 to June 2019 of 2432 U.S. industrial firms. We compare the prediction results of the frailty-correlated default risk model with uniform and subjective prior distributions together. The findings show that the 1-year prediction for both models are pretty good and the prediction accuracy of models decrease considerably as the prediction horizons increase. The results also indicate that prediction accuracy ratios for frailty-correlated default models with non-informative and subjective prior distributions over various prediction horizons are not significantly different. Specifically, the out-of-sample prediction accuracy for the frailty-correlated default models with uniform distribution is slightly higher than that of the frailty-correlated default models with informative prior distribution over three out-of-sample prediction horizons, including 2012–2018 for the 1-year default distribution, 2013–2018 for the 2-year default prediction, and 2014–2018 for the 3-year default prediction.
The frailty-correlated default model with expert opinion has been designed to estimate and predict the default risk of corporations. The model can be adapted to accommodate any context. However, the model also has its limitations. Firstly, one of the main limitations is that we cannot access inputs of data for expert opinion; therefore, to some certain extent, our results also depend on how we assume the values of priors. Accordingly, the prediction accuracy can be slightly different. It is observed that Bayes factors are sensitive to the selection of unknown parameters from informative prior distributions, which has a greater likelihood of influencing the posterior distribution. As a consequence, it generates debates regarding the selection of priors. According to
Kass and Raftery (
1995), non-informative priors may also contribute to posterior estimate instability and convergence of the sampler algorithm. Choosing informative priors and establishing a connection with expert opinion are still the subject of debate in academic research, and interesting stories about them are still being continued. Therefore, future work should use actual data of expert opinion, which may be feasibly conducted in the age of big data. Recently, there have been research on the default prediction combined with expert opinion using machine learnings, such as
Lin and McClean (
2001),
Kim and Han (
2003),
Zhou et al. (
2015), and
Gepp et al. (
2018). However, these studies adopt machine learning techniques with single classifiers and observable variables. Future work can adopt a meta-learning framework to examine and predict defaults with expert opinion at the firm level.