Next Article in Journal
Financial Risk Disclosure and Financial Attributes among Publicly Traded Manufacturing Companies: Evidence from Bangladesh
Next Article in Special Issue
Risk, Return and Volatility Feedback: A Bayesian Nonparametric Analysis
Previous Article in Journal
Can Bitcoin Replace Gold in an Investment Portfolio?
Previous Article in Special Issue
On the Performance of Wavelet Based Unit Root Tests
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Monte Carlo Comparison for Nonparametric Threshold Estimators

Department of Economics and Finance, University of Guelph, Guelph, ON N1G 2W1, Canada
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2018, 11(3), 49; https://doi.org/10.3390/jrfm11030049
Submission received: 17 July 2018 / Revised: 13 August 2018 / Accepted: 15 August 2018 / Published: 17 August 2018
(This article belongs to the Special Issue Nonparametric Econometric Methods and Application)

Abstract

:
This paper compares the finite sample performance of three non-parametric threshold estimators via the Monte Carlo method. Our results indicate that the finite sample performance of the three estimators is not robust to the position of the threshold level along the distribution of the threshold variable, especially when a structural change occurs at the tail part of the distribution.

1. Introduction

Popularly used to describe structural changes in economic relationships, threshold models have seen many applications, especially in macro fields (e.g., Hansen 2011; Potter 1995). Typical examples include the nonlinearity in public debt to GDP ratio (e.g., Afonso and Jalles 2013; Caner et al. 2010; Cecchetti et al. 2011). A number of threshold estimators for threshold models have been proposed in the literature, and the asymptotic results of these estimators can be categorized into two groups based on different assumptions. The first group is based on the “fixed threshold effect” assumption. The second group imposes a “diminishing threshold effect” assumption introduced by Hansen (2000). For example, it is well known that, for the least-squares estimator, the threshold estimator is super-consistent with the convergence rate n under the “fixed threshold effect” assumption and n 1 2 α under the “diminishing threshold effect” assumption, respectively, where α measures the diminishing rate of the threshold effect.
The asymptotic theory and statistical inference have been well developed for the least-squares estimator exogenous regressors and exogenous threshold variable (e.g., Chan 1993; Hansen 2000; Seo and Linton 2007). Recently, there has been a growing interest in studying threshold models with endogenous regressors and/or a threshold variable. Extending the framework of Hansen (2000), Caner and Hansen (2004) applied the two-step least-squares method to estimate threshold models with endogenous slope regressors. In the spirit of the sample selection technique of Heckman (1979), imposing the joint normality assumption, Kourtellos et al. (2016) explored the case that both the threshold variable and slope regressors are endogenous. The work in Seo and Shin (2016) proposed a two-step GMM estimator for a dynamic panel threshold model with fixed effects, which allows endogeneity in both the slope regressors and threshold variable. It is worth noticing that the GMM method allows both a fixed and diminishing threshold effect, and the convergence rate for the GMM threshold estimator is not super-consistent. By relaxing the joint normality assumption of Kourtellos et al. (2016, 2017), a two-step least square estimator based on a nonparametric control function approach to correct the threshold endogeneity was proposed. The semiparametric threshold model separates the threshold effect into two parts, namely the exogenous threshold effect and endogenous threshold bias-correction term. Therefore, with a “small threshold” effect, the convergence rate for the threshold variable depends on diminishing rates of the threshold effect and the bias-correction term.
However, few studies have worked on the estimation and statistical inference of threshold estimators based on nonparametric estimation methods, which do not rely on the least square method. The work in Delgado and Hidalgo (2000) suggested a difference kernel estimator (or DKE), which depends on a chosen point. The convergence rate of Delgado and Hidalgo (2000) DKE is n h d 1 , which depends on both the bandwidth, h, and the dimensionality of regressors in their threshold model, d 1 . Built upon the method of Delgado and Hidalgo (2000), Yu et al. (2018) introduced an integrated difference kernel estimator (or IDKE). The work in Yu et al. (2018) argued that the IDKE can be applied to the case with the endogenous threshold variable. The convergence rate of the IDKE is not related to either the bandwidth or the dimensionality of regressors and is super-consistent with the rate n. Using recently-developed discrete smoothing methods, Henderson et al. (2017) introduced a semiparametric M-estimator of a nonparametric threshold regression model. The threshold estimator of Henderson et al. (2017) can be estimated at the rate n / h (h is the bandwidth), which is faster than the parametric convergence rate of n . One may notice that the aforementioned convergence rate is the same as that of the smoothed least squares estimator in Seo and Linton (2007). However, they are entirely different. The work in Henderson et al. (2017) focussed on the nonparametric threshold model, and their proposed estimator was based on a non-smooth objective function. On the contrary, Seo and Linton (2007) worked on a linear threshold model, and the proposed estimator was based on a smooth objective function with the indicator function replaced by a CDF-type smooth function.
With many applications and simulations available for comparing the parametric threshold estimators in the literature, little guidance is available for researchers to apply as to the choice of nonparametric threshold estimators. Moreover, to avoid the boundary effect of the threshold estimator, most simulations are designed deliberately with the true threshold level chosen at the middle point of the threshold variable distribution, which can be highly doubted in reality. Therefore, the purpose of this paper is to carefully compare the three nonparametric threshold estimators mentioned above using the Monte Carlo method. More importantly, we consider the case that the true threshold level is not only at the middle, but also at the two tails of the threshold variable distribution.
The rest of the paper is organized as follows. In Section 2, we briefly review the estimation procedure of three nonparametric threshold estimators such as DKE, IDKE and the M-estimator, where threshold models have exogenous regressors and a threshold variable. In Section 3, we illustrate the possible theoretical reason for the conjecture of the poor finite sample performance of the difference kernel-type estimators. Section 4 presents the design of the Monte Carlo simulations. Section 5 reports the finite sample performance. Section 6 concludes.

2. Three Nonparametric Threshold Estimators

In this paper, we aim to compare the finite sample performance of three nonparametric threshold estimators: Henderson et al. (2017) the semiparametric M-estimator, Delgado and Hidalgo (2000) the difference kernel estimator (DKE) and Yu et al. (2018) the integrated difference kernel estimator (IDKE).
Following Henderson et al. (2017), we consider a generalized threshold regression model:
y i = α 0 ( X i ) + β 0 I { q i > γ 0 } + ε i ,
for i = 1 , , n , where α 0 ( · ) is an unknown smooth function, X i is a vector of d regressors, q i is the threshold variable, γ 0 is the threshold level, I ( · ) is the indicator function and β 0 measures the jump size of the regression function at q > γ . Furthermore, X i and q i are both exogenous and may have a common variable.

2.1. Semiparametric M-Estimator

If γ 0 is known a priori, Model (1) is known as a partially linear model. The conventional method to estimate the unknown γ 0 is minimizing the sum of squared errors, which can be iterated by the grid search. Therefore, Henderson et al. (2017) suggested the semi-parametric M-estimator of the nonparametric threshold model, which can be obtained in three steps.
In Step 1, given ( β , γ ) , Model (1) becomes a standard nonparametric model. Therefore, we can obtain the Nadaraya–Watson (NW) estimator of α 0 x at an interior point, x, i.e.,
α ^ ( x ; β , γ ) = a r g m i n α Θ α n 1 i = 1 n [ y i α β I { q i > γ } ] 2 K h ( X i x ) ,
where K h ( X i x ) = h d j = 1 d k ( X i j x j h ) , X i = [ X i , 1 , , X i , d ] , x = [ x 1 , , x d ] , k ( · ) is a second order kernel function, h is the bandwidth and d is the dimension of x.
In Step 2, given γ , Model (1) becomes a partially linear model. Then, β 0 can be estimated as:
β ^ ( γ ) = a r g m i n β Θ β n 1 i = 1 n [ y i α ^ ( X i ; β , γ ) β I { q i > γ } ] 2 f ^ h 2 ( X i ) ,
where f ^ h ( X i ) = n 1 i = 1 n K h ( X i x ) works as the weighting function.
The work in Henderson et al. (2017) shows that β ^ ( γ ) has the following mathematical expression:
β ^ ( γ ) = n 1 i = 1 n j = 1 n K h ( X i X j ) ( I i I j ) 2 1 n 1 i = 1 n j = 1 n K h ( X i X j ) ( I i I j ) j = 1 n K h ( X i X j ) ( y i y j ) ,
where we denote I i = I ( q i > γ ) .
In Step 3, we can estimate the threshold level γ 0 by solving the following optimization problem,
γ ^ = a r g m i n γ Θ γ n 1 i = 1 n y i α ^ ( X i ; β ( γ ) , γ ) β ^ ( γ ) I { q i > γ } w ( X i ) ,
where w ( · ) is a weighting function and is application dependent.
As mentioned in Section 1, the convergence rate of the threshold estimator of Henderson et al. (2017) is n / h , which explodes faster than the usual parametric n rate. However, the unknown function α 0 · and the jump size β 0 converge at standard nonparametric rates of n h d and n h , respectively.

2.2. DKE and IDKE

Instead of using the absolute value of the weighted average of the sum of errors as the objective function, Delgado and Hidalgo (2000) considered using the difference between E ^ [ y | x 0 , q = γ ] and E ^ [ y | x 0 , q = γ + ] as the objective function. Ideally, the closer γ approaches the true value, the larger the absolute value of the above difference should be. As a result, we are able to estimate the threshold level by choosing γ , which gives the most considerable gap between the two one-sided expectations. Therefore, the difference kernel estimator (DKE) can be obtained by:
γ ^ D K E = a r g m a x γ Θ γ 1 n i = 1 n y i K h , i γ 1 n i = 1 n y i K h , i γ + 2
where we have:
K h , i γ + = K h ( X i x 0 ) · k h + ( q i γ ) ,
K h , i γ = K h ( X i x 0 ) · k h ( q i γ ) ,
if q i is not part of X i , and
K h , i γ + = K h ( X 1 i x 10 ) · k h + ( q i γ ) ,
K h , i γ = K h ( X 1 i x 10 ) · k h ( q i γ ) ,
if q i is part of X i , i.e., X i = [ X 1 i , q i ] , and x 0 = [ x 10 , q 0 ] . Furthermore, k h + / ( · ) is the one-sided kernel function with:
k h + ( q i γ ) = k ( q i γ h ) I ( q i > γ ) ,
k h ( q i γ ) = k ( q i γ h ) I ( q i γ ) ,
and k ( · ) is a second order kernel function.
Obviously, it is reasonable to expect that the DKE estimator is sensitive to the choice of x 0 . Furthermore, the DKE suffers the curse of dimensionality problem as the convergence rate of the DKE, n h d 1 , depends on the dimension of the regressor. To fix these potential weaknesses, Yu et al. (2018) proposed an integrated difference kernel estimator, which allows γ ^ not to rely on the single choice in x 0 , but the expectation of all X. The γ ^ I D K E can be derived as follows:
γ ^ I D K E = a r g m a x γ Θ γ n 1 i = 1 n 1 n 1 j = 1 , j i n y j K h , i j γ 1 n 1 j = 1 , j i n y j K h , i j γ + 2 ,
where:
K h , i j γ + = K h ( X i x j ) · k h + ( q i γ ) ,
K h , i γ = K h ( X i x j ) · k h ( q i γ ) ,
if q i is not part of X i , and
K h , i γ + = K h ( X 1 i x 1 j ) · k h + ( q i γ ) ,
K h , i γ = K h ( X 1 i x 1 j ) · k h ( q i γ ) ,
if q i is part of X i , i.e., X i = [ X 1 i , q i ] , and x j = [ x 1 j , q j ] . k h + / ( · ) is defined the same as above.
The IDKE is super-consistent with convergence rate n. The work in Yu et al. (2018) showed that IDKE is consistent even if the threshold variable is endogenous. They explain that the role of the instruments of the endogenous regressors and the endogenous threshold variable is improving only the efficiency of the IDKE.

3. Estimation Difficulties in the Difference Kernel-Type Estimator with Near Boundary γ 0

In this section, we use a simple version of Model (1) to explain the estimation difficulties of the difference kernel-type estimators when γ 0 lies at the tails of the threshold variable distribution. This estimation difficulty motivates us to investigate the position effect of the true threshold level on the finite sample performance. Specifically, we consider the true model as:
y i = I ( X i γ 0 ) ,
where X i is randomly drawn from a uniform distribution over the interval of [ 0.5 , 0.5 ] for i = 1 , , n .
The model above can be regarded as Model (1) with α 0 ( x ) 0 , β 0 = 1 , and ε i = 0 for all i = 1 , , n . Therefore, the DKE is based on the objective function:
Q ^ n ( γ ) D K E = 1 n i = 0 n k X i γ h I ( X i < γ ) y i 1 n i = 0 n k X i γ h I ( X i γ ) y i 2 .
Letting u x = X i γ / h and applying the change of variables, we have the probability limit of Q ^ n ( γ ) equal to:
Q n ( γ ) D K E = h 2 0.5 γ h 0.5 γ h k ( u x ) I ( u x < 0 ) I u x γ 0 γ h d u x 0.5 γ h 0.5 γ h k ( u ) I ( u x 0 ) I u x γ 0 γ h d u x 2 ,
where h is the bandwidth.
If γ < γ 0 , we obtain:
Q n ( γ ) D K E = h 2 0.5 γ h 0.5 γ h k ( u x ) d u x 2 ,
and:
Q n ( γ ) D K E γ = 2 h γ 0 γ h 0.5 γ h k ( u x ) d u x k γ 0 γ h k 0.5 γ h > 0 ,
where the positive sign follows for all γ 0 < 0.5 for any second-order kernel function with a bell shape.
It is worth noting that as γ 0 approaches 0.5 from the left side, the difference between k ( γ 0 γ h ) k ( 0.5 γ h ) becomes smaller. As a result, for all γ , the above derivative goes to zero, which makes the objective function flat and leads to the estimation difficulty.
Similarly, if γ > γ 0 , we have:
Q n ( γ ) D K E = h 2 γ 0 γ h 0 k ( u x ) d u x 0 0.5 γ h k ( u x ) d u x 2 ,
and:
Q n ( γ ) D K E γ = 2 h γ 0 γ h 0 k ( u x ) d u x 0 0.5 γ h k ( u x ) d u x k γ 0 γ h + k 0.5 γ h < 0 ,
where the negative sign follows for all γ 0 > 0.5 for any second-order kernel function with a bell shape.
Therefore, we observe that as γ 0 approaches 0.5 from the right side, for all γ , the difference between γ 0 γ h 0 k ( u x ) d u x 0 0.5 γ h k ( u x ) d u x becomes smaller, which makes the derivative go to zero, and this results in a flat objective function.
In summary, the DKE is asymptotically consistent with γ 0 ( 0.5 , 0.5 ) . However, it is reasonable to suspect that DKE may have poor finite performance with the true threshold level lying at the tails of the threshold variable distribution due to the estimation difficulty of the flat objective function.
Next, we assume that there are additional covariates, Z i , which are randomly drawn from uniform distribution over the interval of [ 0.5 , 0.5 ] , for all i = 1 , , n , and { X i } and { Z i } are independent. Therefore, the probability limit of the objective function of the IDKE is (with the same bandwidth):
Q n ( γ ) I D K E = h 4 0.5 0.5 0.5 γ h 0.5 γ h k ( u z ) k ( u x ) I ( u x < 0 ) I ( u x γ 0 γ h ) d u x d u z 0.5 z 0 h 0.5 z 0 h 0.5 γ h 0.5 γ h k ( u z ) k ( u x ) I ( u x 0 ) I ( u x γ 0 γ h ) d u x d u z 2 d z 0
where u z = Z i z 0 h .
Note that:
Q n ( γ ) I D K E γ = h 2 0.5 0.5 0.5 z 0 h 0.5 z 0 h k ( u z ) d u z 2 d z 0 Q n ( γ ) D K E γ .
Consequently, in this typical example, Q n ( γ ) I D K E γ can be interpreted as a rescaled Q n ( γ ) D K E γ , which implies the IDKE will suffer the same boundary problem as the DKE estimator.

4. Monte Carlo Designs

To assess the finite sample performance of the three nonparametric threshold estimators, we consider seven data-generating mechanisms, which are similar to those studied in Henderson et al. (2017); Yu et al. (2018).
  • DGP 1:
    y i = 2 I ( x i γ 0 ) + ε i
  • DGP 2:
    y i = x i + 2 I ( x i γ 0 ) + ε i
  • DGP 3:
    y i = s i n ( x i ) + 2 I ( x i γ 0 ) + ε i
  • DGP 4:
    y i = x i 2 + 2 I ( x i γ 0 ) + ε i
  • DGP 5:
    y i = x 1 i + x 2 i + x 3 i + 2 I ( x 1 i γ 0 ) + ε i
  • DGP 6:
    y i = x 1 i 2 + x 2 i x 3 i + 2 I ( x 1 i γ 0 ) + ε i
  • DGP 7:
    y i = s i n ( x 1 i ) + c o s ( x 2 i ) + s i n ( x 3 i ) + 2 I ( x 1 i γ 0 ) + ε i
where x i is randomly drawn from a uniform distribution over the interval of [ 0.5 , 0.5 ] for all i = 1 , , n ,1 and ε i is randomly drawn from the N ( 0 , 1 ) distribution. All DGPs are based on the fixed threshold effect framework of Chan (1993) with both the exogenous threshold variable and exogenous regressors.
DGPs 1–4 are univariate threshold models. More specifically, DGPs 1–2 are typical linear threshold models. DGPs 3–4 are nonlinear threshold models modelling the periodicity and the quadraticity, respectively. DGPs 5–7 are multivariate threshold models. DGP 5 characterizes the multivariate linear threshold model. DGPs 6–7 are nonlinear threshold models extending DGPs 3–4 to multivariate specifications.
To examine the position effect of the true threshold level on the finite sample performance, we set γ 0 at different segments of the threshold variable distribution. Specifically, we set the true threshold, γ 0 , as the p th quantile of the threshold variable with p = 25 , 50 and 75 to place the true threshold level to the left tail, middle and the right tail of the threshold variable distribution, respectively.
We set x 0 = x m a x for the DKE estimate of Delgado and Hidalgo (2000), where x m a x is the data with the greatest empirical density among all generated x i ’s for each simulation of each DGP.2 We use the rule of thumb bandwidth, h = C σ ^ x n 1 / ( d + 4 ) , where C = 4 d + 2 1 d + 4 , d is the dimension of x i and σ ^ x is the sample standard deviation of { x i } . We use the Gaussian kernel function. As suggested by Yu et al. (2018), we use the one-sided rescaled Epanechnikov kernel with k ( q , 0 ) = 3 4 ( 1 q 2 ) I ( q < 0 ) and k + ( q , 0 ) = k ( q , 0 ) to estimate the DKE and the IDKE.
We repeat 2000 times for each simulation.3 We set the sample size n = 100 , 300 and 500. For each simulation, we report the average bias, mean squared error (or MSE) and the standard deviation (or stdev) of the threshold estimates. Table 1, Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7 contain the details of the simulation results. Table 8 shows the realized convergence rate of the semi-parametric M-estimator of Henderson et al. (2017) and IDKE of Yu et al. (2018).

5. Monte Carlo Results

For the semi-parametric M-estimator introduced by Henderson et al. (2017), our results show that the performance was slightly affected by the position of the true threshold level. Meanwhile, as the sample size increased, this position effect gradually vanished.4 In addition, we observed that the bias was smaller for multivariate models than univariate models. Using the bandwidth as defined in Section 4, which behaved roughly as O ( n 1 / 5 ) for univariate models and O ( n 1 / 7 ) for multivariate models, the theoretical convergence rates were O ( n 1.2 ) and O ( n 1.14 ) accordingly. From Table 8, the super-consistency was confirmed with the estimated convergence rate of γ ^ . Consistent with the theory, the realized convergence rate decreased as the dimension increased. It is quite interesting that, for almost all univariate models, the realized convergence rate of γ ^ was faster when γ 0 was at the left- and right-tail position than when γ 0 was at the median position. However, for multivariate models, the realized rates seemed to be stable with the position of γ 0 .
For the DKE, as we conjectured, it was severely affected by the position of the true threshold value for all DGPs, which may result from the estimation difficulties, as we argued in Section 3. Furthermore, even with the middle-positioned γ 0 , the bias still showed a non-decreasing pattern with the increasing sample size under some multivariate specifications.5 Intuitively, this may result from the choice of x 0 , which distorts the result by providing useless information. According to the comment in the Supplementary Material of Yu et al. (2018), the choice of x 0 is crucial in identifying the DKE estimator. On the one hand, the optimal x 0 should make [ E ( y | x 0 , q = γ 0 ) E ( y | x 0 , q = γ 0 + ) ] 2 as large as possible. On the other hand, one needs the conditional density f ( x 0 | q = γ 0 ) to be large enough to provide sufficient information. Therefore, theoretically, with a uniform distribution and univariate linear threshold model as in DGP2, the ideal x 0 should be at the middle of its distribution with the value of zero. However, in the simulation, we set x 0 equal to the value with the largest empirical density, which may appear at the two tails. This may lead to [ E ( y | x 0 , q = γ 0 ) E ( y | x 0 , q = γ 0 + ) ] 2 approaching zero. Moreover, with the multivariate and nonlinear specification, we can expect more distortion involved. As a result, the DKE performs the worst among all three competitors for all DGPs.
For the IDKE, our results reveal several features. Firstly, the IDKE was affected by the position of the actual threshold value. The influence was not as substantial as the DKE. Indeed, the integration allowed more local information to be used and alleviated the possible distortion due to the choice of x 0 . Surprisingly, unlike the DKE, this position effect seemed to be asymmetric for the IDKE. For most of the DGPs, we observed that the absolute value of the average bias and MSE was larger with the left-tailed γ 0 than the right-tailed γ 0 . The theoretical convergence rate of the IDKE estimator, n, is not related to either the bandwidth or the dimension, which is faster than the semi-parametric M-estimator of Henderson et al. (2017). This is consistent with our realized convergence rates, which are shown in Table 8. Moreover, for all DGPs, the realized convergence rates were faster with two-sided tailed γ 0 than the median γ 0 .
In summary, the simulation results give some evidence that the finite sample performances were affected by the position of the true threshold level for all three nonparametric threshold estimators. However, this effect was heterogeneous. The position effect least influenced the semi-M estimator of Henderson et al. (2017). Meanwhile, the difference kernel-type estimators were severely distorted by the tailed γ 0 , which confirms our conjecture made in Section 3. Furthermore, our results show that the position of the true threshold level also affects the realized convergence rate. We also found, for the semi-M estimator of Henderson et al. (2017) and the IDKE estimator, the tail distortion tended to be reduced in multivariate models.
As a robustness check of our findings, Figure 1, Figure 2, Figure 3 and Figure 4 show the simulation results of DGP 2 and DGP 5 with γ 0 taking different positions along the threshold variable distribution. It is obvious that, for all figures, semi-M had lower average bias in absolute value than difference kernel-type estimators with tail γ 0 . Furthermore, we found the gap between the average bias of the semi-M estimator and the average bias of the difference kernel-type estimators to drop greatly with γ 0 approaching the middle position of the threshold variable distribution.

6. Conclusions

In this paper, we evaluated the finite sample performance of three non-parametric threshold estimators and identified the relationship between the performances of different estimators and the position of the true threshold level with Monte Carlo methods.
The study shows, with all three estimators affected by the tail position of the true threshold value, that the semi-M estimator of Henderson et al. (2017) outperformed DKE and IDKE for roughly all DGPs considered in the paper. Interestingly, there appears to be some evidence that the distortion can be reduced if there are other covariates besides the threshold variable for the semi-M estimator and the IDKE. Consistent with the theory, we find that the realized convergence rates support the super-consistency in the threshold estimate for all three estimators. However, we find that the realized convergence rates are also affected by the position of the true threshold value. We therefore conclude that, in applied works, using the difference kernel-type estimation, researchers must be careful when the threshold estimate is at the left-tail or the right-tail of the threshold variable distribution.

Author Contributions

The two authors both contribute to the project formulation and paper preparation.

Funding

This research received no external funding.

Acknowledgments

We thank three anonymous referees for their helpful and constructive comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Afonso, Antonio, and Joao Tovar Jalles. 2013. Growth and productivity: The role of government debt. International Review of Economics and Finance 25: 384–407. [Google Scholar] [CrossRef]
  2. Caner, Mehmet, and Bruce Hansen. 2004. Instrumental variable estimation of a threshold model. Econometric Theory 20: 813–43. [Google Scholar] [CrossRef]
  3. Caner, Mehmet, Thomas J. Grennes, and Friederike N. Koehler-Geib. 2010. Finding the Tipping Point—When Sovereign Debt Turns Bad. Policy Research WP, No. 5391. New Delhi: Policy Research, pp. 1–13. [Google Scholar]
  4. Cecchetti, Stephen G., Madhusudan Mohanty, and Fabrizio Zampolli. 2011. The Real Effects of Debt. Working paper. Basel, Switzerland: Bank for International Settlements. [Google Scholar]
  5. Chan, Kung-Sik. 1993. Consistency and Limiting Distribution of the Least Squares Estimator of a Threshold Autoregressive Model. Annals of Statistics 21: 520–33. [Google Scholar] [CrossRef]
  6. Delgado, Miguel A., and Javier Hidalgo. 2000. Nonparametric inference on structural breaks. Journal of Econometrics 96: 113–44. [Google Scholar] [CrossRef] [Green Version]
  7. Hansen, Bruce E. 2000. Sample splitting and threshold estimation. Econometrica 68: 575–603. [Google Scholar] [CrossRef]
  8. Hansen, Bruce E. 2011. Threshold Autoregression in Economics. Statistics and Its Interface 4: 123–27. [Google Scholar] [CrossRef]
  9. Heckman, James J. 1979. Sample Selection Bias as a Specification Error. Econometrica 47: 153–61. [Google Scholar] [CrossRef]
  10. Henderson, Daniel J., Christopher F. Parmeter, and Liangjun Su. 2017. Nonparametric Threshold Regression: Estimation and Inference. Working paper. Singapore: Research Collection School of Economics. [Google Scholar]
  11. Kourtellos, Andros, Thanasis Stengos, and Chih Ming Tan. 2016. Structural Threshold Regression. Econometric Theory 32: 827–60. [Google Scholar] [CrossRef]
  12. Kourtellos, Andros, Thanasis Stengos, and Yiguo Sun. 2017. Endogeneity in Semiparametric Threshold Regression. Working paper. Guelph, ON, Canada: University of Cyprus and University of Guelph. [Google Scholar]
  13. Potter, Simon M. 1995. A Nonlinear Approach to US GNP. Journal of Applied Econometrics 10: 109–25. [Google Scholar] [CrossRef]
  14. Seo, Myung Hwan, and Oliver Linton. 2007. A smoothed least squares estimator for threshold regression models. Journal of Econometrics 141: 704–35. [Google Scholar] [CrossRef]
  15. Seo, Myung Hwan, and Yongcheol Shin. 2016. Dynamic Panels with Threshold Effect and Endogeneity. Journal of Econometrics 195: 169–86. [Google Scholar] [CrossRef]
  16. Yu, Ping, and Peter C. B. Phillips. 2018. Threshold Regression with Endogeneity. Journal of Econometrics 203: 50–68. [Google Scholar] [CrossRef]
1.
With the uniform distribution, the intensity of the Poisson process would not change with the change in the true threshold location. Therefore, the limiting distribution of both the DKE and the IDKE is not affected given γ 0 is not on the boundary of Θ γ .
2.
The theoretical density should be the same for all x due to the uniform distribution. The reason we use the data-driven choice of x 0 is because we do not know the true density in reality.
3.
All programming is finished in Matlab.
4.
With n = 100, the bias, MSE and standard deviation were larger with γ 0 placed at two tails and γ 0 placed at the median point. However, with n = 500, there was no apparent difference between tail position γ 0 estimation and the median position γ 0 estimation.
5.
For example, in Table 6, the bias monotonically increases with the in sample size.
Figure 1. Average bias with γ 0 as various quantiles of the threshold variable, DGP 2, n = 100. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 2. The sample size is 100.
Figure 1. Average bias with γ 0 as various quantiles of the threshold variable, DGP 2, n = 100. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 2. The sample size is 100.
Jrfm 11 00049 g001
Figure 2. Average bias with γ 0 as various quantiles of the threshold variable, DGP 2, n = 300. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 2. The sample size is 300.
Figure 2. Average bias with γ 0 as various quantiles of the threshold variable, DGP 2, n = 300. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 2. The sample size is 300.
Jrfm 11 00049 g002
Figure 3. Average bias with γ 0 as various quantiles of the threshold variable, DGP 5, n = 100. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 5. The sample size is 100.
Figure 3. Average bias with γ 0 as various quantiles of the threshold variable, DGP 5, n = 100. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 5. The sample size is 100.
Jrfm 11 00049 g003
Figure 4. Average bias with γ 0 as various quantiles of the threshold variable, DGP 5, n = 300. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 5. The sample size is 300.
Figure 4. Average bias with γ 0 as various quantiles of the threshold variable, DGP 5, n = 300. This figure shows absolute values of the average bias with the true threshold level being several quantiles of the threshold variable (5th, 10th, 20th, 40th, 50th, 60th, 80th, 90th, 95th). The simulation is based on DGP 5. The sample size is 300.
Jrfm 11 00049 g004
Table 1. Simulation results of nonparametric threshold estimators, Data-generating Mechanism 1 (DGP 1). IDKE, integrated difference kernel estimator.
Table 1. Simulation results of nonparametric threshold estimators, Data-generating Mechanism 1 (DGP 1). IDKE, integrated difference kernel estimator.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.03360.27050.06790.01440.09130.02250.11520.13450.1338
3000.00150.29290.08700.00060.09860.03080.02410.11330.1525
5000.00020.26320.15300.00010.09200.05440.00970.15090.1760
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.0056−0.0346−0.01830.00840.01540.00120.09160.11910.0288
3000.0007−0.0346−0.00830.00090.02090.00020.03020.14060.0126
5000.0008−0.0347−0.00550.00030.02330.00010.01660.14880.0080
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0397−0.2485−0.06660.01630.10820.00870.12150.21560.0650
300−0.0028−0.2590−0.03770.00090.11430.00290.02990.21740.0391
500−0.0004−0.2841−0.02870.00010.12880.00180.01180.21930.0308
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the simple jump function defined as Equation (17). The first column gives the sample size that the simulation used. The third to fifth columns report the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 2. Simulation results of nonparametric threshold estimators, DGP 2.
Table 2. Simulation results of nonparametric threshold estimators, DGP 2.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.03590.22720.08130.01540.08230.02500.11900.17520.1357
3000.00530.26800.10190.00200.09540.03240.04420.15360.1485
5000.00020.26320.15300.00010.09200.05440.00970.15090.1760
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0008−0.0246−0.01510.00820.01220.00090.09070.10770.0257
3000.0002−0.0147−0.00670.00090.01300.00020.03060.11300.0107
5000.0002−0.0131−0.00440.00000.01540.00010.00680.12330.0073
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0307−0.2465−0.10310.01190.10490.01590.10480.21010.0730
300−0.0059−0.2564−0.07860.00230.10090.00860.04770.18760.0494
500−0.0008−0.2651−0.06990.00030.10600.00650.01770.18910.0397
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the univariate linear threshold model defined as Equation (18). The first column gives the sample size that the simulation used. The third to fifth columns report the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 3. Simulation results of nonparametric threshold estimators, DGP 3.
Table 3. Simulation results of nonparametric threshold estimators, DGP 3.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.03030.22110.07850.01280.07910.02330.10920.17390.1310
3000.00220.27250.11370.00140.09800.03730.03760.15410.1561
5000.00050.26940.15700.00020.09610.05460.01310.15350.1730
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.0017−0.0236−0.01370.00730.01110.00080.08520.10270.0257
3000.0002−0.0220−0.00610.00040.01320.00010.01960.11280.0101
500−0.0003−0.0114−0.00410.00010.01490.00010.01120.12150.0067
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0358−0.2471−0.10360.01600.10310.01600.12120.20510.0725
300−0.0027−0.2592−0.08220.00130.10410.00910.03600.19240.0482
500−0.0007−0.2637−0.06860.00040.10310.00650.02030.18320.0422
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the univariate threshold periodic model defined as Equation (19). The first column gives the sample size that the simulation used. The third to fifth report propose the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 4. Simulation results of nonparametric threshold estimators, DGP 4.
Table 4. Simulation results of nonparametric threshold estimators, DGP 4.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.03710.27540.10380.01680.09220.03480.12420.12780.1551
3000.00650.28170.14790.00300.09210.05260.05450.11310.1754
5000.00100.28840.21460.00050.09740.07940.02210.11960.1826
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.0050−0.0324−0.01730.00860.01560.00160.09300.12050.0355
300−0.0010−0.0408−0.00710.00120.02120.00020.03410.14000.0135
5000.0000−0.0340−0.00510.00000.02220.00010.00380.14510.0086
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0378−0.2562−0.06940.01570.11050.00890.11960.21200.0640
300−0.0025−0.2622−0.04450.00070.11310.00370.02660.21070.0411
500−0.0007−0.2709−0.03580.00040.11620.00240.02030.20700.0334
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the univariate threshold quadratic model defined as Equation (20). The first column gives the sample size that the simulation used. The third to fifth report propose the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 5. Simulation results of nonparametric threshold estimators, DGP 5.
Table 5. Simulation results of nonparametric threshold estimators, DGP 5.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.01410.25600.07510.00600.10050.02130.07620.18710.1253
3000.00050.25870.04210.00060.09700.01040.02530.17330.0931
5000.00000.26960.03330.00000.09770.00850.00380.15830.0862
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0035−0.0232−0.01670.00500.02480.00140.07100.15590.0335
3000.0000−0.0176−0.00820.00010.02050.00030.01180.14200.0136
5000.0001−0.0330−0.00570.00000.02220.00010.00410.14520.0106
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0203−0.2778−0.11730.00850.12390.02120.09000.21610.0864
300−0.0007−0.2878−0.09580.00020.12560.01330.01540.20690.0639
5000.0000−0.2883−0.09440.00000.12530.01190.00350.20560.0544
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the multivariate linear threshold model defined as Equation (21). The first column gives the sample size that the simulation used. The third to fifth report propose the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 6. Simulation results of nonparametric threshold estimators, DGP 6.
Table 6. Simulation results of nonparametric threshold estimators, DGP 6.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.01970.24950.07040.00820.09720.01880.08820.18710.1177
3000.00020.26520.03640.00010.09970.00940.01140.17140.0898
5000.00000.27380.02970.00000.10030.00740.00320.15940.0807
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.0019−0.0107−0.01580.00510.02420.00130.07110.15530.0323
300−0.0004−0.0251−0.00740.00020.02160.00020.01380.14500.0125
5000.0001−0.0280−0.00540.00000.02100.00010.00360.14220.0094
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0184−0.2709−0.11640.00820.11770.02070.08860.21050.0846
300−0.0007−0.2717−0.09750.00040.11570.01310.01940.20480.0600
5000.0002−0.2647−0.08890.00000.10800.01040.00420.19490.0497
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the multivariate threshold quadratic model defined as Equation (22). The first column gives the sample size that the simulation used. The third to fifth columns report the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 7. Simulation results of nonparametric threshold estimators, DGP 7.
Table 7. Simulation results of nonparametric threshold estimators, DGP 7.
γ0 Is the 25th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
1000.02070.29360.12920.00970.10860.04190.09640.14980.1588
3000.00050.29150.12750.00030.10310.03930.01680.13470.1517
5000.00030.29470.13780.00010.10480.04270.01050.13410.1542
γ0 Is the 50th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.00340.0004−0.03730.00510.02650.00740.07160.16300.0778
3000.00130.0049−0.03660.00030.02290.00290.01780.15140.0398
5000.00030.0077−0.03150.00010.01800.00190.00810.13390.0294
γ0 Is the 75th Quantile of the Threshold Variable
BiasMSEStdev
nSemi-MDKEIDKESemi-MDKEIDKESemi-MDKEIDKE
100−0.0244−0.2830−0.22420.01060.11370.05750.09980.18340.0849
3000.0000−0.2798−0.20680.00010.10740.04570.00840.17080.0539
5000.0000−0.2823−0.19630.00000.10390.04030.00360.15580.0424
This table reports the simulation results of three estimators, the semiparametric M-estimator of Henderson et al. (2017), the DKE of Delgado and Hidalgo (2000) and the IDKE of Yu et al. (2018) for the multivariate threshold periodic model defined as Equation (23). The first column gives the sample size that the simulation used. The third to fifth columns report the average bias. The sixth to eighth columns give the mean squared errors of the threshold estimates. The last three columns present the standard deviations.
Table 8. Estimated convergence rate of the nonparametric threshold estimators.
Table 8. Estimated convergence rate of the nonparametric threshold estimators.
Semiparametric M-Estimator of Henderson et al. (2017)
DGP 1DGP 2DGP 3DGP 4DGP 5DGP 6DGP 7
p = 25−1.235−1.202−1.209−1.280−1.224−1.347−1.307
p = 50−1.162−1.195−1.171−1.234−1.349−1.335−1.347
p = 75−1.215−1.251−1.203−1.205−1.227−1.234−1.331
IDKE of Yu et al. (2018)
DGP 1DGP 2DGP 3DGP 4DGP 5DGP 6DGP 7
p = 25−2.207−2.126−2.164−2.541−1.556−1.436−2.557
p = 50−1.352−1.287−1.305−1.335−1.428−1.348−1.982
p = 75−1.758−1.949−1.966−1.757−1.876−2.115−2.626
This table reports the realized convergence rates of the semiparametric M-estimator of Henderson et al. (2017) and the IDKE of Yu et al. (2018). The realized convergence rates are shown as the coefficient estimate by regressing the logarithm of RMSE on the logarithm of the sample size for each DGP. Samples sizes used are n = 100, 200, 300, 400, 500, 600 and 700.

Share and Cite

MDPI and ACS Style

Chen, C.; Sun, Y. Monte Carlo Comparison for Nonparametric Threshold Estimators. J. Risk Financial Manag. 2018, 11, 49. https://doi.org/10.3390/jrfm11030049

AMA Style

Chen C, Sun Y. Monte Carlo Comparison for Nonparametric Threshold Estimators. Journal of Risk and Financial Management. 2018; 11(3):49. https://doi.org/10.3390/jrfm11030049

Chicago/Turabian Style

Chen, Chaoyi, and Yiguo Sun. 2018. "Monte Carlo Comparison for Nonparametric Threshold Estimators" Journal of Risk and Financial Management 11, no. 3: 49. https://doi.org/10.3390/jrfm11030049

APA Style

Chen, C., & Sun, Y. (2018). Monte Carlo Comparison for Nonparametric Threshold Estimators. Journal of Risk and Financial Management, 11(3), 49. https://doi.org/10.3390/jrfm11030049

Article Metrics

Back to TopTop