3.2. Finite Sample Properties
The finite sample size and power properties of the jackknife-corrected test statistics were investigated using the simulation model adopted by Cavaliere, Rahbek and Taylor [
15] (denoted CRT12) for the purpose of evaluating their bootstrap procedure. The model takes
and is given by
where
is a vector of normally distributed independent random variables with covariance matrix
, the sample size
, and the initial condition is
. The short-run adjustment matrices are defined as
and
where
a,
γ and
δ are scalar parameters defined below for each of the three data generation processes (DGPs) considered:
DGP1: , , , .
DGP2: , , , .
DGP3: , , .
In DGPs 1 and 2 there is a single cointegrating vector while in DGP3 there is no cointegration and
is an I(1) VAR(2) process (or, equivalently,
is an I(0) VAR(1) process). The form of cointegration in DGPs 1 and 2 was considered in CRT12 and implies that
is I(0). The value of
δ that appears in the matrix
was used by CRT12 because it is related to some auxiliary conditions relevant for the bootstrap procedure of Swensen [
21]. These conditions are satisfied for
but not for
. CRT12 also included two additional values of
δ (equal to
and
) but, as will be seen, the value of
δ does not have a major impact on the test procedures under consideration and so we restrict attention to just two of the four values used in CRT12. Note that, in DGP3,
and the matrix
is diagonal with the scalar
γ forming the diagonal elements.
It is also necessary for the DGPs to satisfy the three parts of Assumption 1. The first requires the roots of the equation
to have modulus greater than or equal to one, where in this case
. In DGPs 1 and 2 there are three unit roots and in DGP3 there are four; the moduli of the non-unit roots are reported in
Table 3, where it can be seen that Assumption 1(i) is satisfied in all cases. Comparing DGP1 with DGP2, the effect of reducing
γ from 0.8 to 0.5 is to increase the modulus of each of the non-unit roots. In DGP3, increasing the parameter
γ reduces the non-unit roots towards unity, and in the extreme case of
,
becomes an I(2) process. It is well known that the rank test performs poorly as this extreme case is approached; see, for example, the simulation evidence in Johansen (2002). Note that, when
, there are no roots in addition to the four unit roots because, in this case,
and hence
. Assumption 1(ii) is obviously satisfied, while it can be shown that
and hence Assumption 1(iii) is satisfied provided
.
A total of seven test statistics for cointegration rank were considered. The first is the standard (unadjusted) trace statistic
defined in Equation (
5). The second uses the small sample correction proposed by Reinsel and Ahn [
12]; the resulting statistic, denoted
, is defined by
The third statistic is the Bartlett-corrected statistic defined in Equation (
9); full details concerning computation of the correction factors can be found in Johansen [
11]. The fourth method is based on the bootstrap procedures of CRT12. The bootstrap samples are obtained by estimating the VECM under the null hypothesis, checking that the roots of the estimated matrix polynomial equation
satisfy Assumption 1(i), and then generating a total of
samples recursively using an appropriate method. In the simulations reported here a wild bootstrap was employed in which, if
denotes element
i of the residual vector
, then the residuals used for the bootstrap samples were of the form
where the
are independent normal variates. The statistic
is computed in each bootstrap sample and the critical value is obtained from the distribution of
over the
boostrap samples. We denote this test by
but emphasise that the test statistic is actually
which is compared with the critical value from the finite sample bootstrap distribution rather than the critical value from the asymptotic distribution. Full details of the procedure can be found in CRT12.
Table 3.
Moduli of non-unit roots in simulations.
Table 3.
Moduli of non-unit roots in simulations.
| Moduli |
---|
DGP1 | |
0.0 | 1.1180 | 1.1180 | 1.2500 | 1.2500 | 1.2500 |
0.2 | 1.1335 | 1.1335 | 1.2500 | 1.2500 | 1.2972 |
DGP2 | |
0.0 | 1.4142 | 1.4142 | 2.0000 | 2.0000 | 2.0000 |
0.2 | 1.3639 | 1.3639 | 2.0000 | 2.0000 | 2.5599 |
DGP3 | |
0.5 | 2.0000 | 2.0000 | 2.0000 | 2.0000 |
0.8 | 1.2500 | 1.2500 | 1.2500 | 1.2500 |
0.9 | 1.1111 | 1.1111 | 1.1111 | 1.1111 |
In addition to the above statistics, three versions of the jackknife statistic are considered. The first is
defined in Equation (
10). This was computed for a range of values of
m where practicable, although we report mainly the results for
; details of how the tests perform for other values of
m are also provided. The remaining two jackknife statistics are based on small sample adjustments to either the full sample statistic
and/or the sub-sample statistics
upon which the jackknife is based. In particular the two additional jackknife statistics are defined by
in which the small sample adjusted sub-sample statistics are defined analagously to Equation (
11) by
The first of these statistics uses the small sample adjustment purely on
while the second also uses it on the sub-sample statistics.
A total of
replications were performed for each combination of parameter values for each DGP and, as in CRT12, the VAR model was fitted with a restricted intercept (case 2). The bootstrap procedure is the most computationally intensive component in the simulations, requiring a sufficiently large number (
) of bootstrap samples in each replication in order to compute the critical value from the bootstrap distribution; CRT12, for example, set
. The bootstrap computations are therefore
compared to
for the other statistics. While using a large number of bootstrap samples poses no problem in a single empirical application, it is more of a computational (and time) burden when computing a large number of bootstrap samples for each one of a large number of Monte Carlo replications. We therefore employed the approach of Davidson and MacKinnon [
22] and Giacomini, Politis and White [
23] and used only one bootstrap sample per replication,
i.e.,
. Instead of using a large number of bootstrap samples to determine the critical value from the bootstrap distribution in each replication, the critical value is obtained from the bootstrap distribution across the
R Monte Carlo replications. This `warp-speed’ method reduces substantially the number of bootstrap computations from
to
in accordance with the other non-bootstrap statistics.
The simulation results are summarised in
Table 4,
Table 5,
Table 6 and
Table 7; in all cases the tests are based on a nominal size of 5%.
Table 4 contains the empirical size of each of the seven test statistics in the case of the VAR with a single cointegrating vector. The value of
δ has a relatively small impact on the performance of the tests but the reduction in
γ from 0.8 to 0.5 has a much larger impact. It is apparent that, in all DGPs, the unadjusted Johansen statistic
has large size distortions, with the empirical size being as large as 45% in DGP1 with
and
. The small sample adjustment that results in
reduces the size closer to its nominal level in all cases, particularly in DGP 2 (where
) but less so in DGP 1 (
) where size distortions remain. The Bartlett correction has a tendency to over-compensate, leading to empirical sizes below 5% (and around 2% for
) in most cases. The bootstrap produces sizes around 5% in DGP 1 but in DGP 2 the empirical size tends to be slightly lower than the nominal size. The jackknife statistic
manages to reduce the size towards the 5% level compared to the unadjusted statistic
with empirical sizes around 6%–7% in DGP 2 and a bit higher in DGP 1. The small sample adjustment in
reduces the empirical size in all cases, compared to
, while
produces sizes close to the nominal level in DGP 2 but shows little improvement (if any) over
in DGP 1.
The power performance of the tests in the cointegrated VAR is summarised in
Table 5 in which the probability of rejecting the null hypothesis that
is reported. Beginning with the unadjusted statistic
, the high power at the smaller sample sizes in DGP 1 is a reflection of the large size distortions reported in
Table 4. All of the adjusted statistics are less powerful than
but it should be remembered that they do have better size properties. The statistic
has particularly low power for
in DGP 2.
The size properties of the tests in a non-cointegrated VAR are reported in
Table 6. As the value of
γ increases from 0 to 0.9 the unadjusted statistic
suffers from huge size distortions, rising to 92% for
when
. The size properties of the adjusted statistcs are all better than
with the bootstrap test controlling size best over this range of parameters. The Bartlett adjustment again tends to reduce empirical size to below its nominal level as
γ increases while, for the jackknife statistics,
performs best for smaller values of
γ while
produces the best performance of the three for larger values of
γ.
Table 4.
Empirical size: cointegrated VAR.
Table 4.
Empirical size: cointegrated VAR.
δ | T | | | | | | | |
---|
DGP1 |
0.0 | 50 | 44.68 | 18.80 | 2.10 | 5.49 | 14.26 | 2.53 | 14.37 |
| 100 | 23.02 | 13.36 | 3.91 | 4.64 | 10.04 | 4.83 | 9.37 |
| 200 | 13.03 | 9.87 | 4.73 | 5.22 | 7.85 | 5.03 | 7.28 |
0.2 | 50 | 45.26 | 19.07 | 2.42 | 4.98 | 14.53 | 2.30 | 14.61 |
| 100 | 22.39 | 13.28 | 4.38 | 5.24 | 9.99 | 4.67 | 9.38 |
| 200 | 12.61 | 9.73 | 5.02 | 5.35 | 8.01 | 5.46 | 7.49 |
DGP2 |
0.0 | 50 | 14.35 | 3.15 | 2.11 | 2.64 | 6.00 | 0.59 | 5.01 |
| 100 | 10.44 | 5.38 | 4.68 | 4.62 | 7.62 | 3.31 | 6.58 |
| 200 | 7.14 | 5.21 | 4.75 | 5.11 | 6.03 | 3.95 | 5.50 |
0.2 | 50 | 15.40 | 3.42 | 2.21 | 2.69 | 6.27 | 0.60 | 5.16 |
| 100 | 10.50 | 5.37 | 4.90 | 4.58 | 7.30 | 3.18 | 6.42 |
| 200 | 7.50 | 5.29 | 4.94 | 4.86 | 5.87 | 3.88 | 5.29 |
Table 5.
Empirical power: cointegrated VAR.
Table 5.
Empirical power: cointegrated VAR.
δ | T | | | | | | | |
---|
DGP1 |
0.0 | 50 | 97.57 | 85.20 | 30.93 | 51.76 | 70.05 | 27.76 | 74.00 |
| 100 | 99.99 | 99.92 | 98.78 | 99.11 | 99.59 | 98.26 | 99.62 |
| 200 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
0.2 | 50 | 97.03 | 83.09 | 30.90 | 46.41 | 65.77 | 23.63 | 69.81 |
| 100 | 99.99 | 99.93 | 98.21 | 99.02 | 99.40 | 97.45 | 99.45 |
| 200 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 | 100.00 |
DGP2 |
0.0 | 50 | 62.52 | 27.08 | 18.39 | 17.79 | 26.14 | 3.63 | 26.36 |
| 100 | 92.80 | 83.40 | 77.70 | 78.20 | 81.03 | 61.70 | 79.93 |
| 200 | 100.00 | 100.00 | 100.00 | 100.00 | 99.98 | 99.95 | 99.98 |
0.2 | 50 | 66.70 | 29.74 | 19.45 | 18.88 | 27.87 | 4.19 | 28.26 |
| 100 | 94.25 | 77.06 | 82.00 | 82.14 | 84.40 | 66.73 | 83.67 |
| 200 | 100.00 | 100.00 | 99.99 | 99.99 | 99.99 | 99.98 | 99.99 |
Table 6.
Empirical size: non-cointegrated VAR (DGP3).
Table 6.
Empirical size: non-cointegrated VAR (DGP3).
γ | T | | | | | | | |
---|
0.0 | 50 | 17.30 | 2.75 | 5.40 | 3.76 | 6.33 | 0.27 | 5.19 |
| 100 | 9.37 | 4.03 | 5.22 | 4.36 | 6.18 | 2.00 | 5.29 |
| 200 | 7.12 | 4.71 | 5.30 | 4.98 | 5.89 | 3.36 | 5.41 |
0.5 | 50 | 37.19 | 9.93 | 4.44 | 4.10 | 8.92 | 0.77 | 8.94 |
| 100 | 16.94 | 8.07 | 4.90 | 4.73 | 7.73 | 2.62 | 6.82 |
| 200 | 9.45 | 6.26 | 4.75 | 4.58 | 6.32 | 3.48 | 5.74 |
0.8 | 50 | 78.48 | 41.96 | 1.67 | 6.33 | 22.52 | 2.88 | 25.87 |
| 100 | 44.30 | 27.61 | 3.64 | 5.90 | 13.84 | 5.33 | 13.77 |
| 200 | 21.33 | 15.45 | 4.82 | 5.48 | 8.97 | 5.77 | 8.61 |
0.9 | 50 | 92.73 | 66.06 | 0.76 | 8.56 | 39.16 | 7.08 | 44.76 |
| 100 | 75.26 | 58.00 | 1.10 | 7.96 | 27.61 | 12.09 | 28.55 |
| 200 | 44.69 | 35.42 | 3.09 | 5.89 | 14.69 | 9.78 | 14.49 |
Table 7.
Empirical size of for varying m.
Table 7.
Empirical size of for varying m.
| | m |
---|
| | 2 | 4 | 5 | 8 | 10 |
---|
DGP1 |
0.0 | 100 | 10.04 | 9.95 | 10.12 | | |
| 200 | 7.85 | 7.64 | 8.00 | 7.83 | 7.92 |
0.2 | 100 | 9.99 | 9.87 | 9.94 | | |
| 200 | 8.01 | 7.70 | 8.01 | 7.95 | 8.01 |
DGP2 |
0.0 | 100 | 7.62 | 6.60 | 6.01 | | |
| 200 | 6.03 | 5.90 | 5.85 | 5.50 | 5.32 |
0.2 | 100 | 7.30 | 6.32 | 5.87 | | |
| 200 | 5.87 | 5.94 | 5.95 | 5.68 | 5.44 |
DGP3 |
0.0 | 100 | 6.18 | 5.18 | 4.56 | | |
| 200 | 5.89 | 5.64 | 5.35 | 5.17 | 4.81 |
0.5 | 100 | 7.73 | 6.56 | 5.90 | | |
| 200 | 6.32 | 5.76 | 5.52 | 5.37 | 5.05 |
0.8 | 100 | 13.84 | 16.46 | 16.74 | | |
| 200 | 8.97 | 9.55 | 9.78 | 10.83 | 10.97 |
0.9 | 100 | 27.61 | 38.04 | 40.20 | | |
| 200 | 14.69 | 19.45 | 21.55 | 25.45 | 26.62 |
The results for the jackknife tests in
Table 4,
Table 5 and
Table 6 are based on
sub-samples, but it is of interest to ascertain how the performance of the tests is affected using different values of
m. For
there is little scope to increase
m much further; with
each sub-sample has only
observations, so increasing
m soon makes sub-sample estimation infeasible. However, for larger sample sizes some experimentation is possible, and so
Table 7 reports the empirical size of
for
when
and
when
; in each case, for the largest value of
m, the sub-samples contain just
observations.
Table 7 shows that the empirical size of
is remarkably robust to the value of
m, with the exception of DGP3 when
.
To summarise the simulation results, it appears that rank test statistics based on some form of correction factor can provide size improvements over the unadjusted Johansen statistic while still maintaining good power properties, although a bootstrap approach offers the most consistent performance over the range of DGPs considered. It should be stressed, however, that the corrected statistics and the bootstrap operate in rather different ways. All of the corrected statistics—whether the correction is a simple small sample adjustment, a (parametric) Bartlett correction or a (nonparametric) jackknife correction—aim to adjust the raw statistic in such a way that the distribution of the corrected statistic matches better the asymptotic distribution, the critical values of which the corrected statistic is compared with. The bootstrap, on the other hand, uses as the test statistic the unadjusted statistic itself, but by generating bootstrap samples whose size is equal to the given finite number of observations, uses critical values from the finite sample bootstrap distribution against which to compare the statistic. The evidence obtained here suggests that the latter approach is the most robust in practice.