2.1. Large Sample Theory of
(i) Asymptotic unbiasedness: if as then E
Proof. Note that
E and let
. Thus, the characteristic function of
is as follows:
where
denotes the characteristic function of
(F). Since
,
as
and, thus,
. Hence,
as
at each continuity point of F. The Lebesgue dominated convergence theorem was applied to obtain the result. □
(ii) Weak (and) consistency: if as , then with the same probability as .
Proof. From Part (i), we only need to show that
as
. We will show a stronger result for sufficiently large values of
m and
n, as follows:
Since
as
, using an argument similar to Part (i), we can show that
as
at each continuity point
y of
F1. Thus, it follows that
as
. In fact,
. Next, for sufficiently large values of
m and
n, we can see the following by the same reasoning:
For sufficiently large values of m and n, The proof for is similar. Finally, for , the desired conclusion is reached by collecting terms. □
(iii) Strong consistency: if as , then with a probability of one.
Proof. Since, according to Part (i),
E as
, we only need to look at
. However, since
is a distribution function, by integrating the parts, we obtain the following:
We use the standard law of iterated logarithm for empirical distribution functions. □
(iv) Asymptotic normality: if , and if as , is asymptotically standard normal, where is as given in Equation (9).
Proof. Under the above conditions, it is sufficient to point out that
is asymptotically standard normal. To this end, we obtain the following:
Clearly,
B is the difference between the two independent sample averages; thus, we have proved that
B is asymptotically normal and has a mean of zero with the variance
Additionally, note that
, and if
m and
n are large enough according to the methods in Part (i), then we obtain the following:
Hence, thus . The conclusion is now obtained. □
Some remarks:
(I) A sufficient condition for is that as , provided that F is the Lipschitz of order .
Proof. Note that
, where
. Thus, by integrating the parts, the following are obtained:
□
If we know that with a known K(·) and , , and , and , and , then the above condition is met for and . Of course, if , then .
(II) If one wishes to find an asymptotic confidence interval for
a consistent estimate of
is needed. However, this estimate can be easily obtained using the following:
where
and
is the empirical df of
F, with
and
defined analogously. Thus, we obtain the confidence bounds as follows:
In addition, note that . Thus, in Part (iii) above, we can write that is asymptotically standard normal.
(III) The kernel method provides an easy way to generate , at least for the special cases and , by defining as a kernel estimate , , where K is a random df and is a sequence of constants, such that as ( as . As is now well known in the literature, the choices of and will, generally speaking, depend on the data. Thus, in such cases, the conditions on will have to be adjusted.
(IV) Mann–Whitney–Wilcoxon statistics for paired data.
is drawn from a bivariate df
F(
x,
y). In this case, we may be interested in estimating either (i)
or (ii)
We propose estimating
using
where
is a sequence of df’s converging to
as
Note that
is an average of independent random variables. Thus, one can, without much difficulty, study its properties. We have left in the details for readers who are interested. If it is necessary to estimate
we propose estimating
Thus,
where
The asymptotic properties of
are not trivial to deduce, but can be obtained by the methods described in this paper. The consistency (weak and strong) of
when we have a random sample
is obtained under the condition that
as
. The asymptotic normality is obtained with the following approximate variance:
To estimate
we propose the following:
Thus, is asymptotically normal provided that as
2.2. Robustness of against Dependence
In this section, we assume that
(
denotes the first m(n) units in the sequence
, satisfying the following strong mixing condition c.f. [
28]. Let
denote the σ-field generated by
, then
is said to be strong mixing if there is a function with the integer value
such that
as
, and the following is obtained:
For all . Throughout this section, we shall assume that are strictly stationary. To establish the results of this section, we need some definitions.
Let
and
The following are obtained:
The next lemma is instrumental in the development of this section.
Lemma 1. If and , then as .
Proof. For
and
the following is obtained:
Then, we can easily see the following:
We shall consider each term alone. Since
, the first sum of the order
as
. From now on, we drop the
m and
n suffix from
Next, by the Lemma of [
28], since
and
are strictly stationary, we can see the following:
Thus, the second sum in Equation (20) is bounded above by the following:
which converges to zero, since
. Thus, by Kronecker’s lemma,
as
In a similar way, we can show that the third term in Equation (20) converges to zero as
. Finally, the fourth term is less than or equal to
However, since
and
(note that here and elsewhere,
C denotes a generic positive constant that is not necessarily the same from place to place),
Thus, the last term in Equation (20) is less than or equal to
To show Equation (23), let
be an n integer, such that
and
as
and
. Then, the following is obtained:
Next, since , as Thus, Equation (23) is proved, as is the lemma. □
In light of Lemma 1, the consistency and asymptotic normality of
can be analyzed just by looking at
and
. In addition, note that
Now, by stationarity, the following is obtained:
Similarly, we can express
Thus, we obtain the following for sufficiently large values of
m and
n:
Let us find large sample values for the covariances. For large values of
m and
n, the following is obtained:
where
denotes the joint df of
. This is similar for the second covariance term. Hence, for
m and
n values that are sufficiently large, the following is obtained:
Now, let us write
, where
is the approximate variance of
Note that
and
. Thus, by Lemma 2.1 of [
29], the following is obtained:
where
and
are integer-valued functions, such that
.
Now we can state and prove the properties of if samples are drawn from strictly stationary strong mixing processes.
(i) Weak consistency: if and , and if as , then with the same probability as .
Proof. This directly follows the fact that and by applying the law of large numbers for Ergodic sequences, as according to Lemma 1. □
(ii) Strong consistency: Assume that the conditions of Part (i) hold and, in addition, assume that there are integer-valued functions M(m) and , such that and for any then with a probability of one as .
Proof. Again, from Lemma 1, for some (by choosing based on the proof of that lemma, , . Thus, with and by writing for , we can see that for any thus, with a probability of one as . The conclusion is obtained by applying Equation (27) to and . □
(iii) Asymptotic normality: note that we obtain the following for large values of
m,
n:
where
. Similarly, we define
provided, of course, that limits exist. In this case, we can write that
is asymptotically standard normal.