1. Introduction
Let
be a probability space and
a bivariate stochastic process defined on
. We consider the differenced process
, where Δ is the first-difference operator. Following Caporale and Pittis [
1] and Hassapis
et al. [
2], we say that the process
is jointly unpredictable if
where
is the
σ-field generated by past vectors
.
The goal of this paper is to show that the notion of joint unpredictability, in a particular parametric framework, can be characterized by a geometric condition. This characterization is given in terms of distance between information sets in an Hilbert space. In particular, we will show that the process is jointly unpredictable if and only if the information contained in its past is `much distant’ from the information contained in its future. Even if our result is not as general as might seem desirable, we think that the intuition gained from this characterization makes the notion of joint unpredictability more clear.
The rest of the paper is organized as follows.
Section 2 presents the utilized mathematical framework.
Section 3 presents the geometric characterization.
Section 4 concludes.
2. Preliminaries
Definitions, notation, and preliminary results from Hilbert space theory will be presented prior to establish the main result. An excellent overviews of the applications of Hilbert space methods to time series analysis can be found in Brockwell and Davis [
3].
We use the following notations and symbols. Let be a probability space. We consider the Hilbert space of all real square integrable random variables on The inner product in is defined by for any The space is a normed space and the norm is given by The distance between z, is ,. A sequence {} is said to converge to a limit point if as . A point z is a limit point of a set M (subset of ) if it is a limit point of a sequence from M. In particular, M is said to be closed if it contains all its limit points. If S is a arbitrary subset of , then the set of all (; arbitrary real numbers; arbitrary elements of S) is called a linear manifold spanned by S and is symbolized by (S). If we add to (S) all its limit points we obtain a closed set that we call the closed linear manifold or subspace spanned by S, symbolized by (S). Two elements ∈ are called orthogonal, and we write , if If S is any subset of , then we write if for all similarly, the notation , for two subsets S and T of , indicates that all elements of S are orthogonal to all elements of T. For a given and a closed subspace M of , we define the orthogonal projection of z on M, denoted by , as the unique element of such that for any We remember that if , then
If
M and
N are two arbitrary subsets of
, then the quantity
is called distance between
M and
N.
We close this section introducing some further definitions, concerning discrete stochastic processes in .
Let be a univariate stochastic process. We say that is integrated of order one (denoted ) if the process is stationary whereas is not stationary. We say that the bivariate stochastic process is integrated of order one if and .
A stochastic process
Granger causes another stochastic process
, with respect to a given information set
that contains at least
,
,
, if
can be better predicted by using past values of
y than by not doing so, all other information in
(including the past of
x) being used in either case. More formally, we say that
is Granger causal for
with respect to
if
where
.
Two stochastic processes, , and , both of which are individually are said to be cointegrated if there exists a non-zero constant β such that is a stationary () process.
It is important to note that cointegration between two variables implies the existence of causality (in the Granger sense) between them in at least one direction (see Granger [
4]).
3. A Geometric Characterization
In this section we assume that
be a bivariate stochastic process defined on
integrated of order one, with
that has a VAR(1) representation
where
is a fixed
coefficient matrix and
is i.i.d. with
and
for all
t and
for
.
In this framework we have that does not Granger cause if and only if Similarly, does not Granger cause if and only if .
We observe that the VAR residuals are usually correlated and hence the covariance matrix Σ is seldom a diagonal matrix. However, because the main aim of this study is pedagogical, we assume that Σ is diagonal for analytical convenience.
We consider the following information sets: , , and .
Theorem 3.1. Let be a VAR(1) process defined as in (2). The differenced process is jointly unpredictable if and only if Theorem 1 provides a geometric characterization of the notion of joint unpredictability of a bivariate process in term of distance between information sets. It is important to note that
Thus we have that the process
is jointly unpredictable if and only if the distances
and
achieve their maximum value, respectively.
It is intuitive to think that if these distances achieve their maximum value, then does not contain any valuable information about the future of the differenced series, and hence these are jointly unpredictable with respect to the information set , that is E.
We recall that Theorem 1 holds only in a bivariate setting.
3.1. Lemmas
In order to prove Theorem 1, we need the following lemmas.
Lemma 3.2. Let V be a closed subspace of and a subset of such that , . if and only if
Proof.
Focker and Triacca ([
5], p. 767).
Lemma 1 establishes a relationship between the orthogonality of sets/spaces in the Hilbert space and their distance. We note that the orthogonality between G and V holds if and only if the distance achieves the maximum value. In fact, can not be greater than η since V.
Lemma 3.3. The processes and are not cointegrated if and only if
These equations must be balanced, that is the order of integration of and must be zero.
(⇒) If since and , we can have three cases.
Case (1) with and
Case (2)
with
and
.
Case (3)
with
and
.
In all three cases, there exists at least a not trivial linear combination of the processes and that is stationary. Thus we can conclude that and are cointegrated.
(⇐) If then and so does not Granger cause and does not Granger cause . It follows that and are not cointegrated.
Lemma 3.4. If and are cointegrated, then
Proof.
We subtract
from both sides of Equation (2) by obtaining
If
and
are cointegrated, we have
where
is the cointegration coefficient and
and
are the speed of adjustment coefficients.
By rearranging Equation (3) we obtain an AR(1) model for
where
Since
and
are cointegrated,
is a stationary process and so
Lemma 3.5. The process is jointly unpredictable if and only if Proof.
(⇒) process
is jointly unpredictable, then
On the other hand, since
with
and
for all
t and
for
we have that
Hence we have
and so
(⇐) If
then
with
and
for all
t and
for
, and hence we have
Thus we can conclude that the process
is jointly unpredictable.
Before to conclude this subsection we observe that Equation (2) can be written in lag operator notation. The lag operator
L is defined such that
. We have that
or
3.2. Proof of Theorem 1
Sufficiency. If
then, by Lemma 1, we have
and
Now we assume that
and
are not both equal to zero. We can have three cases.
Case (1)
and
This implies that
and
Thus
Now, we note that
Thus
but this is absurd since
Case (2)
and
In this case we have
Again this is absurd since
Case (3)
and
We note that
where
By Lemma 2, we have that
and
are cointegrated and hence the matrix
has rank 1. It follows that
Thus
where
Since
and
are cointegrated, by Lemma 3 we have that
and hence
Now, we can have two cases.
Case (a)
In this case we have
and
Thus
and
but this is absurd since
and
Case (b)
In this case we have
and
Thus
and
Now, we consider the system
The determinant of the matrix
is
Since
and
we have that
Thus
implies that
or
but this is absurd since
and
In all Cases (1–3) we obtain an absurd conclusion, thus we can state that
Now, we prove that
. We have that
Since the error term
is stationary these equations must be balanced, that is the order of integration of
and
must be the same. By the hypothesis that
it follows that
(
i.e., stationary) and
is
I(1), hence
implies that
Thus
and hence, by Lemma 4, it follows that the process
is jointly unpredictable.
Necessity. If the process
is is jointly unpredictable, then by Lemma 4 it follows that
and hence
and
. This implies that
and
. Therefore we have that
and
. Thus, by Lemma 1, it follows that
Theorem 1 is proved.
4. Conclusions
In this paper we have considered the following geometric condition concerning the distance between information sets
It says that the distances
and
achieve their maximum value, respectively. Theorem 1 tells us that, under the hypothesis that the process
follows a bivariate VAR(1) model, the condition Equation (4) represents a geometric characterization of the notion of joint unpredictability. If this condition holds, the processes
and
are jointly unpredictable since the past of the bivariate process
does not contain any valuable information about the future of the differenced series. The information in the past is too far from the future information.
Even if the bivariate VAR(1) assumption is far from general, we think that this geometric characterization is useful in order to throw light on the concept of joint unpredictability of a stochastic process.