1. Introduction
Throughout this article, we use
to stand for the collection of all
matrices with real numbers;
,
, and
to stand for the transpose, the rank, and the range (column space) of a matrix
, respectively; and
to denote the identity matrix of order
m. For two symmetric matrices
and
of the same size, they are said to satisfy the inequality
in the Löwner partial ordering if
is positive semi-definite. The Kronecker product of any two matrices
and
is defined to be
. The vectorization operator of a matrix
is defined to be
. A well-known property on the vec operator of a triple matrix product is
. The Moore–Penrose inverse of
, denoted by
, is defined to be the unique solution
of the four matrix equations
,
,
, and
. We also denote by
,
, and
the three orthogonal projectors induced from
, respectively, which will help in briefly denoting calculation processes related to generalized inverses of matrices. We also adopt the notation
when
is a block matrix. Further information about the orthogonal projectors
,
, and
with their applications in the linear statistical models can be found e.g., in [
1,
2,
3].
In this paper, we consider the multivariate general linear model
where
is an observable random matrix (a longitudinal data set),
is a known model matrix of arbitrary rank (
,
is a matrix of fixed but unknown parameters,
and
denote the expectation vector and the dispersion matrix of the random error matrix
,
and
are two known positive semi-definite matrices of arbitrary ranks, and
is an arbitrary positive scaling factor. As we know, the multivariate general linear model (for short, MGLM) as such in (
1) is a relative direct extension of the most welcome type of univariate general linear models.
The assumption in (
1) is typical in the estimation and statistical inference under a multivariate linear regression framework. In statistical practice, we may meet with the situation where a true regression model is misspecified in some other forms due to certain unforeseeable reasons, and, therefore, we face with the task of comparing estimation and inference results and establishing certain links between them for the purpose of reasonably explaining and utilizing the misspecified regression model. In this light, one of the situations in relation to model misspecification problems appears by adding or deleting regressors in the model. As such an example, if taking (
1) as a true model and misspecifically adding a multiple new regressor part
in (
1), we obtain an over-parameterized (over-fitted) form of
as
where
is a known matrix of arbitrary rank, and
is a matrix of fixed but unknown parameters. Given (
1) and (
2), we proposed and studied some research problems in [
4] on the equivalence of inference results that are obtained from the two competing MGLMs.
As we know, a commonly-used technique of handling a partitioned model is to multiply a certain annihilating matrix and to transform the the model equation into a reduced model form. As a new exploration regarding the equivalence problem, we introduce the commonly-used technique into the study of (
1) and (
2). To do so, we pre-multiply
to the both sides of the model equation and noting that
to obtain a reduced model as follows:
It should be pointed out that estimation and inference results that we derive from the triple models in (
1)–(
3) are not necessarily identical. Thus, it is a primary requirement to describe the links between the models and to propose and describe possible equalities among estimation and inference results under three MGLMs.
Before approaching comparison problems of estimation and inference results under the triple models in (
1)–(
3), we mention a well-known and effective method that was widely used in the investigation of multivariate general linear models. Recall that the Kronecker products and vec operations of matrices are popular tools in dealing with matrix operations in relation to multivariate general linear models. Referring to these operations, we can alternatively represent the triple models in (
1)–(
3) in the following three standard linear statistical models:
As a common fact in statistical analysis, we know that the first step in the inference of (
1) is to estimate/predict certain functions of the unknown parameter matrices
and
. Based on this consideration, it is of great interest to identify their estimators and predictors simultaneously. For this purpose, we construct a general parametric matrix that involves both
and
as follows:
where
and
are
and
matrices, respectively. In this situation, we easily obtain that
hold. Under the assumptions in the triple models in (
1)–(
3), the corresponding predictions of
in (
7) are not necessarily identical, and we even use the same optimality criterion to derive the predictors of
under the triple competing models, and therefore, this fact leads us to propose and study a series of research problems regarding the comparison and equivalence issues about inference results obtained from the triple models. In order to obtain general results and facts under (
1)–(
8), we do not require probability distributions of the random variables in the MGLMs although they are necessary for further discussing identification and test problems.
The purpose of this paper is to consider some concrete problems on the comparisons of the best linear unbiased estimators derived from (
1) and those derived from (
2) and (
3). Historically, there were some previous investigations on establishing possible equalities of estimations of unknown parameter matrices in two competing linear models; see e.g., [
5,
6], while equalities of estimations of unknown parameter vectors under linear models with new regressors (augmentation by nuisance parameters) were approached in [
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17]. Particularly, the present two authors studied in [
4] the equivalences of estimation and inference results under (
1), (
2), (
4), and (
5). As an updated work on this subject, we introduce the two reduced models in (
3) and (
6), and carry out a new analysis of the equivalences of estimators under (
1)–(
6).
The remaining of this paper is constructed as follows: In
Section 2, we introduce some matrix analysis tools that can be used to characterize equalities that involve algebraic operations of matrices and their generalized inverses. In
Section 3, the authors present a standard procedure to describe the predictability and estimability of parametric matrices under the triple models in (
1)–(
3), and then show how to establish analytical expressions for calculating best linear unbiased predictors and best linear unbiased estimators of parametric matrices under the triple models in (
1)–(
3). In
Section 4, the authors discuss a group of problems on the equivalences of the BLUEs under (
1)–(
3).
3. The Precise Theory of Predictability, Estimability, and BLUP/BLUE
In this section, we present a standard procedure of establishing predictability, estimability, and BLUP theory under an MGLM for the purpose of solving the comparison problems proposed in
Section 1. Most of the materials given below are routine illustrations of various known conceptions, definitions, and fundamental results and facts on MGLMs; see e.g., [
4].
Definition 1. Let be as given in (7). Then, - (a)
is said to be predictable under (1) if there exists a matrix such that ; - (b)
is said to be predictable under (4) if there exists an matrix such that .
Definition 2. Let be as given in (7). Then, - (a)
Given that is predictable under (1), if there exists a matrix such thatholds in the Löwner partial ordering, the linear statistic is defined to be the best linear unbiased predictor (for short, BLUP) of under (1), and is denoted by If or in (7), the satisfying (23) is called the best linear unbiased estimator (for short, BLUE) of and the BLUP of under (1), respectively, and is denoted byrespectively. - (b)
Given that is predictable under (4), if there exists a matrix such thatholds in the Löwner partial ordering, the linear statistic is defined to be the BLUP of under (4), and is denoted by If or in (7), the satisfying (24) is called the BLUE of and the BLUP of under (4), respectively, and is denoted byrespectively.
Recall that the unbiasedness of given predictors/estimators and the lowest covariance matrices formulated in (
23) and (
24) are intrinsic requirements in statistic analysis of parametric regression models, which can be regarded as some special cases of mathematical optimization problems on constrained quadratic matrix-valued functions in the Löwner partial ordering. Note from (
1) and (
7) that
and
can be rewritten as
Hence, the expectations of
and
can be expressed as
The dispersion matrix of
can be expressed as
where
.
Concerning the predictability of
in (
7), we have the following known result.
Lemma 5 ([
4])
. Let Φ
be as given in (
7).
Then, the following three statements are equivalent:- (a)
Φ
is predictable by in (
1).
- (b)
- (c)
Theorem 1. Assume Φ
in (7) is predictable. Then,The matrix equation in (
29),
called the BLUP equation associated with is consistent as well, i.e.,holds under Lemma 5(c)
, while the general expressions of and the corresponding can be written aswhere is arbitrary. In particular,where is arbitrary. Furthermore, the following results hold. - (a)
and
- (b)
is unique if and only if
- (c)
is unique if and only if holds with probability
- (d)
The expectation, the dispersion matrices of and as well as the covariance matrix between and are unique, and are given by - (e)
and satisfy - (f)
holds for any matrix
Proof. We obtain from (
28) that the constrained minimization problem in (
23) is equivalent to
which is further reduced to
Since
is a non-null nonnegative definite matrix, we apply Lemma 4 to (
43) to yield the matrix equation
as required for (
29). Equations (
32) and (
33) follow directly from (
31). Result (a) is well known on the matrix
; see, e.g., [
1,
2].
Note that
by (
10). Combining this fact with (
31) leads to (b). Setting the term
in (
31) leads to (c).
From (
1) and (
31),
thus establishing (
35).
From (
1) and (
31),
establishing (
36). Combining (
8) and (
35) yields (
37). Substituting (
31) into (
28) and simplifying, we obtain
Rewrite the arbitrary matrix
in (
31) as
, and
in (
31) as
. Then, (
31) can equivalently be represented as
thus establishing (
39).
From (
32) and (
33), the covariance matrix between
and
is
Applying (
15) to the matrix product on the right-hand side of (
44) and simplifying, we obtain
thus, the right-hand side of (
44) is null, establishing (
40). Equation (
41) follows from (
39) and (
40). □
Concerning the BLUEs of
the mean matrix
, and the BLUP of the error matrix
in (
1), we have the following results.
Corollary 1. Let be as given in (
1).
Then, the following facts hold. - (i)
is estimable under (
1) ⇔
- (ii)
The mean matrix is always estimable under (
1).
In this case, the matrix equationis consistent, and the following results hold. - (a)
The general expression of can be written aswithwhere is arbitrary. - (b)
The general expression of can be written aswithwhere is arbitrary. - (c)
The general expression of can be written aswithwhere is arbitrary. - (d)
and satisfy
The BLUEs under the over-parameterized model (
2) can be formulated from the standard results on the BLUEs under the true model as follows.
Theorem 2. Let be as given in (
2),
and denote and Then, the following facts hold. - (a)
is estimable under (
2) ⇔
⇔
- (b)
is estimable under (
2) ⇔
In this case, the matrix equationis consistent, and a BLUE of under (
2)
iswhere is arbitrary. In particular,where is arbitrary. Proof. Note that
can be rewritten as
under (
2). Hence,
is estimable under (
2) if and only if
by Lemma 5, or equivalently,
In addition, note that
and
by (
10). Hence, (
65) is further equivalent to
, as required for (a). Let
in (
65). Then, we obtain from (
65) that
Hence, (
66) is equivalent to
as required for (b). Equations (
58)–(
64) follow from the standard results on BLUEs in Corollary 1. □
The BLUEs of unknown parameter matrices under the transformed model in (
3) can be formulated from the standard results on the BLUEs under the true model as follows.
Theorem 3. Let be as given in (
3).
Then, the following results hold: - (a)
is estimable under (
3) ⇔
- (b)
is estimable under (
3) ⇔
In this case, the matrix equationis consistent, andwhere is arbitrary.