Next Article in Journal
Characterization of Transitivity in L-Tolerance Spaces by Convergence and Closure
Next Article in Special Issue
Is Football/Soccer Purely Stochastic, Made Out of Luck, or Maybe Predictable? How Does Bayesian Reasoning Assess Sports?
Previous Article in Journal
A Novel Color Recognition Model for Improvement on Color Differences in Products via Grey Relational Grade
Previous Article in Special Issue
Statistical Riemann and Lebesgue Integrable Sequence of Functions with Korovkin-Type Approximation Theorems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Meta-Analysis for Simultaneously Estimating Individual Means with Shrinkage, Isotonic Regression and Pretests

1
Biostatistics Center, Kurume University, Kurume 830-0011, Fukuoka, Japan
2
Department of Mathematical and Physical Sciences, Japan Women’s University, Tokyo 112-8681, Japan
3
Department of Social Information, Mejiro University, Tokyo 161-8539, Japan
*
Author to whom correspondence should be addressed.
Axioms 2021, 10(4), 267; https://doi.org/10.3390/axioms10040267
Submission received: 23 August 2021 / Revised: 4 October 2021 / Accepted: 12 October 2021 / Published: 20 October 2021

Abstract

:
Meta-analyses combine the estimators of individual means to estimate the common mean of a population. However, the common mean could be undefined or uninformative in some scenarios where individual means are “ordered” or “sparse”. Hence, assessments of individual means become relevant, rather than the common mean. In this article, we propose simultaneous estimation of individual means using the James–Stein shrinkage estimators, which improve upon individual studies’ estimators. We also propose isotonic regression estimators for ordered means, and pretest estimators for sparse means. We provide theoretical explanations and simulation results demonstrating the superiority of the proposed estimators over the individual studies’ estimators. The proposed methods are illustrated by two datasets: one comes from gastric cancer patients and the other from COVID-19 patients.

1. Introduction

Meta-analysis is a statistical method used to summarize results from published studies [1]. While it has been applied to all areas of science, meta-analysis has been a particularly important tool in educational studies [2,3] and medical studies [4,5]. Especially, meta-analysis has played an important role in studies on the impact of COVID-19 [6,7,8].
The most basic method in meta-analyses uses a fixed-effect model, where all of the studies have a common mean [5,9]. This model has long been recognized in mathematical statistics and stratified sampling designs, where the common mean estimator has rigorously been investigated (pp. 55–103 of [10]; [11,12,13]).
Another basic model is the random-effects model [14], where the studies have different means. This model imposes normally distributed random-effects to account for between-study deviations from the common mean [15]. In either model, the goal of meta-analyses is to estimate the common mean by combining the estimators of individual means.
However, in some scenarios, the common mean is undefined or uninformative. In these scenarios, the focus of meta-analyses can be on individual studies’ means. While individual means are estimated by published studies, they could be improved by statistical techniques. For instance, Raudenbush and Bryk [16] pointed out the possibility for improving individual studies’ estimators in meta-analyses by using empirical Bayes estimators under the normal random-effects. Schmid [17] also considered similar Bayes estimators that improve upon individual studies’ estimators. Röver and Friede [18,19] tried to improve an individual estimator by another individual estimator. All these Bayes estimators are defined by posterior means under the normal priors (random-effects). Evidently, the posterior means are related to but different from the individual studies’ estimates. However, there are some cases where the normal priors (random-effects) do not apply (Section 2.2). This motivates us to develop non-Bayesian estimation.
In this article, we consider some improved estimation methods of individual means without normal random-effects. First, we will explain how to employ shrinkage estimators to obtain improved estimators of individual means. Unlike the existing Bayes estimators, our frequentist shrinkage estimators do not involve any prior distribution. Next, we will provide theoretical and numerical results to show the superiority of the proposed estimators over the individual studies’ estimators. Finally, we will analyze two datasets to illustrate the proposed methods.
This article is organized as follows. Section 2 provides a background including a review of meta-analyses. Section 3 proposes the estimators of individual means and studies their theoretical properties. Section 4 conducts simulation studies, and Section 5 analyzes real datasets. Section 6 concludes the article.

2. Background

2.1. Meta-Analysis

This subsection reviews the basic models for meta-analysis under the normal distribution, including the fixed-effect model and random-effects model [1,5].
First, we introduce some mathematical notations and basic assumptions for meta-analyses. Let G > 0 be the number of studies ( G stands for Groups) in a meta-analysis. For i = 1 , 2 , , G , the i-th study provides an estimate Y i for an unknown mean μ i with a known variance σ i 2 > 0 . This implies that each study provides σ i as the standard error (SE) of the estimate Y i . Hence, in meta-analyses, the variance σ i 2 is known. We note that there could be a different setting, where the variance is unknown [20,21].
Next, we further assume the normal model Y i ~ N ( μ i , σ i 2 ) . This model’s special case, the fixed-effect model, imposes the common mean assumption. That is, μ μ 1 = = μ G . The random-effects model imposes normally distributed means given by μ i ~ N ( μ , τ 2 ) for i = 1 , 2 , , G . In either case, the main goal of meta-analyses is to estimate   μ by optimally combining Y i s [1,5,9].
In a meta-analysis, observed data consist of { Y i ; i = 1 , 2 , , G } . An example is the two-sample t-test for a continuous outcome, where Y i is the (standardized) mean difference estimator from the i-th study, and σ i 2 is the pooled variance estimator [5]. Another example is the 2 × 2 table analysis for association between an event and a risk factor, where Y i is the logarithm of the risk ratio (or odds ratio), and σ i 2 is estimated variance [5]. Another example is a survival analysis for censored data, where Y i can be the logarithm of an estimate of relative risks from the Cox model. Without loss of generality, we assume that μ j = 0 corresponds to the null value.
Finally, we describe the main idea of meta-analyses. The idea is to optimally combine the results from the G studies, based on either the fixed-effect model or random-effects model. The first model leads to the fixed-effect meta-analysis under the assumption μ μ 1 = = μ G . The second model leads to the random-effects meta-analysis under the assumption μ i ~ N ( μ , τ 2 ) . In either case, the target is to estimate the common (overall) mean μ by
μ ^ Fixed = i = 1 G ( σ i 2 ) 1 Y i i = 1 G ( σ i 2 ) 1 ,
μ ^ Rand = i = 1 G ( σ i 2 + τ ^ 2 ) 1 Y i i = 1 G ( σ i 2 + τ ^ 2 ) 1 ,
where the value of τ ^ 2 is an estimator from the data [5,14].

2.2. Does the Estimation of the Common Mean Make Sense?

Given the data of { Y i :   i = 1 , 2 , , G } , one usually conducts the fixed-effect meta-analysis and/or random-effects meta-analysis. The Q test for homogeneity can help select between the fixed-effect and random-effects model [1,5]. If the fixed-effect model is rejected (i.e., the random-effects model is selected), one might explore a systematic source of heterogeneity by meta-regression. One might also test the goodness-of-fit of the normal random-effects model [22]. In any case, statistical inference for the common mean depends on some statistical models.
However, the above statistical models do not always fit the data at hand. If the studies have a monotonic-effect, e.g., μ 1 μ G , the model is neither fixed nor random. Thus, there is no general way to estimate the common parameter unless some specific models are imposed. If a covariate is available to explain some structure among μ i s, a meta-regression may be used (Chapter 20 of [1]). However, we do not wish to assume the presence of covariates to conduct meta-regression in this article.
While the common mean estimate can be an informative summary of individual estimates, this single value rarely describes the whole results. Thus, meta-analysis typically displays individual estimates ( y 1 , , y G ) , providing an opportunity for learning the data. For instance, forest plots show the individual estimates ( y 1 , , y G ) along with their 95% confidence intervals (CIs): ( y 1 , , y G ) ± 1.96 × ( σ 1 , , σ G ) . Funnel plots show ( y 1 , , y G ) against   ( σ 1 , , σ G ) ; see [1,5,23,24] for these plots. In some cases, one is interested in the μ i that has the largest size [16]. Therefore, reporting individual estimates is an important part of meta-analysis.

2.3. Individual Estimates Could Be Improved

Despite the potential advantages of improving the individual estimators by the shrinkage methods [11,12,25,26], the development of such estimators has rarely been discussed in the literature of meta-analyses. In other words, the goal of meta-analyses has been the estimation of the common mean by disregarding the individual means. This causes a gap between meta-analyses and shrinkage estimation methods. Therefore, there is room for improving the standard individual estimators ( y 1 , , y G ) for ( μ 1 , , μ G ) in some scenarios.
We raise scenarios where the common mean is undefined or uninformative.
The first example is the case of the monotonic-effect, μ 1 μ 2 μ G , where the common mean seems to be not well-defined. In this case, an alternative is to perform meta-regression with the aid of observable covariates. We however do not assume the availability of covariates in this article.
The second example is the case of sparse normal means [27,28], where most of μ i s are equal to zero, e.g., ( μ 1 , , μ 10 ) = ( 5 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 5 ) . This is relevant to meta-analyses that fail to reject many null hypotheses. In this case, the interest may be in the estimation of the nonzero means rather than the common mean.

3. Proposed Estimators

We propose a variety of shrinkage estimators for simultaneously estimating individual means   ( μ 1 , , μ G ) , and discuss their properties. In particular, we theoretically prove that the proposed estimators have better precision than the individual studies’ estimates ( y 1 , , y G ) in terms of a mean squared error criterion under some conditions.
The problem setting is as follows: Let Y i ~ N ( μ i , σ i 2 ) be an estimator for an unknown mean μ i under a known variance σ i 2 > 0 for i = 1 , 2 , , G . As mentioned in Section 2, the variance σ i 2 is known from the i -th study. Observed data consist of { Y i ; i = 1 , 2 , , G } . We regard Y ( Y 1 , , Y G ) as a simultaneous estimator of μ ( μ 1 , , μ G ) . That is, we regard Y as an observable random vector to estimate the unknown vector μ .
Let δ ( Y ) ( δ 1 ( Y ) , , δ G ( Y ) ) be a simultaneous estimator for μ . We state that δ ( Y ) improves upon Y if
E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] < E [ i = 1 G ( Y i μ i ) 2 σ i 2 ] = G ,         ( μ 1 , , μ G ) ,
E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] E [ i = 1 G ( Y i μ i ) 2 σ i 2 ] = G ,         ( μ 1 , , μ G ) .
We then call δ ( Y ) an improved estimator of Y .
We call δ ( Y ) = Y the standard unbiased estimator because of the unbiasedness E ( Y ) = μ . It is a benchmark estimator for comparing different estimators. Recall that Y is simply the collection of the individual studies’ estimates. If “ ( μ 1 , , μ G ) ” in the criterion (2) holds only for a restricted parameter space, δ ( Y ) is only locally improved.
The criteria (1) and (2) employ the weighted mean squared errors (WMSEs). The inverse variance weights guarantee the equal contributions of all studies to the estimation error. The weighs also make it convenient to apply the classical decision theory. The special case of σ 1 = = σ G = 1 gives the total MSE (TMSE), defined as E [ i = 1 G ( δ i ( Y ) μ i ) 2 ] [11,25,29]. We use the WMSEs for theoretical convenience while the TMSE criterion is also a relevant criterion. We will use both the WMSE and TMSE in our numerical assessments.
Indeed, there are infinitely many estimators δ ( Y ) that improve upon Y , including very complex and unrealistic ones [25]. In addition, it is quite easy to find an estimator that locally improves upon Y , such as δ ( Y ) = 0 . While the problem of deriving/assessing estimators has been intensively discussed in the statistical decision theory, it has rarely been appreciated in meta-analytical settings. The goal of this article is to facilitate the applications of shrinkage estimators in the context of meta-analyses.
In the subsequent sections, we will introduce three estimators that help reduce the WMSE and TMSE by shrinking Y toward a restricted space of ( μ 1 , , μ G ) . Section 3.1 will discuss the shrinkage towards the zero vector 0 = { ( μ 1 , , μ G ) : μ 1 = = μ G = 0 } , the most traditional shrinkage scheme. Section 3.2 will consider the shrinkage toward 0 under constraints { ( μ 1 , , μ G ) : μ 1 μ G } . Section 3.3 will explore the shrinkage towards the sparse space { ( μ 1 , , μ G ) : i = 1 G I ( μ i 0 ) < G } .

3.1. Shrinkage Estimation for Means

To estimate μ = ( μ 1 , , μ G ) , we propose the James–Stein (JS) estimator of the form
δ JS ( δ 1 JS , , δ G JS ) ( 1 G 2 i = 1 G Y i 2 / σ i 2 ) Y .
This estimator is a modification of the original JS estimator [29] that was derived under the unit variances ( σ i = 1 for i ). See Appendix A.1 for the details. The JS estimator reduces variance by shrinking the vector Y toward 0 while it produces bias. The degree of shrinkage is determined by the factor ( G 2 ) / ( i = 1 G Y i 2 / σ i 2 ) that usually ranges from 0 (0% shrinkage) to 1 (100% shrinkage), and occasionally becomes greater than 1 (overshrinkage).
It can be shown in Appendix A.1 that δ JS has the following WMSE
E [ i = 1 G ( δ i JS ( Y ) μ i ) 2 σ i 2 ] = G ( G 2 ) 2 E [ 1 χ G 2 ( λ ) ] ,
where χ G 2 ( λ ) is a random variable having a noncentral χ 2 -distribution with the noncentral parameter λ = i = 1 G μ i 2 / σ i 2 and the degrees of freedom G . Thus, δ JS has a smaller WMSE than Y when G 3 . Indeed, the WMSE is minimized at μ = 0 at which the WMSE is G 1 / ( G 2 ) by the inverse moment of the central χ 2 -distribution with   λ = 0 . Thus, the JS estimator gains the greatest advantage if all the individual means are zero. This gain is appealing for meta-analyses for small individual effects (true means close to zero). Even if μ 0 , the JS estimator has a smaller WMSE than Y . The reduction of the WMSE diminishes as λ departs from zero.
One might ask where the special formula of the JS estimator comes from. The JS estimator δ JS can be derived as an empirical Bayes estimator under the prior μ i ~ N ( 0 , σ i 2 τ 2 ) :
δ JS = ( 1 ( 1 1 + τ 2 ) ^ ) Y ,
where the shrinkage factor ( 1 / ( 1 + τ 2 ) ) ^ ( G 2 ) / i = 1 G Y i 2 / σ i 2 is the estimator of 1 / ( 1 + τ 2 ) . See Appendix A.2 for the detailed derivations. Thus, if μ i ~ N ( 0 , σ i 2 τ 2 ) , the JS estimator minimizes the Bayes risk, and hence, is the optimal estimator under the prior.
A minor modification to the JS estimator can reduce the WMSE further. The modification is made in order to avoid the effect of an overshrinkage, ( G 2 ) / ( i = 1 G Y i 2 / σ i 2 ) > 1 , by which all the signs of Y are reverted. The overshrinkage phenomenon occurs with a small probability, and therefore, the modification is minor in the majority of cases. A modified estimator is the positive-part JS estimator
δ JS + ( δ 1 JS + , , δ G JS + )   ( 1 G 2 i = 1 G Y i 2 / σ i 2 ) + Y ,
where ( . ) + max ( 0 , . ) . Consequently, δ JS + has a smaller WMSE than δ JS (p. 356, Theorem 5.4 of [25].
In summary, this subsection proposes two estimators ( δ JS and δ JS + ) that improve upon the standard unbiased estimator Y .

3.2. Estimation under Ordered Means

Our next proposal is a shrinkage estimator under ordered means. We consider the case where the ordering constraints μ 1 μ G are known by the study design. Thus, the parameter μ is known to be restricted on the space { ( μ 1 , , μ G ) : μ 1 μ G } . For instance, suppose that i   ( = 1 , 2 , , G ) represents the time index ( i = 1 for the oldest study, and i = G for the newest study) at which a treatment effect μ i is estimated. Then, one may assume a trend μ 1 μ G due to the improvements of treatments over time.
For instance, the true means may be ( μ 1 , , μ 5 ) = ( 2 , 1 ,   0 ,   1 ,   2 ) . This trend could be modeled by a meta-regression with μ i = a + b i , where values a and b are unknown. In practice, one does not know any structure of the means (e.g., linear regression) except for μ 1 μ 5 . If some knowledge, such as a linear model on covariates, is true, one could use meta-regression. However, we do not adopt any model, permitting various non-linear settings such as ( 2 , 2 ,   0 ,   0 , 0 ) and ( 2 , 1 ,   4 ,   4 , 5 ) .
The use of the standard unbiased estimator Y = ( Y 1 , , Y G ) is not desirable under the ordering constraints. Due to random variations, the estimator Y = ( Y 1 , , Y G ) can be outside the parameter space, namely, Y { ( μ 1 , , μ G ) : μ 1 μ G } . Under this setting, an estimator accounting for the parameter restriction improves upon the unrestricted estimator Y , even though the former is a biased estimator [30].
The restricted maximum likelihood (RML) estimator satisfying δ 1 δ G is calculated by the pool-adjacent-violators algorithm (PAVA)
δ i RML = max s i min t i j = s t Y j t s + 1 .
This gives the RML estimator δ RML ( δ 1 RML , , δ G RML ) , which has a smaller WMSE than Y . For an example of G = 3 , one has the data of ( Y 1 , Y 2 , Y 3 ) , and the PAVA results in
δ 1 RML = min ( Y 1 , Y 1 + Y 2 2 ,   Y 1 + Y 2 + Y 3 3 ) ,  
δ 2 RML = max { min ( Y 1 + Y 2 2 ,   Y 1 + Y 2 + Y 3 3 ) ,   min (   Y 2 ,   Y 2 + Y 3 2   ) } ,
δ 3 RML = max (   Y 1 + Y 2 + Y 3 3 ,   Y 2 + Y 3 2 , Y 3   ) .
Hence, δ i RML is equal to Y i itself or an average including Y i . Of course, δ i RML = Y i   i if Y 1 Y G . For theories and applications of the PAVA, we refer to [31,32,33]. The max min formula written above can be found in Chapter 8 of [30] or [34].
It is clear that δ RML is different from order statistics Y ( 1 ) Y ( G ) that are a permutation of ( Y 1 , , Y G ) and also improve the WMSE in some cases [34]. However, the permuted estimator loses the information of individual studies’ identifications, and therefore, is not considered in this article.
Below, we further improve δ RML with the aid of the JS estimator. Let I ( . ) be the indicator function; I ( A ) = 1 or I ( A ) = 0 if A is true or false, respectively. We adjust the estimator of Chang [35] who proposed the JS-type estimator under the order restriction as follows:
δ RJS ( 1 G 2 i = 1 G Y i 2 / σ i 2 ) Y I ( Y 1 Y G ) + δ RML ( 1 I ( Y 1 Y G ) ) ,
where “RJS” stands for “Restricted JS”. Note that δ RJS has a smaller WMSE than δ RML , and hence, the former improves upon the latter [35].
We further improve the RJS estimator by the positive-part RJS estimator given by
δ RJS + ( 1 G 2 i = 1 G Y i 2 / σ i 2 ) + Y I ( Y 1 Y G ) + δ RML ( 1 I ( Y 1 Y G ) )
Consequently, δ RJS + has a smaller WMSE than δ RJS (Theorem 5.4 of Lehmann and Casella [25]). Note that, if Y 1 Y G is not satisfied, then δ RML = δ RJS = δ RJS + .
In summary, this subsection proposes three estimators ( δ RML , δ RJS , and δ RJS + ) that improve upon the standard unbiased estimator Y under μ 1 μ G .

3.3. Estimation under Sparse Means

Our third proposal is a shrinkage estimator under sparse normal means where most of the μ i s are zero [27,28]. The vector ( μ 1 , , μ G ) is called sparse if the number i = 1 G I ( μ i 0 ) is much smaller than G , e.g., ( μ 1 , , μ 10 ) = ( 5 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 5 ) . In practice, one does not know which components are zeros, and how many components are zero. Nonetheless, one could assume that many of ( μ 1 , , μ G ) are zero. However, the elements of ( Y 1 , , Y G ) are almost always nonzero, which disagree with the true values ( μ 1 , , μ G ) .
Under the sparse means, it is quite reasonable to estimate μ i as exactly zero if Y i is close to zero. Accordingly, one can use a thresholding estimator Y i I ( | Y i | > c i ) for a critical value c i > 0 . The idea was proposed by Bancroft [36] who formulated pretest estimators that incorporate a preliminary hypothesis test into estimation. Judge and Bock [37] extensively studied pretest estimators with applications to econometrics; see also more recent works [38,39,40,41,42,43]. Among all, we particularly note that Shih et al. [41] proposed the general pretest (GPT) estimator that includes empirical Bayes and Types I-II shrinkage pretest estimators for the univariate normal mean.
We modify the GPT estimator to be adopted to meta-analyses as follows:
δ i GPT = Y i I ( | Y i σ i | > z α 1 ) + q ( Y i σ i ) Y i I ( z α 2 < | Y i σ i | z α 1 ) ,             i = 1 , 2 , , G .
for 0 α 1 α 2 1 , q : ( , ) ( 0 , 1 ) , and z p is the upper p-th quantile of N ( 0 , 1 ) for 0 < p < 1 . To implement the GPT estimator, the values of α 1 and α 2 , and the probability function q ( . ) must be chosen. They cannot be chosen to minimize the WMSE and TMSE criteria since pretest estimators do not permit tractable forms of MSEs [38,41]. Fortunately, for any value of α 1 and α 2 , and a function q , one can show that δ GPT ( δ 1 GPT , , δ G GPT ) has smaller WMSE and TMSE values than Y provided μ 0 ; see Appendix A.3 for the proof.
For the above reasons, we apply statistically interpretable choices of α 1 , α 2 , and q ( . ) . The special case of α 1 = α 2 = α = 0.025 leads to the usual pretest estimator
δ i PT = Y i I ( | Y i σ i | > 1.96 ) ,
for which q ( . ) is arbitrary. Thus, we retain Y i if H 0 : μ i = 0 is rejected in favor of H 1 : μ i 0 at the 5% level. Otherwise, we discard Y i , and conclude μ i = 0 .
For the GPT estimator, we set q ( z ) = 1 / 2 (50% shrinkage) as suggested by Shih et al. [41]. To facilitate the interpretability of pretests, we chose α 1 = 0.025 (5% level) and α 2 = 0.05 (10% level). Thus, we propose the estimator
δ i GPT = Y i I ( | Y i σ i | > 1.96 ) + Y i 2 I ( 1.645 < | Y i σ i | 1.96 ) ,             i = 1 , 2 , , G .      
Thus, if H 0 : μ i = 0 is rejected at the 5% level, we set δ i GPT = Y i . If H 0 : μ i = 0 is accepted at the 5% level, but rejected at the 10% level, we set δ i GPT = Y i / 2 . If H 0 : μ i = 0 is accepted at the 10% level, we set δ i GPT = 0 . Thus, δ i GPT gives a weaker shrinkage than δ i PT does. Obviously, we obtain a relationship | δ i PT | | δ i GPT | | Y i | .
The GPT estimator introduced above is not an empirical Bayesian estimator. If α 1 = 0 , α 2 = 1 , and q ( y ) = 1 σ i 2 / max { σ i 2 , y 2 } were chosen, the resultant GPT estimator would be an empirical Bayes estimator [41]. However, we do not consider this estimator in our analysis.
In summary, this subsection proposes two estimators ( δ PT and δ GPT ) that improve upon the standard unbiased estimator Y provided μ 0 .

4. Simulation

We conducted Monte Carlo simulations to examine the performance of the proposed estimators. The purpose of the simulations is to clarify how the proposed estimators improve upon the standard unbiased estimator Y in terms of the WMSE and TMSE, defined as
WMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] ,
TMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 ] .
We also examined the bias defined as,
Bias E [ δ i ( Y ) μ i ] ,     i = 1 , 2 , , G .

4.1. Simulation Design

We generated normally distributed data Y i ~ N ( μ i , σ i 2 ) , i = 1 , 2 , , G , under the following scenarios:
Scenario (a):
Common means: μ = ( μ 1 , , μ G ) = ( 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 ) ,
Scenario (b):
Zero means: μ = ( μ 1 , , μ G ) = ( 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ) ,
Scenario (c):
Ordered means: μ = ( μ 1 , , μ G ) = ( 2.5 , 2 , 1.5 , 1 , 0.5 , 0 , 0.5 , 1 , 1.5 , 2 , 2.5 ) ,
Scenario (d):
Sparse means: μ = ( μ 1 , , μ G ) = ( 5 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 5 ) .
For the known variances, we generated σ i 2 ~ χ df = 1 2 / 4 , restricted in the interval [ 0.009 , 0.6 ] , as previously considered [15,44], leading to E [ σ i 2 ] = 0.173 . Figure 1 depicts simulated data ( Y 1 , , Y G ) and ( σ 1 , , σ G ) under Scenarios (a) and (c). Using the data, we computed the proposed estimators δ JS , δ JS + , δ RML , δ RJS , δ RJS + , δ PT , and δ GPT for estimating μ . Our interest is to see how and when these estimators improve upon the standard unbiased estimator Y . The WMSE, TMSE and Bias of all the estimators were computed by the Monte Carlo average of 10,000 replications of such data.

4.2. Simulation Results

Table 1, Table 2, Table 3 and Table 4 summarize the simulation results for comparing eight estimators: Y , δ JS , δ JS + ,   δ RML , δ RJS , δ RJS + , δ PT , and δ GPT .
The JS estimators ( δ JS and δ JS + ) improved upon Y in all the scenarios (see the WMSE and TMSE in Table 1, Table 2, Table 3 and Table 4). Scenario (b) produced the greatest reduction in the values of the WMSE and TMSE, where all the true means are equal to zero (Table 2). This is exactly what the theoretical results imply (Section 3.1). The two estimators δ JS and δ JS + showed similar performance except for Scenario (b), where δ JS + exhibited smaller values for the WMSE and TMSE than δ JS (Table 2). In summary, we observed a clear advantage of the two JS estimators ( δ JS and δ JS + ) over the standard estimator ( Y ) in all the scenarios examined.
The order-restricted estimators ( δ RML , δ RJS and δ RJS + ) were even more advantageous than δ JS and δ JS + according to the smallest values of the WMSE and TMSE in Table 1 and Table 3. The substantially reduced values of the WMSE and TMSE were brought by the strict constraints by isotonic regression. We found that δ RJS + had the smallest value of the WMSE and TMSE under Scenarios (a) and (c) (Table 1 and Table 3). However, we observed large and systematic biases of the order-restricted estimators, which are caused by isotonic regression.
The pretest estimator δ PT showed the best performance under Scenario (d) (Table 4), followed by the general pretest estimator δ GPT . However, the comparison between δ PT and δ GPT depends on the scenarios: δ PT was better in Scenarios (b) and (d) while δ GPT was better in Scenarios (a) and (c). Both δ PT and δ GPT did not perform well in Scenarios (a) and (c), where the sparsity hypothesis H 0 : μ i = 0 fails. Naturally, the sparsity of individual means is essential for the pretest estimators to perform well.
In summary, our simulations have shown that the proposed estimators have their own advantageous scenarios: δ JS + for Scenario (b); δ RJS + for Scenarios (a) and (c); δ PT or δ GPT for Scenario (d). These proposed estimators substantially reduced the values of the WMSE and TMSE over the standard estimator   Y . Furthermore, the proposed estimators δ JS , δ JS + ,   δ RML , δ RJS , and δ RJS + improved upon Y across all the scenarios. We therefore conclude that there are good reasons to apply the proposed estimators to estimate μ in order to improve the accuracy of estimation.

5. Data Analyses

We analyze two datasets to illustrate the proposed method. One dataset comes from gastric cancer patients and the other from COVID-19 pneumonia patients.

5.1. Gastric Cancer Data

We illustrate the proposed methods using a dataset from the Global Advanced/Adjuvant Stomach Tumor Research International Collaboration (GASTRIC) [45]. The data provide survival times for gastric cancer patients from 14 studies. One of the goals of this collaboration is to estimate the effects of chemotherapy on disease-free survival (DFS) for gastric cancer patients [46]. We used the data available in the R package surrosurv [47].
Let Y i be the estimator for a Cox regression coefficient for a treatment indicator (1=treatment vs. 0=control) and let σ i be the SE for i = 1 , , 14 . Table 5 summarizes Y i , σ i , and sample sizes for the 14 studies. It is reasonable to assume the asymptotic normal approximation Y i ~ N ( μ i , σ i 2 ) since the sample size of each study is large enough. Table 5 also shows the p-values for testing H 0 :   μ i = 0 through Z-values Y i / σ i .
Table 5 shows that only two studies reached 5% significance. Thus, it is reasonable to assume μ i 0   i , and therefore, the JS estimators are suitable. It is also reasonable to assume μ i = 0 for most of i s, and therefore, the PT and GPT estimators are suitable. We therefore calculated δ JS , δ JS + , δ PT , and δ GPT to estimate μ as shown in Table 6.
Table 6 shows that δ JS and δ JS + shrink the values of Y by 45%, namely, δ JS = δ JS + = ( 1 0.45 ) × Y . We also observe that δ PT and δ GPT cause 0%, 50% or 100% shrinkage of Y . The values of δ PT suggested the null effects ( μ i = 0 ) for 12 studies while the values of δ GPT suggested the null effects for 10 studies. By shrinkage, the estimators δ JS , δ JS + , δ PT , and δ GPT more accurately estimate μ = ( μ 1 , , μ 14 ) than Y do. The reason is that the data structure resembles Scenario (b), where δ JS and δ JS + performed the best (Table 2) or Scenario (d), where δ PT and δ GPT performed the best (Table 4).

5.2. COVID-19 Data

We illustrate the proposed methods using data from Pranata et al., [7] who examined the effect of hypertension on mortality for patients with COVID-19 pneumonia. One of their conclusions was that hypertension increases mortality. We obtained mortality data from Pranata et al. [7], consisting of 11 published studies (Table 7).
Let Y i , i = 1 , , 11 , be the log of the risk ratio (RR) calculated from two-by-two contingency tables examining the association (mortality vs. hypertension). Let σ i be the SE for i = 1 , , 11 . Table 7 summarizes the 11 studies. Table 7 also shows the p-values for testing the no association ( H 0 :   μ i = 0 ) by Z-values Y i / σ i assuming Y i ~ N ( μ i , σ i 2 ) .
Before applying the proposed methods, we changed the order of the 11 studies by the percentages of male patients (Table 8). The largest percentage (82%) gave the smallest RR ( Y 1 = 0.46 ) while the smallest percentage (45%) gave the largest RR ( Y 11 = 2.89 ). However, the values of Y i s are not monotonically increasing ( Y 1 < Y 2 < Y 3 < Y 4 > Y 5 < Y 6 < Y 7 > Y 8 > Y 9 < Y 10 < Y 11   in Table 8). While Pranata et al. [7] regressed Y i s on the percentages by meta-regression, we prefer not to assume a linear model. Alternatively, we considered the order-restricted estimators ( δ RML , δ RJS , and δ RJS + ) by merely assuming that the true means are ordered by the percentages of male patients.
Table 8 shows the proposed estimators δ JS , δ JS + , δ RML , δ RJS , and δ RJS + . The JS shrinkage estimators were δ JS = δ JS + = ( 1 0.05 ) × Y . They reduced only 5% of Y , and hence gave very weak shrinkage toward zero. This is because μ i s are unlikely to be zero. The order-restricted estimator δ RML successfully enforced order restrictions, δ 1 RJS δ 11 RJS , which was not achieved by Y . Since the order restrictions were not met for Y , we had δ RML =   δ RJS = δ RJS + by definition. Thus, the last three estimators can clearly incorporate the knowledge that hypertension-related mortality is monotonically influenced by gender without imposing a linear model.
In summary, it would be advantageous and informative to add the order restriction to individual studies’ estimators on the association (mortality vs. hypertension). The resultant order-restricted estimators are expected to provide more accurate estimates than individual studies’ estimators. The reason is that the data structure resembles Scenario (c), where δ RML , δ RJS , and δ RJS + performed the best (Table 3).

6. Conclusions and Future Extensions

We have shown that the proposed shrinkage estimators can be more precise than the standard unbiased estimators for estimating individual means in meta-analyses. Theoretical evidence (Section 3) and simulation studies (Section 4) identified the remarkable advantages of the proposed estimators under certain scenarios of the true individual means. However, no single estimator is the best for all the scenarios. Hence, it is important to choose a suitable estimator for each scenario. We have provided two data examples to demonstrate the scenarios under which the proposed estimators are beneficial (Section 5). However, the choice of estimators depends on the goal of research, and hence, it is difficult to set a purely statistical strategy to choose the estimator. What we have concluded in this article is that there is good potential for the proposed estimators by their careful applications to real meta-analyses.
Recall that the shrinkage effects on the general pretest estimators, ( δ 1 GPT , , δ G GPT ) , are component-wise, and hence, do not involves the JS-type shrinkage as in Section 3.1 and Section 3.2. In the presence of pretests, it is unclear how the JS-type shrinkage can be introduced. This issue is an interesting topic, yet it is beyond the scope of the article.
This article focused on the shrinkage toward zero because a meta-analysis is typically a collection of individual studies of weak effects. However, other shrinkage schemes can also be considered, such as the shrinkage toward an arbitrary mean vector, a truncated space, a polyhedral convex, or a positive halfplane [12,25,48,49,50,51,52,53,54]. Different shrinkage schemes could be considered depending on the goals of meta-analyses. Therefore, there is good potential for shrinkage estimation methods to work with meta-analyses.
This article focuses on univariate response. One may consider extensions of this article to multiple (e.g., bivariate) responses. For instance, educational research may involve meta-analyses of bivariate test scores on verbal and mathematics tests [55], and mathematics and statistics tests [44]. In our data example of gastric cancer patients (Section 5.1), bivariate correlated endpoints are of great interest [55]. Correlations between responses should be employed to improve the efficiency and reduce the bias of univariate analyses [56,57,58,59,60,61,62,63,64,65] and to predict a primary outcome by the secondary outcome [66,67,68,69]. Multivariate shrinkage estimators of multivariate normal means, such as [70,71], can be considered for this extension.
One may also consider extensions of this article to non-normal (e.g., gamma distributed) responses. For non-normal distributions, shrinkage estimators exist, such as [72] for gamma distribution. Thus, introduction of shrinkage estimators to multivariate and/or non-normal models, and even their regression models, is a promising future direction.
Finally, we note that all meta-analyses are valid under the questionable assumption of no publication bias. Frequentist and Bayesian models for checking the publication biases are often considered [73,74]. It would be interesting to see if the proposed shrinkage methods can mitigate the effect of publication biases.

Author Contributions

Conceptualization, N.T. and T.E.; methodology, N.T. and T.E.; data curation, N.T. and T.E.; writing, N.T., T.E., Y.K. and Y.-T.C.; supervision, T.E., Y.K., and Y.-T.C.; funding acquisition, Y.-T.C. All authors have read and agreed to the published version of the manuscript.

Funding

Chang YT is financially supported by JSPS KAKENHI Grant Number JP26330047 and JP18K11196.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank referees for their valuable suggestions that improved the article. The authors kindly thank Leila Zhang from MDPI for her invitation to submit our work to Axioms free of article processing charge.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Derivations of the JS Estimator

Define X ( Y 1 / σ 1 , , Y G / σ G ) and θ ( θ 1 , , θ G ) ( μ 1 / σ 1 , , μ G / σ G ) . Then, we obtain X ~ N G ( θ , I G ) , where I G is the identity matrix of dimension G. The random vector X having the unit variance is the traditional setup of the simultaneous inference (pp. 346–355 of [25]). Now, X is a maximum likelihood estimator of θ . Consider an arbitrary estimator θ ( X ) ( θ 1 ( X ) , , θ G ( X ) ) of θ under the square error loss θ ( X ) θ 2 i = 1 G | θ i ( Y ) θ i | 2 . Let the JS estimator be
θ JS ( X ) ( 1 G 2 i = 1 G X i 2 ) X .
It can be shown that
E θ θ JS ( X ) θ 2 = G ( G 2 ) 2 E θ [ 1 χ G 2 ( λ ) ] ,
where χ G 2 ( λ ) is the noncentral χ 2 -distribution with the noncentral parameter λ = θ 2 ; see p. 355 of [25] for the proofs. From this equation,
E θ θ JS ( X ) θ 2 < G = E θ | | X θ | | 2 ,
for G 3 . Thus, θ JS ( X ) improves upon X for G 3 . Rewriting the above results, we arrive at
E θ θ JS ( X ) θ 2 = E θ [ i = 1 G ( δ i JS ( Y ) μ i ) 2 σ i 2 ] < E [ i = 1 G ( Y i μ i ) 2 σ i 2 ] = G ,         ( μ 1 , , μ G ) .  

Appendix A.2. Derivations of the Empirical Bayes Estimator

Under the prior μ i ~ N ( 0 , σ i 2 τ 2 ) , the Bayes estimator is derived as
δ Bayes = E ( μ | Y ) = ( 1 1 1 + τ 2 ) Y .
It can be shown that the marginal distribution of Y i is Y i ~ N ( 0 , σ i 2 + σ i 2 τ 2 ) . Thus, i = 1 G Y i 2 / ( σ i 2 + σ i 2 τ 2 ) = 1 / ( 1 + τ 2 ) i = 1 G Y i 2 / σ i 2 ~ χ G 2 , a central χ 2 -distribution with G degrees of freedom. Taking the inverse,
1 + τ 2 i = 1 G Y i 2 / σ i 2 ~ 1 χ G 2 .
Taking the expectation for left and right sides,
E [ 1 + τ 2 i = 1 G Y i 2 / σ i 2 ] = 1 G 2 .
Thus, by the method of moment estimation,
( 1 1 + τ 2 ) ^ = G 2 i = 1 G Y i 2 / σ i 2 .
This estimator is unbiased since
  E ( 1 1 + τ 2 ) ^ E [ G 2 i = 1 G Y i 2 / σ i 2 ] = 1 1 + τ 2 E [ G 2 i = 1 G Y i 2 / ( σ i 2 + σ i 2 τ 2 ) ] = 1 1 + τ 2 .

Appendix A.3. Proof for the Pretest Estimators

The proofs for the WMSE and TMSE are similar. Thus, we focus on the latter. Then, it suffices to show
E [ ( Y i μ i ) 2 ] > E [ ( δ i GPT μ i ) 2 ]   i   for   μ i 0 .
It follows that
E [ ( Y i μ i ) 2 ] = E [ ( Y i μ i ) 2 I ( | Y i | > σ i z α 1 ) ] + E [ ( Y i μ i ) 2 I ( σ i z α 2 < | Y i | σ i z α 1 ) ] + E [ ( Y i μ i ) 2 I ( | Y i | σ i z α 2 ) ]
There exists a small value ε i > 0 such that for μ i ( ε i , ε i )
E [ ( Y i μ i ) 2 I ( | Y i | σ i z α 2 ) ] > E [ ( μ i ) 2 I ( | Y i | σ i z α 2 ) ] ,
E [ ( Y i μ i ) 2 I ( σ i z α 2 < | Y i | σ i z α 1 ) ] > E [ ( q ( Y i / σ i ) Y i μ i ) 2 I ( σ i z α 2 < | Y i | σ i z α 1 ) ]
Thus, for μ i ( ε i , ε i ) ,
E [ ( Y i μ i ) 2 ] > E [ ( Y i μ i ) 2 I ( | Y i | > σ i z α 1 ) ] + E [ ( q ( y σ i ) y μ i ) 2 I ( σ i z α 2 < | Y i | σ i z α 1 ) ] + E [ ( μ i ) 2 I ( σ i z α 2 | Y i | ) ] = E { Y i I ( | Y i σ i | > z α 1 ) + q ( Y i σ i ) Y i I ( z α 2 < | Y i σ i | z α 1 ) μ i } 2 = E [ ( δ i GPT μ i ) 2 ] .

References

  1. Borenstein, M.; Hedges, L.V.; Higgins, J.P.; Rothstein, H.R. Introduction to Meta-Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  2. Kaiser, T.; Menkhoff, L. Financial education in schools: A meta-analysis of experimental studies. Econ. Educ. Rev. 2020, 78, 101930. [Google Scholar] [CrossRef] [Green Version]
  3. Leung, Y.; Oates, J.; Chan, S.P. Voice, articulation, and prosody contribute to listener perceptions of speaker gender: A systematic review and meta-analysis. J. Speech Lang. Hear. Res. 2018, 61, 266–297. [Google Scholar] [CrossRef]
  4. DerSimonian, R.; Laird, N. Meta-analysis in clinical trials revisited. Contemp. Clin. Trials 2015, 45, 139–145. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Fleiss, J.L. Review papers: The statistical basis of meta-analysis. Stat. Methods Med. Res. 1993, 2, 121–145. [Google Scholar] [CrossRef]
  6. Batra, K.; Singh, T.P.; Sharma, M.; Batra, R.; Schvaneveldt, N. Investigating the psychological impact of COVID-19 among healthcare workers: A meta-analysis. Int. J. Environ. Res. Public Health 2020, 17, 9096. [Google Scholar] [CrossRef] [PubMed]
  7. Pranata, R.; Lim, M.A.; Huang, I.; Raharjo, S.B.; Lukito, A.A. Hypertension is associated with increased mortality and severity of disease in COVID-19 pneumonia: A systematic review, meta-analysis and meta-regression. J. Renin-Angiotensin-Aldosterone Syst. 2020, 21, 1470320320926899. [Google Scholar] [CrossRef] [PubMed]
  8. Wang, Y.; Kala, M.P.; Jafar, T.H. Factors associated with psychological distress during the coronavirus disease 2019 (COVID-19) pandemic on the predominantly general population: A systematic review and meta-analysis. PLoS ONE 2020, 15, e0244630. [Google Scholar] [CrossRef] [PubMed]
  9. Rice, K.; Higgins, J.P.; Lumley, T. A re-evaluation of fixed effect (s) meta-analysis. J. R. Stat. Soc. Ser. A 2018, 181, 205–227. [Google Scholar] [CrossRef]
  10. Lehmann, E.L. Elements of Large-Sample Theory; Springer Science & Business Media: Berlin, Germany, 2010. [Google Scholar]
  11. Shinozaki, N.; Chang, Y.T. Minimaxity of empirical Bayes estimators of the means of independent normal variables with unequal variances. Commun. Stat.-Theor. Methods 1993, 8, 2147–2169. [Google Scholar] [CrossRef]
  12. Shinozaki, N.; Chang, Y.T. Minimaxity of empirical Bayes estimators shrinking toward the grand mean when variances are unequal. Commun. Stat.-Theor. Methods 1996, 25, 183–199. [Google Scholar] [CrossRef]
  13. Singh, H.P.; Vishwakarma, G.K. A family of estimators of population mean using auxiliary information in stratified sampling. Commun. Stat. -Theor. Methods 2008, 37, 1038–1050. [Google Scholar] [CrossRef]
  14. DerSimonian, R.; Laird, N. Meta-analysis in clinical trials. Control Clin. Trials 1986, 7, 177–188. [Google Scholar] [CrossRef]
  15. Kontopantelis, E.; Reeves, D. Performance of statistical methods for meta-analysis when true study effects are non-normally distributed: A simulation study. Stat. Methods Med. Res. 2012, 21, 409–426. [Google Scholar] [CrossRef] [PubMed]
  16. Raudenbush, S.W.; Bryk, A.S. Empirical bayes meta-analysis. J. Educ. Stat. 1985, 10, 75–98. [Google Scholar] [CrossRef]
  17. Schmid, C. Using bayesian inference to perform meta-analysis. Eval. Health Prof. 2001, 24, 165–189. [Google Scholar] [CrossRef] [PubMed]
  18. Röver, C.; Friede, T. Dynamically borrowing strength from another study through shrinkage estimation. Stat. Methods Med. Res. 2020, 29, 293–308. [Google Scholar] [CrossRef] [Green Version]
  19. Röver, C.; Friede, T. Bounds for the weight of external data in shrinkage estimation. Biom. J. 2021, 63, 1131–1143. [Google Scholar] [CrossRef]
  20. Malekzadeh, A.; Kharrati-Kopaei, M. Inferences on the common mean of several normal populations under hetero-scedasticity. Comput. Stat. 2018, 33, 1367–1384. [Google Scholar] [CrossRef]
  21. Shinozaki, N. A note on estimating the common mean of k normal distributions and the stein problem. Commun. Stat. -Theory Methods 1978, 7, 1421–1432. [Google Scholar] [CrossRef]
  22. Chen, Z.; Zhang, G.; Li, J. Goodness-of-fit test for meta-analysis. Sci. Rep. 2015, 5, 16983. [Google Scholar] [CrossRef]
  23. Everitt, B. Modern Medical Statistics: A Practical Guide; Wiley: Hoboken, NJ, USA, 2003. [Google Scholar]
  24. Lin, L. Hybrid test for publication bias in meta-analysis. Stat. Methods Med. Res. 2020, 29, 2881–2899. [Google Scholar] [CrossRef] [PubMed]
  25. Lehmann, E.L.; Casella, G. Theory of Point Estimation, 2nd ed.; Springer: New York, NY, USA, 1998. [Google Scholar]
  26. Shao, J. Mathematical Statistics; Springer: New York, NY, USA, 2003. [Google Scholar]
  27. Johnstone, I.M. On minimax estimation of a sparse normal mean vector. Ann. Stat. 1994, 22, 271–289. [Google Scholar] [CrossRef]
  28. van der Pas, S.; Salomond, J.-B.; Schmidt-Hieber, J. Conditions for posterior contraction in the sparse normal means problem. Electron. J. Stat. 2016, 10, 976–1000. [Google Scholar] [CrossRef]
  29. James, W.; Stein, C. Estimation with quadratic loss. In Breakthroughs in Statistics; Springer: New York, NY, USA, 1992; Volume 1, pp. 443–460. [Google Scholar]
  30. van Eeden, C. Restricted Parameter Space Estimation Problems; Springer: New York, NY, USA, 2006. [Google Scholar]
  31. Li, W.; Li, R.; Feng, Z.; Ning, J. Semiparametric isotonic regression analysis for risk assessment under nested case-control and case-cohort designs. Stat. Methods Med. Res. 2020, 29, 2328–2343. [Google Scholar] [CrossRef] [PubMed]
  32. Robertson, T.; Wright, F.T.; Dykstra, R. Order Restricted Statistical Inference; Wiley: Chichester, UK, 1988. [Google Scholar]
  33. Silvapulle, M.J.; Sen, P.K. Constrained Statistical Inference: Inequality, Order, and Shape Restrictions; Wiley: Hoboken, NJ, USA, 2005. [Google Scholar]
  34. Tsukuma, H. Simultaneous estimation of restricted location parameters based on permutation and sign-change. Stat. Pap. 2012, 53, 915–934. [Google Scholar] [CrossRef]
  35. Chang, Y.T. Stein-type estimators for parameters restricted by linear inequalities. Keio Sci. Technol. Rep. 1981, 34, 83–95. [Google Scholar]
  36. Bancroft, T.A. On biases in estimation due to the use of preliminary tests of significance. Ann. Math. Stat. 1944, 15, 190–204. [Google Scholar] [CrossRef]
  37. Judge, G.G.; Bock, M.E. The Statistical Implications of Pre-Test and Stein-Rule Estimators in Econometrics; Elsevier: Amsterdam, The Netherland, 1978. [Google Scholar]
  38. Khan, S.; Saleh, A.K.M.E. On the comparison of the pre-test and shrinkage estimators for the univariate normal mean. Stat. Pap. 2001, 42, 451–473. [Google Scholar] [CrossRef] [Green Version]
  39. Magnus, J.R. The traditional pretest estimator. Theory Probab. Its Appl. 2000, 44, 293–308. [Google Scholar] [CrossRef]
  40. Magnus, J.R.; Wan, A.T.; Zhang, X. Weighted average least squares estimation with nonspherical disturbances and an application to the Hong Kong housing market. Comput. Stat. Data Anal. 2011, 55, 1331–1341. [Google Scholar] [CrossRef]
  41. Shih, J.-H.; Konno, Y.; Chang, Y.-T.; Emura, T. A class of general pretest estimators for the univariate normal mean. Commun. Stat. -Theory Methods 2021. [Google Scholar] [CrossRef]
  42. Shih, J.-H.; Lin, T.-Y.; Jimichi, M.; Emura, T. Robust ridge M-estimators with pretest and Stein-rule shrinkage for an intercept term. Jpn. J. Stat. Data Sci. 2021, 4, 107–150. [Google Scholar] [CrossRef]
  43. Kibria, B.G.; Saleh, A.M.E. Optimum critical value for pre-test estimator. Commun. Stat. -Simul. Comput. 2006, 35, 309–319. [Google Scholar] [CrossRef]
  44. Shih, J.-H.; Konno, Y.; Chang, Y.-T.; Emura, T. Estimation of a common mean vector in bivariate meta-analysis under the FGM copula. Statistics 2019, 53, 673–695. [Google Scholar] [CrossRef]
  45. GASTRIC (Global Advanced/Adjuvant Stomach Tumor Research International Collaboration) Group. Role of chemotherapy for advanced/recurrent gastric cancer: An individual-patient-data meta-analysis. Eur. J. Cancer 2013, 49, 1565–1577. [Google Scholar] [CrossRef]
  46. Oba, K.; Paoletti, X.; Alberts, S.; Bang, Y.J.; Benedetti, J.; Bleiberg, H.; Catalano, P.; Lordick, F.; Michiels, S.; Morita, S.; et al. Disease-free survival as a surrogate for overall survival in adjuvant trials of gastric cancer: A meta-analysis. J. Natl. Cancer Inst. 2013, 105, 1600–1607. [Google Scholar] [CrossRef] [Green Version]
  47. Rotolo, F.; Paoletti, X.; Michiels, S. Surrosurv: An R package for the evaluation of failure time surrogate endpoints in individual patient data meta-analyses of randomized clinical trials. Comput. Methods Programs Biomed. 2018, 155, 189–198. [Google Scholar] [CrossRef]
  48. Kubokawa, T. Shrinkage and modification techniques in estimation of variance and the related problems: A review. Commun. Stat. -Theory Methods 1999, 28, 613–650. [Google Scholar] [CrossRef]
  49. Fourdrinier, D.; Strawderman, W.E.; Wells, M.T. Shrinkage Estimation; Springer: New York, NY, USA, 2018. [Google Scholar]
  50. Fourdrinier, D.; Ouassou, I.; Strawderman, W.E. Estimation of a parameter vector when some components are re-stricted. J. Multivar. Anal. 2003, 86, 14–27. [Google Scholar] [CrossRef] [Green Version]
  51. Kuriki, S.; Takemura, A. Shrinkage estimation towards a closed convex set with a smooth boundary. J. Multivar. Anal. 2000, 75, 79–111. [Google Scholar] [CrossRef] [Green Version]
  52. Tripathi, Y.M.; Kumar, S. Estimating a positive normal mean. Stat. Pap. 2007, 48, 609–629. [Google Scholar] [CrossRef]
  53. Tsukuma, H. Shrinkage estimation in elliptically contoured distribution with restricted parameter space. Stat. Decis. 2009, 27, 25–35. [Google Scholar] [CrossRef]
  54. Tsukuma, H.; Kubokawa, T. Stein’s phenomenon in estimation of means restricted to a polyhedral convex cone. J. Multivar. Anal. 2008, 99, 141–164. [Google Scholar] [CrossRef] [Green Version]
  55. Gleser, L.J.; Olkin, L. Stochastically dependent effect sizes. In the Handbook of Research Synthesis; Russel Sage Foundation: New York, NY, USA, 1994. [Google Scholar]
  56. Emura, T.; Matsui, S.; Rondeau, V. Survival Analysis with Correlated Endpoints: Joint Frailty-Copula Models; Springer: Singapore, 2019. [Google Scholar]
  57. Emura, T.; Nakatochi, M.; Murotani, K.; Rondeau, V. A joint frailty-copula model between tumour progression and death for meta-analysis. Stat. Methods Med. Res. 2017, 26, 2649–2666. [Google Scholar] [CrossRef] [PubMed]
  58. Emura, T.; Sofeu, C.L.; Rondeau, V. Conditional copula models for correlated survival endpoints: Individual patient data meta-analysis of randomized controlled trials. Stat. Methods Med. Res. 2021. [Google Scholar] [CrossRef]
  59. Mavridis, D.; Salanti, G. A practical introduction to multivariate meta-analysis. Stat. Methods Med. Res. 2013, 22, 133–158. [Google Scholar] [CrossRef]
  60. Peng, M.; Xiang, L. Correlation-based joint feature screening for semi-competing risks outcomes with application to breast cancer data. Stat. Methods Med. Res. 2021, 09622802211037071. [Google Scholar]
  61. Riley, R.D. Multivariate meta-analysis: The effect of ignoring within-study correlation. J. R. Stat. Soc. Ser. A 2009, 172, 789–811. [Google Scholar] [CrossRef]
  62. Copas, J.B.; Jackson, D.; White, I.R.; Riley, R.D. The role of secondary outcomes in multivariate meta-analysis. J. R. Stat. Soc. Ser. C 2018, 67, 1177–1205. [Google Scholar] [CrossRef]
  63. Sofeu, C.L.; Emura, T.; Rondeau, V. A joint frailty-copula model for meta-analytic validation of failure time surrogate endpoints in clinical trials. Biom. J. 2021, 63, 423–446. [Google Scholar] [CrossRef]
  64. Wu, B.H.; Michimae, H.; Emura, T. Meta-analysis of individual patient data with semi-competing risks under the Weibull joint frailty–copula model. Comput. Stat. 2020, 35, 1525–1552. [Google Scholar] [CrossRef]
  65. Yamaguchi, Y.; Maruo, K. Bivariate beta-binomial model using Gaussian copula for bivariate meta-analysis of two binary outcomes with low incidence. Jpn. J. Stat. Data Sci. 2019, 2, 347–373. [Google Scholar] [CrossRef] [Green Version]
  66. Emura, T.; Nakatochi, M.; Matsui, S.; Michimae, H.; Rondeau, V. Personalized dynamic prediction of death according to tumour progression and high-dimensional genetic factors: Meta-analysis with a joint model. Stat. Methods Med. Res. 2018, 27, 2842–2858. [Google Scholar] [CrossRef] [PubMed]
  67. Kawakami, R.; Michimae, H.; Lin, Y.-H. Assessing the numerical integration of dynamic prediction formulas using the exact expressions under the joint frailty-copula model. Jpn. J. Stat. Data Sci. 2021, 1–29. [Google Scholar] [CrossRef]
  68. Noughabi, M.S.; Kayid, M. Bivariate quantile residual life: A characterization theorem and statistical properties. Stat. Pap. 2019, 60, 2001–2012. [Google Scholar] [CrossRef]
  69. Shinohara, S.; Lin, Y.-H.; Michimae, H.; Emura, T. Dynamic lifetime prediction using a Weibull-based bivariate failure time model: A meta-analysis of individual-patient data. Commun. Stat. -Simul. Comput. 2020, 1–20. [Google Scholar] [CrossRef]
  70. Bilodeau, M.; Kariya, T. Minimax estimators in the normal MANOVA model. J. Multivar. Anal. 1989, 28, 260–270. [Google Scholar] [CrossRef] [Green Version]
  71. Konno, Y. On estimation of a matrix of normal means with unknown covariance matrix. J. Multivar. Anal. 1991, 36, 44–55. [Google Scholar] [CrossRef] [Green Version]
  72. Vishwakarma, G.K.; Gupta, S. Shrinkage estimator for scale parameter of gamma distribution. Commun. Stat. -Simul. Comput. 2020, 1–8. [Google Scholar] [CrossRef]
  73. Copas, J.B.; Shi, J.Q. A sensitivity analysis for publication bias in systematic reviews. Stat. Methods Med Res. 2001, 10, 251–265. [Google Scholar] [CrossRef]
  74. Larose, D.T.; Dey, D.K. Modeling publication bias using weighted distributions in a Bayesian framework. Comput. Stat. Data Anal. 1998, 26, 279–302. [Google Scholar] [CrossRef]
Figure 1. Simulated values ( Y 1 , , Y G ) and the 95%CI ( Y 1 , , Y G ) ± 1.96 × ( σ 1 , , σ G ) , where Y i ~ N ( μ i , σ i 2 ) , and σ i 2 ~ χ df = 1 2 / 4 being restricted in the interval [ 0.009 , 0.6 ] , for   i = 1 , 2 , , G . The “FE Model” is the common mean estimator and its 95%CI under the fixed-effect (FE) model.
Figure 1. Simulated values ( Y 1 , , Y G ) and the 95%CI ( Y 1 , , Y G ) ± 1.96 × ( σ 1 , , σ G ) , where Y i ~ N ( μ i , σ i 2 ) , and σ i 2 ~ χ df = 1 2 / 4 being restricted in the interval [ 0.009 , 0.6 ] , for   i = 1 , 2 , , G . The “FE Model” is the common mean estimator and its 95%CI under the fixed-effect (FE) model.
Axioms 10 00267 g001
Table 1. Simulation results under Scenario (a): μ = ( μ 1 , , μ 10 ) = ( 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 ) .
Table 1. Simulation results under Scenario (a): μ = ( μ 1 , , μ 10 ) = ( 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 , 1 ) .
Y δ JS δ JS + δ RML δ RJS δ RJS + δ PT δ GPT
TMSE for μ 1.743 1.608 1.608 0.503 0.503 0.503 3.579 3.137
WMSE for μ 10.045 9.635 9.635 5.761 5.761 5.761 17.491 15.586
Bias for μ 1 0.01 −0.04 −0.04 −0.21 −0.21 −0.21 −0.15 −0.12
Bias for μ 2 0.00 −0.05 −0.05 −0.12 −0.12 −0.12 −0.15 −0.12
Bias for μ 3 0.00 −0.05 −0.05 −0.07 −0.07 −0.07 −0.16 −0.13
Bias for μ 4 0.00 −0.05 −0.05 −0.04 −0.04 −0.04 −0.16 −0.13
Bias for μ 5 0.00 −0.05 −0.05 −0.01 −0.01 −0.01 −0.16 −0.13
Bias for μ 6 0.00 −0.05 −0.05 0.01 0.01 0.01 −0.16 −0.13
Bias for μ 7 0.00 −0.05 −0.05 0.04 0.04 0.04 −0.16 −0.13
Bias for μ 8 0.00 −0.05 −0.05 0.07 0.07 0.07 −0.16 −0.13
Bias for μ 9 0.00 −0.05 −0.05 0.12 0.12 0.12 −0.15 −0.13
Bias for μ 10 0.00 −0.05 −0.05 0.21 0.21 0.21 −0.16 −0.13
Notes:   TMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 ] , WMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] ,   Bias E [ δ i ( Y ) μ i ] .
Table 2. Simulation results under Scenario (b): μ = ( μ 1 , , μ 10 ) = ( 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ).
Table 2. Simulation results under Scenario (b): μ = ( μ 1 , , μ 10 ) = ( 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ).
Y δ JS δ JS + δ RML δ RJS δ RJS + δ PT δ GPT
TMSE for μ 1.7430.3530.2230.5030.5030.5030.4890.558
WMSE for μ 10.0452.0511.2875.7615.7615.7612.8373.236
Bias for μ 1 0.010.000.00−0.21−0.21−0.210.000.00
Bias for μ 2 0.000.000.00−0.12−0.12−0.120.000.00
Bias for μ 3 0.000.000.00−0.07−0.07−0.070.000.00
Bias for μ 4 0.000.000.00−0.04−0.04−0.040.000.00
Bias for μ 5 0.000.000.00−0.01−0.01−0.010.000.00
Bias for μ 6 0.000.000.000.010.010.010.000.00
Bias for μ 7 0.000.000.000.040.040.040.000.00
Bias for μ 8 0.000.000.000.070.070.070.000.00
Bias for μ 9 0.000.000.000.120.120.120.000.00
Bias for μ 10 0.000.000.000.210.210.210.000.00
Notes:   TMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 ] ,   WMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] ,   Bias E [ δ i ( Y ) μ i ] .
Table 3. Simulation results under Scenario (c): μ = ( μ 1 , , μ 11 ) = ( 2.5 , 2 , 1.5 , 1 , 0.5 , 0 , 0.5 , 1 , 1.5 , 2 , 2.5 ) .
Table 3. Simulation results under Scenario (c): μ = ( μ 1 , , μ 11 ) = ( 2.5 , 2 , 1.5 , 1 , 0.5 , 0 , 0.5 , 1 , 1.5 , 2 , 2.5 ) .
Y δ JS δ JS + δ RML δ RJS δ RJS + δ PT δ GPT
TMSE for μ 1.9141.8471.8471.3141.3091.3092.8262.547
WMSE for μ 11.04110.83110.8319.2599.2409.24014.60013.494
Bias for μ 1 0.010.070.07−0.03−0.02−0.020.020.02
Bias for μ 2 0.000.050.050.000.010.010.030.03
Bias for μ 3 0.000.030.030.000.000.000.080.06
Bias for μ 4 0.000.020.020.000.000.000.160.13
Bias for μ 5 0.000.010.010.000.000.000.200.17
Bias for μ 6 0.000.000.000.000.000.000.000.00
Bias for μ 7 0.00−0.01−0.010.000.000.00−0.20−0.17
Bias for μ 8 0.00−0.02−0.020.000.000.00−0.16−0.13
Bias for μ 9 0.00−0.04−0.040.000.000.00−0.08−0.06
Bias for μ 10 0.00−0.05−0.050.00−0.01−0.01−0.03−0.02
Bias for μ 11 0.00−0.06−0.060.030.030.03−0.01−0.01
Notes:   TMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 ] ,   WMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] ,   Bias E [ δ i ( Y ) μ i ] .
Table 4. Simulation results under Scenario (d): μ = ( μ 1 , , μ 10 ) = ( 5 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 5 ) .
Table 4. Simulation results under Scenario (d): μ = ( μ 1 , , μ 10 ) = ( 5 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 5 ) .
Y δ JS δ JS + δ RML δ RJS δ RJS + δ PT δ GPT
TMSE for μ 1.7431.7191.7190.8230.8230.8230.7390.794
WMSE for μ 10.10459.8979.8977.2727.2727.2724.2814.599
Bias for μ 1 0.010.100.100.010.010.010.010.01
Bias for μ 2 0.000.000.00−0.21−0.21−0.210.000.00
Bias for μ 3 0.000.000.00−0.11−0.11−0.110.000.00
Bias for μ 4 0.000.000.00−0.06−0.06−0.060.000.00
Bias for μ 5 0.000.000.00−0.02−0.02−0.020.000.00
Bias for μ 6 0.000.000.00−0.02−0.02−0.020.000.00
Bias for μ 7 0.000.000.000.020.020.020.000.00
Bias for μ 8 0.000.000.000.050.050.050.000.00
Bias for μ 9 0.000.000.000.100.100.100.000.00
Bias for μ 10 0.000.000.000.210.210.210.000.00
Bias for μ 11 0.00−0.09−0.090.000.000.000.000.00
Notes:   TMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 ] ,   WMSE E [ i = 1 G ( δ i ( Y ) μ i ) 2 σ i 2 ] ,   Bias E [ δ i ( Y ) μ i ] .
Table 5. The 14 studies on gastric cancer patients for estimating the effects of chemotherapy on DFS.
Table 5. The 14 studies on gastric cancer patients for estimating the effects of chemotherapy on DFS.
Sample Size Treatment   Effect = Y i SE = σ i p-Value
Study 1269−0.183120.153720.234
Study 2190−0.722660.286860.012 **
Study 3252−0.485070.331920.144
Study 4536−0.239610.215580.266
Study 5219−0.132260.146910.368
Study 6306−0.272280.144160.059 *
Study 788−0.58670.248850.018 **
Study 8281−0.139690.145420.337
Study 9271−0.10040.164040.541
Study 10178−0.311430.170380.068 *
Study 11126−0.049490.198180.803
Study 12180−0.116850.164760.478
Study 13206−0.130440.192680.498
Study 141860.043910.176320.803
* 10% significance; ** 5% significance.
Table 6. Estimates of the treatment effects on DFS from the 14 studies on gastric cancer patients.
Table 6. Estimates of the treatment effects on DFS from the 14 studies on gastric cancer patients.
Y δ JS δ JS + δ PT δ GPT
μ 1 −0.18−0.10−0.1000
μ 2 −0.72−0.40−0.40−0.72−0.72
μ 3 −0.49−0.27−0.2700
μ 4 −0.24−0.13−0.1300
μ 5 −0.13−0.07−0.0700
μ 6 −0.27−0.15−0.150−0.14
μ 7 −0.59−0.32−0.32−0.59−0.59
μ 8 −0.14−0.08−0.0800
μ 9 −0.10−0.06−0.0600
μ 10 −0.31−0.17−0.170−0.16
μ 11 −0.05−0.03−0.0300
μ 12 −0.12−0.06−0.0600
μ 13 −0.13−0.07−0.0700
μ 14 0.040.020.0200
All values are rounded up to 2 digits.
Table 7. Summary of the 11 studies (shown by the author and publication year) on COVID-19 patients for examining the effects of hypertension on mortality.
Table 7. Summary of the 11 studies (shown by the author and publication year) on COVID-19 patients for examining the effects of hypertension on mortality.
StudySample SizeMale (%)Log (RR)SEp-Value
Akbari 202044056.40.68810.67320.307
Bai 200012763.00.59330.27540.031 **
Cao 202010252.01.17560.28210.000 **
Chen 202012349.00.53650.24930.031 **
Chen T 202027462.00.67800.17130.000 **
Fu 202020049.50.58780.33020.075 *
Grasselli 2020159182.00.46370.09560.000 **
Li 202010258.00.52470.32720.109
Luo 202040347.91.23260.14890.000 **
Yuan 20202745.02.89041.42630.043 **
Zhou 202019162.01.13780.20970.000 **
* 10% significance; ** 5% significance.
Table 8. Estimated effects of hypertension on mortality from the 11 studies on COVID-19 patients.
Table 8. Estimated effects of hypertension on mortality from the 11 studies on COVID-19 patients.
StudyMale (%)Mean Y δ JS δ JS + δ RML δ RJS δ RJS +
Grasselli 202082.0 μ 1 0.460.440.440.460.460.46
Bai 200063.0 μ 2 0.590.560.560.590.590.59
Chen T 202062.0 * μ 3 0.680.640.640.680.680.68
Zhou 202062.0 * μ 4 1.141.081.080.780.780.78
Li 202058.0 μ 5 0.520.500.500.780.780.78
Akbari 202056.4 μ 6 0.690.650.650.780.780.78
Cao 202052.0 μ 7 1.181.121.120.780.780.78
Fu 202049.5 μ 8 0.590.560.560.780.780.78
Chen 202049.0 μ 9 0.540.510.510.780.780.78
Luo 202047.9 μ 10 1.231.171.171.231.231.23
Yuan 202045.0 μ 11 2.892.742.742.892.892.89
All values are rounded up to 2 digits. * ties are sorted by the age (62 years-old in Chen T 2020; 56 years-old in Zhou 2020).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Taketomi, N.; Konno, Y.; Chang, Y.-T.; Emura, T. A Meta-Analysis for Simultaneously Estimating Individual Means with Shrinkage, Isotonic Regression and Pretests. Axioms 2021, 10, 267. https://doi.org/10.3390/axioms10040267

AMA Style

Taketomi N, Konno Y, Chang Y-T, Emura T. A Meta-Analysis for Simultaneously Estimating Individual Means with Shrinkage, Isotonic Regression and Pretests. Axioms. 2021; 10(4):267. https://doi.org/10.3390/axioms10040267

Chicago/Turabian Style

Taketomi, Nanami, Yoshihiko Konno, Yuan-Tsung Chang, and Takeshi Emura. 2021. "A Meta-Analysis for Simultaneously Estimating Individual Means with Shrinkage, Isotonic Regression and Pretests" Axioms 10, no. 4: 267. https://doi.org/10.3390/axioms10040267

APA Style

Taketomi, N., Konno, Y., Chang, Y. -T., & Emura, T. (2021). A Meta-Analysis for Simultaneously Estimating Individual Means with Shrinkage, Isotonic Regression and Pretests. Axioms, 10(4), 267. https://doi.org/10.3390/axioms10040267

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop