Next Article in Journal
LSTM Gate Disclosure as an Embedded AI Methodology for Wearable Fall-Detection Sensors
Previous Article in Journal
Security with Wireless Sensor Networks in Smart Grids: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Symmetry, Asymmetry and Studentized Statistics

by
Maria de Fátima Brilhante
1,2,
Dinis Pestana
2,3,4,* and
Maria Luísa Rocha
2,5,6
1
Faculdade de Ciências e Tecnologia, Universidade dos Açores, Rua da Mãe de Deus, 9500-321 Ponta Delgada, Portugal
2
Centro de Estatística e Aplicações, Universidade de Lisboa (CEAUL), Campo Grande, 1749-016 Lisboa, Portugal
3
Departamento de Estatística e Investigação Operacional, Faculdade de Ciências, Universidade de Lisboa, Campo Grande, 1749-016 Lisboa, Portugal
4
Instituto de Investigação Científica Bento da Rocha Cabral, Calçada Bento da Rocha Cabral 14, 1250-012 Lisboa, Portugal
5
Faculdade de Economia e Gestão, Universidade dos Açores, Rua da Mãe de Deus, 9500-321 Ponta Delgada, Portugal
6
Centro de Estudos de Economia Aplicada do Atlântico (CEEAplA), Universidade dos Açores, Rua da Mãe de Deus, 9500-321 Ponta Delgada, Portugal
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(10), 1297; https://doi.org/10.3390/sym16101297
Submission received: 12 September 2024 / Revised: 25 September 2024 / Accepted: 29 September 2024 / Published: 2 October 2024
(This article belongs to the Section Mathematics)

Abstract

:
Inferences on the location parameter λ in location-scale families can be carried out using Studentized statistics, i.e., considering estimators λ ˜ of λ and δ ˜ of the nuisance scale parameter δ , in a statistic T = g ( λ ˜ , δ ˜ ) with a sampling distribution that does not depend on ( λ , δ ) . If both estimators are independent, then T is an externally Studentized statistic; otherwise, it is an internally Studentized statistic. For the Gaussian and for the exponential location-scale families, there are externally Studentized statistics with sampling distributions that are easy to obtain: in the Gaussian case, Student’s classic t statistic, since the sample mean λ ˜ = X ¯ and the sample standard deviation δ ˜ = S are independent; in the exponential case, the sample minimum λ ˜ = X 1 : n and the sample range δ ˜ = X n : n X 1 : n , where the latter is a dispersion estimator, which are independent due to the independence of spacings. However, obtaining the exact distribution of Student’s statistic in non-Gaussian populations is hard, but the consequences of assuming symmetry for the parent distribution to obtain approximations allow us to determine if Student’s statistic is conservative or liberal. Moreover, examples of external and internal Studentizations in the asymmetric exponential population are given, and an ANalysis Of Spacings (ANOSp) similar to an ANOVA in Gaussian populations is also presented.

1. Introduction

Let X   = X 1 , , X n be a random sample from a population X Gaussian ( μ , σ ) . If the location parameter μ and the scale parameter σ are unknown, the unbiased estimators μ ˜ = X ¯ = 1 n k = 1 n X k Gaussian μ , σ n and σ ˜ 2 = S 2 = 1 n 1 k = 1 n ( X k X ¯ ) 2 , with ( n 1 ) S 2 σ 2 χ n 1 2 , can be used to estimate these parameters.
The distribution of the estimator X ¯ depends on the nuisance parameter σ , which is a problem for making inferences on the location parameter μ . Student’s [1] pathbreaking paper has shown that the statistic
t n 1 = n X ¯ μ S
has a distribution that does not depend on the nuisance scale parameter, since its probability density function is
f t n 1 ( t ) = 1 B 1 2 , n 1 2 1 n 1 1 + t 2 n 1 n / 2 I R ( t ) ,
where B ( p , q ) = 0 1 x p 1 ( 1 x ) q 1 d x , p , q > 0 , is Euler’s Beta function. The probability density function defined in (2) is for a random variable with Student’s distribution with n 1 degrees of freedom. Basically, the “Studentization” defined in (1), as opposed to the common standardization X ¯ μ σ / n , uses the estimators of the parameter of interest (location) and of the nuisance parameter, so that the sampling distribution of t n 1 = g ( X ¯ , S ) does not depend on the location parameter μ or on the nuisance parameter σ , thus being a pivot statistic that can be used for inferences on μ .
However, Student’s exceptional result depends heavily on the fact that in the location-scale Gaussian family, X ¯ and S 2 are independent random variables, which is an exclusive property of the Gaussian family. This characterization of the Gaussian distribution appeared in Geary’s work [2], and has been proved, independently, by Darmois [3] and by Skitovich [4]. Additionally, in the context of the Koopman–Darmois–Pitman k-parameter exponential family [5,6,7], the Gaussian family is the only one with support R and with a pair of sufficient statistics for ( μ , σ ) , namely k = 1 n X k , k = 1 n X k 2 , which captures all available information in the sample to estimate these two parameters.
When working with samples from non-Gaussian location-scale families, the dependence structure between the sample mean X ¯ n = 1 n k = 1 n X k and the sum of squares S S n = k = 1 n ( X k X ¯ n ) 2 = ( n 1 ) S n 2 is hard to discern, except for the special case n = 2 (note that an index n is now being used to emphasize the size of the random sample). For n > 2 , the computation of the probability density function of Student’s statistic
T n 1 = n ( n 1 ) X ¯ n λ S S n
is a more difficult problem to handle. In Section 2, the joint probability density function of ( X ¯ n , S S n ) is investigated in a general context, i.e., not restricted to the Gaussian family. Its explicit expression for n = 2 is given in Section 2.1, and examples for the probability density function of T 1 , i.e., for n = 2 in (3), for several symmetric parent distributions are given in Section 2.2. For n 3 , a recurrence formula for the joint probability density function of ( X ¯ n , S S n ) is obtained in Section 2.3, and a detailed study of the recurrence formula for the Gaussian case is seen in Section 2.4.
An interesting alternative for Studentization was proposed in Logan et al.’s work [8], where a self-normalized statistic W n 1 = k = 1 n X k k = 1 n | X k | α 1 / α was considered, especially when X has heavier tails than the Gaussian law, and it is in the domain of attraction of a stable law for sums with index α ( 0 , 2 ) . Peña et al.’s [9] monograph on self-normalized processes is a thorough overview, with Chapter 15 dealing with the classical t-statistic and Studentized statistics.
Another interesting fact is that Efron [10] showed that T n 1 = d t n 1 when a rotational symmetry of the unit vector U = X X = X 1 X , , X n X over the surface of the unit sphere in the Euclidean n-space E n is assumed. However, Efron [10] pointed out the following:
Unfortunately the usual sampling procedures almost never yield rotational symmetry for the normalized vector U except in the case X Gaussian ( 0 , σ 2 ) .
Efron also stated that “A very special “lucky” case is given in Section 9, namely Inverted normal error for n = 2 . It is possible to construct examples where T n is t-distributed with n degrees of freedom, without U having rotational symmetry”. Efron’s [10] pioneering work also investigated the consequences, in what regards Student’s statistic, by assuming weaker orthant symmetry, i.e., when the random vector U = U 1 , , U n has the same distribution as U δ = δ 1 U 1 , , δ n U n for every choice of δ i = ± 1 , i = 1 , 2 , , n .
This led us to observe, in Section 2.4, that the joint probability density function of ( X ¯ n , S S n ) , in the exceptional rotational symmetry of the Gaussian case, is proportional to the product of n Gaussian ( 0 , σ ) probability density functions, computed at some special points that form a symmetric arithmetic progression with a null sum and a sum of squares equal to one.
In Section 3, we shall investigate the simplifications that result from an additional symmetry assumption, namely when the joint probability density function of ( X ¯ n , S S n ) is proportional to the product of n probability density functions of X, as in the Gaussian case. The approximation obtained from the smoothness hypothesis 1 (Section 3.2) is further investigated in Section 3.3 for the worst case possible, i.e., uniform distribution (see Hendriks et al. [11]), by comparing the approximate expression for T 2 with the exact distribution given by Perlo [12] for a Uniform ( 1 , 1 ) parent.
Aside from the Gaussian family, the exponential location-scale family is also a remarkable member of the Koopman–Darmois–Pitman exponential family. It is the only member of this family with support in a half-line [ λ , ) and with a pair of sufficient statistics to estimate ( λ , δ ) , namely ( X 1 : n , X ¯ n X 1 : n ) , where X k : n , k = 1 , , n , denotes the k-th order statistic of the random sample ( X 1 , , X n ) .
Moreover, the spacings X k : n X k 1 : n Exponential 0 , 1 n + 1 k , k = 1 , , n , are independent, with the convention X 0 : n = λ . This property characterizes the exponential distribution, and it is a consequence of Pexider’s [13] functional equation f ( x + y ) = ϕ 1 ( x ) ϕ 2 ( y ) , which is an extension of Cauchy’s functional equation, with the solution f ( x ) = k 1 e k 2 x .
Using the independence of the spacings of the exponential model, we shall also obtain in Section 4 the probability density function of some externally and internally Studentized statistics, according to David’s [14] definition, for inferences on the exponential location parameter. The independence of spacings is further used to compare the location parameters of two exponential populations in Section 4.4, and also to establish an ANalysis Of Spacings (ANOSp), similar to a one-way ANOVA in Gaussian populations, which is presented in Section 4.5. A Satterthwaite [15] approximation for the case of unequal dispersions is given in Section 4.6.
Finally, Section 5 summarizes the main findings regarding approximate solutions for non-Gaussian symmetric populations and for a very asymmetric exponential population, either for exact solutions resulting from the independence of spacings or for approximate solutions for inferences on the comparison of k 2 location parameters without assuming equal dispersions.

2. Joint Probability Density Function of  ( X ¯ n , SS n ) and Probability Density Function of T n 1

Let X = X 1 , , X n be a random sample of size n from a population X. If X Gaussian ( μ , σ ) , then from the independence of X ¯ n and S S n , the joint probability density function of ( X ¯ n μ , S S n ) is
f ( X ¯ n μ , S S n ) ( w , s ) = f X ¯ n μ ( w ) f S S n ( s ) = 1 2 π ( σ / n ) e 1 2 ( w σ / n ) 2 s n 1 2 1 e s 2 σ 2 ( 2 σ 2 ) n 1 2 Γ ( n 1 2 ) I R × R + ( w , s ) ,
where Γ ( α ) = 0 x α 1 e x d x , α > 0 , is Euler’s Gamma function.

2.1. Joint Probability Density Function of ( X ¯ 2 , S S 2 )

The independence of X ¯ n and S S n does not hold for a non-Gaussian population X. However, for samples of size n = 2 , it is easy to obtain the probability density function of Student’s statistic T 1 .
Since
X ¯ 2 = X 1 + X 2 2 and S S 2 = ( X 1 X 2 ) 2 2 ,
it follows that if X 1 X 2 , then
X 1 = X ¯ 2 + S S 2 2 and X 2 = X ¯ 2 S S 2 2 ,
with a similar result also holding true if X 1 < X 2 , namely X 1 = X ¯ 2 S S 2 2 and X 2 = X ¯ 2 + S S 2 2 . Therefore, if we denote f X the probability density function of X, as the absolute value for the Jacobian determinant of both transformations is | J | = 1 2 S S 2 , then the joint probability density function of ( X ¯ 2 , S S 2 ) is
f ( X ¯ 2 , S S 2 ) ( w , s ) = f X 1 w + s 2 f X 2 w s 2 1 2 s + f X 1 w s 2 f X 2 w + s 2 1 2 s = 2 s f X w + s 2 f X w s 2 = 2 s f X w + 1 2 s f X w 1 2 s I S ( s , w ) ,
where the support
S = R × R + , if the support of X is R R + × ( 0 , 2 w 2 ) , if the support of X is R + [ a , b ] × ( 0 , 2 min ( w a ) 2 , ( b w ) 2 ) , if the support of X is [ a , b ] .
Note that if in (5), X Gaussian ( 0 , σ ) , then
f ( X ¯ 2 , S S 2 ) ( w , s ) = 2 s e s + 2 w 2 2 σ 2 2 π σ 2 = e 1 2 w σ / 2 2 2 π ( σ / 2 ) s 1 2 1 e s 2 σ 2 ( 2 σ 2 ) 1 2 Γ ( 1 2 ) I R × R + ( w , s ) ,
i.e., the same expression as the one given by (4) for n = 2 because Γ 1 2 = π .
As for the coefficients of s in the arguments of the functions in (5), they satisfy the conditions
1 2 + 1 2 = 0 and 1 2 2 + 1 2 2 = 1 .

2.2. Probability Density Function of Student’s Statistic T 1

To obtain the probability density function of Student’s statistic T 1 , let
T 1 = 2 X ¯ 2 S S 2 and U = S S 2 ,
from which the inverse transformation is X ¯ 2 = U T 1 2 and S S 2 = U 2 , with | J | = 2 U 2 . Hence, the joint probability density function of ( T 1 , U ) is
f ( T 1 , U ) ( t , u ) = 2 u f X u ( t + 1 ) 2 f X u ( t 1 ) 2 ,
and therefore, if X has support R , the probability density function of T 1 is given by
f T 1 ( t ) = 2 0 u f X u ( t + 1 ) 2 f X u ( t 1 ) 2 d u .
Some simple examples for symmetric parent distributions are the following:
  • If X Gaussian ( 0 , 1 ) ,
    f T 1 ( t ) = 1 π ( 1 + t 2 ) I R ( t ) ,
    i.e., T 1 Cauchy ( 0 , 1 ) (note that a standard Cauchy random variable is Student t-distributed with one degree of freedom).
  • If X Uniform ( 1 , 1 ) ,
    f T 1 ( t ) = 1 2 ( 1 + | t | ) 2 I R ( t ) .
  • If X Beta ( 2 , 2 ; 1 , 1 ) ,
    f T 1 ( t ) = 3 8 1 + 4 | t | + t 2 ( 1 + | t | ) 4 I R ( t ) .
  • If X Laplace ( 0 , 1 ) ,
    f T 1 ( t ) = 1 / 4 | t | 1 1 / ( 4 t 2 ) | t | > 1 .
  • If X Cauchy ( 0 , 1 ) ,
    f T 1 ( t ) = 1 2 π 2 t ln t + 1 t 1 2 I R 1 , 1 ( t ) .
(Note that a random variable X has a symmetric distribution around a parameter θ R if its cumulative distribution function F X satisfies F X ( θ x ) = F X ( θ + x ) , x R . In particular, if X is an absolutely continuous random variable, its distribution is symmetric around θ if f X ( θ x ) = f X ( θ + x ) , x R .)
The graphics in Figure 1 show that the probability density function of T 1 with a non-Gaussian parent can be quite different from that of t 1 , i.e., with a Gaussian parent. Note that with a Uniform ( 1 , 1 ) parent and a Beta ( 2 , 2 ; 1 , 1 ) parent, the corresponding densities of T 1 are unimodal and have heavier tails than those of t 1 (see Figure 2). In the case of a Cauchy parent, which is known for having very heavy tails, T 1 has an antimodal probability density function. The observation that heavier tails of the underlying parent distribution result in less heavy tails for the distribution of Student’s T-statistic, and vice-versa, was proved by van Zwet [16,17]. This indicates that the tail weight of the distribution is very important to determine the conservative or liberal behavior of Student’s statistic.

2.3. Joint Probability Density Function of ( X ¯ n , S S n )

Let X = ( X 1 , , X n + 1 ) , n 2 , be a random sample from the parent distribution X. From ( X ¯ n , S S n ) and X n + 1 being independent, the joint probability density function of ( X ¯ n , S S n , X n + 1 ) is
f ( X ¯ n , S S n , X n + 1 ) ( w , s , x ) = f ( X ¯ n , S S n ) ( w , s ) f X n + 1 ( x ) = f ( X ¯ n , S S n ) ( w , s ) f X ( x ) .
On the other hand, as
X ¯ n + 1 = n n + 1 X ¯ n + X n + 1 n + 1 and S S n + 1 = S S n + n n + 1 ( X ¯ n X n + 1 ) 2 ,
using the auxiliary random variable Y = X ¯ n X n + 1 , the inverse transformation is
X ¯ n = X ¯ n + 1 + Y n + 1 , S S n = S S n + 1 n n + 1 Y 2 and X n + 1 = X ¯ n + 1 n n + 1 Y ,
with | J | = 1 . Hence, the joint probability density function of ( X ¯ n + 1 , S S n + 1 , Y ) is
f ( X ¯ n + 1 , S S n + 1 , Y ) ( w , s , y ) = f ( X ¯ n , S S n ) w + y n + 1 , s n y 2 n + 1 f X w n y n + 1 ,
and therefore, the joint probability density function of ( X ¯ n + 1 , S S n + 1 ) is given by
f ( X ¯ n + 1 , S S n + 1 ) ( w , s ) = n + 1 n s n + 1 n s f ( X ¯ n , S S n ) w + y n + 1 , s n y 2 n + 1 f X w n y n + 1 d y .
We emphasize that in integral (8), it is assumed that X has support R , which we shall continue to assume for what follows, but if this is not the case, then the integration limits should be defined according to its support (see (6)).
In making the replacement v = y n + 1 n s in (8), it follows that
f ( X ¯ n + 1 , S S n + 1 ) ( w , s ) = 1 1 n + 1 n s f ( X ¯ n , S S n ) w + v s n ( n + 1 ) , s ( 1 v 2 ) f X w v n s n + 1 d v .
For example, if n + 1 = 3 , then in using Formulas (9) and (5),
f ( X ¯ 3 , S S 3 ) ( w , s ) = 1 1 3 2 s f ( X ¯ 2 , S S 2 ) w + v s 6 , s ( 1 v 2 ) f X w v 2 3 s d v = 3 1 1 1 1 v 2 f X w + v + 3 ( 1 v 2 ) s 6 × f X w + v 3 ( 1 v 2 ) s 6 f X w 2 v s 6 d v = 3 1 1 1 1 v 2 f X w + v + 3 ( 1 v 2 ) 6 s × f X w + v 3 ( 1 v 2 ) 6 s f X w 2 v 6 s d v .
Notice that the coefficients of s in the arguments of the functions in the last integrand in (10) satisfy the following two conditions:
v + 3 ( 1 v 2 ) 6 + v 3 ( 1 v 2 ) 6 2 v 6 = 0
and
v + 3 ( 1 v 2 ) 6 2 + v 3 ( 1 v 2 ) 6 2 + 2 v 6 2 = 1 .

2.4. The Gaussian Case

In applying (10), in particular, to the case X Gaussian ( 0 , σ ) , for w R and s > 0 ,
f ( X ¯ 3 , S S 3 ) ( w , s ) = 3 1 1 1 1 v 2 e s + 3 w 2 2 σ 2 2 2 π 3 / 2 σ 3 d v = 3 e s + 3 w 2 2 σ 2 2 2 π 3 / 2 σ 3 1 1 1 1 v 2 d v = 1 2 π σ 3 e 3 w 2 2 σ 2 I R ( w ) e s 2 σ 2 2 σ 2 I R + ( s ) f X w + s 2 f X ( w ) f X w s 2 ,
where 1 2 , 0 , 1 2 is a set of equidistant points such that their sum is 0 and the sum of their squares is 1.
As for the joint probability density function of ( X ¯ 4 , S S 4 ) , using the recurrence Formula (9), we obtain
f ( X ¯ 4 , S S 4 ) ( w , s ) = 4 s 3 1 1 f ( X ¯ 3 , S S 3 ) w + s 12 v , ( 1 v 2 ) s f X w 3 s 4 v d v ,
and in view of what was established in (11), the joint probability density function (12) can be approximated using the product of f X computed at the points w + α i , 4 s , i = 1 , 2 , 3 , 4 , i.e.,
f ( X ¯ 4 , S S 4 ) ( w , s ) K s i = 1 4 f X w + α i , 4 s ,
with K as a norming constant, provided that i = 1 4 α i , 4 = 0 and i = 1 4 α i , 4 2 = 1 .
In determining the points w + α i , 4 s , since the set A = { α 1 , 4 , α 2 , 4 , α 3 , 4 , α 3 , 4 } is symmetric, for its elements to fulfill the first condition, they must satisfy a 1 , 4 = a 4 , 4 and a 2 , 4 = a 3 , 4 . As the points are also equidistant, i.e., α i , 4 α i 1 , 4 = r , i = 2 , 3 , 4 , they form an arithmetic progression with common distance r, and thus, A = 3 r 2 , r 2 , r 2 , 3 r 2 . In order to satisfy the second condition, we must also have r 2 = 1 5 , and hence,
A = { α 1 , 4 , α 2 , 4 , α 3 , 4 , α 3 , 4 } = 3 20 , 1 20 , 1 20 , 3 20 .
Therefore, from (13), and for w R and s > 0 , we obtain
f ( X ¯ 4 , S S 4 ) ( w , s ) e 4 w 2 2 σ 2 s e s 2 σ 2 f X ¯ 4 ( w ) f S S 4 ( s ) .
More generally, we have the following:
  • If n = 2 k + 1 , then
    α 1 , n , , α n , n = n 1 2 r , n 1 2 1 r , , r , 0 , r , , n 1 2 1 r , n 1 2 r ,
    and from 2 n 1 2 n + 1 2 n 6 r 2 = 1 , we obtain r 2 = 12 n ( n 2 1 ) .
  • If n = 2 k , then
    α 1 , n , , α n , n = 2 n 1 2 r , 2 n 3 2 r , , r 2 , r 2 , , 2 n 3 2 r , 2 n 1 2 r ,
    and from 2 k = 0 n / 2 2 k 1 2 2 r 2 = 1 , we obtain r 2 = 12 n ( n 2 1 ) .
It follows that α 1 , n = 3 ( n 1 ) n ( n + 1 ) , and since α i , n = n 1 2 i + 1 r , i = 1 , , n , we have
α i , n = ( n + 1 2 i ) 3 n ( n 2 1 ) , i = 1 , , n .
In particular, α i , n 1 = ( n 2 i ) 3 ( n 1 ) n ( n 2 ) , and hence,
n 2 n + 1 α i , n 1 = ( n 2 i ) 3 n ( n 2 1 ) .
Therefore,
α i , n n 2 n + 1 α i , n 1 = 3 n ( n 2 1 ) = r 2 .
For example, for n = 2 and n = 3 we obtained the values
n = 2 : α 1 , 2 = 1 2 , α 2 , 2 = 1 2 ; n = 3 : α 1 , 3 = 1 2 , α 2 , 3 = 0 , α 3 , 3 = 1 2 .
As a recurrence formula,
α i , n = 3 n ( n 2 1 ) + n 2 n + 1 α i , n 1 , i = 1 , , n 1 , α n , n = 3 ( n 1 ) n ( n + 1 ) .
By expressing the multiplier n 2 n + 1 as n 2 n + 1 = 1 ξ n 2 , we obtain ξ n = 3 n + 1 .

3. Symmetry and Studentization

3.1. Symmetric Random Variables

There are useful results that enable the identification or characterization of symmetric random variables, such as the following:
(i)
If X and Y are independent random variables, then W = X Y has a characteristic function φ W ( t ) = E e i t X Y = E e i t X y d F Y ( y ) = φ X ( t y ) d F Y ( y ) .
(ii)
If X Bernoulli 1 2 , then B = 2 X 1 has a probability mass function P ( B = 1 ) = P ( B = 1 ) = 1 2 , and its characteristic function is φ B ( t ) = E e i t B = 1 2 e i t + 1 2 e i t = cos t .
(iii)
If X and B (B as defined in (ii)) are independent, the characteristic function of W = B X is φ W ( t ) = cos ( t x ) d F X ( x ) .
(iv)
A random variable is symmetric if and only if its characteristic function is φ X ( t ) = cos ( t x ) d F X ( x ) .
From the above, a random variable X is symmetric if and only if X = d X B , with X and B being independent. Therefore, if X and Y are independent random variables and X is symmetric, then X Y is also symmetric. In particular, if X is symmetric, then X ¯ n is symmetric, which leads to T n 1 being symmetric as well.

3.2. An Approximate Joint Probability Density Function of ( X ¯ n , S S n ) with a Symmetric Parent Distribution

Using the Gaussian case as a guideline, for n = 3 , we assume that f X is a continuous and smooth probability density function of a symmetric random variable X, in the sense that there is a value ξ 3 ( f X ) ( 1 , 1 ) such that, from the mean value theorem, the integral in (10) can be computed as
f ( X ¯ 3 , S S 3 ) ( w , s ) = 1 1 ξ 3 2 f X w + ξ 3 + 3 ( 1 ξ 3 2 ) 6 s × f X w ξ 3 + 3 ( 1 ξ 3 2 ) 6 s f X w 2 ξ 3 6 s = 1 1 ξ 3 2 i = 1 3 f X w + α i , 3 ( ξ 3 ) s ,
where α 1 , 3 ( ξ 3 ) = ξ 3 + 3 ( 1 ξ 3 2 ) 6 , α 2 , 3 ( ξ 3 ) = ξ 3 3 ( 1 ξ 3 2 ) 6 , and α 2 , 3 ( ξ 3 ) = 3 ξ 3 6 , so that i = 1 3 α i , 3 ( ξ 3 ) = 0 and i = 1 3 α i , 3 2 ( ξ 3 ) = 1 .
More generally, we investigate the consequences of a smoothness hypothesis, which is based on the Gaussian case.
Smoothness Hypothesis 1.
Let X 1 , X 2 , be independent replicas of a random variable X, with a smooth probability density function f X that allows the integral mean value theorem to be used. Therefore, for ν = 2 , , n , there are values ξ ν ( f X ) ( 1 , 1 ) and α i , ν ( ξ ν ) of the form
α i , ν = ξ ν ν ( ν 1 ) + α i , ν 1 1 ξ ν 2 , i = 1 , , ν 1 , and α ν , ν = ξ ν ν 1 ν ,
satisfying the conditions i = 1 ν α i , ν ( ξ ν ) = 0 and i = 1 ν α i , ν 2 ( ξ ν ) = 1 such that the approximation
f ( X ¯ ν , S S ν ) ( w , s ) K s ( ν 3 ) / 2 i = 1 ν f X ( w + α i , ν s ) I S ( w , s ) ,
where K is a norming constant, is valid.
Then, the integral mean value theorem also holds for f ( X ¯ n + 1 , S S n + 1 ) , i.e., there are values ξ n + 1 ( f X ) ( 1 , 1 ) and α i , n + 1 ( ξ n + 1 ) such that the approximation
f ( X ¯ n + 1 , S S n + 1 ) ( w , s ) K s ( n + 1 3 ) / 2 i = 1 n + 1 f X ( w + α i , n + 1 ( ξ n + 1 ) s ) I S ( w , s )
is also valid, with ( X ¯ n + 1 , S S n + 1 ) having support S defined in (6), and α i , n + 1 = ξ n + 1 n ( n + 1 ) + α i , n 1 ξ n + 1 2 , i = 1 , , n , and α n + 1 , n + 1 = ξ n + 1 n n + 1 .
If we further assume that X is a symmetric random variable, which implies that the points α 1 , n , , α n , n form an arithmetic progression with common distance r n = 2 3 n ( n 2 1 ) and α i , n = α n + 1 i , n , i = 1 , , n , then ξ n = 3 n + 1 and
α i , n = ( n + 1 2 i ) 3 n ( n 2 1 ) , i = 1 , , n .
Equation (19) is easily obtained. In fact, if n = 2 k + 1 , then α 1 , n = k | r n | . Moreover, the abcissas of the positive points α i , n are | r | , 2 | r | , , k | r | , and since j = 1 k ( j r ) 2 = 1 2 , this implies that | r | = 2 3 n ( n 2 1 ) . For n = 2 k , the results are similar, with the abcissas of the positive α i , n now being | r | 2 , 3 | r | 2 , , ( k 1 2 ) | r | .

3.3. An Approximate Expression for the Probability Density Function of T n 1 with a Smooth Symmetric Parent

Considering the transformation
T n 1 = n ( n 1 ) X ¯ n S S n and U = S S n ,
from which X ¯ n = U T n 1 n ( n 1 ) and S S n = U 2 , with | J | = 2 U 2 n ( n 1 ) , we obtain
f ( T n 1 , U ) ( t , u ) = 2 u 2 n ( n 1 ) f ( X ¯ n , S S n ) u t n ( n 1 ) , u 2 ,
and therefore, for t R , and X with support R ,
f T n 1 ( t ) = 2 n ( n 1 ) 0 u 2 f ( X ¯ n , S S n ) u t n ( n 1 ) , u 2 d u .
Using approximation (18), and denoting α i , n ( ξ n ) = α i , n , it follows that
f T n 1 ( t ) K 0 u n 1 i = 1 n f X t n ( n 1 ) + α i , n u d u .
For example, if X Gaussian ( 0 , 1 ) , then
f t n 1 ( t ) K 0 u n 1 i = 1 n exp 1 2 t n ( n 1 ) + α i , n u 2 d u = = K 0 u n 1 exp u 2 2 t 2 n 1 + 1 d u = K 1 1 + t 2 n 1 n / 2 ,
as expected.
If X Laplace ( 0 , 1 ) , where f X ( x ) = 1 2 e | x | , then
f T n 1 ( t ) K 0 u n 1 i = 1 n exp t n ( n 1 ) + α i , n u d u = = K 0 u n 1 exp u i = 1 n t n ( n 1 ) + α i , n d u ,
and therefore,
f T n 1 ( t ) K i = 1 n t n ( n 1 ) + α i , n n I R ( t ) .
In particular, if n = 3 ,
f T 2 ( t ) K t 6 + 1 2 + | t | 6 + t 6 1 2 3 I R ( t ) .
Computing the norming constant so that the right-hand side of (21) is transformed into a probability density function, we obtain
f T 2 ( t ) 9 6 19 t 6 + 1 2 + | t | 6 + t 6 1 2 3 I R ( t ) .
In Figure 3, the approximate probability density function of T 2 with a Laplace ( 0 , 1 ) parent is represented.
For the non-Gaussian parent case, to assess how well approximation (20) works, we shall compare the exact probability density function of T 2 with a Uniform ( 1 , 1 ) parent, given by Perlo [12], namely
f T 2 ( t ) = 3 1 9 t 2 4 t 2 2 ( 4 t 2 ) 1 t 2 + 3 3 ( t 2 + 2 ) ( 4 t 2 ) 5 / 2 arctanh 1 t 2 4 t 2 , | t | 1 2 9 1 | t | + 1 ) + 3 | t | t 2 4 4 ( | t | + 1 ) ( 4 t 2 ) + 3 3 ( t 2 + 2 ) ( t 2 4 ) 5 / 2 arctan t 2 4 3 ( t + 2 ) , | t | > 1 2 ,
with the approximation using (20).
As we are now dealing with a parent distribution with a limited support, the integration limits for (20) must be defined according to (6). In this case, the condition
6 < u t < 6 0 < u 2 < 2 min u t 6 + 1 2 , 1 u t 6 2
must be satisfied, or equivalently,
6 < u t < 6 0 < u < min 2 u t 6 + 1 , 2 1 u t 6 .
Hence, in denoting a = min 2 u t 6 + 1 , 2 1 u t 6 , it follows that
f T 2 ( t ) K 0 a u 2 1 2 3 d u 0 a u 2 d u = 2 6 ( 3 + | t | ) 3 I R ( t ) .
Computing the norming constant K so that the function in the right-hand side of (23) becomes a probability density function, we obtain
f T 2 ( t ) 3 ( 3 + | t | ) 3 I R ( t ) .
In Figure 4, the exact and approximate probability density functions of T 2 with a Uniform ( 1 , 1 ) parent, defined in (22) and (24), respectively, are plotted together. As can be observed, the approximation is not good in the central region of the distributions, but it is quite good in the tails of the distributions, which is where it actually matters because inferences are made using the tails of the distributions.
Observe that Hendriks et al. [11] analyzed why the Uniform parent is the least favorable one, as far as approximations are concerned, for the probability density function of T n 1 .

4. Externally Studentized Statistics Using Spacings of an Exponential Parent

4.1. External Studentization Using the Maximum Likelihood Scale Estimator

Let X = X 1 , , X n be a random sample of size n from X E x p o n e n t i a l ( λ , δ ) , λ R , δ > 0 , i.e., with probability density function
f X ( x ) = 1 δ e x λ δ I ( λ , ) ( x ) .
The maximum likelihood estimators λ ^ = X 1 : n Exponential λ , δ n and δ ^ = X ¯ n X 1 : n = 1 n k = 1 n ( X k : n X 1 : n ) = 1 n k = 2 n ( n + 1 k ) ( X k : n X k 1 : n ) Gamma n 1 , δ n are independent due to the independence of the spacings X k : n X k 1 : n Exponential 0 , δ n + 1 k , k = 1 , , n , with the usual convention X 0 : n = λ .
Therefore, the externally Studentized statistic
T n 1 = X 1 : n λ X ¯ n X 1 : n ,
where T n 1 = d Z 1 : n Z ¯ n Z 1 : n , with ( Z 1 , , Z n ) being a random sample from Z = X λ δ Exponential ( 0 , 1 ) , has probability density function
f T n 1 ( t ) = f Z 1 : n ( t x ) f Z ¯ n Z 1 : n ( x ) | x | d x = 0 n e n t x n n 1 x n 2 e n x Γ ( n 1 ) x d x = n n Γ ( n 1 ) 0 x n 1 e n ( t + 1 ) x d x = n 1 ( 1 + t ) n I ( 0 , ) ( t ) ,
and hence, T n 1 Pareto ( n 1 , 0 ) .

4.2. External Studentization Using the Sample Range as a Dispersion Estimator

Another externally Studentized statistic that can be used to make inferences on the location parameter λ is
τ n 1 = X 1 : n λ X n : n X 1 : n .
In noticing that τ n 1 = d Z 1 : n Z n : n Z 1 : n and considering that Z n : n Z 1 : n = d Z n 1 : n 1 , the joint probability density function of ( Z 1 : n , Z n : n Z 1 : n ) is
f ( Z 1 : n , Z n : n Z 1 : n ) ( x , y ) = f Z 1 : n ( x ) f Z n : n Z 1 : n ( y ) = n e n x ( n 1 ) 1 e y n 2 e y I R + × R + ( x , y ) .
Using the transformation S = Z 1 : n and τ n 1 , with | J | = S τ n 1 2 , the joint probability density function of ( S , τ n 1 ) is defined as
f ( S , τ n 1 ) ( s , t ) = n ( n 1 ) e n s e s / t ( 1 e s / t ) n 2 s t 2 I R + × R + ( s , t ) .
Therefore, for t > 0 ,
f τ n 1 ( t ) = 1 t 0 n s e n s ( n 1 ) e s / t ( 1 e s / t ) n 2 1 t d s = 1 t 0 n ( 1 n s ) e n s ( 1 e s / t ) n 1 d s = 1 t 0 n e n s ( 1 e s / t ) n 1 d s + 1 t 0 n 2 s e n s ( 1 e s / t ) n 1 d s .
By making the replacement u = 1 e s / t in the integrals, we obtain
f τ n 1 ( t ) = n 0 1 ( 1 u ) n t 1 u n 1 d u n 2 t 0 1 ln ( 1 u ) ( 1 u ) n t 1 u n 1 d u = n B ( n , n t ) + n t B ( n , n t ) ( n t ) .
Since
B ( n , n t ) ( n t ) = B ( n , n t ) ψ ( n t ) ψ ( n + n t ) ,
where ψ ( z ) = d d z ln z = Γ ( z ) Γ ( z ) is the digamma function, it follows that
f τ n 1 ( t ) = n B ( n , n t ) n t ψ ( n + n t ) ψ ( n t ) 1 I ( 0 , ) ( t ) .
Using the recurrence formula for the digamma function,
ψ ( n + z ) = k = 1 n 1 n k + z + ψ ( z ) , n N ,
(cf. Abramowitz and Stegun [18]), the probability density function of τ n 1 , given in (27), can be further simplified to
f τ n 1 ( t ) = n B ( n , n t ) k = 1 n 1 n t n k + n t I ( 0 , ) ( t ) .
In Table A1, in Appendix A, critical values of τ n 1 are displayed. The critical values were obtained using Mathematica v12. In Figure 5, the probability density function of τ n 1 is plotted for some values of n.
For large values of n, an approximation can be used to obtain critical values of the statistic τ n 1 . In recalling that Z n : n Z 1 : n = d Z n 1 : n 1 , the limit distribution of the sequence of maxima ( Z n 1 : n 1 ln ( n 1 ) ) n 2 is a standard Gumbel distribution, and therefore,
Z n : n Z 1 : n ln ( n 1 ) n d V Gumbel ( 0 , 1 ) .
Moreover, as lim n ln n ln ( n 1 ) = 1 , in using the fact that Z n 1 : n 1 ln n P 1 and n Z 1 : n = d Z , it follows from Slutsky’s theorem [19] that
n ln n τ n 1 n d Z Exponential ( 0 , 1 ) .

4.3. Internal Studentization Using Sums of Spacings

As the minimum X 1 : n and/or the maximum X n : n can be outliers, according to Hampel’s [20] breakdown point concept, the statistic τ n 1 , defined in (26), can be unreliable. For an overview of robustness issues, see Rochetti [21], and in what concerns the duality of perspectives for extreme values/outliers, see Bhattacharya et al. [22]. For these reasons, it is more conservative to use instead the internally Studentized statistic
τ n 1 ; i , k = X ¯ n λ X k : n X i : n ,
with 1 < i < k < n .
The sample mean X ¯ n is a sufficient and complete estimator of λ , and the distribution of τ n 1 ; i , k does not depend on the nuisance scale parameter δ . Hence, from Basu’s [23] theorem, we conclude that X ¯ n and τ n 1 ; i , k are independent.
Noticing also that τ n 1 ; i , k = d Z ¯ n Z k : n Z i : n , from the equality
Z k : n Z i : n = Z ¯ n τ n 1 ; i , k ,
we must have f Z k : n Z i : n = f Z ¯ n τ n 1 ; i , k . Hence, from Z ¯ n Gamma n , 1 n and Z k : n Z i : n = d Z k i : n i , we have for x > 0 ,
Γ ( n + 1 i ) Γ ( k i ) Γ ( n + 1 k ) e x ( n + 1 k ) 1 e x k i 1 = 0 f Z ¯ n ( x t ) f τ n 1 ; i . k ( t ) t d t ,
i.e.,
Γ ( n + 1 i ) Γ ( k i ) Γ ( n + 1 k ) e x ( n + 1 k ) 1 e x k i 1 = 0 n ( n x t ) n 1 e n x t Γ ( n ) f τ n 1 ; i . k ( t ) t d t .
Thus, denoting L ( f ; x ) the Laplace transform of function f at x, we obtain
L ( t n f τ n 1 ; i , k ( t ) ; x ) = Γ ( n ) Γ ( n + 1 i ) n Γ ( k i ) Γ ( n + 1 k ) e x ( n + 1 k ) / n ( 1 e x / n ) k i 1 x n 1 ,
and in denoting L 1 ( g ; t ) the inverse Laplace transform of function g at t, it follows that
f τ n 1 ; i , k ( t ) = Γ ( n ) Γ ( n + 1 i ) n Γ ( k i ) Γ ( n + 1 k ) t n L 1 e x ( n + 1 k ) / n ( 1 e x / n ) k i 1 x n 1 ; t .
For example, if n = 2 , i = 1 and k = 2 ,
f τ 1 ; 1 , 2 ( t ) = 1 2 t 2 L 1 e x / 2 x ; t = 1 2 t 2 I ( 1 2 , ) ( t ) ,
i.e., τ 1 ; 1 , 2 Pareto ( 1 , 1 2 ) .
The most interesting scenario for (29) is when the integers n, i, and k satisfy the condition n + 1 k = k i 1 , which occurs in the following cases:
  • n = 3 j 1 , i = j 1 and k = 2 j ;
  • n = 3 j , i = j and k = 2 j + 1 ;
  • n = 3 j + 1 , i = j + 1 and k = 2 j + 2 .
If this happens, then
f τ n 1 ; i , k ( t ) = Γ ( n ) Γ ( 2 j + 1 ) n Γ ( j ) Γ ( j + 1 ) 1 t n L 1 1 x k 2 e x / n ( 1 e x / n ) x j ; t .
Since
L 1 1 x k 2 ; t = t k 3 Γ ( k 2 ) and L 1 e x / n ( 1 e x / n ) x ; t = I ( 1 n , 2 n ) ( t ) ,
from the fact that the inverse Laplace transform of a product is the convolution of the inverse Laplace transforms of the factors, we obtain, for example, for n = 3 , i = 1 , and k = 3 ,
f τ 2 ; 1 , 3 ( t ) = 0 , t < 1 3 4 ( t 1 3 ) 3 t 3 , 1 3 t < 2 3 4 9 t 3 , t 2 3 ,
and for n = 4 , i = 2 , and k = 4 ,
f τ 3 ; 2 , 4 ( t ) = 0 , t < 1 4 3 ( 4 t 1 ) 2 32 t 4 , 1 4 t < 1 2 3 ( 8 t 3 ) 32 t 4 , t 1 2 .
In Table A2, in Appendix A, critical values of τ n 1 ; i , k are supplied for integers n, i, and k satisfying n + 1 k = k i 1 . The critical values of τ n 1 ; i , k were also obtained with Mathematica v12. The probability density functions of τ 2 ; 1 , 3 and τ 3 ; 2 , 4 are shown in Figure 6.

4.4. Comparing the Locations of Two Exponential Populations with Equal Dispersions

For two exponential populations X Exponential ( λ 1 , δ 1 ) and Y Exponential ( λ 2 , δ 2 ) , it may be of interest to make inferences on λ 1 λ 2 . For this purpose, Theorem 1 can be helpful.
Theorem 1.
Let X 1 and X 2 be two independent random variables such that X 1 Exponential ( λ 1 , δ 1 ) and X 2 Exponential ( λ 2 , δ 2 ) . Then, X 1 X 2 has an asymmetric Laplace distribution with probability density function
f X 1 X 2 ( x ) = 1 δ 1 + δ 2 e x ( λ 1 λ 2 ) δ 2 , x λ 1 λ 2 1 δ 1 + δ 2 e x ( λ 1 λ 2 ) δ 1 , x > λ 1 λ 2 .
Proof. 
Let Y = ( X 1 λ 1 ) ( X 2 λ 2 ) . Since X 1 λ 1 Exponential ( 0 , δ 1 ) and X 2 λ 2 Exponential ( 0 , δ 2 ) , for y 0 ,
f Y ( y ) = y f X 1 λ 1 ( y + x ) f X 2 λ 2 ( x ) d x = 1 δ 1 δ 2 y e y + x δ 1 e x δ 2 d x = 1 δ 1 + δ 2 e y δ 2 ,
and for y > 0 ,
f Y ( y ) = 0 f X 1 λ 1 ( y + x ) f X 2 λ 2 ( x ) d x = 1 δ 1 δ 2 0 e y + x δ 1 e x δ 2 d x = 1 δ 1 + δ 2 e y δ 1 .
Hence, from X 1 X 2 = d Y + ( λ 1 λ 2 ) , it follows that
f X 1 X 2 ( x ) = f Y ( x ( λ 1 λ 2 ) ) = 1 δ 1 + δ 2 e x ( λ 1 λ 2 ) δ 2 , x λ 1 λ 2 1 δ 1 + δ 2 e x ( λ 1 λ 2 ) δ 1 , x > λ 1 λ 2 .
If in Theorem 1 it is considered that δ 1 = δ 2 = δ , the symmetric Laplace distribution with location parameter λ 1 λ 2 R , and scale parameter δ > 0 is obtained. For more details on asymmetric Laplace random variables, see Brilhante and Kotz [24].
When dealing with two populations with equal dispersions, inferences on the difference between the locations of the two populations can be carried out using an externally Studentized statistic as described below.
Let X = X 1 , , X n 1 and Y = Y 1 , , Y n 2 be two independent random samples with parent distributions X Exponential ( λ 1 , δ ) and Y Exponential ( λ 2 , δ ) , respectively. Considering that
X 1 : n 1 λ 1 δ Exponential 0 , 1 n 1 and Y 1 : n 2 λ 2 δ Exponential 0 , 1 n 2 ,
from Theorem 1,
U = X 1 : n 1 Y 1 : n 2 ( λ 1 λ 2 ) δ
has probability density function
f U ( x ) = n 1 n 2 N e n 2 x , x 0 n 1 n 2 N e n 1 x , x > 0 ,
with N = n 1 + n 2 .
On the other hand, as
n 1 X ¯ n 1 X 1 : n 1 = k = 2 n 1 ( n 1 + 1 k ) X k : n 1 X k 1 : n 1 Gamma ( n 1 1 , δ ) ,
and, similarly, n 2 Y ¯ n 2 Y 1 : n 2 Gamma ( n 2 1 , δ ) , we have
V = δ ^ δ = n 1 X ¯ n 1 X 1 : n 1 + n 2 Y ¯ n 2 Y 1 : n 2 / N δ Gamma N 2 , 1 N .
The independence of spacings in exponential populations ensures that U and V are independent, and therefore, the externally Studentized statistic W n 1 1 , n 2 1 = U V , i.e.,
W n 1 1 , n 2 1 = X 1 : n 1 Y 1 : n 2 ( λ 1 λ 2 ) n 1 ( X ¯ n 1 X 1 : n 1 ) + n 2 ( Y ¯ n 2 Y 1 : n 2 ) / N ,
can be used to make inferences on λ 1 λ 2 .
As for the probability density function of W n 1 1 , n 2 1 , for w 0 ,
f W n 1 1 , n 2 1 ( w ) = f U ( w x ) f V ( x ) | x | d x = n 1 n 2 N N 3 Γ ( N 2 ) 0 x N 2 e ( N n 2 w ) x d x = n 1 n 2 ( N 2 ) N 2 1 n 2 N w N 1 ,
and for w > 0 ,
f W n 1 1 , n 2 1 ( w ) = f U ( w x ) f V ( x ) | x | d x = n 1 n 2 N N 3 Γ ( N 2 ) 0 x N 2 e ( N + n 1 w ) x d x = n 1 n 2 ( N 2 ) N 2 1 + n 1 N w N 1 .
Therefore,
f W n 1 1 , n 2 1 ( w ) = n 1 n 2 ( N 2 ) N 2 1 n 2 N w N 1 , w 0 n 1 n 2 ( N 2 ) N 2 1 + n 1 N w N 1 , w > 0 .
From (30), P ( W 0 ) = n 1 N and P ( W > 0 ) = n 2 N = 1 n 1 N . Notice that n 1 N is the proportion of observations in the combined sample that comes from the population X.
The critical value W n 1 1 , n 2 1 ; α , 0 < α < 1 , of W n 1 1 , n 2 1 is
W n 1 1 , n 2 1 ; α = N n 2 1 n 1 N α 1 N 2 , 0 < α n 1 N N n 1 n 2 N ( 1 α ) 1 N 2 1 , n 1 N < α < 1 .
If a balanced design is considered instead, i.e., n 1 = n 2 = n in (30), denoting, for simplicity, W n 1 , n 1 = W n 1 , then
f W n 1 ( w ) = n 1 2 1 + | w | 2 2 n 1 I R ( w ) .
The ( 1 α ) -th critical value of W n 1 is W n 1 ; 1 α = 2 1 ( 2 α ) 1 / 2 ( n 1 ) 1 , α < 0.5 . (The distribution of W n 1 is symmetric around zero.)
In Figure 7, the probability density function of W n 1 , n 2 is plotted for some values of n 1 and n 2 .
Observe that if the two populations have unequal dispersions, i.e., δ 1 δ 2 , the distribution of W n 1 1 , n 2 1 depends on the nuisance scale parameters δ 1 and δ 2 , and so, it cannot be used to make inferences on λ 1 λ 2 , unless a Satterthwaite [15] approximation is considered (for more details, see Section 4.6).

4.5. Analysis of Spacings (ANOSp) for Testing Homogeneity of k > 2 Locations of Exponential Populations with Equal Dispersions

Let X 1 = X 11 , , X 1 n 1 , , X k = X k 1 , , X k n k be k > 2 independent random samples, with X i j Exponential ( λ i , δ i ) , i = 1 , , k , j = 1 , , n i . For what follows, we shall use the notations X ¯ i for the sample mean and X j : n i ( i ) for the j-th ascending order statistic, j = 1 , , n i , of the i-th random sample, i = 1 , , k . We shall also use the notations X ¯ = 1 N i = 1 k j = 1 n i X i j and X 1 : N = min 1 i k 1 j n i X i j = min 1 i k X 1 : n i ( i ) , with N = n 1 + + n k denoting the size of the combined samples.
The maximum likelihood estimators of the location and scale parameters of the individual populations are
λ ^ i = X 1 : n i ( i ) Exponential λ i , δ i n i and δ ^ i = X ¯ i X 1 : n i ( i ) Gamma n i 1 , δ i n i ,
i = 1 , , k . Our interest lies in testing the homogeneity of locations of the populations, i.e., testing
H 0 : λ 1 = = λ k ( = λ ) vs . H A : λ i λ j , for some i , j { 1 , , , k } , i j .
For the time being, we shall assume that the populations have equal dispersions, i.e., δ 1 = = δ k = δ .
In this setting, a parallelism with a one-way ANOVA scheme can be made, since the Total Sum of Spacings (TSSp) can be split into a Between Sum of Spacings (BSSp) and a Within Sum of Spacings (WSSp) as follows:
TSSp = i = 1 k j = 1 n i X i j X 1 : N = i = 1 k j = 1 n i X i j X 1 : n i ( i ) + X 1 : n i ( i ) X 1 : N = i = 1 k n i X 1 : n i ( i ) X 1 : N + i = 1 k j = 1 n i X i j X 1 : n i ( i ) = BSSp + WSSp .
In noticing that
WSSp = i = 1 k j = 2 n i X j : n i ( i ) X 1 : n i ( i ) = i = 1 k j = 2 n i ( n i + 1 j ) X j : n i ( i ) X j 1 : n i ( i ) Gamma ( N k , δ ) ,
the statistic WSSp N k is an unbiased estimator of δ , regardless of the validity of H 0 .
On the other hand, under H 0 , the random variables Y i = n i X 1 : n i ( i ) Exponential ( λ , δ ) , i = 1 , , k , are independent. Thus, from
BSSp = i = 1 k n i X 1 : n ( i ) X 1 : N = i = 1 k n i X 1 : n i ( i ) N X 1 : N
and N X 1 : N = d k Y 1 : k , with Y 1 : k = min { Y 1 , , Y k } , it follows that
BSSp = d i = 1 k Y i k Y 1 : k = k Y ¯ k Y 1 : k Gamma k 1 , δ .
Thus, under H 0 , the statistic BSSp k 1 is an unbiased estimator of δ .
From the above, and under H 0 , the F-statistic
F = MBSSp MWSSp = BSSp k 1 WSSp N k = BSSp / ( 2 δ ) 2 ( k 1 ) WSSp / ( 2 δ ) 2 ( N k ) F 2 ( k 1 ) , 2 ( N k )
can detect gross departures from the null hypothesis. Notice that the independence of spacings in exponential populations guarantees the independence of BSSp and WSSp.
An ANOSp table, similar to a one-way ANOVA table, can be shown in this context (see Table 1).

4.6. Analysis of Spacings (ANOSp) for Testing Homogeneity of k > 2 Locations of Exponential Populations with Unequal Dispersions

If the exponential populations have unequal dispersions, the statistic defined in (32) is useless, because its distribution depends now on the nuisance scale parameters of the individual populations. However, a Satterthwaite approximation can be considered to eliminate its distribution’s dependence on those nuisance parameters.
Satterthwaite [15] showed that if Y 1 , , Y n are independent random variables such that Y i χ ν i 2 , i = 1 , , n , the linear combination i = 1 n a i Y i , with a i R , i = 1 , , n , can be approximated using Y ν , where Y χ v 2 , i.e., i = 1 n a i Y i Y ν , where the degree of freedom ν is estimated using the estimator
ν ˜ = i = 1 n a i Y i 2 i = 1 n a i 2 ν i Y i 2 .
Since under H 0 , the distribution of MBSSp does not depend on λ , we shall assume, without loss of generality, that λ = 0 . We shall also assume that from X 1 : N = min 1 i k X 1 : n i ( i ) , in particular, X 1 : N = X 1 : n 1 ( 1 ) , and therefore,
MBSSp = 1 k 1 i = 1 k n i X 1 : n i ( i ) X 1 : N = 1 k 1 i = 1 k n i X 1 : n i ( i ) X 1 : n 1 ( 1 ) = n 1 N k 1 X 1 : n 1 ( 1 ) + i = 2 k n i k 1 X 1 : n i ( i ) = n 1 N k 1 λ ^ 1 + i = 2 k n i k 1 λ ^ i ,
which can be expressed as
MBSSp = δ 1 ( n 1 N ) 2 n 1 ( k 1 ) Y 1 + i = 2 k δ i 2 ( k 1 ) Y i U ν 1 ,
with Y i = 2 n i δ i λ ^ i χ 2 2 , i = 1 , , k , and U χ ν 1 2 . In using a 1 = δ 1 ( n 1 N ) 2 n 1 ( k 1 ) and a i = δ i 2 ( k 1 ) , i = 2 , , k , in Formula (33), the parameter ν 1 is estimated by
ν ˜ 1 = 2 i = 1 k n i λ ^ i N λ ^ 1 ( n 1 N ) 2 λ ^ 1 2 + i = 2 k n i 2 λ ^ i 2 .
On the other hand, noticing that
MWSSp = 1 N k i = 1 k j = 1 n i X i j X 1 : n i ( i ) = 1 N k i = 1 k n i X ¯ i X 1 : n i ( i ) = i = 1 k n i N k δ ^ i ,
with δ ^ i Gamma ( n i 1 , δ i ) , i = 1 , , k , and then considering Y i = 2 δ i δ ^ i χ 2 ( n i 1 ) 2 , i = 1 , , k , we have
MWSSp = i = 1 k n i δ i 2 ( N K ) Y i V ν 2 ,
with V χ ν 2 2 . In using a i = n i δ i 2 ( N K ) , i = 1 , , n , in Formula (33), the parameter ν 2 is estimated by
ν ˜ 2 = 2 i = 1 k n i δ ^ i 2 i = 1 k n i 2 n i 1 δ ^ i 2 .
Consequently, under H 0 , U and V are independent, and therefore,
F = MBSSp MWSSp U / ν 1 V / ν 2 F ν 1 , ν 2 ,
where the degrees of freedom ν 1 and ν 2 of the F-statistic are estimated using (34) and (35), respectively.

5. Conclusions

The exact distribution of externally Studentized statistics in Gaussian samples and in exponential samples are readily obtained. The members of the location-scale exponential family with a pair of sufficient statistics for the parameters are as follows:
  • X Gaussian ( μ , σ ) , when the support is the real line;
  • Y Exponential ( λ , δ ) , when the support is the half-line ( λ , ) (or the reverted exponential when the support is ( , λ ) , in which case the pair of sufficient statistics is Y n : n , k = 1 n Y k , and the maximum likelihood estimators of the parameters are λ ^ = Y n : n and δ ^ = Y n : n Y ¯ n );
  • W Uniform ( λ , λ + δ ) when the support is a segment, in which case the pair of sufficient statistics is ( W 1 : n , W n : n ) and the maximum likelihood estimators of the parameters are λ ^ = W 1 : n and δ ^ = W n : n W 1 : n .
While for Gaussian samples and for exponential samples, the probability density functions of externally Studentized statistics are easily obtained, in the case of uniform samples, this is not possible. However, as uniform random variables are symmetric by default, the approximation under the smoothness hypothesis 1 holds, although it is known from Hendriks [11] that this is the worst framework to consider. Alternatively, as uniform samples can be logarithmically transformed into exponential data, inferences on the location parameter(s) can be dealt with using the results of Section 4.
On the other hand, it is, in general, impossible to obtain exact results for internally Studentized statistics. As Hotelling [25], Efron [10], and Lehman [26] have shown, symmetry for the parent distribution is a useful property for Studentizations. The investigation of the approximation resulting from the smoothness hypothesis (17) for symmetric parents, inspired by the Gaussian case, shows that even in the worst Uniform ( 1 , 1 ) case, which was identified by Hendriks et al. [11] as such, the approximation in the tails is quite good, and therefore, it is usable for inferences on the location parameter.
With regard to asymmetric exponential parent distributions, from the independence of spacings, exact results either for one sample inferences on the location parameter or for the comparison of k 2 location parameters assuming equal dispersions are straightforward. An approximate solution for the unequal dispersions case, inspired by Satterthwaite’s [15] treatment for comparing means in heteroscedastic Gaussian parents, has been presented.

Author Contributions

Conceptualization, M.d.F.B., D.P. and M.L.R.; methodology, M.d.F.B. and D.P.; investigation, M.d.F.B., D.P. and M.L.R.; writing—original draft preparation, M.d.F.B. and D.P.; writing—review and editing, M.d.F.B., D.P. and M.L.R.; supervision, M.d.F.B. and D.P.; project administration, M.d.F.B. and D.P.; funding acquisition, M.d.F.B., D.P. and M.L.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially financed by national funds through FCT (Fundação para a Ciência e a Tecnologia), Portugal, under project UIDB/00006/2020 (https://doi.org/10.54499/UIDB/00006/2020), and research grant UIDB/00685/2020 of the CEEAplA (Centre of Applied Economics Studies of the Atlantic), School of Business and Economics of the University of the Azores.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Tables with Critical Values

Table A1. Critical values of τ n 1 = X 1 : n λ X n : n X 1 : n .
Table A1. Critical values of τ n 1 = X 1 : n λ X n : n X 1 : n .
n0.0010.0050.010.0250.050.10.90.950.9750.990.9950.999
20.0005010.0025130.0050510.0128210.0263160.0555564.59.519.549.599.5499.5
30.0002220.0011150.0022400.0056660.0115620.02411011.6147632.4860794.2169916.16875014.408050
40.0001360.0006840.0013730.0034700.0070680.0146770.50.751.0670301.6184602.1644904.047390
50.0000960.0004820.0009660.0024410.0049660.0102910.3191260.4629480.6357820.9177451.1797501.999380
60.0000730.0003660.0007350.0018550.0037710.0078050.2289900.3259320.4388210.6162370.7750771.244660
70.0000580.0002920.0005870.0014810.0030100.0062240.1760150.2474670.3289730.4539850.5632220.874548
80.0000480.0002420.0004850.0012240.0024870.0051400.1415630.1973180.2599960.3544950.4356710.661185
90.0000410.0002050.0004110.0010380.0021080.0043540.1175660.1628210.2131430.2880520.3515920.524816
100.0000350.0001770.0003560.0008970.0018220.0037620.10.1378050.1794880.2409300.2925400.431222
110.0000310.0001560.0003120.0007880.0015990.0033010.0866470.1189300.1542830.2059860.2490790.363559
120.0000280.0001380.0002780.0007000.0014220.0029340.0761930.1042390.1347820.1791650.2159230.312669
130.0000250.0001240.0002490.0006290.0012770.0026340.0678110.0925180.1193000.1580090.1898980.273188
140.0000220.0001120.0022920.0005700.0011570.0023860.0609580.0829740.1067450.1409450.1689950.241782
150.0000210.0001030.0002060.0005200.0010560.0021770.0552610.0750690.0963810.1269260.1518810.216280
160.0000190.0000940.0001890.0004780.0009700.0019990.0504590.0684260.0876990.1152280.1376440.195213
170.0000170.0000870.0001750.0004410.0008960.0018470.0463630.0627740.0803320.1053360.1256360.177552
180.0000160.0000810.0001620.00040100.0008310.0017140.0428320.0579130.0740110.0968740.1153880.162561
190.0000150.0000760.0001510.0003820.0007750.0015970.0397610.0536940.0685350.0895640.1065530.149697
200.0000140.0000710.0001420.0003570.0007250.0014950.0370670.050000.0637510.0831910.0988650.138550
210.0000130.0000660.0001330.0003360.0006810.0014040.0346870.0467430.0595380.0775930.0921220.128812
220.0000120.0000630.0001250.0003160.0006420.0013220.0325710.0438510.0558040.0726410.0861660.120238
230.0000120.0000590.0001180.0002990.0006060.0012490.0306790.0412680.0524740.0682320.0808710.112640
240.0000110.0000560.0001120.0002830.0005740.0011830.0289780.0389500.0494880.0642850.0761370.105864
250.0000110.0000530.0001070.0002690.0005450.0011230.0274410.0368570.0467960.0607330.0718800.099790
300.0000080.0000420.0000850.0002130.0004330.0008910.0215680.0288830.0365670.0472810.0558050.076986
500.0000040.0000220.0000450.0001130.0002300.0004720.0111860.0148870.0187320.0240310.0282000.038399
Table A2. Critical values of τ n 1 ; i , k = X ¯ n λ X k : n X i : n .
Table A2. Critical values of τ n 1 ; i , k = X ¯ n λ X k : n X i : n .
nik0.0010.0050.010.0250.050.10.90.950.9750.990.9950.999
3130.3409570.3508770.3586970.3752920.3959360.4293361.490712.108192.981424.714056.6666714.9071
4240.2715530.2892580.3015670.3255550.3533080.3958221.796722.603833.741355.994678.5324419.2387
5140.4512750.4818720.5012860.5362410.5731790.6243371.791632.312812.964664.088595.196179.00251
6250.4043040.4398660.4618140.5006230.5410230.5970811.936352.528283.267274.540055.7935910.0999
7360.3732820.4121450.4357590.4771170.5200690.5801652.035452.675763.474584.849796.2039210.855
8260.4953620.5401840.5661930.6102030.6545550.7150251.939232.402432.945013.814464.61367.09137
9370.4694720.5165240.5437430.5899120.6366740.7008132.002482.492993.06723.986914.832027.4518
10480.4496040.4984040.5266180.5745740.6233110.6903882.051942.563823.16284.121965.003187.73456
11380.5409560.5919630.6208430.6691810.7175260.7829971.963772.364652.815473.505894.113695.88409
12490.5231830.5757480.6055640.6555870.705750.7738181.999832.415152.882033.596864.226046.05843
135100.5085820.5624810.5931000.6445580.6962480.7664742.029862.45722.937493.672704.319746.20397
144100.5807150.635320.6659430.7168820.7674720.8353821.957162.309282.69423.265533.753775.11685
155110.5673440.6232250.6546160.7069030.7588950.8287381.980742.341842.736463.322083.822485.21933
166120.5559210.6129270.6449890.6984480.7516490.8231492.001132.369982.77303.371013.881935.30806
175120.615110.6720860.7038320.7563510.8081530.8771141.941472.256032.592693.080983.489214.59421
186130.6045160.6625770.6949650.7485860.8015050.8719731.958252.278962.622143.119813.535834.66184
197140.5952260.6542590.6872180.7418130.7957180.867511.973122.299272.648233.154223.577174.72183
206140.6451590.7037580.7362310.7896830.8420740.9113081.923542.20862.508712.936253.287694.21667
217150.6364660.6960050.7290220.7833940.8367030.9071561.936172.225742.530552.964733.321584.26483
228160.6286990.6890890.7225970.7777950.8319260.9034711.947572.241182.5550232.99043.352164.3083
237160.6717060.7314260.7643590.818320.8709140.9399711.905722.167092.438632.819953.129153.93126
248170.6643890.7249310.7583340.8130780.8664430.9365141.915632.180462.455562.841833.155033.96745
259180.6577590.7190530.7528830.808340.8624050.9333971.924692.192662.471022.861823.178674.00052
3010210.7112620.7729020.8066140.8614270.9143460.983101.879842.107792.339842.658642.911873.55044
5016340.8850940.9178360.9700731.019431.0820801.792971.952052.107732.312772.469332.84359

References

  1. Student. The probable error of a mean. Biometrika 1908, 6, 1–25. [Google Scholar] [CrossRef]
  2. Geary, R.C. Distribution of Student’s ratio for nonnormal samples. Suppl. J. R. Stat. Soc. 1936, 3, 178–184. [Google Scholar] [CrossRef]
  3. Darmois, G. Analyse générale des liaisons stochastiques: Étude particuière de l’analyse factorielle lineéaire. Rev. L’Institut Int. Stat./Rev. Int. Stat. Inst. 1953, 21, 2–8. [Google Scholar] [CrossRef]
  4. Skitovich, V.P. Linear forms of independent random variables and the normal distribution law. Izv. Akad. Nauk SSSR Ser. Mat. 1954, 18, 185–200, (English Translation Sel. Transl. Math. Stat. Probab. 1962, 2, 211–218). [Google Scholar]
  5. Koopman, B.O. On distributions admitting a sufficient statistic. Trans. Am. Math. Soc. 1936, 39, 399–409. [Google Scholar] [CrossRef]
  6. Darmois, G. Sur les lois de probabilités à estimation exhaustive. Comptes Rendus L’Acad. Sci. Paris 1935, 200, 1265–1266. [Google Scholar]
  7. Pitman, E. Sufficient statistics and intrinsic accuracy. Math. Proc. Camb. Philos. Soc. 1936, 32, 567–579. [Google Scholar] [CrossRef]
  8. Logan, B.F.; Mallows, C.L.; Rice, S.O.; Shepp, L.A. Limit distributions of self-normalized sums. Ann. Probab. 1973, 1, 788–809. [Google Scholar] [CrossRef]
  9. Peña, V.H.D.L.; Lai, T.L.; Shao, Q.M. Self-Normalized Processes. Limit Theory and Statistical Applications; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  10. Efron, B. Student’s t-test under symmetry conditions. J. Am. Stat. Assoc. 1969, 64, 1278–1302. [Google Scholar] [CrossRef]
  11. Hendriks, H.W.M.; Ijzerman-Boon, P.C.; Klaassen, C.A.J. Student’s t-statistic under unimodal densities. Austrian J. Stat. 2006, 35, 131–141. [Google Scholar] [CrossRef]
  12. Perlo, V. On the distribution of Student’s ratio for samples of three drawn from the rectangular distribution. Biometrika 1933, 25, 203–204. [Google Scholar] [CrossRef]
  13. Pexider, J.V. Notiz über Funkcionaltheoreme. Monatsh. Math. Phys. 1903, 14, 293–301. [Google Scholar] [CrossRef]
  14. David, H.A. Order Statistics; Wiley: New York, NY, USA, 1981. [Google Scholar]
  15. Satterthwaite, F.E. An approximate distribution of estimates of variance components. Biom. Bull. 1946, 2, 110–114. [Google Scholar] [CrossRef]
  16. van Zwet, W. Convex Transformations of Random Variables, 7th ed.; Mathematical Centre: Amsterdam, The Netherlands, 1964. [Google Scholar]
  17. van Zwet, W. Convex transformations: A new approach to skewness and kurtosis. Statist. Neerl. 1964, 18, 433–441. [Google Scholar] [CrossRef]
  18. Abramowitz, M.; Stegun, I.A. Handbook of Mathematical Functions: With Formulas, Graphs, and Mathematical Tables, 8th ed.; Dover: New York, NY, USA, 1972. [Google Scholar]
  19. Slutsky, E. Über stochastische Asymptoten und Grenzwerte. Metron 1964, 5, 3–89. [Google Scholar]
  20. Hampel, F.R. Contributions to the Theory of Robust Estimation. Ph.D. Thesis, University of California, Berkeley, CA, USA, 1986. [Google Scholar]
  21. Ronchetti, E. The main contributions of robust statistics to statistical science and a new challenge. Metron 2021, 79, 127–135. [Google Scholar] [CrossRef]
  22. Bhattacharya, S.; Kamper, F.; Beirlant, J. Outlier detection based on extreme value theory and applications. Scand. J. Stat. 2024, 50, 1466–1502. [Google Scholar] [CrossRef]
  23. Basu, D. On statistics independent of a complete sufficient statistic. Sankhyā Indian J. Stat. 1955, 15, 377–380. [Google Scholar] [CrossRef]
  24. Brilhante, M.F.; Kotz, S. Infinite divisibility of the spacings of a Kotz–Kozubowski–Podgórski generalized Laplace model. Stat. Probab. Lett. 2008, 78, 2433–2436. [Google Scholar] [CrossRef]
  25. Hotelling, H. The behavior of some standard statistical tests under nonstandard conditions. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability (Volume 1: Contributions to the Theory of Statistics); Neyman, J., Ed.; University of California Press: Berkeley, CA, USA, 1961; pp. 319–359. [Google Scholar]
  26. Lehmann, E.L. “Student” and small-sample theory. Stat. Sci. 1999, 14, 418–426. [Google Scholar] [CrossRef]
Figure 1. Probability density function of T 1 with symmetric parent distributions.
Figure 1. Probability density function of T 1 with symmetric parent distributions.
Symmetry 16 01297 g001
Figure 2. Right tail of the probability density function of T 1 for the symmetric parent distributions in Figure 1.
Figure 2. Right tail of the probability density function of T 1 for the symmetric parent distributions in Figure 1.
Symmetry 16 01297 g002
Figure 3. Approximate probability density function of T 2 with a Laplace ( 0 , 1 ) parent.
Figure 3. Approximate probability density function of T 2 with a Laplace ( 0 , 1 ) parent.
Symmetry 16 01297 g003
Figure 4. Probability density function of T 2 with a Uniform ( 1 , 1 ) parent: (a) Exact and approximate densities. (b) Zoom in of right tails.
Figure 4. Probability density function of T 2 with a Uniform ( 1 , 1 ) parent: (a) Exact and approximate densities. (b) Zoom in of right tails.
Symmetry 16 01297 g004
Figure 5. Probability density function of τ n 1 .
Figure 5. Probability density function of τ n 1 .
Symmetry 16 01297 g005
Figure 6. Probability density functions of τ 2 ; 1 , 3 and τ 3 ; 2 , 4 .
Figure 6. Probability density functions of τ 2 ; 1 , 3 and τ 3 ; 2 , 4 .
Symmetry 16 01297 g006
Figure 7. Probability density function of W n 1 1 , n 2 1 .
Figure 7. Probability density function of W n 1 1 , n 2 1 .
Symmetry 16 01297 g007
Table 1. ANOSp table.
Table 1. ANOSp table.
Sum of SpacingsdfMean of Sum of SpacingsF-Statistic
BSSp = i = 1 k n i X 1 : n i ( i ) X 1 : N 2 ( k 1 ) MBSSp = BSSp 2 ( k 1 ) F = MBSSp MWSSp = BSSp k 1 WSSp N k
WSSp = i = 1 k j = 1 n i X i j X 1 : n i ( i ) 2 ( N k ) MWSSp = WSSp 2 ( N k )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Brilhante, M.d.F.; Pestana, D.; Rocha, M.L. Symmetry, Asymmetry and Studentized Statistics. Symmetry 2024, 16, 1297. https://doi.org/10.3390/sym16101297

AMA Style

Brilhante MdF, Pestana D, Rocha ML. Symmetry, Asymmetry and Studentized Statistics. Symmetry. 2024; 16(10):1297. https://doi.org/10.3390/sym16101297

Chicago/Turabian Style

Brilhante, Maria de Fátima, Dinis Pestana, and Maria Luísa Rocha. 2024. "Symmetry, Asymmetry and Studentized Statistics" Symmetry 16, no. 10: 1297. https://doi.org/10.3390/sym16101297

APA Style

Brilhante, M. d. F., Pestana, D., & Rocha, M. L. (2024). Symmetry, Asymmetry and Studentized Statistics. Symmetry, 16(10), 1297. https://doi.org/10.3390/sym16101297

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop