Next Article in Journal / Special Issue
Skewness-Based Projection Pursuit as an Eigenvector Problem in Scale Mixtures of Skew-Normal Distributions
Previous Article in Journal
Initial Value Problems of Linear Equations with the Dzhrbashyan–Nersesyan Derivative in Banach Spaces
Previous Article in Special Issue
Distribution-Based Entropy Weighting Clustering of Skewed and Heavy Tailed Time Series
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula

Institute of Mathematics and Statistics, University of Tartu, Narva mnt 18, 51009 Tartu, Estonia
*
Author to whom correspondence should be addressed.
Symmetry 2021, 13(6), 1059; https://doi.org/10.3390/sym13061059
Submission received: 20 May 2021 / Revised: 7 June 2021 / Accepted: 8 June 2021 / Published: 11 June 2021
(This article belongs to the Special Issue Symmetry and Asymmetry in Multivariate Statistics and Data Science)

Abstract

:
Symmetric elliptical distributions have been intensively used in data modeling and robustness studies. The area of applications was considerably widened after transforming elliptical distributions into the skew elliptical ones that preserve several good properties of the corresponding symmetric distributions and increase possibilities of data modeling. We consider three-parameter p-variate skew t-distribution where p-vector μ is the location parameter, Σ : p × p is the positive definite scale parameter, p-vector α is the skewness or shape parameter, and the number of degrees of freedom ν is fixed. Special attention is paid to the two-parameter distribution when μ = 0 that is useful for construction of the skew t-copula. Expressions of the parameters are presented through the moments and parameter estimates are found by the method of moments. Asymptotic normality is established for the estimators of Σ and α . Convergence to the asymptotic distributions is examined in simulation experiments.

1. Introduction

The idea of constructing multivariate skewed distributions from symmetric elliptical distributions has attracted many statisticians over last 25 years. This generalization preserves several good properties of corresponding symmetric distributions and in the same time increases possibilities of data modeling. In 1996, Azzalini and Dalla Valle introduced multivariate skew normal distribution [1]. Their idea of transforming multivariate normal distribution into a skewed distribution has been carried over to several other distributions and many generalizations have been made since 1996 (see for instance, Theodossiou [2], Arellano-Valle and Genton [3]). From these developments multivariate skew t-distribution has become most frequently used [4]. This distribution became an important tool in many data modeling applications because of heavier tail area than the normal and skew normal distributions (see, for instance, Lee and McLachan [5], Marchenko [6], Parisi and Liseo [7], Fernandez and Steel [8], Jones [9]). A review of applications in finance and insurance mathematics is given in Adcock et al. [10]. Multivariate skew t-distribution has also been successfully used as a data model in environmental studies and modelling coastal floodings (Ghizzolini et al. [11], Thompson and Shen [12]). Several other applications of skew elliptical distributions can be found in the collective monograph edited by Genton [13] and in the book by Azzalini and Capitanio [14] which includes a comprehensive list of references on the topic. Another useful property that makes multivariate t-distribution and skew t-distribution attractive in financial and actuarial applications is the possibility to take tail dependence between marginals into account. Tail dependence of multivariate t- and skew t-distributions has been intensively studied in recent years, see Fung and Seneta [15] or Joe and Li [16], for example. In Kollo et al. [17] it is shown that the upper tail dependence coefficient of the skew t-distribution can be several times bigger than the corresponding value of multivariate t-distribution. Moments and inferential aspects related to the multivariate skew t-distribution have been also actively studied in Padoan [18], Galarza et al. [19], Zhou and He [20], Lee and McLachlan [5], DiCiccio and Monti [21], for instance. A construction based on the skew t-distribution—skew t-copula—has made this distribution even more attractive in applications where correlated differently distributed marginals are joined into a multivariate distribution. Demarta and McNeil introduced skew t-copula based on multivariate generalized hyperbolic distribution [22], Kollo and Pettere [23] defined skew t-copula on the basis of the multivariate skew t-distribution following Azzalini and Capitanio [4].
The paper is organized as follows. In Section 2, we introduce necessary notation and notions from matrix algebra. In Section 3, skew t-distribution is defined through skew normal and inverse χ -distribution and the first four central moments are found via the moments of skew normal and inverse χ -distribution. The asymptotic normality for parameter estimators in the two-parameter case is also established in this section. In Section 4, obtained analytical results are illustrated by simulations for different scenarios and the speed of convergence to asymptotic distribution is also addressed. In Section 5, multivariate skew t-copula is defined and some properties are discussed. Technical proofs are presented in Appendix A. A comparison of density contours of Gaussian, skew normal and skew t-copula is presented in Supplementary Materials.

2. Notation and Preliminaries

To derive asymptotic normal distribution for the parameter estimators of skew t-distribution, we use matrix technique. Necessary notions and notation are presented in what follows.
Let us first denote the moments of a random vector x in the following matrix representation:
m k ( x ) = E x x x x k times = E x x x x .
The corresponding central moment is
m ¯ k ( x ) = m k ( x E x ) .
We use notions vec -operator, commutation matrix, Kronecker product, matrix derivative and their properties. For deeper insight into this technique we refer to Magnus and Neudecker [24], or Kollo and von Rosen [25]. If A is a p × q matrix, A = ( a 1 , , a q ) , the p q column vector vec A is formed by stacking the columns a i under each other; that is
vec A = a 1 a q .
We denote a p × p identity matrix by I p and a p q × p q commutation matrix by K p , q . Commutation matrix K p , q is an orthogonal p q × p q partitioned matrix consisting of q × p blocks with the property
K p , q vec A = vec A
where A is a p × q -matrix. The Kronecker product A B of a p × q matrix A and an r × s matrix B is a p r × q s partitioned matrix consisting of the r × s blocks
A B = [ a i j B ] , i = 1 , , p ; j = 1 , , q .
The following properties of the notions above are repeatedly used. It is assumed that the dimensions of the matrices match with the operations.
1.
( A B ) ( C D ) = ( A C ) ( B D ) ;
2.
a b = a b = b a ;
3.
vec ( A B C ) = ( C A ) vec B ;
4.
K p , 1 = I p , K 1 , q = I q ;
5.
for A : p × q and B : r × s , A B = K p , r ( B A ) K s , q .
Further, when deriving asymptotic distributions of the estimators of Σ and α in Section 3, we use the matrix derivative where partial derivatives y k l x i j are ordered by the Kronecker product (see Magnus and Neudecker, [24])
d Y d X = d d vec X vec Y ,
with X : p × q and Y = Y ( X ) : r × s . It follows from the definition that
d Y d X = d vec Y d vec X = d vec Y d X = d Y d vec X .
The main properties of the matrix derivative are listed below.
(1)
d X d X = I p q ;
(2)
d ( Y + Z ) d X = d Y d X + d Z d X , Z = Z ( X ) : r × s ;
(3)
d A X B d X = B A ;
(4)
If Z = Z ( Y ) , Y = Y ( X ) then d Z d X = d Z d Y d Y d X ;
(5)
If W = Y Z , Z = Z ( X ) : s × t then
d W d X = ( Z I r ) d Y d X + ( I t Y ) d Z d X ;
(6)
If X : p × p then d X 1 d X = ( X ) 1 X 1 ;
(7)
If X R p × p then d | X | d X = | X | vec ( X 1 ) , where | · | denotes the determinant;
(8)
If W = Y Z , Z = Z ( X ) : m × n , then
d W d X = ( I s K r , n I m ) ( I r s vec Z ) d Y d X + ( vec Y I m n ) d Z d X .
(9)
If y = y ( X ) , Z = Z ( X ) : m × n , then
d ( y Z ) d X = vec Z d y d X + y d Z d X .

3. Multivariate Skew t ν -Distribution

3.1. Definition

Multivariate skew t-distribution has been defined in different ways, see Kotz and Nadarajah [26] (Chapter 5). Kshirsagar [27] was the first to introduce a noncentral multivariate t-distribution, but the density expression was complicated and the distribution was not used in applications. From more recent definitions we refer to Fang et al. [28], but as the marginals in their definition have different degrees of freedom, the distribution had problems with practical use. Gupta [29] defined the multivariate skew t-distribution in a similar way like Azzalini and Capitanio [4], but their definition had several advantages and has become widely used in applications (for example, it is possible to use explicit expression of the moment generating function of skew normal distribution for calculating moments of skew t-distribution). We follow the approach by Azzalini and Capitanio [4].
A p-vector x has multivariate skew t-distribution with ν degrees of freedom if it is the ratio of a p-variate skew normal vector and a chi-distributed random variable with ν degrees of freedom. We have slightly modified their definition of the skew normal vector and use the covariance matrix as a parameter instead of the correlation matrix. Let y be a p-variate skew normal vector with density
f ( y ) = 2 f N p ( 0 , Σ ) ( y ) Φ ( α y ) ,
where f N p ( 0 , Σ ) ( y ) is the density function (density) of the p-variate normal distribution N p ( 0 , Σ ) and Φ ( · ) is the cumulative distribution function (distribution function) of N ( 0 , 1 ) . Denote the distribution of vector y with density (1) by y S N p ( Σ , α ) , where Σ : p × p is the scale parameter and α R p is the shape or skewing parameter.
Definition 1. 
Let y be skew normally distributed with density (1) and let Z be χ-distributed with ν degrees of freedom, independent from y . Then
x = μ + ν y Z
is skew t ν -distributed, x S t ν ( μ , Σ , α ) .
The density function of x S t ν ( μ , Σ , α ) is of the form (a modification of Azzalini and Capitanio [4])
f x ( x ) = 2 f ν , p ( x ) F p + ν α ( x μ ) p + ν Q + ν ,
where Q = ( x μ ) Σ 1 ( x μ ) , F k is the distribution function of the univariate t-distribution with k degrees of freedom and f ν , p is the density function of the p-variate t-distribution with ν degrees of freedom and scale matrix Σ
f ν , p ( x ) = Γ ν + p 2 ( ν π ) p / 2 Γ ν 2 | Σ | 1 / 2 1 + ( x μ ) Σ 1 ( x μ ) ν ν + p 2 .

3.2. Moments

By Formula (2), a skew t-distributed x is constructed using independent skew normal random vector y and χ -distributed Z. Because of the independence of y and Z, the moments of x can be found as products of the moments of the skew normal and inverse χ -distribution. The k-th moment, central moment and cumulant of a skew-normal random vector y are denoted by m k ( y ) , m ¯ k ( y ) and c k ( y ) , respectively. Explicit expressions of the raw moments, central moments and cumulants of y can be obtained by differentiating the moment generating function
M ( t ) = 2 e 1 2 t Σ t Φ α Σ t 1 + α Σ α
and the cumulant generating function
C ( t ) = ln M ( t ) .
The first four cumulants for a skew-normal random vector y are (Kollo et al. [30])
c 1 ( y ) = E y = 2 π δ , c 2 ( y ) = D y = Σ 2 π δ δ ,
c 3 ( y ) = m ¯ 3 ( y ) = 2 π 4 π 1 δ δ δ ,
c 4 ( y ) = 8 π 1 3 π δ δ δ δ ,
where δ equals
δ = Σ α 1 + α Σ α .
From general relation between the fourth order cumulants and central moments we have (see, e.g., Kollo and von Rosen [25] (p. 187))
m ¯ 4 ( y ) = c 4 ( y ) + ( I p 2 + K p , p ) ( D y D y ) + vec ( D y ) vec ( D y ) .
After some algebra, we get the expression for α from (5):
α = Σ 1 δ 1 δ Σ 1 δ .
The quantity under square root in (7) influences the stability of parameter estimates and will be used later in a related discussion. Thus, we add a notation for it:
η = 1 δ Σ 1 δ .
Note also that each component δ i in vector δ = ( δ 1 , , δ p ) is directly connected to the corresponding marginal distribution, whereas this property does not hold for vector α = ( α 1 , , α p ) . For a more comprehensive discussion on parametrizations of skew-normal vector see Käärik et al. [31].
To derive the central moments of the multivariate skew t ν -distribution, we also need the moments of the skew normally distributed vector y
m 2 ( y ) = Σ ;
m 3 ( y ) = 2 π ( vec Σ δ + δ Σ + Σ δ δ δ δ ) .
Recall that a random variable V ν = 1 Z is inverse χ ν -distributed if its probability density function is (Lee [32] (p. 378))
f ( x ) = 1 2 ν / 2 1 Γ ν 2 x ν 1 exp 1 2 x 2 .
The moments of the inverse χ ν -distribution can be obtained by integration.
Lemma 1. 
The moments of the inverse χ ν -distribution are:
E V ν = Γ ν 1 2 2 Γ ν 2 , ν > 1 ; E V ν k = Γ ν k 2 2 k / 2 Γ ν 2 , ν > k .
Proof. 
By standard integration technique, we can derive the expectation:
E V ν = 0 x 1 2 ν / 2 1 Γ ν 2 x ν 1 exp 1 2 x 2 d x = 1 Γ ν 2 0 Γ ν 1 2 2 ν 1 2 1 2 Γ ν 1 2 x ν exp 1 2 x 2 d x = Γ ν 1 2 Γ ν 2 2 .
Following the same approach, the k-th order moment can be derived as follows
E V ν k = 0 x k 1 2 ν / 2 1 Γ ν 2 x ν 1 exp 1 2 x 2 d x = 2 ν 1 2 1 Γ ν 1 2 2 ν 2 1 Γ ν 2 0 1 2 ν 1 2 1 Γ ν 1 2 x ( ν 1 ) 1 x k 1 exp 1 2 x 2 d x = i = 0 k 1 E V ν i = Γ ν k 2 2 k 2 Γ ν 2 , ν > k .
The expressions of the second, third and fourth moment are special cases of this result:
E V ν 2 = 1 ν 2 , ν > 2 ;
E V ν 3 = E V ν ν 3 , ν > 3 ;
E V ν 4 = 1 ( ν 2 ) ( ν 4 ) , ν > 4 .
After some elementary algebra we get expressions of the corresponding central moments
m ¯ 3 ( V ν ) = E V ν 7 2 ν ( ν 3 ) ( ν 2 ) + 2 ( E V ν ) 2 ;
m ¯ 4 ( V ν ) = 1 ( ν 2 ) ( ν 4 ) + 2 ν 5 ( ν 2 ) ( ν 3 ) ( E V ν ) 2 3 ( E V ν ) 4 .
To keep the notation simpler, we denote V = V ν .
In order to derive the asymptotic normality for parameter estimators of S t ν , we need the first 4 central moments of the distribution. Let us start with general expressions for central moments m ¯ 3 ( x ) and m ¯ 4 ( x ) .
Lemma 2. 
Let x be a p-dimensional random vector with finite fourth order moments. Then the third and fourth central moment of x can be expressed as follows:
m ¯ 3 ( x ) = m 3 ( x ) ( I p 2 + K p , p ) ( m 2 ( x ) E x ) vec m 2 ( x ) E x + 2 E x E x E x , m ¯ 4 ( x ) = m 4 ( x ) C ( x ) ( I p 2 + K p , p ) ( I p 2 + K p , p ) ( C ( x ) ) 3 E x E x E x E x + ( I p 2 + K p , p ) A ( x ) + A ( x ) K p , p + B ( x ) + ( A ( x ) + B ( x ) ) ,
where
A ( x ) = E x m 2 ( x ) E x ; B ( x ) = E x vec m 2 ( x ) E x ; C ( x ) = m 3 ( x ) E x .
Proof of the lemma is presented in Appendix A.1. Applying this lemma to skew t-distributed x , the next result follows.
Lemma 3. 
Let x S t ν ( μ , Σ , α ) . Then its mean E x , covariance matrix D x , and central moments m ¯ 3 ( x ) and m ¯ 4 ( x ) can be expressed through the moments of inverse χ ν -distributed V and y S N p ( Σ , α ) in the following form:
E x = μ + ν E V E y ;
D x = ν m 2 ( V ) m 2 ( y ) E x E x = ν ( m 2 ( V ) D y + D V E y E y ) ;
m ¯ 3 ( x ) = ν 3 / 2 { E V 3 m ¯ 3 ( y ) 2 ( E V 3 ( E V ) 3 ) ( E y E y E y ) + ( E V 3 E V E V 2 ) [ ( I p 2 + K p , p ) ( m 2 ( y ) E y ) + vec m 2 ( y ) E y ] } ;
m ¯ 4 ( x ) = ν 2 { E V 4 m ¯ 4 ( y ) + ( E V 4 E V E V 3 ) ( C ( y ) ( I p 2 + K p , p ) + ( I p 2 + K p , p ) C ( y ) ) + 3 ( E V 4 ( E V ) 4 ) E y E y E y E y ( E V 4 E V 2 ( E V ) 2 ) [ ( I p 2 + K p , p ) A ( y ) + A ( y ) K p , p + B ( y ) + ( A ( y ) + B ( y ) ) ] } .
Proof is straightforward and follows from Lemma 2 after substituting in the required moments of skew normal distribution from Formulas (9) and (10) and the moments of inverse χ -distribution given in Lemma 1.

3.3. Point Estimates

Let us now focus on the case when μ = 0 and estimate the parameters of the two-parameter distribution S t ν ( Σ , α ) using the method of moments.
Let us have a sample of size n, n > p , from S t ν ( Σ , α ) with sample mean x ¯ and unbiased sample covariance matrix S . Now, the estimates for parameters Σ and α through the moments of V and y can be obtained from Expressions (11) and (12):
E x = Γ ν 1 2 Γ ν 2 ν π δ ,
D x = ν 1 ν 2 Σ Γ ν 1 2 Γ ν 2 2 1 π δ δ ,
where δ is defined in (5). From here we get the following estimates
Σ ^ = ν 2 ν S + x ¯ x ¯ ;
α ^ = ν ν 2 ( S + x ¯ x ¯ ) 1 x ¯ ν π Γ ν 1 2 Γ ν 2 2 ν ν 2 x ¯ ( S + x ¯ x ¯ ) 1 x ¯ .

3.4. Asymptotic Normality

As the estimators depend on both, the sample mean and the sample covariance matrix, we need their joint distribution to find the asymptotic distributions of the parameter estimators. In general, the following convergence in distribution takes place for x ¯ vec S if m ¯ 4 ( x ) exists (Parring [33]):
n x ¯ vec S E x ¯ vec S D N ( 0 , D a s ) ,
where
D a s = D x m ¯ 3 ( x ) m ¯ 3 ( x ) m ¯ 4 ( x ) vec ( D x ) vec ( D x ) .
For the asymptotic covariance matrix of the joint distribution of sample mean and covariance matrix from skew t-distribution, D a s S t , we get the expressions of necessary central moments from Equations (12)–(14). Derivation of asymptotic normality for estimators of parameters is based on Theorem 1 by Anderson [34] (pp. 132–133).
Theorem 1. 
Assume that for { x n }
n ( x n a ) D N p ( 0 , Σ )
and
x n P a ,
if n . Let the function g ( x ) : R p R q have continuous partial derivatives in a neighborhood of a . Then, if n
n ( g ( x n ) g ( a ) ) D N q ( 0 , ξ Σ ξ ) ,
where
ξ = d g ( x ) d x | x = a 0
is the matrix derivative.
The vector ( x ¯ , vec S ) satisfies assumptions of x n and we consider Σ ^ and α ^ as the functions of ( x ¯ , vec S ) and derive asymptotic normality for the estimators. For Σ ^ the following statement holds.
Theorem 2. 
Let us have a p-vector x S t ν ( Σ , α ) , where ν is fixed. Then, for the estimator Σ ^ in (15), the following convergence takes place, if n
n vec ( Σ ^ Σ ) D N p 2 ( 0 , D a s Σ ) ,
where
D a s Σ = d ( ν 2 ν ( S + x ¯ x ¯ ) ) d ( x ¯ , vec S ) E x ¯ vec S D a s S t d ( ν 2 ν ( S + x ¯ x ¯ ) ) d ( x ¯ , vec S ) E x ¯ vec S ,
the matrix D a s S t is the asymptotic covariance matrix of x ¯ vec S and the derivative
d ( ν 2 ν ( S + x ¯ x ¯ ) ) d ( x ¯ , vec S ) E x ¯ vec S = ν 2 ν ( I p 2 + K p , p ) I p 2 π δ I p 2 .
Proof. 
To find the matrix derivative, we can rely on the corresponding expression for the skew normal distribution, derived by Kollo et al. [30] (Proposition 2) as the expression of the derivative for skew t-distribution differs only by a constant.
For y S N p ( Σ , α ) , the derivative of Σ ^ is:
( I p 2 + K p , p ) I p 2 π δ I p 2 .
From here, the derivative of Σ ^ for x S t ν ( Σ , α ) is:
d Σ ^ d ( x ¯ , vec S ) E x ¯ vec S = ν 2 ν ( I p 2 + K p , p ) I p 2 π δ I p 2 .
This completes the proof. □
The proof of asymptotic normality of estimator for parameter α ,
n ( α ^ α ) D N p ( 0 , D a s α ) ,
is similar. In order to shorten the notation from (16), we define
M = ν π Γ ν 1 2 Γ ν 2 2 ν ν 2 x ¯ ( S + x ¯ x ¯ ) 1 x ¯ .
This allows us to express the estimator α ^ as
α ^ = Σ ^ 1 x ¯ M .
Theorem 3. 
Let us have a p-vector x S t ν ( Σ , α ) , where ν is fixed. Then, for the estimator α ^ in (18), the following convergence takes place, if n
n ( α ^ α ) D N p ( 0 , D a s α ) ,
where
D a s α = d α ^ d ( x ¯ , vec S ) E x ¯ vec S D a s S t d α ^ d ( x ¯ , vec S ) E x ¯ vec S ,
matrix D a s S t is the asymptotic covariance matrix and the derivative
d α ^ d ( x ¯ , vec S ) = d α ^ d x ¯ d α ^ d vec S ,
where
d α ^ d x ¯ = M 1 2 Σ ^ 1 1 ν 2 ν x ¯ Σ ^ 1 x ¯ I p + M 1 1 ν 2 ν x ¯ Σ ^ 1 x ¯ + M x ¯ x ¯ Σ ^ 1 .
and
d α ^ d vec S = ν 2 ν M 1 2 ( x ¯ Σ ^ 1 Σ ^ 1 ) + 1 2 M 1 ( Σ ^ 1 x ¯ x ¯ Σ ^ 1 x ¯ Σ ^ 1 ) .
The derivatives in (19) for the asymptotic covariance matrix D a s α are obtained when the estimators x ¯ and Σ ^ are replaced by their expected values.
Derivations in the proof of Theorem 3 are more technical than in the proof of Theorem 2 and therefore the proof is presented in Appendix A.2.

4. Simulations and Discussion

To illustrate theoretical results, we consider a bivariate skew t-distribution x S t ν ( Σ , α ) and study the convergence speed of the estimates using several parameter sets. First we fix two values of η in (8), since, by the construction, this quantity has effect on the stability of the moments-based estimate for α . Namely, values of η close to zero make the moment-based estimates unstable (see, e.g., Formula (7)). Taking this into account, we choose one “problematic” value of η , η 0.03 , and one “safe” value, η 0.3 . Different values for Σ and α are chosen to obtain given values of η . In both cases, two α values of different magnitude are considered: small value is set to ( 0.5 , 1 ) and moderate to ( 3 , 2 ) . Furthermore, the off-diagonal values in matrix parameter Σ are pre-set to achieve low or high absolute value of σ 12 / σ 11 σ 22 with both signs. Thus, altogether 16 sets of parameters are used, the exact values can be found in Appendix A.3, Table A1. The number of degrees of freedom ν is taken equal to 4 in all cases considered.
With all parameter sets, the data is simulated using sample sizes n from 100 to 10 6 (using sequence for the powers of 10 from 2 to 6 with 50 equally spaced points) and k = 3000 repetitions for each sample size. The mean absolute percentage deviation between the estimate and the theoretical asymptotic value (MAPD) is used as an indicator of convergence. Given k generated samples, MAPD is calculated as
MAPD ( θ ) = 1 k j = 1 k θ ^ j θ θ ,
where θ is the theoretical asymptotic value of estimate and θ ^ j is the parameter estimate from the jth sample.
It must be noted that depending on the sample size and the value of η there are occasions when it is not possible to find an eligible estimate for α since the sample value for η is negative. To achieve the planned 3000 estimates for parameter sets with η 0.03 and small sample sizes there is up to 2.75 times more samples needed altogether. For parameter sets with η 0.3 the number of failed estimates is considerably smaller. Figure A1 in Appendix A.4 shows the ratio of the number of failed samples to the number of planned estimates from our simulations.
Figure 1 shows resulting MAPD values for parameters depending on the logarithm of sample size. Red tones represent η 0.03 group and blue tones η 0.3 . Lighter tone is for smaller α and dark for moderate size α . Line types represent different sign and magnitude combinations of σ 12 / σ 11 σ 22 .
As one can see, the convergence for diagonal elements of Σ is fastest and shows no remarkable differences for different parameter sets. However, MAPD for σ ^ 11 and σ ^ 22 is slightly bigger for sample sizes between 100 to 1000 if η is small.
For the off-diagonal element of Σ , the convergence speed varies with different parameter sets. If the absolute value of σ 12 / σ 11 σ 22 is big, the pattern is actually similar to the convergence of diagonal elements, but for small values, the convergence is slower and MAPD can be close to or exceed value 1 even with sample sizes near 1000.
Most recognizable distinction is seen with the MAPD values of the estimates of α and it is due to the η value. If η is small ( η 0.03 ), the MAPD value retains almost constantly high level with sample sizes from 100 up to 10,000, then starts to decrease with larger samples, but is still bigger than the MAPD value for the group with η 0.3 with sample sizes near 10 6 . For η 0.3 , the convergence is faster and reaches smaller values.
Another aspect is the stability of estimates: with smaller η ( η 0.03 ), the estimates are also more unstable for sample sizes under 10 4 than the estimates corresponding to η 0.3 . Other variations (low and moderate α elements and different σ 12 / σ 11 σ 22 values) do not seem to have an effect on stability of the estimates.
Next we compare asymptotic covariance matrices of the estimators found in Theorems 2 and 3 with their estimates from simulated samples. Parameter sets 1 and 9 from Table A1 in Appendix A.3 were used. The former (set 1) corresponds to safe η 0.3 value and the latter (set 9) to problematic η 0.03 value and these parameter sets have the same value of α . For both cases α ^ and Σ ^ were calculated 1000 times from samples of size 500. These parameter estimates were used to estimate D a s α and D a s Σ . For the simulations, the number of degrees of freedom ν = 5 was used, as it is the minimal value for calculating the asymptotic covariance matrices.
For the parameter set 9 with η 0.03 , the theoretical covariance matrix for n α ^ is
D a s α = 280.18 503.84 503.84 993.65
and its estimate
D a s α ^ = 230.93 372.54 372.54 756.02 .
As the theoretical expected value is n α ( 11.2 , 22.4 ) we can say that the variance of estimates is high in this case, but the theoretical and estimated covariance matrices are in agreement, although the theoretical values are higher. For the estimator n vec Σ ^ , the results are similar. In this case, the theoretical covariance matrix is
D a s Σ = 200.00 460.00 460.00 1213.50 460.00 1291.25 1291.25 3864.00 460.00 1291.25 1291.25 3864.00 1213.50 3864.00 3864.00 14,112.00
and the corresponding estimate:
D a s Σ ^ = 182.98 425.81 425.81 1044.01 425.81 1191.52 1191.52 3293.59 425.81 1191.52 1191.52 3293.59 1044.01 3293.59 3293.59 11,819.03 .
For the parameter set with η 0.3 , the same applies: theoretical and estimated covariance matrices are close while theoretical values are higher. Here the theoretical covariance matrix for n α ^ is
D a s α = 12.11 13.01 13.01 21.38
and the estimate
D a s α ^ = 9.59 9.23 9.23 14.71 .
As these two parameter sets have the same value of α , we can see that for the second case the estimates vary much less. For η 0.3 , the theoretical covariance matrix for n vec Σ ^ is
D a s Σ = 32.00 35.20 35.20 45.04 35.20 48.20 48.20 70.40 35.20 48.20 48.20 70.40 45.04 70.40 70.40 128.00
and its estimate
D a s Σ ^ = 29.14 30.29 30.29 36.03 30.29 39.30 39.30 54.62 30.29 39.30 39.30 54.62 36.03 54.62 54.62 97.27 .
Figure 2 presents scatterplots for estimates of α 1 and α 2 with theoretical 95% confidence ellipses. On the left panel ( η 0.03 ), the estimates of α are more scattered as compared to the right panel, where η 0.3 , as expected. In addition, the placing of points does not correspond to an ellipse on the left panel, indicating the asymptotics does not work with sample size n = 500 for distribution with η 0.03 . At the same time the result for η 0.3 can be considered good.
Similar plots for the elements of Σ ^ are presented in Figure 3. Here the estimates and ellipses again fit well for η 0.3 case. For η 0.03 the fit is also better compared to estimates of α in the previous figure.

5. Skew t-Copula

5.1. Definition

The skew t-copula is introduced in Kollo and Pettere [23] (Chapter 15). When the location parameter μ = 0 , the copula has the form
C p , ν ( u 1 , , u p ; Σ , α ) = F p , ν ( F 1 , ν 1 ( u 1 ; σ 11 , α 1 * ) , , F 1 , ν 1 ( u p ; σ p p , α p * ) ; Σ , α ) ,
where F 1 , ν 1 ( u i ; σ i i , α i * ) denotes the inverse of the univariate skew t 1 , ν -distribution function and F p , ν is the distribution function of the p-variate skew t p , ν -distribution. The arguments σ i i are the main diagonal elements of the covariance matrix Σ and α i * can be derived by the following simplification of (7):
α i * = σ i i 1 δ i 1 σ i i 1 δ i 2 , i = 1 , , p .
Note that α i * -s are not equal to the components α i of the parameter vector α of the multivariate distribution. The corresponding copula density is
c p , ν ( u ; Σ , α ) = f p , ν ( F 1 , ν 1 ( u 1 ; σ 11 , α 1 * ) , , F 1 , ν 1 ( u p ; σ p p , α p * ) ) ; Σ , α i = 1 p f 1 , ν F 1 , ν 1 ( u i ; σ i i , α i * ) ; σ i i , α i * ,
where f p , ν ( u ; Σ , α ) is the density function of S t p , ν and the inverse functions F 1 , ν 1 ( u i ; σ i i , α i * ) are as before.
In the standard case when μ = 0 and the covariance matrix Σ is equal to the correlation matrix R , the density of the skew t-copula is of the form
c p , ν ( u ; R , α ) = f p , ν ( F 1 , ν 1 ( u 1 ; α 1 * ) , , F 1 , ν 1 ( u p ; α p * ) ) ; R , α i = 1 p f 1 , ν F 1 , ν 1 ( u i ; α i * ) ; α i * ,
where F 1 , ν ( u i , α i * ) , i = 1 , , p are the distribution functions of the univariate standard skew t 1 , ν -distribution. Density (21) is the most often used version of the skew t-copula density in data analysis.
As the numerator of the copula density consists of the density of multivariate skew t-distribution, the resulting copula is distribution function of a skewed multivariate distribution. When we want to apply skew t-copula in data analysis to join (possibly) differently distributed marginals into a multivariate distribution, we have to take into account the correlations between marginals. Pearson correlation matrix can easily be expressed through the parameter Σ . Unfortunately Pearson correlations are not invariant under copula transformations (inverse distribution functions). We have to switch over to rank correlations. Fortunately for continuous elliptical distributions, Kendall’s τ can be expressed through the linear correlation (see Lindskog et al. [35])
τ ( X i , X j ) = 2 π arcsin ( r i j ) ,
so we can switch over from linear correlations to rank correlations and vice versa.

5.2. Examples of Bivariate Skew t-Copula Densities

In the following we present some examples of two-dimensional copula densities to compare Gaussian, skew-normal and skew t-copulas using contour plots. First we present the formulas for the before mentioned copulas in bivariate case. The bivariate Gaussian copula has one parameter ρ and using the general formula for multivariate Gaussian copula from Cherubini et al. [36] (p. 148) we can express bivariate density as
c N ( u 1 , u 2 ; ρ ) = 1 1 ρ 2 exp ρ 2 ( z 1 2 + z 2 2 ) 2 ρ z 1 z 2 2 ( 1 ρ 2 ) ,
where z i = Φ 1 ( u i ) and u i [ 0 , 1 ] , i = 1 , 2 . Skew-normal copula has in addition to ρ parameter vector α and following Käärik et al. [37] its density in bivaraite case is of form
c S N ( u 1 , u 2 ; ρ , α ) = 1 2 1 ρ 2 exp ρ 2 ( w 1 2 + w 2 2 ) 2 ρ w 1 w 2 2 ( 1 ρ 2 ) Φ ( α ( w 1 , w 2 ) ) Φ ( α 1 * w 1 ) Φ ( α 2 * w 2 ) ,
where α i * are given by (20) and w i = F 1 ( u i ; α i * ) denotes inverse of the univariate skew-normal distribution function with parameter α i * , as u i [ 0 , 1 ] , i = 1 , 2 . The bivariate skew t-copula has parameters ρ , α and ν and using the general Formula (21) we get the bivariate density as
c 2 , ν ( u 1 , u 2 ; ρ ; α ) = ν 4 1 ρ 2 Γ ν 2 Γ ν + 1 2 2 1 + Q ( 1 ρ 2 ) ν ν + 2 2 i = 1 2 ( 1 + q i 2 ν ) ν + 1 2 F 2 + ν ( α ( q 1 , q 2 ) 2 + ν Q + ν ) i = 1 2 F 1 + ν ( α i * q i 1 + ν q i 2 + ν ) ,
where q i = F 1 , ν 1 ( u i ; α i * ) , u i [ 0 , 1 ] , i = 1 , 2 , denotes inverse of the univariate skew t distribution function with parameter α i * , Q = q 1 2 2 q 1 q 2 ρ + q 2 2 and F k is the distribution function of the univariate t k -distribution.
For a first example, we consider here the case were parameter ρ is zero. Gaussian copula reduces to the product copula when ρ = 0 . If the corresponding parameter for skew-normal or skew t-copula equals zero, the resulting copula is not a product copula, this can be seen in Figure 4.
In the figures, contours on the left correspond to skew-normal and on the right to skew t-copulas with different α values (values are given in the header of the figure, note that the same number of degrees of freedom ν = 4 is used for all examples here). Contours are given with step size of 0.2 starting from 0 and ending with 3. Dark blue, yellow and dark red colors represent the lowest, middle and highest contour levels for chosen scale.
If ρ = 0 and both components of α are positive, the densities of skew-normal and skew t-copulas tend to show negative dependence that grows stronger if α i are bigger (see first and second row of Figure 4). If one α i is negative, the copula represents positive dependence, but the direction of dependence also depends on the magnitude of α i , i.e., the direction is not so clear for α i close to zero. When comparing contours in the second and third row we see that adding negative sign to the second component of α results in rotation of the density by 90o to the right. Rotation to the left would be the result if first element of α is negative and the other positive. We note that the two skew copulas give the product copula if in addition to ρ = 0 one element in the α vector is also zero.
As a second example, we consider the case when ρ 0 . For Gaussian copula, the parameter ρ gives the direction and strength of dependence between marginals. For the two skewed copulas, the additional parameter α adds complexity to this relation. We can conclude from the Figures S1–S4. that if both α i are positive and ρ is negative as given here, the dependence for skewed copulas can be stronger as compared to Gaussian copula. For positive ρ , the opposite may happen: the skew-normal and skew t represent weaker dependence as compared to Gaussian copula or the dependence changes direction, see Figure S2 for ρ = 0.2 , for example. The effect of parameter α is more notable if its elements have high absolute values.

6. Conclusions

Our main theoretical results, presented in Theorems 2 and 3, assure the asymptotic normality for the moment based parameter estimators of the two-parameter multivariate skew t-distribution. We can make the following observations based on the simulations used to illustrate the analytical results. First, the behaviour of the estimates of α depends on the value of η : if the value of η is close to zero (we used value 0.03 ), the estimates of α are unstable and the convergence is really slow, MAPD values are greater than 0.5 with sample sizes up to 10 4 . For parameter sets with η value 0.3 , the results are more stable and, for the aforementioned sample size, the MAPD values were already less than 0.1 . Second, it occurs that the combination of small sample size and low η value can result in ineligible estimates of α . As a third point, we note that the estimates of Σ are stable and the convergence is faster than the convergence of the estimates of α . Furthermore, the value of η does not affect the estimates of Σ . As the two considered values of η were chosen arbitrarily, its effect needs further investigation. Finally we also define skew t-copula, a powerful and flexible tool that can be used to model dependent data with differently distributed marginals in fields like finance, actuarial science and more.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/sym13061059/s1, Figure S1: Gaussian copula and examples of skew-normal and skew t-copulas when the components of α have small values, Figure S2: Gaussian copula and examples of skew-normal and skew t-copulas when the components of α have relatively high values, Figure S3: Gaussian copula and examples of skew-normal and skew t-copulas when the components of α have different magnitude, Figure S4: Gaussian copula and examples of skew-normal and skew t-copulas when components of α are equal.

Author Contributions

Conceptualization, T.K. and M.K.; methodology, T.K., M.K. and A.S.; simulations and data analysis, A.S.; writing—original draft preparation, T.K.; writing—review and editing, T.K., M.K. and A.S.; visualization, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Estonian Research Council (grant PRG 1197).

Acknowledgments

The authors are thankful to the editor and three anonymous referees for their constructive feedback that certainly improved the quality of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Proof of Lemma 2

Proof. 
We have to express the third and fourth central moments through raw moments.
m ¯ 3 ( x ) = E [ ( x E x ) ( x E x ) ( x E x ) ] = E [ ( x x E x x x E x + E x E x ) ( x E x ) ] = E [ x x x x x E x E x x x + E x x E x x E x x + x E x E x + E x E x x E x E x E x ] = E ( x x x ) E ( x x ) E x E x E ( x x ) + E x E x E x E ( x E x x ) + E x E x E x = m 3 ( x ) m 2 ( x ) E x E x m 2 ( x ) + 2 E x E x E x E x E ( x x ) = m 3 ( x ) m 2 ( x ) E x E x m 2 ( x ) + 2 E x E x E x E x vec m 2 ( x ) .
Finally
m ¯ 3 ( x ) = m 3 ( x ) ( I p 2 + K p , p ) ( m 2 ( x ) E x ) vec m 2 ( x ) E x + 2 E x E x E x .
Derivation of the fourth central moment is longer. We can use the intermediate expression from derivation of m ¯ 3 ( x ) after the third equality sign above
m ¯ 4 ( x ) = E [ ( x E x ) ( x E x ) ( x E x ) ( x E x ) ] = E { [ x x x x x E x E x x x + E x x E x x E x x + x E x E x + E x E x x E x E x E x ] x } E { [ x x x x x E x E x x x + E x x E x x E x x + x E x E x + E x E x x E x E x E x ] E x } = m 4 ( x ) m ¯ 3 ( x ) E x E ( x x E x x ) E ( E x x x x ) + E ( E x x E x x ) E ( x E x x x ) + E ( x E x E x x ) + E x E x m 2 ( x ) E x E x E x E x .
The expectations in the last equality can be presented through moments when we use the property of the Kronecker product for vectors a b = a b = b a
E ( x x E x x ) = E ( x x E x x ) = E ( x x x E x ) = m 3 ( x ) E x ; E ( E x x x x ) = E x m 3 ( x ) ; E ( E x x E x x ) = E ( E x x x E x ) = E x vec m 2 ( x ) E x ; E ( x E x x x ) = E ( x E x x x ) = E ( E x x x x ) = E x m 3 ( x ) ; E ( x E x E x x ) = E ( x E x E x x ) = E ( E x x x E x ) = E x m 2 ( x ) E x .
After plugging in the expression of m ¯ 3 ( x ) we have
m ¯ 4 ( x ) = m 4 ( x ) m 3 ( x ) E x + m 2 ( x ) E x E x + E x m 2 ( x ) E x + E x vec m 2 ( x ) E x 3 E x E x E x E x m 3 ( x ) E x E x m 3 ( x ) + E x vec m 2 ( x ) E x E x m 3 ( x ) + E x m 2 ( x ) E x + E x E x m 2 ( x ) .
Applying notation in Lemma 2
A ( x ) = E x m 2 ( x ) E x ; B ( x ) = E x vec m 2 ( x ) E x ; C ( x ) = m 3 ( x ) E x
we get the statement after using property 5 of the Kronecker product
m ¯ 4 ( x ) = m 4 ( x ) C ( x ) ( I p 2 + K p , p ) ( I p 2 + K p , p ) ( C ( x ) ) 3 E x E x E x E x + A ( x ) + B ( x ) + ( A ( x ) + B ( x ) ) + m 2 ( x ) E x E x + E x E x m 2 ( x ) .

Appendix A.2. Proof of Theorem 3

Proof. 
Let us start from Expression (18) for α ^ :
α ^ = Σ ^ 1 x ¯ M ,
where M is given in (17), which, taking (15) into account, can also be expressed as follows:
M = ν π Γ ν 1 2 Γ ν 2 2 x ¯ Σ ^ 1 x ¯ .
Let us now derive the needed derivatives step by step.
d α ^ d x ¯ = d ( Σ ^ 1 x ¯ M 1 2 ) d x ¯ = M 1 2 d ( Σ ^ 1 x ¯ ) d x ¯ + Σ ^ 1 x ¯ d M 1 2 d x ¯ .
Let us find the two derivatives in (A1).
(a)
The first derivative can be expressed as
d ( Σ ^ 1 x ¯ ) d x ¯ = Σ ^ 1 + ( x ¯ I p ) d Σ ^ 1 d x ¯ .
Now the derivative d Σ ^ 1 d x ¯ can be found as
d Σ ^ 1 d x ¯ = d Σ ^ 1 d Σ ^ d Σ ^ d x ¯ = ( Σ ^ 1 Σ ^ 1 ) d ( ν 2 ν ( S + x ¯ x ¯ ) ) d x ¯ = ν 2 ν Σ ^ 1 Σ ^ 1 x ¯ I p + I p x ¯ = ν 2 ν Σ ^ 1 x ¯ Σ ^ 1 + Σ ^ 1 Σ ^ 1 x ¯ .
The Equations (A2) and (A3) imply that
d ( Σ ^ 1 x ¯ ) d x ¯ = Σ ^ 1 ( x ¯ I p ) ν 2 ν Σ ^ 1 x ¯ Σ ^ 1 + Σ ^ 1 Σ ^ 1 x ¯ = Σ ^ 1 ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 + x ¯ Σ ^ 1 Σ ^ 1 x ¯ = Σ ^ 1 ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 + Σ ^ 1 x ¯ x ¯ Σ ^ 1 .
(b)
The second derivative in (A1) is
d M 1 2 d x ¯ = d M 1 2 d M d M d x ¯ = 1 2 M 3 2 d ( x ¯ Σ ^ 1 x ¯ ) x ¯ = 1 2 M 3 2 x ¯ Σ ^ 1 + x ¯ d ( Σ ^ 1 x ¯ ) d x ¯ = 1 2 M 3 2 x ¯ Σ ^ 1 + x ¯ Σ ^ 1 x ¯ ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 + Σ ^ 1 x ¯ x ¯ Σ ^ 1 = 1 2 M 3 2 2 x ¯ Σ ^ 1 ν 2 ν x ¯ Σ ^ 1 x ¯ x ¯ Σ ^ 1 + x ¯ Σ ^ 1 x ¯ x ¯ Σ ^ 1 = M 3 2 1 ν 2 ν x ¯ Σ ^ 1 x ¯ x ¯ Σ ^ 1 .
Now, applying the results of parts (a) and (b) into Equation (A1), yields
d α ^ d x ¯ = M 1 2 Σ ^ 1 ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 + Σ ^ 1 x ¯ x ¯ Σ ^ 1 + Σ ^ 1 x ¯ M 3 2 1 ν 2 ν x ¯ Σ ^ 1 x ¯ x ¯ Σ ^ 1 = M 1 2 [ Σ ^ 1 ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 + Σ ^ 1 x ¯ x ¯ Σ ^ 1 + M 1 1 ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 x ¯ x ¯ Σ ^ 1 ] = M 1 2 [ 1 ν 2 ν x ¯ Σ ^ 1 x ¯ Σ ^ 1 + M 1 1 ν 2 ν x ¯ Σ ^ 1 x ¯ + M Σ ^ 1 x ¯ x ¯ Σ ^ 1 ] .
Taking the derivative of α ^ by vec S is similar, but simpler. First of all, following Equality (A1), we can write
d α ^ d vec S = M 1 2 d ( Σ ^ 1 x ¯ ) d vec S + Σ ^ 1 x ¯ d M 1 2 d vec S .
Let us find the two derivatives in the last sum.
(a)
The first derivative can be expressed as
d ( Σ ^ 1 x ¯ ) d vec S = ( x ¯ I p ) d Σ ^ 1 d Σ ^ d Σ ^ d vec S = ν 2 ν ( x ¯ Σ ^ 1 Σ ^ 1 ) ,
(b)
The second derivative is
d M 1 2 d vec S = d M 1 2 d M d M d vec S = 1 2 M 3 2 d ( x ¯ Σ ^ 1 x ¯ ) d vec S = 1 2 M 3 2 ( x ¯ x ¯ ) ( Σ ^ 1 Σ ^ 1 ) ν 2 ν = ν 2 2 ν M 3 2 ( x ¯ Σ ^ 1 x ¯ Σ ^ 1 ) .
The full expression of the derivative d α ^ d vec S is
d α ^ d vec S = ν 2 ν M 1 2 ( x ¯ Σ ^ 1 Σ ^ 1 ) + 1 2 M 1 ( Σ ^ 1 x ¯ x ¯ Σ ^ 1 x ¯ Σ ^ 1 ) .

Appendix A.3. Parameter Sets

Table A1. Parameter sets for simulations.
Table A1. Parameter sets for simulations.
η α 1 α 2 σ 11 σ 22 σ 12 σ 12 / σ 11 σ 22
Set 10.30300.512.04.00−2.20−0.78
Set 20.30300.511.02.20−0.15−0.10
Set 30.29990.511.51.000.960.78
Set 40.30300.513.01.350.200.10
Set 50.30493.020.61.50−0.76−0.80
Set 60.30493.020.20.30−0.06−0.24
Set 70.30673.020.10.100.080.80
Set 80.30123.020.20.100.010.07
Set 90.03050.515.042.00−11.50−0.79
Set 100.03010.5128.028.00−2.80−0.10
Set 110.03000.5116.016.0012.300.77
Set 120.02980.5115.025.003.800.20
Set 130.03013.027.022.00−9.90−0.80
Set 140.03053.023.03.00−0.60−0.20
Set 150.03013.021.12.001.190.80
Set 160.03023.022.22.000.360.17

Appendix A.4. Samples with Failed Estimates

Figure A1. Ratio of the number of failed samples to the number of planned estimates from our simulations for all parameter sets.
Figure A1. Ratio of the number of failed samples to the number of planned estimates from our simulations for all parameter sets.
Symmetry 13 01059 g0a1

References

  1. Azzalini, A.; Valle, A.D. The multivariate skew-normal distribution. Biometrika 1996, 83, 715–726. [Google Scholar] [CrossRef]
  2. Theodossiou, P. Financial data and the skewed generalized t distribution. Manag. Sci. 1998, 44, 1650–1661. [Google Scholar] [CrossRef]
  3. Arellano-Valle, R.B.; Genton, M.G. Multivariate unified skew-elliptical distributions. Chil. J. Stat. 2010, 1, 17–31. [Google Scholar]
  4. Azzalini, A.; Capitanio, A. Distributions generated by perturbation of symmetry with emphasis on a multivariate skew t-distribution. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2003, 65, 367–389. [Google Scholar] [CrossRef]
  5. Lee, S.X.; McLachlan, G.J. Finite mixtures of multivariate skew t-distributions: Some recent and new results. Stat. Comput. 2014, 24, 181–202. [Google Scholar] [CrossRef]
  6. Marchenko, Y.V. Multivariate Skew-T Distributions in Economeetrics and Environmetrics. Ph.D. Thesis, Texas A&M University, College Station, TX, USA, 2010; p. 137. [Google Scholar]
  7. Parisi, A.; Liseo, B. Objective Bayesian analysis for the multivariate skew-t model. Stat. Methods Appl. 2018, 27, 277–295. [Google Scholar] [CrossRef] [Green Version]
  8. Fernandez, C.; Steel, M. Multivariate Student-t regression models: Pitfalls and inference. Biometrika 1999, 86, 153–167. [Google Scholar] [CrossRef]
  9. Jones, M.C. On families of distributions with shape parameters. Int. Stat. Rev. 2015, 83, 175–192. [Google Scholar] [CrossRef]
  10. Adcock, C.; Eling, M.; Loperfido, N. Skewed distributions in finance and actuarial science: A review. Eur. J. Financ. 2015, 21, 1253–1281. [Google Scholar] [CrossRef]
  11. Ghizzoni, T.; Roth, G.; Rudari, R. Multivariate skew-t approach to the design of accumulation risk scenarios for the flooding hazard. Adv. Water Resour. 2010, 33, 1243–1255. [Google Scholar] [CrossRef]
  12. Thompson, K.; Shen, Y. Coastal flooding and the multivariate skew-t distribution. In Skew-Elliptical Distributions and Their Applications: A Journey beyond Normality; Genton, M.G., Ed.; Chapman & Hall/CRC: Boca Raton, FL, USA, 2004; Chapter 14; pp. 243–258. [Google Scholar] [CrossRef]
  13. Genton, M.G. (Ed.) Skew-Elliptical Distributions and Their Applications. A Journey beyond Normality; Chapman & Hall/CRC: Boca Raton, FL, USA, 2004. [Google Scholar] [CrossRef]
  14. Azzalini, A.; Capitanio, A. The Skew-Normal and Related Families; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar] [CrossRef]
  15. Fung, T.; Seneta, E. Tail dependence for two skew t distributions. Stat. Probab. Lett. 2010, 80, 784–791. [Google Scholar] [CrossRef]
  16. Joe, H.; Li, H. Tail densities of skew-elliptical distributions. J. Multivar. Anal. 2019, 171, 421–435. [Google Scholar] [CrossRef] [Green Version]
  17. Kollo, T.; Pettere, G.; Valge, M. Tail dependence of skew t-copulas. Commun. Stat. Simul. Comput. 2017, 46, 1024–1034. [Google Scholar] [CrossRef]
  18. Padoan, S.A. Multivariate extreme models based on underlying skew-t and skew-normal distributions. J. Multivar. Anal. 2011, 102, 977–991. [Google Scholar] [CrossRef] [Green Version]
  19. Galarza, C.E.; Matos, L.A.; Lachos, V.H. Moments of the doubly truncated selection elliptical distributions with emphasis on the unified multivariate skew-t distribution. arXiv 2020, arXiv:2007.14980. [Google Scholar]
  20. Zhou, T.; He, X. Three-step estimation in linear mixed models with skew-t distributions. J. Stat. Plan. Inference 2008, 138, 1542–1555. [Google Scholar] [CrossRef]
  21. DiCiccio, T.J.; Monti, A.C. Inferential aspects of the skew exponential power distribution. J. Am. Stat. Assoc. 2004, 99, 439–450. [Google Scholar] [CrossRef]
  22. Demarta, S.; McNeil, A.J. The t copula and related copulas. Int. Stat. Rev. 2005, 73, 111–129. [Google Scholar] [CrossRef]
  23. Kollo, T.; Pettere, G. Parameter estimation and application of the multivariate skew t-copula. In Copula Theory and Its Applications; Jaworski, P., Durante, F., Härdle, W.K., Rychlik, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 289–298. [Google Scholar] [CrossRef]
  24. Magnus, J.R.; Neudecker, H. Matrix Differential Calculus with Applications in Statistics and Econometrics, 2nd ed.; Wiley: Chichester, UK, 1999. [Google Scholar]
  25. Kollo, T.; von Rosen, D. Advanced Multivariate Statistics with Matrices; Springer: Dordrecht, The Netherlands, 2010. [Google Scholar] [CrossRef]
  26. Kotz, S.; Nadarajah, S. Multivariate t Distributions and Their Applications; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar] [CrossRef]
  27. Kshirsagar, A.M. Some extensions of the multivariate t-distribution and the multivariate generalization of the distribution of the regression coefficient. Math. Proc. Camb. Philos. Soc. 1961, 57, 80–85. [Google Scholar] [CrossRef]
  28. Fang, H.B.; Fang, K.T.; Kotz, S. The meta-elliptical distributions with given marginals. J. Multivar. Anal. 2002, 82, 1–16. [Google Scholar] [CrossRef] [Green Version]
  29. Gupta, A.K. Multivariate skew t-distribution. Statistics 2003, 37, 359–363. [Google Scholar] [CrossRef]
  30. Kollo, T.; Käärik, M.; Selart, A. Asymptotic normality of estimators for parameters of a multivariate skew-normal distribution. Commun. Stat. Theory Methods 2018, 47, 3640–3655. [Google Scholar] [CrossRef]
  31. Käärik, M.; Selart, A.; Käärik, E. On Parametrization of multivariate skew-normal distribution. Commun. Stat. Theory Methods 2015, 44, 1869–1885. [Google Scholar] [CrossRef]
  32. Lee, P.M. Bayesian Statistics: An Introduction, 4th ed.; Wiley: Chichester, UK, 2012. [Google Scholar]
  33. Parring, A.M. Computation of asymptotic characteristics of sample functions. Acta Comment. Univ. Tartu. Math. 1979, 492, 86–90. (In Russian) [Google Scholar]
  34. Anderson, T.W. An Introduction to Multivariate Statistical Analysis, 3rd ed.; Wiley: New York, NY, USA, 2003. [Google Scholar]
  35. Lindskog, F.; McNeil, A.; Schmock, U. Kendall’s tau for elliptical distributions. In Credit Risk; Bol, G., Nakhaeizadeh, G., Rachev, S.T., Ridder, T., Vollmer, K.H., Eds.; Physica-Verlag HD: Heidelberg, Germany, 2003; pp. 149–156. [Google Scholar] [CrossRef]
  36. Cherubini, U.; Luciano, E.; Vecchiato, W. Copula Methods in Finance; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar] [CrossRef]
  37. Käärik, M.; Selart, A.; Käärik, E. The use of copulas to model conditional expectation for multivariate data. In ISI World Statistics Congress Proceedings. Bulletin of the ISI 2011 58th WSC, Dublin; ISI: The Hague, The Netherlands, 2011; pp. 5533–5538. [Google Scholar]
Figure 1. Convergence speed of parameter estimates.
Figure 1. Convergence speed of parameter estimates.
Symmetry 13 01059 g001
Figure 2. Comparison of estimates and theoretical confidence region: α .
Figure 2. Comparison of estimates and theoretical confidence region: α .
Symmetry 13 01059 g002
Figure 3. Comparison of estimates and theoretical confidence region: Σ .
Figure 3. Comparison of estimates and theoretical confidence region: Σ .
Symmetry 13 01059 g003
Figure 4. Examples of skew-normal and skew t-copulas when ρ = 0 .
Figure 4. Examples of skew-normal and skew t-copulas when ρ = 0 .
Symmetry 13 01059 g004
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kollo, T.; Käärik, M.; Selart, A. Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry 2021, 13, 1059. https://doi.org/10.3390/sym13061059

AMA Style

Kollo T, Käärik M, Selart A. Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry. 2021; 13(6):1059. https://doi.org/10.3390/sym13061059

Chicago/Turabian Style

Kollo, Tõnu, Meelis Käärik, and Anne Selart. 2021. "Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula" Symmetry 13, no. 6: 1059. https://doi.org/10.3390/sym13061059

APA Style

Kollo, T., Käärik, M., & Selart, A. (2021). Multivariate Skew t-Distribution: Asymptotics for Parameter Estimators and Extension to Skew t-Copula. Symmetry, 13(6), 1059. https://doi.org/10.3390/sym13061059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop