Next Article in Journal
Sustaining Economic Growth in Sub-Saharan Africa: Do FDI Inflows and External Debt Count?
Next Article in Special Issue
Multiscale Stochastic Volatility Model with Heavy Tails and Leverage Effects
Previous Article in Journal
Impact Factors on Portuguese Hotels’ Liquidity
Previous Article in Special Issue
Application of Discriminant Analysis for Avoiding the Risk of Quarry Operation Failure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Analysis of Intraday Stochastic Volatility Models of High-Frequency Stock Returns with Skew Heavy-Tailed Errors

1
Centre for Finance, Technology and Economics at Keio (FinTEK), Keio University, Tokyo 108-8345, Japan
2
Faculty of Economics, Keio University, Tokyo 108-8345, Japan
*
Author to whom correspondence should be addressed.
J. Risk Financial Manag. 2021, 14(4), 145; https://doi.org/10.3390/jrfm14040145
Submission received: 26 January 2021 / Revised: 15 March 2021 / Accepted: 23 March 2021 / Published: 29 March 2021
(This article belongs to the Special Issue Volatility Modelling and Forecasting)

Abstract

:
Intraday high-frequency data of stock returns exhibit not only typical characteristics (e.g., volatility clustering and the leverage effect) but also a cyclical pattern of return volatility that is known as intraday seasonality. In this paper, we extend the stochastic volatility (SV) model for application with such intraday high-frequency data and develop an efficient Markov chain Monte Carlo (MCMC) sampling algorithm for Bayesian inference of the proposed model. Our modeling strategy is two-fold. First, we model the intraday seasonality of return volatility as a Bernstein polynomial and estimate it along with the stochastic volatility simultaneously. Second, we incorporate skewness and excess kurtosis of stock returns into the SV model by assuming that the error term follows a family of generalized hyperbolic distributions, including variance-gamma and Student’s t distributions. To improve efficiency of MCMC implementation, we apply an ancillarity-sufficiency interweaving strategy (ASIS) and generalized Gibbs sampling. As a demonstration of our new method, we estimate intraday SV models with 1 min return data of a stock price index (TOPIX) and conduct model selection among various specifications with the widely applicable information criterion (WAIC). The result shows that the SV model with the skew variance-gamma error is the best among the candidates.

1. Introduction

It is well documented that (a) probability distributions of stock returns are heavy-tailed (both tails of the probability density function go down to zero much slower than in the case of the normal distribution, and as a result, the kurtosis of the distribution exceeds 3), (b) they are often asymmetric around the mean (the skewness of the distribution is either positive or negative), (c) they exhibit volatility clustering (positive autocorrelation among the day-to-day variance of returns) and (d) the leverage effect (the current volatility and the previous return are negatively correlated so that downturns in the stock market tend to predate sharper spikes in the volatility). In the practice of financial risk management, it is imperative to develop a statistical model that can capture these characteristics of stock returns because they are thought to be related to steep drops and rebounds in stock prices during the periods of financial turmoil. Without factoring them into risk management, financial institutions might unintentionally take on a higher risk and as a result would be faced with grave consequences, which we already observed during the Global Financial Crisis.
As a time-series model with the aforementioned characteristics, a family of time-series models called the stochastic volatility (SV) model has been developed in the field of financial econometrics. The standard SV model is a simple state-space model in which the measurement equation is a mere distribution of stock returns with the time-varying variance (volatility) and the system equation is an AR(1) process of the latent log volatility. In the standard setting, both measurement and system errors are supposed to be Gaussian and negatively correlated in order to incorporate the leverage effect into the model. The standard SV model can explain three stylized facts: heavy-tailed distribution, volatility clustering and the leverage effect, but it cannot make the distribution of stock returns asymmetric. Furthermore, although in theory the standard SV model incorporates the heavy-tail behavior of stock returns, many empirical studies demonstrated that it was insufficient to explain extreme fluctuations of stock prices that were caused by large shocks in financial markets.
Based on the plain-vanilla SV model, researchers have developed numerous variants that are designed to capture all aspects of stock returns sufficiently well. The SV model has been pioneered by Taylor (1982), and numerous studies related to the SV model have been conducted so far. The Markov chain Monte Carlo (MCMC) algorithms for SV models, which can be analyzed by numerical method, have been introduced by (Jacquier et al. 1994, 2004). Ghysels et al. (1996) also survey and develop statistical inferences of the SV model including a Bayesian approach. A direct way to introduce a more heavy-tailed distribution to the SV model is to assume that the error term of the measurement equation follows a distribution with much heavier tails than the normal distribution. The Student’s t distribution is a popular choice ( Berg et al. 2004; Omori et al. 2007; Nakajima and Omori 2009; Nakajima 2012 among others). In the literature, the asymmetry in stock returns can be handled by assuming that the error term follows an asymmetric distribution (Nakajima and Omori 2012; Tsiotas 2012; Abanto-Valle et al. 2015 among others). In particular, the generalized hyperbolic (GH) distribution proposed by Barndorff-Nielsen (1977) has recently drawn increasing attention among researchers (e.g., Nakajima and Omori 2012), since it is regarded as a broad family of heavy-tailed distributions such as variance-gamma and Student’s t, as well as their skewed variants such as skew variance-gamma and skew Student’s t.
As an alternative to the SV model, the realized volatility (RV) model (e.g., Andersen and Bollerslev 1997, 1998) is often applied to evaluation of daily volatility. A naive RV estimator is defined as the sum of squared intraday returns. It converges to the daily integrated volatility as the time interval of returns becomes shorter. Due to this characteristic, RV is suitable for foreign exchange markets, which are open for 24 h a day continuously, though this may not be the case for stock markets. Most stock markets close at night, and some of them, including the Tokyo Stock Exchange, have lunch breaks when no transactions take place. It is well known that the naive RV estimator is biased for such stock markets. Nonetheless, since RV is a convenient tool for volatility estimation, researchers have developed various improved estimators of RV as well as robust estimators of its standard error. For example, Mykland and Zhang (2017) proposed a general nonparametric method called the observed asymptotic variance for assessing the standard error of RV.
Traditionally, empirical studies with the SV model as well as the RV model focused on daily volatility of asset returns. However, the availability of high-frequency tick data and the advent of high-frequency trading (HFT), which is a general term for algorithmic trading in full use of high-performance computing and high speed communication technology, has shifted the focus of research on volatility from closing-to-closing daily volatility to intraday volatility in a very short interval (e.g., 5 min or shorter). This shift paved the way for a new type of SV model. In addition to the traditional stylized facts on daily volatility, intraday volatility is known to exhibit a cyclical pattern during trading hours. On a typical trading day, the volatility tends to be high immediately after the market opens, but it gradually declines in the middle of trading hours. In the late trading hours, the volatility again becomes higher as it nears the closing time. This U-shaped trend in volatility is called intraday seasonality in the literature (see Chan et al. 1991 among others). Although it is crucial to take the intraday seasonality into consideration in estimation of any intraday volatility models, only a few studies (e.g., Stroud and Johannes 2014; Fičura and Witzany 2015a, 2015b) explicitly incorporate it into their volatility models.
In this paper, we propose to directly embed intraday seasonality into the SV model by approximating the U-shaped seasonality pattern with a linear combination of Bernstein polynomials. In order to capture skewness and excess kurtosis in high-frequency stock returns, we employ two distributions (variance-gamma and Student’s t) and their skewed variants (skew variance-gamma and skew Student’s t) in the family of GH distributions as the distribution of stock returns in the SV model. The complicated SV models generally tend to be inefficient for analyzing in a primitive form. In order to solve the problem, numerous studies concerned with efficiency of the SV model have been developed. Omori and Watanabe (2008) introduce a sampling method with block unit for asymmetric SV models, which can sample disturbances from their conditional posterior distribution simultaneously. As another approach to optimize computation, a Sequential Monte Carlo (SMC) algorithm for Bayesian semi-parametric SV model was designed by Virbickaite et al. (2019). The ancillarity-sufficiency interweaving strategy (ASIS) proposed by Yu and Meng (2011) is highly effective to improve MCMC sampling effeciency. We discuss ASIS in detail in Section 3. Needless to say, since the proposed SV model is intractably complicated, we develop an efficient Markov chain Monte Carlo (MCMC) sampling algorithm for full Bayesian estimation of all parameters and state variables (latent log volatilities in our case) in the model.
The rest of this paper is organized as follows. In Section 2, we introduce a reparameterized Gaussian SV model with leverage and intraday seasonality and derive an efficient MCMC sampling algorithm for its Bayesian estimation. In addition, we show the conditional posterior distributions and prepare for application of ASIS. In Section 3, we extend the Gaussian SV model to the case of variance gamma and Student’s t error as well as their skewed variants. In Section 4, we report the estimation results of our proposed SV models with 1 min return data of TOPIX. Finally, conclusions are given in Section 5.

2. Stochastic Volatility Model with Intraday Seasonality

Consider the log difference of a stock price in a short interval (say, 1 or 5 min). We divide trading hours evenly into T periods and normalize them so that the length of the trading hours is equal to 1; that is, the length of each period is 1 T and the time stamp of the t-th period is t T ( t = 1 , , T ) . Note that the market opens at time 0 and closes at time 1 in our setup. Let y t ( t = 1 , , T ) denote the stock return in the t-th period (at time t T in the trading hours) and consider the following stochastic volatility (SV) model of y t with intraday seasonality:
y t = exp ( x t β + h t ) ϵ t , h t + 1 = ϕ h t + η t , ϵ t η t Normal 0 0 , 1 ρ τ ρ τ τ 2 , | ρ | < 1 , τ > 0 ,
and
h 1 Normal 0 , τ 2 1 ϕ 2 , | ϕ | < 1 .
It is well known that the estimate of the correlation coefficient ρ is negative in most stock markets. This negative correlation is often referred to as the leverage effect. Note that the stock volatility in the t-th period (the natural logarithm of the conditional standard deviation of y t ) is
log Var [ y t | F t 1 ] = x t β + h t ,
where F t 1 is the filtration that represents all available information at time t 1 T . Hence, the stock volatility in the SV model (1) is decomposed into two parts: a linear combination of covariates x t β and the unobserved AR(1) process h t . In this paper, we regard x t β as the intraday seasonal component of the stock volatility, though it can be interpreted as any function of covariates x t in a different situation. On the other hand, h t is supposed to capture volatility clustering. We call h t the latent log volatility since it is unobservable.
Although the intraday seasonal component x t β is likely to be a U-shaped function of time stamps (the stock volatility is higher right after the opening or near the closing, but it is lower in the middle of the trading hours), we have no information about the exact functional form of the intraday seasonality. To make it in a flexible functional form for the intraday seasonality, we assume that x t β is a Bernstein polynomial
x t β = k = 0 n β k x k , t = k = 0 n β k b k , n t T ,
where b k , n ( · ) is called a Bernstein basis polynomial of degree n:
b k , n ( v ) = n C k v k ( 1 v ) n k , k = 0 , , n , v [ 0 , 1 ] .
According to the Weierstrass approximation theorem, the Bernstein polynomial (2) can approximate any continuous function on [ 0 , 1 ] as n goes to infinity. In practice, however, the number of observations T is finite. Thus, we need to choose a finite n via a model selection procedure. We will discuss this issue in Section 4.
Although the parameterization of the SV model in (1) is widely applied in the literature, we propose an alternative parameterization that facilitates MCMC implementation in non-Gaussian SV models. By replacing the covariance matrix in (1) with
Var [ ϵ t ] Cov [ η t , ϵ t ] Cov [ ϵ t , η t ] Var [ η t ] = 1 + γ 2 τ 2 γ τ 2 γ τ 2 τ 2 , γ R ,
we obtain an alternative formulation of the SV model:
y t = exp ( x t β + h t ) ϵ t , h t + 1 = ϕ h t + η t , ϵ t η t Normal 0 0 , 1 + γ 2 τ 2 γ τ 2 γ τ 2 τ 2 .
Since in (4) the variance of ϵ t is no longer equal to one, the interpretation of β and h t in (4) is slightly different from the original one in (1). Nonetheless, the SV model (4) has essentially the same characteristics as (1). Since the correlation coefficient in (3) is
Corr [ ϵ t , η t ] = γ τ 1 + γ 2 τ 2 ,
the sign of γ always coincides with the correlation coefficient and the leverage effect exists if γ < 0 . To distinguish γ in (4) from the correlation parameter ρ in (1), we call γ the leverage parameter in this paper.
Note that the inverse of (3) is
Var [ ϵ t ] Cov [ η t , ϵ t ] Cov [ ϵ t , η t ] Var [ η t ] 1 = 1 γ γ γ 2 + τ 2 = 1 0 γ τ 1 1 γ 0 τ 1 ,
and the determinant of (3) is τ 2 . Using
ϵ t η t 1 γ γ γ 2 + τ 2 ϵ t η t =   ϵ t η t 1 0 γ τ 1 1 γ 0 τ 1 ϵ t η t = ϵ t γ η t 2 + η t 2 τ 2 ,
we can easily show that the SV model (4) is equivalent to
y t = exp ( x t β + h t ) ( z t + γ η t ) , h t + 1 = ϕ h t + η t ,
where
z t Normal ( 0 , 1 ) , η t Normal ( 0 , τ 2 ) , z t η t .
In the alternative formulation of the SV model (5), we can interpret η t as a common shock that affects both the stock return y t and the log volatility h t + 1 and z t as an idiosyncratic shock that affects y t only.
The likelihood for the SV model (5) given the observations y 1 : T = [ y 1 ; ; y T ] , and the latent log volatility h 1 : T + 1 = [ h 1 ; ; h T + 1 ] is
p ( y 1 : T , h 1 : T + 1 | θ ) = t = 1 T p ( y t | h t , h t + 1 , θ ) p ( y 1 : T | h 1 : T + 1 , θ ) · p ( h 1 | θ ) t = 1 T p ( h t + 1 | h t , θ ) p ( h 1 : T + 1 | θ ) ,
where
p ( y t | h t , h t + 1 , θ ) = 1 2 π exp x t β h t y t exp x t β h t γ ( h t + 1 ϕ h t ) 2 2 , p ( h t + 1 | h t , θ ) = 1 2 π τ 2 exp ( h t + 1 ϕ h t ) 2 2 τ 2 , t = 1 , , T , p ( h 1 | θ ) = 1 ϕ 2 2 π τ 2 exp ( 1 ϕ 2 ) h 1 2 2 τ 2 ,
and θ = ( β , γ , τ 2 , ϕ ) . Since h t follows a stationary AR(1) process, the joint probability distribution of h 1 : T + 1 is Normal ( 0 , τ 2 V 1 ) , where
V = 1 ϕ ϕ 1 + ϕ 2 ϕ ϕ 1 + ϕ 2 ϕ ϕ 1 + ϕ 2 ϕ ϕ 1 + ϕ 2 ϕ ϕ 1 ,
is a tridiagonal matrix, and it is positive definite as long as | ϕ | < 1 . Thus, the joint p.d.f. of h 1 : T + 1 is
p ( h 1 : T + 1 | θ ) = ( 2 π τ 2 ) T + 1 2 | V | 1 2 exp 1 2 τ 2 h 1 : T + 1 V h 1 : T + 1 , | V | = 1 ϕ 2 .
The prior distributions for ( β , γ , τ 2 , ϕ ) in our study are
β Normal ( μ ¯ β , Ω ¯ β 1 ) , γ Normal ( μ ¯ γ , ω ¯ γ 1 ) , τ 2 Inv . Gamma ( a τ , b τ ) , ϕ + 1 2 Beta ( a ϕ , b ϕ ) .
Then the joint posterior density of ( h 1 : T + 1 , θ ) for the SV model (5) is
p ( h 1 : T + 1 , θ | y 1 : T ) t = 1 T p ( y t | h t , h t + 1 , θ ) · p ( h 1 : T + 1 | θ ) · p ( θ ) ,
where p ( θ ) is the prior density of the parameters in (10).
Since analytical evaluation of the joint posterior distribution (11) is impractical, we apply an MCMC method to generate a random sample { ( h 1 : T + 1 ( r ) , β ( r ) , γ ( r ) , τ 2 ( r ) , ϕ ( r ) ) } r = 1 R from the joint posterior distribution (11) and numerically evaluate the posterior statistics necessary for Bayesian inference with Monte Carlo integration. The outline of the standard MCMC sampling scheme for the posterior distribution (11) is given as follows:
Jrfm 14 00145 i001
Although the above MCMC sampling scheme is ubiquitous in the literature of the SV model, the generated Monte Carlo sample { ( h 1 : T + 1 ( r ) , β ( r ) , γ ( r ) , τ 2 ( r ) , ϕ ( r ) ) } r = 1 R tends to exhibit strongly positive autocorrelation. To improve efficiency of MCMC implementation, Yu and Meng (2011) proposed an ancillarity-sufficiency interweaving strategy (ASIS). In the literature of the SV model, Kastner and Frühwirth-Schnatter (2014) applied ASIS to the SV model of daily US-dollar/Euro exchange rate data with the Gaussian error. Their SV model did not include either intraday seasonality or the leverage effect since they applied it to daily exchange rate data that exhibited no leverage effect in most cases. We extend the algorithm developed by Kastner and Frühwirth-Schnatter (2014) to facilitate the converge of the sample path in the SV model (5). The basic principle of ASIS is to construct MCMC sampling schemes for two different but equivalent parameterizations of a model with missing/latent variables ( h 1 : T + 1 in our case) and generate the parameters alternately with each of them.
According to Kastner and Frühwirth-Schnatter (2014), the SV model (5) is in a non-centered parameterization (NCP). On the other hand, we may transform h t as
h ˜ t = x t β + h t ,
and rearrange the SV model (5) as
y t = exp ( h ˜ t ) ( z t + γ η t ) , h ˜ t + 1 x t + 1 β = ϕ ( h ˜ t x t β ) + η t .
The above SV model (13) is in a centered parameterization (CP).
The posterior distribution in the CP form (13) is equivalent to the one in the NCP form (5) in the sense that they give us the same posterior distribution of θ . Let us verify this claim. The likelihood for the SV model (13) given the observations y 1 : T and the latent log volatility h ˜ 1 : T + 1 = [ h ˜ 1 ; ; h ˜ T + 1 ] is
p ( y 1 : T , h ˜ 1 : T + 1 | θ ) = t = 1 T p ( y t | h ˜ t , h ˜ t + 1 , θ ) p ( y 1 : T | h ˜ 1 : T + 1 , θ ) · p ( h ˜ 1 | θ ) t = 1 T p ( h ˜ t + 1 | h ˜ t , θ ) p ( h ˜ 1 : T + 1 | θ ) ,
where
p ( y t | h ˜ t , h ˜ t + 1 , θ ) = 1 2 π exp h ˜ t y t e h ˜ t γ ( ( h ˜ t + 1 x t + 1 β ) ϕ ( h ˜ t x t β ) ) 2 2 , p ( h ˜ t + 1 | h ˜ t , θ ) = 1 2 π τ 2 exp { ( h ˜ t + 1 x t + 1 β ) ϕ ( h ˜ t x t β ) } 2 2 τ 2 , t = 1 , , T , p ( h ˜ 1 | θ ) = 1 ϕ 2 2 π τ 2 exp ( 1 ϕ 2 ) ( h ˜ 1 x 1 β ) 2 2 τ 2 .
Note that the joint p.d.f. of h ˜ 1 : T + 1 is
p ( h ˜ 1 : T + 1 | θ ) = ( 2 π τ 2 ) T + 1 2 | V | 1 2 exp 1 2 τ 2 ( h ˜ 1 : T + 1 X β ) V ( h ˜ 1 : T + 1 X β ) ,
where X = [ x 1 ; ; x T + 1 ] . With the prior of θ in (10), the joint posterior density of ( h ˜ 1 : T + 1 , θ ) for the SV model (13) is obtained as
p ( h ˜ 1 : T + 1 , θ | y 1 : T ) t = 1 T p ( y t | h ˜ t , h ˜ t + 1 , θ ) · p ( h ˜ 1 : T + 1 | θ ) · p ( θ ) .
Note that θ is unchanged between the NCP form (11) and the CP form (17). Although the latent variables are transformed with (12), the “marginal” posterior p.d.f. of θ is unchanged, because
p ( h ˜ 1 : T + 1 , θ | y 1 : T + 1 ) d h ˜ 1 : T + 1 = p ( h 1 : T + 1 , θ | y 1 : T + 1 ) | J | d h 1 : T + 1 ,
where the Jacobian | J | = 1 .
With this fact in mind, we can incorporate ASIS into the MCMC sampling scheme by replacing Step 1 with
Jrfm 14 00145 i002
Note that we generate a new latent log volatility h 1 : T + 1 from its conditional posterior distribution in the NCP form (11) only once at the beginning of Step 1. This is the reason we call it the NCP-based ASIS step. After this update, we merely shift the location of h 1 : T + 1 by x t β ( r + 0.5 ) (Step 1) or by x t β ( r + 1 ) (Step 1.5). In ASIS, these shifts are applied with probability 1 even if all elements in h 1 : T + 1 are not updated at the beginning of Step 1, which is highly probable in practice because we need to use the MH algorithm to generate h 1 : T + 1 . Although we also utilize the MH algorithm to generate β , as explained later, the acceptance rate of β in the MH step is much higher than that of h 1 : T + 1 in our experience. Thus, we expect that both x t β ( r + 0.5 ) and x t β ( r + 1 ) will be updated more often than h 1 : T + 1 itself. As a result, the above ASIS step may improve mixing of the sample sequence of h 1 : T + 1 . Conversely, we may apply the following CP-based ASIS step:
Jrfm 14 00145 i003
In the CP-based ASIS step, we generate h ˜ 1 : T + 1 from its conditional posterior distribution in the CP form (17) once. The rest is the same as in the NCP-based ASIS step except that the order of sampling is reversed.
In the NCP form, the conditional posterior distributions for ( β , γ , τ 2 , ϕ ) are
β Normal μ β ( β ) , Σ β ( β ) ,
where
Σ β ( β ) = Q ( β ) + Ω ¯ β 1 , μ β ( β ) = Σ β ( β ) g ( β ) + Q ( β ) β + Ω ¯ β μ ¯ β .
γ | h 1 : T + 1 , θ γ , y 1 : T Normal t = 1 T η t ϵ t + ω ¯ γ μ ¯ γ t = 1 T η t 2 + ω ¯ γ , 1 t = 1 T η t 2 + ω ¯ γ .
τ 2 | h 1 : T + 1 , θ τ 2 , y 1 : T Inv . Gamma T + 1 2 + a τ , 1 2 h 1 : T + 1 V h 1 : T + 1 + b τ .
ϕ Normal t = 1 T h t + 1 h t t = 2 T h t 2 , τ 2 t = 2 T h t 2 1 < ϕ < 1 .
In the CP form, the conditional posterior distributions for ( β , γ , τ 2 , ϕ ) are
β Normal μ ˜ β , Σ ˜ β ,
where
Σ ˜ β = X ˜ X ˜ + 1 τ 2 X V X + Ω ¯ β 1 , μ ˜ β = Σ ˜ β X ˜ y ˜ + 1 τ 2 X V h ˜ 1 : T + 1 + Ω ¯ β μ ¯ β .
γ | h ˜ 1 : T + 1 , θ γ , y 1 : T Normal t = 1 T η ˜ t ϵ ˜ t + ω ¯ γ μ ¯ γ t = 1 T η ˜ t 2 + ω ¯ γ , 1 t = 1 T η ˜ t 2 + ω ¯ γ .
τ 2 | h ˜ 1 : T + 1 , θ τ 2 , y 1 : T Inv . Gamma T + 1 2 a τ , 1 2 ( h ˜ 1 : T + 1 X β ) V ( h ˜ 1 : T + 1 X β ) + b τ .
ϕ Normal t = 1 T ( h ˜ t + 1 x t + 1 β ) ( h ˜ t x t β ) t = 2 T ( h ˜ t x t β ) 2 , τ 2 t = 2 T ( h ˜ t x t β ) 2 1 < ϕ < 1 .
Derivations of the conditional posterior distributrions are shown in Appendix A.

3. Extension: Skew Heavy-Tailed Distributions

3.1. Mean-Variance Mixture of the Normal Distribution

It is a well-known stylized fact that probability distributions of stock returns are almost definitely heavy-tailed (the probability density goes down to zero much slower than the normal distribution) and often have non-zero skewness (they are not symmetric around the mean). Although introduction of stochastic volatility and leverage makes the distribution of y t skew and heavy-tailed, it may not be sufficient to capture those characteristics of real data. For this reason, instead of the normal distribution, we introduce a skew heavy-tailed distribution to the SV model.
In our study, we suppose that z t in (5) is expressed as a mean-variance mixture of the standard normal distribution:
z t = α δ t + δ t u t , u t Normal ( 0 , 1 ) , δ t GIG ( λ , ψ , ξ ) ,
where GIG ( λ , ψ , ξ ) stands for the generalized inverse Gaussian distribution with the probability density:
p ( δ t ) = ( ψ / ξ ) λ / 2 2 K λ ( ψ ξ ) δ t λ 1 exp 1 2 ψ δ t + ξ δ t ,
where
λ R , ( ψ , ξ ) { ( ψ , ξ ) : ψ > 0 , ξ 0 } if λ > 0 , { ( ψ , ξ ) : ψ > 0 , ξ > 0 } if λ = 0 , { ( ψ , ξ ) : ψ 0 , ξ > 0 } if λ < 0 ,
and K λ ( · ) is the modified Bessel function of the second kind. The family of generalized inverse Gaussian distributions includes
  • exponential distribution ( λ = 1 , ξ = 0 ),
  • gamma distribution ( λ > 0 , ξ = 0 ),
  • inverse gamma distribution ( λ < 0 , ψ = 0 ),
  • inverse Gaussian distribution ( λ = 1 2 )
Under the assumption (26), the distribution of z t belongs to the family of generalized hyperbolic distributions proposed by Barndorff-Nielsen (1977), which includes many well-known skew heavy-tailed distributions such as
  • skew variance gamma (VG) distribution( λ = ν 2 , ψ = ν , ξ = 0 ),
  • skew t distribution ( λ = ν 2 , ψ = 0 , ξ = ν ),
where ν > 0 . In general, the skew VG distribution is a mean-variance mixture of the standard normal distribution with GIG ( λ , ψ , 0 ) . To make the estimation easier, we set λ = ν 2 and ψ = ν so that the skew VG distribution has only two free parameters ( α , ν ) . Thus, we have two additional parameters ( α , ν ) in the SV model. Since α determines whether the distribution of y t is symmetric or not while ν determines how heavy-tailed the distribution is, we call α the asymmetry parameter and ν the tail parameter, respectively. In our study, we use the above three skew heavy-tailed distributions as alternatives to the normal distribution. To distinguish each model specification, we use the following abbreviations:
SV-N:stochastic volatility model with the normal error,
SV-G:stochastic volatility model with the VG error,
SV-SG:stochastic volatility model with the skew VG error,
SV-T:stochastic volatility model with the Student-t error,
SV-ST:stochastic volatility model with the skew t error.
In this setup, the SV model with heavy-tailed error is formulated as
y t = exp ( x t β + h t ) ϵ t , h t + 1 = ϕ h t + η t , ϵ t η t δ t Normal α δ t 0 , δ t + γ 2 τ 2 γ τ 2 γ τ 2 τ 2 .
It is straightforward to show that the conditional probability density of y t given ( h t , h t + 1 ) is given by
p ( y t | h t , h t + 1 , θ ) = 0 p ( y t | h t , h t + 1 , δ t , θ ) p ( δ t | ν ) δ t ,
where θ = ( β , γ , τ 2 , ϕ , α , ν ) ,
p ( y t | h t , h t + 1 , δ t , θ ) = 1 2 π δ t exp x t β h t y t exp x t β h t α δ t γ ( h t + 1 ϕ h t ) 2 2 δ t ,
and
p ( δ t | ν ) = ( ν / 2 ) ν / 2 Γ ( ν / 2 ) δ t ν 2 1 exp ν 2 δ t ( SV - SG ) , ( ν / 2 ) ν / 2 Γ ( ν / 2 ) δ t ν 2 1 exp ν 2 δ t ( SV - ST ) .
Since it is impractical to evaluate the multiple integral in (29), we generate δ 1 : T = ( δ 1 , , δ T ) along with h 1 : T + 1 and θ form their joint posterior distribution. In this setup, the likelihood used in the posterior simulation is
p ( y 1 : T , h 1 : T + 1 , δ 1 : T | θ ) = p ( y 1 : T | h 1 : T + 1 , δ 1 : T , θ ) p ( h 1 : T + 1 | θ ) = t = 1 T p ( y t | h t , h t + 1 , θ ) · p ( h 1 : T + 1 | θ ) .
We suppose that the prior distributions for α and ν are
α Normal ( μ ¯ α , ω ¯ α 1 ) , ν Gamma ( a ν , b ν ) .
As for the other parameters, we keep the same ones as in (10).

3.2. Conditional Posterior Distributions

3.2.1. Latent Log Volatility h 1 : T + 1

Our sampling scheme for h 1 : T + 1 is basically the same as before. We first approximate the log likelihood with the second-order Taylor expansion around the mode and construct a proposal distribution of h 1 : T + 1 with the approximated log likelihood. Then, we apply a multi-move MH sampler for generating h 1 : T + 1 from the conditional posterior distribution. The sole differences are the functional form of g ( h 1 : T + 1 ) and Q ( h 1 : T + 1 ) .
g t ( h 1 : T + 1 ) = 1 + 1 δ t ϵ t α δ t γ η t ϵ t γ ϕ 1 ( t T ) + γ δ t 1 ϵ t 1 α δ t 1 γ η t 1 1 ( t 2 ) , ( t = 1 , , T + 1 ) ,
where 1 ( · ) is the indicator function. Each diagonal element of Q ( h 1 : T + 1 ) is
q t , t ( h 1 : T + 1 ) = 1 δ t ϵ t ( ϵ t α δ t γ η t ) + ϵ t γ ϕ 2 1 ( t T ) + γ 2 δ t 1 1 ( t 2 ) , ( t = 1 , , T + 1 ) ,
and the off-diagonal element is
q t , t + 1 ( h 1 : T + 1 ) = γ δ t ( ϵ t γ ϕ ) , ( t = 1 , , T ) .
For the NCP form, we use ϵ t and η t in (A3). For the CP form, we replace them with ϵ ˜ t and η ˜ t in (A23).

3.2.2. Regression Coefficients β

The sampling scheme for β is the same as before. For the NCP form, g ( β ) and Q ( β ) are given by
g ( β ) = t = 1 T ϵ t δ t ϵ t α δ t γ η t 1 x t ,
Q ( β ) = t = 1 T ϵ t δ t 2 ϵ t α δ t γ η t x t x t ,
respectively. For the CP form, the conditional posterior distribution of β are given by
β Normal μ ˜ β , Σ ˜ β ,
where
Σ ˜ β = X ˜ D 1 X ˜ + 1 τ 2 X V X + Ω ¯ β 1 , μ ˜ β = Σ ˜ β X ˜ D 1 y ˜ + 1 τ 2 X V h ˜ 1 : T + 1 + Ω ¯ β μ ¯ β , y ˜ t = ϵ ˜ t α δ t γ ( h ˜ t + 1 ϕ h ˜ t ) , y ˜ = [ y ˜ 1 ; ; y ˜ T ] , D = diag { δ 1 , , δ T } .

3.2.3. Leverage Parameter γ

Their conditional posterior distribution of γ is given by
γ | h 1 : T + 1 , δ 1 : T , θ γ , y 1 : T Normal t = 1 T η t ( ϵ t / δ t α ) + ω ¯ γ μ ¯ γ t = 1 T η t 2 / δ t + ω ¯ γ , 1 t = 1 T η t 2 / δ t + ω ¯ γ .
For the NCP form, we use ϵ t and η t in (A3). For the CP form, we replace them with ϵ ˜ t and η ˜ t in (A23).

3.2.4. Random Scale δ 1 : T

Using the Bayes theorem, we obtain the conditional posterior distribution of δ t as
δ t | h 1 : T + 1 , θ , y 1 : T GIG ( λ t , ψ t , ξ t ) , t = 1 , , T ,
where
( λ t , ψ t , ξ t ) = ν 1 2 , α 2 + ν , ( ϵ t γ η t ) 2 , ( SV - SG ) , ν + 1 2 , α 2 , ( ϵ t γ η t ) 2 + ν , ( SV - ST ) .
For the NCP form, we use ϵ t and η t in (A3). For the CP form, we replace them with ϵ ˜ t and η ˜ t in (A23).
To improve the performance of the MCMC algorithm, we apply a generalized Gibbs sampler by Liu and Sabatti (2000) to { δ t } t = 1 T after we generate them from the conditional posterior distribution (41). This is rather simple. All we need to do is to multiply each of { δ t } t = 1 T by a random number c that is generated from
c GIG ( ν 1 ) T 2 , ( α 2 + ν ) t = 1 T δ t , t = 1 T ( ϵ t γ η t ) 2 δ t , ( SV - SG ) , GIG ( ν + 1 ) T 2 , α 2 t = 1 T δ t , t = 1 T ( ϵ t γ η t ) 2 + ν δ t , ( SV - ST ) .

3.2.5. Asymmetry Parameter α

Using the Bayes theorem, we obtain the conditional posterior distribution of α as
α | h 1 : T + 1 , δ 1 : T , θ α , y 1 : T Normal t = 1 T ( ϵ t γ η t ) + ω ¯ α μ ¯ α t = 1 T δ t + ω ¯ α , 1 t = 1 T δ t + ω ¯ α .
For the NCP form, we use ϵ t and η t in (A3). For the CP form, we replace them with ϵ ˜ t and η ˜ t in (A23).

3.2.6. Tail Parameter ν

The explicit form of the conditional posterior density of ν is not available. Therefore, we apply the MH algorithm for generating ν . Note that the gamma density for SV-SG in (31) is identical to the inverse gamma density for SV-ST in (31) as a function of ν if we exchange δ t with δ t 1 . Since we use the same gamma prior for ν in either case, the resultant conditional posterior density should be the same in both SV-SG and SV-ST. Therefore, it suffices to derive the MH algorithm for SV-ST.
The sampling strategy for ν is basically the same as for β , which was originally proposed by Watanabe (2001). We first consider the second-order Taylor expansion of the log conditional posterior density of ν :
f ( ν ) = t = 1 T log p ( δ t | ν ) + log p ( ν ) + constant
= ν T 2 log ν 2 T log Γ ν 2 ν 1 2 t = 1 T log δ t + 1 δ t + b ν + ( a ν 1 ) log ν + constant ,
with respect to ν in the neighborhood of ν > 0 , i.e.,
f ( ν ) f ( ν ) + g ( ν ) ( ν ν ) 1 2 q ( ν ) ( ν ν ) 2 ,
where
g ( ν ) ν f ( ν ) = T 2 + T 2 log ν 2 T 2 ψ ( 0 ) ν 2 1 2 t = 1 T log δ t + 1 δ t b ν + a ν 1 ν , q ( ν ) ν 2 f ( ν ) = T 2 ν + T 4 ψ ( 1 ) ν 2 + a ν 1 ν 2 ,
and ψ ( s ) is the polygamma function of order s. Note that q ( ν ) > 0 if T + 2 a ν > 2 . See Theorem 1 in Watanabe (2001) for the proof. By applying the completing-the-square technique to (46), we obtain the proposal distribution for the MH algorithm:
ν Normal μ ν ( ν ) , σ ν 2 ( ν ) ,
where
σ ν 2 ( ν ) = 1 q ( ν ) , μ ν ( ν ) = ν + g ( ν ) q ( ν ) .
If we use the mode of f ( ν ) as ν , g ( ν ) = 0 always holds due to the global concavity of f ( ν ) . Thus, μ ν ( ν ) is effectively identical to ν .

4. Empirical Study

As an application of our proposed models to real data, we analyze high-frequency data of the Tokyo Stock Price Index (TOPIX), a market-cap-weighted stock index based on all domestic common stocks listed in the Tokyo Stock Exchange (TSE) First Section, which is provided by Nikkei Media Marketing. We use the data in June 2016, when the referendum for the UK’s withdrawal from the EU (Brexit) was held on the 23rd of the month. The result of the Brexit referendum was announced during the trading hours of the TSE on that day. That news made the Japanese Stock Market plunge significantly. The Brexit referendum is arguably one of the biggest financial events in recent years. We can thus analyze the effect of the Brexit referendum on the volatility of the Japanese stock market. Another reason for this choice is that Japan has no holiday in June, so all weekdays are trading days. There are five weeks in June 2016. Since the first week of June 2016 includes 30 and 31 May and the last week includes 1 July, we also include them in the sample period.
The morning session of TSE starts at 9:00 a.m. and ends at 11:30 a.m. while the afternoon session of TSE starts at 12:30 a.m. and ends at 3:00 p.m., so both sessions last for 150 min. We treat the morning session and the afternoon session as if they are separated trading hours, and normalize the time stamps so that they take values within [ 0 , 1 ] . As a result, t = 0 corresponds to 9:00 a.m. for the morning session, while it corresponds to 12:30 a.m. for the afternoon session. In the same manner, t = 1 corresponds to 11:30 a.m. for the morning session, while it corresponds to 3:00 p.m. for the afternoon session. In this empirical study, we estimate the Bernstein polynomial of the intraday seasonality in each session by allowing β in (2) to differ from session to session.
We pick prices at every 1 min and compute 1 min log difference of prices as 1 min stock returns. Thus, the number of observations per session is 150. Furthermore, we put together all series of 1 min returns in each week. As a result, the total number of observations per week is 150 × 2 × 5 = 1500 . In addition, to simplify the interpretation of the estimation results, we standardize each week-long series of 1 min returns so that the sample mean is 0 and the sample variance is 1. Table 1 shows the descriptive statistics of the standard 1 min returns of TOPIX in each week, while Figure 1, Figure 2, Figure 3, Figure 4 and Figure 5 show the time series plots of the standardized 1 min returns for each week.
We consider five candidates (SV-N, SV-G, SV-SG, SV-T, SV-ST) in the SV model (28) and set the prior distributions as follows:
β Normal ( 0 , 100 I ) , γ Normal ( 0 , 100 ) , τ 2 Inverse Gamma ( 1 , 0.04 ) , ϕ + 1 2 Beta ( 1 , 1 ) , α Normal ( 0 , 100 ) , ν Gamma ( 0 , 0.1 ) .
We vary the order of the Bernstein polynomial from 5 to 10. In sum, we try 30 different model specifications for the SV model (28). In the MCMC implementation, we generate 10,000 draws after the first 5000 draws are discarded as the burn-in periods. To select the best model among the candidates, we employ the widely applicable information criterion (WAIC, Watanabe 2010). We compute the WAIC of each model specification with the formula by Gelman et al. (2014). The results are reported in Table 2, Table 3, Table 4, Table 5 and Table 6. According to these tables, SV-G or SV-SG is the best model in all months. It may be a notable finding since the SV model with the variance-gamma error has hardly been applied in the previous studies.
For the selected models, we compute the posterior statistics (posterior means, standard deviations, 95% credible intervals and inefficiency factors) of the parameters and report them in Table 7, Table 8, Table 9, Table 10 and Table 11. The inefficiency factor measures how inefficient the MCMC sampling algorithm is (see e.g., Chib 2001). In these tables, the 95% credible intervals of the leverage parameter γ and the asymmetric parameter α contain 0 for all specifications. Thus, we may conclude that the error distribution of 1 min returns of TOPIX is not asymmetric. In addition, most of the marginal posterior distributions of ϕ are concentrated near 1, even though the uniform prior is assumed for ϕ . This suggests that the latent log volatility is strongly persistent, which is consistent with findings by previous studies on the stock markets. Regarding the tail parameter ν , its marginal posterior distribution is centered around 2–6 in most models, which indicates that the excess kurtosis of the error distribution is high.
As for the intraday seasonality, the estimates of β themselves are not of our interest. Instead we show the posterior mean and the 95% credible interval of the Bernstein polynomial x t β in Figure 6. These figures show that some of the trading days exhibit the well-known U-shaped curve of intraday volatility, but others slant upward or downward. At the beginning on the day of Brexit (23 June), the market began with a highly volatile situation, but the volatility gradually became lower. During the afternoon session, the volatility was kept in a stable condition.

5. Conclusions

In this paper, we extended the standard SV model into a more general formulation so that it could capture key characteristics of intraday high-frequency stock returns such as intraday seasonality, asymmetry and excess kurtosis. Our proposed model uses a Bernstein polynomial of time stamps as the intraday seasonal component of the stock volatility, and the coefficients in the Bernstein polynomial are simultaneously estimated along with the rest of the parameters in the model. To incorporate asymmetry and excess kurtosis into the standard SV model, we assume that the error distribution of stock returns in the SV model belongs to a family of generalized hyperbolic distributions. In particular, we focus on two sub-classes of this family: skew Student’s t distribution and skew variance-gamma distribution. Furthermore we developed an efficient MCMC sampling algorithm for Bayesian inference on the proposed model by utilizing all without a loop (AWOL), ASIS and the generalized Gibbs sampler.
As an application, we estimated the proposed SV models with 1 min return data of TOPIX in various specifications and conducted model selection with WAIC. The model selection procedure chose the SV model with the variance-gamma-type error as the most suitable one. The estimated parameters indicated strong excess kurtosis in the error distribution of 1 min returns, though the asymmetry was not supported since both leverage parameter γ and asymmetry parameter α were not significantly different from zero. Furthermore our proposed model successfully extracted intraday seasonal patterns in the stock volatility with Bernstein polynomial approximation, though the shape of the intraday seasonal component was not necessarily U-shaped.

Author Contributions

M.N. had full access to the data in the study and takes responsibility for the accuracy and integrity of the data and its analyses. Study concept and design: All authors. Acquisition, analysis, or interpretation of data: All authors. Drafting of the manuscript: M.N. Critical revision of the manuscript for important intellectual content: T.N. Statistical analysis: All authors. Administrative, technical or material support: All authors. Study supervision: T.N. All authors have read and agreed to the published version of the manuscript

Funding

This study was funded by the Japan Society for the Promotion of Science (JSPS) KAKENHI Grant Number 19K01592.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Conditional Posterior Distributions

In Appendix A, we derive the conditional posterior distribution of the latent log volatility and that of each parameter in the SV model for both NCP and CP.

Appendix A.1. NCP Form

Appendix A.1.1. Latent Log Volatility h1:T+1

The conditional posterior density of the latent log volatility h 1 : T + 1 is
p ( h 1 : T + 1 | θ , y 1 : T ) t = 1 T p ( y t | h t , h t + 1 , θ ) · p ( h 1 : T + 1 | θ ) .
We apply the Metropolis-Hastings (MH) algorithm to generate h 1 : T + 1 from (A1). To derive a suitable proposal distribution for the MH algorithm, we first consider consider the second-order Taylor approximation of ( h 1 : T + 1 ) = log p ( y 1 : T | h 1 : T + 1 , θ ) in the neighborhood of h 1 : T + 1 :
( h 1 : T + 1 ) ( h 1 : T + 1 ) + g ( h 1 : T + 1 ) ( h 1 : T + 1 h 1 : T + 1 ) 1 2 ( h 1 : T + 1 h 1 : T + 1 ) Q ( h 1 : T + 1 ) ( h 1 : T + 1 h 1 : T + 1 ) ,
where g ( h 1 : T + 1 ) is the gradient vector of ( h 1 : T + 1 ) :
g ( h 1 : T + 1 ) = g 1 ( h 1 : T + 1 ) g t ( h 1 : T + 1 ) g T ( h 1 : T + 1 ) g T + 1 ( h 1 : T + 1 ) = 1 log p ( y 1 : T | h 1 : T + 1 , θ ) t log p ( y 1 : T | h 1 : T + 1 , θ ) T log p ( y 1 : T | h 1 : T + 1 , θ ) T + 1 log p ( y 1 : T | h 1 : T + 1 , θ ) ,
and Q ( h 1 : T + 1 ) is the Hessian matrix of log p ( y 1 : T | h 1 : T + 1 , θ ) times 1 :
Q ( h 1 : T + 1 ) = q 11 ( h 1 : T + 1 ) q 12 ( h 1 : T + 1 ) 0 q 21 ( h 1 : T + 1 ) q 22 ( h 1 : T + 1 ) q 23 ( h 1 : T + 1 ) 0 0 q T , T 1 ( h 1 : T + 1 ) q T , T ( h 1 : T + 1 ) q T , T + 1 ( h 1 : T + 1 ) 0 q T + 1 , T ( h 1 : T + 1 ) q T + 1 , T + 1 ( h 1 : T + 1 ) .
which is a ( T + 1 ) × ( T + 1 ) band matrix.
Let us derive the explicit form of each element in g ( h 1 : T + 1 ) and Q ( h 1 : T + 1 ) . By defining
ϵ t = y t exp ( x t β h t ) , η t = h t + 1 ϕ h t ,
the log density of y t (7) is rewritten as
log p ( y t | h t , h t + 1 , θ ) = x t β h t 1 2 ( ϵ t γ η t ) 2 + constant .
Note that
t ϵ t = ϵ t , t η t = ϕ , t η t 1 = 1 .
where t = h t . Each element of g ( h 1 : T + 1 ) is derived as
g t ( h 1 : T + 1 ) = t log p ( y 1 : T | h 1 : T + 1 , θ ) = t log p ( y t | h t , h t + 1 , θ ) + t log p ( y t 1 | h t 1 , h t , θ ) = 1 ( ϵ t γ η t ) ( ϵ t γ ( ϕ ) ) ( ϵ t 1 γ η t 1 ) ( γ ) = 1 + ( ϵ t γ η t ) ( ϵ t γ ϕ ) + γ ( ϵ t 1 γ η t 1 ) ,
for t = 2 , , T ,
g 1 ( h 1 : T + 1 ) = 1 log p ( y 1 : T | h 1 : T + 1 , θ ) = 1 log p ( y 1 | h 1 , h 2 , θ ) = 1 ( ϵ 1 γ η 1 ) ( ϵ 1 γ ( ϕ ) ) = 1 + ( ϵ 1 γ η 1 ) ( ϵ 1 γ ϕ ) ,
for t = 1 , and
g T + 1 ( h 1 : T + 1 ) = T + 1 log p ( y 1 : T | h 1 : T + 1 , θ ) = T + 1 log p ( y T | h T , h T + 1 , θ ) = ( ϵ T γ η T ) ( γ ) = γ ( ϵ T γ η T ) ,
for t = T + 1 . The diagonal element in Q ( h 1 : T + 1 ) is given as
q t , t ( h 1 : T + 1 ) = ( 1 ) × t 2 log p ( y 1 : T | h 1 : T + 1 , θ ) = ( ϵ t γ ( ϕ ) ) ( ϵ t γ ϕ ) ( ϵ t γ η t ) ( ϵ t ) γ ( γ ) = ( ϵ t γ ϕ ) 2 + ϵ t ( ϵ t γ η t ) + γ 2 ,
for t = 2 , , T ,
q 11 ( h 1 : T + 1 ) = ( 1 ) × 1 2 log p ( y 1 : T | h 1 : T + 1 , θ ) = ( ϵ 1 γ ( ϕ ) ) ( ϵ 1 γ ϕ ) ( ϵ 1 γ η 1 ) ( ϵ 1 ) = ( ϵ 1 γ ϕ ) 2 + ϵ 1 ( ϵ 1 γ η 1 ) ,
for t = 1 , and
q T + 1 , T + 1 ( h 1 : T + 1 ) = ( 1 ) × T + 1 2 log p ( y 1 : T | h 1 : T + 1 , θ ) = γ ( γ ) = γ 2 ,
for t = T + 1 . Furhtermore the first off-diagonal element of Q ( h 1 : T + 1 ) is derived as
q t , t + 1 ( h 1 : T + 1 ) = ( 1 ) × t , t + 1 log p ( y 1 : T | h 1 : T + 1 , θ ) = ( γ ) ( ϵ t γ ϕ ) = γ ( ϵ t γ ϕ ) ,
for t = 1 , , T . In summary,
g t ( h 1 : T + 1 ) = { 1 + ( ϵ t γ η t ) ( ϵ t γ ϕ ) } 1 ( t T ) + γ ( ϵ t 1 γ η t 1 ) 1 ( t 2 ) ,
q t ( h 1 : T + 1 ) = { ( ϵ t γ ϕ ) 2 + ϵ t ( ϵ t γ η t ) } 1 ( t T ) + γ 2 1 ( t 2 ) ,
q t , t + 1 ( h 1 : T + 1 ) = γ ( ϵ t γ ϕ ) .
Since the log prior density of h 1 : T + 1 is
p ¯ ( h 1 : T + 1 ) = T + 1 2 log ( 2 π τ 2 ) + 1 2 log | V | 1 2 τ 2 h 1 : T + 1 V h 1 : T + 1 ,
the conditional posterior density of h 1 : T + 1 (A1) can be approximated by
p ( h 1 : T + 1 | θ , y 1 : T ) = C exp ( h 1 : T + 1 ) + p ¯ ( h 1 : T + 1 ) C exp [ ( h 1 : T + 1 ) + g ( h 1 : T + 1 ) ( h 1 : T + 1 h 1 : T + 1 ) 1 2 ( h 1 : T + 1 h 1 : T + 1 ) Q ( h 1 : T + 1 ) ( h 1 : T + 1 h 1 : T + 1 ) + p ¯ ( h 1 : T + 1 ) ] = C exp ( h 1 : T + 1 ) T + 1 2 log ( 2 π τ 2 ) + 1 2 log | V | + f ( h 1 : T + 1 ) ,
where C is the normalizing constant of the conditional posterior density and
f ( h 1 : T + 1 ) = g ( h 1 : T + 1 ) ( h 1 : T + 1 h 1 : T + 1 ) 1 2 ( h 1 : T + 1 h 1 : T + 1 ) Q ( h 1 : T + 1 ) ( h 1 : T + 1 h 1 : T + 1 ) 1 2 τ 2 h 1 : T + 1 V h 1 : T + 1 ,
By completing the square in (A9), we have
f ( h 1 : T + 1 ) = 1 2 h 1 : T + 1 μ h ( h 1 : T + 1 ) Σ h ( h 1 : T + 1 ) 1 h 1 : T + 1 μ h ( h 1 : T + 1 ) + constant ,
where
Σ h ( h 1 : T + 1 ) = Q ( h 1 : T + 1 ) + 1 τ 2 V 1 , μ h ( h 1 : T + 1 ) = Σ h ( h 1 : T + 1 ) g ( h 1 : T + 1 ) + Q ( h 1 : T + 1 ) h 1 : T + 1 .
Therefore the right-hand side of (A8) is approximately proportional to the pdf of the following normal distribution:
h 1 : T + 1 Normal μ h ( h 1 : T + 1 ) , Σ h ( h 1 : T + 1 ) .
Recall that both Q ( h 1 : T + 1 ) and V are tridiagonal matrices. Thus, Σ h ( h 1 : T + 1 ) 1 = Q ( h 1 : T + 1 ) + 1 τ 2 V is also tridiagonal. Since the Cholesky decomposition of a tridiagonal matrix and the inverse of a triangular matrix can be efficiently computed if they exist, h 1 : T + 1 is readily generated from (A11) with
h 1 : T + 1 = L 1 L 1 g ( h 1 : T + 1 ) + Q ( h 1 : T + 1 ) h 1 : T + 1 + z ˜ , z ˜ Normal ( 0 , I ) ,
where L is a lower triangular matrix obtained by the Cholesky decomposition as
L L = Q ( h 1 : T + 1 ) + 1 τ 2 V .
The above algorithm, which is called the all without a loop (AWOL) in Kastner and Frühwirth-Schnatter (2014), has been applied to Gaussian Markov random fields (e.g., Rue 2001) and state-space models (e.g., Chan and Jeliazkov 2009; McCausland et al. 2011).
Hoping that the approximation (A8) is sufficiently accurate, we use (A11) as the proposal distribution in the MH algorithm. In practice, however, we need to address two issues:
1
the choice of h 1 : T + 1 is crucial to make the approximation (A8) workable.
2
the acceptance rate of the MH algorithm tends to be too low when h 1 : T + 1 is a high-dimensional vector.
We address the former issue by using the mode of the conditional posterior density as h 1 : T + 1 . The search of the mode is performed by the following recursion:
Step 1:
Initialize h 1 : T + 1 ( 0 ) and set the counter r = 1 .
Step 2:
Update h 1 : T + 1 ( r ) by h 1 : T + 1 ( r ) = μ h ( h 1 : T + 1 ( r 1 ) ) .
Step 3:
Let r = r + 1 and go to Step 2 unless max t = 1 , , T + 1 | h t ( r ) h t ( r 1 ) | is less than the preset tolerance level.
In our experience, it mostly attains convergence in a few iterations.
We address the latter issue by applying a so-called block sampler. In the block sampler, we randomly partition h 1 : T + 1 into several sub-vectors (blocks), generate each block from its conditional distribution given the rest of the blocks and apply the MH algorithm to each generated block. Without loss of generality, suppose the proposal distribution (A11) is partitioned as
h 1 h 2 h 1 : T Normal μ h 1 μ h 2 μ h , Σ h 11 Σ h 12 Σ h 21 Σ h 11 Σ h ,
where μ h = μ h h 1 : T + 1 and Σ h = Σ h h 1 : T + 1 (we ignore the dependence on h 1 : T + 1 for brevity) and h 1 is the block to be updated in the current MH step, while h 2 contains either elements that were already updated in the previous MH steps or those to be updated in the following MH steps. It is well known that the conditional distribution of h 1 given h 2 is given by
h 1 | h 2 Normal μ h 1 + Σ h 11 Σ 12 1 h 2 μ h 2 , Σ h 11 Σ h 12 Σ h 22 1 Σ h 21 .
Note that the inverse of the covariance matrix Σ h in (A12) is
Σ h 1 = Σ h 11 Σ h 12 Σ h 21 Σ h 22 1 = Ω h 11 Ω h 11 Σ h 12 Σ h 22 1 Σ h 22 1 Σ h 21 Ω h 11 Σ h 22 1 + Σ h 22 1 Σ h 21 Ω h 11 Σ h 12 Σ h 22 1 , Ω h 11 = Σ h 11 Σ h 12 Σ h 22 1 Σ h 21 1 .
Furthermore, if we let Ω h 12 denote the upper-right block of Σ h 1 , we have
Ω h 12 = Ω h 11 Σ h 12 Σ h 22 1 .
Therefore the conditional distribution of h 1 given h 2 in (A13) is rearranged as
h 1 | h 2 Normal μ h 1 Ω h 11 1 Ω h 12 h 2 μ h 2 , Ω h 11 1 .
Recall that Σ h 1 is tridiagonal and so is Ω h 11 by construction. Thus, we can apply the AWOL algorithm:
h 1 = μ h 1 L 1 1 L 1 1 Ω h 12 ( h 2 μ h 2 ) z ˜ 1 , z ˜ 1 Normal ( 0 , I ) , L 1 L 1 = Ω h 11 ,
to generate h 1 from (A14). In essence, our approach is an AWOL variant of the block sampler proposed by Omori and Watanabe (2008).

Appendix A.1.2. Regression Coefficients β

The sampling scheme for the regression coefficients β is almost identical to the one for the log volatility h 1 : T + 1 . Let ( β ) denote log p ( y 1 : T | h 1 : T + 1 , θ ) given y 1 : T and the parameters other than β . In the same manner as (A2), consider the second-order Taylor approximation of ( β ) in the neighborhood of β :
( β ) ( β ) + g ( β ) ( β β ) 1 2 ( β β ) Q ( β ) ( β β ) ,
where g ( β ) is the gradient vector of ( β ) and Q ( β ) is the Hessian matrix of ( β ) times 1 . Since β ϵ t = ϵ t x t , we have
β log p ( y t | h t , h t + 1 , θ ) = x t + ( ϵ t 2 γ η t ϵ t ) x t , β β log p ( y t | h t , h t + 1 , θ ) = ( 2 ϵ t 2 + γ η t ϵ t ) x t x t .
Therefore, g ( β ) and Q ( β ) are obtained as
g ( β ) = t = 1 T ( ϵ t ( ϵ t γ η t ) 1 ) x t , Q ( β ) = t = 1 T ϵ t ( 2 ϵ t γ η t ) x t x t .
With the prior β Normal ( μ ¯ β , Ω ¯ β 1 ) , the conditional posterior density of β can be approximated by
p ( β | h 1 : T + 1 , θ β , y 1 : T ) = C exp ( β ) + log p ( β ) C exp ( β ) 1 2 log ( 2 π ) + 1 2 log | Ω ¯ β | × exp g ( β ) ( β β ) 1 2 ( β β ) Q ( β ) ( β β ) 1 2 ( β μ ¯ β ) Ω ¯ β ( β μ ¯ β ) .
By completing the square as in (A16), the proposal distribution for the MH algorithm is derived as
β Normal μ β ( β ) , Σ β ( β ) ,
where
Σ β ( β ) = Q ( β ) + Ω ¯ β 1 , μ β ( β ) = Σ β ( β ) g ( β ) + Q ( β ) β + Ω ¯ β μ ¯ β .
The search algorithm for β is the same as h 1 : T + 1 .
Since the dimension of β is considerably smaller than h 1 : T + 1 , it is not necessary to apply the block sampler in our experience.

Appendix A.1.3. Leverage Parameter γ

Since we use the standard conditionally conjugate prior distributions for γ , the conditional posterior distribution is given by
γ | h 1 : T + 1 , θ γ , y 1 : T Normal t = 1 T η t ϵ t + ω ¯ γ μ ¯ γ t = 1 T η t 2 + ω ¯ γ , 1 t = 1 T η t 2 + ω ¯ γ .

Appendix A.1.4. Variance τ2

Since we use the standard conditionally conjugate prior distribution for τ 2 , the conditional posterior distribution is given by
τ 2 | h 1 : T + 1 , θ τ 2 , y 1 : T Inv . Gamma T + 1 2 + a τ , 1 2 h 1 : T + 1 V h 1 : T + 1 + b τ .

Appendix A.1.5. AR(1) Coefficient ϕ

Once the state variables h 1 : T + 1 are generated, the conditional posterior density of ϕ is given by
p ( ϕ | h 1 : T + 1 , θ ϕ , y 1 : T ) 1 ϕ 2 exp ( 1 ϕ 2 ) h 1 2 + t = 1 T ( h t + 1 ϕ h t ) 2 2 τ 2 × ( 1 + ϕ ) a ϕ 1 ( 1 ϕ ) b ϕ 1 1 ( 1 , 1 ) ( ϕ ) .
By completing the square, we have
( 1 ϕ 2 ) h 1 2 + t = 1 T ( h t + 1 ϕ h t ) 2 = ( 1 ϕ 2 ) h 1 2 + t = 1 T h t + 1 2 2 ϕ t = 1 T h t + 1 h t + ϕ 2 t = 1 T h t 2 = t = 1 T + 1 h t 2 2 ϕ t = 1 T h t + 1 h t + ϕ 2 t = 2 T h t 2 = t = 2 T h t 2 ϕ t = 1 T h t + 1 h t t = 2 T h t 2 2 + t = 1 T + 1 h t 2 t = 1 T h t + 1 h t 2 t = 2 T h t 2 .
With the above expression in mind, we use the following truncated normal distribution:
ϕ Normal t = 1 T h t + 1 h t t = 2 T h t 2 , τ 2 t = 2 T h t 2 1 < ϕ < 1 ,
as the proposal distribution for ϕ in the MH algorithm.

Appendix A.2. CP Form

Appendix A.2.1. Latent Log Volatility h ˜ 1 : T + 1

The sampling scheme from the conditional posterior distribution of h ˜ 1 : T + 1 :
p ( h ˜ 1 : T + 1 | θ , y 1 : T ) t = 1 T p ( y t | h ˜ t , h ˜ t + 1 , θ ) · p ( h ˜ 1 : T + 1 | θ ) ,
is based on the MH algorithm, which is similar to the case of the NCP form. To construct the proposal distribution of h ˜ 1 : T + 1 , we consider the second-order Taylor approximation of ( h ˜ 1 : T + 1 ) = log p ( h ˜ 1 : T + 1 | θ , y 1 : T ) in the neighborhood of h ˜ 1 : T + 1 as in (A2). We first derive the explicit form of each element in g ( h ˜ 1 : T + 1 ) and Q ( h ˜ 1 : T + 1 ) . By defining
ϵ ˜ t = y t exp ( h ˜ t ) , η ˜ t = h ˜ t + 1 ϕ h ˜ t ( x t + 1 ϕ x t ) β ,
the log density of y t in (15) is rewritten as
log p ( y t | h ˜ t , h ˜ t + 1 , θ ) = h ˜ t 1 2 ( ϵ ˜ t γ η ˜ t ) 2 + constant .
Since
t ϵ ˜ t = ϵ ˜ t , t η ˜ t = ϕ , t η ˜ t 1 = 1 ,
g t ( h ˜ 1 : T + 1 ) , q t ( h ˜ 1 : T + 1 ) and q t , t + 1 ( h ˜ 1 : T + 1 ) are identical to (A4)–(A6) except that ϵ t and η t are replaced with ϵ ˜ t and η ˜ t , respectively.
Since the log prior density of h ˜ 1 : T + 1 is
p ¯ ( h ˜ 1 : T + 1 ) = T + 1 2 log ( 2 π τ 2 ) + 1 2 log | V | 1 2 τ 2 ( h ˜ 1 : T + 1 X β ) V ( h ˜ 1 : T + 1 X β ) ,
the conditional posterior density of h ˜ 1 : T + 1 (A22) can be approximated by
p ( h ˜ 1 : T + 1 | θ , y 1 : T ) = C exp ( h 1 : T + 1 ) + p ¯ ( h 1 : T + 1 ) C exp ( h ˜ 1 : T + 1 ) T + 1 2 log ( 2 π τ 2 ) + 1 2 log | V | + f ( h ˜ 1 : T + 1 ) ,
where C is the normalizing constant of the conditional posterior density and
f ( h ˜ 1 : T + 1 ) = g ( h ˜ 1 : T + 1 ) ( h ˜ 1 : T + 1 h ˜ 1 : T + 1 ) 1 2 ( h ˜ 1 : T + 1 h ˜ 1 : T + 1 ) Q ( h ˜ 1 : T + 1 ) ( h ˜ 1 : T + 1 h ˜ 1 : T + 1 ) 1 2 τ 2 ( h ˜ 1 : T + 1 X β ) V ( h ˜ 1 : T + 1 X β ) ,
By completing the square in (A26), we have
f ( h ˜ 1 : T + 1 ) = 1 2 h ˜ 1 : T + 1 μ h ˜ ( h ˜ 1 : T + 1 ) Σ h ˜ ( h ˜ 1 : T + 1 ) 1 h ˜ 1 : T + 1 μ h ˜ ( h ˜ 1 : T + 1 ) + constant ,
where
Σ h ˜ ( h ˜ 1 : T + 1 ) = Q ( h ˜ 1 : T + 1 ) + 1 τ 2 V 1 , μ h ˜ ( h ˜ 1 : T + 1 ) = Σ h ˜ ( h ˜ 1 : T + 1 ) g ( h ˜ 1 : T + 1 ) + Q ( h ˜ 1 : T + 1 ) h ˜ 1 : T + 1 + 1 τ 2 V X β .
Therefore, the right-hand side of (A25) is approximately proportional to the pdf of the following normal distribution:
h ˜ 1 : T + 1 Normal μ h ˜ ( h ˜ 1 : T + 1 ) , Σ h ˜ ( h ˜ 1 : T + 1 ) ,
which we use as the proposal distribution in the MH algorithm. We obtain h ˜ 1 : T + 1 in (A28) with the same search algorithm as in the case of the NCP form and apply the block sampler to improve the acceptance rate in the MH algorithm.

Appendix A.2.2. Regression Coefficients β

By ignoring the terms that do not depend on β , we can rearrange the density of Y t in (15) as
p ( y t | h ˜ t , h ˜ t + 1 , β , θ β ) exp 1 2 ( y ˜ t x ˜ t β ) 2 ,
where
y ˜ t = y t exp ( h ˜ t ) γ ( h ˜ t + 1 ϕ h ˜ t ) , x ˜ t = γ ( x t + 1 ϕ x t ) .
By defining y ˜ = [ y ˜ 1 ; ; y ˜ T ] and X ˜ = [ x ˜ 1 ; ; x ˜ T ] , we have
p ( y 1 : T | h ˜ 1 : T + 1 , β , θ β ) = t = 1 T p ( y t | h ˜ t , h ˜ t + 1 , β , θ β ) exp 1 2 ( y ˜ X ˜ β ) ( y ˜ X ˜ β ) .
Then the conditional posterior distribution of β is given by
p ( β | h ˜ 1 : T + 1 , θ β , y 1 : T ) p ( y 1 : T | h ˜ 1 : T + 1 , β , θ β ) p ( h ˜ 1 : T + 1 | β , θ β ) p ( β ) exp [ 1 2 ( y ˜ X ˜ β ) ( y ˜ X ˜ β ) 1 2 τ 2 ( h ˜ 1 : T + 1 X β ) V ( h ˜ 1 : T + 1 X β ) 1 2 ( β μ ¯ β ) Ω ¯ β ( β μ ¯ β ) ] .
By completing the square, we have the conditional posterior distribution of β as
β Normal μ ˜ β , Σ ˜ β ,
where
Σ ˜ β = X ˜ X ˜ + 1 τ 2 X V X + Ω ¯ β 1 , μ ˜ β = Σ ˜ β X ˜ y ˜ + 1 τ 2 X V h ˜ 1 : T + 1 + Ω ¯ β μ ¯ β .
Since X ˜ X ˜ = γ 2 X V X ,
Σ ˜ β = γ 2 + 1 τ 2 X V X + Ω ¯ β 1 , μ ˜ β = Σ ˜ β X ˜ ϵ ˜ + γ 2 + 1 τ 2 X V h ˜ 1 : T + 1 + Ω ¯ β μ ¯ β ,
where ϵ ˜ = [ ϵ ˜ 1 ; ; ϵ ˜ T ] .

Appendix A.2.3. Leverage Parameter γ

Replacing ϵ t and η t in (19) with ϵ ˜ t and η ˜ t , respectively, we have
γ | h ˜ 1 : T + 1 , θ γ , y 1 : T Normal t = 1 T η ˜ t ϵ ˜ t + ω ¯ γ μ ¯ γ t = 1 T η ˜ t 2 + ω ¯ γ , 1 t = 1 T η ˜ t 2 + ω ¯ γ .

Appendix A.2.4. Variance τ2

It is straightforward to show that the conditional posterior distribution of τ 2 is
τ 2 | h ˜ 1 : T + 1 , θ τ 2 , y 1 : T Inv . Gamma T + 1 2 a τ , 1 2 ( h ˜ 1 : T + 1 X β ) V ( h ˜ 1 : T + 1 X β ) + b τ .

Appendix A.2.5. AR(1) Coefficient ϕ

Replacing h t in derivation of (A21) with h ˜ t x t β , we have
ϕ Normal t = 1 T ( h ˜ t + 1 x t + 1 β ) ( h ˜ t x t β ) t = 2 T ( h ˜ t x t β ) 2 , τ 2 t = 2 T ( h ˜ t x t β ) 2 1 < ϕ < 1 .
We use (A34) as the proposal distribution for ϕ in the MH algorithm.

References

  1. Abanto-Valle, C. A., V. H. Lachos, and Dipak K. Dey. 2015. Bayesian estimation of a skew-Student-t stochastic volatility model. Methodology and Computing in Applied Probability 17: 721–38. [Google Scholar] [CrossRef]
  2. Andersen, Torben G., and Tim Bollerslev. 1997. Intraday periodicity and volatility persistence in financial markets. Journal of Empirical Finance 4: 115–58. [Google Scholar] [CrossRef]
  3. Andersen, Torben G., and Tim Bollerslev. 1998. Answering the skeptics: Yes, standard volatility models do provide accurate forecasts. International Economic Review 39: 885–905. [Google Scholar] [CrossRef]
  4. Barndorff-Nielsen, Ole. 1977. Exponentially decreasing distributions for the logarithm of particle size. Proceedings of the Royal Society A. Mathematical, Physical and Engineering Sciences 353: 401–19. [Google Scholar]
  5. Berg, Andreas, Renate Meyer, and Jun Yu. 2004. DIC as a model comparison criterion for stochastic volatility models. Journal of Business and Economic Statistics 22: 107–20. [Google Scholar] [CrossRef]
  6. Chan, Joshua CC, and Ivan Jeliazkov. 2009. Efficient simulation and integrated likelihood estimation in state space models. International Journal of Mathematical Modelling and Numerical Optimisation 1: 101–20. [Google Scholar] [CrossRef]
  7. Chan, Kalok, Kakeung C. Chan, and G. Andrew Karolyi. 1991. Intraday volatility in the stock index and stock index futures markets. The Review of Financial Studies 4: 657–84. [Google Scholar] [CrossRef]
  8. Chib, Siddhartha. 2001. Markov chain Monte Carlo methods: Computation and inference. In Handbook of Econometrics. Edited by James J. Heckman and Edward E. Leamer. Amsterdam: North Holland, pp. 3569–649. [Google Scholar]
  9. Fičura, Milan, and Jiri Witzany. 2015a. Estimating Stochastic Volatility and Jumps Using High-Frequency Data and Bayesian Methods. Available online: https://ssrn.com/abstract=2551807 (accessed on 8 December 2020).
  10. Fičura, M., and J. Witzany. 2015b. Applications of Mathematics & Statistics in Economics. Available online: http://amse2015.cz/doc/Ficura_witzany.pdf (accessed on 8 December 2020).
  11. Gelman, Andrew, Jessica Hwang, and Aki Vehtari. 2014. Understanding predictive information criteria for Bayesian models. Statistics and Computing 24: 997–1016. [Google Scholar] [CrossRef]
  12. Ghysels, Eric, Andrew C. Harvey, and Eric Renault. 1996. Stochastic volatility. In Handbook of Statistics. Edited by Maddala G. S and Calyampudi Radhakrishna Rao. Amsterdam: Elsevier Science, pp. 119–91. [Google Scholar]
  13. Jacquier, Eric, Nicholas G. Polson, and Peter E. Rossi. 1994. Bayesian analysis of stochastic volatility models. Journal of Business & Economic Statistics 12: 371–89. [Google Scholar]
  14. Jacquier, Eric, Nicholas G. Polson, and Peter E. Rossi. 2004. Bayesian analysis of stochastic volatility models with fat-tail and correlated errors. Journal of Econometrics 122: 185–212. [Google Scholar] [CrossRef]
  15. Kastner, Gregor, and Sylvia Frühwirth-Schnatter. 2014. Ancillarity-sufficiency interweaving strategy (ASIS) for boosting MCMC estimation of stochastic volatility models. Computational Statistics & Data Analysis 76: 408–23. [Google Scholar]
  16. Liu, Jun S., and Chiara Sabatti. 2000. Generalised Gibbs sampler and multigrid Monte Carlo for Bayesian computation. Biometrika 87: 353–69. [Google Scholar] [CrossRef]
  17. McCausland, William J., Shirley Miller, and Denis Pelletier. 2011. Simulation smoothing for state-space models: A computational efficiency analysis. Computational Statistics & Data Analysis 55: 199–212. [Google Scholar]
  18. Mykland, Per A., and Lan Zhang. 2017. Assessment of Uncertainty in High Frequency Data: The Observed Asymptotic Variance. Econometrica 85: 197–231. [Google Scholar] [CrossRef] [Green Version]
  19. Nakajima, Jouchi. 2012. Bayesian analysis of generalized autoregressive conditional heteroskedasticity and stochastic volatility: Modeling leverage, jumps and heavy-tails for financial time series. Japanese Economic Review 63: 81–103. [Google Scholar] [CrossRef]
  20. Nakajima, Jouchi, and Yasuhiro Omori. 2009. Leverage, heavy-tails and correlated jumps in stochastic volatility models. Computational Statistics & Data Analysis 53: 2335–53. [Google Scholar]
  21. Nakajima, Jouchi, and Yasuhiro Omori. 2012. Stochastic volatility model with leverage and asymmetrically heavy-tailed error using GH skew Student’s t-distribution. Computational Statistics & Data Analysis 56: 3690–704. [Google Scholar]
  22. Omori, Yasuhiro, Siddhartha Chib, Neil Shephard, and Jouchi Nakajima. 2007. Stochastic volatility with leverage: Fast and efficient likelihood inference. Journal of Econometrics 140: 425–49. [Google Scholar] [CrossRef] [Green Version]
  23. Omori, Yasuhiro, and Toshiaki Watanabe. 2008. Block sampler and posterior mode estimation for asymmetric stochastic volatility models. Computational Statistics & Data Analysis 52: 2892–910. [Google Scholar]
  24. Rue, Håvard. 2001. Fast sampling of Gaussian Markov random fields. Journal of the Royal Statistical Society 63: 325–38. [Google Scholar] [CrossRef]
  25. Stroud, Jonathan R., and Michael S. Johannes. 2014. Bayesian modeling and forecasting of 24-hour high-frequency volatility. Journal of the American Statistical Association 109: 1368–84. [Google Scholar] [CrossRef]
  26. Taylor, Stephen John. 1982. Financial returns modelled by the product of two stochastic processes-A study of the daily sugar prices 1961–1975. In Time Series Analysis: Theory and Practice 1. Edited by Anderson O. D. Amsterdam: North Holland, pp. 203–26. [Google Scholar]
  27. Tsiotas, Georgios. 2012. On generalised asymmetric stochastic volatility models. Computational Statistics & Data Analysis 56: 151–72. [Google Scholar]
  28. Watanabe, Toshiaki. 2001. On sampling the degree-of-freedom of Student’s-t disturbances. Statistics & Probability Letters 52: 177–81. [Google Scholar]
  29. Watanabe, Sumio. 2010. Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory. Journal of Machine Learning Research 11: 3571–94. [Google Scholar]
  30. Virbickaitė, Audronė, Hedibert F. Lopes, M. Concepción Ausín, and Pedro Galeano. 2019. Particle learning for Bayesian semi-parametric stochastic volatility model. Econometric Reviews 38: 1007–23. [Google Scholar] [CrossRef]
  31. Yu, Yaming, and Xiao-Li Meng. 2011. To center or not to center: That is not the question—An ancillarity-sufficiency interweaving strategy (ASIS) for boosting MCMC efficiency. Journal of Computational and Graphical Statistics 20: 531–70. [Google Scholar] [CrossRef]
Figure 1. Standardized returns of the TOPIX data in the first week of June 2016.
Figure 1. Standardized returns of the TOPIX data in the first week of June 2016.
Jrfm 14 00145 g001
Figure 2. Standardized returns of the TOPIX data in the second week of June 2016.
Figure 2. Standardized returns of the TOPIX data in the second week of June 2016.
Jrfm 14 00145 g002
Figure 3. Standardized returns of the TOPIX data in the third week of June 2016.
Figure 3. Standardized returns of the TOPIX data in the third week of June 2016.
Jrfm 14 00145 g003
Figure 4. Standardized returns of the TOPIX data in the fourth week of June 2016.
Figure 4. Standardized returns of the TOPIX data in the fourth week of June 2016.
Jrfm 14 00145 g004
Figure 5. Standardized returns of the TOPIX data in the fifth week of June 2016.
Figure 5. Standardized returns of the TOPIX data in the fifth week of June 2016.
Jrfm 14 00145 g005
Figure 6. Intraday seasonality with Bernstein polynomial approximation.
Figure 6. Intraday seasonality with Bernstein polynomial approximation.
Jrfm 14 00145 g006
Table 1. Descriptive statistics of standardized TOPIX 1 min returns in June 2016.
Table 1. Descriptive statistics of standardized TOPIX 1 min returns in June 2016.
DateSkewnessKurtosisMin.Max
Week 1−0.10817.2296−7.25695.4520
Week 20.24947.7468−5.98865.7911
Week 30.35347.4500−6.44155.8413
Week 4−0.01257.1031−6.42125.6074
Week 50.03464.9146−4.64335.4437
Table 2. Widely applicable information criterion (WAIC) values of TOPIX returns (week 1).
Table 2. Widely applicable information criterion (WAIC) values of TOPIX returns (week 1).
Order 5Order 6Order 7Order 8Order 9Order 10
SV-N3965.13966.03962.03975.13969.63971.3
SV-G3701.03698.53702.23697.43700.03701.1
SV-SG3701.93701.03698.93699.03702.83700.8
SV-T3813.23813.33813.83816.63813.93813.0
SV-ST3819.33813.83816.53816.23815.73817.5
Note: Bold highlight means the best model according to WAIC.
Table 3. WAIC values of TOPIX returns (week 2).
Table 3. WAIC values of TOPIX returns (week 2).
Order 5Order 6Order 7Order 8Order 9Order 10
SV-N3811.03813.73814.53815.73818.03819.6
SV-G3621.03622.73621.03619.73621.23621.8
SV-SG3618.63621.33622.03623.73623.43619.7
SV-T3685.43684.13681.53859.03860.43684.6
SV-ST3686.93684.93683.33684.13684.33686.4
Note: Bold highlight means the best model according to WAIC.
Table 4. WAIC values of TOPIX returns (week 3).
Table 4. WAIC values of TOPIX returns (week 3).
Order 5Order 6Order 7Order 8Order 9Order 10
SV-N3976.93991.13996.83975.13969.63971.3
SV-G3750.23752.43752.03755.23752.93754.3
SV-SG3753.63753.33755.83754.13759.93752.8
SV-T3862.03859.43861.63859.03860.43861.6
SV-ST3862.13861.23861.33860.63861.53862.1
Note: Bold highlight means the best model according to WAIC.
Table 5. WAIC values of TOPIX returns (week 4).
Table 5. WAIC values of TOPIX returns (week 4).
Order 5Order 6Order 7Order 8Order 9Order 10
SV-N3927.53924.53924.83925.83924.73927.1
SV-G3670.33680.13679.03662.83675.73670.4
SV-SG3663.53668.23670.33664.63672.63669.1
SV-T3750.93751.93752.33750.83753.13753.9
SV-ST3753.53754.33753.43755.33752.83751.7
Note: Bold highlight means the best model according to WAIC.
Table 6. WAIC values of TOPIX returns (week 5).
Table 6. WAIC values of TOPIX returns (week 5).
Order 5Order 6Order 7Order 8Order 9Order 10
SV-N4114.64114.34111.34115.24113.74114.5
SV-G3961.23959.63958.33960.03961.23960.0
SV-SG3953.73955.63960.03963.33957.73953.1
SV-T4033.74033.84032.74031.24032.84033.1
SV-ST4032.54033.04033.44030.44032.04031.5
Note: Bold highlight means the best model according to WAIC.
Table 7. Estimation results for TOPIX returns (week 1).
Table 7. Estimation results for TOPIX returns (week 1).
γ ϕ τ α ν
SV-N (7) a -0.5742 b 0.89090.1903
[−1.4270, −0.1241] c [0.8214, 0.9431][0.1306, 0.2494]
3.42 d 4.014.54
SV-G (8)−1.35650.96080.0798 2.5248
[−3.1743, 0.8347][0.9320, 0.9815][0.0604, 0.1034] [2.0444, 3.2647]
4.133.174.45 3.40
SV-SG (7)−1.45200.95980.08120.00142.5489
[−3.2233, 0.4995][0.9316, 0.9802][0.0630, 0.1077][−0.0504, 0.0544][2.0425, 3.3462]
3.993.014.451.233.48
SV-T (10)−1.42090.98640.0766 4.2950
[−4.3298, 1.5726][0.9745, 0.9955][0.0594, 0.1010] [3.3294, 5.5097]
3.852.904.42 3.44
SV-ST (6)−1.61370.98700.07530.00064.2338
[−5.0449, 1.6958][0.9751, 0.9957][0.0586, 0.0962][−0.0400, 0.0407][3.3247, 5.5254]
3.942.784.381.623.54
a: the selected Bernstein polynomial order. b: posterior mean. c: 95% credible interval. d: inefficiency factor.
Table 8. Estimation results for TOPIX returns (week 2).
Table 8. Estimation results for TOPIX returns (week 2).
γ ϕ τ α ν
SV-N (5) a −0.2127 b 0.71490.3435
[−0.5550, 0.0923] c [0.5967, 0.8106][0.2750, 0.4187]
3.04 d 3.874.39
SV-G (8)0.03970.91910.0917 2.2475
[−1.7042, 2.0695][0.8122, 0.9694][0.0604, 0.1442] [2.0089, 2.7055]
4.274.054.58 2.6736
SV-SG (5)0.03800.89650.0991−0.00012.2622
[−1.5878, 1.9299][0.6761, 0.9664][0.0647, 0.1610][−0.0530, 0.0531][2.0102, 2.7308]
4.244.464.611.242.85
SV-T (7)−0.78620.9140.0673 3.30
[−4.7784, 2.8361][0.9825, 0.9979][0.0511, 0.0886] [2.7141, 4.0299]
3.912.834.43 3.17
SV-ST (7)−0.68620.99090.0690−0.00273.3647
[−4.4268, 2.9433][0.9816, 0.9976][0.0535, 0.0902][−0.0376, 0.0323][2.7611, 4.0958]
3.932.774.421.503.14
a: the selected Bernstein polynomial order. b: posterior mean. c: 95% credible interval. d: inefficiency factor.
Table 9. Estimation results for TOPIX returns (week 3).
Table 9. Estimation results for TOPIX returns (week 3).
γ ϕ τ α ν
SV-N (9) a 0.0514 b 0.62490.2950
[−0.3345, 0.4436] c [0.3506, 0.8190][0.2042, 0.3872]
2.71 d 4.394.54
SV-G (5)0.06390.45330.0919 2.2888
[−1.8134, 1.9573][0.1344, 0.7393][0.0632, 0.1413] [2.0155, 2.7501]
4.204.334.57 2.65
SV-SG (10)−0.20230.79920.0851−0.00312.3419
[−2.1985, 1.7331][0.2317, 0.9511][0.0596, 0.1250][−0.0546, 0.0485][2.0232, 2.9008]
4.204.594.551.172.99
SV-T (8)−0.23690.98710.0661 4.0539
[−3.6943, 3.4453][0.9755, 0.9960][0.0514, 0.0837] [3.2156, 5.1327]
3.862.834.39 3.39
SV-ST (8)−0.33130.98660.0670−0.00344.1237
[−4.2240, 3.6835][0.9730, 0.9957][0.0521, 0.0900][−0.0426, 0.0361][3.2114, 5.2538]
3.983.074.421.593.44
a: the selected Bernstein polynomial order. b: posterior mean. c: 95% credible interval. d: inefficiency factor.
Table 10. Estimation results for TOPIX returns (week 4).
Table 10. Estimation results for TOPIX returns (week 4).
γ ϕ τ α ν
SV-N (6) a −0.6502 b 0.93360.1526
[−1.8098, 0.2658] c [0.8831, 0.9704][0.1038, 0.2117]
3.53 d 4.014.57
SV-G (8)−1.44350.96890.0853 2.9487
[−3.6134, 0.6862][0.9454, 0.9859][0.0657, 0.1098] [2.1879, 3.9161]
4.113.104.43 3.63
SV-SG (5)−1.71460.96940.0846−0.00682.93
[−4.0432, −0.4361][0.9458, 0.9868][0.0650, 0.1116][−0.0599, 0.0471][2.14, 3.94]
4.183.114.461.543.64
SV-T (8)−1.40430.98690.0824 4.5805
[−4.4147, −1.6809][0.9747, 0.9960][0.0634, 0.1060] [3.5659, 5.8472]
3.962.894.41 3.43
SV-ST (10)−1.64820.98820.0788−0.00454.4738
[−5.1451, 1.6390][0.9777, 0.9964][0.0623, 0.0982][−0.0460, 0.0362][3.4944, 5.7573]
4.052.644.361.703.39
a: the selected Bernstein polynomial order. b: posterior mean. c: 95% credible interval. d: inefficiency factor.
Table 11. Estimation results for TOPIX returns (week 5).
Table 11. Estimation results for TOPIX returns (week 5).
γ ϕ τ α ν
SV-N (7) a −0.2602 b 0.85290.1428
[−1.5977, 0.8891] c [0.6617, 0.9395][0.0858, 0.2348]
3.47 d 4.384.62
SV-G (7)−1.11750.86380.0873 3.6578
[−3.7543, 1.4340][0.6609, 0.9457][0.0642, 0.1144] [2.4812, 5.1414]
4.214.244.45 3.86
SV-SG (10)−1.05620.82260.0828−0.00423.5728
[−4.0335, 2.0096][0.1974, 0.9485][0.0587, 0.1170][−0.0565, 0.0490][2.4627, 4.8738]
4.274.604.521.193.73
SV-T (8)−1.28530.97270.0730 6.2435
[−4.7632, 1.8139][0.9448, 0.9898][0.0547, 0.0998] [4.5496, 8.6440]
3.923.594.50 3.71
SV-ST (8)−1.57440.97260.0735−0.00846.2388
[−5.5186, 1.9908][0.9459, 0.9898][0.0566, 0.0992][−0.0536, 0.0378][4.6016, 8.6696]
4.043.564.461.843.74
a: the selected Bernstein polynomial order. b: posterior mean. c: 95% credible interval. d: inefficiency factor.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nakakita, M.; Nakatsuma, T. Bayesian Analysis of Intraday Stochastic Volatility Models of High-Frequency Stock Returns with Skew Heavy-Tailed Errors. J. Risk Financial Manag. 2021, 14, 145. https://doi.org/10.3390/jrfm14040145

AMA Style

Nakakita M, Nakatsuma T. Bayesian Analysis of Intraday Stochastic Volatility Models of High-Frequency Stock Returns with Skew Heavy-Tailed Errors. Journal of Risk and Financial Management. 2021; 14(4):145. https://doi.org/10.3390/jrfm14040145

Chicago/Turabian Style

Nakakita, Makoto, and Teruo Nakatsuma. 2021. "Bayesian Analysis of Intraday Stochastic Volatility Models of High-Frequency Stock Returns with Skew Heavy-Tailed Errors" Journal of Risk and Financial Management 14, no. 4: 145. https://doi.org/10.3390/jrfm14040145

APA Style

Nakakita, M., & Nakatsuma, T. (2021). Bayesian Analysis of Intraday Stochastic Volatility Models of High-Frequency Stock Returns with Skew Heavy-Tailed Errors. Journal of Risk and Financial Management, 14(4), 145. https://doi.org/10.3390/jrfm14040145

Article Metrics

Back to TopTop