- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Journal of Probability and Statistics

VolumeΒ 2011Β (2011), Article IDΒ 718647, 39 pages

http://dx.doi.org/10.1155/2011/718647

## Estimation and Properties of a Time-Varying GQARCH(1,1)-M Model

Athens University of Economics and Business, Department of International and European Economic Studies, 10434 Athens, Greece

Received 16 May 2011; Accepted 14 July 2011

Academic Editor: MikeΒ Tsionas

Copyright Β© 2011 Sofia Anyfantaki and Antonis Demos. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

Time-varying GARCH-M models are commonly used in econometrics and financial economics. Yet the recursive nature of the conditional variance makes exact likelihood analysis of these models computationally infeasible. This paper outlines the issues and suggests to employ a Markov chain Monte Carlo algorithm which allows the calculation of a classical estimator via the simulated EM algorithm or a simulated Bayesian solution in only computational operations, where is the sample size. Furthermore, the theoretical dynamic properties of a time-varying GQARCH(1,1)-M are derived. We discuss them and apply the suggested Bayesian estimation to three major stock markets.

#### 1. Introduction

Time series data, emerging from diverse fields appear to possess time-varying second conditional moments. Furthermore, theoretical results seem to postulate quite often, specific relationships between the second and the first conditional moment. For instance, in the stock market context, the first conditional moment of stock marketβs excess returns, given some information set, is a possibly time-varying, linear function of volatility (see, e.g., Merton [1], Glosten et al. [2]). These have led to modifications and extensions of the initial ARCH model of Engle [3] and its generalization by Bollerslev [4], giving rise to a plethora of dynamic heteroscedasticity models. These models have been employed extensively to capture the time variation in the conditional variance of economic series, in general, and of financial time series, in particular (see Bollerslev et al. [5] for a survey).

Although the vast majority of the research in conditional heteroscedasticity is being processed aiming at the stylized facts of financial stock returns and of economic time series in general, Arvanitis and Demos [6] have shown that a family of time-varying GARCH-M models can in fact be consistent with the sample characteristics of time series describing the temporal evolution of velocity changes of turbulent fluid and gas molecules. Despite the fact that the latter statistical characteristics match in a considerable degree their financial analogues (e.g., leptokurtosis, volatility clustering, and quasi long-range dependence in the squares are common), there are also significant differences in the behavior of the before mentioned physical systems as opposed to financial markets (examples are the anticorrelation effect and asymmetry of velocity changes in contrast to zero autocorrelation and the leverage effect of financial returns) (see Barndorff-Nielsen and Shephard [7] as well as Mantegna and Stanley [8, 9]). It was shown that the above-mentioned family of models can even create anticorrelation in the means as far as an AR(1) time-varying parameter is introduced.

It is clear that from an econometric viewpoint it is important to study how to efficiently estimate models with partially unobserved GARCH processes. In this context, our main contribution is to show how to employ the method proposed in Fiorentini et al. [10] to achieve MCMC likelihood-based estimation of a time-varying GARCH-M model by means of feasible algorithms, where is the sample size. The crucial idea is to transform the GARCH model in a first-order Markovβs model. However, in our model, the error term enters the in-mean equation multiplicatively and not additively as it does in the latent factor models of Fiorentini et al. [10]. Thus, we show that their method applies to more complicated models, as well.

We prefer to employ a GQARCH specification for the conditional variance (Engle [3] and Sentana [11]) since it encompasses all the existing restricted quadratic variance functions (e.g., augmented ARCH model), its properties are very similar to those of GARCH models (e.g., stationarity conditions) but avoids some of their criticisms (e.g., very easy to generalize to multivariate models). Moreover, many theories in finance involve an explicit tradeoff between the risk and the expected returns. For that matter, we use an in-mean model which is ideally suited to handling such questions in a time series context where the conditional variance may be time varying. However, a number of studies question the existence of a positive mean/variance ratio directly challenging the mean/variance paradigm. In Glosten et al. [2] when they explicitly include the nominal risk free rate in the conditioning information set, they obtain a negative ARCH-M parameter. For the above, we allow the conditional variance to affect the mean with a possibly time varying coefficient which we assume for simplicity that it follows an AR(1) process. Thus, our model is a time-varying GQARCH-M-AR(1) model.

As we shall see in Section 2.1, this model is able to capture the, so-called, stylized facts of excess stock returns. These are (i) the sample mean is positive and much smaller than the standard deviation, that is, high coefficient of variation, (ii) the autocorrelation of excess returns is insignificant with a possible exception of the 1st one, (iii) the distribution of returns is nonnormal mainly due to excess kurtosis and may be asymmetry (negative), (iv) there is strong volatility clustering, that is, significant positive autocorrelation of squared returns even for high lags, and (v) the so-called leverage effect; that is, negative errors increase future volatility more than positive ones of the same size.

The structure of the paper is as follows. In Section 2, we present the model and derive the theoretical properties the GQARCH(1,1)-M-AR(1) model. Next, we review Bayesian and classical likelihood approaches to inference for the time-varying GQARCH-M model. We show that the key task (in both cases) is to be able to produce consistent simulators and that the estimation problem arises from the existence of two unobserved processes, causing exact likelihood-based estimations computationally infeasible. Hence, we demonstrate that the method proposed by Fiorentini et al. [10] is needed to achieve a first-order Markovβs transformation of the model and thus, reducing the computations from to . A comparison of the efficient ( calculations) and the inefficient ( ones) simulator is also given. An illustrative empirical application on weekly returns from three major stock markets is presented in Section 4, and we conclude in Section 5.

#### 2. GQARCH(1,1)-M-AR(1) Model

The definition of our model is as follows.

*Definition 2.1. *The time-varying parameter GQARCH(1,1)-M-AR(1) model:
where
, , , are independent for all, and where are the observed excess returns, is the sample size, is an unobserved AR(1) process independent (with ) of , and is the conditional variance (with equal to the unconditional variance and ) which is supposed to follow a GQARCH(1,1). It is obvious that is the market price of risk (see, e.g., Merton [1] Glosten at al. [2]). Let us call the sequence of natural filtrations generated by the past values of and .

Modelling the theoretical properties of this model has been a quite important issue. Specifically, it would be interesting to investigate whether this model can accommodate the main stylized facts of the financial markets. On other hand, the estimation of the model requires its transformation into a first-order Markovβs model to implement the method of Fiorentini et al. [10]. Let us start with the theoretical properties.

##### 2.1. Theoretical Properties

Let us consider first the moments of the conditional variance , needed for the moments of . The proof of the following lemma is based on raising to the appropriate power, in (2.3), and taking into account that , and .

Lemma 2.2. *If , the first four moments of the conditional variance of (2.3) are given by
*

Consequently, the moments of are given in the following theorem taken from Arvanitis and Demos [6].

Theorem 2.3. *The first two moments of the model in (2.1), (2.2), and (2.3) are given by
**
whereas the skewness and kurtosis coefficients are
**
where
**
and , , ,and are given in Lemma 2.2.*

In terms of stationarity, the process is 4th-order stationary if and only if These conditions are the same as in Arvanitis and Demos [6], indicating that the presence of the asymmetry parameter, , does not affect the stationarity conditions (see also Bollerslev [4] and Sentana [11]). Furthermore, the 4th-order stationarity is needed as we would like to measure the autocorrelation of the squared βs (volatility clustering), as well as the correlation of and (leverage effect). The dynamic moments of the conditional variance and those between the conditional variance and the error are given in the following two lemmas (for a proof see Appendix B).

Lemma 2.4. *Under the assumption of Lemma 2.2, one has that
**
where and .*

Lemma 2.5. *
where and are given in Lemma 2.4.*

Furthermore, from Arvanitis and Demos [6] we know that the following results hold.

Theorem 2.6. *The autocovariance of returns for the model in (2.1)β(2.3) is given by
**
and the covariance of squaresβ levels and the autocovariance of squares are
**
where all needed covariances and expectations of the right-hand sizes are given in Lemmas 2.4 and 2.5,
*

From the above theorems and lemmas it is obvious that our model can accommodate all stylized facts. For example, negative asymmetry is possible; volatility clustering and the leverage effect (negative ) can be accommodated, and so forth. Furthermore, the model can accommodate negative autocorrelations, , something that is not possible for the GARCH-M model (see Fiorentini and Sentana [12]). Finally, another interesting issue is the diffusion limit of the time-varying GQARCH-M process. As already presented by Arvanitis [13], the weak convergence of the Time-varying GQARCH(1,1)-M coincides with the general conclusions presented elsewhere in the literature. These are that weak limits of the endogenous volatility models are exogenous (stochastic volatility) continuous-time processes. Moreover, Arvanitis [13] suggests that there is a distributional relation between the GQARCH model and the continuous-time Ornstein-Uhlenbeck models with respect to appropriate nonnegative Levyβs processes.

Let us turn our attention to the estimation of our model. We will show that estimating our model is a hard task and the use of well-known methods such as the EM-algorithm cannot handle the problem due to the huge computational load that such methods require.

#### 3. Likelihood-Inference: EM and Bayesian Approaches

The purpose of this section is the estimation of the time-varying GQARCH(1,1)-M model. Since our model involves two unobserved components (one from the time-varying in-mean parameter and one from the error term), the estimation method required is an EM and more specifically a simulated EM (SEM), as the expectation terms at the step cannot be computed. The main modern way of carrying out likelihood inference in such situations is via a Markov chain Monte Carlo (MCMC) algorithm (see Chib [14] for an extensive review). This simulation procedure can be used either to carry out Bayesian inference or to classically estimate the parameters by means of a simulated EM algorithm.

The idea behind the MCMC methods is that in order to sample a given probability distribution, which is referred to as the target distribution, a suitable Markov chain is constructed (using a Metropolis-Hasting (M-H) algorithm or a Gibbs sampling method) with the property that its limiting, invariant distribution is the target distribution. In most problems, the target distribution is absolutely continuous, and as a result the theory of MCMC methods is based on that of the Markov chains on continuous state spaces [15]. This means that by simulating the Markov chain a large number of times and recording its values a sample of (correlated) draws from the target distribution can be obtained. It should be noted that the Markov chain samplers are invariant by construction, and, therefore, the existence of the invariant distribution does not have to be checked in any particular application of MCMC method.

The Metropolis-Hasting algorithm (M-H) is a general MCMC method to produce sample variates from a given multivariate distribution. It is based on a candidate generating density that is used to supply a proposal value that is accepted with probability given as the ratio of the target density times the ratio of the proposal density. There are a number of choices of the proposal density (e.g., random walk M-H chain, independence M-H chain, tailored M-H chain) and the components may be revised either in one block or in several blocks. Another MCMC method, which is special case of the multiple block M-H method with acceptance rate always equal to one, is called the Gibbs sampling method and was brought into statistical prominence by Gelfand and Smith [16]. In this algorithm, the parameters are grouped into blocks, and each block is sampled according to the full conditional distribution denoted as . By Bayes' theorem, we have , the joint distribution of all blocks, and so full conditional distributions are usually quite simply derived. One cycle of the Gibbs sampling algorithm is completed by simulating , where is the number of blocks, from the full conditional distributions, recursively updating the conditioning variables as one moves through each distribution. Under some general conditions, it is verified that the Markov chain generated by the M-H or the Gibbs sampling algorithm converges to the target density as the number of iterations becomes large.

Within the Bayesian framework, MCMC methods have proved very popular, and the posterior distribution of the parameters is the target density (see [17]). Another application of the MCMC is the analysis of hidden Markovβs models where the approach relies on augmenting the parameter space to include the unobserved states and simulate the target distribution via the conditional distributions (this procedure is called data augmentation and was pioneered by Tanner and Wong [18]). Kim et al. [19] discuss an MCMC algorithm of the stochastic volatility (SV) model which is an example of a state space model in which the state variable (log-volatility) appears non-linearly in the observation equation. The idea is to approximate the model by a conditionally Gaussian state space model with the introduction of multinomial random variables that follow a seven-point discrete distribution.

The analysis of a time-varying GQARCH-M model becomes substantially complicated since the log-likelihood of the observed variables can no longer be written in closed form. In this paper, we focus on both the Bayesian and the classical estimation of the model. Unfortunately, the non-Markovian nature of the GARCH process implies that each time we simulate one error we implicitly change all future conditional variances. As pointed out by Shephard [20], a regrettable consequence of this path dependence in volatility is that standard MCMC algorithms will evolve in computational load (see [21]). Since this cost has to be borne for each parameter value, such procedures are generally infeasible for large financial datasets that we see in practice.

##### 3.1. Estimation Problem: Simulated EM Algorithm

As mentioned already, the estimation problem arises because of the fact that we have two unobserved processes. More specifically, we cannot write down the likelihood function in closed form since we do not observe both and . On the other hand, the conditional log-likelihood function of our model assuming that were observed would be the following: where , and .

However, the βs are unobserved, and, thus, to classically estimate the model, we have to rely on an EM algorithm [22] to obtain estimates as close to the optimum as desired. At each iteration, the EM algorithm obtains , where is the parameter vector, by maximizing the expectation of the log-likelihood conditional on the data and the current parameter values, that is, with respect to keeping fixed.

The step, thus, requires the expectation of the complete log-likelihood. For our model, this is given by:

It is obvious that we cannot compute such quantities. For that matter, we may rely on a simulated EM where the expectation terms are replaced by averages over simulations, and so we will have an SEM or a simulated score. The SEM log-likelihood is: Consequently, we need to obtain the following quantities: , , , ,ββ and , , where is the number of simulations.

Thus, to classically estimate our model by using an SEM algorithm, the basic problem is to sample from where is the vector of the unknown parameters and also sample from .

In terms of identification, the model is not, up to second moment, identified (see Corollary 1 in Sentana and Fiorentini [23]). The reason is that we can transfer unconditional variance from the error, , to the price of risk, , and vice versa. One possible solution is to fix such that is 1 or to set to a specific value. In fact in an earlier version of the paper, we fixed to be 1 (see Anyfantaki and Demos [24]). Nevertheless, from a Bayesian viewpoint, the lack of identification is not too much of a problem, as the parameters are identified through their proper priors (see Poirier [25]).

Next, we will exploit the Bayesian estimation of the model, and, since we need to resort to simulations, we will show that the key task is again to simulate from .

##### 3.2. Simulation-Based Bayesian Inference

In our problem, the key issue is that the likelihood function of the sample is intractable which precludes the direct analysis of the posterior density . This problem may be overcome by focusing instead on the posterior density of the model using Bayesβ rule: where

Now,

On the other hand, is the full-information likelihood. Once we have the posterior density, we get the parameters' marginal posterior density by integrating the posterior density. MCMC is one way of numerical integration.

The Hammersley-Clifford theorem (see Clifford [26]) says that a joint distribution can be characterized by its complete conditional distribution. Hence, given initial values , we draw from and then from . Iterating these steps, we finally get , and under mild conditions it is shown that the distribution of the sequence converges to the joint posterior distribution .

The above simulation procedure may be carried out by first dividing the parameters into two blocks: Then the algorithm is described as follows.(1)Initialize .(2)Draw from .(3)Draw from in the following blocks:(i)draw from using the Gibbs sampling. This is updated in one block;(ii)draw from by M-H. This is updated in a second block.(4)Go toββ(2).

We review the implementation of each step.

###### 3.2.1. Gibbs Sampling

The task of simulating from an AR model has been already discussed. Here, we will follow the approach of Chib [27], but we do not have any MA terms which makes inference simpler.

Suppose that the prior distribution of is given by: which means that is a priory independent of .

Also the following holds for the prior distributions of the parameter subvector : where ensures that lies outside the unit circle, is the inverted gamma distribution, and the hyperparameters have to be defined.

Now, the joint posterior is proportional to From a Bayesian viewpoint, the right-hand side of the above equation is equal to the βaugmentedβ prior, that is, the prior augmented by the latent (We would like to thank the associate editor for bringing this to our attention.) We proceed to the generation of these parameters.

*Generation of *

First we see how to generate . Following again Chib [27], we may write
or, otherwise,
Under the above and using Chib's (1993) notation, we have that the proposal distribution is the following Gaussian distribution (see Chib [27] for a proof).Proposition 3.1. *The proposal distribution of is
**
where
*

Hence, the generation of is completed, and we may turn on the generation of the other parameters.

*Generation of *

For the generation of and using [27] notation, we have the following.

Proposition 3.2. *The proposal distribution of is
**
where
*

Finally, we turn on the generation of .

*Generation of *

For the generation of , we follow again Chib [27] and write
We may now state the following proposition (see Chib [27] for a proof).

Proposition 3.3. *The proposal distribution of is
**
where
*

The Gibbs sampling scheme has been completed, and the next step of the algorithm requires the generation of the conditional variance parameters via an M-H algorithm which is now presented.

###### 3.2.2. Metropolis-Hasting

Step (3)-(ii) is the task of simulating from the posterior of the parameters of a GQARCH-M process. This has been already addressed by Kim et al. [19], Bauwens and Lubrano [28], Nakatsuma [29], Ardia [30] and others.

First, we need to decide on the priors. For the parameters , we use normal densities as priors: and similarly where are the indicators ensuring the constraints and , respectively. are the hyperparameters.

We form the joint prior by assuming prior independence between , and the joint posterior is then obtained by combining the joint prior and likelihood function by Bayes' rule:

For the M-H algorithm, we use the following approximated GARCH model as in Nakatsuma [29] which is derived by the well-known property of GARCH models [4]: where with .

Then the corresponding approximated likelihood is written as and the generation of is based on the above likelihood where we update each time after the corresponding parameters are updated. The generation of the four variance parameters is given.

*Generation of *

For the generation of , we first note that in (3.32), below, can be written as a linear function of :
where with

Now, let the two following vectors be

Then the likelihood function of the approximated model is rewritten as

Using this we have the following proposal distribution of (see Nakatsuma [29] or Ardia [30] for a proof).

Proposition 3.4. *The proposal distribution of is
**
where , , and . imposes the restriction that and . **Hence a candidate is sampled from this proposal density and accepted with probability:
**
where is the previous draw.*

Similar procedure is used for the generation of and .

*Generation of *

Following Nakatsuma [29], we linearize by the first-order Taylor expansion
where is the first-order derivative of evaluated at the previous draw of the M-H sampler.

Define as
where which is computed by the recursion:
for [30]. Then,

Let the following two vectors be

Then, the likelihood function of the approximated model is rewritten as

We have the following proposal distribution for (for a proof see Nakatsuma [29] or Ardia [30]).

Proposition 3.5. *The proposal distribution for is
**
where , , and . imposes the restriction that and . **Hence, a candidate is sampled from this proposal density and accepted with probability:
*

Finally, we explain the generation of .

*Generation of *

As with , we linearize by a first-order Taylor, expansion at a point the previous draw in the M-H sampler. In this case,
where which is computed by the recursion:
and for .

Then

Let again
and the likelihood function of the approximated model is rewritten as

Thus, we have the following proposal distribution for (for proof see Nakatsuma [29] and Ardia [30]).

Proposition 3.6. *The proposal distribution of is
**
where , , and . A candidate is sampled from this proposal density and accepted with probability:
*

The algorithm described above is a special case of a MCMC algorithm, which converges as it iterates, to draws from the required density . Posterior moments and marginal densities can be estimated (simulation consistently) by averaging the relevant function of interest over the sample variates. The posterior mean of is simply estimated by the sample mean of the simulated values. These estimated values can be made arbitrarily accurate by increasing the simulation sample size. However, it should be remembered that sample variates from an MCMC algorithm are a high dimensional (correlated) sample from the target density, and sometimes the serial correlation can be quite high for badly behaved algorithms.

All that remains, therefore, is step (2). Thus, from the above, it is seen that the main task is again as with the classical estimation of the model, to simulate from .

###### 3.2.3. MCMC Simulation of

For a given set of parameter values and initial conditions, it is generally simpler to simulate for and then compute than to simulate directly. For that matter, we concentrate on simulators of given and . We set the mean and the variance of equal to their unconditional values, and, given that is a sufficient statistic for and the unconditional variance is a deterministic function of , can be eliminated from the information set without any information loss.

Now sampling from is feasible by using an M-H algorithm where we update each time only one leaving all the other unchanged [20]. In particular, let us write the th iteration of a Markov chain as . Then we generate a potential new value of the Markov chain by proposing from some candidate density where which we accept with probability

If it is accepted then, we set and otherwise we keep . Although the proposal is much better since it is only in a single dimension, each time we consider modifying a single error we have to compute: where for , while

Nevertheless, each time we revise one , we have also to revise conditional variances because of the recursive nature of the GARCH model which makes depend upon for . And since , it is obvious that we need to calculate normal densities, and so this algorithm is . And this should be done for every . To avoid this huge computational load, we show how to use the method proposed by Fiorentini et al. [10] and so do MCMC with only calculations. The method is described in the following subsection.

##### 3.3. Estimation Method Proposed: Classical and Bayesian Estimation

The method proposed by Fiorentini et al. [10] is to transform the GARCH model into a first-order Markovβs model and so do MCMC with only calculations. Following their transformation, we augment the state vector with the variables and then sample the joint Markov process where so that with probability one. The mapping is one to one and has no singularities. More specifically, if we know and , then we know the value of

Hence the additional knowledge of the signs of would reveal the entire path of so long as (which equals the unconditional value in our case) is known, and, thus, we may now reveal also the unobserved random variable .

Now we have to sample from where the second and the third term come from the model, and the first comes from the fact that but takes values where

From the above, it is seen that we should first simulate since we do not alter the volatility process when we flip from to (implying that the signs do not cause the volatility process), but we do alter and then simulate . The second step is a Gibbs sampling scheme whose acceptance rate is always one and also conditional on the elements of are independent which further simplifies the calculations. We prefer to review first the Gibbs sampling scheme and then the simulation of the conditional variance.

###### 3.3.1. Simulations of ,

First, we see how to sample from . To obtain the required conditionally Bernoulli distribution, we establish first some notation. We have the following (see Appendix A):

Using the above notation, we see that the probability of drawing conditional on is equal to the probability of drawing conditional on , where is given by (3.70), which is given by

Similarly for the probability of drawing . Both of these quantities are easy to compute; for example, and so we may simulate using a Gibbs sampling scheme. Specifically, since conditional on the elements of are independent, we actually draw from the marginal distribution, and the acceptance rate for this algorithm is always one.

The Gibbs sampling algorithm for drawing may be described as below.(1)Specify an initial value .(2)Repeat for .(a)Repeat for .(i)Draw with probability, and with probability, (3)Return the values .

###### 3.3.2. Simulations of , (Single Move Samplers)

On the other hand, the first step involves simulating from . To avoid large dependence in the chain, we use an M-H algorithm where we simulate one at a time leaving the others unchanged [20, 31]. So if is the current value of the th iteration of a Markov chain, then we draw a candidate value of the Markov chain by proposing it from a candidate density (proposal density) where . We set with acceptance probability where we have used the fact that

However, we may simplify further the acceptance rate. More specifically, we have that

Now, since the following should hold: and similarly we have the support of the conditional distribution of given that is bounded from below by , and the same applies to the distribution of given (lower limit corresponds to and the upper limit to ). This means that the range of values of compatible with and in the GQARCH case is bounded from above and below; that is:

From the above, we understand that it makes sense to make the proposal to obey the support of the density, and so it is seen that we can simplify the acceptance rate by setting appropriately truncated from above (since the truncation from below will automatically be satisfied). But the above proposal density ignores the information contained in , and so according to Fiorentini et al. [10] we can achieve a substantially higher acceptance rate if we propose from

A numerically efficient way to simulate from is to sample an underlying Gaussian random variable doubly truncated by using an inverse transform method. More specifically, we may draw doubly truncated so that it remains within the following bounds: where using an inverse transform method and then compute which in turn implies a real value for and so guarantees that lies within the acceptance bounds.

The inverse transform method to draw the doubly truncated Gaussian random variable first draws a uniform random number and then computes the following:

A draw is then given by

However, if the bounds are close to each other (the degree of truncation is small) the extra computations involved make this method unnecessarily slow, and so we prefer to use the accept-reject method where we draw and accept the draw if , and otherwise we repeat the drawing (this method is inefficient if the truncation lies in the tails of the distribution). It may be worth assessing the degree of truncation first, and, depending on its tightness, choose one simulation method or the other.

The conditional density of will be given according to the definition of a truncated normal distribution: where is the cdf of the standard normal.

By using the change of variable formula, we have that the density of will be

Using Bayes theorem we have that the acceptance probability will be

Since the degree of truncation is the same for old and new, the acceptance probability will be where is a mixture of two univariate normal densities so

Hence, and the acceptance probability becomes where

Overall the MCMC of includes the following steps.(1)Specify an initial value .(2)Repeat for .(a)Repeat for .(i)Use an inverse transform method to simulate doubly truncated.(ii)Calculate Steps (2)(a)(i) and (2)(a)(ii) are equivalent to draw appropriately truncated.(iii)Calculate (iv)Set

*Remark 3.7. *Every time we change , we calculate only one normal density since the transformation is Markovian, and, since , we need calculations.

Notice that if we retain , then is retained and we will not need to simulate at a later stage. In fact we only need to simulate at since we need to know . The final step involves computing