Abstract

The estimation of the Smoothed Conditional Scale Function for time series was taken out under the conditional heteroscedastic innovations by imitating the kernel smoothing in nonparametric QAR-QARCH scheme. The estimation was taken out based on the quantile regression methodology proposed by Koenker and Bassett. And the proof of the asymptotic properties of the Conditional Scale Function estimator for this type of process was given and its consistency was shown.

1. Introduction

Consider a Quantile Autoregressive model,where is the Conditional Quantile Function of given and the innovation is assumed to be independent and identically distributed with zero quantile and constant scale function; see [1]. A kernel estimator of has been determined and its consistency is shown [2]. A bootstrap kernel estimator of was determined and shown to be consistent [3]. This research will extend [3] by assuming that the innovations follow Quantile Autoregressive Conditional Heteroscedastic process similar to Autoregressive-Quantile Autoregressive Conditional Heteroscedastic process proposed in [1]:where is the conditional -quantile function of given ; is a conditional scale function at -level, and is independent and identically distributed (i.i.d.) error with zero -quantile and unit scale. The function can be expressed aswhere is the so-called volatility found in [4, 5] which are papers of reference on Engle’s ARCH models among many others and is a positive constant depending on [see [6]]. An example of this kind of function is Autoregressive-Generalized Autoregressive Conditional Heteroscedastic AR(1)-GARCH(1,1),where , , , , , , , , and with 0 mean and variance 1. Note that may also be an ARMA (see [7]). The specifications for model (4) are given in Section 4.2.

Considering other financial time series models, the model (1) can be seen as a robust generalization of AR-ARCH-models, introduced in [7], and their nonparametric generalizations reviewed by [8]. For instance, consider a financial time series model of AR()-ARCH()-type,where and and are arbitrary functions representing, respectively, the conditional mean and conditional variance of the process.

The focus of this paper is to determine a smoothed estimator of the conditional scale function (CSF) and its asymptotic properties. This study is essential since volatility is inherent in many areas, for example, hydrology, finance, and weather. The volatility needs to be estimated robustly even when the moments of distribution do not exist.

A partitioned stationary -mixed time series , where the and the variate are, respectively, -measurable and -measurable, is considered. For some , the conditional -quantile of given the past assumed to be determined by is estimated. For simplicity, we assume that throughout the rest of the discussion.

We derive a smoothed nonparametric estimator of and show its consistency using standard estimate of Nadaraya [9]-Watson [10] type. This estimate is obtained from the estimate of the conditional scale function in [11] which is a type of estimator that has some disadvantages of not being adaptive and having some boundary effects but can be fixed by well-known techniques ([12]). It is though a constrained estimator in and a monotonically increasing function. This is very important to our estimation of the conditional distribution function and its inverse.

2. Methods and Estimations

Let and denote the probability density function (pdf) of and the joint pdf of . The dependence between the exogenous and the endogenous variables is described by the following conditional probability density function (CPDF):and the conditional cumulative distribution function (CCDF)The estimation of the conditional scale function is derived through the CCDF. However, the following assumptions and definitions (these assumptions are commonly used for kernel density estimation (KDE), bias reduction [13], asymptotic properties, and normality proof) are necessary (see Table 1).

Assumption 1. (i) and exist.(ii)For fixed , and are continuous in the neighborhood of where the estimator is to be estimated.(iii)The derivatives and for exist.(iv) is a convex function in for fixed .(v)The conditional density exists and is continuous in the neighborhood of .(vi).

Assumption 2. The kernel function is(i)Symmetrical: with (ii)Nonnegative and bounded: for , , (iii)Lipschitz: , such that for all (iv)A pdf: with .

Assumption 3. The process is strong mixing with , ; see [14, Theorem 1.7].

Assumption 4. The sequence of the smoothing parameters is such that , as and .

Definition 5 (strong mixing). Let be a stationary time series endowed with -algebras and . Define as If as , then the process is strong mixing.

The results in this section are about the case when the Autoregressive part of the model (4) for any . We therefore consider the modelDefine the check-function asHere, is the indicator function. Therefore, is a piece-wise monotone increasing function. is a function of any real random variable with distribution function , and a real value, , is the asymmetric absolute value function whose amount of asymmetry depends on ; see [15]. In case where is symmetric and , then we have the fact that is an absolute value function and is the conditional median absolute deviation (CMAD) of . When became 0 in model (5), we have a purely heteroscedastic ARCH model introduced in [16] and for , which, in this particular case, can be seen as a conditional scale function at -level.

The check-function in (10) is Lipschitz continuous by the following theorem.

Theorem 6. Let be defined as in (10) and . Then, satisfies the Lipschitz continuity condition:with the Lipschitz constant and for all ,

Proof of Theorem 6. See the proof of Lemma 3.1 in [1, p. 74-75].

By the next theorem we show clearly why the errors in model (2) are assumed to be zero -quantile and unit scale.

Theorem 7. Consider model (5) and the so-called check-function in (10); then, for ,is zero -quantile and unit scale. And the following equations are verifiable:

Proof of Theorem 7. The -quantile operator iswith well-defined properties in [1, p. 9-10]. From model (5), the conditional -quantile of iswhere is the -quantiles of . Then, using model (5) and (16), we getAnd the -quantile of (18) iswhere is the -quantile of . Note that, from (17), . The quotientis zero -quantile and unit scale and can be seen as model (2) if , , and
Now, assuming that (independent of ) in model (2) is zero -quantile, it is equivalent to writeThis proves (13) for . Also, is unit scale, which means

Assuming , the estimator, , of the conditional scale function, , is obtained through the minimization of the objective functionThus, the conditional scale function may be obtained by minimizing with respect to ; that is,The kernel estimator of (24) at is given byWe can express the estimate of in the random design as it was developed in [17]. Let be a nonnegative function of and a random vector in , . In the random design, the conditional expectation (23) can be rewritten as follows:where represents the conditional pdf of given , is the joint pdf of the two random variables and , and is the pdf of . Using [9, 10] with , a 1-dimensional rescaled kernel with bandwidth , we have the following estimates of and [18]:From the estimations above, , the estimate of , isand considering the regularity conditions of in Assumption 2 and also the fact that , , we havewhere is the estimate of the marginal pdf of at point and can be rewritten asand the derivative of with respect to isThe minimizer of (30) is obtained from . This leads to the following equation:wherefor all , . Note that in (27). The left part of (33) is a (unsmoothed) conditional cumulative distribution function (CCDF),that needs to be estimated and our estimator is thereforewhich is equivalent to .

An algorithm for estimating is proposed in the following section. This estimator suffers from the problem of boundary effects as we can see it on Figure 2 due to outliers. We obtain unsmoothed curves of the CCDF because the smoothness is only in the direction. A method is proposed by [19] to smooth it in the . The form of Smoothed Conditional Distribution Estimator iswhere is an integrated kernel with the smoothing parameter in the direction. This estimate is smooth rather than the NW which is a jump function in . To deal with boundary effects, one may think of the Weighted Nadaraya-Watson (WNW) estimate of the CDF discussed in [12, 20], [21, p. 3–18] among others. The WNW estimator’s expression iswith conditions and Lambda is determined using the Newton-Raphson iteration. Smoothing the CDF does not smooth the estimator in (36).

2.1. Algorithm

This algorithm estimates the empirical CCDF, , and its inverse . Starting with the estimation of the former, the denominator is easy to compute as the estimator of the probability density function of as vector of points .(1)Obtain , , for all .(2)Check if each is less than or equal to each observation of the whole sequence . The result determines which can be expressed in -matrix of order .(3)Construct from the sequence of i.i.d random variable with observation . is the number of from which the probability density function (pdf) of is to be estimated.(4)Determine the matrix of kernels which isThe row sums of over give the estimator of the pdf of at , , . We obtain the matrix of weights by the ration of and (element-wise), where is a matrix of ones. Note that the row sums of are 1.Let be the -matrix from 2. The estimator of the Conditional Cumulative Distribution Function (CCDF) is

2.2. Nadaraya-Watson Smoothing Method

We can make smooth by using NW regression (one can also use LOWESS (LOcally WEighted Scatter plot Smoother) regression introduced by [22] to smooth the estimator in (36) and it solves the problem of boundary effects). This will provide a smoothed curve at each level . We write the regression equation aswith and and the errors satisfy , , and for . Note that can be derived using joint pdf aswhere and are estimated as in (28).

We can perform some transformations on (42) in order to show that it is actually better than the unsmoothed one. By Assumption 1 (iv) and the fact that , we have We have used Jensen’s theorem for conditional expectation found in [23] and stated as follows.

Theorem 8 (Jensen’s inequality). For any convex function ,

Proof of Theorem 8. Suppose that is differentiable. The function is convex ifLet and . The inequality is true for all and taking its expectation on both sides proves the theorem.

This inequality is applicable when is a conditional convex function and when is a conditional expectation. The estimator is also element of the set to which the unsmoothed estimator belongs. This means that . The estimator is empirically given by

2.2.1. Asymptotic Properties

To show the asymptotic properties of our estimator, we compute its expectation and variance. Assuming the data is i.i.d, the expectation of the numerator is given byWe assume that the first and the second derivatives of at point exist. That is, by Taylor’s expansion of and given bywe getSimilarly, the expectation of the numerator isFor small enough, . Thus,The variance of the numerator, say , isNote that . Similarly, the variance of the denominator, , is .

The covariance of the numerator and the denominator of the estimator in (46) are given byThe variance of the estimator in (46) is the variance of a ratio of correlated variables that can be calculated using the approximation found in [24]:If Assumption 3 for strong mixing processes holds, then from the Central Limit Theorem (CLT) we have

2.3. Asymptotic Normality of QARCH

The CCDF in (35) can be written in the form of an arithmetic mean of a random variable :and the approximation of the expectation of is[see [24]]. Using the i.i.d assumption over the data, the numerator isWe have used the change of variables , the definition of the conditional density function turned into , and Fubuni’s theorem for multiple integrals. Taylor series expansions of and yieldand, for the denominator, we haveThus,From the assumption that , the denominator is approximated to . Hence,Some authors assumed that, in this case, the first derivative of the true pdf of at point can be zero [19] as the one for the fixed design and, therefore, the bias can be given byWe haveUsing the same approximation in (54), the variance of isand by the Central Limit Theorem, using Assumption 3 for ,Notice that the expectation of is the same as the one of and the variance is . To show the asymptotic normality of , we use the following theorem.

Theorem 9 (delta method). Suppose has the asymptotic normal distribution as in (67). Suppose is a continuous function that has a derivative at . Then

Proof of Theorem 9. The first-order Taylor expansion of about the point , and evaluated at the random variable , isand subtracting from both sides and multiplying by , we getwhich tends to in distribution.

For , thus, . In the next section, it is shown that the AMSE (Asymptotic Mean Squared Error) of is equal to which tends to 0 as and . This shows the consistency of the CCDF estimate, that is, , and we haveat points ’s that satisfy (36). Using again the first-order Taylor expansion, we also havefor ’s satisfying (36) and replacing by using the uniqueness assumption of , (68) becomeswith .

This result can be used to calculate the optimal bandwidth to compute the good estimation of the CSF.

3. Bandwidth Selections

3.1. Optimal Bandwidth for Density Estimations

In nonparametric estimations, specially in Kernel Density Estimations, computing a curve of an arbitrary function from the data without guessing the shape in advance requires an adequate choice of the smoothing parameter. The most used method is the “plug-in” method which consists of assigning a pilot bandwidth in order to estimate the derivatives of . We choose the bandwidth that minimizes the AMISE (Asymptotic Mean Integrated Squared Error) below.The general form of the derivatives of the AMISE with respect to was studied in [25], considering that the unknown functions in (74) are also functions of the smoothing parameter.The optimal smoothing parameter minimizing (75) isUsing this result, we came up with the optimal version of optimal bandwidth for CCDF. The aim of derivation of the AMISE in (74) is to get the optimal bandwidth for each directly. As an example, we consider the Epanechnikov Kernel function in order to compute , , and the efficiency of the kernel function given by . Epanechnikov’s kernel function is and its efficiency is measured bywhich is the smallest of all the other kernel functions.

3.2. Optimal Bandwidth for CCDF

The optimal bandwidth for the CCDF estimate is the one that minimizes the AMSE. It is shown below that the AMSE is actually the summation of the variance and the bias of the CCDF estimate. This is useful because the two are linked. When the variance is big, the bias also is big and when the variance is small, the bias is small.which is given by (66) and (64). Therefore,and leads toThis result is practically possible by estimating the unknown functions which are dependent on the smoothing parameter. is the second derivative of the CCDF from (35) at point . The estimator of the derivatives of (35) iswiththe function of weights. Thus, the first derivative is given byand the second derivative is alsowith and Note that the estimation of the CCDF is function of the estimation of the empirical pdf of . An optimal bandwidth that minimizes the AMISE of can also be the one that is optimal for the estimation of the CCDF.

Recent findings on the estimation of an optimal bandwidth for KDE (Kernel Density Estimation) are numerous ([2527]) but the estimation of an optimal smoothing parameter remains irksome due to computation issue and time consuming routines. To do so, we adopt what had been done by [27] to estimate the derivatives of the pdf of with respect to . We extend the idea to estimate the first and the second derivative of the CCDF with respect to .

4. Simulation Study

4.1. Model Specification

The ARCH() models introduced by [16] are widely used in financial applications. An AR(1)-ARCH(1) is a mixed model from an AR() and GARCH() for , , and . In time series, an observation at one time can be correlated with the observations in the previous time. That is,Note that the operator means the element-wise division between matrices.(5)For each row of , find the smallest such that .(6)The quantiles are the ’s which satisfy (36). This gives an unsmoothed estimator curve with bad shape at boundaries (see Figure 2).

The data to be simulated is given by , .(i)autoregressive process of order ,(ii)autoregressive ()-General Autoregressive Conditional Heteroscedastic process of order ,

4.2. Specifications for AR(1)-GARCH(1,1)
4.2.1. Unconditional Expectation

The unconditional expectation isNote that is used to ensure the stationarity of the process. That is, the expectation is therefore given by

4.2.2. Unconditional Variance

The unconditional variance of the model is given by the law of total varianceWe haveUsing the i.i.d. assumption on the sequence of random variables , the expected value of can be calculated as follows:which is independent of time. In another way,Equation (92) becomesWe obtainThe expectation of is given byIt follows thatand the variance in (91) becomesThis variance is positive and finite for , , , , , and .

4.3. Model Simulation

We simulated the data from (1) with , , for the AR(1) part and , for the ARCH(1) and . The data plot is represented by Figure 1.

Our algorithm gives the estimation of the conditional scale function which suffers from boundary effects as it is seen from Figure 2. This issue is recurrent while performing Kernel Density Estimations. The reason is that, at the boundaries, is underestimated because of the minimal number of points [28]. The consistency of our estimator is dependent on this problem of big variations at the boundaries. This increases the Average Squared Error between two different estimations from the same model.

4.4. Boundary Correction

To correct the boundary effects, we use the method of box-plot fences proposed by [29] to detect the extreme values that make the estimation too rough at the extremities of the CCDF estimations’ curves. Our estimator, being the inverse of the CCDF, is naturally rough at extremities. Among the Kernel functions, only the Gaussian can handle the sparseness of points at boundaries because its domain is . The other kernel functions can bring zero at extremities and make the estimation of the CCDF wrong. What we do is to omit the points that are extremely far from the others by the box-plot fences method. The method consist of determining the first and the third quantiles from the ’s. Outliers are the points that are located outside the intervalwhere and are the first and the third quantiles. Figure 3 is the representation of and the transformed response variable defined in (34) at level .

The gray points are outliers from (100). We lose some information by deleting them but we get the possibility of performing the estimation of a continuous curve of the CSF. Figure 4 is the estimations of the CSF at levels 0.25, 0.5 (median), 0.75, and 0.9. As we can see on the graphic, despite the optimal bandwidth for the empirical pdf of at point , we get unsmoothed curves at high level .

The curves represent the estimations of the CSF at from up to down. As it is seen in Figure 4, some curves are not smooth; that is why the NW method is discussed in Section 2.2 which requires that unsmoothed estimator and the bins . We obtain Figure 5 which combines the two estimations.

The next section discusses how precise is our estimation with the optimal bandwidth selection with the calculation of the MASE (Mean Average Squared Errors).

4.5. Consistency

The consistency of the estimator can be shown with the calculation of the Mean Average Squared Error providing the quantitative assessment of the accuracy of our estimator. This is a kind of bootstrap method to calculate the average gap between estimated CSFs. The formula isTable 2 shows that the estimator of the CSF is more precise at level for both the smoothed and the LOWESS versions.

5. Conclusion

We have derived an estimator for the conditional scale function in an AR(1)-GARCH(1) and despite the heavy-tail of the data, we could deal with the boundary effect and were able to show the consistency of the estimator through a Monte Carlo study. We assumed that the QAR(1) is known and is zero and, along with the regularity assumptions, we derived the estimator which can be improved in some next papers. The very next paper will focus on the estimation when the QAR(1) is unknown.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors thank the Pan-African University of Basic Sciences, Technology and Innovation (PAUISTI) for funding this research.