Abstract

In this paper, we consider a general nonparametric regression estimation model with the feature of having multiplicative noise. We propose a linear estimator and nonlinear estimator by wavelet method. The convergence rates of those regression estimators under pointwise error over Besov spaces are proved. It turns out that the obtained convergence rates are consistent with the optimal convergence rate of pointwise nonparametric functional estimation.

1. Introduction

The regression estimation plays important roles in practical applications. The classical regression estimation model considers a strictly stationary random process , which defined on and has a common density function . Then, a regression function is defined by where be a known function and stands for the density function of random variable . This classical model is aimed at estimating the unknown regression function by the data .

However, the above data is not easily observed in practical applications. In other words, the observed data contains some noise. In view of this case, this paper considers a regression estimation model with multiplicative noise. Let be a strictly stationary random process with a common density function

where be a multiplicative noise function, stands for the density function of the unobserved random variable , and . Then, we want to estimate the unknown regression function based on the observed noisy data .

This paper is aimed at considering the pointwise error of wavelet regression estimators under some mild assumptions. Chaubey et al. [1, 2] construct linear and nonlinear wavelet estimators based on dependent data and evaluate its rate of convergence under the mean integrated square error, respectively. The convergence rates of wavelet estimators in independent case are discussed by Chesneau and Shirazi [3]. An optimal convergence rate of regression estimator in independent case is proved by Kou and Liu [4]. Guo and Kou [5] consider wavelet estimations for negatively associated case. The mean integrated square error of regression derivative estimators based on strong mixing data are discussed by Kou et al. [6]. But those above results all only focus on global error. There is a lack of result on pointwise error for the regression model with multiplicative noise ((1) and (2)). Hence, this paper will study the convergence rates of wavelet estimators under pointwise risk.

1.1. Strong Mixing and Assumptions

In many practical applications, the observed data is usual not independent. This paper will consider strong mixing case, which is an important dependent relationship in time series. Now we will introduce the definition of strong mixing in the following.

Definition 1. Let be a strict stationary sequence of random vectors and where denotes the field generated by and does by . Then, the sequence is strong mixing.

Note that when the random sample is independent, and . Hence, the independent and identically distributed data must be strong mixing. Gorodetskii [7] shows that linear processes are strong mixing under certain conditions. Genon-Catahot et al. [8] prove that some continuous time diffusion models and stochastic volatility models are strong mixing, which include some most popular models in the pricing theories of financial assets, such as the Black Scholes pricing theory of options. Moreover, strong mixing also naturally appears in a large class of stationary time series such as ARMA and GARCH models.

Next, we formulate some assumptions for the above regression model. Note that the assumptions C1 and C2 are the standard conditions for nonparametric regression estimation model with multiplicative noise [1, 3]. C3 is a technical condition, which plays an important role in the proof of some lemmas. For the assumptions C4 and C5, the functions and when the random sample is independent. Hence, C4 and C5 may be considered as other explanation of strong mixing.

C1. The density function satisfies with a positive constant .

C2. For any , there exist two positive constant such that

C3. The function satisfies .

C4. The strong mixing coefficient of satisfies with and .

C5. Let be the density function of and be the density function of . Then, for , with and .

1.2. Wavelets and Besov Space

In this paper, we construct regression estimators by wavelet method. Now let us briefly present the wavelet basis of by tensor product method. For a scaling function , there is wavelet function such that for each , holds in sense, where , , and

If a scaling function satisfies Condition , i.e., , then the functions and converge absolutely almost everywhere. In this paper, we use the Daubechies wavelet, which has compact supports and satisfies Condition . Then, it can be shown that for , holds almost everywhere on .

Besov spaces are important in theory and applications, which contain Hölder and lebesgue spaces as special examples [911]. The next lemma provides equivalent definitions of Besov space [12].

Lemma 2. Let be a scaling function, be the corresponding wavelet function, and . For and , then the following assertions are equivalent: The Besov norm of can be defined by with and .
In this paper, we assume that the regression function belongs to Besov ball with , i.e.,

2. Wavelet Estimators and Theorem

Let us define linear wavelet estimator as follows: where

Using hard thresholding method, our nonlinear wavelet estimator is defined by with , the indicator function and

The positive parametrics , , and will be given in later discussion. In addition, and . Because of the compactly supported of regression functions and , the cardinality of and satisfies that and , respectively. Here and after, denotes for some positive constant , and stands for both and .

Now we state our main theorem in the following, which investigates the performance of linear and nonlinear wavelet estimators in terms of rates of convergence under pointwise risk over Besov space.

Theorem 3. Consider the regression problems (1) and (2) under the assumptions C1-C5 and . Then, the linear wavelet estimator with satisfies The nonlinear one with and does

Remark 4. Note that the linear and nonlinear wavelet estimators all can attain the optimal convergence rate for pointwise risk [13].

Remark 5. Compared with the linear wavelet estimator , the convergence rate of the nonlinear estimator keeps the same as the linear one up to a factor. Moreover, the nonlinear wavelet estimator is adaptive, which means that it only depends on the observed data.

3. Auxiliary Results

This section will provide some results for the proof of Theorem 3.

Lemma 6. For the problem defined in (1) and (2), we have

Lemma 7. Let be a stationary strong mixing process. If C1-C5 hold and ,

Proof. According to the properties of variance, we know Now we estimate the first term. It follows from the property of variance and C2 that Using the property of strictly stationary, Then, we need the following Davydov inequality [14]. Let be strong mixing with mixing coefficient and and be two measurable functions. If and exist for and , Using the Davydov inequality with , This with C2 shows . Furthermore, it follows from C4 that Combining this result with and (22)–(24) and ((27)), one gets that For the estimation of (20), we know that Then, it is easy to see from the property of variance and the assumptions C1-C3 that Now it remains to prove Similar to the arguments of (24), one gets that . Because of , we can obtain that First, estimate . By the property of covariance, . Furthermore, using the assumptions C1-C3 and C5, one can get that Then, one have Now we estimate . Using the above Davydov inequality with , It follows from C1-C3 that On the other hand, Hence, Now the above results (31), (35), and (39) imply the desired inequality The prove of the last inequality (21) is similar to (20).

Lemma 8. Let assumptions C1-C5 hold and . Then,

Proof. By the definition of , Note that C3 and the definition of regression function imply the boundedness of . Furthermore, by Hölder inequality and the orthonormality of . On the other hand, and because of C2. Then, one can obtain that It is easy to see from Lemmas 6 and 7 that For the second inequality, according to the definition of , Then, it follows from the similar arguments of (43) that Using Lemmas 6 and 7, one can get

Lemma 9. Let be defined by (15). If C1-C5 hold, and ; then for each , there exists a constant such that

Proof. According to the arguments of (43) and the definition of , Then, we only need to prove the following two inequalities, respectively. (i)Proof of (50). Define . Then, thanks to Lemma 6. Moreover, are strong mixing with the mixing coefficients . Using the assumptions C1-C3, one gets that . On the other hand, it follows from Lemma 7 thatBecause strong mixing is much more complicated than independent, we need to use a new technique. In this paper, we use the following Bernstein type inequality [15]. Let be a strong mixing process with mixing coefficient , , , and . Then, for and with , According to the above Bernstein type inequality with and , Then, it follows from and that Obviously, there exists sufficiently large such that On the other hand, note that Choosing such that , then Together with the above results, one can obtain the desired result (50) (ii)Proof of (51). Define . Then, thanks to Lemma 6. Moreover, it follows from the assumption C4 that are strong mixing with the mixing coefficients . Note that with C2. According to Lemma 7, . Then, similar to the proof of (50), one can get the desired result (51).

4. Proof of Theorem

This last section is devoted to proving Theorem 3.

Proof of (16). It is easy to see that (i)Upper bound of . Note thatThen, it follows from Cauchy-Schwarz inequality and Condition that Using Lemma 8 and , (ii)Upper bound of . It is easy to see from Hölder inequality thatFurthermore, Condition and the boundedness of imply that Now using , Lemma 2, and , one can get that Combining with (59), (62), and (65), the desired result is obtained by

Proof of (17). By the definition of , Hence, According to the arguments of (62), and , Similar to the arguments of (65) and , Then, we only need to estimate the upper bound of Note that where (i)Upper Bound ofand. It follows from Cauchy-Schwarz inequality thatUsing Lemmas 8 and 9, one gets Then, choosing such that , one has that (ii)Upper Bound of. To estimate , one definesNote that can be decomposed into the following two terms: Using Cauchy-Schwarz inequality and Lemma 8, On the other hand, it is easy to see that Then, one knows with Lemma 2. Now together with (78), (79), and (81), (iii)Upper Bound of. Note thatBy Condition and , Moreover, according to Hölder inequality and Lemma 2, we can get Combining this with (83) and (84), Now the results (72), (76), (82), and (86) imply that This with (68)–(70) shows

Data Availability

No data were used in this study.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This paper is supported by the National Natural Science Foundation of China (Nos. 12001133, 12001132, and 71961004), Guangxi Natural Science Foundation (Nos. 2018GXNSFBA281076 and 2019GXNSFFA245012), Guangxi Science and Technology Project (Nos. Guike AD18281058 and Guike AD18281019), and Guangxi Colleges and Universities Key Laboratory of Data Analysis and Computation. The authors would like to thank the two reviewers for their thoughtful comments.