Abstract

Generalized method of moments (GMM) has been widely applied for estimation of nonlinear models in economics and finance. Although generalized method of moments has good asymptotic properties under fairly moderate regularity conditions, its finite sample performance is not very well. In order to improve the finite sample performance of generalized method of moments estimators, this paper studies higher-order mean squared error of two-step efficient generalized method of moments estimators for nonlinear models. Specially, we consider a general nonlinear regression model with endogeneity and derive the higher-order asymptotic mean square error for two-step efficient generalized method of moments estimator for this model using iterative techniques and higher-order asymptotic theories. Our theoretical results allow the number of moments to grow with sample size, and are suitable for general moment restriction models, which contains conditional moment restriction models as special cases. The higher-order mean square error can be used to compare different estimators and to construct the selection criteria for improving estimator’s finite sample performance.

1. Introduction

It is a stylized fact that plenty of relationships are dynamic and nonlinear in nature and society, especially in economic and financial systems [18]. These relationships are usually depicted by nonlinear models. Generalized method of moments (GMM) has been widely applied for analysis for these nonlinear models since it was first introduced by Hansen [9] and gradually became a fundamental estimation method in econometrics [10]. Nevertheless, although GMM has good asymptotic properties under fairly moderate regularity conditions, its finite sample performance is not very well [1113]. Similar to the maximum likelihood estimation (MLE), GMM does not have an exact finite sample distribution. In practice, we generally use the asymptotic distribution to approximate this finite sample distribution, but many applications of GMM reveal that this approximation has low precision [14].

When traditional asymptotic theory cannot precisely approximate the finite sample distributions of estimators or tests, we need higher-order asymptotic expansion for these estimators or tests to get more accurate approximation [15]. Nagar [16] studied the small sample properties of the general -class estimators of simultaneous equations and gave the higher-order asymptotic expansion of the first- and second-order moments for two-stage least squares (2SLS) estimator. Donald and Newey [17] gave a theoretical derivation of the higher-order MSE for 2SLS based on Nagar [16]; however, their MSE formula applied to the case where the number of instruments grows with but at a smaller rate than the sample size, while Nagar [16] considered the cases where the number of instruments is fixed. Kuersteiner [18] derived the higher-order asymptotic properties of GMM estimators for linear time series models using many lags as instruments.

Besides the linear models, Rilstone et al. [19] derived and examined the second-order bias and MSE of a fairly wide class of nonlinear estimators, which included nonlinear least squares, maximum likelihood, and GMM estimators as special cases. Bao and Ullah [20] extended the second-order bias and MSE results of Rilstone et al. [19] for time series dependent observations. In addition, Bao and Ullah [21] derived the higher-order bias and mean squared error of a large class of nonlinear estimators to order and , respectively. However, although these papers gave the high-order bias and MSE for nonlinear estimators, they were not suitable for two-step efficient GMM estimators.

Newey and Smith [22] studied the higher-order bias for two-step GMM estimators, empirical likelihood (EL) and generalized empirical likelihood (GEL) estimators through higher-order asymptotic expansions. But this paper needs to be improved in the following aspects. First, the data generating process considered in this paper was independently identically distributed. Second, the number of moments is fixed. Third, the MSE of GMM was not given. Anatolyev [23] extended Newey and Smith [22] to stationary time series models with serial correlation. Again, the number of moments in this paper was fixed, and this paper only gave the higher-order bias for the estimators, but not the MSE.

Donald et al. [24] examined higher-order asymptotic MSE for conditional moment restriction models. Based on this MSE, they developed moment selection criteria for two-step GMM estimator, a bias corrected version, and GEL estimators. Donald et al. [24] allowed the number of instruments to grow with sample size. However, this paper constructed moment conditions through instrumental variables, which was not suitable for general moment restriction models. Thus, our paper tends to fill an important lacuna in the literature about higher-order asymptotic expansion of nonlinear estimators. Specially, we consider a general nonlinear regression model with endogeneity, and our theoretical results are suitable for general moment restriction models, which contain conditional moment restriction models as a special case.

The remainder of the paper proceeds as follows: Section 2 introduces the model and notations. Section 3 discusses the estimation for the threshold and slope coefficients. Section 4 concludes.

2. Model

Many economic and financial models can be written as nonlinear functions of data and parameters. Consider the following nonlinear regression model with endogeneity: where is dependent variable, is a vector of explanatory variables, which contains endogenous variables, is a vector of parameters, is a compact subset of , and .

For model (1), the usual way to estimate is nonlinear least squares (NLS). However, contains endogenous variables, which means that . In this case, NLS estimator of is not consistent. We have to look for another consistent estimator, such as GMM estimator. Let be    instrumental variables of , and    are independent random vectors. The orthogonality conditions can be written as And the sample moments are Then, the two-step efficient GMM estimator of is given by where , , and is a random weighting matrix that almost surely converges to a nonstochastic symmetric positive definite matrix .

Our goal is to obtain the MSE of . However, formula (4) does not have an analytical solution. We have to obtain it through higher-order asymptotic theory.

3. Higher-Order MSE of GMM Estimators

This part uses the iterative idea [19] to derive the higher-order MSE of the GMM estimator, which can be seen as a generalization of Donald et al. [24] to the unconditional moment restriction models. For the convenience of discussion, we use the following notations: where is a derivation operator defined by the following recursive method: That is, can be seen as a block matrix whose entry in the th row and th column is a vector , in which is the th row and th column element of matrix .

To derive the higher-order expansion of the GMM estimator , the following assumptions for the moment function are required.

Assumption 1. For some neighborhood of , is, at least, three times continuously differentiable and , , , and .

Assumption 2. For some neighborhood of , , in which , , and  ,  .

Assumption 3. , in which , and are nonnegative integers, ,  ,  , and  .

Assumption 4. The smallest eigenvalues of and are bounded away from zero, in which belongs to the neighborhood of .

Assumption 5. There is and for some finite constant , such that .

Assumption 1 is a necessary condition for a higher-order Taylor expansion. Assumption 2 is a common condition for the moments of remainder terms to bound (see also [19, 20, 2325]). Assumption 3 requires that the third moments are zero, which can simplify the MSE calculations (see also [24, 26, 27]). Assumption 4 is a further identification condition. The purpose of Assumption 5 is to control the remainder terms of higher-order expansions (see [25] for details).

The first order condition for the optimization problem in (4) is given by Define an auxiliary vector ; the above first order condition can be rewritten as Let , , and . Equation (8) can be written as

For (9), the number of parameters is equal to the number of equations. A second-order Taylor expansion of (9) at yields where , , , , and , , in which and are the th element of and , respectively.

From (10), we have where and .

Since the right-hand side of (11) contains terms and , we use iterative techniques to remove these terms. For GMM estimator under Assumptions 15, if and , then can be decomposed as follows: where and, for ,

Before deriving the higher-order MSE of , we need the following lemmas.

Lemma 6. Consider

Proof. Consider in which
Similarly, ,

Lemma 7. Consider

Proof. By definition, and ; then and according to the independence assumption, Let , and ; then . Now, we examine the th row and th column element of , :   . By Assumption 3, . Then we have ; that is, . Finally, we have . So, .

Lemma 8. Consider

Proof. By definition of and , . And by the definition of , . Take mathematical expectation for this formula, According to the independence assumption, And by , we have . To sum up,

Lemma 9. Consider

Proof. By definition of and , .
For the first term, .
Since and , we have .
According to the independence assumption, According to the assumption for and Assumption 1, the mathematical expectation of the first term is zero.
For the second term, . By Lemma 6, .
For the third term, .
For the fourth term, .
For the fifth term, .
Similarly, according to the independence assumption and Assumption 1, the mathematical expectation of the third, fourth, and fifth terms are zero.
To sum up, .

Lemma 10. Consider

Proof. By definition of and , Similar to the proof of Lemma 8, we have And according to the independence assumption and Assumption 1, To sum up, .

Using these lemmas, then we can get the higher-order MSE of as follows.

Theorem 11. For GMM estimator in (4), under Assumptions 15, if and , then the higher-order MSE of is given by in which

Proof. By (12), By Lemma 6, , the above formula can be rewritten as By Lemmas 710, In addition, by definition of , . Consider Denote ; then . . Then, According to the independence assumption,

In (32), is asymptotic variance of , and can be seen as the asymptotic bias terms. In practice, and can be substituted by their consistent estimators.

4. Conclusions

In this paper, we consider a general nonlinear regression model with endogeneity and derive the higher-order mean square error of two-step efficient generalized method of moments estimators for this nonlinear model. The theoretical results in this paper allow the number of moments to grow with but at smaller rate than the sample size. And the derivations are suitable for general moment restriction models, which contain conditional moment restriction models and linear models as special cases. The higher-order mean squared error got in this paper has many uses. For example, it can be used to compare among different estimators or to construct the selection criteria of moments for improving the finite sample performance of GMM estimators. This paper considered a restrictive condition in which the data generating process is independent. It would be valuable to extend the results to the dynamic panel data models, in which the moments are going with the time dimension. It is saved for future research.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was partially supported by the National Natural Science Foundation of China (Grant nos. 71301160, 71203224, and 71301173), Beijing Planning Office of Philosophy and Social Science (13JGB018), and Program for Innovation Research in Central University of Finance and Economics.