Abstract

The ridge regression-type (Hoerl and Kennard, 1970) and Liu-type (Liu, 1993) estimators are consistently attractive shrinkage methods to reduce the effects of multicollinearity for both linear and nonlinear regression models. This paper proposes a new estimator to solve the multicollinearity problem for the linear regression model. Theory and simulation results show that, under some conditions, it performs better than both Liu and ridge regression estimators in the smaller MSE sense. Two real-life (chemical and economic) data are analyzed to illustrate the findings of the paper.

1. Introduction

To describe the problem, we consider the following linear regression model:where y is an vector of the response variable, is a known full rank matrix of predictor or explanatory variables, is an vector of unknown regression parameters, is an vector of errors such that , and , is an identity matrix. The ordinary least squares estimator (OLS) of β in (1) is defined aswhere is the design matrix.

The OLS estimator dominates for a long time until it was proven inefficient when there is multicollinearity among the predictor variables. Multicollinearity is the existence of near-to-strong or strong-linear relationship among the predictor variables. Different authors have developed several estimators as an alternative to the OLS estimator. These include Stein estimator [1], principal component estimator [2], ridge regression estimator [3], contraction estimator [4], modified ridge regression (MRR) estimator [5], and Liu estimator [6]. Also, some authors have developed two-parameter estimators to combat the problem of multicollinearity. The authors include Akdeniz and Kaçiranlar [7]; Özkale and Kaçiranlar [8]; Sakallıoğlu and Kaçıranlar [9]; Yang and Chang [10]; and very recently Roozbeh [11]; Akdeniz and Roozbeh [12]; and Lukman et al. [13, 14], among others.

The objective of this paper is to propose a new one-parameter ridge-type estimator for the regression parameter when the predictor variables of the model are linear or near-to-linearly related. Since we want to compare the performance of the proposed estimator with ridge regression and Liu estimator, we will give a short description of each of them as follows.

1.1. Ridge Regression Estimator

Hoerl and Kennard [3] originally proposed the ridge regression estimator. It is one of the most popular methods to solve the multicollinearity problem of the linear regression model. The ridge regression estimator is obtained by minimizing the following objective function:with respect to β, will yield the normal equationswhere k is the nonnegative constant. The solution to (4) gives the ridge estimator which is defined aswhere , and k is the biasing parameter. Hoerl et al. [15] defined the harmonic-mean version of the biasing parameter for the ridge regression estimator as follows:where is the estimated mean squared error form OLS regression using equation (1) and is ith coefficient of and is defined under equation (17). There are a high number of techniques suggested by various authors to estimate the biasing parameters. To mention a few, McDonald and Galarneau [16]; Lawless and Wang [17]; Wichern and Churchill [18]; Kibria [19]; Sakallıoğlu and Kaçıranlar [9]; Lukman and Ayinde [20]; and recently, Saleh et al. [21], among others.

1.2. Liu Estimator

The Liu estimator of is obtained by augmenting to (1) and then applying the OLS estimator to estimate the parameter. The Liu estimator is obtained to bewhere . The biasing parameter d for the Liu estimator is defined as follows:where is the ith eigenvalue of the matrix and which is defined under equation (17). If is negative, Özkale and Kaçiranlar [8] adopt the following alternative biasing parameter:where is the ith component of .

For more on the Liu [6] estimator, we refer our readers to Akdeniz and Kaçiranlar [7]; Liu [22]; Alheety and Kibria [23]; Liu [24]; Li and Yang [25]; Kan et al. [26]; and very recently, Farghali [27], among others.

In this article, we propose a new one-parameter estimator in the class of ridge and Liu estimators, which will carry most of the characteristics from both ridge and Liu estimators.

1.3. The New One-Parameter Estimator

The proposed estimator is obtained by minimizing the following objective function:with respect to β, will yield the normal equationswhere k is the nonnegative constant. The solution to (11) gives the new estimator aswhere , and . The new proposed estimator will be called the Kibria–Lukman (KL) estimator and denoted by .

1.3.1. Properties of the New Estimator

The proposed estimator is a biased estimator unless k = 0.and the mean square error matrix (MSEM) is defined as

To compare the performance of the four estimators (OLS, RR, Liu, and KL), we rewrite (1) in the canonical form which giveswhere and . Here, Q is an orthogonal matrix such that Z’Z = QX’XQ = Λ = diag (λ1, λ2, …, λp). The OLS estimator of is

The ridge estimator (RE) of iswhere and k is the biasing parameter.where .

The Liu estimator of iswhere .where .

The proposed one-parameter estimator of iswhere and .

The following notations and lemmas are needful to prove the statistical property of :

Lemma 1. Let n × n matrices M > 0 and N > 0 (or N ≥ 0); then, M > N if and only if λ1 (NM−1) < 1, where λ1 (NM−1) is the largest eigenvalue of matrix NM−1 [28].

Lemma 2. Let M be an n × n positive definite matrix, that is, M > 0 and be some vector; then, if and only if [29].

Lemma 3. Let , i = 1, 2, be two linear estimators of . Suppose that , where denotes the covariance matrix of and . Consequently,if and only if , where [30].

The other parts of this article are as follows. The theoretical comparison among the estimators and estimation of the biasing parameters are given in Section 2. A simulation study has been constructed in Section 3. We conducted two numerical examples in Section 4. This paper ends up with concluding remarks in Section 5.

2. Comparison among the Estimators

2.1. Comparison between and

The difference between and is

We have the following theorem.

Theorem 1. If k> 0, estimator is superior to estimator using the MSEM criterion, that is, if and only if

Proof. The difference between (15) and (19) iswhere will be positive definite (pd) if and only if We observed that, for k > 0, .
Consequently, is pd.

2.2. Comparison between and

The difference between and is

Theorem 2. When estimator is superior to in the MSEM sense if and only ifwhere

Proof. Using the dispersion matrix difference,It is obvious that, for k > 0, G > 0 and H > 0. According to Lemma 1, it is clear that G-H > 0 if and only if , where is the maximum eigenvalue of the matrix Consequently, is pd.

2.3. Comparison between and

The difference between and is

We have the following theorem.

Theorem 3. If k> 0 and 0 < d < 1, estimator is superior to estimator using the MSEM criterion, that is, if and only ifwhere .

Proof. Using the difference between the dispersion matrix,where and We observed that is pd if and only if or . Obviously for k> 0 and 0 < d < 1, Consequently, is pd.

2.4. Determination of Parameter k

There is a need to estimate the parameter of the new estimator for practical use. The ridge biasing parameter and the Liu shrinkage parameter were determined by both Hoerl and Kennard [3] and Liu [6], respectively. Different authors have developed other estimators of these ridge parameters. To mention a few, these include Hoerl et al. [15]; Kibria [19]; Kibria and Banik [31]; and Lukman and Ayinde [20], among others. The optimal value of k is the one that minimizes

Differentiating with respect to k gives and setting , we obtain

The optimal value of k in (39) depends on the unknown parameter and . These two estimators are replaced with their unbiased estimate. Consequently, we have

Following Hoerl et al. [15], the harmonic-mean version of (40) is defined as

According to Özkale and Kaçiranlar [8], the minimum version of (41) is defined as

3. Simulation Study

Since theoretical comparisons among the estimators, ridge regression, Liu and KL in Section 2 give the conditional dominance among the estimators, a simulation study has been conducted using the R 3.4.1 programming languages to see a better picture about the performance of the estimators.

3.1. Simulation Technique

The design of the simulation study depends on factors that are expected to affect the properties of the estimator under investigation and the criteria being used to judge the results. Since the degree of collinearity among the explanatory variable is of central importance, following Gibbons [32] and Kibria [19], we generated the explanatory variables using the following equation:where are independent standard normal pseudo-random numbers and represents the correlation between any two explanatory variables. We consider and 7 in the simulation. These variables are standardized so that and are in correlation forms. The n observations for the dependent variable y are determined by the following equation:where are i.i.d N (0, σ2), and without loss of any generality, we will assume zero intercept for the model in (44). The values of β are chosen such that  = 1 [33]. Since our main objective is to compare the performance of the proposed estimator with ridge regression and Liu estimators, we consider k=d = 0.1, 0.2, …, 1. We have restricted k between 0 and 1 as Wichern and Churchill [18] have found that the ridge regression estimator is better than the OLS when k is between 0 and 1. Kan et al. [26] also suggested a smaller value of k (less than 1) is better. Simulation studies are repeated 1,000 times for the sample sizes n = 30 and 100 and σ2 = 1, 25, and 100. For each replicate, we compute the mean square error (MSE) of the estimators by using the following equation:where would be any of the estimators (OLS, ridge, Liu, or KL). Smaller MSE of the estimators will be considered the best one.

The simulated results for n = 30, , and ρ = 0.70, 0.80 and ρ = 0.90, 0.99 are presented in Tables 1 and 2, respectively, and for n = 100, , and ρ = 0.7, 0.80 and ρ = 0.90, 0.99 are presented in Tables 3 and 4, respectively. The corresponding simulated results for n = 30, 100 and are presented in Tables 58. For a better visualization, we have plotted MSE vs. d for n = 30, σ = 10, and ρ = 0.70, 0.90, and 0.99 in Figures 13, respectively. We also plotted MSE vs σ for n = 30, d = .50, and ρ = 0.90 and 0.99, which is presented in Figures 4 and 5, respectively. Finally, to see the effect of sample size on MSE, we plotted MSE vs. sample size for d = 0.5 and ρ = 0.90 and presented in Figure 6.

3.2. Simulation Results and Discussion

From Tables 18 and Figures 16, it appears that, as the values of σ increase, the MSE values also increase (Figure 3), while the sample size increases as the MSE values decrease (Figure 4). Ridge, Liu, and proposed KL estimators uniformly dominate the ordinary least squares (OLS) estimator. In general, from these tables, an increase in the levels of multicollinearity and the number of explanatory variables increase the estimated MSE values of the estimators. The figures consistently show that the OLS estimator performs worst when there is multicollinearity. From Figures 16 and simulation Tables 18, it clearly indicated that, for or less, the proposed estimator uniformly dominates the ridge regression estimator, while Liu performed much better than both proposed and ridge estimators for small d, say 0.3 or less. When , the ridge regression performs the best for higher k, while the proposed estimator performs the best for say k (say 0.3 or less). When d = k = 0.5 and , both ridge and KL estimators outperform the Liu estimator. None of the estimators uniformly dominates each other. However, it appears that our proposed estimator, KL, performs better in the wider space of d = k in the parameter space. If we review all Tables 18, we observed that the conclusions about the performance of all estimators remain the same for both and .

4. Numerical Examples

To illustrate our theoretical results, we consider two datasets: (i) famous Portland cement data originally adopted by Woods et al. [34] and (ii) French economy data from Chatterjee and Hadi [35], and they are analyzed in the following sections, respectively.

4.1. Example 1: Portland Data

These data are widely known as the Portland cement dataset. It was originally adopted by Woods et al. [34]. It has also been analyzed by the following authors: Kaciranlar et al. [36]; Li and Yang [25]; and recently by Lukman et al. [13]. The regression model for these data is defined aswhere  = heat evolved after 180 days of curing measured in calories per gram of cement,  = tricalcium aluminate,  = tricalcium silicate,  = tetracalcium aluminoferrite, and  = -dicalcium silicate. The correlation matrix of the predictor variables is given in Table 9.

The variance inflation factors are VIF1 = 38.50, VIF2 = 254.42, VIF3 = 46.87, and VIF4 = 282.51. Eigenvalues of are , and , and the condition number of is approximately 424. The VIFs, the eigenvalues, and the condition number all indicate the presence of severe multicollinearity. The estimated parameters and MSE are presented in Table 10. It appears from Table 11 that the proposed estimator performed the best in the sense of smaller MSE.

4.2. Example 2: French Economy Data

The French economy data in Chatterjee and Hadi [37] are considered in this example. It has been analyzed by Malinvard [38] and Liu [6], among others. The variables are imports, domestic production, stock formation, and domestic consumption. All are measured in milliards of French francs for the years 1949 through 1966.

The regression model for these data is defined aswhere  = IMPORT,  = domestic production,  = stock formation, and  = domestic consumption. The correlation matrix of the predicted variable is given in Table 12.

The variance inflation factors are and . The eigenvalues of the matrix are λ1 = 161779, λ2 = 158, and λ3 = 49.61, and the condition number is 32612. If we review the above correlation matrix, VIFs, and condition number, it can be said that there is presence of severe multicollinearity existing in the predictor variables.

The biasing parameter for the new estimator is defined in (41) and (42). The biasing parameter for the ridge and Liu estimator is provided in (6), (8), and (9), respectively.

We analyzed the data using the biasing parameters for each of the estimators and presented the results in Tables 10 and 11. It can be seen from Tables 10 and 11 that the proposed estimator performed the best in the sense of smaller MSE.

5. Summary and Concluding Remarks

In this paper, we introduced a new biased estimator to overcome the multicollinearity problem for the multiple linear regression model and provided the estimation technique of the biasing parameter. A simulation study has been conducted to compare the performance of the proposed estimator and Liu [6] and ridge regression estimators [3]. Simulation results evidently show that the proposed estimator performed better than both Liu and ridge under some condition on the shrinkage parameter. Two sets of real-life data are analyzed to illustrate the benefits of using the new estimator in the context of a linear regression model. The proposed estimator is recommended for researchers in this area. Its application can be extended to other regression models, for example, logistic regression, Poisson, ZIP, and related models, and those possibilities are under current investigation [37, 39, 40].

Data Availability

Data will be made available on request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

We are dedicating this article to those who lost their lives because of COVID-19.