797
Views
9
CrossRef citations to date
0
Altmetric
Original Article

Performance of a new ridge regression estimator

Pages 23-26 | Published online: 27 Mar 2018

Abstract

Ridge regression estimator has been introduced as an alternative to the ordinary least squares estimator (OLS) in the presence of multicollinearity. Several studies concerning ridge regression have dealt with the choice of the ridge parameter. Many algorithms for the ridge parameter have been proposed in the statistical literature. In this article, a new method for estimating ridge parameter is proposed. A simulation study has been made to evaluate the performance of the proposed estimator based on the mean squared error (MSE) criterion. The evaluation has been done by comparing the MSEs of the proposed estimator with other well-known estimators. In the presence of multicollinearity, the simulation study indicates that under certain conditions the proposed estimator performs better than other estimators.

1 Introduction

Consider the standard model for multiple linear regression (Draper and Smith, Citation1998)(1) where is an n × 1 column vector of observations on the dependent variable, is an n × p fixed matrix of observations on the explanatory variables and is of full rank p (p ⩽ n), β is a p × 1 unknown column vector of regression coefficients, and is an n × 1 vector of random errors; , where denotes the n × n identity matrix and the prime denotes the transpose of a matrix. The variables are assumed to be standardized so that is in the form of correlation matrix, and the vector is the vector of correlation coefficients of the dependent variable with each explanatory variable. The ordinary least squares (OLS) estimator, , of the parameters is given by (Draper and Smith, Citation1998)(2) Clearly, is an unbiased estimator of β. Let denote the eigenvalues of . The mean squared error (MSE) of the components of is given by (Draper and Smith, Citation1998)(3) In application of multiple linear regression, the matrix might be nearly singular, that is, is small for some value of i. This is due to some inter-relation between the explanatory variables. The relation is technically called multicollinearity. The OLS estimator of regression coefficients tends to become “unstable” in the presence of multicollinearity. More precisely, the variance of the estimates of some of the regression coefficients becomes large. This is clear from Equation(3).

Table 1 Estimated MSEs and the values of with p = 5 and n = 15.

Table 2 Estimated MSEs and the values of with p = 10 and n = 25.

Table 3 Estimated MSEs and the values of with p = 20 and n = 30.

Many attempts have been made to improve the OLS estimation procedure. In general, there are two approaches. One approach centers on finding (biased) estimators which have smaller MSE than the OLS estimators. Ridge regression, as well as many shrinkage type of estimators (Stein, Citation1960; Scolve, Citation1968"), is one example. This approach does not directly address itself to the issue of multicollinearity, even though multicollinearity is often the situation where the aforementioned procedures (or estimators) are used.

Among these estimators, the ridge estimator points indirectly to the issue of multicollinearity by constraining the length of the coefficient estimator (Hocking, Citation1976). In contrast, the second approach deals straightforward with the dependency nature of the explanatory variables. The principal components regression, as well as the latent root regression and the factor analysis approach, is one such example.

Hoerl and Kennard (Citation1970a,Citationb)" proposed the ridge estimator as an alternative to the OLS estimator for use in the presence of multicollinearity. The ridge estimator is given by(4) where denotes an identity matrix and k is a positive number known as ridge parameter. The corresponding MSE is given by(5) Though this estimator results in bias, for a certain value of k, it yields minimum MSE compared to the OLS estimator (Hoerl and Kennard, Citation1970a). However, the depends on unknown parameters k, β and , which can’t be calculated in practice, but k has to be estimated from the real data instead.

Several methods for estimating k have been proposed (Hoerl and Kennard, Citation1970a; Hoerl et al., Citation1975; McDonald and Galarneau, Citation1975; Lawless and Wang, Citation1976; Hocking et al., Citation1976; Wichern and Churchill, Citation1978; Nordberg, Citation1982; Saleh and Kibria, Citation1993; Singh and Tracy, Citation1999; Wencheko, Citation2000; Kibria, Citation2003; Khalaf and Shukur, Citation2005; Alkhamisi et al., Citation2006" and Alkhamisi and Shukur, Citation2007).

Alkhamisi and Shukur (Citation2007) suggested a new approach to estimate the ridge parameter. They also proposed some new estimators by adding to some well-known estimators, where is the largest eigenvalue of . The authors used Monte Carlo experiments and the MSE criterion to compare the proposed estimators with some well-known estimators.

The purpose of this study is to apply the modification mentioned in Alkhamisi and Shukur (Citation2007) to the estimator proposed by Hocking et al. (Citation1976) in order to define a new estimator. A Monte Carlo comparison will be made using the MSE criterion to compare the performances of the proposed estimator with the OLS estimator and the estimators of Hocking et al., Citation1976 and Hoerl and Kennard, Citation1970 .

2 Methodology

It is convenient to express the regression model Eq. Equation(1) in the canonical form. Suppose that there exists an orthogonal matrix such that , where , and contains the eigenvalues of the matrix C, then the canonical form of the model Equation(1) is(6) where and . Then the OLS estimator is given as follows(7) and so we can write the ridge estimator as(8) where . Eq. Equation(8) is called the general form of ridge regression (Hoerl and Kennard, Citation1970a). It follows from Hoerl and Kennard (Citation1970a) that the value of which minimizes the , where(9) is(10) where represents the error variance of model Eq. Equation(1), is the ithelement of α.

Eq. Equation(10) gives a value of that fully depends on the unknowns and and must be estimated from the observed data. Hoerl and Kennard (Citation1970a) suggested the replacement of and by their corresponding unbiased estimators, that is,(11) where is the residual mean square estimate, which is an unbiased estimator of , and is the element of , which is an unbiased estimator of α. They found that the best method for achieving a better estimate is to use for all i, and they suggested k to be (or HK) where(12) If and α are known, then is sufficient to give ridge estimators having smaller MSE than the OLS estimator.

Hocking et al. (Citation1976) suggested the following estimator, (or HSL), for k:(13) We now apply the modification mentioned in Alkhamisi and Shukur (Citation2007) to the estimator proposed by Hocking et al. (Citation1976), , to obtain our new estimator (or NHSL):(14) Since is grater than .

3 The simulation study

In this section, we use Monte Carlo simulation to investigate the properties of OLS, HK, HSL and NHSL. A comparison is then made based on the MSE criterion. Although many estimators can be considered in this simulation study (Kibria, Citation2003; Khalaf and Shukur, Citation2005"; Alkhamisi et al., Citation2006; Alkhamisi and Shukur, Citation2007" and Al-Hassan, Citation2008), we will only consider OLS, HK and HSL estimators and compare them with NHSL. We made these choices for the following reasons:

1.

Our interest herein lies in studying the properties of NHSL as an alternative of OLS in the presence of multicollinearity.

2.

HK estimator is the first ridge estimator that was proposed among all other estimators. Moreover, most of studies concerned with proposing new ridge estimators or comparing ridge estimators to each other take HK estimator in consideration.

3.

By construction, NHSL is a modified version of HSL, so we thought that it is necessary to make a comparison between them.

Therefore, we can say that it is convenient to make the comparison among OLS, HK, HSL and NHSL estimators. But, at the same time, we have to note that more investigation of NHSL is needed in future. This may be done by making comparisons between NHSL and other ridge estimators.

Following McDonald and Galarneau (Citation1975), Wichern and Churchill (Citation1978)", Gibbons (Citation1981) and Kibria (Citation2003), the explanatory variables were generated using the device(15) where are independent standard normal pseudo-random numbers, γ is specified so that the correlation between any two explanatory variables is given by , and p is the number of explanatory variables. The variables are then standardized so that and are in correlation forms. Different sets of correlation are considered corresponding to γ = 0.7, 0.8, 0.9 and 0.99. Using the condition number , it can be shown that these values of γ will include a wide range of low, moderate and high correlations between variables. The n observations for the dependent variable y are determined by(16) where are independent normal pseudo-numbers and is taken to be identically zero, and p is defined as in Eq. Equation(15). We used three different sample sizes: 15, 25 and 30 with 5, 10 and 20 explanatory variables, respectively. These choices of p are taken to study the behavior of the estimators for small, moderate and large number of explanatory variables.

For each set of explanatory variables, one choice for the coefficient vectors is considered. The MSE function depends on the explanatory variables (through , on and on . It was noted (Newhouse and Oman, Citation1971) that if MSE is regarded as a function of with , k and the explanatory variables are fixed, then, subject to the constraint that , the MSE is minimized when is the normalized eigenvector corresponding to the largest eigenvalue of matrix . We didn’t use normalized eigenvectors corresponding to the smallest eigenvalue because the conclusion about the performance of estimators in both cases will not change greatly (Kibria, Citation2003).

For given values of p, n and γ, the experiment was repeated 1000 times by generating 1000 samples. For each replicate r (r = 1, 2,…, 1000), the values of k of different proposed estimators and the corresponding ridge estimators were calculated using(17) Then the MSEs for estimators are calculated as follows(18)

4 Results of the simulation study

In this section we present the results of our Monte Carlo experiments. Our primary interest herein lies in comparing the MSEs of the considered estimators. The main results of simulation are summarized in below. To compare the performances of the considered estimators, we calculate the MSEs of each one. We consider the estimator that leads to the minimum MSE to be the best. It is worth mentioning herein that we used the statistics package Minitab14 to do all calculations that were made in this article.

Looking at , we can see that the HK, HSL and NHSL are better than the OLS, and the NHSL performs better than the HK and HSL. The results also reveal that for high correlations, i.e., when γ = 0.9 and 0.99, the HK and HSL perform almost equivalently. However, the HK produces somewhat lower MSEs than the HSL for all sets of correlation. Moreover, it is observed that for given n and p, the MSEs for all estimators increase as the correlation among the explanatory variables increases. In an opposite manner, for given, as the sample size and the number of explanatory variables increase, the MSEs of all estimators decrease.

5 Conclusion

In this article we have investigated the properties of a new proposed method for estimating the ridge parameter in the presence of multicollinearity. The investigation has been done using Monte Carlo experiments, where levels of correlation, the numbers of explanatory variables and the sample sizes have been varied. For each combination we have 1000 replications. The evaluation of our estimator has been done by comparing the MSEs of our proposed estimator with the OLS estimator and the estimators of Hocking et al. (Citation1976) and Hoerl and Kennard (Citation1970) . We found that our estimator uniformly dominates the other estimators.

References

  • Y.Al-HassanA Monte Carlo evaluation of some ridge estimatorsJordan Journal of Applied Science (JJAS)1022008101110
  • M.A.AlkhamisiG.KhalafG.ShukurSome modifications for choosing ridge parameterCommunications in Statistics-Theory and Methods35200620052020
  • M.A.AlkhamisiG.ShukurA Monte Carlo study of recent ridge parametersCommunications in Statistics, Simulation and Computation3632007535547
  • N.R.DraperH.SmithApplied Regression Analysisthird ed.1998John Wiley and SonsNew York
  • D.G.GibbonsA simulation study of some ridge estimatorsJournal of the American Statistical Association763731981131139
  • R.R.HockingThe analysis and selection of variables in linear regressionBiometrics321976149
  • R.R.HockingF.M.SpeedM.J.LynnA class of biased estimators in linear regressionTechnometrics1841976425437
  • A.E.HoerlR.W.KennardRidge regression: biased estimation for nonorthogonal problemsTechnometrics1219705567
  • A.E.HoerlR.W.KennardRidge regression: applications to nonorthogonal problemsTechnometrics1219706982
  • A.E.HoerlR.W.KennardK.F.BaldwinRidge regression: some simulationCommunications in Statistics41975105123
  • G.KhalafG.ShukurChoosing ridge parameter for regression problemsCommunications in Statistics-Theory and Methods34200511771182
  • B.M.KibriaPerformance of some new ridge regression estimatorsCommunications in Statistics- Simulation and Computation3222003419435
  • J.F.LawlessP.WangA simulation study of ridge and other regression estimatorsCommunications in StatisticsA51976307323
  • G.C.McDonaldD.I.GalarneauA Monte Carlo evaluation of some ridge-type estimatorsJournal of the American Statistical Association703501975407412
  • J.P.NewhouseS.D.OmanAn evaluation of ridge estimatorsRand CorporationP-716-PR1971128
  • L.NordbergA procedure for determination of a good ridge parameter in linear regressionCommunications in StatisticsA111982285309
  • A.K.SalehB.M.KibriaPerformances of some new preliminary test ridge regression estimators and their propertiesCommunications in Statistics-Theory and Methods22199327472764
  • S.L.ScolveImproved estimators for coefficients in linear regressionJournal of American Statistical Association631968597606
  • S.SinghD.S.TracyRidge-regression using scrambled responsesMetrika1999147157
  • C.M.SteinMultiple Regression, Contributions to Probability and Statistics1960Stanford University PressUSA pp. 424–443
  • E.WenchekoEstimation of the signal-to-noise in the linear regression modelStatistical Papers412000327343
  • D.WichernG.ChurchillA comparison of ridge estimatorsTechnometrics201978301311

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.