1,824
Views
20
CrossRef citations to date
0
Altmetric
Article

New ridge estimators in the inverse Gaussian regression: Monte Carlo simulation and application to chemical data

ORCID Icon, ORCID Icon, ORCID Icon &
Pages 6170-6187 | Received 05 Oct 2019, Accepted 14 Jul 2020, Published online: 04 Aug 2020

Abstract

In numerous application areas, when the response variable is continuous, positively skewed, and well fitted to the inverse Gaussian distribution, the inverse Gaussian regression model (IGRM) is an effective approach in such scenarios. The problem of multicollinearity is very common in several application areas like chemometrics, biology, finance, and so forth. The effects of multicollinearity can be reduced using the ridge estimator. This research proposes new ridge estimators to address the issue of multicollinearity in the IGRM. The performance of the new estimators is compared with the maximum likelihood estimator and some other existing estimators. The mean square error is used as a performance evaluation criterion. A Monte Carlo simulation study is conducted to assess the performance of the new ridge estimators based on the minimum mean square error criterion. The Monte Carlo simulation results show that the performance of the proposed estimators is better than the available methods. The comparison of proposed ridge estimators is also evaluated using two real chemometrics applications. The results of Monte Carlo simulation and real applications confirmed the superiority of the proposed ridge estimators to other competitor methods.

MATHEMATICS SUBJECT CLASSIFICATION:

1. Introduction

In real life we often deal with the datasets in which the form of dependent variable is continuous and positively skewed. In such scenario, the inverse Gaussian regression model (IGRM) is more suitable than the linear regression model (LRM). The most common maximum likelihood estimation (MLE) is used to estimate the regression coefficients of the IGRM. The applications of IGRM are mostly observed in the fields of physical sciences, health sciences, chemical sciences, and engineering (Amin, Amanullah, and Aslam Citation2016; Kinat, Amin, and Mahmood Citation2020; Akram, Amin, and Qasim Citation2020; Naveed et al. Citation2020; Amin, Amanullah, and Qasim Citation2020). Multicollinearity is an imperative problem particularly in fields of chemometrics, health sciences, and biostatistics. This problem occurs when the regressors are linearly correlated. In the presence of multicollinearity problem, it is not suitable to estimate the regression coefficients of the IGRM using MLE. The MLE has many drawbacks when the regressors are linearly correlated. The MLE may be unstable with wrong signs of the coefficients and the variances of the coefficients become inflated in the presence of multicollinearity. In addition, the interpretation of the estimated coefficients also becomes difficult (Qasim et al. Citation2019; Amin, Akram, and Amanullah Citation2020). To resolve the issue of multicollinearity, many alternative biased estimation methods are provided in the literature. One way to address the problem of multicollinearity is ridge regression (RR). The method of RR was firstly introduced by Hoerl and Kennard (Citation1970) for the LRM. One benefit of RR is that the mean squared error (MSE) can be reduced using an optimal value of the shrinkage parameter k, with the decrease in variance and increase in bias. Many research studies have been conducted for the LRM which is estimated using ordinary least square and various shrinkage parameters, k have been suggested. For the LRM, some popular studies have been done by Hoerl and Kennard (Citation1970), Hoerl, Kennard, and Baldwin (Citation1975), McDonald and Galarneau (Citation1975), Kibria (Citation2003), Khalaf and Shukur (Citation2005), Muniz and Kibria (Citation2009), and many others.

The literature on ridge estimators for the generalized linear model (GLM) is very limited. However, a few studies have been conducted where different ridge estimators are suggested for some specific cases of the GLM. For instance, Månsson and Shukur (Citation2011) proposed a RR approach in the Poisson regression model. Kibria, Månsson, and Shukur (Citation2015) evaluated the performance of some ridge estimators in the logistic regression model. Furthermore, Månsson (Citation2012) suggested ridge estimators for the negative binomial regression model. Recently, Amin et al. (Citation2020a) and Algamal (Citation2018a) proposed some ridge estimators and recommended best ridge estimator for the gamma RR. Algamal (Citation2018b) presented an efficient estimation algorithm of biasing parameter of the ridge estimator. The working process of kidney in human body is the inspiration behind the proposed algorithm. Some new shrinkage estimators for gamma regression model were suggested by Algamal (Citation2018c). The adjusted biased estimator proposed by Amin et al. (Citation2020b) to overcome the problem of inflated bias in the gamma RR estimator. Algamal (Citation2019) proposed ridge estimator in the IGRM. Shamany, Alobaidi, and Algamal (Citation2019) introduced a two-parameter estimator for the IGRM which was applied to a chemometric dataset. Amin, Qasim, and Amanullah (Citation2019) and Akram, Amin, and Amanullah (Citation2020) proposed two-parameter estimators for the Gamma regression model and IGRM, respectively. Different studies have been proposed related to ridge estimators and recommended a best ridge parameter for different form of the GLM. From the available literature, it is determined that there is a need of study to choose suitable ridge parameter for the inverse Gaussian ridge regression (IGRR).

This article introduces some new estimating methods for choosing best ridge parameter for the IGRR to decrease the MSE and acquire efficient results. In addition, we derived the MSE properties of the IGRR. The performance of the proposed ridge estimators is compared with the existing best ridge estimators for different type of models. The Monte Carlo simulation is used to evaluate the performance of the proposed and existing estimators in sense of MSE. The advantage of the proposed estimators is shown using two different applications where one can see that the newly proposed estimator is an effective alternative to the MLE and other existing biased estimators in the presence of multicollinearity.

The rest of the article is structured as follows. In Sec. 2, we discuss the estimation methods of the IGRM, IGRR, and derive MSE properties. In Sec. 3, the new ridge estimators are defined and modified the existing estimators for the IGRR. Details of the Monte Carlo simulation and simulated results are presented in Sec. 4. The advantage of the proposed estimator is shown by real life applications in Sec. 5. Finally, concluding remarks of the article are given in Sec. 6.

2. Theory and method

2.1. The IGRM

The IGRM is very common in applied research work, when the dependent variable (yi) is continuous, enormously positively skewed and well fitted to the inverse Gaussian (IG) distribution with location parameter, μ and scale parameter σ2, denoted by IG(μ,σ2). The required probability density function is given as(1) f(y;μ,σ2)=12πy3σ2exp{(yμ)22μ2yσ2};y>0(1) with mean μ, and variance σ2μ3. Suppose y be the dependent variable and distributed as IG(μ,σ2). In the IGR model g(μ)=1η is the link function where η=Xtβ be the linear predictor of the explanatory variables, where Xt=(xi1,,xip)t which is n×(p+1) data matrix of the explanatory variables and β is (p+1)×1 vector of the regression parameters. In case of the GLM, the IG distribution is a member of exponential family for which the general density is given as(2) f(y;θ,ϕ)=exp[yθb(θ)ϕ+c(y,ϕ)],(2) where θ is the location parameter but not necessarily mean, b(θ) is cumulant function and ϕ is the dispersion parameter. The IG distribution in exponential density form is written as(3) f(y:θ,ϕ)=exp[y2μ2ϕ+1μϕ12yϕ12ln(2πy3)ln(ϕ)](3)

By equating EquationEqs. (2) and Equation(3), we have θ=12μ2,b(θ)=2θ and ϕ=σ2. Thus, the mean and variance of the IG density are given in EquationEqs. (4) and Equation(5), respectively.(4) E(y)=b(θ)=(1μ2)(μ3)=μ=1η,(4) (5) V(y)=ϕb(θ)=ϕV(μ)=ϕdiag(μ13,,μn3),(5) where prime denotes the partial derivative. The final form of log likelihood function of EquationEq. (3) is given in EquationEq. (6).li(yi;μi,ϕ)=i=1n{[(yi)2μi21μi]ϕ+1(2yiϕ)12ln(2πyi3ϕ)}or equivalently(6) li(yi;xitβ,ϕ)=i=1n{[yixitβ2xitβ]ϕ+1(2yiϕ)12ln(2πyi3ϕ)}.(6)

The MLE ofβ can be obtained by solving the following equation,(7) U(βj)=liβj=12ϕ(yi1xitβj)xit=0, i=1,2,,n, j=1,2,,p+1.(7)

Since the solution of EquationEq. (7) is non-linear, so the Newton-Raphson iterative procedure is used to estimate the unknown parameters. For iterative procedure of the IGRM, initial values and full algorithm for the estimation of unknown parameters can be found in Hardin and Hilbe (Citation2012). Let β(m) be the approximated maximum likelihood value of β at the mth iteration with convergence in deviance, the iterative method gives the relation as(8) β(m+1)=β(m)+{I(β(m))}1U(β(m)),(8) whereI(β(m)) is the fisher information matrix and U(β(m)) is the score vector with dimension (p+1)×1 and both fisher information matrix and score vector are evaluated at β(m). At convergence in deviance, the unknown regression parameters can be calculated as(9) β̂ML=(S)1XtŴẑ,(9) where S=XtŴX, ẑi=η̂i+(yiμ̂i)μ̂i3 be the adjusted response variable and Ŵ=diag(μ̂13,μ̂23.,μ̂n3). Here, μ̂i=1xitβ̂ML,i=1,2,,n. and η̂i=xitβ̂ML. Both ẑ and Ŵ are found by the iterative methods and for the detail derivations, readers are referred to Hardin and Hilbe (Citation2012). Moreover, the MSE of the MLE is given as(10) E(LML2)=E(β̂MLβ)t(β̂MLβ)=ϕ̂tr(S)1=ϕ̂j=1p+11λj,(10) where λj is the jth eigenvalue of the XtŴX matrix and ϕ̂ is the estimated dispersion parameter computed asϕ̂=1n(p+1)i=1n(yiμ̂i)2V(μ̂i)=1n(p+1)i=1n(yi1xitβ̂ML)2(1xitβ̂ML)3. One disadvantage of MLE is that the MSE becomes large when the explanatory variables are linearly correlated since some of the eigenvalues will be decreased. The EquationEq. (10) undoubtedly shows that the MSE of the MLE becomes overstated due to the problem of multicollinearity.

2.2. The IGRR

The IGRR is an effective alternative to the MLE in the presence of multicollinearity. The MSE of the MLE becomes inflated and results are inefficient when the explanatory variables are correlated. To solve the problem of collinearity among the explanatory variables, Segerstedt (Citation1992) recommended RR method for the GLM based on the work of Hoerl and Kennard (Citation1970). The regression coefficients of the IGRM are estimated by MLE using iterative weighted least square approach. Let β* be an estimate of vector β. Then, the weighted sum of squared error (WSSE) can be written as(11) WSSE=(yβ*)t(yβ*)=(yXβ̂ML)t(yXβ̂ML)+(β*β̂ML)tS(β*β̂ML),(11)

According to the condition (β*β̂ML)tS(β*β̂ML) is the increased of the WSSE when β̂ML is substituted by β*. In order to find the IGRR estimator, the length of β*tβ* should be minimized subject to (β*β̂ML)tS(β*β̂ML)=0, which can be described as a Lagrangian problem(12) Minimize F=β*tβ*+1k[(β*β̂ML)tS(β*β̂ML)0],(12) where 1k represents the Lagrange multiplier and when it is differentiated with respect to β* by resulting zero then it can be shown as(13) Fβ*=2β*+1k(2Sβ*2Sβ̂ML)=0.(13)

Thus, the final form of the IGRR estimator may be written after simplifying the EquationEq. (13) as(14) β*=β̂RR=(S+kIp)1Sβ̂ML,(14) where k represents the ridge parameter of the IGRR.

2.3. Bias and MSE of the IGRR

The bias and MSE of the IGRR estimator are derived as(15) Bias(β̂RR)=E(β̂RR)β=E[(S+kIp)1Sβ̂ML]β.=[(S+kIp)1]Sββ.(15) (16) Bias(β̂RR)=(S+kIp)1(S+kIpkIp)ββ.=[(S+kIp)1(S+kIp)k(S+kIp)1]ββ.=[Ipk(S+kIp)1]ββ.(16)

The bias of IGRR is written as(17) Bias(β̂RR)=[k(S+kI)1]β,(17)

The matrix MSE (MMSE) of β̂RRof β can be written asMMSE(β̂RR)=E(β̂RRβ)t(β̂RRβ)=Var(β̂RR)+Bias(β̂RR)Bias(β̂RR)t(18) MMSE(β̂RR)=(S+kIp)1S(S+kIp)1+k2[(S+kI)1]ββt[(S+kI)1].(18)

The scalar MSE of β̂RR can be found by applying trace operator on EquationEq. (18), we have(19) MSE(β̂RR)=tr[MSE(β̂RR)]=tr[Var(β̂RR)]+Bias(β̂RR)Bias(β̂RR)t.(19) (20) MSE(β̂RR)=ϕ̂j=1p+1λj(λj+kIp)2+k2j=1p+1αj2(λj+kIp)2=γ1(k)+γ2(k),(20) where λj is the jth eigenvalue of matrix S and α=QtβML and Q is the eigenvector that can be defined as S=QtΛQ, where Λ=diag(λ1,λ2,,λp+1).

2.4. The superiority of the IGRR to the MLE

Hoerl and Kennard (Citation1970) derived three theorems about the MSE properties of RR estimator in the LRM. Here, we show that these theorems also hold for the IGRR estimator. In addition, we also show the superiority of the IGRR estimator to the MLE.

Theorem 2.4.1: The total variance γ1(k) is a continuous, monotonically decreasing function of k.

Proof.

The first derivative of γ1(k) w.r.t k using EquationEq. (20) is given by

(21) γ1(k)=dγ1(k)dk=2ϕ̂j=1p+1λj(λj+k)3(21)

The EquationEq. (21) shows that γ1(k) is a continuous, monotonically decreasing function of k, since λj>0  and k>0.

Theorem 2.4.2: The squared bias γ2(k) is a continuous, monotonically increasing function of k.

Proof.

The first derivative of γ2(k) with respect to k using EquationEq. (20), we have

γ2(k)=dγ2(k)dk=2k2j=1p+1αj2(λj+k)3+2kj=1p+1αj2(λj+k)3=2kj=1p+1λjαj2(λj+k)3

EquationEquation (22) clearly shows that γ2(k) is a continuous, monotonically increasing function of k, since λj>0 and k>0.

Theorem 2.4.3: There always exist a k>0 such that E(LRR2)<E(LML2)=ϕ̂j=1p+11λj.

Proof.

The first derivative of EquationEq. (20) with respect to k equals

(23) E(LRR2)=dE(LRR2)dk=γ1(k)+γ2(k)=2ϕ̂j=1p+1λj(λj+k)3+2kj=1p+1λjαj2(λj+k)3(23)

First note that from EquationEq. (20), if k=0, then γ1(0)=ϕ̂j=1p+11λj and γ2(0)=0. In Theorem 2.4.1 and 2.4.2, it was also confirmed that γ1(k) and γ2(k)  are monotonically decreasing and increasing functions, respectively. Moreover, γ1(k) and γ2(k) are always non-positive and non-negative, correspondingly. Hence, to verify the theorem, it is only necessary to show that there always exist a k>0 such that E(LRR2)<0. As noted from EquationEq. (23), one can easily observe that the condition for E(LRR2)<0 to be hold if k<ϕ̂αmax2.

3. Optimal choice of ridge parameters

In this section, we modify the existing ridge parameters for the IGRR. In addition, we proposed new ridge parameters for the IGRR.

3.1. Existing best ridge parameters in the GLM

Though, various methods have been recommended for the LRM. Such methods are generalized for the GLM. First, we consider the most popular one, which is consider in various GLM based on the work of Hoerl, Kennard, and Baldwin (Citation1975) and we modify it for the IGRR as:k1=kHKB=pϕ̂α̂tα̂.

Månsson and Shukur (Citation2011) recommended the best ridge parameter for the Poisson RR and we modify this parameter for the IGRR as:k2=max(αj2ϕ̂)

Månsson (Citation2012) suggested the similar ridge parameter (k2) with addition of some other best ridge parameters for the negative binomial RR and we modified these parameters for the IGRR as:k3=(Πj=1p+1αj2ϕ̂)1p+1,k4=median(αj2ϕ̂)Kibria, Månsson, and Shukur (Citation2012) recommend the best ridge parameter for the logistic RR and we modify for the IGRR as:k5=Πj=1p+1(1qj)1p+1,k6=median(qj),where qj=λmaxϕ̂(np)ϕ̂+λmaxα̂j2. Amin et al. (Citation2020a) have found the following best ridge parameters for the Gamma RR model ask7=diag(ϕ̂αj2[1+{1+(ϕ̂αj2)2}])k8=ϕ̂(Πj=1p+1αj2)1p+1

3.2. Proposed ridge parameters for the IGRR

To find the optimal value of k, we equate EquationEq. (20) to zero and solve for the ridge parameter. After simplification, we can define the general optimal ridge parameter as(24) kj=ϕ̂α̂j2(24) Using EquationEq. (24), we propose the following new ridge parametersk9=diag(kj),k10=max(kj),k11=min(kj),k12=j=1p+1kjp+1,k13=median(kj),k14=(j=1p+1kj)1p+1,k15=ϕ̂j=1p+1λ̂jj=1p+1α̂j2λ̂j,whereλ̂j is the jth  estimated eigenvalue of XtŴX such that λ1λ2,,λj.

Now, we compare the performance of proposed ridge parameters with the available best ridge parameters for the IGRR model to determine which one performs better in terms of minimum MSE.

4. The Monte Carlo simulation

The main aim of this study is to propose the ridge estimators for the IGRM. In addition, we propose some new estimation methods for ridge parameter in the IGRR. The performance of the proposed estimators is compared with the existing ridge estimators via Monte Carlo simulation under consider different factors in sense of MSE.

4.1. The design of an experiment

We used MSE as performance criterion to gauge the performance of estimators. The MSE is defined as(25) MSE=i=1M(β̂iβ)t(β̂iβ)M,(25) where M is number of replications and are set to be 5000 and β̂i is the estimated value of β in ith replication. The dependent variable of the IGRM is obtained from the IG distribution with mean is given below(26) μi=E(yi)=(β0+β1xi1+β2xi2++βpxip)12,i=1,2,,nj=1,2,,p.(26)

The IGRM in EquationEq. (26) is generated for p = 3, 6, and 9 explanatory variables, respectively. The regression parameter values of β are selected so that j=1p+1βj2=1, which are common restrictions in simulation studies. However, the correlated explanatory variables are generated as(27) xij=(1α02)1/2Zij+α0Zi(p+1),i=1,2,, n j=1,2,,p.(27) where α02 is the correlation among the explanatory variables and Zip represents the independent standard normal pseudorandom number. The Monte Carlo simulation are designed under different factors and these factors are presented in .

Table 1. Assumed values of different factors for Monte Carlo simulation.

4.2. Results and discussion

The MSE of the ML and IGRR estimators under different factors are obtained from simulation experiments and simulated results are given in . presents the estimated MSEs of proposed ridge parameters for p = 3. It can be observed from the simulation results that the MSE is inflated under different values of collinearity, dispersion, and sample sizes. It can be viewed from the results that MSE of the ML estimator is large while the MSE of the IGRR estimator is small. We considered different ridge parameters for the IGRR model and different patterns in terms of MSEs under different conditions can be observed. The MSEs of the ridge parameters i.e. k1k15 have increasing pattern with increasing degree of collinearity from lower to a higher level by taking the constant values of n and ϕ. The above statement is also true for all p, n, andϕ. The MSEs of all the ridge parameters are increasing as the value of dispersion is increases. Moreover, the MSE of the ridge parameters is affected inversely by the sample sizes. All ridge parameters attain smaller MSE with an increase in sample sizes. This statement is true for all levels of collinearity, number of explanatory variables, and the dispersion. Another factor which affects the MSE of ridge parameters is the number of explanatory variables. This factor has a direct effect on the MSE’s of different ridge parameters of the IGRR model. So, the MSE of the ridge parameters is increased with the increase in the number of explanatory variables on fitting the IGRR model. On comparing the ridge parameters, it is found that for p=3, n=25, ϕ=0.10 and all collinearity levels, the performance of ridge parameter k10 is well than the other ridge choices because it attains smaller MSE among all other ridge parameters. While for n50, p=3, ϕ=0.10  and all collinearity levels, the ridge parameter k6 is the best choice than others. This ridge parameter is also performed well for ϕ2  with the other similar conditions as stated above. Moreover, the ridge parameter k10 looks good when 0.10<ϕ<2. From and , we observed that for p = 6 and p = 9, with moderate multicollinearity and small sample size, the ridge parameter k7 found to be better than other ridge parameters with minimum MSE. While for these explanatory variables, sample size, high multicollinearity, and ϕ=0.10, the ridge parameter k10 gives the smaller MSE as other ridge parameters. On the whole, for these explanatory variables and other parametric conditions, the performance of two ridge parameters i.e. k7 and k10 found to be better than other ridge parameters. So, we can say that k6, k7, and k10 are the best ridge parameters for the IGRR model to overcome the issue of multicollinearity.

Table 2. Simulated MSE when p=3.

Table 3. Simulated MSE when p=6.

Table 4. Simulated MSE when p=9.

5. Applications

In this section, we evaluate the performance of proposed methods using two data sets from real life i.e. the stack loss data and nitrogen dioxides data.

5.1. Application 1: stack loss data

Brownlee (Citation1965) first used the stack loss data for the model inferences. This data set consists of n=21 operations of a plant for the oxidation of ammonia to nitric acid with dependent (stack loss) and p=3 explanatory variables (air flow, cooling water inlet temperature, and acid concentration). The representation of these variables are as follows, y= the percent of the ingoing ammonia that is lost by escaping in the unabsorbed nitric oxides, x1= air flow (which reflects the rate of operation of the plant), x2= temperature of the cooling water in the coils of the absorbing tower for the nitric oxides, x3= concentration of nitric acid in the absorbing liquid. We use this data set to compare the applicability of proposed ridge parameters for the IGRR because stack loss data is well fitted to the IG distribution. The distribution fitting results using goodness of fit tests are given in .

Table 5. Goodness of fit tests.

The problem of multicollinearity is tested using condition index (CI). We observed that the CI=λmax/λmin=145.16. This indicates the existence of severe multicollinearity among the explanatory variables. So, we use the IGRR to overcome the effect of multicollinearity in the IGRM. After fitting the IGRM, the MSEs of the MLE and IGRR with different ridge parameters are computed using EquationEqs. (10) and Equation(20), respectively. The estimated regression coefficients and MSEs of different ridge parameters of the IGRR and the MLE are given in . On comparing the performance of proposed ridge parameters, it is observed that our purposed ridge parameters perform better than that of those which were found to be better for other GLMs. The ridge parameters k10 and k15 for this data are the best ridge parameters because these attains the minimum MSE as compared to other ridge parameters.

Table 6. The estimated regression coefficients and MSE for stack loss data.

Table 7. Goodness of fit tests.

5.2. Application 2: nitrogen dioxides data

For the evaluation of proposed estimator, we consider another example i.e. nitrogen dioxides data. The main aim of this application is to check the effect of three explanatory variables (p = 3) on the response variable (nitrogen dioxides measured in ppm). The explanatory variables are: x1 represents the humidity in the air, x2 denotes the temperature, and x3 indicated the barometric pressure. First, it is necessary to find the distribution of response variable to find out the appropriate regression model. showed that three tests are used to test the distribution of the response variable. Results of Crammer-Von Misses test indicated that the most suitable one is the IG distribution with test statistic (p-value) equals to 0.0518 (0.6266). Moreover, the estimated dispersion parameter found to be 0.0042.

Second, is to test the multicollinearity via CI. The estimated value of CI equals to 58.0951 indicating severe multicollinearity among explanatory variables. The nitrogen dioxides data is well fitted to the IG distribution and the explanatory variables are collinear. So, we apply the IGRR to measure the effect of humidity, temperature, and barometric pressure on nitrogen oxides and check the supremacy of proposed ridge parameters.

The estimated regression coefficients with different estimators and respective MSEs are given in . From , we observed that IGRR with new shrinkage parameters attains minimum MSE as compared to the best ridge parameters which are proposed for the other different GLMs. On comparing the all ridge parameters of the IGRR, we observed that the performance of ridge parameters k10, k12, and k15 is better than the other shrinkage parameters

Table 8. The estimated regression coefficients and MSE for nitrogen dioxides data.

6. Conclusion

The choice of the ridge parameter for regression modeling with correlated explanatory variables is necessary for better decision making. As the available literature shows that GLM with different exponential distributions has different best ridge parameters. In this article, some new ridge parameters for the IGRR model are therefore proposed to overcome the problem of multicollinearity. The proposed ridge parameters are evaluated by the simulation study and chemometric data sets. The performance of these ridge parameters is assessed based on MSE. The simulation results demonstrated that the MSE properties of the MLE and IGRR estimators are affected by changing the factors such as multicollinearity, sample size, number of explanatory variables and dispersion. It is shown that under multicollinearity IGRR estimators outperform the ML method. On comparing the performance of our proposed ridge parameters with the other GLM best ridge parameters for the IGRR, it is observed that mostly the proposed ridge parameter k10 have minimum MSE as compared to other GLM best ridge parameters. The chemometric data set results also support the simulation study results. Based on Monte Carlo simulation and chemometric examples results, the ridge parameter k10 is suggested as the best option for applying IGRM with correlated explanatory variables.

Acknowledgments

The authors are grateful to anonymous referees for their valuable comments that led to improvements in the paper.

References

  • Akram, M. N., M. Amin, and M. Amanullah. 2020. Two-parameter estimator for the inverse Gaussian regression model. Communications in Statistics-Simulation and Computation 1:31. doi:10.1080/03610918.2020.1797797.
  • Akram, M. N., M. Amin, and M. Qasim. 2020. A new Liu-type estimator for the inverse Gaussian regression model. Journal of Statistical Computation and Simulation 90 (7):1153–72. doi:10.1080/00949655.2020.1718150.
  • Algamal, Z. Y. 2018a. Developing a ridge estimator for the gamma regression model. Journal of Chemometrics 32 (10):e3054. doi:10.1002/cem.3054.
  • Algamal, Z. Y. 2018b. A new method for choosing the biasing parameter in ridge estimator for generalized linear model. Chemometrics and Intelligent Laboratory Systems 183:96–101. doi:10.1016/j.chemolab.2018.10.014.
  • Algamal, Z. Y. 2018c. Shrinkage estimators for gamma regression model. Electronic Journal of Applied Statistical Analysis 11 (01):253–68.
  • Algamal, Z. Y. 2019. Performance of ridge estimator in inverse Gaussian regression model. Communications in Statistics-Theory and Methods 48 (15):3836–49.
  • Amin, M., M. Amanullah, and M. Aslam. 2016. Empirical evaluation of the inverse Gaussian regression residuals for the assessment of influential points. Journal of Chemometrics 30 (7):394–404. doi:10.1002/cem.2805.
  • Amin, M., M. Amanullah, and M. Qasim. 2020. Diagnostic techniques for the inverse Gaussian regression model. Communications in Statistics - Theory and Methods. Advance online publication. doi:10.1080/03610926.2020.1777308.
  • Amin, M., M. N. Akram, and M. Amanullah. 2020. On the James-Stein estimator for the Poisson regression model. Communications in Statistics - Simulation and Computation. Advance online publication. doi:10.1080/03610918.2020.1775851.
  • Amin, M., M. Qasim, A. Yasin, and M. Amanullah. 2020a. Almost unbiased ridge estimator in the gamma regression model. Communications in Statistics-Simulation and Computation. Advance online publication. doi:10.1080/03610918.2020.1722837.
  • Amin, M., M. Qasim, and M. Amanullah. 2019. Performance of Asar and Genç and Huang and Yang’s two-parameter estimation methods for the gamma regression model. Iranian Journal of Science and Technology, Transactions A: Science 43 (6):2951–63. doi:10.1007/s40995-019-00777-3.
  • Amin, M., M. Qasim, M. Amanullah, and S. Afzal. 2020b. Performance of some ridge estimators for the gamma regression model. Statistical Papers 61 (3):997–1026. doi:10.1007/s00362-017-0971-z.
  • Brownlee, K. A. 1965. Statistical theory and methodology in science and engineering, vol. 150, 120–31. New York: Wiley.
  • Hardin, J. W., and J. M. Hilbe. 2012. Generalized estimating equations. USA: Chapman and Hall/CRC.
  • Hoerl, A. E., and R. W. Kennard. 1970. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 12 (1):55–67. doi:10.1080/00401706.1970.10488634.
  • Hoerl, A., R. Kennard, and K. Baldwin. 1975. Ridge regression: Some simulations. Communications in Statistics - Simulation and Computation 4 (2):105–23. doi:10.1080/03610917508548342.
  • Khalaf, G., and G. Shukur. 2005. Choosing ridge parameter for regression problems. Communications in Statistics - Theory and Methods 34 (5):1177–82. doi:10.1081/STA-200056836.
  • Kibria, B. M. G. 2003. Performance of some new ridge regression estimators. Communications in Statistics - Simulation and Computation 32 (2):419–35. doi:10.1081/SAC-120017499.
  • Kibria, B. M. G., K. Månsson, and G. Shukur. 2012. Performance of some logistic ridge regression estimators. Computational Economics 40 (4):401–14. doi:10.1007/s10614-011-9275-x.
  • Kibria, B. M. G., K. Månsson, and G. Shukur. 2015. A simulation study of some biasing parameters for the ridge type estimation of Poisson regression. Communications in Statistics - Simulation and Computation 44 (4):943–57. doi:10.1080/03610918.2013.796981.
  • Kinat, S., M. Amin, and T. Mahmood. 2020. GLM-based control charts for the inverse Gaussian response variable. Quality and Reliability Engineering International 36 (2):765–83. doi:10.1002/qre.2603.
  • Månsson, K. 2012. On ridge estimators for the negative binomial regression model. Economic Modelling 29 (2):178–84. doi:10.1016/j.econmod.2011.09.009.
  • Månsson, K., and G. Shukur. 2011. A Poisson ridge regression estimator. Economic Modelling 28 (4):1475–81. doi:10.1016/j.econmod.2011.02.030.
  • McDonald, G. C., and D. I. Galarneau. 1975. A Monte Carlo evaluation of some ridge-type estimators. Journal of the American Statistical Association 70 (350):407–16. doi:10.1080/01621459.1975.10479882.
  • Muniz, G., and B. M. G. Kibria. 2009. On some ridge regression estimators: An empirical comparisons. Communications in Statistics - Simulation and Computation 38 (3):621–30. doi:10.1080/03610910802592838.
  • Naveed, K., M. Amin, S. Afzal, and M. Qasim. 2020. New shrinkage parameters for the inverse Gaussian liu regression. Communications in Statistics-Theory and Methods. Advance online publication. doi:10.1080/03610926.2020.1791339.
  • Qasim, M., B. M. G. Kibria, K. Månsson, and P. Sjölander. 2019. A new Poisson Liu regression estimator: Method and application. Journal of Applied Statistics. Advance online publication. doi:10.1080/02664763.2019.1707485.
  • Segerstedt, B. 1992. On ordinary ridge regression in generalized linear models. Communications in Statistics - Theory and Methods 21 (8):2227–46. doi:10.1080/03610929208830909.
  • Shamany, R. E., N. Z. Alobaidi, and Z. Y. Algamal. 2019. A new two-parameter estimator for the inverse Gaussian regression model with application in chemometrics. Electronic Journal of Applied Statistical Analysis 12 (02):453–64.