1,073
Views
2
CrossRef citations to date
0
Altmetric
Articles

Efficient GMM estimation with singular system of moment conditions

Pages 172-178 | Received 21 Apr 2019, Accepted 05 Aug 2019, Published online: 23 Aug 2019

Abstract

Standard generalised method of moments (GMM) estimation was developed for nonsingular system of moment conditions. However, many important economic models are characterised by singular system of moment conditions. This paper shows that efficient GMM estimation of such models can be achieved by using the reflexive generalised inverses, in particular the Moore–Penrose generalised inverse, of the variance matrix of the sample moment conditions as the weighting matrix. We provide a consistent estimator of the optimal weighting matrix and establish its consistency. Potential issues of using generalised inverse and some remedies are also discussed.

JEL Classifications:

1. Introduction

Over the past several decades, a great deal of statisticians' efforts has been devoted to the statistical inference of moment condition models, i.e., models where the linkage between parameter and data is specified through a set of moment restrictions (also known as estimating equations). Technically, a moment condition model specifies that the data generating process of observations Z1,,Zn satisfies (1) E[g(Zi,β0)]=0,(1) where g is a RK-valued known function and β0 is the Rp-valued parameter of interest, and Kp. The popularity of moment condition models is partially due to that a parametric likelihood form may be too strong for many real applications or scientific theories. When the dimension of parameter of interest equals the number of moment conditions, the parameter is said to be just-identified, and the classical approach of the method of moments can be applied for parameter estimation. In practice, a majority of the moment condition models investigated by applied researchers, such as the models for assets pricing and dynamic panel data, are over-identified. The generalised method of moments (GMM) of Hansen (Citation1982) is one of the most popular techniques that are designed for the estimation of over-identified moment condition models (see, e.g., Hansen & West, Citation2002 and Hall, Citation2005).

Like many other classical statistical methods, GMM comes with the price of a set of regularity conditions which warrant its validity. Although in most applications those regularity conditions are not binding, some of them can be violated in interesting circumstances. This paper is concerned with the efficient GMM estimation when one of the regularity conditions of standard GMM, that the covariance matrix of the moment vector evaluated at the true parameter be of full rank, is violated. A typical such kind of violation appears when the system of moment conditions is singular, i.e., some components of the moment functions are linear combinations of each other.

Singular systems of moment conditions exist in a wide variety of economic studies, such as the consumer expenditure function analysis (Barten, Citation1969Citation1977), the market share analysis (Rao, Citation1972; Weiss, Citation1968), the production function estimation (Dhrymes, Citation1962), the translog utility function analysis (Berndt & Christensen, Citation1974), the linearised dynamic stochastic general equilibrium (DSGE) modelling (Bierens, Citation2007; Ireland, Citation2004), the errors-in-variables analysis with panel data (Biørn, Citation2000; Biørn & Klette, Citation1998; Wansbeek, Citation2001; Xiao, Shao, & Palta, Citation2010aCitation2010b; Xiao, Shao, Xu, & Palta, Citation2007), the multivariate random-effects meta-analysis models (Chen, Hong, & Riley,Citation2014; Riley, Abrams, Lambert, Sutton, & Thompson, Citation2007) and the non-Gaussian ARMA models (Alessi, Barigozzi, & Capasso, Citation2011; Leeper, Walker, & Yang, Citation2013; Mountford & Uhlig, Citation2009; Velasco & Lobato, Citation2018).

In a linear regression model with known singular disturbance covariance matrix, Theil (Citation1971) showed that a generalised Aitken-like estimator using the Moore–Penrose generalised inverse is best linear unbiased. Following Theil (Citation1971), Kreijger and Neudecker (Citation1977) proposed two optimality criteria to obtain best linear unbiased estimators. Within the same context, Dhrymes and Schwarz (Citation1987) discussed the existence issue of the estimators using generalised inverses. Haupt and Oberhofer (Citation2006) proposed an estimator which does not use the generalised inverses and allows for additional exogenous restrictions, collinearities and generalised adding-up. Bierens and Swanson (Citation2000) and Bierens (Citation2007) suggested that one can obtain parameter estimate by maximising the information content of the singular system. Ireland (Citation2004) and Lai (Citation2008) proposed adding random noises to the singular system to implement maximum likelihood estimation.

In the GMM literature, White (Citation1986) showed that if the estimating function g is of the form g=(g1,g2) such that: (i) Ω1=E[g1(Zi,β0)g1(Zi,β0)] is nonsingular and (ii) components of g2 are linear combinations of g1, then the efficient GMM estimator is the minimiser of (2) Jn(β)=ng¯n(β)Ωg¯n(β),(2) where g¯n(β)=1nni=1g(Zi,β) and Ω is a reflexive generalised inverse of (3) Ω=E[g(Zi,β0)g(Zi,β0)].(3) However, in practice, the aforementioned representation of g is generally not readily obtainable (see, e.g., Schneeweiss, Citation2014; Velasco & Lobato, Citation2018; Xiao et al., Citation2010b).

The purpose of this article is to develop an efficient GMM estimator for a singular system of moment conditions with general form. An earlier effort appeared in Xiao (Citation2008), which is proposed using the reflexive generalised inverses to deal with the singularity. Schneeweiss (Citation2014) independently discussed similar ideas.

The rest of the paper is organised as follows. In Section 2, we briefly review the GMM methodology, the concepts of generalised inverses and some results of the reflexive generalised inverses. We present our main result in Section 3. Section 4 discusses further issues such as the estimation of optimal weighting matrix and the method of adding noises, and Section 5 concludes. Proofs of results are relegated to the Appendix.

2. GMM and generalised inverses

We first make a brief introduction of the standard GMM method. For book-length detailed account, see Hall (Citation2005). For simplicity we assume that the data Z1,,Zn are i.i.d. Assume also that K>p, i.e., the model is over-identified. Since the number of restrictions on parameter is greater than the dimension of parameter, in general it is impossible to obtain an estimator of the parameter by using method of moments, i.e., by setting the sample moment g¯n equal to zero. The idea of GMM by Hansen (Citation1982) is to minimise a quadratic norm of g¯n: (4) Jn(β)=ng¯n(β)Wng¯n(β),(4) where Wn is a positive semidefinite matrix. Under a set of regularity conditions including that Ω being positive definite, and assuming Wn converges in probability to a positive semi-definite matrix W, βˆGMM, the minimiser of (Equation4), is a consistent estimator for β0 and has limiting distribution n(βˆGMMβ0)dN(0,V(W)), where V(W)=(GWG)1GWΩWG(GWG)1 with G=E[g(Zi,β0)β]. The lower bound of V(W) is achieved at W=Ω1, i.e., V(W)V(Ω1) in the sense of being nonnegative definite, for any W. In practice, a consistent estimator of Ω1 can be set as (5) Ωˆ1=[1nni=1[g(Zi,β~)g(Zi,β~)]]1,(5) where β~ is a consistent estimator of β0. A typical choice of β~ is a GMM estimator with W=IK, the identity matrix of order K. Note that Ωˆ1 converges in probability to Ω1 because 1nni=1[g(Zi,β~)g(Zi,β~)] converges in probability to Ω, and more importantly, Ω is positive definite.

Next we review the concepts of generalised inverses of a matrix and some of their properties.

Definition 2.1

Let A be a real l×s matrix. An s×l real matrix A may have one or all of the following properties:

  1. AAA=A;

  2.  AAA=A;

  3. (AA)=AA;

  4. (AA)=AA.

If A satisfies (i), it is called a generalised inverse of A; if A satisfies (i) and (ii), it is called a reflexive generalised inverse (or g2-inverse) of A; if A satisfies (i) –(iv), it is called the Moore–Penrose generalised inverse of A. The Moore–Penrose generalised inverse of a matrix A is unique and is denoted by A+ hereafter.Footnote1

We list some of the important properties of the generalised inverses by the following two propositions, proof of which can be achieved by direct verification and therefore is omitted.Footnote2 Proposition 1 states that when the matrix of interest has natural factorisation with certain structure, some of its generalised inverses can be easily derived.

Proposition 2.1

(i) Let Ω=[A1  A2]Ω1[A1  A2], where A1 and Ω1 are nonsingular square matrices. Then Ω=[(A11)Ω11A11000] is a reflexive generalised inverse of Ω.

(ii) Let Ω=AΩ1A, where A is of full column rank and Ω1 is nonsingular, then any reflexive generalised inverse Ω of Ω satisfies AΩA=Ω11. Moreover, we have A+=(AA)1A, and Ω+=(A+)Ω11A+=A(AA)1Ω11(AA)1A.

Proposition 2.2 points out that the generalised inverses (including the reflexive generalised inverses) are not unique and can be obtained by using the singular value decomposition .Footnote3

Proposition 2.2

Let Ω be an m×n real valued matrix with rank r>0. Suppose that the singular value decomposition of Ω is Ω=SΣT, where S is m×m with SS=Im, T is n×n with TT=In, and Σ=[ΣrOOO], with Σr the diagonal matrix of singular values of Ω and O the matrices of zeros. Then

(i) (6) G=T[Σr1XYZ]S(6) is a generalised inverse of Ω, where X, Y and Z are arbitrary real valued matrices with appropriate dimension.

(ii) (7) G=T[Σr1XYYΣr1X]S(7) is a reflexive generalised inverse of Ω, where X, Y are arbitrary real valued matrices with appropriate dimension.

(iii) (8) G=T[Σr1OOO]S(8) is the Moore–Penrose generalised inverse of Ω.

White (Citation1986) result on GMM estimation with singular moment conditions can be stated as follows:

Theorem 2.1

White, Citation1986

Suppose there exists a matrix Δ such that G=[Il1  Δ]G1 and Ω=[Il1  Δ]Ω1[Il1  Δ], where G1 is of full column rank and Ω1 is l1×l1 positive definite with l1p, then

(i) For any reflexive generalised inverse Ω of Ω, ΩG(GΩG)1G is independent of the choice of Ω, and ΩG(GΩG)1G0.

(ii) For any reflexive generalised inverse Ω of Ω, and for any W, (GWG)1GWΩWG(GWG)1(GΩG)1. Hence Ω is the optimal weighting matrix. In practice, one may choose the Moore–Penrose generalised inverse of Ω, or a special reflexive generalised inverse Ω=[Ω11000].

Remark

Note that (GΩG)1 – the asymptotic covariance matrix of the optimal GMM estimator – does not depend on Δ. The basic idea of Theorem 2.1 is as follows. Suppose we have two sets of instrumental variables, say Z1 and Z2, such that Z1 is linearly independent and Z2 is a linear combination of Z1, i.e., Z2=Z1α for some constant vector α, then one can ignore Z2 and use Z1 only as instruments, in doing so we achieve the same asymptotic efficiency as using Z=(Z1Z2). The limitation of this result is that to apply this method we have to sort all instrumental variables into two groups, such that instrumental variables in one group are linear combinations of those in the other group. This can be very tedious in practice. For example, in panel data models, there are often a very large number of instruments and it is in general impossible to sort out them.

3. Main results

We now establish some basic results about random vectors with singular covariance matrices.

Lemma 3.1

Let Y=[Y1Ym] be an m×1 random vector and r be the rank of the covariance matrix of Y. Suppose that r<m. Then

(i) There exist a r-dimensional subvector Y(r)=[Yi1Yir] of Y such that its covariance matrix var(Y(r)) is positive definite. The vector Y(r) is called an essential subvector of Y.

(ii) Let Y(r) be the (mr)×1 vector consisting of the remaining components of Y. Then there exist an (mr)×(mr) constant matrix C and an (mr)×1 constant vector d such that Y(r)=CY(r)+d,w.p.1., where w.p.1. means ‘with probability one’. Hence there exist an m×r constant matrix B of full column rank and an m×1 constant vector d~ such that Y=BY(r)+d~,w.p.1. (iii) If EY=0, then d~ in (ii) is the zero vector. i.e., Y=BY(r), w.p.1.

Theorem 3.1 is the main result of this paper.

Theorem 3.1

Consider GMM estimation for model (Equation1) with Ω defined by (Equation3). Suppose it is known that the components of g(Zi,β0) are linearly dependent (with probability one). Then any reflexive inverse of Ω is an optimal weighting matrix. Particularly, we can use the Moore–Penrose generalised inverse Ω+.

Let Ω be an arbitrary reflexive generalised inverse of Ω. Then the asymptotic variance matrix of the GMM estimator using Ω as the weighting matrix is V(Ω)=(GΩG)1GΩΩΩG(GΩG)1=(GΩG)1, where G=E[g(Zi,β0)β]. A natural question is whether V(Ω) is a constant matrix independent of the choice of Ω. The answer is yes. To see this, suppose the essential subvector of g(Zi,β0) is g(r)(Zi,β0)=(gi1(Zi,β0),,gir(Zi,β0)), with g(Zi,β0)=Bg(r)(Zi,β0) (w.p.1.). Let G1=E[g(r)(Zi,β0)β] and Ω1=var(g(r)(Zi,β0)). Then we have G=BG1, BΩB=Ω11, hence V(Ω)=(GΩG)1=(G1BΩBG1)1=(G1Ω11G1)1, which is independent of the choice of Ω, as the essential vector and the corresponding matrices G1 and Ω1 are unrelated to Ω. We also see that V(Ω) remains the same if we use another essential vector, as (GΩG)1=(GΩ+G)1 is unrelated to the choice of the essential vectors. More details can be found in the proof of Theorem 3.1 in the Appendix.

Judging from the asymptotic distributions we can see that GMM estimation using moment conditions (Equation1) and the reflexive generalised inverses as weighting matrix is asymptotically equivalent to the efficient GMM using moment conditions E[g(r)(Zi,β0)]=0. In some situations, one can figure out the essential subvector g(r)(Zi,β), then efficient GMM estimation can be based on E[g(r)(Zi,β0)]=0 directly. For instance, in the errors-in-variables analysis of panel data, Xiao et al. (Citation2010a) and Xiao et al. (Citation2010b) found that one can obtain g(r)(Zi,β) by using singular value decomposition. However, such simple decomposition is not available in general and it can be very inconvenient, if not impossible, to find the essential vector g(r)(Zi,β). Theorem 2 tells us that whatever this subvector is, the GMM estimator using any of the reflexive generalised inverses, and the Moore–Penrose generalised inverse in particular, as the weighting matrix will always have the same asymptotic variance as the efficient GMM based on E[g(r)(Zi,β0)]=0.

4. Further issues

4.1. Optimal weighting matrix estimation

Now we discuss consistent estimation of Ω+. Let β~ be a consistent estimator of β0 and Ωˆn=1nni=1[g(Zi,β~)g(Zi,β~)]. Then ΩˆnΩ in probability under normal regularity conditions. A natural candidate estimator of Ω+ is Ωˆn+.

It is well known that if a sequence of nonsingular square matrices {An} converges to a nonsingular square matrix A, then An1A1.Footnote4 However, if A is singular, and AnA, we may not necessarily have that An+A+. Footnote5 Assuming AnA and A is singular, then a necessary and sufficient condition for An+A+ is:

Theorem 4.1

Stewart, Citation1969

Let {An} be a sequence of real m×n matrices converging to a m×n matrix A. Then An+A+ if and only rank(An)=rank(A) for n large enough.

We now prove that Ωˆn+ is a consistent estimator for Ω+. By Theorem 4.1, we need only to show that rank(Ωˆn)=rank(Ω) when n is large enough. By Lemma 3.1, there exists a constant matrix B of full column rank, such that w.p.1., Ωˆn=1nni=1[Bgr(Zi,β~)g(r)(Zi,β~)B]=B(1nni=1[g(r)(Zi,β~)g(r)(Zi,β~)])B. Since the components of g(r)(Zi,β) are linear independent for any β, rank(1nni=1[g(r)(Zi,βˆ)g(r)(Zi,βˆ)])=r=rank(Ω), for any n. Therefore rank(Ωˆn)=rank(1nni=1[g(r)(Zi,βˆ)g(r)(Zi,βˆ)])=rank(Ω), and Ωˆn+ converges to Ω+ in probability.

Even though using generalised inverses is theoretically sound, it can be unstable, i.e., small perturbation of a singular matrix may result in large deviation from its generalised inverses.Footnote6 Therefore, one must be cautious when using generalised inverses. We suggest that one should first try to find the essential subvector g(r). In case that g(r) is not easily obtainable, the method introduced below can be used as an alternative to generalised inverses.

4.2. Imposing random noises

To avoid the potential bias caused by generalised inverses, we can add randomly generated noises to the system to make it nonsingular, as Bierens (Citation2007) and Lai (Citation2008) did in the maximum likelihood estimation of singular system of equations. Specifically, let U1,,Un be i.i.d. K×1 random vectors generated from the multivariate normal distribution with mean zero and covariance matrix σ2IK, and assume that U1,,Un are independent from Z1,,Zn. Define h(Zi,Ui,β)=g(Zi,β)+Ui, for i=1,,n. Then β0 is the solution of the set of moment conditions (9) E[h(Zi,Ui,β)]=0.(9) The set of moment conditions (Equation9) is nonsingular, since Σ=E[h(Zi,Ui,β0)h(Zi,Ui,β0)]=Ω+σ2IK>0. Let β~GMM be an efficient GMM estimator of β0 based on (Equation9), then the asymptotic distribution of β~GMM is n(β~GMMβ0)dN(0,(GΣ1G)1). Since (GΣ1G)1>(GΩ1G)1, for any σ>0, β~GMM is asymptotically less efficient than βˆGMM. However, the loss of efficiency can be controlled since (GΣ1G)1(GΩ1G)1, as σ0. Similar to Lai (Citation2008), one can also generate m independent samples of U1,,Un, obtain m GMM estimators β~GMM1,,β~GMMm and then construct a new estimator by β~GMMA=1mmj=1β~GMMj.Footnote7 Since β~GMMA combines information in β~GMM1,,β~GMMm, in theory it is asymptotically more efficient than any of β~GMM1,,β~GMMm. It is of interest to investigate the asymptotic distribution and finite sample performance of β~GMMA in a future study.

5. Concluding remarks

Since the moment condition models do not require researchers to specify the likelihood function of the data generating process, they have been widely used by econometricians to model economic theories. Though it is desirable that the moment conditions constructed from economic theory are linearly independent, in practice this may not always be the case. Sometimes singularity is inherent in the model or is caused by some singular transformations. In this paper, we extended the efficient GMM estimation to linearly dependent moment condition models. The result can be viewed as a natural extension of the standard GMM theory, since the generalised inverse of a matrix is a natural extension of the inverse of a matrix. Though in theory using generalised inverses yields efficient GMM estimators, in practice one must be cautious of using them, in light of the following two concerns. First, using generalised inverses ignores the intrinsic structure of the moment conditions, which sometimes contains important information. Second, the generalised inverses of a singular matrix are unstable, which could induce serious bias of the resulting GMM estimator. Therefore when there is singularity in the system, a practical strategy is to obtain an essential moment vector and apply GMM to it. In case an essential moment vector is not available, we can add random noises to the moment conditions and obtain GMM estimators based on the new set of moment conditions. We suggest using generalised inverses with discretion. The results in this paper might also shed light on other popular statistical methods (such as the empirical likelihood) for estimating equations with singularity.

Disclosure statement

No potential conflict of interest was reported by the author.

Additional information

Funding

The research is supported by the National Natural Science Foundation of China (NSFC grant: 71661137005, 71473040 and 11571081).

Notes on contributors

Zhiguo Xiao

Dr. Zhiguo Xiao received his PhD in statistics from the University of Wisconsin-Madison. His research interest includes both theoretical and applied econometrics.

Notes

3 Similar results for square matrices can be found in Bapat (Citation2012, pp. 47–48).

2 More results on reflexive generalised inverses can be found in Rao and Mitra (Citation1971), Rao (2001), Bapat (Citation2012), Fampa and Lee (Citation2018), and Xu, Fampa, and Lee (Citation2019).

1 For the existence and uniqueness of the Moore–Penrose generalised inverse, see, e.g., Penrose (Citation1955) and Abadir and Magnus (Citation2005, pp. 284–285).

5 For example, consider An=[11n11n211n21] and A=[1111]. Then AnA. Since An is invertible, An+=An1[]. Hence An+A+=[12121212].

4 A sequence of real m×n matrices {An} is said to converge to a m×n matrix A if AnA0, where is a matrix norm, such as the Euclidean norm or A=supx=1{Ax}.

6 For example, consider a matrix A=[100]. Its Moore–Penrose generalised inverse is A+=A. Adding a small number ε=103 to the first two diagonal entries of A, we obtain a new Moore–Penrose generalised inverse [0.99910000].

7 This idea is similar to bootstrap aggregating (bagging).

References

  • Abadir, K. M., & Magnus, J. R. (2005). Matrix algebra. New York: Cambridge University Press.
  • Alessi, L., Barigozzi, M., & Capasso, M. (2011). Non-fundamentalness in structural econometric models: A review. International Statistical Review, 79, 16–47. doi: 10.1111/j.1751-5823.2011.00131.x
  • Bapat, R. B. (2012). Linear algebra and linear models (3rd ed.). London: Springer-Verlag.
  • Barten, A. (1969). Maximum likelihood estimation of a complete system of demand equations. European Economic Review, 1, 7–73. doi: 10.1016/0014-2921(69)90017-8
  • Barten, A. (1977). The system of consumer demand function approach: A review. Econometrica, 45, 23–51. doi: 10.2307/1913286
  • Berndt, E. R., & Christensen, L. R. (1974). Testing for the existence of a consistent aggregate index of labor inputs. American Economics Review, 64, 391–403.
  • Bierens, H. J. (2007). Econometric analysis of linearized singular dynamic stochastic general equilibrium models. Journal of Econometrics, 136, 595–627. doi: 10.1016/j.jeconom.2005.11.008
  • Bierens, H. J., & Swanson, N. R. (2000). The econometric consequences of the ceteris paribus condition in economic theory. Journal of Econometrics, 95, 223–253. doi: 10.1016/S0304-4076(99)00038-X
  • Biørn, E. (2000). Panel data with measurement errors: Instrumental variables and GMM procedures combining levels and differences. Econometric Reviews, 19(4), 391–424. doi: 10.1080/07474930008800480
  • Biørn, E., & Klette, T. (1998). Panel data with errors-in-variables: Essential and redundant orthogonal conditions in GMM-estimation. Economics Letters, 59, 275–282. doi: 10.1016/S0165-1765(98)00053-6
  • Chen, Y., Hong, C., & Riley, R. D. (2014). An alternative pseudolikelihood method for multivariate random-effects meta-analysis. Statistics in Medicine, 34(3), 361–380. doi: 10.1002/sim.6350
  • Dhrymes, P. J. (1962). On devising unbiased estimators for the parameters of a Cobb–Douglas production function. Econometrica, 30, 297–304. doi: 10.2307/1910218
  • Dhrymes, P. J., & Schwarz, S. (1987). On the existence of generalized inverse estimators in a singular system of equations. Journal of Forecasting, 6, 181–192. doi: 10.1002/for.3980060304
  • Fampa, M., & Lee, J. (2018). On sparse reflexive generalized inverses. Operations Research Letters, 46(6), 605–610. doi: 10.1016/j.orl.2018.09.005
  • Hall, A. (2005). Generalized method of moments. Oxford University Press: New York.
  • Hansen, L. (1982). Large sample properties of generalized method of moments estimators. Econometrica, 50, 1029–1054. doi: 10.2307/1912775
  • Hansen, B., & West, K. (2002). Generalized method of moments and macroeconomics. Journal of Business & Economic Statistics, 20, 460–469. doi: 10.1198/073500102288618603
  • Haupt, H., & Oberhofer, W. (2006). Generalized adding-up in systems of regression equations. Economics Letters, 92, 263–269. doi: 10.1016/j.econlet.2006.03.001
  • Ireland, P. N. (2004). A method for taking models to the data. Journal of Economic Dynamics and Control, 28, 1205–1226. doi: 10.1016/S0165-1889(03)00080-0
  • Kreijger, R. G., & Neudecker, H. (1977). Exact linear restrictions on parameters in the general linear model with a singular covariance matrix. Journal of the American Statistical Association, 72, 430–432. doi: 10.1080/01621459.1977.10481014
  • Lai, H. (2008). Maximum likelihood estimation of singular systems of equations. Economics Letters, 99, 51–54. doi: 10.1016/j.econlet.2007.05.027
  • Leeper, E. M., Walker, T. B., & Yang, S.-C. S. (2013). Fiscal foresight and information flows. Econometrica, 81, 1115–1145. doi: 10.3982/ECTA8337
  • Mountford, A., & Uhlig, H. (2009). What are the effects of fiscal policy shocks?. Journal of Applied Econometrics, 24, 960–992. doi: 10.1002/jae.1079
  • Penrose, R. (1955). A generalized inverse for matrices. Mathematical Proceedings of the Cambridge Philosophical Society, 51, 406–413. doi: 10.1017/S0305004100030401
  • Rao, C. R. (1972). Alternative econometric models of sales advertising relationships. Journal of Marketing Research, 9, 171–181. doi: 10.1177/002224377200900209
  • Rao, C. R., & Mitra, S. K. (1971). Generalized inverse of matrices and its applications. John Wiley & Sons, Inc.: New York.
  • Riley, R., Abrams, K., Lambert, P., Sutton, A., & Thompson, J. (2007). An evaluation of bivariate random-effects meta-analysis for the joint synthesis of two correlated outcomes. Statistics in Medicine, 26(1), 78–97. doi: 10.1002/sim.2524
  • Schneeweiss, H. (2014). The linear GMM model with singular covariance matrix due to the elimination of a nuisance parameter. Technical Report 165, Dept. Statistics, Univ. Munich.
  • Stewart, G. W. (1969). On the continuity of the generalized inverse. SIAM Journal of Applied Mathematics, 17, 33–45. doi: 10.1137/0117004
  • Theil, H. (1971). Principle of econometrics. John Wiley & Sons: New York.
  • Velasco, C., & Lobato, I. N. (2018). Frequency domain minimum distance inference for possibly noninvertible and noncausal ARMA models. The Annals of Statistics, 46(2), 555–579. doi: 10.1214/17-AOS1560
  • Wansbeek, T. J. (2001). GMM Estimation in panel data models with measurement error. Journal of Econometrics, 104, 259–268. doi: 10.1016/S0304-4076(01)00079-3
  • Weiss, D (1968). Determinants of market share. Journal of Marketing Research, 5, 290–295. doi: 10.1177/002224376800500307
  • White, H. (1986). Instrumental variables analogs of generalized least squares estimators. Advances in Statistical Analysis and Statistical Computing, 1, 173–227.
  • Xiao, Z. (2008). Generalized inverses in GMM estimation with redundant moment conditions. In Topics in Generalized Method of Moments Estimation with Application to Panel Data with Measurement Error, University of Wisconsin-Madison PhD thesis.
  • Xiao, Z., Shao, J., & Palta, M. (2010a). GMM in linear regression for longitudinal data with multiple covariates measured with error. Journal of Applied Statistics, 37, 791–805. doi: 10.1080/02664760902890005
  • Xiao, Z., Shao, J., & Palta, M. (2010b). Instrumental variable and GMM estimation of panel data with measurement error. Statistica Sinica, 20(4), 1725–1747.
  • Xiao, Z., Shao, J., Xu, R., & Palta, M. (2007). Efficiency of GMM estimation in panel data models with measurement error. Sankhya: The Indian Journal of Statistics, 69, 101–118.
  • Xu, L., Fampa, M., & Lee, J. (2019). Aspects of symmetry for sparse reflexive generalized inverses. arXiv:1903.05744v1[math.OC].

Appendix

Proofs of results

Proof

Proof of Lemma 1

Let Γ=var(Y), and Γ=T(Λr000)T be the spectrum decomposition of Γ, with TT=TT=Im. Let T=[T1 T2], then TT=[T1T2][T1 T2]=[T1T1T1T2T2T1T2T2]=Im,Γ=T[Λr000]T=T1ΛrT1, hence T2T1=0, and var(T2Y)=T2VT2=T2T1ΛrT1T2=0. Hence there exists an (mr)×1 constant vector c such T2Y=c. Let T2=[t1tm], with rank(T2)=q:=mr. Suppose tj1 tjq are linearly independent, then T2Y=[t1tm][Y1Ym]=t1Y1++tmYm=[tj1tjq][Yj1Yjq]+TˆY(q), hence [tj1tjq][Yj1Yjq]+TˆY(q)=c, i.e., Y(q)=[Yj1Yjq]=[tj1tjq]1TˆY(q)+[tj1tjq]1c. Let C=[tj1tjq]1Tˆ,d=[tj1tjq]1c, we get Y(q)=CY(q)+d, i.e., Y(r)=CY(r)+d. Since Y(r) and Y(r) are subvectors of Y, there exists a m×m nonsingular matrix A such that Y=A[Y(r)Y(r)]=A[Y(r)CY(r)+d]=A[IC]Yr+A[0d], i.e., Y=BY(r)+d~, with B=A[IC],d~=A[0d]. Hence var(Y)=Bvar(Y(r))B. Since B=A[IC] is of full column rank, rank(var(Y))=rank(var(Y(r)))=r. This shows that var(Y(r)) is nonsingular.

Proof

Proof of Theorem 3.1

Let V(W) denote the asymptotic variance of the GMM estimator using weighting matrix W. Then V(W)=(GWG)1GWΩWG(GWG)1. Let Ω be a reflexive generalised inverse of Ω. Then we have V(Ω)=(GΩG)1GΩΩΩG(GΩG)1=(GΩG)1. Hence V(W)V(Ω)=(GWG)1GW[ΩG(GΩG)1G]WG(GWG)1. To establish V(W)V(Ω)0, we just need to show that ΩG(GΩG)1G0. Let rank(Ω)=r. By Lemma 3.1, there exist a subvector g(r)(Zi,β0)=(gi1(Zi,β0),,gir(Zi,β0)) and a matrix B of full column rank such that g(Zi,β0)=Bgr(Zi,β0) a.s., with Ω1=var(g(r)(Zi,β0)) positive definite. Then Ω=var(g(Zi,β0))=BΩ1B, and G=E[g(Zi,β0)β]=BG1, with G1=E[g(r)(Zi,β0)β]. Hence ΩG(GΩG)1G=BΩ1BBG1(GΩG)1G1B=B[Ω1G1(GΩG)1G1]B.So we just need to show that Ω1G1(GΩG)1G10. By Proposition 2.1, BΩB=Ω11, hence Ω1G1(GΩG)1G1=Ω1G1(G1BΩBG1)1G1=Ω1G1(G1Ω11G1)1G1=Ω112[IΩ112G1(G1Ω11G1)1G1Ω112]Ω1120, since IΩ112G1(G1Ω11G1)1G1Ω112 is idempotent and symmetric.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.