692
Views
0
CrossRef citations to date
0
Altmetric
Research Article

A method to evaluate the rank condition for CCE estimators

, &

Abstract.

We develop a binary classifier to evaluate whether the rank condition (RC) is satisfied or not for the Common Correlated Effects (CCE) estimator. The RC postulates that the number of unobserved factors, m, is not larger than the rank of the unobserved matrix of average factor loadings, ϱ. When this condition fails, the CCE estimator is inconsistent, in general. Despite its importance, to date this rank condition could not be verified. The difficulty lies in the fact that factor loadings are unobserved, such that ϱ cannot be directly determined. The key insight in this article is that ϱ can be consistently estimated with existing techniques through the matrix of cross-sectional averages of the data. Similarly, m can be estimated consistently from the data using existing methods. Thus, a binary classifier, constructed by comparing estimates of m and ϱ, correctly determines whether the RC is satisfied or not as (N,T). We illustrate the practical relevance of testing the RC by studying the effect of the Dodd-Frank Act on bank profitability. The RC classifier reveals that the rank condition fails for a subperiod of the sample, in which case the estimated effect of bank size on profitability appears to be biased upwards.

1. Introduction

In a seminal paper, Pesaran (Citation2006) put forward the Common Correlated Effects (CCE) approach for NT-consistent estimation of panel data models with a multifactor error structure. The method aims to control for the unobserved common factors by augmenting the regression model with cross-sectional averages (CSA) of the observables. The CCE estimator has been applied to a large range of fieldsFootnote1, and it has also been extended to several theoretical settings.Footnote2 Such popularity of CCE can be attributed to the computational simplicity as well as the excellent finite-sample performance of the estimator in stylised setups.

Nevertheless, CCE comes at a cost. In particular, the CSA of the observables are valid proxies for the unobserved factors only if the number of factors, m, does not exceed the rank of the matrix of averaged factor loadings, ϱ. This so-called “rank condition” (RC) implies that there exist at least as many observables holding linearly independent information about the unobserved factors as there are factors. Westerlund and Urbain (Citation2013) demonstrate that the CCE estimator is inconsistent when the RC fails and the factor loadings are correlated with the regressors. Furthermore, Karabiyik et al. (Citation2019) and Juodis et al. (Citation2021) show that even when the factor loadings are uncorrelated with the regressors, failure of the RC leads to a lower rate of consistency for the CCE estimator.

Despite the importance of the RC for the asymptotic properties of the CCE estimator, practitioners typically take this assumption for granted. The main reason is that the matrix of average factor loadings is unobserved and therefore its rank cannot be directly evaluated or estimated.

This article puts forward a binary classifier that evaluates the rank condition. The key insight is that the rank of the unobserved matrix of average factor loadings, ϱ, can be established from the matrix of CSA of the data. We shall show that ϱ can be estimated consistently using existing procedures developed for determining the true rank of an unknown matrix; see e.g., Camba-Mendez and Kapetanios (Citation2009) and Al-Sadoon (Citation2017) for an overview of this literature. Similarly, the number of factors, m, can be estimated from the data in a straightforward manner using existing methods, such as those developed by Onatski (Citation2010), Ahn and Horenstein (Citation2013), and Kapetanios (Citation2010), among many others. Comparing consistent estimates of m and ϱ, m^ and ϱ^, respectively, the rank condition is deemed to be satisfied when the classifier RC^1\mathbbm1{ϱ^<m^}=1, where 𝟙{⋅} is an indicator function that returns 1 when the argument inside the curly brackets holds true and 0 otherwise. RC^ is shown to be consistent, i.e., it correctly determines whether the rank condition is satisfied or not, with probability 1 as (N,T).

When the RC is violated for the standard CCE approach, one can augment the model with additional CSA that contain new information about the factors. Several potential augmentations have been suggested (see e.g., Chudik and Pesaran, Citation2015; Juodis, Citation2022; Karabiyik et al., Citation2019). However, it is not always clear which set of additional CSA to choose, and whether the selected augmentation is sufficient to restore the RC.Footnote3 To address these issues, we put forward a strategy that combines the classifier proposed in this article and the IC criterion of Karabiyik et al. (Citation2019). The resulting procedure enables consistent CCE estimation of panel data models with a multifactor error structure, even in cases where the rank condition fails for the original CCE estimator.

We illustrate the practical relevance of our RC classifier and augmentation strategy by studying the effect of the Dodd-Frank Act of 2010 on bank profitability. In particular, based on a random sample of 450 banks, we analyze bank profitability conditional on several potential drivers, controlling for macro-risk factors and common shocks. To examine the impact of the Dodd-Frank Act, we estimate the model separately over two subperiods, namely 2006:Q1-2010:Q4 and 2011:Q1-2019:Q4. The RC classifier reveals that the rank condition fails for the first subperiod. By augmenting the standard set of CSA using external variables, our procedure is able to restore the rank condition. This proves to be important because the estimated effect of bank size on profitability is significantly lower when the RC is restored.

In what follows, we will use 𝐀 to denote the Moore-Penrose pseudo-inverse of the matrix 𝐀, rk(𝐀) for its rank, |A| for the determinant and A=[tr(AA)]1/2 for its Euclidean (Frobenius) matrix norm. For an n×n matrix 𝐀, the λ1(A)λ2(A)λn(A) denote its n ordered eigenvalues. A vec(.) denotes the vectorization operation. Finally, ⌊a⌋ (⌈a⌉) is the floor (ceiling) function, which yields the largest (smallest) integer less than (greater than) or equal to a.

2. A multifactor panel data model and CCE

2.1. Model and assumptions

We study the following linear regression model with unobserved common factors (1) yi=Xiβ+Fλi+εi,(1) where yi=[yi1,,yiT] denotes a T×1 vector of observations on the dependent variable for individual i, Xi=[xi1,,xiT] denotes a T×K matrix of covariates, where 𝐱it is K×1, and 𝜷 is a K×1 vector of unknown parameters of interest with β<. The error term is composite, such that F=[f1,,fT] denotes a T×m matrix of unobserved common factors, where 𝐟t is m×1, and 𝝀i denotes an m×1 vector of factor loadings. The dimension m is fixed and finite. Finally, εi=[εi1,,εiT] is a T×1 vector of purely idiosyncratic disturbances.

Following Pesaran (Citation2006), we assume that the covariates are also subject to a common factor structure, such that the data generating process (DGP) for 𝐗i is given by (2) Xi=FΓi+Vi,(2) where 𝚪i denotes an m×K matrix of factor loadings, and Vi=[vi1,,viT] is a T×K matrix of idiosyncratic errors.

Replacing 𝐗i in Eq. (Equation1) by the expression in Eq. (Equation2), and stacking the observables into a T×(1+K) matrix Zi=[yi,Xi][zi1,...,ziT], yields (3) Zi=FCi+Ui,(3) where Ci=[δi,Γi] is of order m×(1+K) with δi=λi+Γiβ, and Ui=[εi+Viβ,Vi]. In what follows, it is important to note that 𝐂i can be written as Ci=C~iB, with (4) C~i=[λi,Γi];B=[101×KβIK].(4)

Therefore, since 𝐁 has full rank, the rank of 𝐂i is solely determined by the matrix of factor loadings C~i.

The following assumptions are made throughout the article:

Assumption 1.

(Idiosyncratic errors) εit and vit are mean zero, covariance-stationary, and independent across i, with E(εit4)< and E(vit4)< for all i and t. Let U=[U1,,UN] be such that λ1(UU/M)=Op(1) and λdcM~(UU/M)c+op(1) for some real c > 0 and dc[0,1), where M=max{N,T} and M~=min{N,T}.

Assumption 2.

(Common factors) ft is covariance-stationary with E(ft4)< and absolute summable autocovariances. In addition, rk(F) = m and T1FFΣF as T→∞, where ΣF is positive definite.

Assumption 3.

(Factor loadings) C~i is generated according to (5) C~i=C~+Ξi;ξii.i.d.(0m(K+1),Ωξ),(5)

where C~=E(C~i)[λ,Γ] such that C~<, 𝛏i = 𝑣𝑒𝑐(Ξi) and Ωξ=E(ξiξi) with Ωξ<. In addition, 1Ni=1NCiCiΣC as N→∞, with ΣC positive definite.

Assumption 4.

(Independence) ft,εis,vjl,ξh are mutually independent for all t,i,s,j,l,h.

The setup described by the DGP in Eq. (Equation3) together with Assumptions 14, is similar to that in Pesaran (Citation2006) but deviates in the following respects. First, we focus on a model with homogeneous slope coefficients and without fixed effects. This is for ease of exposition only, as the results below also follow through under the assumption of independent random coefficients with a common mean, as in Pesaran (Citation2006). See Section 3.5 for a discussion, and also Appendix C.2 for simulation evidence. Second, following Westerlund and Urbain (Citation2013) and Karabiyik et al. (Citation2019), Assumption 3 generalizes Pesaran (Citation2006) by allowing 𝝀i and 𝚪i to be mutually correlated, although uncorrelated across cross-sectional units i. Third, we introduce more explicit regularity conditions on the innovations, factors and their loadings compared to what is typically the case in the CCE literature. In particular, Assumption 1 places restrictions on the eigenvalues of the innovation covariance matrix as in Ahn and Horenstein (Citation2013). They exclude the presence of factors in the innovations and bound the largest dcM~ eigenvalues away from zero. For the factors and loadings, the non central second moments are assumed to converge to a positive definite matrix. These regularity conditions are common in the factor literature (see e.g., Ahn and Horenstein, Citation2013; Bai and Ng, Citation2002) and allow us to consistently estimate m. Lastly, note that rk(𝐅) = m of Assumption 2 implies that Tm.

2.2. CCE and the rank condition

Since 𝐅 enters into the data generating process of both 𝐲i and 𝐗i, and 𝝀i and 𝚪i are allowed to be mutually correlated, 𝐗i is endogenous. Therefore, standard panel data estimators, such as the two-way fixed effects estimator, fail to be consistent for the parameters of interest, 𝜷. The key idea of CCE is to replace 𝐅 with CSA of the observables in Eq. (Equation3).

In particular, taking sample averages over i in Eq. (Equation3), we obtain (6) ZT×(K+1)=FT×m Cm×(K+1)+UT×(K+1),(6) where Z=[z1,,zT], U=[u1,,uT] and bars denote CSA as in Z=1Ni=1NZi.

Under Assumptions 14 it is easy to show that C=C+Op(N1/2), where 𝐂 = E(𝐂i), and ut=Op(N1/2) for all t=1,,T. As a result, the observed CSA converge to a linear combination of the m common factors at every t=1,,T: (7) zt=Cft+(CC)ft+ut=Cft+Op(N1/2).(7)

Suppose that 𝐂 has full rank such that 𝐂𝐂 is invertible and bounded by Assumption 3. Pre-multiplying Eq. (Equation7) by 𝐂 and solving for 𝐟t yields (8) ft=(CC)1Czt+Op(N1/2).(8)

Hence, as N→∞, the common factor component at time t can be controlled for (or estimated) with the cross-section averages zt.

The pooled CCE estimator for 𝜷 is the least-squares estimator given by (9) β^=(i=1NXiMXi)1i=1NXiMyi,(9) where M=ITZ(ZZ)Z.Footnote4

The above idea of estimating factors with CSA crucially relies upon the assumption that 𝐂 has full rank. This restriction, known as the “rank condition” (RC), corresponds to (10) ϱ=m(10) where ϱ = rk(𝐂). When ϱ < m, the RC fails and the CCE estimator is generally inconsistent. This is because the CSA do not contain enough information on 𝐟t, which implies that the factor estimator in Eq. (Equation8) does not exist.

There are several cases where the RC may fail. To begin with, such failure occurs when m>K+1, i.e., the number of factors is larger than the number of CSA, in which case ϱmin{m,K+1}=K+1<m. In addition, although K+1m is a necessary condition for the RC to hold, it is by no means sufficient. For example, certain columns of Z can be asymptotically uninformative because the corresponding observables: (i) do not load on the common factors (e.g., some of the columns in 𝚪i equal zero); (ii) have factor loadings that average out (e.g., Γ=Op(N1/2)); or (iii) do not contain information on the common factors that is distinct from that already provided by other observables. In all these cases, the number of columns that are informative to estimate 𝐟t, as measured by ϱ, can be lower than m.Footnote5

3. Evaluating the rank condition

Despite the importance of the RC for the properties of the CCE estimator, this assumption is typically taken for granted. The main reason is that the population mean of the matrix of factor loadings, 𝐂, is unobserved and therefore its rank cannot be directly evaluated or estimated. The key insight of this article is that ϱ can be determined by estimating the rank of Z using existing techniques. Given a consistent estimate of ϱ, the RC is evaluated by direct comparison of that value with a consistent estimate for m. The latter can be determined from the observed data in a straightforward manner, based on e.g., Bai and Ng (Citation2002), Alessi et al. (Citation2010), Onatski (Citation2010), and Ahn and Horenstein (Citation2013).

The following two sections provide details for consistent estimation of ϱ and m as (N,T). Section 3.3 puts forward a binary classifier that evaluates the rank condition correctly with probability 1 as (N,T). Section 3.4 discusses a strategy for obtaining a consistent CCE estimator when the RC fails.

3.1. Consistent estimation of ϱ

We make use of the fact that the rank of an unobserved matrix 𝐀 can be determined through a N-consistent estimator of that matrix AN=A+Op(N1/2); see e.g., Robin and Smith (Citation2000) and Kleibergen and Paap (Citation2006). Noting that Z=FC+Op(N1/2) is N-consistent for 𝐅𝐂 (for T fixed), it follows from the rank equivalence (11) rk(FC)=rk(C)=ϱ,(11) that ϱ can be consistently estimated by applying a rank estimator to Z.Footnote6

Many popular rank estimators are based either on sequential testing procedures, as e.g., in Chen and Fang (Citation2019), or on information criteria (IC), see e.g., Cragg and Donald (Citation1997). Camba-Mendez and Kapetanios (Citation2009) provide an overview and conclude that sequential testing procedures have an advantage over IC methods under several modeling scenarios. Therefore, in the remainder of this section, we closely follow the sequential testing procedure developed by Robin and Smith (Citation2000). This is easy to implement and relies on relatively mild assumptions, in that it does not require the variance-covariance of the estimator of the unknown matrix 𝐅𝐂 to be full rank, or its rank to be known.

A major complication that arises in our setting, however, is that unlike Robin and Smith (Citation2000), where the dimensions of the target matrix are fixed as the sample size grows, here 𝐅𝐂 and its estimator Z are of order T×n.Footnote7 Therefore, the number of rows increases with the time dimension such that Z=FC+Op(TN1/2) is not N-consistent when (N,T). To circumvent this issue, we introduce a narrow matrix 𝚿 of order n×T, such that ΨZ is n×n and rk(𝚿𝐅𝐂) = rk(𝐅𝐂). That is, 𝚿 has the role of reducing the dimension of Z without altering the rank of the matrix it estimates. The following assumption is imposed:

Assumption 5.

(Dimension reduction matrix) As (N,T), Ψ satisfies (i) ΨF=Op(1);(ii)ΨU=Op(N1/2);(iii) Nvec(ΨZΨFC)LN(0,Ω).

Assumption 5 places additional restrictions on the potential choices for 𝚿, besides it being rank preserving. Assumption 5(i) requires that the entries of 𝚿 are sufficiently bounded. Assumption 5(ii) states that 𝚿 is asymptotically uncorrelated with U, the error term in Eq. (Equation6). Assumption 5(iii) ensures that, by application of a suitable central limit theorem, ΨZ remains a N-consistent estimator for 𝚿𝐅𝐂 and is asymptotically normally distributed with variance 𝛀 as (N,T). This assumption is identical to Assumption 2.2 in Robin and Smith (Citation2000), except that it is imposed on ΨZ rather than Z itself.

In practice, there exist several options for 𝚿 that satisfy the rank preservation condition and Assumption 5. One option is to set Ψ=T1/2Φ, with the entries of 𝚽 drawn from the standard normal distribution. The following theorem confirms that this choice is rank-preserving and satisfies the required consistency and boundedness conditions of Assumption 5. In turn, Assumption 5 (iii) is easily seen to hold by application of a CLT.

Theorem 1.

Let T > n and 𝚽 be a n×T random matrix with i.i.d. standard normal entries. (i) For a T×n matrix A, it holds that Pr[rk(ΦA)=rk(A)]=1.

(ii) Let Ψ=T1/2Φ. Under Assumptions 1-4, as (N,T), it follows that ΨZ=ΨFC+Op(N1/2),

where ΨFC=Op(1).

The proof of Theorem 1 is in Appendix B.

Remark 3.1.

An alternative stochastic option would be Ψ=Z/T. However, this is ruled out because even though ΨZ=ZZ/T is stochastically bounded and has the same rank as Z, it does not have an asymptotic normal distribution, i.e., it violates Assumption 5(iii).

The 𝚿 matrix can also be deterministic. Since n time periods contain the same information on the rank of 𝐂 as do T observations, an obvious candidate is Ψ=[0n×(Tn), In], which considers only the last n rows of Z. One can also take averages over every n-th row in Z by setting Ψ=1T/n[ιT/nIn]IT/nn,T, where 𝜾a is an a×1 vector of ones.

Given the above, we propose estimating the rank of the n×n matrix ΨZ by sequentially testing the null hypothesis H0:ϱ=ϱ against the alternative Ha:ϱ>ϱ, using the following statistic: (12) τ=N=ϱ+1nλ(A),(12) where λ1(A)λn(A) are the ordered eigenvalues of AΨZZΨ. The procedure is implemented sequentially for ϱ=0,,n1 and the estimated rank ϱ^ corresponds to the smallest value of ϱ* for which the null hypothesis is not rejected. Under the null, τ has a limiting distribution which is a weighted sum of independent χ12 variables, with weights given by the (nϱ)2 largest eigenvalues of (DϱRϱ)Ω(DϱRϱ), where Dϱ and Rϱ denote the eigenvectors corresponding to the nϱ smallest eigenvalues of ZΨΨZ and 𝐀, respectively. We shall assume in accordance with Robin and Smith (Citation2000) that rk[(DϱRϱ)Ω(DϱRϱ)]>0.

The asymptotic variance 𝛀 of ΨZ is unknown but can be estimated consistently byFootnote8: (13) Ω^=1Ni=1Nvec(ΨZiΨZ)vec(ΨZiΨZ).(13)

As discussed in Robin and Smith (Citation2000), the estimator for ϱ obtained from the test sequence is consistent when the employed significance level αN vanishes at an appropriate rate with N. This is because αN is the probability of over-estimating the true rank, Pr(ϱ^>ϱ), which must tend to zero for consistency. The authors show that αN = o(1) and N1lnαN=o(1) are sufficient for consistency. This is summarized in the following proposition:

Proposition 1.

Let Ass.1-5 hold and rk[(DϱRϱ)Ω(DϱRϱ)]>0. Provided that αN = o(1) and N1lnαN=o(1) as N→∞, it follows that ϱ^ϱ=op(1).

The proposition follows from the arguments of Theorem 5.2 in Robin and Smith (Citation2000), mutatis mutandis. We omit the proof to save space.

Remark 3.2.

Clearly, αN has to vanish sufficiently fast with N to limit the over-estimation frequency, but not too fast, as this results in under-estimation when N is small. We suggest to specify the nominal level as αN=αcN1/γ. This way, for a given choice of α and γ, the small N significance level is controlled through c > 1, whereas the speed at which αN decreases with N is governed by γ > 0. For instance, choosing α=5% and setting c = 20, γ = 1 fixes the nominal level to 5% for N = 20 and lets it decrease at rate N. Given that over-estimating ϱ may lead to false conclusions that the rank condition holds, we prefer a conservative estimator through a fast decrease with N (i.e., requiring strong evidence against the null before rejecting it in favor of a higher rank estimate).

3.2. Consistent estimation of m

Existing methods to estimate the number of factors from observed data rely on one of the following three approaches: looking at differences or ratios of adjacent eigenvalues (Ahn and Horenstein, Citation2013; Onatski, Citation2010), specifying threshold functions to separate bounded from unbounded eigenvalues (Alessi et al., Citation2010; Bai and Ng, Citation2002) or sequential tests to determine which eigenvalues are unbounded (Kapetanios, Citation2010; Trapani, Citation2018). Preliminary simulation evidence conducted for this article shows that the Growth Ratio (GR) by Ahn and Horenstein (Citation2013) performs well in finite samples and outperforms other estimators.Footnote9 Therefore, in what follows we propose estimating m using the GR statistic.

In particular, let Z=[Z1,,ZN] denote a T×(K+1)N matrix, where 𝐙i (defined in Eq. (Equation3)) collects all observables for individual i in a T×(K+1) matrix. Also, let mmax denote the maximum value of m considered in estimation, such that mmaxm. We define (14) m^=argmaxj{1,,mmax}GR(j);GR(j)=ln(V(j1)/V(j))ln(V(j)/V(j+1)),(14) where V(j)=k=j+1hλk(ZZ/NT) with h=min{T,N(K+1)}, and λk(𝐙𝐙NT) denotes the kth largest eigenvalue of (𝐙𝐙NT).

The GR statistic is easy to compute because it involves maximizing the “growth ratio” of two adjacent eigenvalues arranged in descending order. The main intuition is that the growth ratios of two adjacent eigenvalues of 𝐙𝐙NT are asymptotically bounded, except for the growth ratio involving the mth and (m+1)th eigenvalues, which diverges to infinity.

Under regularity conditions implied by Assumptions 14, Ahn and Horenstein (Citation2013) showFootnote10 (15) limmin{N,T}Pr(m^=m)=1,(15) for any mmax{m,(dcmin{N,T})m1}, where dc(0,1].

Remark 3.3.

In exactly the same way as described above, the number of factors can also be estimated based on the T×T matrix 𝐘𝐘NT, where Y=[y1,,yN] is of dimension T×N. However, since both 𝐲i and 𝐗i share the same factors by assumption, it is natural to combine them together in order to increase the information set used to construct proxies for 𝐅. This strategy is in line with the rationale behind the CCE approach, which involves solving a system of equations, such that Eq. (Equation3) includes LHS variables (observables) that are solely driven by a common factor component and purely idiosyncratic noise. Moreover, this strategy is consistent with Westerlund and Urbain (Citation2015), who also estimate factors based on 𝐙𝐙NT.

3.3. A consistent classifier for the rank condition

Given consistent estimates ϱ^ and m^ of the rank of 𝐂 and the number of factors, the rank condition is deemed to be violated when ϱ^<m^. We define the following classifier: (16) RC^11{ϱ^<m^},(16) where 𝟙{⋅} is an indicator function that returns 1 when the argument inside the curly brackets holds true, and 0 otherwise. Hence, if RC^=1 the rank condition is considered to be satisfied, whereas RC^=0 indicates that (Equation10) may be violated. The definition in (Equation16) shows that we also take ϱ^>m^ as a sign that (Equation10) is satisfied.Footnote11

Let RC1\mathbbm1{ϱ<m} denote the indicator of the true state of the rank condition in the population. The following proposition summarizes the asymptotic properties of the classifier.

Proposition 2.

Let Assumptions 15 hold true. Suppose also that ϱ is determined based on the sequential testing procedure outlined in Section 3.1, with αN = o(1), and N1lnαN=o(1), and m is determined by Eq. (Equation14) with mmaxm. Then, Pr[RC^=RC]1 as (N,T).

That is, the probability that the classifier correctly identifies whether the rank condition is satisfied or not, converges to unity. The result follows directly from the consistency of ϱ^ as N→∞ under Assumptions 1-5, given an appropriate rate of decay for αN, and the consistency of m^ as (N,T) given appropriate specification of mmax.

3.4. What if the rank condition is violated?

When RC^=0, the standard CCE estimator is generally inconsistent unless the regressors are uncorrelated with the unobserved factor loadings. One may seek to restore the RC by augmenting the model with additional CSA (see Appendix A for several options). This brings about two important issues.

The first one is how to choose relevant additional CSA from a set of candidate expansions, as not all candidates are necessarily informative about 𝐅. The second one is whether the selected additional CSA are also able to restore the RC.

To tackle the first question, Karabiyik et al. (Citation2019) have proposed an IC selection procedure. To illustrate, let Z+ be the matrix of available expansions (17) Z+={Z+(1),Z+(2),Z+(3)},(17) where (say) Z+(1)=Z(e) contains CSA of new exogenous variables, Z+(2)=Zw1 contains a matrix of CSA arising from a new weighting variable w1, and similarly Z+(3)=Zw2 for a weight w2 (see Appendix A for details). The appropriate set of expansion CSA can be selected from Z+ by minimizing (18) =argminIC(),(18) where CNT=min{N,T} in (19) IC()=ln|Σi=1NZiMA()Zi/NT|+g();g()=cols(ZA())(K+1)ln(CNT)CNT,(19) where cols(⋅) denotes the number of columns of the matrix within brackets and ={1,2,...} gathers the indices of the considered expansions from Z+. As such, for (say) ={1,3}, we have ZA()=[Z,Z+(1),Z+(3)]=[Z,Z(e),Zw2], and MA()=ITZA()(ZA()ZA())ZA().

A desirable property of the IC selection procedure is that it identifies the CSA that bring in new information about the factors in 𝐙i given what is already present in Z. Candidates that are uninformative, or informative on factors that do not feature in 𝐙i, will be excluded (asymptotically). However, the IC does not by itself signal whether the additional CSA are sufficient to restore the RC. For example, if the IC does not select additional CSA besides Z, this could be either because the rank condition is satisfied with Z, or because no further informative CSA are available in the proposal set Z+. To overcome this problem, we propose combining the IC with our RC classifier, as outlined in Algorithm 1.

Remark 3.4.

An alternative strategy would be to combine our classifier with the regularization approach proposed by Juodis (Citation2022). The latter makes use of the Singular Value Decomposition in order to remove the asymptotically redundant singular values of appropriately normalized CSA. We leave this possibility for future research.

Algorithm 1.

CCEA algorithm

  1. Estimate the model parameters using the standard CCE approach and calculate IC0 = IC(∅) (no expansions). Proceed to step 2;

  2. Evaluate the rank condition for Z. If RC^=1, no further steps are required. If RC^=0, proceed to step 3;

  3. Employ the IC in Eq. Equation19 to select from Z+={Z+(1),Z+(2),Z+(3),} the set of CSA that are relevant for the factors in 𝐙i. That is, define =arg minIC();

  4. If IC()IC0, evaluate the rank condition for ZA=[Z,Z+()] and proceed to step 5, else proceed to step 6;

  5. If RC^(ZA)=1, estimate the model with the CCEA estimator based on ZA. No further steps are required. If RC^(ZA)=0, proceed to step 6;

  6. Z+ does not contain sufficient informative expansions to restore the rank condition in the model. Add new potential expansions to Z+ and return to step 3;

The key benefit of Algorithm 1 is that it either returns a consistent CCE estimator or signals to the researcher that alternative CSA need to be sought. If Z or the potential expansions Z+ contain sufficient informative CSA to satisfy or restore the RC, they will (asymptotically) be identified and the CCE estimates emerging from Algorithm 1 are consistent. In this case, the algorithm will end in step (5) (or step (2)). If on the other hand Z does not satisfy the RC and Z+ does not contain the right (or insufficient) expansions to restore it, then this will be signaled by landing in step (6). The researcher then knows that alternative CSA candidates need to be sought and fed into the algorithm before consistent CCE estimates can be obtained.

Remark 3.5.

It is also possible to evaluate the RC for all potential augmentations until RC^=1. However, this strategy bares the risk of selecting CSA that load on different factors than those in 𝐙i, and so they are irrelevant for approximating the factor space. This is because such CSA will increase the rank of the augmented loading matrix, despite being irrelevant, and they will therefore be incorrectly favored by the classifier. A preliminary pass-through by the IC selection, as in Algorithm 1, eliminates such irrelevant options.

Remark 3.6.

We refer to Appendix D for a discussion and results pertaining to inference with the augmented CCE estimator after the application of Algorithm 1.

3.5. Heterogeneous slopes

The preceding results and theorems apply similarly to heterogeneous random slope models as in Assumption 4 of Pesaran (Citation2006), where βi=β+νi, νiIID(0K×1,Ων), with 𝛀ν nonnegative definite and 𝝂g is independent of ft,εis,vjl,υh for all t,i,s,j,l,g,h. Note that in this setting, the first column of Ci=[δi,Γi] changes to δi=λi+Γiβi and similarly for the first column of Ui=[εi+Viβi,Vi]. However, since 𝝂i is an i.i.d. random variable with finite variance, expectation zero and independent of the other model components, it straightforwardly follows that Assumptions 15 listed above directly follow through. In particular, the critical eigenvalue bounds and N-consistency of ΨZ for 𝚿𝐅𝐂 remain unaffected. The rank condition can thus be evaluated with the same procedures as outlined above. This is also confirmed by the Monte Carlo simulations reported in Appendix C.

4. Monte Carlo simulation

In this section, we investigate the small sample performance of the rank condition classifier proposed in Section 3 using Monte Carlo simulations.

4.1. Design

Data are generated from Eq. (Equation3), broadly following Westerlund and Urbain (Citation2013). We set m = 2, K = 1, β = 3 and sample the time series in 𝐅, 𝜺i and 𝐕i assuming independent autoregressive processes with a common AR coefficient ρ = 0. 8 and normally distributed mean zero innovations with variance (1−ρ2) for the factors and (1ρ2)/2 for the idiosyncratic errors. For the factor loadings 𝝀i and 𝚪i, we specify the following three scenarios:

  • Experiment 1: λi=[3,2]+ηi, ηiN(02,I2), and Γi=λi+[2,0].

  • Experiment 2: λi={[0,2]+ηifor i=1,,N/2[0,2]+ηifor i=N/2+1,,N with ηiN(02,I2) and 𝚪i = 𝝀i.

  • Experiment 3: λiN(02,I2) and 𝚪i = 𝝀i.

Thus, in Experiment 1 the RC is satisfied for the simple CSA Z (ϱ=m=2). In Experiment 2, the basic CSA contain some information for estimating the factors (ϱ = 1), yet not sufficient to satisfy the RC. Since the loadings in 𝐲i and 𝐗i are (perfectly) correlated, the standard CCE estimator is not consistent. In Experiment 3 the standard CSA contain no information at all about the factors (ϱ=0<m), in which case consistent CCE estimation is also not possible with Z.

We evaluate the RC in each MC iteration, using Algorithm 1 of Section 3.4. The number of factors (m) is estimated by the GR statistic of Ahn and Horenstein (Citation2013), setting mmax = 7. The rank of the loading matrix (ϱ) is estimated as in Section 3.1 with a random dimension reduction Ψ=T1/2Φ, 𝚽 containing i.i.d. standard-normal entries, and the nominal significance level given by αN=cαN1/γ, with c = 20, γ = 1 and α=5%.

Additional CSA are constructed using the following weighting schemes: (20) Zw,1=i=1NZiwi,1;wi,1={1/N1for i=1,,N/2;0for i=N/2+1,,N,(20) (21) Zw,2=i=1NZiwi,2;wi,2={0for i=1,,N/2;1/(NN1)for i=N/2+1,,N,(21) which results in CSA calculated over the first (Zw,1) and second (Zw,2) group of N∕2 cross-sectional units. This choice of weights presumes the existence of an exogenous grouping of the cross-sectional units, as in Experiment 2. It is not an appropriate RC-restoring expansion for experiments 1 and 3, as no such grouping exists for these experiments. We also consider candidate CSA originating from the T×2 matrix of external variables Zi(e)=FCi(e)+ϵi(e), where the columns of ϵi(e) are generated as AR(1) processes with autoregressive coefficient ρ = 0. 8 and mean zero normally distributed innovations with variance (1ρ2)/2, while Ci(e)=[2.5112.5]+ηi(e);vec(ηi(e))N(04,I4).

As the Zi(e) load on the same factors as those in 𝐙i, the matrix Z(e)=1Ni=1NZi(e) is an informative, RC-restoring, expansion in experiments 2 and 3. We also accommodate in our simulations the fact that in practice not all external variables will load on the same factors as those in 𝐙i. These irrelevant candidates are generated from Zi(g)=GCi(g)+ϵi(g), where the factors 𝐆, loadings Ci(g) and innovations ϵi(g) follow the same DGP as 𝐅, Ci(e) and ϵi(e) but are independently generated from the latter. As such, Zi(g) is informative about 𝐆 but not 𝐅, and Z(g) is therefore not an appropriate expansion in any of the considered experiments. The total set of candidate expansions that is fed into Algorithm 1 is thus a mixture of both relevant and uninformative candidates, and is given by (22) Z+=[Zw,1,Zw,2,Z(e),Z(g)].(22)

In accordance with Algorithm 1, the augmented estimator CCEA selects expansions from Z+ using the Information Criterion by Karabiyik et al. (Citation2019) given in Eq. (Equation19). The RC is re-evaluated when expansions are selected.

We generate 10000 datasets for each combination of N=(20,50,100,200,500,1000) and T=(20,50,100,200), and calculate the under/over-estimation frequencies for ϱ^ and m^, and the classification accuracy of RC^, i.e., the % of MC draws where the RC is correctly evaluated. When the RC is not satisfied for the standard CCE estimator (experiments 2 and 3), we also consider the CCEA estimator and compute the ‘RC satisfied rate’ as the % of MC draws where Algorithm 1 selects expansions that restore the rank condition.

4.2. Estimating ϱ and m

Results for the performance of the estimators for ϱ and m are presented in in AB format, with A and B the percentage of MC iterations where ϱ or m are, respectively, under- and over-estimated. The left panel contains results for estimating the rank ϱ of the loading matrix and reveals that both the over-and under-estimation frequencies tend to zero as N→∞. This is consistent with the main result of the paper that ϱ can be estimated consistently from Z. As expected, we find that the rank estimator is somewhat sensitive to the size of the cross-section dimension, which needs to be sufficiently large (i.e., N of at least 50) to achieve an accuracy of 75%. In contrast, its performance is largely invariant to the size of T, which supports the projection strategy to guarantee computability of the estimator and large N consistency when also T→∞. Note that the rank estimator is conservative in the sense that the true rank is more likely to be under-estimated than over-estimated. This is a consequence of our chosen significance level αN=20αN1, of which its fast decay in N implies that strong evidence against the null ϱ=ϱ is required before it is rejected in favor of a higher rank ϱ>ϱ. Yet, the observed under-estimation frequency is reasonable and vanishes sufficiently fast with N.

Table 1. Under/over-estimation frequency of the estimators for ϱ and m.

The right panel of reports results for estimating the number of factors m = 2 with the GR estimator of Ahn and Horenstein (Citation2013). The estimator performs very well despite the high serial dependence in the generated data, in which case many of its competitors in the literature tend to behave more poorly. The finite sample performance of the GR approach appears to be primarily driven by the time series dimension T. Yet, its small-sample performance is more than adequate as the approach displays low error frequencies even when T = 20, and identifies m without error when T > 50.

4.3. Evaluating the rank condition

4.3.1. Experiment 1: rank condition satisfied

In Experiment 1, the RC is satisfied for the CCE estimator that uses the standard set of CSA in Z. The classification accuracy reported in shows that the RC^ classifier is reasonably accurate in detecting that the rank condition is indeed satisfied. Even for smaller samples, the RC is correctly confirmed for at least 70% of the MC iterations, the only exception being the smallest N = 20 setting where the lowest rate is 64%. As the sample size grows, the accuracy improves and we find that it tends to 1 as both (N,T), as required. The results also show that the main determinant for finite sample performance is the cross-section dimension N, rather than T. This is as expected from the results in , which show that (i) ϱ^ is more prone to finite sample error than m^. The latter is practically error-less when T≥50; and (ii) ϱ^ converges at a slower rate and only as N grows. Hence, ϱ^ is the main driver of the finite sample performance of RC^. Therefore, in line with the properties of the CCE estimator itself, it will mainly be N that needs to be sufficiently large to be able to correctly assess the rank condition in practice. Finally, note that the samples where we incorrectly obtained RC^=0 for the CCE estimator prompted the application of the augmentation strategy outlined in Algorithm 1 of Section 3.4. As shown in in the appendix, an expansion was only selected in the smallest samples and in at most 2% of the MC iterations. Hence, the rank evaluation results for the augmented CCEA estimator reported on the right panel of are almost identical to those for the CCE estimator. This also confirms the effectiveness of the IC selector.

Table 2. Evaluating the rank condition: Experiment 1.

4.3.2. Experiment 2: rank condition violated for basic weights

In Experiment 2, the RC is violated when using the standard set of CSA Z, and since factor loadings are (perfectly) correlated, the CCE estimator is inconsistent for β in this setting.Footnote12 The left panel of shows that the RC-classifier strongly signals that the RC is violated for the CCE estimator. The proportion of samples where the classifier wrongly concludes that the RC holds quickly diminishes as (N,T).

Table 3. Evaluating the rank condition: Experiment 2.

When the RC is found to be violated, Algorithm 1 is applied by letting the IC search among the proposal expansions for additional CSA. In this experiment, this leads to the selection of at least one of the valid augmentations (Zw,1,Zw,2,Z(e)) in the majority of combinations of N and T (see in Appendix B). Accordingly, the rank condition was successfully restored in around 90% of the MC iterations, even in the smallest samples T=N=20. The proportion of samples where the RC is restored is given in the lower panel of for the CCEA estimator, and can be seen to converge to 1 as (N,T). Hence, Algorithm 1 leads to a consistent CCEA estimator as (N,T), when provided with appropriate rank-increasing CSA.Footnote13 Note that the algorithm performs well in finite samples. The cases where the RC is not satisfied for CCEA are mostly due to the miss-classification as RC^=1 in the ‘CCE’ panel, which vanishes as the sample size grows.

In practice, selecting expansion CSA with the IC does not guarantee that the rank condition is also satisfied, leaving the researcher unsure about the state of the RC. Hence, Algorithm 1 incorporates a re-evaluation with the classifier after expansions have been chosen. The top right panel of reveals that this re-evaluation is able to confirm with good accuracy that the rank condition is satisfied in those cases where the right expansions have been selected. The overall classification accuracy is over 70% in the smallest samples and gradually converges to 1 as (N,T). Finite T fluctuations in the table are induced by the estimator of m^ (which we have demonstrated to be sensitive to T), and the variability of the averages selected by the IC estimator for finite T settings. See to that end the selection probabilities reported in of the Appendix.

4.3.3. Experiment 3: rank condition violated

In Experiment 3, the loading matrix 𝐂 for the standard CSA has rank zero. Intuitively, the effect of the factors is averaged out in Z such that the CSA are uninformative for estimating the factor space. The top panel of reveals that our RC evaluation method is highly accurate in this setting even for very small N. This is due to the large discrepancy between m = 2 and ϱ = 0. Since we have specified a conservative estimator for ϱ, such a large over-estimation error almost never occurred (recall the bottom panel of ).

Table 4. Evaluating the rank condition: Experiment 3.

Given the strong signal by the classifier that the RC is violated, Algorithm 1 in the ‘CCEA’ panel has led to a search for expansion CSA in nearly all MC samples. We find that the sole rank-restoring expansion Z(e) was selected with high probability, as indicated by the high proportion of samples for which the RC has been restored (see the bottom panel of ). This again confirms the performance of the IC approach to select only relevant candidates (i.e., loading on the same factors) for expansion. Note, however, that compared to Experiment 2, the classifier appears less capable to confirm that the rank condition is restored when N is very small. Accuracy is only 40% when N = 20. Closer analysis reveals that this is caused by a relatively large under-estimation rate (60%) of the true rank in N = 20 samples when the correct expansion was selected. A possible cause is that the expanded matrix ZA=[Z,Z(e)] has a potential rank (number of columns = 4) which is twice the true rank (2). This suggests a relatively high level of estimation noise, and a cross-section dimension of N = 20 appears too small to estimate the rank accurately in such cases. Yet, the performance of the estimator improves quickly with N, and classification accuracy recovers to 85% or higher for N = 100.

We consider also the empirically relevant scenario where the proposal set Z+ does not contain sufficient informative CSA to restore the rank condition. To that end, we report in the CCEA,sub panel of the outcomes of Algorithm 1 when the set of proposal expansions is Z+,sub={Zw,1,Zw,2,Z(g)} instead of Z+. Hence, Z+,sub contains insufficient valid expansions to restore the RC, and Algorithm 1 should signal that the RC remains violated even when expansions have been selected from it. It is furthermore important that the IC does not select Z(g), the CSA that load on factors other than those in 𝐙i, as it would lead to false conclusions that the RC is satisfied by the classifier (see Remark 3.5). This makes the setting particularly challenging. However, the results summarized in the CCEA, sub panel of show that the classifier confirms with high accuracy that the RC fails even after expansions were chosen.

4.3.4. Experiments with heterogeneous slopes

To confirm the validity of the RC classifier in settings with heterogeneous slopes, we also run Experiments 1-3 with the slopes in Eq. (Equation1) generated as βi=β+νi, where νiN(0,1). We generally find that slope heterogeneity has little effect on the performance of the classifier, or Algorithm 1, save that the rank estimator incurs a slightly higher error rate due to the additional noise (max +5%). Conclusions are thus practically identical to the common slope results reported above, and we refer to Appendix C.2 for the results in the interest of space. This confirms the consistency of the procedures posited in Section 3.5.

5. Application: the impact of the Dodd-Frank Act on the profitability of U.S. banks

Studies on the profitability of banking institutions are vital for obtaining better understanding of the causes of financial crises, economic recessions, and growth. Profits constitute the first line of defense against losses from credit impairment, since retained earnings are an important source of capital. When it comes to large banks, high profitability may also signal excessive market power through stronger brand image or implicit regulatory protection; this is the so-called “too-big-to-fail” (TBTF) hypothesis, which postulates that large financial institutions may be so widely interconnected to the rest of the economy that their failure would generate a disastrous domino effect for the whole economy. To the extent that governments effectively subsidize downsize risk for financial institutions with TBTF status, large banks face artificially lower costs of capital, and thus reap more profits.

A large number of studies analyze drivers of bank profits (see e.g., Baker and Wurgler, Citation2015; Goddard et al., Citation2011; Iannotta et al., Citation2007; Lee and Hsieh, Citation2013; Staikouras and Wood, Citation2004). There is also a fairly substantial literature focusing on the TBTF hypothesis (see e.g., Gropp and Vesala, Citation2004; Hakenes and Schnabel, Citation2011; Morgan and Stiroh, Citation2005; Sironi, Citation2003; Stern and Feldman, Citation2009; Völz and Wedow, Citation2011). The bulk of this literature provides evidence that government bailout guarantees may distort market discipline, inducing excessive risk-taking and morally hazardous behavior (Mattana et al., Citation2015).

The present illustration contributes to this literature by examining the impact of the well-known “Dodd-Frank Act” (DFA) on profitability in the U.S. banking sector. The DFA is a U.S. federal law enacted in 2010 that has instituted a new failure-resolution regime, which seeks to ensure that losses resulting from bad decisions by managers are absorbed by equity and debt holders, thus potentially reducing moral hazard. Existing empirical evidence on the extent to which the DFA has alleviated the TBTF is relatively sparse and not in agreement. For example, while Baily et al. (Citation2020) conclude on a positive influence of the DFA towards resolving moral hazard, other studies point in the opposite direction (see e.g., Bordo and Duca, Citation2018). In what follows, we apply the CCE estimator and rank test methodology developed in the present paper to shed further light on this important topic.

5.1. Data and model specification

We make use of a panel data set consisting of 450 U.S. banking institutions over the period 2006:Q1–2019:Q4.Footnote14 We analyze the impact of major drivers of bank profitability, with emphasis on bank size. Thus, we specify the following model: ROAit=β1()SIZEit+β2()CARit+β3()LIQUIDITYit+β4()QUALITYit+β5()RISKit+uit;uit=ηi+λift+εit, where i=1,,N, t=1,,T, =τ1\mathbbm1{t<2011:Q1}+τ2\mathbbm1{t2011:Q1}. Essentially, the model is estimated for two sub-periods, namely 2006:Q1–2010:Q4 and 2011:Q1–2019:Q4. The first sub-period belongs to the Basel I-II period, whereas the second corresponds to the DFA and coincides with the introduction of the Basel III internationally.Footnote15

The variables of the model are defined as follows: ROAit is the return on assets (annualized net income expressed as a percentage of average total assets on a consolidated basis); SIZEit denotes the natural logarithm of bank total assets; CARit stands for “capital adequacy ratio” (ratio of Tier 1 capital over average total assets minus ineligible intangibles). Higher values of this ratio imply higher levels of capitalization; LIQUIDITYit is proxied by the loan-to-deposit ratio. Higher values imply a lower level of liquidity; QUALITYit is computed as the total amount of loan loss provisions expressed as a percentage of assets; and RISKit denotes the ratio of non performing loans to total loans. Higher values of RISK indicate that banks ex-ante took higher lending risk and therefore they have accumulated ex-post more bad loans.

The error term uit is composite. In particular, ηi captures bank-specific effects, such as ownership and location. The m×1 vector 𝐟t denotes unobserved economy-wide factors that influence bank profits, albeit with heterogeneous intensities 𝝀i. Last, εit is an idiosyncratic error.

The above set of explanatory variables originate from bank accounts (balance sheets and/or profit and loss accounts) and are tied to management decisions. As such, they are viewed as “internal”. Bank profitability is also driven by “external” factors that lie beyond the control of management, such as business cycle effects, monetary shocks, and financial innovation. These are absorbed in our model by the common factor component specified in the error term, λift. Although in some cases external drivers can be measured and included directly in the model, often the details of measurement may be difficult and/or contentious.Footnote16

We note that internal and external drivers of bank profitability are likely to be mutually correlated. For example, asset quality may depend on the position of the business cycle, since contractionary phases are typically associated with a higher level of default risk. Therefore, standard panel data approaches that fail to control for external drivers are likely to face an endogeneity problem and, hence, to yield inconsistent parameter estimates. The CCE approach allows for consistent estimation, provided that the rank condition is satisfied such that the external drivers are adequately controlled for.

For notational convenience, let 𝐙i denote the T×6 matrix with the observables (23) Zi=[yi,xi(1),,xi(5)],(23) where yi=[yi1,,yiT] is a T×1 vector such that yitROAit, and similarly for the remaining variables, where xi(k) denotes the covariate with coefficient βk().

5.2. Evaluating the RC

Before looking at the CCE estimation results, it is important to examine whether the RC is satisfied. The number of factors m is estimated from the T×(K+1)N matrix Z=[Z1,,ZN], using the Growth Ratio statistic of Ahn and Horenstein (Citation2013). The rank of the matrices of CSA that we consider, to be defined shortly, is determined based on the sequential testing procedure of Robin and Smith (Citation2000). Since T is small in both sub-samples, there is no need to reduce the row-dimensionality using a projection matrix.

We start with the standard CCE estimator based on the unweighted CSA Z=1Ni=1NZi. reports results for evaluating the RC. The first and second columns correspond to the standard CCE estimator applied to the periods 2006:Q1–2010:Q4 (Basel I-II) and 2011:Q1–2019:Q4 (Dodd-Frank Act). For the period under Basel I-II, m^=3. The standard CSA Z appear unable to proxy these factors as the RC is found to be violated, RC^=0. For the period under the Dodd-Frank Act, we obtain m^=2 and the rank condition now appears to hold for the standard CCE estimator.

Table 5. US bank profitability: Evaluating the rank condition.

Given that the RC is violated for the standard CCE approach in the first sub-period of the sample, we consider a set of potential expansion of CSA, given by (24) Z+={Z+(1),Z+(2),Z+(3),Z+(4)}.(24)

Z+(1)[x(6),x(7)] is a T×2 matrix, where x(6) and x(7) denote the simple CSA of two external variables, namely the return to equity (ROE), and the tier 1 risk-based capital ratio. ROE is defined as annualized net income expressed as a percent of average total equity on a consolidated basis. The risk-based capital ratio is defined as the tier 1 (core) capital expressed as a percent of risk-weighted assets. As these variables present alternative measures of profitability (𝐲i) and capitalization (xi(2)), respectively, they are expected to be driven by the same common factors as those entering into the regression model.

Z+(2) and Z+(3) denote T×(K+1) matrices of weighted CSA, computed from 𝐙i in Eq. (Equation23). Z+(2) is calculated using as aggregation weight the initial value of the bank-specific debt ratio (defined as total liabilities over total assets). This variable has been employed in the literature as a measure of interconnectedness of banks (Fernandez, Citation2011). Thus, banks with similar levels of debt ratio may be hit by common shocks in an alike manner and therefore they take a similar weight in the computation of the CSA of 𝐙i. Z+(3) uses the size of each bank in the beginning of the sample as averaging weight. This implies that banks of similar size get a similar weight in the computation of Z+(3).

Finally, Z+(4) is a T×2(K+1) matrix of CSA, obtained using two weights that are constructed by grouping banks according to their size. In particular, we take CSA over the bottom and the top quintile of banks.

reports IC results for each of the suggested additional CSA. Under Basel I-II, where the RC was found to be violated, the IC selects Z+(1) as a relevant expansion. The other expansions (Z+(2),Z+(3),Z+(4)) do not provide new information about the factor space. In the DFA period none of the expansions are selected, since the RC was already satisfied.

Table 6. US bank profitability: IC for additional CSA.

Given the IC results, we consider the augmented CCEA estimator with CSA ZA=[Z,Z+(1)]. Whether this augmented set of CSA is also sufficient to restore the rank condition needs to be verified with the RC classifier. Results are reported in the right panel of . As we can see, the augmentation has restored the rank condition (RC^=1) for the first sub-period. As expected, the RC remains satisfied in the second sub-period should we also augment the CCE estimator with Z+(1).

5.3. CCE and CCEA estimation results

reports CCE and CCEA estimates for the two sub-periods 2006:Q1–2010:Q4 and 2011:Q1–2019:Q4. The RC evaluation results imply that in the first sub-period the CCEA estimator is consistent, whereas CCE is not. Such discrepancy is mainly noticeable in the estimated coefficients of SIZE and LIQUIDITY. In both cases, the inconsistent CCE appears to overestimate the impact of these variables on bank profitability. For the period 2011:Q1–2019:Q4, RC holds for both CCE and CCEA. Hence, there is no need to augment the model with additional CSA. Notably, the estimated coefficients obtained by the two estimators are not statistically different.

Table 7. US bank profitability: CCE and CCEA estimation results

Turning to a comparison of the results across the two sub-periods, SIZE appears to be substantially less important in terms of driving profitability of banks under the DFA period.Footnote17 More specifically, the difference between β^1(τ1)=0.647 and β^1(τ2)=0.267 equals 0.38 and is statistically significant at the 10% level of significance, with a p-value that is roughly equal to 0.061 (one-tailed test).Footnote18 That is, if large banks exercised market power and implicitly relied on regulatory protection based on a “too-big-to-fail” presumption, such type of behavior seems to be less prevalent after the introduction of the Dodd-Frank Act. Further, note that if we use the standard CCE estimator in both sub-periods, the difference between β^1(τ1) and β^1(τ2) amounts to 0.9590.267=0.692. Hence, the impact of the DFA is estimated to be twice as large as that obtained based on our approach. This further highlights the importance of evaluating the rank condition for CCE-type estimators.

6. Conclusion

It is well known that the so-called Rank Condition is crucial for the statistical properties of the CCE approach developed by Pesaran (Citation2006). However, to date this rank condition could not be verified as it relates to the rank of the unobserved matrix of factor loadings. Therefore, in practice the rank condition is typically assumed to hold true.

In this article, we have outlined a procedure to evaluate whether the rank condition holds in the model of interest given a chosen set of cross-sectional averages. If the rank condition is found to be violated, the procedure can be applied in an augmentation strategy, which combines our proposed classifier with an Information Criterion, to determine the set of CSA that restores the rank condition. Therefore, our approach is generally applicable for checking whether the chosen cross-section averages are sufficient to satisfy the rank condition, or whether additional variables should be explored.

Acknowledgments

The authors thank Alexander Chudik, Arturas Juodis, George Kapetanios and Joakim Westerlund for helpful comments and discussions. This paper has also benefited from presentations at the 2017, 2018 International Panel Data Conference, the 2018 Asian and European Meetings of the Econometric Society and the 2019 Panel Data Workshop in Amsterdam.

Additional information

Funding

Ignace De Vos acknowledges financial support from the Ghent University BOF research fund. Ignace De Vos and Gerdie Everaert further acknowledge financial support from the National Bank of Belgium.

Notes

1. A recent search on Google Scholar indicated that the number of empirical applications based on CCE estimation currently exceeds one thousand.

2. See e.g., Kapetanios et al. (Citation2011), Su and Jin (Citation2012), Chudik and Pesaran (Citation2015), De Vos and Everaert (Citation2021), Juodis and Sarafidis (Citation2022b) and Cui et al. (Citation2022), to mention a few.

3. The strategy of choosing all possible additional CSA at hand, in the hope of satisfying the RC, can lead to a different problem. In particular, as shown by Karabiyik et al. (Citation2017) and Juodis et al. (Citation2021), when too many CSA are employed, the CCE estimator can suffer from bias or it may have a reduced convergence rate.

4. When the model contains fixed effects, then one may set M=ITH(HH)H, where H=[ιT,Z] and 𝜾T is a T×1 vector of ones.

5. Strictly speaking, the dimension of the vector space spanned by Z will be lower than m in these cases.

6. Note that the first equality in (Equation11) follows from the fact that 𝐅 has full column rank by Assumption 2. See exercise 4.25 on pg. 85 in Abadir and Magnus (Citation2005).

7. Hereafter, we change the notation for the number of columns of the CSA matrix Z from K+1 to a general n. This is in order to accommodate for the need to augment Z with additional CSA when the rank condition fails, as described in Section 3.4.

8. 𝛀 can also be estimated using bootstrap techniques. When the model contains fixed constants, 𝐙i should be time-demeaned, i.e., pre-multiplied with Q=ITιTιT/T.

9. Juodis and Sarafidis (Citation2022b) provide additional evidence that confirms the good performance of the GR statistic in finite samples.

10. Note that our Ass. 2-3 and the independence in Ass.4 imply Assumptions A and B in Ahn and Horenstein (Citation2013), and our eigenvalue condition in Ass.1 is sufficient for Assumptions C and D in that paper (see in particular Eqs. (2)–(3) and the surrounding discussion in Ahn and Horenstein, Citation2013).

11. ϱ^>m^ can only occur in finite samples due to estimation error but not at the population level.

12. This can also be seen from the estimation results in in Appendix B.

13. This is also confirmed by the estimation results for β in of Appendix B.

14. All data are publicly available and they have been downloaded from the Federal Deposit Insurance Corporation (FDIC) website. See https://www.fdic.gov/.

15. Basel III is an international regulatory framework for capital standards, which incorporates a set of reforms within the banking sector, designed to improve the regulation, supervision and risk management. It requires banks to maintain proper leverage ratios and meet certain minimum capital requirements.

16. For example, how does one measure monetary shocks? Does one look at interest rates or monetary aggregates? Which monetary aggregates? Similarly, how does one proxy financial innovation? For instance, how does one measure embedded leverage in new financial instruments?

17. In practice, we note that β1() captures not only the effect of market power and implicit regulatory protection via TBTF, but also the effect of economies of scale. Such scale effects will be positive (negative) if there exist economies (diseconomies) of scale. However, to the extent that the degree of returns to scale in the banking sector has remained unaltered during the sampling period of the analysis, the difference in the coefficient of SIZEit between the two sub-periods, i.e., β1(τ2)β1(τ1), will measure the impact of the Dodd-Frank Act on TBTF, conditional on the remaining covariates.

18. The t-statistic is t=(0.6470.267)/0.1962+0.1492=1.54. Note that since the CCEA and CCE estimates are based on different samples, it is natural to assume that their covariance equals zero.

19. For example, Karabiyik et al. (Citation2019) estimate a gravity equation of bilateral trade flows and construct weights based on different measures of trade cost.

20. This property is also utilized by Juodis and Sarafidis (Citation2022a), who propose the use of aggregation weights in the context of GMM estimation in panels with T fixed or large.

21. See Section 2 in Karabiyik et al. (Citation2019) for the formal set of assumptions required to ensure the validity of such weights.

References

  • Abadir, K. M., Magnus, J. R. (2005). Matrix Algebra. Cambridge: Cambridge University Press.
  • Ahn, S. C.,Horenstein, A. R. (2013). Eigenvalue ratio test for the number of factors. Econometrica 81(3):1203–1227.
  • Al-Sadoon, M. M. (2017). A unifying theory of tests of rank. Journal of Econometrics 199(1):49–62.
  • Alessi, L., Barigozzi, M., Capasso, M. (2010). Improved penalization for determining the number of factors in approximate factor models. Statistics & Probability Letters 80(23):1806–1813.
  • Bai, J.,, Ng, S. (2002). Determining the number of factors in approximate factor models. Econometrica 70(1):191–221.
  • Baily, M. N., Klein, A., Schardin, J. (2020). The impact of the Dodd- Frank act on financial stability and economic growth. The Russell Sage Foundation Journal of the Social Sciences 3:20–47.
  • Baker, M., Wurgler, J. (2015). Do strict capital requirements raise the cost of capital? bank regulation, capital structure, and the low-risk anomaly. American Economic Review 105(5):315–320. doi:10.1257/aer.p20151092
  • Bordo, M. D., Duca, J. V. (2018). The impact of the Dodd-Frank act on small business. Staff Reports 1806:1–40.
  • Camba-Mendez, G., Kapetanios, G. (2009). Statistical tests and estimators of the rank of a matrix and their applications in econometric modelling. Econometric Reviews 28(6):581–611. doi:10.1080/07474930802473785
  • Chen, Q., Fang, Z. (2019). Improved inference on the rank of a matrix. Quantitative Economics 10(4):1787–1824. doi:10.3982/QE1139
  • Chudik, A., Pesaran, M. H. (2015). Common correlated effects estimation of heterogeneous dynamic panel data models with weakly exogenous regressors. Journal of Econometrics 188(2):393–420. doi:10.1016/j.jeconom.2015.03.007
  • Cragg, J. G., Donald, S. G. (1997). Inferring the rank of a matrix. Journal of Econometrics 76(1-2):223–250. doi:10.1016/0304-4076(95)01790-9
  • Cui, G., Norkutė, M., Sarafidis, V., Yamagata, T. (2022). Two-stage instrumental variable estimation of linear panel data models with interactive effects. The Econometrics Journal 25(2):340–361. doi:10.1093/ectj/utab029
  • De Vos, I., Everaert, G. (2021). Bias-corrected common correlated effects pooled estimation in dynamic panels. Journal of Business & Economic Statistics 39(1):294–306. doi:10.1080/07350015.2019.1654879
  • Fan, J., Liao, Y. (2020). Learning latent factors from diversified projections and its applications to over-estimated and weak factors. Journal of the American Statistical Association 117(538):909–24.
  • Fernandez, V. (2011). Spatial linkages in international financial markets. Quantitative Finance 11(2):237–245. doi:10.1080/14697680903127403
  • Goddard, J., Liu, H., Molyneux, P., Wilson, J. O. (2011). The persistence of bank profit. Journal of Banking & Finance 35(11):2881–2890. doi:10.1016/j.jbankfin.2011.03.015
  • Gropp, R., Vesala, J. (2004). Deposit insurance, moral hazard and market monitoring. Review of Finance 8(4):571–602. doi:10.1093/rof/8.4.571
  • Hakenes, H., Schnabel, I. (2011). Bank size and risk-taking under basel II. Journal of Banking & Finance 35(6):1436–1449. doi:10.1016/j.jbankfin.2010.10.031
  • Iannotta, G., Nocera, G., Sironi, A. (2007). Ownership structure, risk and performance in the european banking industry. Journal of Banking & Finance 31(7):2127–2149. doi:10.1016/j.jbankfin.2006.07.013
  • Juodis, A. (2022). A regularization approach to common correlated effects estimation. Journal of Applied Econometrics 37(4):788–810. doi:10.1002/jae.2899
  • Juodis, A., Karabiyik, H., Westerlund, J. (2021). On the robustness of the pooled CCE estimator. Journal of Econometrics 220(2):325–348. doi:10.1016/j.jeconom.2020.06.002
  • Juodis, A., Sarafidis, V. (2022). An incidental parameters free inference approach for panels with common shocks. Journal of Econometrics 229(1):19–54. doi:10.1016/j.jeconom.2021.03.011
  • Juodis, A., Sarafidis, V. (2022). A linear estimator for factor-augmented fixed-T panels with endogenous regressors. Journal of Business & Economic Statistics 40(1):1–15. doi:10.1080/07350015.2020.1766469
  • Kapetanios, G. (2010). A testing procedure for determining the number of factors in approximate factor models with large datasets. Journal of Business & Economic Statistics 28(3):397–409. doi:10.1198/jbes.2009.07239
  • Kapetanios, G., Pesaran, M. H., Yamagata, T. (2011). Panels with non-stationary multifactor error structures. Journal of Econometrics 160(2):326–348. doi:10.1016/j.jeconom.2010.10.001
  • Karabiyik, H., Reese, S., Westerlund, J. (2017). On the role of the rank condition in CCE estimation of factor-augmented panel regressions. Journal of Econometrics 197(1):60–64. doi:10.1016/j.jeconom.2016.10.006
  • Karabiyik, H., Urbain, J., Westerlund, J. (2019). CCE estimation of factor-augmented regression models with more factors than observables. Journal of Applied Econometrics 34(2):268–284. doi:10.1002/jae.2661
  • Kleibergen, F., Paap, R. (2006). Generalized reduced rank tests using the singular value decomposition. Journal of Econometrics 133(1):97–126. doi:10.1016/j.jeconom.2005.02.011
  • Lee, C.-C., Hsieh, M.-F. (2013). The impact of bank capital on profitability and risk in asian banking. Journal of International Money and Finance 32:251–281. doi:10.1016/j.jimonfin.2012.04.013
  • Mattana, P., Petroni, F., Rossi, S. P. S. (2015). A test for the too-big-TO-fail hypothesis for european banks during the financial crisis. Applied Economics 47(4):319–332. doi:10.1080/00036846.2014.959654
  • Morgan, D. P., Stiroh, K. J. (2005). Too big to fail after all these years. Staff Reports, 220:1–30.
  • Onatski, A. (2010). Determining the number of factors from empirical distribution of eigenvalues. Review of Economics and Statistics 92(4):1004–1016. doi:10.1162/REST_a_00043.
  • Pesaran, M. H. (2006). Estimation and inference in large heterogeneous panels with a multifactor error structure. Econometrica 74(4):967–1012. doi:10.1111/j.1468-0262.2006.00692.x
  • Robin, J.-M., Smith, R. J. (2000). Tests of rank. Econometric Theory 16(2):151–175. doi:10.1017/S0266466600162012
  • Sironi, A. (2003). Testing for market discipline in the european banking industry: Evidence from subordinated debt issues. Journal of Money, Credit, and Banking 35(3):443–472. doi:10.1353/mcb.2003.0022
  • Staikouras, C. K., Wood, G. E. (2004). The determinants of european bank profitability. International Business & Economics Research Journal. 3(6):57–68.
  • Stern, G. H., Feldman, R. J. (2009). Too Big to Fail: The Hazards of Bank Bailouts, Washington, DC: Brookings Institution Press. ISBN: 0-8157-8152-0 220.
  • Su, L., Jin, S. (2012). Sieve estimation of panel data models with cross section dependence. Journal of Econometrics 169(1):34–47. doi:10.1016/j.jeconom.2012.01.006
  • Trapani, L. (2018). A randomized sequential procedure to determine the number of factors. Journal of the American Statistical Association 113(523):1341–1349. doi:10.1080/01621459.2017.1328359
  • Völz, M., Wedow, M. (2011). Market discipline and too-big-TO-fail in the CDS market: Does banks’ size reduce market discipline?. Journal of Empirical Finance 18(2):195–210. doi:10.1016/j.jempfin.2011.01.001
  • Westerlund, J., Urbain, J.-P. (2013). On the estimation and inference in factor-augmented panel regressions with correlated loadings. Economics Letters 119(3):247–250. doi:10.1016/j.econlet.2013.03.022
  • Westerlund, J., Urbain, J.-P. (2015). Cross-sectional averages versus principal components. Journal of Econometrics 185(2):372–377. doi:10.1016/j.jeconom.2014.09.014

Appendix A

Rank condition not satisfied: potential CSA for expansions

Chudik and Pesaran (Citation2015) advocate expanding Z by adding cross-sectional averages of external variables. This practice requires that these variables load on the same set of factors 𝐅 that operate in 𝐙i, but otherwise have no relation to the dependent variable. To illustrate, consider a setting where m>K+1 so that the rank condition is violated for Z. Let Zi(e) be the T×Ke matrix gathering the exogenous covariates, given by (A.1) Zi(e)=FCi(e)+ϵi(e),(A.1) where Ci(e) denotes an m×Ke matrix of factor loadings with finite mean 𝐂(e), and ϵi(e) is the T×Ke matrix of errors. Assuming that the components of this DGP also satisfy Assumptions 1-3 and 4, the augmented matrix of CSA, ZA=[Z,Z(e)], may satisfy the rank condition, because it can be written as (A.2) ZA=F[C,C(e)]+[U,ϵ(e)]=FCA+Op(N1/2),(A.2) where CA=[C,C(e)]. Given that Zi(e) loads on the same set of factors 𝐅, the augmented loading matrix 𝐂A is now of order m×(1+K+Ke). Therefore, this can restore the RC provided that m1+K+Ke and 𝐂(e) is also sufficiently distinct from 𝐂.

An alternative idea is to make use of external variables as additional weights, in order to construct weighted CSA. Such an approach has been recently advocated by Juodis and Sarafidis (Citation2022b), Fan and Liao (Citation2020), Juodis and Sarafidis (Citation2022a) and, in the present context of CCE estimation, by Karabiyik et al. (Citation2019).

To illustrate, let wi denote an external, time-invariant variable.Footnote19 Multiplying Eq. (Equation3) by wi and summing over i yields (A.3) ZwT×(K+1)=FT×m Cwm×(K+1)+UwT×(K+1),(A.3) where Zw=i=1NZiwi, Cw=i=1NCiwi, and Uw=i=1NUiwi. As shown by Karabiyik et al. (Citation2019), when 𝐂i and wi are correlated, but 𝐔i and wi are not, then Zw=FCw+Op(N1/2) and Cw converges to a non zero matrix.Footnote20 If Cw is also sufficiently distinct from C, the obtained Zw provides new (i.e., rank increasing) information on 𝐅, and the rank of the augmented matrix ZA=[Z,Zw] is increased. As the authors point out, wi effectively acts as an instrument for 𝐂i, and multiple wi can be combined in an attempt to restore the RC.Footnote21

Lastly, one can also employ deterministic averaging weights, such as binary indicators that give rise to group-specific cross-sectional averages. For example, in a panel of countries, individual units may be classified as developed, emerging and developing economies; in a panel of firms, units may be grouped according to their size or sector; and so on. In many cases, such group memberships are known and the group-specific averages can be more informative factor proxies than the simple (overall) average.

Appendix B

Proofs of theoretical results

B.1. Proof of Theorem 1

Let 𝐌 be a given T×n matrix (T > n) and let Φ=[ϕ1,,ϕn] be an n×T random matrix, where ϕ1,,ϕn are i.i.d.MN(𝟎, 𝐈T) in T.

Part (i). We wish to prove that Pr[rank(ΦM)=rank(M)]=1.

Case 1. 𝐌 has full column rank, i.e., rank(𝐌) = n.

Consider the row-matrix representation of a product of two matrices: ΦM=[ϕ1Mϕ2MϕnM].

It suffices to show that (B.1) Pr[{ϕ1M,,ϕnM} are linearly independent]=1Pr[{Mϕ1,,Mϕn} are linearly independent]=1.(B.1)

Let 𝐳i = 𝐌ϕi denote an n×1 vector, for i=1,,n. Since ϕ1,,ϕn are i.i.d.MN(𝟎, 𝐈T), it follows that z1,,zn are i.i.d.MN(𝟎, 𝐌𝐌), with 𝐌𝐌 non singular because 𝐌 has full rank. Let z~i denote a specific realization of 𝐳i, where z~iRn. Define y=vec(z1,,zn) and y~=vec(z~1,,z~n), both n2×1 vectors. Let A={y~Rn2:z~1,,z~n are linearly dependent }.

We have (B.2) Pr[z1,,zn are linearly dependent]=E(I[yA])=E(E(I[yA]|z1,,zn1))=0(B.2) because E(I[yA]|z1,,zn1)=0.

Therefore, we have proved that Pr[z1,,zn are linearly dependent]=0, and thereby Pr[rank(ΦM)=rank(M)=n]=1.

Case 2. 𝐌 has less than full column rank, i.e., rank(M)=n1<n. Partition 𝐌 = [𝐌1 ⋮ 𝐌2], where 𝐌1 is T×n1 and 𝐌2 is T×(nn1), such that rank(𝐌1) = n1. Similarly, partition 𝚽 such that Φ=[Φ1[2pt/2pt]Φ2] where 𝚽1 and 𝚽2 are n1×T and (nn1)×T respectively with rank(𝚽1) = n1 with probability 1. Thus, the product between 𝚽 and 𝐌 can be written as (B.3) 𝚽𝐌=[𝚽1𝐌1𝚽2𝐌1𝚽1𝐌2𝚽2𝐌2]n×n(B.3)

Based on exactly the same arguments as in Case 1, it can be shown that rank(𝚽1𝐌1) = n1 with probability 1. However, since 𝚽1𝐌1 is a submatrix of 𝚽𝐌, rank(𝚽𝐌)≥n1. Therefore, we have n1=rank(Φ1M1) rank(ΦM)min{rank (Φ),rank(M)}=n1.

Hence, rank(ΦM)=n1= rank(𝐌) with probability 1. This completes part (i) of the theorem.

Part (ii). We can write by simple addition and subtraction ΨZ=T1/2ΦZ=T1/2(ΦFC+ΦU)=T1/2ΦFC+T1/2ΦF(CC)+T1/2ΦU

Recall that Φ=[ϕ1,,ϕn], with its rows given by ϕkiidMN(0T×1,IT) for k=1,,n. Consider then the kth row of T1/2ΦU, and write it as T1/2t=1Tϕktut, with ϕk=[ϕk1,,ϕkT] and U=[u1,,uT]. By the independence of 𝚽 and U we have for every k=1,,n that E(t=1Tϕktut)=01×n and Var(t=1TϕktutT)=1Tt=1Ts=1TE(ϕktϕks)E(utus)=1Tt=1TE(utut)=O(N1), because E(ϕktϕks) = 0 for st, E(ϕktϕkt)=E(ϕkt2)=1 and E(ut2)=O(N1) by A.4 of Lemma 1 in Pesaran (Citation2006) under Assumptions 1 and 4. Hence, T1/2ΦU=Op(N1/2) as (N,T).

Consider next T1/2ΦF. By the independence of 𝚽 and 𝐅 we have E(ΦF)=0n×m, and since F=[f1,,fT] we can write the k-th row of 𝚽𝐅 as t=1Tϕktft. Therefore, Var(t=1TϕktftT)=1Tt=1TE(ϕktϕkt)E(ftft)=1Tt=1TE(ftft)=O(1), because E(ftft)=O(1) for every t (Assumption 2). Hence, we have T1/2ΦF=Op(1). Noting then that CC=Op(N1/2) under Assumption 3, it follows that T1/2ΦF(CC)T1/2ΦFCC=Op(N1/2).

Thus, combining the results above yields as (N,T), ΨZ=T1/2ΦFC+T1/2ΦF(CC)+T1/2ΦU=T1/2ΦFC+Op(N1/2), where also T1/2ΦFCT1/2ΦFC=Op(1) since C< under Assumption 3. Hence, the proof of part (ii) of the theorem is complete.

Appendix C

Additional simulation results

C.1. Homogeneous slopes

Table C.1. Algorithm 1: Selection percentages for expansion CSA.

Table C.2. Algorithm 1: Sensitivity and specificity.

C.1.1. Estimation results

Table C.3. Estimation results for β in Experiment 1.

Table C.4. Estimation results for β in Experiment 2.

Table C.5. Estimation results for β in Experiment 3.

C.2. Heterogeneous slopes

Table C.6. Algorithm 1: Selection percentages for expansion CSA (heterogeneous slopes).

Table C.7. Algorithm 1: Sensitivity and Specificity (heterogeneous slopes).

C.2.1. Evaluating the rank condition

Table C.8. Under/over-estimation frequency of the estimators for ϱ and m (heterogeneous slopes).

Table C.9. Evaluating the rank condition: Experiment 1 (heterogeneous slopes).

Table C.10. Evaluating the rank condition: Experiment 2 (heterogeneous slopes).

Table C.11. Evaluating the rank condition: Experiment 3 (heterogeneous slopes).

C.2.2. Estimation results

Table C.12. Estimation results for heterogeneous β in Experiment 1.

Table C.13. Estimation results for heterogeneous β in Experiment 2.

Table C.14. Estimation results for heterogeneous β in Experiment 3.

Appendix D

Post-selection inference

One remaining question is how to perform inference with the CCEA estimator after application of Algorithm 1. To address this, let β^A be the estimated parameter vector that follows from Algorithm 1, calculated with the (potentially) augmented set of cross-section averages ZA. We shall assume that Z+ contains sufficient valid augmentations to restore the rank condition through ZA in case the RC is violated with Z. The distribution of β^A is then given by Pr(NT(β^Aβ)δ)=Pr(NT(β^Aβ)δ|ϱ(ZA)=m)+Pr(NT(β^Aβ)δ|ϱ(ZA)m)(1Pr(ϱ(ZA)=m)) where ϱ(ZA) denotes the rank of the loading matrix implied by ZA, such that Pr(ϱ(ZA)=m) is the probability that RC^=0 when the RC was not satisfied (the augmentation sequence is kicked into gear) and an RC restoring set of augmentations was selected from Z+ by the IC criterion. Since the classifier correctly evaluates the rank condition as (N,T) by Proposition 2, and hence leads to the IC selection step in stage (3) and beyond of Algorithm 1 if ϱ(Z)m, we have by the consistency of (Equation19) for selecting the correct rank-restoring averages, established in Karabiyik et al. (Citation2019), that Pr(ϱ(ZA)=m)1 as (N,T). Hence, asymptotically Pr(NT(β^Aβ)δ)=Pr(NT(β^Aβ)δ|ϱ(ZA)=m), such that the distribution of the augmented estimator asymptotically equals that of the CCE estimator when the rank condition is satisfied. As is well known from the CCE literature, this asymptotic distribution is independent of the specific choice of CSA provided that T/N0 (see e.g., Theorem 1 in Karabiyik et al. (Citation2019)). Hence, the distribution of NT(β^Aβ) is asymptotically unaffected by pre-testing and augmentations, and inference can proceed as for the original CCE approach, with rank condition satisfied, provided T/N0.