1,328
Views
0
CrossRef citations to date
0
Altmetric
Corrections

From Fixed-X to Random-X Regression: Bias-Variance Decompositions, Covariance Penalties, and Prediction Error Estimation: Correction

Page 529 | Received 13 Nov 2021, Accepted 29 Nov 2021, Published online: 15 Feb 2022
This article refers to:
From Fixed-X to Random-X Regression: Bias-Variance Decompositions, Covariance Penalties, and Prediction Error Estimation: Rejoinder

Saharon Rosseta and Ryan J. Tibshiranib

aDepartment of Statistics and Operations Research, Tel Aviv University, Tel Aviv, Israel; bDepartment of Statistics and Data Science, Carnegie Mellon University, Pittsburgh, PA

In our article (Rosset and Tibshirani Citation2020) Theorem 3 is incorrect as stated. Recalling the definition earlier in the articleV+=EX,x0var(f̂n(x0) | X,x0)EXvar(f̂n(x1) | X),the original Theorem states:

Assume that xQ is generated as follows: we draw zRp, having iid components ziF,i=1,,p, where F is any distribution with zero mean and unit variance, and then set x=Σ1/2z, where ΣRp×p is positive definite and Σ1/2 is its symmetric square root. Consider an asymptotic setup where p/nγ(0,1) as n. Then, for the OLS estimator,V+σ2γ21γas n.

This theorem is incorrect as stated because V + is defined in terms of an expectation over X. However, under the assumptions of the theorem, the eigenvalues of the empirical covariance matrix XTX may not be bounded away from zero, and consequently the almost sure convergence of the eigenspectrum used in the proof does not guarantee convergence (or existence) of the expectation. One can generate counterexamples along the lines of the classic examples of almost sure convergence not implying convergence in expectation.

We note that under the same conditions (with the added requirement that the fourth moment of F is finite), Hastie et al. (Citation2021) proved (in their proof of Proposition 2) thatEx0var(f̂n(x0) | X,x0)var(f̂n(x1) | X)a.s.Xσ2γ21γas n.

In words, for almost any sequence of training covariate matrices X generated by this mechanism, the random variable whose expectation is V + converges to this same fixed limit.

Thus, we conclude that the failure of Theorem 3 as written in the original article is really somewhat of an “edge case”: Although it may not apply to our formal definition of V +, which integrates over the distribution of X, the excess variance result does apply in asymptotic almost sure sense over X. In terms of practical use of the corrections proposed later in the article, we believe it still supports their wide applicability in estimating Random-X prediction error.

Acknowledgments

We thank an anonymous reviewer of a separate article for pointing out the difficulty with the original result.

ORCID

Saharon Rosset http://orcid.org/0000-0002-4458-9545

References

  • Hastie, T., Montanari, A., Rosset, S., and Tibshirani, R. (2021), “Surprises in High-Dimensional Ridgeless Least Squares Interpolation,” arXiv: 1903.08560.
  • Rosset, S., and Tibshirani, R. J. (2020), “From Fixed-x to Random-x Regression: Bias-Variance Decompositions, Covariance Penalties, and Prediction Error Estimation,” Journal of the American Statistical Association, 115, 138–151. DOI: https://doi.org/10.1080/01621459.2018.1424632.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.