279
Views
2
CrossRef citations to date
0
Altmetric
Articles

Model selection in high-dimensional noisy data: a simulation study

ORCID Icon &
Pages 2031-2050 | Received 11 Apr 2018, Accepted 10 Apr 2019, Published online: 18 Apr 2019

References

  • Tibshirani R. Regression shrinkage and selection via the lasso. J R Stat Soc Ser B (Stat Meth). 1996;58(1):267–288.
  • Candes E, Tao T. The Dantzig selector: statistical estimation when p is much larger than n. Ann Stat. 2007;35(6):2313–2351.
  • Fan J, Li R. Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc. 2001;96(456):1348–1360.
  • Rocke DM, Durbin B. A model for measurement error for gene expression arrays. J Comput Biol. 2001;8(6):557–569.
  • Purdom E, Holmes SP. Error distribution for gene expression data. Stat Appl Genet Mol Biol. 2005;4(1):1–33.
  • Bertrand A, Moonen M. Consensus-based distributed total least squares estimation in ad hoc wireless sensor networks. IEEE Trans Signal Process. 2011;59(5):2320–2330.
  • Slijepcevic S, Megerian S, Potkonjak M. Location errors in wireless embedded sensor networks: sources, models, and effects on applications. ACM SIGMOBILE Mobile Comput Commun Rev. 2002;6(3):67–78.
  • Bound J, Brown C, Mathiowetz N. Measurement error in survey data. In: Handbook of econometrics. Vol. 5. Elsevier; 2001. p. 3705–3843.
  • Kipnis V, Subar AF, Midthune D, et al. Structure of dietary measurement error: results of the open biomarker study. Am J Epidemiol. 2003;158(1):14–21.
  • Carroll RJ, Ruppert D, Stefanski LA, et al. Measurement error in nonlinear models: a modern perspective. Chapter 1.8, Loss of power. New York, NY: Chapman and Hall/CRC; 2006. Monographs on Statistics and Applied Probability 105; p. 18–22.
  • Rosenbaum M, Tsybakov AB. Sparse recovery under matrix uncertainty. Ann Stat. 2010;38(5):2620–2651.
  • Sørensen Ø, Frigessi A, Thoresen M. Measurement error in lasso: impact and likelihood bias correction. Stat Sin. 2015;25(2):809–829.
  • Loh BPL, Wainwright MJ. High-dimensional regression with noisy and missing data: provable guarantees with nonconvexity. Annals Stat. 2012;40(3):1637–1664.
  • Datta A, Zou H. Cocolasso for high-dimensional error-in-variables regression. Ann Stat. 2017;45(6):2400–2426.
  • Rosenbaum M, Tsybakov AB. Improved matrix uncertainty selector. In: From probability to statistics and back: high-dimensional models and processes – a festschrift in honor of Jon A. Wellner. Institute of Mathematical Statistics; 2013. p. 276–290.
  • Belloni A, Rosenbaum M, Tsybakov AB. Linear and conic programming estimators in high dimensional errors-in-variables models. J R Stat Soc: Ser B (Stat Meth). 2017;79(3):939–956.
  • Sørensen Ø, Hellton KH, Frigessi A, et al. Covariate selection in high-dimensional generalized linear models with measurement error. J Comput Graph Stat. 2018;27(4):739–749.
  • Kaul A, Koul HL, Chawla A, et al. Two stage non-penalized corrected least squares for high dimensional linear models with measurement error or missing covariates. arXiv preprint:160503154.2016.
  • Brown B, Weaver T, Wolfson J. Meboost: variable selection in the presence of measurement error. arXiv preprint:170102349.2017.
  • Chen Y, Caramanis C. Noisy and missing data regression: distribution-oblivious support recovery. In: International Conference on Machine Learning; 2013. p. 383–391.
  • Hastie T, Tibshirani R, Friedman J. The elements of statistical learning. Chapter 3.8.3, The dantzig selector. New York, NY: Springer; 2009. Springer Series in Statistics; p. 89–90.
  • Boyd S, Parikh N, Chu E, et al. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found Trends® Mach Learn. 2011;3(1):1–122.
  • Friedman J, Hastie T, Tibshirani R. Regularization paths for generalized linear models via coordinate descent. J Stat Softw. 2010;33(1):1–22.
  • Efron B, Hastie T, Johnstone I, et al. Least angle regression. Ann Stat. 2004;32(2):407–499.
  • Zhao P, Yu B. On model selection consistency of lasso. J Mach Learn Res. 2006;7(Nov):2541–2563.
  • Zou H. The adaptive lasso and its oracle properties. J Am Stat Assoc. 2006;101(476):1418–1429.
  • Reppe S, Refvem H, Gautvik VT, et al. Eight genes are highly associated with BMD variation in postmenopausal caucasian women. Bone. 2010;46(3):604–612.
  • Boulesteix AL, Strobl C, Augustin T, et al. Evaluating microarray-based classifiers: an overview. Cancer Inform. 2008;6:77–97.
  • Tadesse MG, Ibrahim JG, Gentleman R, et al. Bayesian error-in-variable survival model for the analysis of genechip arrays. Biometrics. 2005;61(2):488–497.
  • Hein AMK, Richardson S, Causton HC, et al. Bgx: a fully bayesian integrated approach to the analysis of affymetrix genechip data. Biostatistics. 2005;6(3):349–373.
  • Irizarry RA, Hobbs B, Collin F, et al. Exploration, normalization, and summaries of high density oligonucleotide array probe level data. Biostatistics. 2003;4(2):249–264.
  • Fan J, Lv J. Sure independence screening for ultrahigh dimensional feature space. J R Stat Soc: Ser B (Stat Meth). 2008;70(5):849–911.
  • Bickel PJ, Ritov Y, Tsybakov AB, et al. Simultaneous analysis of lasso and Dantzig selector. Ann Stat. 2009;37(4):1705–1732.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.