References
- Beck, A., and Eldar, Y. C. (2013), “Sparsity Constrained Nonlinear Optimization: Optimality Conditions and Algorithms,” SIAM Journal on Optimization, 23, 1480–1509. DOI: https://doi.org/10.1137/120869778.
- Bertsimas, D., and King, A. (2017), “Logistic Regression: From Art to Science,” Statistical Science, 32, 367–384. DOI: https://doi.org/10.1214/16-STS602.
- Bertsimas, D., King, A., and Mazumder, R. (2016), “Best Subset Selection via a Modern Optimization Lens,” The Annals of Statistics, 44, 813–852. DOI: https://doi.org/10.1214/15-AOS1388.
- Bertsimas, D., and Van Parys, B. (2017), “Sparse High-Dimensional Regression: Exact Scalable Algorithms and Phase Transitions,” arXiv no. 1709.10029.
- Bien, J., and Tibshirani, R. J. (2011), “Sparse Estimation of a Covariance Matrix,” Biometrika, 98, 807–820. DOI: https://doi.org/10.1093/biomet/asr054.
- Chen, D.-S., Batson, R. G., and Dang, Y. (2011), Applied Integer Programming: Modeling and Solution, New York: Wiley.
- Cormen, T. H., Leiserson, C. E., Rivest, R. L., and Stein, C. (2009), Introduction to Algorithms, Cambridge, MA: MIT Press.
- Efron, B., Hastie, T., Johnstone, I., and Tibshirani, R. (2004), “Least Angle Regression,” The Annals of Statistics, 32, 407–499. DOI: https://doi.org/10.1214/009053604000000067.
- Fan, J., Liao, Y., and Liu, H. (2016), “An Overview of the Estimation of Large Covariance and Precision Matrices,” The Econometrics Journal, 19, C1–C32. DOI: https://doi.org/10.1111/ectj.12061.
- Fan, J., and Lv, J. (2008), “Sure Independence Screening for Ultrahigh Dimensional Feature Space,” Journal of the Royal Statistical Society, Series B, 70, 849–911. DOI: https://doi.org/10.1111/j.1467-9868.2008.00674.x.
- Friedman, J., Hastie, T., and Tibshirani, R. (2010), “Regularization Paths for Generalized Linear Models via Coordinate Descent,” Journal of Statistical Software, 33, 1–22. DOI: https://doi.org/10.18637/jss.v033.i01.
- Hastie, T., Tibshirani, R., Friedman, J., and Franklin, J. (2005), “The Elements of Statistical Learning: Data Mining, Inference and Prediction,” The Mathematical Intelligencer, 27, 83–85. DOI: https://doi.org/10.1007/BF02985802.
- Hastie, T., Tibshirani, R., and Tibshirani, R. J. (2017), “Extended Comparisons of Best Subset Selection, Forward Stepwise Selection, and the Lasso,” arXiv no. 1707.08692.
- Hazimeh, H., and Mazumder, R. (2018), “Fast Best Subset Selection: Coordinate Descent and Local Combinatorial Optimization Algorithms,” arXiv no. 1803.01454.
- Kessy, A., Lewin, A., and Strimmer, K. (2018), “Optimal Whitening and Decorrelation,” The American Statistician, 72, 309–314. DOI: https://doi.org/10.1080/00031305.2016.1277159.
- Mazumder, R., Radchenko, P., and Dedieu, A. (2017), “Subset Selection With Shrinkage: Sparse Linear Modeling When the SNR Is Low,” arXiv no. 1708.03288.
- Meier, L., Van De Geer, S., and Bühlmann, P. (2008), “The Group Lasso for Logistic Regression,” Journal of the Royal Statistical Society, Series B, 70, 53–71. DOI: https://doi.org/10.1111/j.1467-9868.2007.00627.x.
- Meinshausen, N. (2007), “Relaxed Lasso,” Computational Statistics & Data Analysis, 52, 374–393.
- Miyashiro, R., and Takano, Y. (2015), “Mixed Integer Second-Order Cone Programming Formulations for Variable Selection in Linear Regression,” European Journal of Operational Research, 247, 721–731. DOI: https://doi.org/10.1016/j.ejor.2015.06.081.
- Patrascu, A., and Necoara, I. (2015), “Random Coordinate Descent Methods for l0 Regularized Convex Optimization,” IEEE Transactions on Automatic Control, 60, 1811–1824. DOI: https://doi.org/10.1109/TAC.2015.2390551.
- Penrose, K. W., Nelson, A., and Fisher, A. (1985), “Generalized Body Composition Prediction Equation for Men Using Simple Measurement Techniques,” Medicine & Science in Sports & Exercise, 17, 189.
- Schrijver, A. (1986), Theory of Linear and Integer Programming, New York: Wiley.
- Tibshirani, R. (1996), “Regression Shrinkage and Selection via the Lasso,” Journal of the Royal Statistical Society, Series B, 58, 267–288. DOI: https://doi.org/10.1111/j.2517-6161.1996.tb02080.x.
- Tibshirani, R. (2011), “Regression Shrinkage and Selection via the Lasso: A Retrospective,” Journal of the Royal Statistical Society, Series B, 73, 273–282.
- Willis, M. J., and von Stosch, M. (2017), “L0-Constrained Regression Using Mixed Integer Linear Programming,” Chemometrics and Intelligent Laboratory Systems, 165, 29–37. DOI: https://doi.org/10.1016/j.chemolab.2016.12.016.
- Zhao, P., and Yu, B. (2006), “On Model Selection Consistency of Lasso,” Journal of Machine Learning Research, 7, 2541–2563.
- Zou, H., and Hastie, T. (2005), “Regularization and Variable Selection via the Elastic Net,” Journal of the Royal Statistical Society, Series B, 67, 301–320. DOI: https://doi.org/10.1111/j.1467-9868.2005.00503.x.