References
- Barber, R. F., and Ha, W. (2018), “Gradient Descent With Non-Convex Constraints: Local Concavity Determines Convergence,” Information and Inference: A Journal of the IMA, 7, 755–806. DOI: https://doi.org/10.1093/imaiai/iay002.
- Bellman, R. (1966), Dynamic Programming Science, 153(3731), 34–37.
- Bertsimas, D., King, A., Mazumder, R. (2016), “Best Subset Selection Via a Modern Optimization Lens,” The Annals of Statistics, 44, 813–852. DOI: https://doi.org/10.1214/15-AOS1388.
- Brisbane, W., Bailey, M. R., and Sorensen, M. D. (2016), “An Overview of Kidney Stone Imaging Techniques,” Nature Reviews Urology, 13, 654–662. DOI: https://doi.org/10.1038/nrurol.2016.154.
- Charbonnel, B., Matthews, D., Schernthaner, G., Hanefeld, M., Brunetti, P., and Group, Q. S. (2005), “A Long-Term Comparison of Pioglitazone and Gliclazide in Patients with Type 2 Diabetes Mellitus: A Randomized, Double-Blind, Parallel-Group Comparison Trial,” Diabetic Medicine, 22, 399–405. DOI: https://doi.org/10.1111/j.1464-5491.2004.01426.x.
- Clark, E., Askham, T., Brunton, S. L., and Nathan Kutz, J. (2019), “Greedy Sensor Placement With Cost Constraints,” IEEE Sensors Journal, 19, 2642–2656. DOI: https://doi.org/10.1109/JSEN.2018.2887044.
- Fan, J. and Li, R. (2001), “Variable Selection via Nonconcave Penalized Likelihood and Its Oracle Properties,” Journal of the American Statistical Association, 96, 1348–1361. DOI: https://doi.org/10.1198/016214501753382273.
- Fouskakis, D. and Draper, D. (2008), “Comparing Stochastic Optimization Methods for Variable Selection in Binary Outcome Prediction, With Application to Health Policy,” Journal of the American Statistical Association, 103, 1367–1381. DOI: https://doi.org/10.1198/016214508000001048.
- Fouskakis, D., Ntzoufras, I., and Draper, D. (2009), “Bayesian Variable Selection Using Cost-adjusted BIC, With Application to Cost-Effective Measurement of Quality of Health Care,” The Annals of Applied Statistics, 663–690. DOI: https://doi.org/10.1214/08-AOAS207.
- Frank, L. E. and Friedman, J. H. (1993), “A Statistical View of Some Chemometrics Regression Tools,” Technometrics, 35, 109–135. DOI: https://doi.org/10.1080/00401706.1993.10485033.
- Friedman, J., Hastie, T., and Tibshirani, R. (2010), “Regularization Paths for Generalized Linear Models Via Coordinate Descent,” Journal of Statistical Software, 33, 1. DOI: https://doi.org/10.18637/jss.v033.i01.
- Glover, F. (1977), “Heuristics for Integer Programming Using Surrogate Constraints,” Decision Sciences, 8, 156–166. DOI: https://doi.org/10.1111/j.1540-5915.1977.tb01074.x.
- Glover, F. (1986), “Future Paths for Integer Programming and Links to Artificial Intelligence,” Computers & Operations Research, 13, 533–549.
- Glover, F. (1989), “Tabu Search-Part I,” ORSA Journal on Computing, 1, 190–206.
- Holland, J. H. (1992), Adaptation in Natural and Artificial Systems: An Introductory Analysis With Applications to Biology, Control, and Artificial Intelligence, Cambridge, MA: MIT Press.
- Kachuee, M., Karkkainen, K., Goldstein, O., Zamanzadeh, D., and Sarrafzadeh, M. (2019), “Cost-Sensitive Diagnosis and Learning Leveraging Public Health Data,” arXiv preprint arXiv:1902.07102.
- Kirkpatrick, S., Gelatt, C. D., Vecchi, M. P. (1983), “Optimization by Simulated Annealing,” Science, 220, 671–680. DOI: https://doi.org/10.1126/science.220.4598.671.
- Krishnapuram, B., Yu, S., and Rao, R. B. (2011), Cost-Sensitive Machine Learning, Boca Raton, FL: CRC Press.
- Martello, S., Pisinger, D., and Toth, P. (1999), “Dynamic Programming and Strong Bounds for the 0-1 Knapsack Problem,” Management Science, 45, 414–424. DOI: https://doi.org/10.1287/mnsc.45.3.414.
- Nauss, R. M. (1976), “An Efficient Algorithm for the 0-1 Knapsack Problem,” Management Science, 23, 27–31. DOI: https://doi.org/10.1287/mnsc.23.1.27.
- Nesterov, Y. (2013), “Gradient Methods for Minimizing Composite Functions,” Mathematical Programming, 140, 125–161. DOI: https://doi.org/10.1007/s10107-012-0629-5.
- Paschos, V. T. (2013), Paradigms of Combinatorial Optimization: Problems and New Approaches (vol. 2), Hoboken, NJ: Wiley.
- Pattuk, E., Kantarcioglu, M., Ulusoy, H., and Malin, B. (2015), “Privacy-aware Dynamic Feature Selection,” in 2015 IEEE 31st International Conference on Data Engineering, Seoul, South Korea: IEEE, pp. 78–88.
- Shapiro, A., Dentcheva, D., and Ruszczyński, A. (2014), Lectures on Stochastic Programming: Modeling and Theory, Philadelphia, PA: SIAM.
- Tibshirani, R. (1996), “Regression Shrinkage and Selection Via the Lasso,” Journal of the Royal Statistical Society, Series B, 58, 267–288. DOI: https://doi.org/10.1111/j.2517-6161.1996.tb02080.x.
- Yang, Y., and Zou, H. (2015), “A Fast Unified Algorithm for Solving Group-lasso Penalize Learning Problems,” Statistics and Computing, 25, 1129–1141. DOI: https://doi.org/10.1007/s11222-014-9498-5.
- Yuan, M., and Lin, Y. (2006), “Model Selection and Estimation in Regression With Grouped Variables,” Journal of the Royal Statistical Society, Series B, 68, 49–67. DOI: https://doi.org/10.1111/j.1467-9868.2005.00532.x.
- Yue, L. H. (2010), “Cost-efficient Variable Selection Using Branching LARS,” Electronic Thesis and Dissertation Repository.
- Zhang, C. (2010), “Nearly Unbiased Variable Selection Under Minimax Concave Penalty,” The Annals of Statistics, 38, 894–942. DOI: https://doi.org/10.1214/09-AOS729.
- Zhao, P., and Yu, B. (2006), “On Model Selection Consistency of Lasso,” The Journal of Machine Learning Research, 7, 2541–2563.
- Zou, H. (2006), “The Adaptive Lasso and Its Oracle Properties,” Journal of the American Statistical Association, 101, 1418–1429. DOI: https://doi.org/10.1198/016214506000000735.
- Zou, H. and Hastie, T. (2005), “Regularization and Variable Selection Via the Elastic Net,” Journal of the Royal Statistical Society, Series B, 67, 301–320. DOI: https://doi.org/10.1111/j.1467-9868.2005.00503.x.