Publication Cover
Optimization
A Journal of Mathematical Programming and Operations Research
Volume 73, 2024 - Issue 7
175
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Mini-batch stochastic subgradient for functional constrained optimization

, &
Pages 2159-2185 | Received 19 Apr 2022, Accepted 28 Feb 2023, Published online: 15 Mar 2023

References

  • Bhattacharyya C, Grate LR, Jordan MI, et al. Robust sparse hyperplane classifiers: application to uncertain molecular profiling data. J Comput Biol. 2004;11(6):1073–1089.
  • Vapnik V. Statistical learning theory. John Wiley; 1998.
  • Nedelcu V, Necoara I, Tran Dinh Q. Computational complexity of inexact gradient augmented Lagrangian methods: application to constrained MPC. SIAM J Control Optim. 2014;52(5):3109–3134.
  • Necoara I. General convergence analysis of stochastic first order methods for composite optimization. J Optim Theory Appl. 2021;189(1):66–95.
  • Tibshirani R. The solution path of the generalized Lasso [PhD thesis]. Stanford Univ.; 2011.
  • Rockafellar RT, Uryasev SP. Optimization of conditional value-at-risk. J Risk. 2000;2(3):21–41.
  • Jacob L, Obozinski G, Vert JP. Group Lasso with overlap and graph Lasso. In: International conference on machine learning; 2009. p. 433–440.
  • Yin X, Büyüktahtakın İ. A multi-stage stochastic programming approach to epidemic resource allocation with equity considerations. Health Care Manag Sci. 2021;24(3):597–622.
  • Zafar M, Valera I, Gomez-Rodriguez M, et al. Fairness constraints: a flexible approach for fair classification. J Mach Learn Res. 2019;20(1):2737–2778.
  • Rosasco L, Villa S, Vu BC. Convergence of stochastic proximal gradient algorithm. Appl Math Optim. 2020;82(3):891–917.
  • Hardt M, Recht B, Singer Y. Train faster, generalize better: stability of stochastic gradient descent. In: International conference on machine learning; 2016.
  • Nemirovski A, Yudin DB. Problem complexity and method efficiency in optimization. New York: Wiley Interscience; 1983.
  • Robbins H, Monro S. A stochastic approximation method. Ann Math Stat. 1951;22(3):400–407.
  • Moulines E, Bach F. Non-asymptotic analysis of stochastic approximation algorithms for machine learning. In: Advances in neural information processing systems conference; 2011.
  • Nemirovski A, Juditsky A, Lan G, et al. Robust stochastic approximation approach to stochastic programming. SIAM J Optim. 2009;19(4):1574–1609.
  • Patrascu A, Necoara I. Nonasymptotic convergence of stochastic proximal point algorithms for constrained convex optimization. J Mach Learn Res. 2018;18(198):1–42.
  • Asi H, Chadha K, Cheng G, et al. Minibatch stochastic approximate proximal point methods. In: Advances in neural information processing systems conference; 2020.
  • Duchi J, Singer Y. Efficient online and batch learning using forward backward splitting. J Mach Learn Res. 2009;10:2899–2934.
  • Nedich A, Necoara I. Random minibatch subgradient algorithms for convex problems with functional constraints. Appl Math Optim. 2019;80(3):801–833.
  • Peng X, Li L, Wang F. Accelerating minibatch stochastic gradient descent using typicality sampling. In: IEEE Trans. Neural Networks Learn. Syst.; 2019.
  • Gower R, Nicolas L, Xun Q, et al. SGD: general analysis and improved rates. In: International conference on machine learning; 2019.
  • Renegar J, Zhou S. A different perspective on the stochastic convex feasibility problem; 2021. arXiv preprint: 2108.12029v1.
  • Polyak BT, Juditsky AB. Acceleration of stochastic approximation by averaging. SIAM J Control Optim. 1992;30(4):838–855.
  • Yang T, Lin Q. RSG: beating subgradient method without smoothness and strong convexity. J Mach Learn Res. 2018;19(6):1–33.
  • Gorbunov E, Hanzely F, Richtarik P. A unified theory of SGD: variance reduction, sampling, quantization and coordinate descent. In: Proceedings of the 23rd international conference on artificial intelligence and statistics; vol. 108, 2020.
  • Johnson R, Zhang T. Accelerating stochastic gradient descent using predictive variance reduction. In: Advances in neural information processing systems; 2013. p. 315–323.
  • Lin H, Mairal J, Harchaoui Z. A universal catalyst for first-order optimization. In: Advances in neural information processing systems conference; 2015.
  • Necoara I, Singh NK. Stochastic subgradient projection methods for composite optimization with functional constraints. J Mach Learn Res. 2022;23:1–35.
  • Necoara I, Nesterov Y, Glineur F. Linear convergence of first order methods for non-strongly convex optimization. Math Program. 2019;175(1):69–107.
  • Lewis A, Pang JS. Error bounds for convex inequality systems. In: Crouzeix J-P, Martinez-Legaz J-E, Volle M, editors. Generalized convexity, generalized monotonicity. Cambridge University Press; 1998. p. 75–110.
  • Nesterov Y. Lectures on convex optimization, vol. 137. Springer Optimization and Its Applications; 2018.
  • Richtarik P, Takac M. On optimal probabilities in stochastic coordinate descent methods. Optim Lett. 2016;10(6):1233–1243.
  • Polyak BT. Minimization of unsmooth functionals. USSR Comput Math Math Phys. 1969;9(3):14–29.
  • Gaines BR, Kim J, Zhou H. Algorithms for fitting the constrained Lasso. J Comput Graph Stat. 2018;27(4):861–871.
  • Hu Q, Zeng P, Lin L. The dual and degrees of freedom of linearly constrained generalized Lasso. Comput Stat Data Anal. 2015;86:13–26.
  • James GM, Paulsonand C, Rusmevichientong P. Penalized and constrained optimization: an application to high-dimensional website advertising. SIAM J Optim. 2019;30(4):3230–3251.
  • Grant M, Boyd S. CVX: matlab software for disciplined convex programming, version 2.0 beta; 2013. Available from: http://cvxr.com/cvx.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.