References
- Banerjee M, Durot C, Sen B. Divide and conquer in nonstandard problems and the super-efficiency phenomenon. Ann Stat. 2019;47(2):720–757.
- Battey H, Fan J, Liu H, et al. Distributed estimation and inference with statistical guarantees; 2015. arXiv preprint arXiv:1509.05457.
- Fan J, Wang D, Wang K, et al. Distributed estimation of principal eigenspaces. Ann Stat. 2019;47(6):3009–3031.
- Lee JD, Sun Y, Liu Q, et al. Communication-efficient sparse regression: a one-shot approach; 2015. arXiv preprint arXiv:1503.04337.
- Van de Geer S, Bühlmann P, Ritov YA, et al. On asymptotically optimal confidence regions and tests for high-dimensional models. Ann Stat. 2014;42(3):1166–1202.
- Volgushev S, Chao SK, Cheng G. Distributed inference for quantile regression processes. Ann Stat. 2019;47(3):1634–1662.
- Zhang Y, Duchi J, Wainwright M. Communication-efficient algorithms for statistical optimization. J Mach Learn Res. 2013;14:3321–3363.
- Jordan MI, Lee JD, Yang Y. Communication-efficient distributed statistical inference. J Am Stat Assoc. 2019;114(526):668–681.
- Little R, Rubin D. Statistical analysis with missing data. 2nd ed. New York: Wiley; 2002.
- Horton NJ, Laird NM. Maximum likelihood analysis of generalized linear models with missing covariates. Stat Methods Med Res. 1999;8(1):37–50.
- Ma X, Wang J. Robust inference using inverse probability weighting. J Am Stat Assoc. 2020;115(532):1851–1860.
- Wooldridge JM. Inverse probability weighted estimation for general missing data problems. J Econom. 2007;141(2):1281–1301.
- Rubin DB. Multiple imputation for nonresponse in surveys. Vol. 81. New York: John Wiley and Sons; 2004.
- Ibrahim JG, Chen MH, Lipsitz SR. Bayesian methods for generalized linear models with covariates missing at random. Can J Stat. 2002;30(1):55–78.
- Shi J, Qin G, Zhu H, et al. Communication-efficient distributed M-estimation with missing data. Comput Stat Data Anal. 2021;161:Article ID 107251.
- Zhang CH, Huang J. The sparsity and bias of the lasso selection in high-dimensional linear regression. Ann Stat. 2008;36(4):1567–1594.
- Chen X, Lee JD, Tong XT, et al. Statistical inference for model parameters in stochastic gradient descent. Ann Stat. 2020;48(1):251–273.
- Negahban SN, Ravikumar P, Wainwright MJ, et al. A unified framework for high-dimensional analysis of M-estimators with decomposable regularizers. Stat Sci. 2012;27(4):538–557.
- Van de Geer SA, Bühlmann P. On the conditions used to prove oracle results for the lasso. Electron J Stat. 2009;3:1360–1392.
- Wu X, Li F, Kumar A, et al. Bolt-on differential privacy for scalable stochastic gradient descent-based analytics. In: Proceedings of the 2017 ACM international conference on management of data; 2017. p. 1307–1322.
- Xu P, Chen J, Zou D, et al. Global convergence of Langevin dynamics based algorithms for nonconvex optimization. Adv Neural Inf Process Syst. 2018;31.
- Fan J, Guo Y, Wang K. Communication-efficient accurate statistical estimation. J Am Stat Assoc. 2021;1–11.
- Fan J, Li R. Variable selection via nonconcave penalized likelihood and its oracle properties. J Am Stat Assoc. 2001;96(456):1348–1360.
- Zou H. The adaptive lasso and its oracle properties. J Am Stat Assoc. 2006;101(476):1418–1429.
- Zhang CH. Nearly unbiased variable selection under minimax concave penalty. Ann Stat. 2010;38(2):894–942.
- Shao J, Wang L. Semiparametric inverse propensity weighting for nonignorable missing data. Biometrika. 2016;103(1):175–187.