References
- Berger, Y. G., and De La Riva Torres, O. (2016), “Empirical Likelihood Confidence Intervals for Complex Sampling Designs,” Journal of the Royal Statistical Society, Series B, 78, 319–314. DOI: https://doi.org/10.1111/rssb.12115.
- Breidt, F. J., and Opsomer, J. D. (2000), “Local Polynomial Regression Estimators in Survey Sampling,” The Annals of Statistics, 28, 1026–1053. DOI: https://doi.org/10.1214/aos/1015956706.
- Chen, K., Hu, I., and Ying, Z. (1999), “Strong Consistency of Maximum Quasi-Likelihood Estimators in Generalized Linear Models With Fixed and Adaptive Designs,” The Annals of Statistics, 27, 1155–1163. DOI: https://doi.org/10.1214/aos/1017938919.
- Chen, X. (2011), Quasi Likelihood Method for Generalized Linear Model (in Chinese), Hefei: Press of University of Science and Technology of China.
- Dhillon, P. S., Lu, Y., Foster, D., and Ungar, L. (2013), “New Subsampling Algorithms for Fast Least Squares Regression,” in International Conference on Neural Information Processing Systems, pp. 360–368.
- Drineas, P., Mahoney, M. W., Muthukrishnan, S., and Sarlós, T. (2011), “Faster Least Squares Approximation,” Numerische Mathematik, 117, 219–249. DOI: https://doi.org/10.1007/s00211-010-0331-6.
- Duchi, J. C., Agarwal, A., and Wainwright, M. J. (2012), “Dual Averaging for Distributed Optimization: Convergence Analysis and Network Scaling,” IEEE Transactions on Automatic Control, 57, 592–606. DOI: https://doi.org/10.1109/TAC.2011.2161027.
- Fahrmeir, L., and Tutz, G. (2001), Multivariate Statistical Modelling Based on Generalized Linear Models, New York: Springer-Verlag.
- Jordan, M. I., Lee, J. D., and Yang, Y. (2019), “Communication-Efficient Distributed Statistical Inference,” Journal of the American Statistical Association, 114, 668–681. DOI: https://doi.org/10.1080/01621459.2018.1429274.
- Kleiner, A., Talwalkar, A., Sarkar, P., and Jordan, M. I. (2015), “A Scalable Bootstrap for Massive Data,” Journal of the Royal Statistical Society, Series B, 76, 795–816. DOI: https://doi.org/10.1111/rssb.12050.
- Li, R., Lin, D. K., and Li, B. (2013), “Statistical Inference in Massive Data Sets,” Applied Stochastic Models in Business and Industry, 29, 399–409. DOI: https://doi.org/10.1002/asmb.1927.
- Lin, N., and Xi, R. (2011), “Aggregated Estimating Equation Estimation,” Statistics & Its Interface, 1, 73–83. DOI: https://doi.org/10.4310/SII.2011.v4.n1.a8.
- Ma, P., Mahoney, M. W., and Yu, B. (2015), “A Statistical Perspective on Algorithmic Leveraging,” Journal of Machine Learning Research, 16, 861–919.
- Mahoney, M. W. (2012), “Randomized Algorithms for Matrices and Data,” Foundations and Trends[textregistered] in Machine Learning, 3, 647–672.
- Mccullagh, P., and Nelder, J. A. (1989), Generalized Linear Models, Monographs on Statistics and Applied Probability (Vol. 37), London: Chapman & Hall.
- Neath, A. A., and Cavanaugh, J. E. (2012), “The Bayesian Information Criterion: Background, Derivation, and Applications,” Wiley Interdisciplinary Reviews: Computational Statistics, 4, 199–203. DOI: https://doi.org/10.1002/wics.199.
- Newey, W. K., and McFadden, D. (1994), “Large Sample Estimation and Hypothesis Testing,” in Handbook of Econometrics (Vol. 4), eds. R. F. Engle and D. L. McFadden, Amsterdam: Elsevier, pp. 2111–2245.
- Pukelsheim, F. (2006), Optimal Design of Experiments, Philadelphia, PA: Society for Industrial and Applied Mathematics.
- Quiroz, M., Kohn, R., Villani, M., and Tran, M.-N. (2019), “Speeding Up MCMC by Efficient Data Subsampling,” Journal of the American Statistical Association, 114, 831–843. DOI: https://doi.org/10.1080/01621459.2018.1448827.
- R Core Team (2018), R: A Language and Environment for Statistical Computing, Vienna, Austria: R Foundation for Statistical Computing, available at https://www.R-project.org/.
- Rao, C. R., Toutenburg, H., Shalabh and Heumann, C. (2007), Linear Models and Generalizations: Least Squares and Alternatives (3rd ed.), Berlin, Heidelberg: Springer Publishing Company, Inc.
- Särndal, C. E., Swensson, B., and Wretman, J. (1992), Model Assisted Survey Sampling, New York: Springer.
- Schifano, E. D., Wu, J., Wang, C., Yan, J., and Chen, M.-H. (2016), “Online Updating of Statistical Inference in the Big Data Setting,” Technometrics, 58, 393–403. DOI: https://doi.org/10.1080/00401706.2016.1142900.
- Tang, J., Zhang, J., Yao, L., Li, J., Zhang, L., and Su, Z. (2008), “Arnetminer: Extraction and Mining of Academic Social Networks,” in KDD’08, pp. 990–998.
- Tzavelas, G. (1998), “A Note on the Uniqueness of the Quasi-Likelihood Estimator,” Statistics & Probability Letters, 38, 125–130. DOI: https://doi.org/10.1016/S0167-7152(97)00162-4.
- van der Vaart, A. (1998), Asymptotic Statistics, New York: Cambridge University Press.
- Wang, H. Y., Yang, M., and Stufken, J. (2019), “Information-Based Optimal Subdata Selection for Big Data Linear Regression,” Journal of the American Statistical Association, 114, 393–405. DOI: https://doi.org/10.1080/01621459.2017.1408468.
- Wang, H. Y., Zhu, R., and Ma, P. (2018), “Optimal Subsampling for Large Sample Logistic Regression,” Journal of the American Statistical Association, 113, 829–844. DOI: https://doi.org/10.1080/01621459.2017.1292914.