Optimal Distributed Subsampling for Maximum Quasi-Likelihood Estimators With Massive Data

Jun Yu, Hai Ying Wang, Mingyao Ai*, Huiming Zhang

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

56 引用 (Scopus)

摘要

Nonuniform subsampling methods are effective to reduce computational burden and maintain estimation efficiency for massive data. Existing methods mostly focus on subsampling with replacement due to its high computational efficiency. If the data volume is so large that nonuniform subsampling probabilities cannot be calculated all at once, then subsampling with replacement is infeasible to implement. This article solves this problem using Poisson subsampling. We first derive optimal Poisson subsampling probabilities in the context of quasi-likelihood estimation under the A- and L-optimality criteria. For a practically implementable algorithm with approximated optimal subsampling probabilities, we establish the consistency and asymptotic normality of the resultant estimators. To deal with the situation that the full data are stored in different blocks or at multiple locations, we develop a distributed subsampling framework, in which statistics are computed simultaneously on smaller partitions of the full data. Asymptotic properties of the resultant aggregated estimator are investigated. We illustrate and evaluate the proposed strategies through numerical experiments on simulated and real datasets. Supplementary materials for this article are available online.

源语言英语
页(从-至)265-276
页数12
期刊Journal of the American Statistical Association
117
537
DOI
出版状态已出版 - 2022

指纹

探究 'Optimal Distributed Subsampling for Maximum Quasi-Likelihood Estimators With Massive Data' 的科研主题。它们共同构成独一无二的指纹。

引用此