| Literature DB >> 30034040 |
Heather Battey1,2, Jianqing Fan1,3, Han Liu1, Junwei Lu1, Ziwei Zhu1.
Abstract
This paper studies hypothesis testing and parameter estimation in the context of the divide-and-conquer algorithm. In a unified likelihood based framework, we propose new test statistics and point estimators obtained by aggregating various statistics from k subsamples of size n/k, where n is the sample size. In both low dimensional and sparse high dimensional settings, we address the important question of how large k can be, as n grows large, such that the loss of efficiency due to the divide-and-conquer algorithm is negligible. In other words, the resulting estimators have the same inferential efficiencies and estimation rates as an oracle with access to the full sample. Thorough numerical results are provided to back up the theory.Entities:
Keywords: 62F10; Divide and conquer; Primary 62F05; debiasing; massive data; secondary 62F12; thresholding
Year: 2018 PMID: 30034040 PMCID: PMC6051757 DOI: 10.1214/17-AOS1587
Source DB: PubMed Journal: Ann Stat ISSN: 0090-5364 Impact factor: 4.028