2016
DOI: 10.1080/01621459.2015.1080709
|View full text |Cite
|
Sign up to set email alerts
|

A Subsampled Double Bootstrap for Massive Data

Abstract: The bootstrap is a popular and powerful method for assessing precision of estimators and inferential methods. However, for massive datasets which are increasingly prevalent, the bootstrap becomes prohibitively costly in computation and its feasibility is questionable even with modern parallel computing platforms. Recently Kleiner, Talwalkar, Sarkar, and Jordan (2014) proposed a method called BLB (Bag of Little Bootstraps) for massive data which is more computationally scalable with little sacrifice of statisti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

3
49
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 39 publications
(52 citation statements)
references
References 40 publications
3
49
0
Order By: Relevance
“…We have presented the BLBB and SDBB as two data-subsetting procedures to approximate the BB. The BLBB and SDBB are analogous to the BLB (Kleiner et al, 2014) and SDB (Sengupta et al, 2016). The proposed procedures have theoretical and computational properties that are comparable to those of their frequentist counterparts.…”
Section: Discussionmentioning
confidence: 92%
See 3 more Smart Citations
“…We have presented the BLBB and SDBB as two data-subsetting procedures to approximate the BB. The BLBB and SDBB are analogous to the BLB (Kleiner et al, 2014) and SDB (Sengupta et al, 2016). The proposed procedures have theoretical and computational properties that are comparable to those of their frequentist counterparts.…”
Section: Discussionmentioning
confidence: 92%
“…The SDBB is the Bayesian analogue to the subsampled double bootstrap for massive data proposed by Sengupta et al (2016), which also provides an approximation of ξ{π φ (·|X n )}. In Sengupta et al (2016), the authors claim that the SDB outperforms the BLB in some scenarios with limited time budget, especially when it is only possible to run s < n/b little bootstraps. Therefore, we would expect the same phenomenon to occur with the BLBB and SDBB.…”
Section: Subsampled Double Bayesian Bootstrapmentioning
confidence: 99%
See 2 more Smart Citations
“…Among the techniques developed for analyzing massive datasets, there are two categories that are most important. One is the "split-and-conquer" method (Zhang, Duchi and Wainwright, 2013;Chen and Xie, 2014;Battey, et al, 2015), and the other one is the resampling-based methods (Kleiner, et al, 2014;Sengupta, Volgushev and Shao, 2016). In this paper, we consider a general class of symmetric statistics (Lai and Wang, 1993;Jing and Wang, 2010) that encompasses many commonly used statistics, for example, the U and L-statistics.…”
Section: Chapter 1 General Introductionmentioning
confidence: 99%