2014
DOI: 10.5705/ss.2013.088
|View full text |Cite
|
Sign up to set email alerts
|

A split-and-conquer approach for analysis of

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

3
161
0

Year Published

2014
2014
2022
2022

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 134 publications
(164 citation statements)
references
References 0 publications
3
161
0
Order By: Relevance
“…A partial list of references covering DC algorithms from a statistical perspective is Chen and Xie (2012), Zhang et al (2013), Kleiner et al (2014), Liu and Ihler (2014) and Zhao et al (2014a). The closest works to ours are Zhang et al (2013), Lee et al (2015) and Rosenblatt and Nadler (2016).…”
Section: Introductionmentioning
confidence: 93%
“…A partial list of references covering DC algorithms from a statistical perspective is Chen and Xie (2012), Zhang et al (2013), Kleiner et al (2014), Liu and Ihler (2014) and Zhao et al (2014a). The closest works to ours are Zhang et al (2013), Lee et al (2015) and Rosenblatt and Nadler (2016).…”
Section: Introductionmentioning
confidence: 93%
“…Although a large number of statistical methods and computational recipes have been developed to address various challenges for big data analytics, such as the subsampling-based methods (Liang et al, 2013;Kleiner et al, 2014;Ma et al, 2015) divide-and-conquer techniques (Lin and Xi, 2011;Guha et al, 2012;Chen and Xie, 2014;Tang et al, 2019;Zhou and Song, 2017), little is known about statistical inference in streaming data analyses under dynamic data storage and incremental updates. This paper has filled the gap with the proposed renewable estimation and incremental inference.…”
Section: Discussionmentioning
confidence: 99%
“…Chen and Xie (2014) consider a divide and conquer approach for generalized linear models (GLM) where both the sample size n and the number of covariates p are large, by incorporating variable selection via penalized regression into a subset processing step. More specifically, for p bounded or increasing to infinity slowly, ( p n not faster than o ( e n k ), while model size may increase at a rate of o ( n k )), they propose to first randomly split the data of size n into K blocks (size n k = O ( n/K )).…”
Section: Methodsmentioning
confidence: 99%
“…Sound statistical procedures that are scalable computationally to massive datasets have been proposed (Jordan, 2013). Examples are subsampling-based approaches (Kleiner et al, 2014; Ma, Mahoney and Yu, 2013; Liang et al, 2013; Maclaurin and Adams, 2014), divide and conquer approaches (Lin and Xi, 2011; Chen and Xie, 2014; Song and Liang, 2014; Neiswanger, Wang and Xing, 2013), and online updating approaches (Schifano et al, 2015). From a computational perspective, much effort has been put into the most active, open source statistical environment, (R Core Team, 2014a).…”
Section: Introductionmentioning
confidence: 99%