2018
DOI: 10.5705/ss.202016.0019
|View full text |Cite
|
Sign up to set email alerts
|

Clusters with random size: maximum likelihood versus weighted estimation

Abstract: The analysis of hierarchical data that take the form of clusters with random size has received considerable attention. The focus here is on samples that are very large in terms of number of clusters and/or members per cluster, on the one hand, as well as on very small samples (e.g., when studying rare diseases), on the other. Whereas maximum likelihood inference is straightforward in medium to large samples, in samples of sizes considered here it may become prohibitive. We propose sample-splitting (Molenberghs… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5

Relationship

3
2

Authors

Journals

citations
Cited by 8 publications
(17 citation statements)
references
References 33 publications
0
17
0
Order By: Relevance
“…In the LMM, it facilitates the estimation process, because with balanced clusters there are complete sufficient statistics and a closed-form MLE exists [13]. For more details on the split-sample methodology, we refer to [12].…”
Section: Split-sample Methodology For Clustered Datamentioning
confidence: 99%
See 2 more Smart Citations
“…In the LMM, it facilitates the estimation process, because with balanced clusters there are complete sufficient statistics and a closed-form MLE exists [13]. For more details on the split-sample methodology, we refer to [12].…”
Section: Split-sample Methodology For Clustered Datamentioning
confidence: 99%
“…However, V β has to be modified accordingly to estimate D. Furthermore, a specific covariance structure for D, e.g., compound-symmetry (CS) or autoregressive (AR), can be considered. Each case implies that a different system of equations needs to be solved to find an estimator of D. Findings by [12] and [11] for normally distributed hierarchical data with CS and AR structure can be extended to non-Gaussian outcomes, but arguable will be more complicated. In most cases, we may rely on approximations to estimate D. Therefore, a biased estimator is expected.…”
Section: Final Remarksmentioning
confidence: 99%
See 1 more Smart Citation
“…While useful for surrogacy evaluation, it is by no means restricted to that setting. Our estimator is based on so-called split-sample methodology and pseudo-likelihood (Molenberghs et al, 2011(Molenberghs et al, , 2014(Molenberghs et al, , 2018. Here, the sample is conveniently divided into subsamples, and the parameters are estimated in each one.…”
Section: Introductionmentioning
confidence: 99%
“…In our case, each subsample contains one single trial, leading to the so-called trial-by-trial or cluster-by-cluster estimator (Molenberghs et al, 2018). This method has been shown to exhibit good statistical performance and computational efficiency to analyze data with different clustering structures, such as autoregressive (AR) (Hermans et al, 2017) and compound-symmetry (Molenberghs et al, 2018). For the latter, the cluster-by-cluster estimator is consistent when the number of replicates per cluster increases more rapidly than the number of clusters.…”
Section: Introductionmentioning
confidence: 99%