You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Right now, the creation of the generalized bootstrap replicate factors requires a lot of computer memory, because:
The algorithm forms the entire quadratic form matrix of the target variance estimator, which is $n \times n$. Even with a survey of only 1,000 cases, that results in a large matrix.
Sampling from the multivariate normal distribution for $B$ replicates results in a $n \times B$ matrix, where $B$ can be quite large.
A couple ideas for making this more feasible are:
Form bootstrap adjustment factors separately by stratum. The covariances of units in different strata are all zero, so we can get correct results by generating adjustment factors separately for each stratum. This means that instead of creating one big $n \times n$ covariance matrix, we can generate $H$ covariance matrices of size $n_h \times n_h$. Within each of the separate strata $H$, the created columns of replicate weights should be shuffled to ensure independence.
Update: (1) is done where possible.
Use quadratic forms whose dimension matches the rank. For example, if a the dataset has 10 rows but three clusters, form a $3 \times 3$ quadratic form, generate adjustment factors for three clusters, and then expand the matrix of replicates to the within-cluster units.
Right now, the creation of the generalized bootstrap replicate factors requires a lot of computer memory, because:
A couple ideas for making this more feasible are:
The text was updated successfully, but these errors were encountered: