-
Notifications
You must be signed in to change notification settings - Fork 194
Indexing Huge Datasets
The index construction guidelines work best with datasets that are no larger than the 1000 Genomes Project (2500 human samples). This page describes techniques that may be useful when trying to index larger datasets.
Many of these techniques use standalone GBWT tools in addition to vg
. See the GBWT wiki for further documentation.
- Version 1.11.0 or later
By default, GBWT stores haplotype identifiers at one out of 1024 positions on each path. The distance can be changed by using option -n N
of vg index -G
. This is a time/space trade-off. Each stored identifier typically takes 3 to 4 bytes of memory, while the time required for determining the haplotypes containing a particular subpath is proportional to the distance between stored identifiers. In a 1000GP GBWT with the default distance, the stored identifiers take roughly as much space as the GBWT itself. If the number of samples is larger, the distance should also be proportionally larger.
In the standalone tool deps/gbwt/build_gbwt
, the corresponding option is -s N
.
- Version 1.11.0 or later
Parsing a massive VCF file can take hours or even days. Because the parse does not depend on GBWT construction parameters other than -P
, parsing the file only once and storing the intermediate data on disk may save time when trying different GBWT construction parameters. By replacing the option -G FILE
with -e FILE
, vg index
will write the intermediate data to files with prefix FILE
instead of building a GBWT index.
File FILE_contig
will contain the reference path and the paths for each alternate allele for VCF contig contig
, while files FILE_contig_X_Y
will contain the phasing information for samples X
to X
+Y
-1. These files can then be used as an input to the standalone GBWT construction tool deps/gbwt/build_gbwt
.
In the following example, we index graph.vg
, which contains VCF contigs chr1
and chr2
. The output is written to graph.gbwt
.
vg index -v variants.vcf.gz -e parse graph.vg
build_gbwt -p -r -o graph parse_chr1 parse_chr2
- Version 1.13.0 or later (PR #2002; November 28, 2018)
GBWT construction for tens of thousands of samples can be 2x to 3x faster, if we first build indexes for batches of e.g. 2500 samples in parallel, and then merge the partial indexes.
The default GBWT merging algorithm (vg gbwt -m
) is mostly suited for merging small indexes, as it is slower than direct construction. The fast algorithm (vg index -f
) only works with GBWT indexes over different contigs. In order to merge GBWT indexes over the same contig quickly, we need the parallel algorithm from the standalone tool deps/gbwt/merge_gbwt
.
In the following, we first parse the VCF file and write the phasing information in batches to parse_chr1_*
. Then we assume that the names of the phasing files have been split into 12 larger batches in files chr1_1.txt
to chr1_12.txt
. We build GBWT indexes for the large batches using 4 parallel jobs. Finally we merge the partial indexes using the parallel algorithm and write the result to chr1.gbwt
.
vg index -v chr1.vcf.gz -e parse chr1.vg
seq 1 12 | parallel -j 4 "./build_gbwt.sh {}"
merge_gbwt -p -o chr1 $(for i in $(seq 1 12); do echo chr1_${i}; done)
Script build_gbwt.sh
builds the GBWT index for the files in chr1_$1.txt
in both orientations (-r
) using 500-million node buffers (-b 500
) and sample interval 16384 (-s 16384
). By adding -S
, we could instruct it to skip overlapping variants, effectively generating chromosome-length haplotypes.
#!/bin/bash
build_gbwt -r -b 500 -s 16384 -p -o chr1_$1 $(for i in $(cat chr1_$1.txt); do echo "-P $i"; done) > /dev/null 2> /dev/null
The thread identifiers are determined by the order of the partial indexes in the arguments of merge_gbwt
. Thread databases cannot be merged at the moment.
Merging takes the first GBWT index as the base. Then, for each subsequent index in the argument list, it builds the rank array for that index relative to the base index, and interleaves the indexes according to the rank array. The merging is heavily parallelized and uses plenty of temporary disk space (2-3 bytes times the total length of the threads in the batch). By default, the temporary files are written to the current directory, but the directory can be changed with -t DIR
.
- Version 1.12.0 or later