UCSC Computational Genomics Lab
vg is a toolkit for DNA sequence analysis using variation graphs. Toil-vg is a Toil-based framework for running common vg pipelines at scale, either locally or on a distributed computing environment:
toil-vg construct
: Create vg graphs and indexes from FASTA and VCF, constructing contigs in parallel. toil-vg index
: Produce a GCSA, GBWT and/or XG index from input vg graphs.toil-vg map
: Produce a graph alignment (GAM) for each chromosome from input reads and indextoil-vg call
: Produce VCF from input XG index and GAM(s).Why use toil-vg?
toil-vg construct
.Installation requires Python and Toil. We recommend installing within a virtualenv as follows
virtualenv -p python3.6 toilvenv
source toilvenv/bin/activate
pip install toil[aws,mesos]==3.24.0
pip install toil-vg
Developers may want to work with the latest master branch from github instead:
git clone https://github.com/vgteam/toil-vg.git
cd toil-vg
virtualenv toilvenv
source toilvenv/bin/activate
make prepare
make develop
# After installing in this fashion, the tests can be run with
make test
toil-vg can run vg, along with some other tools, via Docker. Docker can be installed locally as follows:
docker version
doesn't work, try adding user to docker group with sudo usermod -aG docker $USER
, then log out and back in.docker version
doesn't work, try adding docker environment variables: docker-machine start; docker-machine env; eval "$(docker-machine env default)"
Note If you get an error message of the form APIError: 502 Server Error: Bad Gateway (Mounts denied:...
, you can fix this by either adding /var/folders
to the directories list in the "File Sharing" section of the Docker preferences dialog, or by specifying a temporary working directory that already is shared using the --tempDir
toil-vg option or the TMPDIR
environment variable. --container None
, toil-vg requires the following command line tools to be installed on the system: vg
, pigz
, tabix
, and bcftools
. Note that the HS38D1 example below uses the official 1000 Genomes liftover VCFs, which contain fatal semantic errors, and are rejected by any bcftools
newer than 1.9. jq
, samtools
, rtg vcfeval
, 'hap.py', and 'Platypus.py' are also necessary for certain tests. A configuration file can be used as an alternative to some command line options, as well as to tune Toil job resources. A default configuration file can be generated using
toil-vg generate-config > config.yaml
Pass this file to toil-vg
commands using the --config
option.
For non-trivial inputs, care must be taken to specify the resource requirements for the different pipeline phases (via the command line or by editing the config file), as they all default to single-core and 4G of ram.
To generate a default configuration for running at genome scale on a cluster with 32-core worker nodes, use
toil-vg generate-config --whole_genome > config_wg.yaml
The jobStore and outStore arguments to toil-vg are directories that will be created if they do not already exist. When starting a new job, toil will complain if the jobStore exists, so use toil clean <jobStore>
first. When running on Mesos, these stores should be S3 buckets. They are specified using the following format aws:region:bucket (see examples below).
All other input files can either either be local (best to specify absolute path) or URLs specified in the normal manner, such as: http://address/input_file
or s3://bucket/input_file
. The config file must always be local. When using an S3 jobstore, it is preferable to pass input files from S3 as well, as they load much faster and less cluster time will be wasted importing data.
The following command will construct and index a graph from the 1000 Genomes calls for the HS38D1 (GRCH38 analysis set excluding alt sequences + decoys) reference.
# Toil boilerplate
export TOIL_OPTS="--realTimeLogging --logInfo --realTimeStderr"
# Directory for Toil's temporary files
export TOIL_JS="./my-jobstore"
# All output will be written here
export TOIL_OS="./my-output"
# Construct graph and all (XG, GBWT, GCSA, Snarls, id-ranges) indexes
toil-vg construct $TOIL_JS $TOIL_OS --pangenome --out_name 1kg_hs38d1 --all_index --merge_graphs --add_chr_prefix --whole_genome_config $TOIL_OPTS --fasta ftp://ftp-trace.ncbi.nih.gov/1000genomes/ftp/technical/reference/GRCh38_reference_genome/GRCh38_full_analysis_set_plus_decoy_hla.fa --vcf $(for i in $(seq 1 22; echo X; echo Y); do echo ftp://ftp.1000genomes.ebi.ac.uk/vol1/ftp/release/20130502/supporting/GRCh38_positions/ALL.chr${i}_GRCh38.genotypes.20170504.vcf.gz; done) --regions $(for i in $(seq 1 22; echo X; echo Y) ; do echo chr${i}; done) --fasta_regions --regions_regex 'chr.*_random' 'chrUn_[a-zA-Z0-9]*' 'chr.*decoy'
Filtering out low-frequency variants (often recommended) can be done by replacing --pangenome
by --min_af 0.01
(both options can be used to simultaneously create filtered and unfiltered graphs). This command uses --fasta_regions
to add every sequence from the input fasta file to the graph but --regions_regex
to filter down to just non-alt sequences.
Alt sequences can be added via alignment by way of the --alt_regions_bed
option. Use --alt_regions_bed https://raw.githubusercontent.com/vgteam/toil-vg/master/data/grch38-alt-positions-no-hla-no-chr6_GL000251v2_alt.bed
for example to add the non-HLA alts. I higher AF threshold than above is advisable to reduce running time.
By default, this command will use every core on the system. Use --maxCores
to limit to fewer processes. It is recommended to have 3TB disk, 256GB RAM and 32 cores to run this, and even then it will take several days to complete. See below for how the above command can be adapted to run on Amazon EC2.
wget https://raw.githubusercontent.com/BD2KGenomics/toil-vg/master/scripts/create-ec2-leader.sh
./create-ec2-leader.sh <leader-name> <keypair-name>
Log into the leader with
toil ssh-cluster <leader-name> --zone us-west-2a
In order to log onto a worker node instead of the leader, find its public IP from the EC2 Management Console or command line, and log in using the core username: ssh core@public-ip
Destroy the leader when finished with it. After logging out with exit
:
toil destroy-cluster -z us-west-2a myleader
Log onto the leader node with toil ssh-cluster
as described above and open a screen
session. The same construction example as above can now be run after adapting the input environment variables as follows:
# Use maximum of 8 i3.8xlarge nodes with spot bit of $0.80
export NODE_OPTS="--nodeTypes i3.8xlarge:0.80,i3.8xlarge --maxNodes 8"
# Set Toil to use AWS autoscaling
export AWS_OPTS="--defaultPreemptable --batchSystem mesos --provisioner aws --retryCount 3 --metrics"
# Toil boilerplate
export TOIL_OPTS="$NODE_OPTS $AWS_OPTS --realTimeLogging --logInfo --realTimeStderr"
# S3 Bucket for Toil's temporary files
export TOIL_JS="aws:us-west-2:my-jobstore"
# All output will be written here
export TOIL_OS="aws:us-west-2/my-bucket/hs38d1-output"
For a performance dashboard, browse to localhost:3000
on the computer from which you ran toil ssh-cluster
(this is enabled by the --metrics
option above)