deepvariant

Run GPU-accelerated deepvariant algorithm.

Parabricks has accelerated Google Deepvariant to extensively use GPUs and finish 30x WGS analysis in 25 minutes instead of hours. The Parabricks flavor of Deepvariant is more like other command line tools that users are familiar with: It takes a BAM and reference as inputs and produces variants as outputs.

Currently, Deepvariant is supported for T4, V100, and A100 GPUs out of the box. Please visit Models for additional GPUs section for more details.

Note

In version 3.8, we are introducing the --run-partition option which can lead to significant speedups. However, using the three options --run-partition, --proposed-variants and --gvcf at the same time will lead to a substantial slowdown. A warning will be issued and the --run-partition option will be ignored.

Copy
Copied!
            

$ pbrun deepvariant \ --ref Ref/Homo_sapiens_assembly38.fasta \ --in-bam mark_dups_gpu.bam \ --out-variants output.vcf

The commands below are the Google counterpart of the Parabricks command above. The output from these commands will be identical to the output from the above command. See the Output Comparison page for comparing the results.

Copy
Copied!
            

# Run make_examples in parallel seq 0 $((N_SHARDS-1)) | \ parallel --eta --halt 2 --joblog "${LOGDIR}/log" --res "${LOGDIR}" \ sudo docker run \ -v ${HOME}:${HOME} \ gcr.io/deepvariant-docker/deepvariant:"${BIN_VERSION}" \ /opt/deepvariant/bin/make_examples \ --mode calling \ --ref "${REF}" \ --reads "${BAM}" \ --examples "${OUTPUT_DIR}/examples.tfrecord@${N_SHARDS}.gz" \ --task {} # Run call_variants in parallel sudo docker run \ -v ${HOME}:${HOME} \ gcr.io/deepvariant-docker/deepvariant:"${BIN_VERSION}" \ /opt/deepvariant/bin/call_variants \ --outfile "${CALL_VARIANTS_OUTPUT}" \ --examples "${OUTPUT_DIR}/examples.tfrecord@${N_SHARDS}.gz" \ --checkpoint "${MODEL}" # Run postprocess_variants in parallel sudo docker run \ -v ${HOME}:${HOME} \ gcr.io/deepvariant-docker/deepvariant:"${BIN_VERSION}" \ /opt/deepvariant/bin/postprocess_variants \ --ref "${REF}" \ --infile "${CALL_VARIANTS_OUTPUT}" \ --outfile "${FINAL_OUTPUT_VCF}"

Parabricks Deep Variant supports the following models:

  1. Deep Variant WGS

  2. Deep Variant WES

  3. Deep Trio

    1. Parent

    2. Child

Deep Variant models for T4, V100 and A100 GPUs ship with the software. Additional models for A10, A30, A40 and A6000 GPUs can be downloaded from this NGC resource.

Run DeepVariant to convert BAM/CRAM to VCF.

Input/Output file options

--ref REF

Path to the reference file. (default: None)

Option is required.

--in-bam IN_BAM

Path to the input BAM/CRAM file for variant calling. (default: None)

Option is required.

--interval-file INTERVAL_FILE

Path to a BED file (.bed) for selective access. This option can be used multiple times. (default: None)

--out-variants OUT_VARIANTS

Path of the vcf/g.vcf/g.vcf.gz file after variant calling. (default: None)

Option is required.

--pb-model-file PB_MODEL_FILE

Path to a non-default parabricks model file for deepvariant. (default: None)

--proposed-variants PROPOSED_VARIANTS

Path of the VCF file which has proposed variants for the make examples stage. (default: None)

Tool Options:

--disable-use-window-selector-model

Change the window selector model from Allele Count Linear to Variant Reads. This option will increase the accuracy and runtime. (default: None)

--gvcf

Generate variant calls in .gvcf Format. (default: None)

--norealign-reads

Do not locally realign reads before calling variants. Reads longer than 500 base pairs are never realigned. (default: None)

--sort-by-haplotypes

Reads are sorted by haplotypes (using HP tag) (default: None)

--keep-duplicates

Keep reads that are duplicate. (default: None)

--vsc-min-count-snps VSC_MIN_COUNT_SNPS

SNP alleles occurring at least this many times in the AlleleCount will be advanced as candidates. (default: 2)

--vsc-min-count-indels VSC_MIN_COUNT_INDELS

Indel alleles occurring at least this many times in the AlleleCount will be advanced as candidates. (default: 2)

--vsc-min-fraction-snps VSC_MIN_FRACTION_SNPS

SNP alleles occurring at least this fraction of all counts in the AlleleCount will be advanced as candidates. (default: 0.12)

--vsc-min-fraction-indels VSC_MIN_FRACTION_INDELS

Indel alleles occurring at least this fraction of all counts in the AlleleCount will be advanced as candidates. (default: None)

--min-mapping-quality MIN_MAPPING_QUALITY

By default, reads with any mapping quality are kept. Setting this field to a positive integer i will only keep reads that have a MAPQ >= i. Note this only applies to aligned reads. (default: 5)

--min-base-quality MIN_BASE_QUALITY

Minimum base quality. This option enforces a minimum base quality score for alternate alleles. Alternate alleles will only be considered if all bases in the allele have a quality greater than min_base_quality. (default: 10)

--mode MODE

Value can be one of [shortread, pacbio, ont]. By default, it is shortread. If mode is set to pacbio, the following defaults are used: --norealign-reads, --alt-aligned-pileup diff_channels, --vsc-min-fraction-indels 0.12. If mode is set to ont, the following defaults are used: -norealign-reads, --variant-caller VCF_CANDIDATE_IMPORTER. (default: shortread)

--alt-aligned-pileup ALT_ALIGNED_PILEUP

Value can be one of [none, diff_channels]. Include alignments of reads against each candidate alternate allele in the pileup image. (default: None)

--variant-caller VARIANT_CALLER

Value can be one of [VERY_SENSITIVE_CALLER, VCF_CANDIDATE_IMPORTER]. The caller to use to make examples. If you use VCF_CANDIDATE_IMPORTER, it implies force calling. Default is VERY_SENSITIVE_CALLER (default: None)

--add-hp-channel

Add another channel to represent HP tags per read. (default: None)

--parse-sam-aux-fields

Auxiliary fields of the BAM/CRAM records are parsed. If either --sort-by-haplotypes or --add-hp-channel is set then this option must also be set. (default: None)

--use-hp-information

If passed, both --sort-by-haplotypes and --parse-sam-aux-fields will be set to true. Only used in PACBIO mode. (default: None)

--use-wes-model

If passed, the WES model file will be used. Only used in shortread mode. (default: None)

--run-partition

Divide the whole genome into multiple partition and run multiple processes at the same time, each on 1 partition. (default: None)

-L INTERVAL, --interval INTERVAL

Interval within which to call the variants from the BAM/CRAM file. Overlapping intervals will be combined. Interval files should be passed using the --interval-file option. This option can be used multiple times (e.g. "-L chr1 -L chr2:10000 -L chr3:20000+ -L chr4:10000-20000") (default: None)

Common options:

--logfile LOGFILE

Path to the log file. If not specified, messages will only be written to the standard error output. (default: None)

--tmp-dir TMP_DIR

Full path to the directory where temporary files will be stored.

--with-petagene-dir WITH_PETAGENE_DIR

Full path to the PetaGene installation directory. By default, this should have been installed at /opt/petagene. Use of this option also requires that the PetaLink library has been preloaded by setting the LD_PRELOAD environment variable. Optionally set the PETASUITE_REFPATH and PGCLOUD_CREDPATH environment variables that are used for data and credentials (default: None)

--keep-tmp

Do not delete the directory storing temporary files after completion.

--license-file LICENSE_FILE

Path to license file license.bin if not in the installation directory.

--no-seccomp-override

Do not override seccomp options for docker (default: None).

--version

View compatible software versions.

GPU options:

--num-gpus NUM_GPUS

Number of GPUs to use for a run. GPUs 0..(NUM_GPUS-1) will be used.

--gpu-devices GPU_DEVICES

GPU devices to use for a run. By default, all GPU devices will be used. To use specific GPU devices, enter a comma-separated list of GPU device numbers. Possible device numbers can be found by examining the output of the nvidia-smi command. For example, using --gpu-devices 0,1 would only use the first two GPUs.

© Copyright 2023, Nvidia. Last updated on Jun 28, 2023.