(Internal) Extracts evidence of structural variations from reads
Category Structural Variant Discovery
Overview
(Internal) Extracts evidence of structural variations from readsThis tool is used in development and should not be of interest to most researchers. It repackages the first two steps of the structural variation workflow as a separate tool for the convenience of developers.
This tool examines a SAM/BAM/CRAM for reads, or groups of reads, that demonstrate evidence of a structural variation in the vicinity. It records this evidence as a group of text files in a specified output directory on Spark's HDFS file system.
Inputs
- A file of paired-end, aligned and coordinate-sorted reads.
Output
- A directory of text files describing the evidence for structural variation discovered.
Usage example
gatk ExtractSVEvidenceSpark \ -I input_reads.bam \ -O hdfs://my_cluster-m:8020/output_directory --aligner-index-image ignored --kmers-to-ignore ignored
This tool can be run without explicitly specifying Spark options. That is to say, the given example command without Spark options will run locally. See Tutorial#10060 for an example of how to set up and run a Spark tool on a cloud Spark cluster.
Additional Information
Read filters
This Read Filter is automatically applied to the data by the Engine before processing by ExtractSVEvidenceSpark.
ExtractSVEvidenceSpark specific arguments
This table summarizes the command-line arguments that are specific to this tool. For more details on each argument, see the list further down below the table or click on an argument name to jump directly to that entry in the list.
Argument name(s) | Default value | Summary | |
---|---|---|---|
Required Arguments | |||
--aligner-index-image |
bwa-mem index image file | ||
--input -I |
BAM/SAM/CRAM file containing reads | ||
--kmers-to-ignore |
file containing ubiquitous kmer list. see FindBadGenomicKmersSpark to generate it. | ||
--output -O |
HDFS path for output | ||
Optional Tool Arguments | |||
--adapter-sequence |
Adapter sequence. | ||
--allowed-short-fragment-overhang |
10 | Proper pairs have the positive strand read upstream of the negative strand read, but we allow this much slop for short fragments. | |
--arguments_file |
read one or more arguments files and add them to the command line | ||
--assembled-contigs-output-order -sort |
coordinate | sorting order to be used for the output assembly alignments SAM/BAM file (currently only coordinate or query name is supported) | |
--assembly-to-mapped-size-ratio-guess |
7 | Guess at the ratio of reads in the final assembly to the number reads mapped to the interval. | |
--bam-partition-size |
0 | maximum number of bytes to read from a file into each partition of reads. Setting this higher will result in fewer partitions. Note that this will not be equal to the size of the partition in memory. Defaults to 0, which uses the default split size (determined by the Hadoop input format, typically the size of one HDFS block). | |
--breakpoint-evidence-dir |
directory for evidence output | ||
--breakpoint-intervals |
file for breakpoint intervals output | ||
--cleaner-max-copy-number |
4 | KmerCleaner maximum copy number (not count, but copy number) for a kmer. Kmers observed too frequently are probably mismapped or ubiquitous. | |
--cleaner-max-intervals |
3 | KmerCleaner maximum number of intervals for a localizing kmer. If a kmer occurs in too many intervals, it isn't sufficiently local. | |
--cleaner-min-kmer-count |
4 | KmerCleaner minimum kmer count for a localizing kmer. If we see it less often than this many times, we're guessing it's erroneous. | |
--conf |
Spark properties to set on the Spark context in the format = | ||
--cross-contigs-to-ignore |
file containing alt contig names that will be ignored when looking for inter-contig pairs | ||
--disable-sequence-dictionary-validation |
false | If specified, do not check the sequence dictionaries from our inputs for compatibility. Use at your own risk! | |
--exclusion-interval-padding |
0 | Exclusion interval padding. | |
--exclusion-intervals |
file of reference intervals to exclude | ||
--external-evidence |
external evidence input file | ||
--external-evidence-uncertainty |
150 | Uncertainty in location of external evidence. | |
--external-evidence-weight |
10 | Weight to give external evidence. | |
--fastq-dir |
output dir for assembled fastqs | ||
--gcs-max-retries -gcs-retries |
20 | If the GCS bucket channel errors out, how many times it will attempt to re-initiate the connection | |
--gcs-project-for-requester-pays |
Project to bill when accessing "requester pays" buckets. If unset, these buckets cannot be accessed. User must have storage.buckets.get permission on the bucket being accessed. | ||
--help -h |
false | display the help message | |
--high-coverage-intervals |
file for high-coverage intervals output | ||
--high-depth-coverage-factor |
3 | We filter out contiguous regions of the genome that have coverage of at least high-depth-coverage-factor * avg-coverage and a peak coverage of high-depth-coverage-peak-factor * avg-coverage, because the reads mapped to those regions tend to be non-local and high depth prevents accurate assembly. | |
--high-depth-coverage-peak-factor |
7 | We filter out contiguous regions of the genome that have coverage of at least high-depth-coverage-factor * avg-coverage and a peak coverage of high-depth-coverage-peak-factor * avg-coverage, because the reads mapped to those regions tend to be non-local and high depth prevents accurate assembly. | |
--include-mapping-location |
true | Include read mapping location in FASTQ files. | |
--interval-merging-rule -imr |
ALL | Interval merging rule for abutting intervals | |
--interval-only-assembly |
false | Don't look for extra reads mapped outside the interval. | |
--intervals -L |
One or more genomic intervals over which to operate | ||
--k-size |
51 | Kmer size. | |
--kmer-intervals |
file for kmer intervals output | ||
--kmer-max-dust-score |
49 | Maximum kmer DUST score. | |
--max-fastq-size |
3000000 | Maximum total bases in FASTQs that can be assembled. | |
--max-tracked-fragment-length |
2000 | Largest fragment size that will be explicitly counted in determining fragment size statistics. | |
--min-coherent-evidence-coverage-ratio |
0.1633408753260167 | Minimum weight of the evidence that shares a distal target locus to validate the evidence, as a ratio of the mean coverage in the BAM. The default value is coherent-count / mean coverage ~ 7 / 42.9 ~ 0.163 | |
--min-evidence-coverage-ratio |
0.35001616141289293 | Minimum weight of the corroborating read evidence to validate some single piece of evidence, as a ratio of the mean coverage in the BAM. The default value is overlap-count / mean coverage ~ 15 / 42.9 ~ 0.350 | |
--min-evidence-mapq |
20 | The minimum mapping quality for reads used to gather evidence of breakpoints. | |
--min-evidence-match-length |
45 | The minimum length of the matched portion of an interesting alignment. Reads that don't match at least this many reference bases won't be used in gathering evidence. | |
--min-kmers-per-interval |
5 | Minimum number of localizing kmers in a valid interval. | |
--num-reducers |
0 | For tools that shuffle data or write an output, sets the number of reducers. Defaults to 0, which gives one partition per 10MB of input. | |
--output-shard-tmp-dir |
when writing a bam, in single sharded mode this directory to write the temporary intermediate output shards, if not specified .parts/ will be used | ||
--program-name |
Name of the program running | ||
--qname-intervals-for-assembly |
file for mapped qname intervals output | ||
--qname-intervals-mapped |
file for mapped qname intervals output | ||
--read-metadata |
output file for read metadata | ||
--reference -R |
Reference sequence | ||
--run-without-gaps-annotation |
false | Allow evidence filter to run without gaps annotation (assume no gaps). | |
--run-without-umap-s100-annotation |
false | Allow evidence filter to run without annotation for single-read mappability of 100-mers (assume all mappable). | |
--sharded-output |
false | For tools that write an output, write the output in multiple pieces (shards) | |
--spark-master |
local[*] | URL of the Spark Master to submit jobs to when using the Spark pipeline runner. | |
--spark-verbosity |
Spark verbosity. Overrides --verbosity for Spark-generated logs only. Possible values: {ALL, DEBUG, INFO, WARN, ERROR, FATAL, OFF, TRACE} | ||
--sv-evidence-filter-model-file |
Path to xgboost classifier model file for evidence filtering | ||
--sv-evidence-filter-threshold-probability |
0.92 | Minimum classified probability for a piece of evidence to pass xgboost evidence filter | |
--sv-evidence-filter-type |
DENSITY | Filter method for selecting evidence to group into Assembly Intervals | |
--sv-genome-gaps-file |
Path to file enumerating gaps in the reference genome, used by classifier to score evidence for filtering. To use classifier without specifying gaps file, pass the flag --run-without-gaps-annotation | ||
--sv-genome-umap-s100-file |
Path to single read 100-mer mappability file in the reference genome, used by classifier to score evidence for filtering. To use classifier without specifying mappability file, pass the flag --run-without-umap-s100-annotation | ||
--target-link-file |
output file for non-assembled breakpoints in bedpe format | ||
--unfiltered-breakpoint-evidence-dir |
directory for evidence output | ||
--use-nio |
false | Whether to use NIO or the Hadoop filesystem (default) for reading files. (Note that the Hadoop filesystem is always used for writing files.) | |
--version |
false | display the version number for this tool | |
--write-gfas |
false | Write GFA representation of assemblies in fastq-dir. | |
Optional Common Arguments | |||
--add-output-vcf-command-line |
true | If true, adds a command line header line to created VCF files. | |
--create-output-bam-index -OBI |
true | If true, create a BAM index when writing a coordinate-sorted BAM file. | |
--create-output-bam-splitting-index |
true | If true, create a BAM splitting index (SBI) when writing a coordinate-sorted BAM file. | |
--create-output-variant-index -OVI |
true | If true, create a VCF index when writing a coordinate-sorted VCF file. | |
--disable-read-filter -DF |
Read filters to be disabled before analysis | ||
--disable-tool-default-read-filters |
false | Disable all tool default read filters (WARNING: many tools will not function correctly without their default read filters on) | |
--exclude-intervals -XL |
One or more genomic intervals to exclude from processing | ||
--gatk-config-file |
A configuration file to use with the GATK. | ||
--interval-exclusion-padding -ixp |
0 | Amount of padding (in bp) to add to each interval you are excluding. | |
--interval-padding -ip |
0 | Amount of padding (in bp) to add to each interval you are including. | |
--interval-set-rule -isr |
UNION | Set merging approach to use for combining interval inputs | |
--QUIET |
false | Whether to suppress job-summary info on System.err. | |
--read-filter -RF |
Read filters to be applied before analysis | ||
--read-index |
Indices to use for the read inputs. If specified, an index must be provided for every read input and in the same order as the read inputs. If this argument is not specified, the path to the index for each input will be inferred automatically. | ||
--read-validation-stringency -VS |
SILENT | Validation stringency for all SAM/BAM/CRAM/SRA files read by this program. The default stringency value SILENT can improve performance when processing a BAM file in which variable-length data (read, qualities, tags) do not otherwise need to be decoded. | |
--splitting-index-granularity |
4096 | Granularity to use when writing a splitting index, one entry will be put into the index every n reads where n is this granularity value. Smaller granularity results in a larger index with more available split points. | |
--tmp-dir |
Temp directory to use. | ||
--use-jdk-deflater -jdk-deflater |
false | Whether to use the JdkDeflater (as opposed to IntelDeflater) | |
--use-jdk-inflater -jdk-inflater |
false | Whether to use the JdkInflater (as opposed to IntelInflater) | |
--verbosity |
INFO | Control verbosity of logging. | |
Advanced Arguments | |||
--expand-assembly-graph |
true | Traverse assembly graph and produce contigs for all paths. | |
--pop-variant-bubbles |
false | Aggressively simplify local assemblies, ignoring small variants. | |
--remove-shadowed-contigs |
true | Simplify local assemblies by removing contigs shadowed by similar contigs. | |
--showHidden |
false | display hidden arguments | |
--z-dropoff |
20 | ZDropoff (see Bwa mem manual) for contig alignment. |
Argument details
Arguments in this list are specific to this tool. Keep in mind that other arguments are available that are shared with other tools (e.g. command-line GATK arguments); see Inherited arguments above.
--adapter-sequence
Adapter sequence.
String null
--add-output-vcf-command-line / -add-output-vcf-command-line
If true, adds a command line header line to created VCF files.
boolean true
--aligner-index-image
bwa-mem index image file
R String null
--allowed-short-fragment-overhang
Proper pairs have the positive strand read upstream of the negative strand read, but we allow this much slop for short fragments.
int 10 [ [ -∞ ∞ ] ]
--arguments_file
read one or more arguments files and add them to the command line
List[File] []
--assembled-contigs-output-order / -sort
sorting order to be used for the output assembly alignments SAM/BAM file (currently only coordinate or query name is supported)
The --assembled-contigs-output-order argument is an enumerated type (SortOrder), which can have one of the following values:
- unsorted
- queryname
- coordinate
- duplicate
- unknown
SortOrder coordinate
--assembly-to-mapped-size-ratio-guess
Guess at the ratio of reads in the final assembly to the number reads mapped to the interval.
int 7 [ [ -∞ ∞ ] ]
--bam-partition-size
maximum number of bytes to read from a file into each partition of reads. Setting this higher will result in fewer partitions. Note that this will not be equal to the size of the partition in memory. Defaults to 0, which uses the default split size (determined by the Hadoop input format, typically the size of one HDFS block).
long 0 [ [ -∞ ∞ ] ]
--breakpoint-evidence-dir
directory for evidence output
String null
--breakpoint-intervals
file for breakpoint intervals output
String null
--cleaner-max-copy-number
KmerCleaner maximum copy number (not count, but copy number) for a kmer. Kmers observed too frequently are probably mismapped or ubiquitous.
int 4 [ [ -∞ ∞ ] ]
--cleaner-max-intervals
KmerCleaner maximum number of intervals for a localizing kmer. If a kmer occurs in too many intervals, it isn't sufficiently local.
int 3 [ [ -∞ ∞ ] ]
--cleaner-min-kmer-count
KmerCleaner minimum kmer count for a localizing kmer. If we see it less often than this many times, we're guessing it's erroneous.
int 4 [ [ -∞ ∞ ] ]
--conf
Spark properties to set on the Spark context in the format =
List[String] []
--create-output-bam-index / -OBI
If true, create a BAM index when writing a coordinate-sorted BAM file.
boolean true
--create-output-bam-splitting-index
If true, create a BAM splitting index (SBI) when writing a coordinate-sorted BAM file.
boolean true
--create-output-variant-index / -OVI
If true, create a VCF index when writing a coordinate-sorted VCF file.
boolean true
--cross-contigs-to-ignore
file containing alt contig names that will be ignored when looking for inter-contig pairs
This is a path to a text file of contig names (one per line) that will be ignored when looking for inter-contig pairs.
String null
--disable-read-filter / -DF
Read filters to be disabled before analysis
List[String] []
--disable-sequence-dictionary-validation / -disable-sequence-dictionary-validation
If specified, do not check the sequence dictionaries from our inputs for compatibility. Use at your own risk!
boolean false
--disable-tool-default-read-filters / -disable-tool-default-read-filters
Disable all tool default read filters (WARNING: many tools will not function correctly without their default read filters on)
boolean false
--exclude-intervals / -XL
One or more genomic intervals to exclude from processing
Use this argument to exclude certain parts of the genome from the analysis (like -L, but the opposite).
This argument can be specified multiple times. You can use samtools-style intervals either explicitly on the
command line (e.g. -XL 1 or -XL 1:100-200) or by loading in a file containing a list of intervals
(e.g. -XL myFile.intervals).
List[String] []
--exclusion-interval-padding
Exclusion interval padding.
int 0 [ [ -∞ ∞ ] ]
--exclusion-intervals
file of reference intervals to exclude
This is a file that calls out the coordinates of intervals in the reference assembly to exclude from
consideration when calling putative breakpoints.
Each line is a tab-delimited interval with 1-based inclusive coordinates like this:
chr1 124535434 142535434
String null
--expand-assembly-graph
Traverse assembly graph and produce contigs for all paths.
boolean true
--external-evidence
external evidence input file
String null
--external-evidence-uncertainty
Uncertainty in location of external evidence.
int 150 [ [ -∞ ∞ ] ]
--external-evidence-weight
Weight to give external evidence.
int 10 [ [ -∞ ∞ ] ]
--fastq-dir
output dir for assembled fastqs
String null
--gatk-config-file
A configuration file to use with the GATK.
String null
--gcs-max-retries / -gcs-retries
If the GCS bucket channel errors out, how many times it will attempt to re-initiate the connection
int 20 [ [ -∞ ∞ ] ]
--gcs-project-for-requester-pays
Project to bill when accessing "requester pays" buckets. If unset, these buckets cannot be accessed. User must have storage.buckets.get permission on the bucket being accessed.
String ""
--help / -h
display the help message
boolean false
--high-coverage-intervals
file for high-coverage intervals output
String null
--high-depth-coverage-factor
We filter out contiguous regions of the genome that have coverage of at least high-depth-coverage-factor * avg-coverage and a peak coverage of high-depth-coverage-peak-factor * avg-coverage, because the reads mapped to those regions tend to be non-local and high depth prevents accurate assembly.
int 3 [ [ -∞ ∞ ] ]
--high-depth-coverage-peak-factor
We filter out contiguous regions of the genome that have coverage of at least high-depth-coverage-factor * avg-coverage and a peak coverage of high-depth-coverage-peak-factor * avg-coverage, because the reads mapped to those regions tend to be non-local and high depth prevents accurate assembly.
int 7 [ [ -∞ ∞ ] ]
--include-mapping-location
Include read mapping location in FASTQ files.
boolean true
--input / -I
BAM/SAM/CRAM file containing reads
R List[GATKPath] []
--interval-exclusion-padding / -ixp
Amount of padding (in bp) to add to each interval you are excluding.
Use this to add padding to the intervals specified using -XL. For example, '-XL 1:100' with a
padding value of 20 would turn into '-XL 1:80-120'. This is typically used to add padding around targets when
analyzing exomes.
int 0 [ [ -∞ ∞ ] ]
--interval-merging-rule / -imr
Interval merging rule for abutting intervals
By default, the program merges abutting intervals (i.e. intervals that are directly side-by-side but do not
actually overlap) into a single continuous interval. However you can change this behavior if you want them to be
treated as separate intervals instead.
The --interval-merging-rule argument is an enumerated type (IntervalMergingRule), which can have one of the following values:
- ALL
- OVERLAPPING_ONLY
IntervalMergingRule ALL
--interval-only-assembly
Don't look for extra reads mapped outside the interval.
boolean false
--interval-padding / -ip
Amount of padding (in bp) to add to each interval you are including.
Use this to add padding to the intervals specified using -L. For example, '-L 1:100' with a
padding value of 20 would turn into '-L 1:80-120'. This is typically used to add padding around targets when
analyzing exomes.
int 0 [ [ -∞ ∞ ] ]
--interval-set-rule / -isr
Set merging approach to use for combining interval inputs
By default, the program will take the UNION of all intervals specified using -L and/or -XL. However, you can
change this setting for -L, for example if you want to take the INTERSECTION of the sets instead. E.g. to
perform the analysis only on chromosome 1 exomes, you could specify -L exomes.intervals -L 1 --interval-set-rule
INTERSECTION. However, it is not possible to modify the merging approach for intervals passed using -XL (they will
always be merged using UNION).
Note that if you specify both -L and -XL, the -XL interval set will be subtracted from the -L interval set.
The --interval-set-rule argument is an enumerated type (IntervalSetRule), which can have one of the following values:
- UNION
- Take the union of all intervals
- INTERSECTION
- Take the intersection of intervals (the subset that overlaps all intervals specified)
IntervalSetRule UNION
--intervals / -L
One or more genomic intervals over which to operate
List[String] []
--k-size
Kmer size.
int 51 [ [ -∞ ∞ ] ]
--kmer-intervals
file for kmer intervals output
String null
--kmer-max-dust-score
Maximum kmer DUST score.
int 49 [ [ -∞ ∞ ] ]
--kmers-to-ignore
file containing ubiquitous kmer list. see FindBadGenomicKmersSpark to generate it.
This is a path to a file of kmers that appear too frequently in the reference to be usable as probes to localize
reads. We don't calculate it here, because it depends only on the reference.
The program FindBadGenomicKmersSpark can produce such a list for you.
R String null
--max-fastq-size
Maximum total bases in FASTQs that can be assembled.
int 3000000 [ [ -∞ ∞ ] ]
--max-tracked-fragment-length
Largest fragment size that will be explicitly counted in determining fragment size statistics.
int 2000 [ [ -∞ ∞ ] ]
--min-coherent-evidence-coverage-ratio
Minimum weight of the evidence that shares a distal target locus to validate the evidence, as a ratio of the mean coverage in the BAM. The default value is coherent-count / mean coverage ~ 7 / 42.9 ~ 0.163
double 0.1633408753260167 [ [ -∞ ∞ ] ]
--min-evidence-coverage-ratio
Minimum weight of the corroborating read evidence to validate some single piece of evidence, as a ratio of the mean coverage in the BAM. The default value is overlap-count / mean coverage ~ 15 / 42.9 ~ 0.350
double 0.35001616141289293 [ [ -∞ ∞ ] ]
--min-evidence-mapq
The minimum mapping quality for reads used to gather evidence of breakpoints.
int 20 [ [ -∞ ∞ ] ]
--min-evidence-match-length
The minimum length of the matched portion of an interesting alignment. Reads that don't match at least this many reference bases won't be used in gathering evidence.
int 45 [ [ -∞ ∞ ] ]
--min-kmers-per-interval
Minimum number of localizing kmers in a valid interval.
int 5 [ [ -∞ ∞ ] ]
--num-reducers
For tools that shuffle data or write an output, sets the number of reducers. Defaults to 0, which gives one partition per 10MB of input.
int 0 [ [ -∞ ∞ ] ]
--output / -O
HDFS path for output
R String null
--output-shard-tmp-dir
when writing a bam, in single sharded mode this directory to write the temporary intermediate output shards, if not specified .parts/ will be used
Exclusion: This argument cannot be used at the same time as sharded-output
.
String null
--pop-variant-bubbles
Aggressively simplify local assemblies, ignoring small variants.
boolean false
--program-name
Name of the program running
String null
--qname-intervals-for-assembly
file for mapped qname intervals output
String null
--qname-intervals-mapped
file for mapped qname intervals output
String null
--QUIET
Whether to suppress job-summary info on System.err.
Boolean false
--read-filter / -RF
Read filters to be applied before analysis
List[String] []
--read-index / -read-index
Indices to use for the read inputs. If specified, an index must be provided for every read input and in the same order as the read inputs. If this argument is not specified, the path to the index for each input will be inferred automatically.
List[GATKPath] []
--read-metadata
output file for read metadata
String null
--read-validation-stringency / -VS
Validation stringency for all SAM/BAM/CRAM/SRA files read by this program. The default stringency value SILENT can improve performance when processing a BAM file in which variable-length data (read, qualities, tags) do not otherwise need to be decoded.
The --read-validation-stringency argument is an enumerated type (ValidationStringency), which can have one of the following values:
- STRICT
- LENIENT
- SILENT
ValidationStringency SILENT
--reference / -R
Reference sequence
GATKPath null
--remove-shadowed-contigs
Simplify local assemblies by removing contigs shadowed by similar contigs.
boolean true
--run-without-gaps-annotation
Allow evidence filter to run without gaps annotation (assume no gaps).
boolean false
--run-without-umap-s100-annotation
Allow evidence filter to run without annotation for single-read mappability of 100-mers (assume all mappable).
boolean false
--sharded-output
For tools that write an output, write the output in multiple pieces (shards)
Exclusion: This argument cannot be used at the same time as output-shard-tmp-dir
.
boolean false
--showHidden / -showHidden
display hidden arguments
boolean false
--spark-master
URL of the Spark Master to submit jobs to when using the Spark pipeline runner.
String local[*]
--spark-verbosity
Spark verbosity. Overrides --verbosity for Spark-generated logs only. Possible values: {ALL, DEBUG, INFO, WARN, ERROR, FATAL, OFF, TRACE}
String null
--splitting-index-granularity
Granularity to use when writing a splitting index, one entry will be put into the index every n reads where n is this granularity value. Smaller granularity results in a larger index with more available split points.
long 4096 [ [ 1 ∞ ] ]
--sv-evidence-filter-model-file
Path to xgboost classifier model file for evidence filtering
String null
--sv-evidence-filter-threshold-probability
Minimum classified probability for a piece of evidence to pass xgboost evidence filter
double 0.92 [ [ -∞ ∞ ] ]
--sv-evidence-filter-type
Filter method for selecting evidence to group into Assembly Intervals
The --sv-evidence-filter-type argument is an enumerated type (SvEvidenceFilterType), which can have one of the following values:
- DENSITY
- XGBOOST
SvEvidenceFilterType DENSITY
--sv-genome-gaps-file
Path to file enumerating gaps in the reference genome, used by classifier to score evidence for filtering. To use classifier without specifying gaps file, pass the flag --run-without-gaps-annotation
String null
--sv-genome-umap-s100-file
Path to single read 100-mer mappability file in the reference genome, used by classifier to score evidence for filtering. To use classifier without specifying mappability file, pass the flag --run-without-umap-s100-annotation
String null
--target-link-file
output file for non-assembled breakpoints in bedpe format
String null
--tmp-dir
Temp directory to use.
GATKPath null
--unfiltered-breakpoint-evidence-dir
directory for evidence output
String null
--use-jdk-deflater / -jdk-deflater
Whether to use the JdkDeflater (as opposed to IntelDeflater)
boolean false
--use-jdk-inflater / -jdk-inflater
Whether to use the JdkInflater (as opposed to IntelInflater)
boolean false
--use-nio
Whether to use NIO or the Hadoop filesystem (default) for reading files. (Note that the Hadoop filesystem is always used for writing files.)
boolean false
--verbosity / -verbosity
Control verbosity of logging.
The --verbosity argument is an enumerated type (LogLevel), which can have one of the following values:
- ERROR
- WARNING
- INFO
- DEBUG
LogLevel INFO
--version
display the version number for this tool
boolean false
--write-gfas
Write GFA representation of assemblies in fastq-dir.
boolean false
--z-dropoff
ZDropoff (see Bwa mem manual) for contig alignment.
int 20 [ [ -∞ ∞ ] ]
GATK version 4.2.2.0-SNAPSHOT built at Thu, 19 Aug 2021 09:49:28 -0700.
0 comments
Please sign in to leave a comment.