Genome Analysis Toolkit

Variant Discovery in High-Throughput Sequencing Data

GATK process banner

Need Help?

Search our documentation

Community Forum

Hi, How can we help?

Developed in the Data Sciences Platform at the Broad Institute, the toolkit offers a wide variety of tools with a primary focus on variant discovery and genotyping. Its powerful processing engine and high-performance computing features make it capable of taking on projects of any size. Learn more

FilterLongReadAlignmentsSAMSpark (EXPERIMENTAL) Follow

  • Metagenomics
  • PathSeqBuildKmers
  • PathSeqBuildReferenceTaxonomy
  • PathSeqBwaSpark
  • PathSeqFilterSpark
  • PathSeqPipelineSpark
  • PathSeqScoreSpark
  • Other
  • CreateHadoopBamSplittingIndex
  • FifoBuffer (Picard)
  • FixCallSetSampleOrdering
  • GatherBQSRReports
  • GatherTranches
  • IndexFeatureFile
  • ParallelCopyGCSDirectoryIntoHDFSSpark
  • Read Data Manipulation
  • AddCommentsToBam (Picard)
  • AddOrReplaceReadGroups (Picard)
  • ApplyBQSR
  • ApplyBQSRSpark
  • BQSRPipelineSpark
  • BamToBfq (Picard)
  • BaseRecalibrator
  • BaseRecalibratorSpark
  • BaseRecalibratorSparkSharded
  • BuildBamIndex (Picard)
  • BwaAndMarkDuplicatesPipelineSpark
  • BwaSpark
  • CleanSam (Picard)
  • ClipReads
  • ConvertHeaderlessHadoopBamShardToBam
  • DownsampleSam (Picard)
  • ExtractOriginalAlignmentRecordsByNameSpark
  • FastqToSam (Picard)
  • FilterSamReads (Picard)
  • FixMateInformation (Picard)
  • FixMisencodedBaseQualityReads
  • GatherBamFiles (Picard)
  • LeftAlignIndels
  • MarkDuplicates (Picard)
  • MarkDuplicatesGATK
  • MarkDuplicatesSpark
  • MarkDuplicatesWithMateCigar (Picard)
  • MergeBamAlignment (Picard)
  • MergeSamFiles (Picard)
  • PositionBasedDownsampleSam (Picard)
  • PrintReads
  • PrintReadsSpark
  • ReorderSam (Picard)
  • ReplaceSamHeader (Picard)
  • RevertBaseQualityScores
  • RevertOriginalBaseQualitiesAndAddMateCigar (Picard)
  • RevertSam (Picard)
  • SamFormatConverter (Picard)
  • SamToFastq (Picard)
  • SetNmAndUqTags (Picard)
  • SetNmMdAndUqTags (Picard)
  • SimpleMarkDuplicatesWithMateCigar (Picard)
  • SortSam (Picard)
  • SortSamSpark
  • SplitNCigarReads
  • SplitReads
  • SplitSamByLibrary (Picard)
  • SplitSamByNumberOfReads (Picard)
  • UmiAwareMarkDuplicatesWithMateCigar (Picard)
  • UnmarkDuplicates
  • Reference
  • BaitDesigner (Picard)
  • BwaMemIndexImageCreator
  • CreateSequenceDictionary (Picard)
  • ExtractSequences (Picard)
  • FindBadGenomicKmersSpark
  • NonNFastaSize (Picard)
  • NormalizeFasta (Picard)
  • ScatterIntervalsByNs (Picard)
  • Short Variant Discovery
  • CombineGVCFs
  • GenomicsDBImport
  • GenotypeGVCFs
  • HaplotypeCaller
  • HaplotypeCallerSpark
  • Mutect2
  • ReadsPipelineSpark
  • Structural Variant Discovery
  • DiscoverVariantsFromContigAlignmentsSAMSpark
  • ExtractSVEvidenceSpark
  • FindBreakpointEvidenceSpark
  • StructuralVariationDiscoveryPipelineSpark
  • SvDiscoverFromLocalAssemblyContigAlignmentsSpark
  • Variant Evaluation and Refinement
  • AnnotatePairOrientation
  • AnnotateVcfWithBamDepth
  • AnnotateVcfWithExpectedAlleleFraction
  • CalculateGenotypePosteriors
  • CalculateMixingFractions
  • Concordance
  • CountFalsePositives
  • CountVariants
  • CountVariantsSpark
  • FindMendelianViolations (Picard)
  • Funcotator
  • GenotypeConcordance (Picard)
  • ValidateBasicSomaticShortMutations
  • ValidateVariants
  • VariantsToTable
  • Variant Filtering
  • ApplyVQSR
  • CreateSomaticPanelOfNormals
  • FilterByOrientationBias
  • FilterMutectCalls
  • FilterVcf (Picard)
  • VariantFiltration
  • VariantRecalibrator
  • Variant Manipulation
  • FixVcfHeader (Picard)
  • GatherVcfs (Picard)
  • GatherVcfsCloud
  • LiftoverVcf (Picard)
  • MakeSitesOnlyVcf (Picard)
  • MergeVcfs (Picard)
  • PrintVariantsSpark
  • RemoveNearbyIndels
  • RenameSampleInVcf (Picard)
  • SelectVariants
  • SortVcf (Picard)
  • SplitVcfs (Picard)
  • UpdateVCFSequenceDictionary
  • UpdateVcfSequenceDictionary (Picard)
  • VcfFormatConverter (Picard)
  • VcfToIntervalList (Picard)

  • Base Calling
  • CheckIlluminaDirectory (Picard)
  • CollectIlluminaBasecallingMetrics (Picard)
  • CollectIlluminaLaneMetrics (Picard)
  • ExtractIlluminaBarcodes (Picard)
  • IlluminaBasecallsToFastq (Picard)
  • IlluminaBasecallsToSam (Picard)
  • MarkIlluminaAdapters (Picard)

  • Read Filters
  • AlignmentAgreesWithHeaderReadFilter
  • AllowAllReadsReadFilter
  • AmbiguousBaseReadFilter
  • CigarContainsNoNOperator
  • FirstOfPairReadFilter
  • FragmentLengthReadFilter
  • GoodCigarReadFilter
  • HasReadGroupReadFilter
  • LibraryReadFilter
  • MappedReadFilter
  • MappingQualityAvailableReadFilter
  • MappingQualityNotZeroReadFilter
  • MappingQualityReadFilter
  • MatchingBasesAndQualsReadFilter
  • MateDifferentStrandReadFilter
  • MateOnSameContigOrNoMappedMateReadFilter
  • MetricsReadFilter
  • NonZeroFragmentLengthReadFilter
  • NonZeroReferenceLengthAlignmentReadFilter
  • NotDuplicateReadFilter
  • NotSecondaryAlignmentReadFilter
  • NotSupplementaryAlignmentReadFilter
  • OverclippedReadFilter
  • PairedReadFilter
  • PassesVendorQualityCheckReadFilter
  • PlatformReadFilter
  • PlatformUnitReadFilter
  • PrimaryLineReadFilter
  • ProperlyPairedReadFilter
  • ReadGroupBlackListReadFilter
  • ReadGroupReadFilter
  • ReadLengthEqualsCigarLengthReadFilter
  • ReadLengthReadFilter
  • ReadNameReadFilter
  • ReadStrandFilter
  • SampleReadFilter
  • SecondOfPairReadFilter
  • SeqIsStoredReadFilter
  • ValidAlignmentEndReadFilter
  • ValidAlignmentStartReadFilter
  • WellformedReadFilter
  • Variant Annotations
  • AS_BaseQualityRankSumTest
  • AS_FisherStrand
  • AS_InbreedingCoeff
  • AS_MappingQualityRankSumTest
  • AS_QualByDepth
  • AS_RMSMappingQuality
  • AS_ReadPosRankSumTest
  • AS_StrandOddsRatio
  • BaseQuality
  • BaseQualityRankSumTest
  • ChromosomeCounts
  • ClippingRankSumTest
  • Coverage
  • DepthPerAlleleBySample
  • DepthPerSampleHC
  • ExcessHet
  • FisherStrand
  • FragmentLength
  • GenotypeSummaries
  • InbreedingCoeff
  • LikelihoodRankSumTest
  • MappingQuality
  • MappingQualityRankSumTest
  • MappingQualityZero
  • OxoGReadCounts
  • PossibleDeNovo
  • QualByDepth
  • RMSMappingQuality
  • ReadPosRankSumTest
  • ReadPosition
  • ReferenceBases
  • SampleList
  • StrandArtifact
  • StrandBiasBySample
  • StrandOddsRatio
  • TandemRepeat
  • UniqueAltReadCount
  • Examines alignments of chimeric contigs, attempting to produce an optimal tiling

    Category Diagnostics and Quality Control


    Overview

    (Internal) Examines alignments of chimeric contigs, attempting to produce an optimal tiling

    This tool is used in development and should not be of interest to most researchers. It is a prototype of one aspect of structural variant calling from chimeric contigs produced by local assemblies.

    It takes a SAM/BAM/CRAM containing the alignments of assembled contigs and filters them with the aim of providing "optimal coverage" of the contig, based on an heuristic scoring scheme.

    It saves the result as a text file, formatted as:

    (CONTIG_NAME, [[LIST_OF_PICKED_ALIGNMENT_INTERVALS_AS_A_COMPACT_STRING]])

    Inputs

    Output

    Usage example

       gatk FilterLongReadAlignmentsSAMSpark \
         -I assemblies.sam \
         -O selected_alignments.txt
     

    This tool can be run without explicitly specifying Spark options. That is to say, the given example command without Spark options will run locally. See Tutorial#10060 for an example of how to set up and run a Spark tool on a cloud Spark cluster.

    FilterLongReadAlignmentsSAMSpark specific arguments

    This table summarizes the command-line arguments that are specific to this tool. For more details on each argument, see the list further down below the table or click on an argument name to jump directly to that entry in the list.

    Argument name(s) Default value Summary
    Required Arguments
    --input
     -I
    [] BAM/SAM/CRAM file containing reads
    --output
     -O
    null prefix for output text file for filtered alignment intervals
    Optional Tool Arguments
    --arguments_file
    [] read one or more arguments files and add them to the command line
    --bam-partition-size
    0 maximum number of bytes to read from a file into each partition of reads. Setting this higher will result in fewer partitions. Note that this will not be equal to the size of the partition in memory. Defaults to 0, which uses the default split size (determined by the Hadoop input format, typically the size of one HDFS block).
    --conf
    [] spark properties to set on the spark context in the format =
    --disable-sequence-dictionary-validation
    false If specified, do not check the sequence dictionaries from our inputs for compatibility. Use at your own risk!
    --gcs-max-retries
     -gcs-retries
    20 If the GCS bucket channel errors out, how many times it will attempt to re-initiate the connection
    --help
     -h
    false display the help message
    --interval-merging-rule
     -imr
    ALL Interval merging rule for abutting intervals
    --intervals
     -L
    [] One or more genomic intervals over which to operate
    --non-canonical-contig-names-file
     -alt-tigs
    null file containing non-canonical contig names (e.g chrUn_KI270588v1) in the reference, human reference assumed when omitted
    --num-reducers
    0 For tools that shuffle data or write an output, sets the number of reducers. Defaults to 0, which gives one partition per 10MB of input.
    --old-filtering-too
     -OT
    false whether to run old way of filtering or not
    --program-name
    null Name of the program running
    --reference
     -R
    null Reference sequence
    --sharded-output
    false For tools that write an output, write the output in multiple pieces (shards)
    --spark-master
    local[*] URL of the Spark Master to submit jobs to when using the Spark pipeline runner.
    --version
    false display the version number for this tool
    Optional Common Arguments
    --disable-read-filter
     -DF
    [] Read filters to be disabled before analysis
    --disable-tool-default-read-filters
    false Disable all tool default read filters
    --exclude-intervals
     -XL
    [] One or more genomic intervals to exclude from processing
    --gatk-config-file
    null A configuration file to use with the GATK.
    --interval-exclusion-padding
     -ixp
    0 Amount of padding (in bp) to add to each interval you are excluding.
    --interval-padding
     -ip
    0 Amount of padding (in bp) to add to each interval you are including.
    --interval-set-rule
     -isr
    UNION Set merging approach to use for combining interval inputs
    --QUIET
    false Whether to suppress job-summary info on System.err.
    --read-filter
     -RF
    [] Read filters to be applied before analysis
    --read-index
    [] Indices to use for the read inputs. If specified, an index must be provided for every read input and in the same order as the read inputs. If this argument is not specified, the path to the index for each input will be inferred automatically.
    --read-validation-stringency
     -VS
    SILENT Validation stringency for all SAM/BAM/CRAM/SRA files read by this program. The default stringency value SILENT can improve performance when processing a BAM file in which variable-length data (read, qualities, tags) do not otherwise need to be decoded.
    --TMP_DIR
    [] Undocumented option
    --use-jdk-deflater
     -jdk-deflater
    false Whether to use the JdkDeflater (as opposed to IntelDeflater)
    --use-jdk-inflater
     -jdk-inflater
    false Whether to use the JdkInflater (as opposed to IntelInflater)
    --verbosity
    INFO Control verbosity of logging.
    Advanced Arguments
    --config-score-diff-tolerance
     -cst
    0.0 Maximum difference between alignment configuration scores for which two configurations will be considered equally likely
    --showHidden
    false display hidden arguments

    Argument details

    Arguments in this list are specific to this tool. Keep in mind that other arguments are available that are shared with other tools (e.g. command-line GATK arguments); see Inherited arguments above.


    --arguments_file / NA

    read one or more arguments files and add them to the command line

    List[File]  []


    --bam-partition-size / NA

    maximum number of bytes to read from a file into each partition of reads. Setting this higher will result in fewer partitions. Note that this will not be equal to the size of the partition in memory. Defaults to 0, which uses the default split size (determined by the Hadoop input format, typically the size of one HDFS block).

    long  0  [ [ -∞  ∞ ] ]


    --conf / -conf

    spark properties to set on the spark context in the format =

    List[String]  []


    --config-score-diff-tolerance / -cst

    Maximum difference between alignment configuration scores for which two configurations will be considered equally likely

    Double  0.0  [ [ -∞  ∞ ] ]


    --disable-read-filter / -DF

    Read filters to be disabled before analysis

    List[String]  []


    --disable-sequence-dictionary-validation / -disable-sequence-dictionary-validation

    If specified, do not check the sequence dictionaries from our inputs for compatibility. Use at your own risk!

    boolean  false


    --disable-tool-default-read-filters / -disable-tool-default-read-filters

    Disable all tool default read filters

    boolean  false


    --exclude-intervals / -XL

    One or more genomic intervals to exclude from processing
    Use this argument to exclude certain parts of the genome from the analysis (like -L, but the opposite). This argument can be specified multiple times. You can use samtools-style intervals either explicitly on the command line (e.g. -XL 1 or -XL 1:100-200) or by loading in a file containing a list of intervals (e.g. -XL myFile.intervals).

    List[String]  []


    --gatk-config-file / NA

    A configuration file to use with the GATK.

    String  null


    --gcs-max-retries / -gcs-retries

    If the GCS bucket channel errors out, how many times it will attempt to re-initiate the connection

    int  20  [ [ -∞  ∞ ] ]


    --help / -h

    display the help message

    boolean  false


    --input / -I

    BAM/SAM/CRAM file containing reads

    R List[String]  []


    --interval-exclusion-padding / -ixp

    Amount of padding (in bp) to add to each interval you are excluding.
    Use this to add padding to the intervals specified using -XL. For example, '-XL 1:100' with a padding value of 20 would turn into '-XL 1:80-120'. This is typically used to add padding around targets when analyzing exomes.

    int  0  [ [ -∞  ∞ ] ]


    --interval-merging-rule / -imr

    Interval merging rule for abutting intervals
    By default, the program merges abutting intervals (i.e. intervals that are directly side-by-side but do not actually overlap) into a single continuous interval. However you can change this behavior if you want them to be treated as separate intervals instead.

    The --interval-merging-rule argument is an enumerated type (IntervalMergingRule), which can have one of the following values:

    ALL
    OVERLAPPING_ONLY

    IntervalMergingRule  ALL


    --interval-padding / -ip

    Amount of padding (in bp) to add to each interval you are including.
    Use this to add padding to the intervals specified using -L. For example, '-L 1:100' with a padding value of 20 would turn into '-L 1:80-120'. This is typically used to add padding around targets when analyzing exomes.

    int  0  [ [ -∞  ∞ ] ]


    --interval-set-rule / -isr

    Set merging approach to use for combining interval inputs
    By default, the program will take the UNION of all intervals specified using -L and/or -XL. However, you can change this setting for -L, for example if you want to take the INTERSECTION of the sets instead. E.g. to perform the analysis only on chromosome 1 exomes, you could specify -L exomes.intervals -L 1 --interval-set-rule INTERSECTION. However, it is not possible to modify the merging approach for intervals passed using -XL (they will always be merged using UNION). Note that if you specify both -L and -XL, the -XL interval set will be subtracted from the -L interval set.

    The --interval-set-rule argument is an enumerated type (IntervalSetRule), which can have one of the following values:

    UNION
    Take the union of all intervals
    INTERSECTION
    Take the intersection of intervals (the subset that overlaps all intervals specified)

    IntervalSetRule  UNION


    --intervals / -L

    One or more genomic intervals over which to operate

    List[String]  []


    --non-canonical-contig-names-file / -alt-tigs

    file containing non-canonical contig names (e.g chrUn_KI270588v1) in the reference, human reference assumed when omitted

    String  null


    --num-reducers / NA

    For tools that shuffle data or write an output, sets the number of reducers. Defaults to 0, which gives one partition per 10MB of input.

    int  0  [ [ -∞  ∞ ] ]


    --old-filtering-too / -OT

    whether to run old way of filtering or not

    boolean  false


    --output / -O

    prefix for output text file for filtered alignment intervals

    R String  null


    --program-name / NA

    Name of the program running

    String  null


    --QUIET / NA

    Whether to suppress job-summary info on System.err.

    Boolean  false


    --read-filter / -RF

    Read filters to be applied before analysis

    List[String]  []


    --read-index / -read-index

    Indices to use for the read inputs. If specified, an index must be provided for every read input and in the same order as the read inputs. If this argument is not specified, the path to the index for each input will be inferred automatically.

    List[String]  []


    --read-validation-stringency / -VS

    Validation stringency for all SAM/BAM/CRAM/SRA files read by this program. The default stringency value SILENT can improve performance when processing a BAM file in which variable-length data (read, qualities, tags) do not otherwise need to be decoded.

    The --read-validation-stringency argument is an enumerated type (ValidationStringency), which can have one of the following values:

    STRICT
    LENIENT
    SILENT

    ValidationStringency  SILENT


    --reference / -R

    Reference sequence

    String  null


    --sharded-output / NA

    For tools that write an output, write the output in multiple pieces (shards)

    boolean  false


    --showHidden / -showHidden

    display hidden arguments

    boolean  false


    --spark-master / NA

    URL of the Spark Master to submit jobs to when using the Spark pipeline runner.

    String  local[*]


    --TMP_DIR / NA

    Undocumented option

    List[File]  []


    --use-jdk-deflater / -jdk-deflater

    Whether to use the JdkDeflater (as opposed to IntelDeflater)

    boolean  false


    --use-jdk-inflater / -jdk-inflater

    Whether to use the JdkInflater (as opposed to IntelInflater)

    boolean  false


    --verbosity / -verbosity

    Control verbosity of logging.

    The --verbosity argument is an enumerated type (LogLevel), which can have one of the following values:

    ERROR
    WARNING
    INFO
    DEBUG

    LogLevel  INFO


    --version / NA

    display the version number for this tool

    boolean  false


    Return to top


    GATK version 4.0.2.0 built at 25-48-2019 02:48:40.

    0 comments

    Please sign in to leave a comment.

    Powered by Zendesk