Genome Analysis Toolkit

Variant Discovery in High-Throughput Sequencing Data

GATK process banner

Need Help?

Search our documentation

Community Forum

Hi, How can we help?

Developed in the Data Sciences Platform at the Broad Institute, the toolkit offers a wide variety of tools with a primary focus on variant discovery and genotyping. Its powerful processing engine and high-performance computing features make it capable of taking on projects of any size. Learn more

gatk4.5 markduplicatesspark error spark

0

4 comments

  • Avatar
    Ryan Welch

    This is the command:

    gatk MarkDuplicatesSpark -I /data/BI_data/nbs/aligned_reads/SRR062634.paired.sam -O /data/BI_data/nbs/aligned_reads/SRR062634_sorted_dedup_reads.bam --java-options "-Djava.io.tmpdir=/data/BI_data/nbs/tmp"

    Jave specs:

    openjdk 17.0.9-internal 2023-10-17
    OpenJDK Runtime Environment (build 17.0.9-internal+0-adhoc..src)
    OpenJDK 64-Bit Server VM (build 17.0.9-internal+0-adhoc..src, mixed mode, sharing)

     

     

    0
    Comment actions Permalink
  • Avatar
    Gökalp Çelik

    Hi Ryan Welch

    Can you try running the tool with the below parameter?

    --java-options "-Dsamjdk.use_async_io_write_samtools=false"
    0
    Comment actions Permalink
  • Avatar
    Ryan Welch

    Hello Gökalp Çelik, I ran the follwong command:

    gatk MarkDuplicatesSpark -I /data/BI_data/nbs/aligned_reads/SRR062634.paired.sam -O /data/BI_data/nbs/aligned_reads/SRR062634_sorted_dedup_reads1.bam --java-options "-Dsamjdk.use_async_io_write_samtools=false"

     

    The error I get now is: 

    15:11:51.897 INFO  SparkContext - Created broadcast 14 from broadcast at BamSink.java:76
    15:11:51.918 INFO  PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory
    15:11:51.921 INFO  FileOutputCommitter - File Output Committer Algorithm version is 1
    15:11:51.921 INFO  FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false
    15:11:51.938 INFO  SparkContext - SparkContext is stopping with exitCode 0.
    15:11:51.943 INFO  AbstractConnector - Stopped Spark@1c3e6bf3{HTTP/1.1, (http/1.1)}{0.0.0.0:4040}
    15:11:51.948 INFO  SparkUI - Stopped Spark web UI at http://10.199.221.155:4040
    15:11:51.957 INFO  MapOutputTrackerMasterEndpoint - MapOutputTrackerMasterEndpoint stopped!
    15:11:52.946 INFO  MemoryStore - MemoryStore cleared
    15:11:52.947 INFO  BlockManager - BlockManager stopped
    15:11:52.948 INFO  BlockManagerMaster - BlockManagerMaster stopped
    15:11:52.949 INFO  OutputCommitCoordinator$OutputCommitCoordinatorEndpoint - OutputCommitCoordinator stopped!
    15:11:52.956 INFO  SparkContext - Successfully stopped SparkContext
    15:11:52.956 INFO  MarkDuplicatesSpark - Shutting down engine
    [June 25, 2024 at 3:11:52 PM CDT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 7.88 minutes.
    Runtime.totalMemory()=9412018176
    ***********************************************************************

    A USER ERROR has occurred: Couldn't write file /data/BI_data/nbs/aligned_reads/SRR062634_sorted_dedup_reads1.bam because writing failed with exception chmod: changing permissions of '/dat   a/BI_data/nbs/aligned_reads/SRR062634_sorted_dedup_reads1.bam.parts': Operation not permitted


    ***********************************************************************
    Set the system property GATK_STACKTRACE_ON_USER_EXCEPTION (--java-options '-DGATK_STACKTRACE_ON_USER_EXCEPTION=true') to print the stack trace.
    15:11:52.960 INFO  ShutdownHookManager - Shutdown hook called
    15:11:52.960 INFO  ShutdownHookManager - Deleting directory /tmp/spark-362bdcd6-9251-453e-bb2e-084872bdff0b
    (gatk4) [ryan.welch@dphccfpcdc319 ~]$ [June 25, 2024 at 3:11:52 PM CDT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 7.88 minutes.
    bash: [June: command not found...
    (gatk4) [ryan.welch@dphccfpcdc319 ~]$ Runtime.totalMemory()=9412018176
    -bash: syntax error near unexpected token `=9412018176'
    (gatk4) [ryan.welch@dphccfpcdc319 ~]$ ***********************************************************************
    bash: $: command not found...
    (gatk4) [ryan.welch@dphccfpcdc319 ~]$
    (gatk4) [ryan.welch@dphccfpcdc319 ~]$ A USER ERROR has occurred: Couldn't write file /data/BI_data/nbs/aligned_reads/SRR062634_sorted_dedup_reads1.bam because writing failed with exception chmod: changing permissions of '/dat   a/BI_data/nbs/aligned_reads/SRR062634_sorted_dedup_reads1.bam.parts': Operation not permitted

     

     

    0
    Comment actions Permalink
  • Avatar
    Gökalp Çelik

    Hi again.

    /data folder seems to be under root therefore it is possible that read write permissions are not set properly for that folder. Can you move your files to a local user folder and try again?

    0
    Comment actions Permalink

Please sign in to leave a comment.

Powered by Zendesk