I am trying to prepare my data for freebayes and markdup was the final step before merging. However, that failed with a segmentation fault.
Error
...
[bam_sort_core] merging from 248 files and 8 in-memory blocks...
finding positions of the duplicate reads in the file...
sorted 244119030 end pairs
and 947324 single ends (among them 0 unmatched pairs)
collecting indices of duplicate reads... done in 37330 ms
found 94344096 duplicates
collected list of positions in 9 min 16 sec
marking duplicates...
/var/spool/PBS/mom_priv/jobs/4252591.pbsserver.SC: line 27: 190448 Segmentation fault (core dumped) sambamba markdup -t 8 ch3hs1/aln.sorted.bam ch3hs1/aln.md.bam
I discussed this with the HPC team, here is there reply:
The markdup docs state you need loads of memory. Even using the latest version (0.7.1) and running on the highmem nodes with between 2Tb to 4Tb of mem I get variously
- a segmenation fault
- This message: sambamba-markdup: Memory allocation failed
- This message: sambamba-markdup: Read reference ID is out of range (4Tb) - but someones the Memory allocation failed as well.
The software clearly has a "bug" whereby it can't cope with failed memory allocations - perhaps thread related so it doesn't "print" or exit with the message before the seg fault.
You would have to ask the authors.
I am trying to prepare my data for freebayes and markdup was the final step before merging. However, that failed with a segmentation fault.
Error
I discussed this with the HPC team, here is there reply:
PBS Script