Open theaidenlab opened 4 years ago
Sorry about the delay. Not sure if you're still interested in this, but if 1.6B reads worked maybe just leave it at that, I would say. Some of the heuristics were never update to use large coverages on larger genomes, so anything higher than 100x coverage is pushing it too far in these scenarios as there's a few hardcoded thresholds.
I can't really think top of my mind what could be going wrong but I would bet some threshold is letting way too many elements join a collection that then overflows. If you really need this looked at, I would say run it on GDB and let us know where the overflow is happening. Otherwise, I will just close this in a few days.
Hi, Using a relatively large data set (~3B reads) we are experiencing some problem in Step 5 using the "step7_fix" branch. (Usually this branch is reliable and still gives the best result comparing to the latest master branch) The machine that was in use had 2TB RAM but w2rap-contigger peaked only around 1.2TB before it crashed with "Segmentation fault". These are the last couple of lines before the crash:
We tried to lower the amount of reads to 2.3B reads but had similar result, only when we used around 1.6B reads we managed to pass step 5. In case it helps I've saved the "core dumps". We would really appreciate some help finding a solution to this problem and/or guidance how to debug it further. Thanks,