Hi,
I am using canu on two different PacBio ccs datasets. One is assembled without issues, but the second terminates at Unitiging stage. Both unitiging attempts failed. Removing 4-unitigger/ and re-running canu also fails.
-- Detected Java(TM) Runtime Environment '1.8.0_91' (from 'java').
-- Detected gnuplot version '4.6 patchlevel 6' (from 'gnuplot') and image format 'png'.
-- Detected 8 CPUs and 12 gigabytes of memory.
-- No grid engine detected, grid disabled.
-- Allowed to run 2 jobs concurrently, and use up to 4 compute threads and 6 GB memory for stage 'bogart (unitigger)'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 6 GB memory for stage 'mhap (overlapper)'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 6 GB memory for stage 'mhap (overlapper)'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 6 GB memory for stage 'mhap (overlapper)'.
-- Allowed to run 2 jobs concurrently, and use up to 4 compute threads and 2 GB memory for stage 'read error detection (overlap error adjustment)'.
-- Allowed to run 8 jobs concurrently, and use up to 1 compute thread and 1 GB memory for stage 'overlap error adjustment'.
-- Allowed to run 2 jobs concurrently, and use up to 4 compute threads and 12 GB memory for stage 'utgcns (consensus)'.
-- Allowed to run 8 jobs concurrently, and use up to 1 compute thread and 2 GB memory for stage 'overlap store parallel bucketizer'.
-- Allowed to run 8 jobs concurrently, and use up to 1 compute thread and 8 GB memory for stage 'overlap store parallel sorting'.
-- Allowed to run 8 jobs concurrently, and use up to 1 compute thread and 2 GB memory for stage 'overlapper'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 8 GB memory for stage 'overlapper'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 8 GB memory for stage 'overlapper'.
-- Allowed to run 2 jobs concurrently, and use up to 4 compute threads and 6 GB memory for stage 'meryl (k-mer counting)'.
-- Allowed to run 4 jobs concurrently, and use up to 2 compute threads and 6 GB memory for stage 'falcon_sense (read correction)'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 6 GB memory for stage 'minimap (overlapper)'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 6 GB memory for stage 'minimap (overlapper)'.
-- Allowed to run 1 job concurrently, and use up to 8 compute threads and 6 GB memory for stage 'minimap (overlapper)'.
-- This is canu parallel iteration #1, out of a maximum of 2 attempts.
-- low-coverage 1067 6.61 2141.30 +- 941.69 4.29 +- 1.93 (easy to assemble, potential for lower quality consensus)
-- unique 2893 17.92 2044.14 +- 742.22 21.63 +- 8.89 (easy to assemble, perfect, yay)
-- repeat-cont 5176 32.05 2333.75 +- 757.25 1447.01 +- 1290.48 (potential for consensus errors, no impact on assembly)
-- repeat-dove 11 0.07 3532.36 +- 842.29 716.70 +- 1014.36 (hard to assemble, likely won't assemble correctly or even at all)
-- span-repeat 941 5.83 2698.68 +- 1157.89 1437.14 +- 1217.33 (read spans a large repeat, usually easy to assemble)
-- uniq-repeat-cont 3861 23.91 2242.83 +- 690.59 (should be uniquely placed, low potential for consensus errors, no impact on assembly)
-- uniq-repeat-dove 796 4.93 3001.63 +- 645.58 (will end contigs, potential to misassemble)
-- uniq-anchor 115 0.71 3846.08 +- 991.92 591.47 +- 849.43 (repeat read, with unique section, probable bad read)
Configure RED for 2gb memory with batches of at most 5000 reads and (unlimited) bases.
RED job 1 from read 1 to read 5000 - 0.576 GB for 5000 reads - 0.143 GB for 11795454 olaps - 0.038 GB for evidence
RED job 2 from read 5001 to read 10000 - 0.571 GB for 5000 reads - 0.141 GB for 11642554 olaps - 0.039 GB for evidence
RED job 3 from read 10001 to read 15000 - 0.563 GB for 5000 reads - 0.139 GB for 11495875 olaps - 0.039 GB for evidence
RED job 4 from read 15001 to read 16148 - 0.135 GB for 1148 reads - 0.033 GB for 2740694 olaps - 0.009 GB for evidence
-- read error detection attempt 1 begins with 0 finished, and 4 to compute.
-- Starting concurrent execution on Mon Oct 10 19:53:10 2016 with 468.902 GB free disk space (4 processes; 2 concurrently)
ReadInfo()-- Using 16148 reads, no minimum read length used.
OverlapCache()-- limited to 6144MB memory (user supplied).
PHYS_PAGES = 3051251
PAGE_SIZE = 4096
MEMORY = 12497924096
OverlapCache()-- 0MB for read data.
OverlapCache()-- 0MB for best edges.
OverlapCache()-- 0MB for unitig layouts.
OverlapCache()-- 0MB for tigs.
OverlapCache()-- 0MB for id maps.
OverlapCache()-- 0MB for error profiles.
OverlapCache()-- 0MB for overlap cache pointers.
OverlapCache()-- 56MB for overlap cache initial bucket.
OverlapCache()-- 64MB for overlap cache thread data.
OverlapCache()-- 0MB for number of overlaps per read.
OverlapCache()-- 0MB for other processes.
OverlapCache()-- ---------
OverlapCache()-- 121MB for data structures (sum of above).
OverlapCache()-- 6022MB available for overlaps.
OverlapCache()-- Loading number of overlaps per read.
OverlapCache()-- Initial guess at _maxPer=24443 (max of 4822) from (memLimit=6315325907 - memUsed=0) / (numReads=16148 * sizeof(OVL)=16)
OverlapCache()-- _maxPer= 24443 (numBelow=16148 numEqual=0 numAbove=0 totalLoad=11198720 -- 0 + 11198720 = 179179520 <? 6315325907
BestOverlapGraph()-- allocating best edges (0MB)
BestOverlapGraph()-- analyzing 16148 reads for best contains, with 4 threads.
BestOverlapGraph()-- analyzing 16148 reads for best edges, with 4 threads.
BestOverlapGraph()-- removing suspicious reads from graph, with 4 threads.
BestOverlapGraph()-- marked 378 reads as suspicious.
BestOverlapGraph()-- analyzing 16148 reads for best contains, with 4 threads.
BestOverlapGraph()-- analyzing 16148 reads for best edges, with 4 threads.
BestOverlapGraph()-- analyzing best edges to find useful edge error rate
BestOverlapGraph()-- analyzing 16148 reads for best contains, with 4 threads.
BestOverlapGraph()-- analyzing 16148 reads for best edges, with 4 threads.
BestOverlapGraph()-- removing suspicious edges from graph, with 4 threads.
BestOverlapGraph()-- analyzing 16148 reads for best contains, with 4 threads.
BestOverlapGraph()-- analyzing 16148 reads for best edges, with 4 threads.
BestOverlapGraph()-- detecting spur reads.
BestOverlapGraph()-- detected 1014 spur reads.
BestOverlapGraph()-- analyzing 16148 reads for best contains, with 4 threads.
BestOverlapGraph()-- analyzing 16148 reads for best edges, with 4 threads.
BestOverlapGraph()-- removing best edges for contained reads.
==> BUILDING GREEDY TIGS.
breakSingletonTigs()-- Removed 998 singleton tigs; reads are now unplaced.
==> PLACE CONTAINED READS.
computeErrorProfiles()-- Computing error profiles for 1180 tigs, with 4 threads.
placeContains()-- placing 13059 contained and 1672 unplaced reads, with 4 threads.
placeContains()-- Placed 8465 contained reads and 111 unplaced reads.
placeContains()-- Failed to place 4594 contained reads (too high error suspected) and 1561 unplaced reads (lack of overlaps suspected).
==> MERGE ORPHANS.
computeErrorProfiles()-- Computing error profiles for 1180 tigs, with 4 threads.
Hi, I am using canu on two different PacBio ccs datasets. One is assembled without issues, but the second terminates at Unitiging stage. Both unitiging attempts failed. Removing 4-unitigger/ and re-running canu also fails.
Checked https://github.com/marbl/canu/issues/139 and https://github.com/marbl/canu/issues/154 but commits look diferent in my system.
Any suggestion?
Please see below => canu version => Run details => unitigger.err (in 4-unitigger)
=> canu version
Canu v1.3 (+218 commits) r7834 0049b3bfc0dbe2f1f6d2dd4d96763ad9a4826202.
=> Run details
$ canu -p 9332_mogene_B01 -d 9332_mogene_B01/ genomeSize=2m errorRate=0.01 trimReadsCoverage=2 -pacbio-corrected 9332_mogene_B01.ccs.3pass-Q10.fastq
=> unitigger.err (in 4-unitigger)