Open ChaoXianSen opened 1 year ago
Hi @Puumanamana,
I'm getting the same error with the following command.
coconet run --fasta contigs.fasta --bam bam_files/*.bam --output coconet_results --threads 64
This is the output I get.
15:39:32 (Mem: 0.4 GB) <CoCoNet> INFO: Using 64 threads
15:39:32 (Mem: 0.4 GB) <CoCoNet> INFO: Features: coverage, composition
15:39:34 (Mem: 0.4 GB) <preprocessing> INFO: Processing 571,798 contigs
15:39:38 (Mem: 0.4 GB) <preprocessing> INFO: Length filter (L>2048 bp) -> 31,394 contigs remaining
15:42:22 (Mem: 0.4 GB) <preprocessing> INFO: Complete contigs filtering (DTR > 10 bp) -> 31207 contigs remaining
15:42:22 (Mem: 0.4 GB) <preprocessing> INFO: Processing alignments and converting to h5 format
15:45:10 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 1,000 contigs processed
15:46:22 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 2,000 contigs processed
15:47:07 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 3,000 contigs processed
15:47:43 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 4,000 contigs processed
15:48:13 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 5,000 contigs processed
15:48:39 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 6,000 contigs processed
15:49:03 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 7,000 contigs processed
15:49:25 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 8,000 contigs processed
15:49:44 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 9,000 contigs processed
15:50:03 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 10,000 contigs processed
15:50:19 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 11,000 contigs processed
15:50:35 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 12,000 contigs processed
15:50:50 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 13,000 contigs processed
15:51:03 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 14,000 contigs processed
15:51:17 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 15,000 contigs processed
15:51:28 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 16,000 contigs processed
15:51:40 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 17,000 contigs processed
15:51:51 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 18,000 contigs processed
15:52:02 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 19,000 contigs processed
15:52:13 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 20,000 contigs processed
15:52:26 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 21,000 contigs processed
15:52:36 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 22,000 contigs processed
15:52:45 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 23,000 contigs processed
15:52:54 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 24,000 contigs processed
15:53:03 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 25,000 contigs processed
15:53:11 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 26,000 contigs processed
15:53:19 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 27,000 contigs processed
15:53:27 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 28,000 contigs processed
15:53:35 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 29,000 contigs processed
15:53:43 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 30,000 contigs processed
15:53:50 (Mem: 0.8 GB) <preprocessing> DEBUG: Coverage: 31,000 contigs processed
15:53:55 (Mem: 0.7 GB) <preprocessing> INFO: Coverage filtering summary:
- 32,155,348 total reads
- 94.9% reads (mapped)
- 100.0% reads (primary alignments)
- 89.1% reads (mapq>30)
- 98.2% reads (coverage>50%)
- 89.2% reads (flag & 3596 == 0)
15:54:11 (Mem: 0.7 GB) <preprocessing> INFO: 31,394 contigs are only present in the composition. Taking the intersection (0 contigs)
15:54:11 (Mem: 0.7 GB) <preprocessing> INFO: Prevalence filter (prevalence>=2) -> 0 contigs remaining
15:54:11 (Mem: 0.7 GB) <learning> INFO: Making train/test examples
Traceback (most recent call last):
File "/home/mall0133/miniconda3/envs/coconet/bin/coconet", line 10, in <module>
sys.exit(main())
File "/home/mall0133/.local/lib/python3.7/site-packages/coconet/coconet.py", line 61, in main
make_train_test(cfg)
File "/home/mall0133/.local/lib/python3.7/site-packages/coconet/coconet.py", line 167, in make_train_test
assembly_idx = dict(test=np.random.choice(n_ctg, n_ctg_for_test))
File "mtrand.pyx", line 909, in numpy.random.mtrand.RandomState.choice
ValueError: a must be greater than 0 unless no samples are taken
I tried changing the parameter --min-ctg-len
to 1024, 512 and 256 but all still failed giving the same error.
Appreciate any help to fix this problem.
Thanks!
Hi !@Puumanamana run pipeline : coconet run \ --fasta R63.contigs.fa \ --bam R63_sort.bam \ --output coconet_binning_results \ --threads 16
err report:
00:50:09 (Mem: 0.2 GB) INFO: Using 16 threads
00:50:09 (Mem: 0.2 GB) INFO: Features: coverage, composition
00:50:11 (Mem: 0.2 GB) INFO: Processing 681,530 contigs
00:50:17 (Mem: 0.2 GB) INFO: Length filter (L>2048 bp) -> 43,057 contigs remaining
00:56:50 (Mem: 0.2 GB) INFO: Complete contigs filtering (DTR > 10 bp) -> 42993 contigs remaining
00:56:50 (Mem: 0.2 GB) INFO: Processing alignments and converting to h5 format
01:19:26 (Mem: 0.5 GB) INFO: Coverage filtering summary:
Any idea about this problem?
Thanks!