Closed Sumsarium closed 11 months ago
Hi
Thanks for your attention of SemiBin.
Do you mind sharing the contig id in cannot.txt and contig file? It seems they are inconsistent.
Another way is that maybe you can try:
conda activate SemiBin SemiBin \ single_easy_bin \ -i $metagenome \ --input-bam $temp/$assemblyName/*.bam \ --output $temp/$assemblyName/semibin \ --sequencing-type=long_read --self-supervised
Sincerely Shaojun
@Sumsarium Did you find a solution? Or could you perhaps share some of your inputs with Shaojun so we could check ourselves?
@luispedro @psj1997 Thanks and sorry for the overdue reply - I plan to look into it later this week. I will let you know as soon as I have some info / files.
Trieded the other approach:
SemiBin \ single_easy_bin \ -i $metagenome \ --input-bam $temp/$assemblyName/*.bam \ --output $temp/$assemblyName/semibin \ --sequencing-type=long_read --self-supervised
but got this error:
Traceback (most recent call last):
File "/home/user/miniconda3/envs/SemiBin/bin/SemiBin", line 10, in <module>
sys.exit(main1())
File "/home/user/miniconda3/envs/SemiBin/lib/python3.9/site-packages/SemiBin/main.py", line 1482, in main1
main2(args, is_semibin2=False)
File "/home/user/miniconda3/envs/SemiBin/lib/python3.9/site-packages/SemiBin/main.py", line 1455, in main2
single_easy_binning(
File "/home/user/miniconda3/envs/SemiBin/lib/python3.9/site-packages/SemiBin/main.py", line 1159, in single_easy_binning
training(logger, None,
File "/home/user/miniconda3/envs/SemiBin/lib/python3.9/site-packages/SemiBin/main.py", line 1007, in training
model = train_self(output,
File "/home/user/miniconda3/envs/SemiBin/lib/python3.9/site-packages/SemiBin/self_supervised_model.py", line 87, in train_self
indices1 = np.random.choice(data_length, size=n_samples)
File "mtrand.pyx", line 928, in numpy.random.mtrand.RandomState.choice
ValueError: a must be greater than 0 unless no samples are taken
Different error though. The question is whether the problem is the underlying data (small metagenome with 6 bam files) or the env. I will attempt to set up a singularity-based workflow and test on a different dataset.
Hi, can you check the data.csv file in the output_dir? It seems it is a empty file.
Sincerely Shaojun
We just released version 2, which includes a more error checking and better diagnostics; so I am closing this here, but please feel free to re-open if it still relevant.
Hi. I keep getting the same KeyError (same error on three different HPCs).
My cmd:
The error:
Any thoughts?