Closed Nathhub closed 1 year ago
Hi Nathan,
it could be a problem limiting the maximum number of cores. Could you run sed -i -e 's/getThreads(80)/getThreads(64)/g' Snakefile
in the binny dir and try again?
Best
Oskar
Hi Oskar,
Thanks for your answer and suggestion.
I have tried changing getThreads(80) to getThreads(64) as you suggested but the job still gets stopped at job 9 with the same error... Any other ideas on how to solve this problem?
Cheers
Hi Oskar,
I have tried again to run binny but this time reducing the number of threads to 40, using the command: ./binny -c -r -t 40 config/config.default.yaml and this time it worked.
Thanks
Hi there, I am struggling to identify what is the problem with the configuration of binny. I hope you can help.
The steps seem to be going well until step 9 (see log report below):
The command I use is : ./binny -c -r -t 64 config/config.default.yaml
I am using an HPC cluster having 64 CPUs per nodes and 4G per CPU. The HPC only allows me to send jobs of max 24 hours.
Below you will find the content of some config files.
Thanks in advance for your help.
Cheers, Nathan
---------------------------------------------------------------------- snakemake log -------------------
Config file /home/ndh1n17/.conda/envs/mamba/envs/binny/binny/config/config.default.yaml is extended by additional config specified via the command line. Building DAG of jobs... Using shell: /usr/bin/bash Provided cluster nodes: 64 Job stats: job count min threads max threads
ALL 1 1 1 annotate 1 24 24 binny 1 64 64 call_contig_depth 3 1 1 format_assembly 1 1 1 mantis_checkm_marker_sets 1 64 64 merge_contig_depths 1 1 1 prepare_input_data 1 1 1 total 10 1 64
Select jobs to execute...
[Mon Feb 20 17:30:58 2023] Job 5: Preparing input. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.fa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000001_sorted.bam, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000002_sorted.bam, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000003_sorted.bam
Config file /home/ndh1n17/.conda/envs/mamba/envs/binny/binny/config/config.default.yaml is extended by additional config specified via the command line. Building DAG of jobs... Using shell: /usr/bin/bash Provided cores: 1 (use --cores to define parallelism) Rules claiming more threads will be scaled down. Provided resources: mem_mb=197589, disk_mb=197589 Select jobs to execute... [Mon Feb 20 17:33:54 2023] Finished job 5. 1 of 10 steps (10%) done Select jobs to execute...
[Mon Feb 20 17:33:54 2023] Job 4: Preparing assembly. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.fa
Submitted job 4 with external jobid 'Submitted batch job 229278'. [Mon Feb 20 17:34:08 2023] Finished job 4. 2 of 10 steps (20%) done Select jobs to execute...
[Mon Feb 20 17:34:08 2023] Job 8: annotate: Running prokkaP. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/prokka.faa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/annotation.filt.gff; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa
Submitted job 8 with external jobid 'Submitted batch job 229279'.
[Mon Feb 20 17:34:08 2023] Job 3: /dssgfs01/scratch/ndh1n17/08_bins/aa/call_contig_depth: Getting data on assembly coverage with mg reads. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_fsw.txt; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000001_sorted.bam
Submitted job 3 with external jobid 'Submitted batch job 229280'. [Mon Feb 20 17:34:08 2023] Job 8: annotate: Running prokkaP. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/prokka.faa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/annotation.filt.gff; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa
Submitted job 8 with external jobid 'Submitted batch job 229279'.
[Mon Feb 20 17:34:08 2023] Job 3: /dssgfs01/scratch/ndh1n17/08_bins/aa/call_contig_depth: Getting data on assembly coverage with mg reads. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_fsw.txt; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000001_sorted.bam
Submitted job 3 with external jobid 'Submitted batch job 229280'.
[Mon Feb 20 17:34:08 2023] Job 6: /dssgfs01/scratch/ndh1n17/08_bins/aa/call_contig_depth: Getting data on assembly coverage with mg reads. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_go.txt; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000002_sorted.bam
Submitted job 6 with external jobid 'Submitted batch job 229281'.
[Mon Feb 20 17:34:08 2023] Job 7: /dssgfs01/scratch/ndh1n17/08_bins/aa/call_contig_depth: Getting data on assembly coverage with mg reads. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_ra.txt; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.formatted.fa, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/reads_sample_000003_sorted.bam
Submitted job 7 with external jobid 'Submitted batch job 229282'. [Mon Feb 20 17:58:40 2023] Finished job 7. 3 of 10 steps (30%) done [Mon Feb 20 18:15:01 2023] Finished job 3. 4 of 10 steps (40%) done [Mon Feb 20 18:24:21 2023] Finished job 6. 5 of 10 steps (50%) done Select jobs to execute...
[Mon Feb 20 18:24:21 2023] Job 2: Merging depth files. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly.contig_depth.txt; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_fsw.txt, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_ra.txt, /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/assembly_contig_depth_aa_go.txt
Submitted job 2 with external jobid 'Submitted batch job 229285'. [Mon Feb 20 18:24:51 2023] Finished job 2. 6 of 10 steps (60%) done [Mon Feb 20 19:40:26 2023] Finished job 8. 7 of 10 steps (70%) done Select jobs to execute...
[Mon Feb 20 19:40:26 2023] Job 9: MANTIS: Running MANTIS with CheckM marker sets. Reason: Missing output files: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/mantis_out/consensus_annotation.tsv; Input files updated by another job: /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/prokka.faa
sbatch: error: Batch job submission failed: Requested node configuration is not available Error submitting jobscript (exit code 1):
Exiting because a job execution failed. Look above for error message Complete log: ../../.conda/envs/mamba/envs/binny/binny/.snakemake/log/2023-02-20T173045.483493.snakemake.log
SNAKEMAKE_EXTRA_ARGUMENTS LOADING_MODULES SUBMIT_COMMAND sbatch --time=24:00:00 --mem-per-cpu=4G --output=slurm-%j.out BIND_JOBS_TO_MAIN false NODENAME_VAR SLURMD_NODENAME SCHEDULER slurm_simple MAX_THREADS 64
call: "sbatch " nodes: "" partition: "" qos: "" mem_per_cpu: "--mem-per-cpu " runtime: "-t " threads: "-c " stdout: "-o binny.{rule}.{wildcards}.stdout"
bash: cannot set terminal process group (-1): Inappropriate ioctl for device bash: no job control in this shell Config file /home/ndh1n17/.conda/envs/mamba/envs/binny/binny/config/config.default.yaml is extended by additional config specified via the command line. Building DAG of jobs... Creating report... Missing metadata for file /dssgfs01/scratch/ndh1n17/08_bins/aa/binny.done. Maybe metadata was deleted or it was created using an older version of Snakemake. This is a non critical warning. Missing metadata for file /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/mantis_out/output_annotation.tsv. Maybe metadata was deleted or it was created using an older version of Snakemake. This is a non critical warning. Missing metadata for file /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/mantis_out/integrated_annotation.tsv. Maybe metadata was deleted or it was created using an older version of Snakemake. This is a non critical warning. Missing metadata for file /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/mantis_out/consensus_annotation.tsv. Maybe metadata was deleted or it was created using an older version of Snakemake. This is a non critical warning. Missing metadata for file /dssgfs01/scratch/ndh1n17/08_bins/aa/intermediary/mantis_out. Maybe metadata was deleted or it was created using an older version of Snakemake. This is a non critical warning. Downloading resources and rendering HTML. Report created: report.html. slurm-229277.out (END)
Mask potentially disruptive contig regions (e.g. rRNA and CRISPR elements) from k-mer counting
mask_disruptive_sequences: 'True'
Extract single contig MAGs of at least 90% purity and 92.5% completeness
extract_scmags: 'True'
Will use coassembly mode, starting with contigs >= 500 bp instead of high threshold, decreasing, if set to 'on' or
if 'auto' and multiple depth files are detected
Choose between: 'auto', 'on' , 'off''
coassembly_mode: 'auto'
Binny prefilters assemblies based on N value to try and take as much information as possible into account,
while minimizing the amount of noise. Be aware that, depending on the assembly quality, low values as the N
might results in leaving out a large proportion of the assembly (if the max_cont_length cutoffs are set high as well).
NX_value: 90
Minimum and maximum contig length. Caps value from NX filtering.
min_cont_length_cutoff: 2250 # 500 max_cont_length_cutoff: 2250 # 1000
Minimum and maximum length of contigs containing CheckM markers. Caps value from NX filtering (initial value is NX_value / 2).
min_cont_length_cutoff_marker: 2250 # 250 max_cont_length_cutoff_marker: 2250 # 400
Maximum number of contigs binny uses. If the number of available
contigs after minimum size filtering exceeds this, binny will
increase the minimum size threshold until the maximum is reached.
Prevents use of excessive amounts of memory on large assemblies.
Default should ensure adequate performance, adjust e.g. according
to available memory.
max_n_contigs: 5.0e5 # 3.5e5
Maximum marker set lineage depth to check bin quality with:
0: 'domain', 1: 'phylum', 2: 'class', 3: 'order', 4: 'family', 5: 'genus', 6: 'species'
max_marker_lineage_depth_lvl: 2
Distance metric for opentSNE and HDBSCAN.
distance_metric: 'manhattan' embedding:
Maximum number of binny iterations.
max_iterations: 50 clustering:
Increasing the HDBSCAN cluster selection epsilon beyond 0.5
is not advised as it might massively increase run time, but it might
help recover fragmented genomes that would be missed with lower settings.
hdbscan_epsilon_range: '0.250,0.000'
Adapted from the HDBSCAN manual: 'Measure of how conservative the
clustering should be. With larger values, more points will be declared
as noise, and clusters will be restricted to progressively more dense areas.'.
hdbscan_min_samples_range: '1,5,10' # '1,2,4,6,8'
Use depth as additional dimension during the initial clustering.
include_depth_initial: 'False'
Use depth as additional dimension during the main clusterings.
include_depth_main: 'False' bin_quality:
Minimum value binny will lower completeness to while running.
min_completeness: 72.5
Completeness threshold binny wilt begin with.
start_completeness: 92.5
Minimum purity for bins to be selected.