Open mpiersonsmela opened 3 months ago
@mpiersonsmela
I've tested the pipeline and my nextflow report shows high RAM usage particularly by the deduplication step. I'm not sure if it's optimal but hope it helps
its true that it requires 72.GB mem as the process is labelled with process_high
with config set in base.config
.
I can limit the max mem for the test_full profile but, if any other changes you have to make as per your resource availability by setting institutional cluster specific config settings. Does that sound ok to you ?
Hi @mpiersonsmela, it’s true that the test_full
profile needs 72 GB of RAM since we’re testing real-life samples. However, the test
profile requires only 4 GB of RAM. So, if you’re just testing the pipeline setup, use the test
profile. If you want to test with a real-sized dataset, you can try test_full
, which does require high memory to process these samples.
If convinced with the answer, kindly close this issue. Thank you!
Description of the bug
On my university's cluster, users are penalized (with priority reduction) for requesting more RAM than they actually use. So the fact that the pipeline requires at least 72GB of RAM to run is an issue for me, given than I'm just trying to test it with the example samplesheet.csv from https://nf-co.re/methylseq/2.6.0/
This is the relevant portion of the output. Does bismark genome preparation really need so much RAM?
`ERROR ~ Error executing process > 'NFCORE_METHYLSEQ:METHYLSEQ:PREPARE_GENOME:BISMARK_GENOMEPREPARATION (BismarkIndex/grch38_core+bs_controls.fa)'
Caused by: Process requirement exceeds available memory -- req: 72 GB; avail: 32 GB
Command executed:
bismark_genome_preparation \ --bowtie2 \ BismarkIndex`
Command used and terminal output
Relevant files
No response
System information
No response