Closed sci-kai closed 1 month ago
I solved this error by giving the STAR processes more memory. It would be still good if the workflow produces an exit code so that the process restarts with the respective error strategy with more memory instead of failing like this.
I added the following lines to the run.config
file:
process {
// increase memory usage to avoid error.
withName: 'STAR_ALIGN' {
cpus = 16
memory = 120.GB
}
withName: 'STAR_GENOMEGENERATE' {
cpus = 16
memory = 120.GB
}
}
Now the processes use 16 instead of 12 cores and 120 GB instead of 70 GB memory.
Description of the bug
Hi, I am encountering the following bug running nf-core/rnaseq v3.15.1 on an offline HPC system with PBSPro:
I already investigated a bit and think the error is coming from the ALIGN_SATR:STAR_ALIGN process. It always crashes after this process. The
versions.yml
within the corresponding working directory of this process looks like this, probably causing this error message:The
.command.err
file gives the following message:While
.command.out
shows Successful finish (and.exitcode
is set to 0)This error only occurs with some samples, while other finish successfully on different nextflow runs. I assume based on this it might be a resource-related error, but not 100 % sure. If so, it would be very helpful if the process terminates with an appropriate exit code and is rerun with higher resources in the respective error strategy.
Command used and terminal output
Relevant files
nf-params.json run-rnaseq.config
System information
Nextflow version: 24.04.2 Hardware: HPC Cluster (offline, no internet connection on nodes) Executor: local, run on single node of HPC Cluster Container engine: singularity OS: Linux version nf-core/rnaseq: 3.15.1