soedinglab / MMseqs2

MMseqs2: ultra fast and sensitive search and clustering suite
https://mmseqs.com
MIT License
1.46k stars 197 forks source link

OOM Error #898

Open angelapredolac opened 4 weeks ago

angelapredolac commented 4 weeks ago

Expected Behavior

I tried using easy-clust to perform clustering to generate smaller representative subsets from Uniref50 (Uniref40, Uniref 30, Uniref 20, etc.)

Current Behavior

The clustering dies after a few steps with an Out of Memory Error.

Steps to Reproduce (for bugs)

Please make sure to execute the reproduction steps with newly recreated and empty tmp folders.

MMseqs Output (for bugs)

Please make sure to also post the complete output of MMseqs. You can use gist.github.com for large output. tmp/5521603387764364218/clu_tmp/3834389364932800683/cascaded_clustering.sh: line 102: 208445 Killed $RUNNER "$MMSEQS" prefilter "$INPUT" "$INPUT" "${TMP_PATH}/pref_step$STEP" ${TMP} slurmstepd: error: Detected 1 oom_kill event in StepId=16746209.batch. Some of the step tasks have been OOM Killed.

Context

Providing context helps us come up with a solution and improve our documentation for the future.

SBATCH --output=slurm_mlm.out

SBATCH -e slurm_mlm.err

SBATCH -p scavenger-gpu

SBATCH --gres=gpu:1

SBATCH --mem=200G

SBATCH -c 32

mmseqs easy-cluster /hpc/group/naderilab/eleanor/prose/data/uniref50.fasta uniref40 tmp --min-seq-id 0.4 -c 0.8 --cov-mode 1 --split-memory-limi t 160G --threads 12

Your Environment

Include as many relevant details about the environment you experienced the bug in.

milot-mirdita commented 1 day ago

The excessive memory consumption issue should be fixed in git mmseqs and will be part of release 16 that we are going to release in the next few days. It would be great if you could test if this crash still happens, you can use the precompiled binaries at https://mmseqs.com/latest/