Open jprnz opened 2 weeks ago
Thanks for reporting: I will look into it. Thankfully, you provided the solution along with the report. Meanwhile, I think the unsetting of SLURM
variables will not help at all. According to SchedMD's documentation, the built-in variables are always exported. The solution must be in setting job parameters explicitly, always. Could you test code during a few iterations? I'm afraid, I might only get to it on Thursday or Friday, though.
NB:
Our admins prefer ...
Yes, I got this nonsense a lot. As if it hurts anyone, when someone produces a plot within a few seconds on a login node. (Or runs a workflow manager, which consumes about as much CPU power during the run of a workflow.)
Two things prevent the use of this plugin on our cluster:
In order to make many of the common SLURM tools work, users of our cluster need to have
SLURM_CONFIG
set in their environment. Since all environmental variables prefixed with 'SLURM_*' are wiped if the plugin seesSLURM_JOB_ID
, this results insacctmgr
andsinfo
exiting with an error:This seems like an unintended consequence and could be easily fixed by not removing
SLURM_CONFIG
. The issue can be avoided by running:Personally, I think it would be nice to set the values for
slurm_account
/slurm_partition
via env vars (assrun
/sbatch
do) and, to me this seems like a sensible way to determine a default value.Thanks for your work and continuing to help the community!