Open drewbailey opened 4 years ago
Hm, this is still happening on v1.5.17 apparently:
Task Group: "mygroup" (18446744073709551615 create, 1 create/destroy update, 123 in-place update)
Thanks for the verification @andrey-mazo but just want to point out that 1.5.x is out of support at this point.
Thanks for the verification @andrey-mazo but just want to point out that 1.5.x is out of support at this point.
Yeah, I totally understand. I may be able to retest on a newer version as soon as we upgrade, but don't have a particular timeline for that.
And to be fair, I'm not really concerned about the number itself, but more about what Nomad is actually going to do when running such a job.
Hm, this is still happening on v1.5.17 apparently:
Task Group: "mygroup" (18446744073709551615 create, 1 create/destroy update, 123 in-place update)
This happens not just when updating constraints on a job/group, but also on changing environment vars, templates, etc.
Interestingly, doing edit+plan from Nomad UI shows a slightly different number:
Task Group: "mygroup" ( 1 create/destroy update 18446744073709552000 create 123 in-place update )
And the issue is not job-specific -- changing constraints on another system job (which is normally placed on the same nodes), for example, shows the same 18446744073709551615 create
thing.
This happens not just when updating constraints on a job/group, but also on changing environment vars, templates, etc.
There was a placement failure,
and now after resolving it, only changes to the constraints trigger the 18446744073709551615 create
issue -- changing anything else results in a normal 123 create/destroy update
.
And not even all constraints -- for example, artificially bumping attr.vault.version
to a slightly newer version results in a simple "in-place update". (probably because this doesn't really affect the job placement decision in this case)
To continue my little story here.
I drained one of the nodes -- and it magically resolved the issue. This was a node which would have stopped being eligible for the job with the updated constraints.
So, I suspect that 18446744073709551615 create
really meant to say 1 destroy
.
Nomad version
Output from
nomad version
Nomad v0.10.3 (65af1b9ecff5b55a1dd6e10b8c3224f896d6c9fa)
Operating system and Environment details
ubuntu 19.10
Issue
nomad job plan repro.hcl
for a system job diaplays max uint64 (likely negative number being computed)Reproduction steps
running a nomad dev cluster with different node classes, run a system job without constraints, plan job with constraints.
nomad job run simplerepro.hcl
then
nomad job plan repro.hcl
Job file (if appropriate)
simplerepro.hcl
repro.hcl
Nomad logs (if appropriate)
If possible please post relevant logs in the issue.