Closed js-timbirkett closed 2 years ago
I wonder if this would help with #84 🤔
After looking at this again, I need to understand the behaviours of scaling in various cases first. I don't think it's as simple as the changes that have been made in this PR... I will look more closely at this 🔜
Any update on this thread? We could use this extra parameter as well.
Closing due to lack of time.
Hello 👋 - A few weeks ago I opened #96 but promptly closed it as I could have solved the problem with a different tool. After looking back at this, I think it'd be simpler if solved in
eks-rolling-update
directly.This PR adds a new env variable:
ASG_BUFFER_INSTANCES
which allows an arbitrary number to be given toeks_rolling_update.py
and will cause each ASG to be over-scaled by that number.But why?
The past few rolling upgrades I've done have resulted in some things like workloads with PV/PVC getting stuck in pending as other pods had started, scaleout of HPAs causing pods to get stuck in pending, deployments during rollout causing issues...
Since I've been pre-scaling each ASG by a few instances it hasn't been an issue and
cluster-autoscaler
takes care of scaling in unused compute after rollout.As always, open to any feedback or ideas 😸
Thanks for an awesome tool!