Closed treffynnon closed 2 months ago
Hello! Thank you for filing an issue.
The maintainers will triage your issue shortly.
In the meantime, please take a look at the troubleshooting guide for bug reports.
If this is a feature request, please review our contribution guidelines.
Hey @treffynnon,
This is working as intended. We count failed runners in to signal some problem on the cluster, and to stop creating resources indefinitely.
We are planning to change this in the future, so I will close this issue in favour of https://github.com/actions/actions-runner-controller/issues/2721
Checks
Controller Version
0.9.0
Deployment Method
Helm
Checks
To Reproduce
Describe the bug
When there are three pods with a status of Evicted the controller ceases to provision new runner pods for jobs. This appears to be related, but not confirmed, to the values.yaml having a minScaling of 2. The jobs begin to queue up and new pods are never created. As soon as one of the Evicted pods is deleted using kubectl the jobs start running again and new pods are provisioned.
We're running our cluster on Azure AKS with both Linux and Windows nodes - this issue has only affected the Linux node/pods so far.
Describe the expected behavior
It should not consider Evicted pods to be the same as a running pod. It should provision new pods despite the presence of Evicted pods.
In an ideal world Evicted pods would be deleted after a period of time too so that they are cleaned up too.
Additional Context
This is from after I had deleted one of the Evicted pods, but you can see here that the runners are sitting around for hours.
Controller Logs
Runner Pod Logs