Closed runningman84 closed 1 week ago
I have also raised this in aws/karpenter#2544
This is an interesting feature request. I think this would fit nicely with out recent work on involuntary interruption handling. @jonathan-innis
Interesting proposal
The Kubernetes project currently lacks enough contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle stale
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
AKS Also uses node problem detector to detect events and fire auto-healing actions on the nodes. We use these scheduled events to cordon and drain nodes via node auto drain as part of the node auto healing framework. This is the reason we didn't implement an interupt controller in the aks provider because we already have a component that will respond to interurption events.
See:
/remove-lifecycle stale
Since Draino is deprecated; It make total sense for Karpenter to react to Problem API.
The Kubernetes project currently lacks enough contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle stale
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle rotten
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle rotten
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.
This bot triages issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/reopen
/remove-lifecycle rotten
Please send feedback to sig-contributor-experience at kubernetes/community.
/close not-planned
@k8s-triage-robot: Closing this issue, marking it as "Not Planned".
Tell us about your request
We used to run node problem detector in combination with drainio and cluster-autoscaler. Once a problem was detected drainio/cluster-autoscaler removed the node.
Dranio is not really maintained anymore and cluster-autoscaler is replaced by karpenter in our usecase.
It would be great if karpenter would also look for node conditions like
and drain the corresponding node. Maybe this list could be configured in the karpenter setup...
Tell us about the problem you're trying to solve. What are you trying to do, and why is it hard?
We would like to reduce complexity, because karpenter manages nodes anyway it seems to make sense to also include this feature.
Are you currently working around this issue?
Trying to switch to the DataDog drainio fork.
Additional Context
No response
Attachments
No response
Community Note