kubernetes / autoscaler

Autoscaling components for Kubernetes
Apache License 2.0
8.1k stars 3.98k forks source link

is it possible 502 raise sometimes, during node evict/drain. #6793

Closed johnzheng1975 closed 3 weeks ago

johnzheng1975 commented 6 months ago

Which component are you using?: autoscaler

What version of the component are you using?: cluster-autoscaler:v1.29.0

Component version:

What k8s version are you using (kubectl version)?:

kubectl version Output
$ kubectl version

What environment is this in?:

aws alb, eks

What did you expect to happen?: no 502

What happened instead?: 502 raise sometimes.

How to reproduce it (as minimally and precisely as possible):

Anything else we need to know?:

is it possible 502 raise sometimes, during node evict/drain. After you drain node successfully, this ec2 is not k8 node anymore, however, then api will still be forwarded from aws alb until the node is terminated. 502 may raise since the api cannot be forwarded to k8s service. Is this possible during this time period? Thanks.

Shubham82 commented 5 months ago

/area provider/aws /area cluster-autoscaler

k8s-triage-robot commented 2 months ago

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

You can:

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

k8s-triage-robot commented 1 month ago

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

You can:

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle rotten

k8s-triage-robot commented 3 weeks ago

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues according to the following rules:

You can:

Please send feedback to sig-contributor-experience at kubernetes/community.

/close not-planned

k8s-ci-robot commented 3 weeks ago

@k8s-triage-robot: Closing this issue, marking it as "Not Planned".

In response to [this](https://github.com/kubernetes/autoscaler/issues/6793#issuecomment-2452132211): >The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. > >This bot triages issues according to the following rules: >- After 90d of inactivity, `lifecycle/stale` is applied >- After 30d of inactivity since `lifecycle/stale` was applied, `lifecycle/rotten` is applied >- After 30d of inactivity since `lifecycle/rotten` was applied, the issue is closed > >You can: >- Reopen this issue with `/reopen` >- Mark this issue as fresh with `/remove-lifecycle rotten` >- Offer to help out with [Issue Triage][1] > >Please send feedback to sig-contributor-experience at [kubernetes/community](https://github.com/kubernetes/community). > >/close not-planned > >[1]: https://www.kubernetes.dev/docs/guide/issue-triage/ Instructions for interacting with me using PR comments are available [here](https://git.k8s.io/community/contributors/guide/pull-requests.md). If you have questions or suggestions related to my behavior, please file an issue against the [kubernetes-sigs/prow](https://github.com/kubernetes-sigs/prow/issues/new?title=Prow%20issue:) repository.