Open fanhaouu opened 4 weeks ago
Hi @fanhaouu. Thanks for your PR.
I'm waiting for a kubernetes-sigs member to verify that this patch is reasonable to test. If it is, they should reply with /ok-to-test
on its own line. Until that is done, I will not automatically test new commits in this PR, but the usual testing commands by org members will still work. Regular contributors should join the org to skip this step.
Once the patch is verified, the new status will be reflected by the ok-to-test
label.
I understand the commands that are listed here.
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by: Once this PR has been reviewed and has the lgtm label, please assign a7i for approval. For more information see the Kubernetes Code Review Process.
The full list of commands accepted by this bot can be found here.
@a7i ,master, can you help me review this pr?
Hi @fanhaouu great contribution. Going to copy some of the maintainers for feedback as well: /cc @jklaw90 @ingvagabund @damemi
My feedback:
regarding point 1: For GOMAXPROCS
, as far as I know, we don't run any goroutines in Descheduler. How does this help?
regarding point 2: The idea is to present all NodeFit predicates. The predicate check is not in any particular order, so returning the first one may not present the whole picture to the cluster operator. In a cluster of 35k pods / 15k deployments, a Descheduler run takes a few (single digit) seconds. Given that the shortest frequency can be a minute, I'm not convinced this optimization is worth it. What do you think?
regarding point 3: I like it, that's a great change!
overall: if you could split this into 3 PRs, I think it would make it easier to provide feedback and request changes.
Hi @fanhaouu great contribution. Going to copy some of the maintainers for feedback as well: /cc @jklaw90 @ingvagabund @damemi
My feedback:
- regarding point 1: For
GOMAXPROCS
, as far as I know, we don't run any goroutines in Descheduler. How does this help?- regarding point 2: The idea is to present all NodeFit predicates. The predicate check is not in any particular order, so returning the first one may not present the whole picture to the cluster operator. In a cluster of 35k pods / 15k deployments, a Descheduler run takes a few (single digit) seconds. Given that the shortest frequency can be a minute, I'm not convinced this optimization is worth it. What do you think?
- regarding point 3: I like it, that's a great change!
- overall: if you could split this into 3 PRs, I think it would make it easier to provide feedback and request changes.
Hi @fanhaouu great contribution. Going to copy some of the maintainers for feedback as well: /cc @jklaw90 @ingvagabund @damemi
My feedback:
- regarding point 1: For
GOMAXPROCS
, as far as I know, we don't run any goroutines in Descheduler. How does this help?- regarding point 2: The idea is to present all NodeFit predicates. The predicate check is not in any particular order, so returning the first one may not present the whole picture to the cluster operator. In a cluster of 35k pods / 15k deployments, a Descheduler run takes a few (single digit) seconds. Given that the shortest frequency can be a minute, I'm not convinced this optimization is worth it. What do you think?
- regarding point 3: I like it, that's a great change!
- overall: if you could split this into 3 PRs, I think it would make it easier to provide feedback and request changes.
master, thank you for your reply.
point1. At present, descheduler has some goroutine, but the number is not very large, so gomaxproc can be added or not, but the current go runtime cannot recognize the container environment, I think it is better to add it. For example, jvm runtime has long supported container environments; (As descheduler in our company has a lot of goroutine, the performance can be improved significantly after adding this restriction, so I didn't remove this submission from the pr)
point2. Because the default loop time is 5min, this time can be adjusted. The larger the cluster, the longer the nodeFit time, so the full loop time can easily exceed 5min. What if the user sets the loop time to 1 minute? Comparing one by one I think is very, very unnecessary, too time-consuming, we should be the same as the filter policy in the scheduler.
In short, if we can make the program better, faster, without affecting any functionality, I think it makes a lot of sense.
We are also developing a descheduler cache mechanism internally. Currently, descheduler pulls weight data for filtering at policy runtime, which can actually cache a good part of the data in advance, just like the cache in the scheduler. If our company is stable online, I will also contribute to our descheduler community. We can review it together then.
As Amir mentioned would you please break the PR into three separate PRs? Some of the suggested changes deserves a dedicated discussion. Wrt. NodeFit I am in the process of composing a KEP: https://github.com/kubernetes-sigs/descheduler/issues/1421. This sounds like a good use case to include in the proposal.
hi, masters, i have split this into 3 PRs, looking forward to your feedback:
/cc @a7i @jklaw90 @ingvagabund @damemi
PR needs rebase.
This PR aims to address the following three issues: