Closed damemi closed 1 year ago
/sig testing (not sure which sig this falls under, since it's about the automated CI jobs I assumed sig-testing)
It seems clear that we should parallelize these builds, which we tried in https://github.com/kubernetes-sigs/descheduler/pull/1019 by making each arch its own gcb build stage. However, this didn't seem to have any effect.
GCB stages run in serial unless you set waitFor:
https://cloud.google.com/build/docs/configuring-builds/configure-build-step-order
GCB stages run in serial unless you set waitFor:
Hi @BenTheElder as noted in the Issue, we tried this and it didn't reduce the duration https://github.com/kubernetes-sigs/descheduler/pull/1019
GCB also has different machine sizes, but it takes time to spin up custom machine sizes.
I recommend doing multiple architectures in parallel using buildx and pushing a single multi-arch image directly, FWIW, but that may not improve cold build times.
The Kubernetes project currently lacks enough contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle stale
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle rotten
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle rotten
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.
This bot triages issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/reopen
/remove-lifecycle rotten
Please send feedback to sig-contributor-experience at kubernetes/community.
/close not-planned
@k8s-triage-robot: Closing this issue, marking it as "Not Planned".
Hi, we have an image push postsubmit job for Descheduler at https://github.com/kubernetes/test-infra/blob/master/config/jobs/image-pushing/k8s-staging-descheduler.yaml
This job builds images for multiple arches, and over the past year or so we've noticed an increasing number of timeouts leading to failed image builds. The timeouts don't seem to be due to any issue other than simply taking a long time to build each image.
We've tried increasing the timeout a few times but even at 30 minutes we're still getting failures (the latest increase comes after bumping our k8s dependencies, maybe related).
It seems clear that we should parallelize these builds, which we tried in https://github.com/kubernetes-sigs/descheduler/pull/1019 by making each arch its own gcb build stage. However, this didn't seem to have any effect.
Is there a recommended way to split up these builds? Any docs or examples would be helpful, thanks
cc @a7i @ingvagabund @knelasevero