Closed krzyzacy closed 1 week ago
We should do this for all presubmits :-)
/assign
ping
we moved the blocking ones to this at least, except scalability I think :|
/assign @amwat (who is working on scalability presubmits :-) )
/milestone v1.13
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta. /lifecycle stale
/milestone v1.14
What remains to be done here?
I think scalability presubmits are the only ones left.
cc @amwat @krzysied
Were looking into the scalability jobs. There should be a github issue for this somewhere.
@shyamjvs https://github.com/kubernetes/test-infra/issues/9223 Most of the small scalability jobs were moved to boskos in https://github.com/kubernetes/test-infra/pull/9567
Only the large scalability tests remain. @krzyzacy IIRC the last consensus on this one was to have a boskos pool of one project?
I didn't see the presubmit jobs... maybe it's ok they fails when there's a periodic job running?
/remove-lifecycle stale
will take a stab
/unshrug /kind cleanup /sig testing
@spiffxp: ¯\_(ツ)_/¯
/milestone v1.15 How far along are we on this?
eh, never got to this one, will poke next week
Looks like even the small presubmit jobs are not running on boskos yet (https://github.com/kubernetes/test-infra/blob/master/config/jobs/kubernetes/sig-scalability/sig-scalability-presubmit-jobs.yaml#L29)
cc @amwat do we have enough resource to move all presubmit scalability jobs there?
We only moved the small periodic jobs, is it safe to move the presubmits? I anticipate we'll run out of free projects pretty easily. How do we run multiple presubmits in a fixed project right now?
@amwat I think that's by different cluster names etc and resource leak check was disabled :man_shrugging:
/shrug /unassign /lifecycle frozen
let's see how sig-scalability wanna do this after move the jobs to their own build cluster
/milestone v1.16 What remains to be done here?
any update?
Just guessing based on the project names there are probably still a few remaining
$ ag '\-\-gcp-project'
config/jobs/containerd/cri/containerd-cri-presubmit-jobs.yaml
56: --gcp-project=cri-c8d-pr-node-e2e
100: --gcp-project=cri-c8d-pr-node-e2e
144: --gcp-project=cri-c8d-pr-node-e2e
188: --gcp-project=cri-c8d-pr-node-e2e
config/jobs/kubernetes/sig-scalability/sig-scalability-presubmit-jobs.yaml
85: - --gcp-project=k8s-presubmit-scale
135: - --gcp-project=k8s-presubmit-scale
262: - --gcp-project=k8s-presubmit-scale
config/jobs/kubernetes/sig-cloud-provider/gcp/gpu/gcp-gpu-presubmits.yaml
46: - --gcp-project=k8s-jkns-pr-gce-gpus
88: - --gcp-project=k8s-gke-gpu-pr
config/jobs/kubernetes/sig-node/sig-node-presubmit.yaml
29: - --gcp-project=k8s-jkns-pr-node-e2e
104: - --gcp-project=k8s-c8d-pr-node-e2e
138: - --gcp-project=k8s-jkns-pr-node-e2e
174: - --gcp-project=k8s-jkns-pr-node-e2e
208: - --gcp-project=k8s-jkns-pr-node-e2e
I can help, but I need some guidance. I'm tracing through some recent PRs: from https://github.com/kubernetes/test-infra/pull/17349 it looks we use a different arg, gcp-project to gcp-project-type from https://github.com/kubernetes/test-infra/pull/17359 I wonder if I need to find an entry to slot tests under or if I need to make a new one some how. from https://github.com/kubernetes/test-infra/pull/17452#discussion_r418136073 it looks like maybe the idea is to move the pull job into the same project-type as the ci job, which I think we have for some of these. from https://github.com/kubernetes/kubernetes/issues/89847#issuecomment-621432309 some thoughts are linking up in my head
I may have an idea of what to do. I'm gonna circle back later and maybe push some PRs.
/remove-lifecycle frozen
Jobs that pin to a gcp-project
config/jobs/containerd/containerd/containerd-presubmit-jobs.yaml
63: --gcp-project=cri-c8d-pr-node-e2e
config/jobs/containerd/cri/containerd-cri-presubmit-jobs.yaml
56: --gcp-project=cri-c8d-pr-node-e2e
config/jobs/kubernetes/generated/generated.yaml
26: - --gcp-project=ubuntu-image-validation
64: - --gcp-project=ubuntu-image-validation
104: - --gcp-project=ubuntu-image-validation
142: - --gcp-project=ubuntu-image-validation
182: - --gcp-project=ubuntu-image-validation
220: - --gcp-project=ubuntu-image-validation
260: - --gcp-project=ubuntu-image-validation
298: - --gcp-project=ubuntu-image-validation
338: - --gcp-project=ubuntu-image-validation
376: - --gcp-project=ubuntu-image-validation
416: - --gcp-project=ubuntu-image-validation
454: - --gcp-project=ubuntu-image-validation
494: - --gcp-project=ubuntu-image-validation
532: - --gcp-project=ubuntu-image-validation
572: - --gcp-project=ubuntu-image-validation
610: - --gcp-project=ubuntu-image-validation
651: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
697: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
743: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
782: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
822: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
862: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
901: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
941: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
981: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1020: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1060: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1100: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1139: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1179: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1219: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1258: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1298: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1338: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1377: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1417: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1457: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1496: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1536: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
1576: - --gcp-project=ubuntu-os-gke-cloud-dev-tests
config/jobs/kubernetes/sig-release/release-branch-jobs/1.19.yaml
1038: - --gcp-project=k8s-jkns-pr-gce-gpus
config/jobs/kubernetes/sig-network/sig-network-misc.yaml
81: - --gcp-project=k8s-e2e-gce-alpha-api-access
365: - --gcp-project=k8s-jenkins-gce-gci-ip-aliases
config/jobs/kubernetes/sig-release/release-branch-jobs/1.17.yaml
886: - --gcp-project=k8s-jkns-pr-gce-gpus
config/jobs/kubernetes/sig-release/release-branch-jobs/1.16.yaml
736: - --gcp-project=k8s-jkns-pr-gce-gpus
1013: - --gcp-project=k8s-presubmit-scale
config/jobs/kubernetes/sig-network/ingress-gce-e2e.yaml
246: - --gcp-project=k8s-ingress-e2e-scale-backup
config/jobs/kubernetes/sig-release/release-branch-jobs/1.18.yaml
1087: - --gcp-project=k8s-jkns-pr-gce-gpus
config/jobs/kubernetes/sig-autoscaling/sig-autoscaling-presubmits.yaml
40: - --gcp-project=k8s-jkns-gci-autoscaling
config/jobs/kubernetes/sig-autoscaling/sig-autoscaling-config.yaml
181: - --gcp-project=k8s-jkns-gci-autoscaling
207: - --gcp-project=k8s-jkns-gci-autoscaling
241: - --gcp-project=k8s-jkns-gci-autoscaling-migs
267: - --gcp-project=k8s-jkns-gci-autoscaling-migs
config/jobs/kubernetes/sig-scalability/sig-scalability-presubmit-jobs.yaml
98: - --gcp-project=k8s-presubmit-scale
206: - --gcp-project=k8s-presubmit-scale
347: - --gcp-project=k8s-presubmit-scale
config/jobs/kubernetes/sig-scalability/sig-scalability-periodic-jobs.yaml
233: - --gcp-project=k8s-jenkins-blocking-kubemark
290: - --gcp-project=kubemark-scalability-testing
config/jobs/kubernetes/sig-scalability/sig-scalability-golang.yaml
70: - --gcp-project=k8s-presubmit-scale
config/jobs/kubernetes/sig-scalability/sig-scalability-release-blocking-jobs.yaml
30: - --gcp-project=kubernetes-scale
80: - --gcp-project=kubernetes-scale
config/jobs/kubernetes/sig-cloud-provider/gcp/gpu/gpu-gce.yaml
77: - --gcp-project=k8s-infra-e2e-gpu-project
config/jobs/kubernetes/sig-cloud-provider/gcp/gpu/gcp-gpu-presubmits.yaml
45: - --gcp-project=k8s-jkns-pr-gce-gpus
config/jobs/kubernetes/sig-cloud-provider/gcp/gcp-gce.yaml
888: - --gcp-project=k8s-jkns-e2e-gce-ubelite
916: - --gcp-project=k8s-jkns-gce-gci-soak
946: - --gcp-project=k8s-jkns-gce-soak-1-6
976: - --gcp-project=k8s-jkns-gci-gce-soak-1-4
1005: - --gcp-project=k8s-jkns-gci-gce-soak-1-7
1034: - --gcp-project=k8s-jkns-gci-gce-soak-1-6
config/jobs/kubernetes/sig-cloud-provider/gcp/gce-misc.yaml
17: - --gcp-project=kubernetes-ha-master
config/jobs/kubernetes/sig-api-machinery/sig-api-machinery-config.yaml
108: - --gcp-project=k8s-network-proxy-e2e
config/jobs/kubernetes/sig-node/containerd.yaml
269: - --gcp-project=cri-containerd-node-e2e
299: - --gcp-project=cri-containerd-node-e2e
329: - --gcp-project=cri-containerd-node-e2e
359: - --gcp-project=cri-containerd-node-e2e
392: - --gcp-project=k8s-jkns-e2e-gci-gce-soak-1-4
786: - --gcp-project=cri-containerd-node-e2e
816: - --gcp-project=cri-containerd-node-e2e
846: - --gcp-project=cri-containerd-node-e2e
931: - --gcp-project=cri-containerd-node-e2e
960: - --gcp-project=cri-containerd-node-e2e
config/jobs/kubernetes/sig-node/sig-node-presubmit.yaml
143: - --gcp-project=k8s-jkns-pr-node-e2e
179: - --gcp-project=k8s-jkns-pr-node-e2e
213: - --gcp-project=k8s-jkns-pr-node-e2e
247: - --gcp-project=k8s-jkns-pr-node-e2e
284: - --gcp-project=k8s-jkns-pr-node-e2e
config/jobs/cadvisor/cadvisor.yaml
40: - --gcp-project=cadvisor-e2e
96: - --gcp-project=ci-cadvisor-e2e
/unshrug
/sig node
A number of the jobs that still pin are sig-node jobs. I don't know if there's something special about those projects, but I know we've successfully migrated a few node e2e projects to use the standard gce-project
pool boskos offers up
/sig scalability
/sig network
/sig autoscaling
based on some of the other paths
@spiffxp: ¯\_(ツ)_/¯
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta. /lifecycle stale
Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten
.
Rotten issues close after an additional 30d of inactivity.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta. /lifecycle rotten
Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen
.
Mark the issue as fresh with /remove-lifecycle rotten
.
Send feedback to sig-contributor-experience at kubernetes/community. /close
@fejta-bot: Closing this issue.
/remove-lifecycle rotten
/reopen
/priority backlog
/lifecycle frozen
These should be addressed while migrating kubernetes CI away from google.com gcp projects (ref: https://github.com/kubernetes/k8s.io/issues/1469). If there's no need for a special project, just switch to --gcp-project-type=gce
@spiffxp: Reopened this issue.
/help
@spiffxp: This request has been marked as needing help from a contributor.
Please ensure the request meets the requirements listed here.
If this request no longer meets these requirements, the label can be removed
by commenting with the /remove-help
command.
/unassign @amwat
Hi Amit, I'm unassigning you from this so someone else can take it, feel free to take it again if you plan to work on it.
We're doing this by way of eliminating all jobs on the old clusters and only using boskos for community resources
this is ~done.
We are seeing janitor/quota issues again for newly added scalability suites, also we have the log sink issue for gce presubmit. These are all solvable by using boskos projects, and we should do so to unify our presubmit/CI environment.
/area boskos /assign
cc @shyamjvs @BenTheElder