Closed spiffxp closed 4 years ago
@spiffxp: This request has been marked as needing help from a contributor.
Please ensure the request meets the requirements listed here.
If this request no longer meets these requirements, the label can be removed
by commenting with the /remove-help
command.
FYI @liggitt @BenTheElder since this came up in #sig-testing (ref: https://kubernetes.slack.com/archives/C09QZ4DQB/p1564674490149000)
took a quick poke, noting that python3 is being used for these files under bazel, but the image is pypy2 and probably old. will look at updating that while we're at it...
semi-related: https://github.com/kubernetes/test-infra/pull/13732
triage is now broken entirely with https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-test-infra-triage/1157331701362331652
+ bq extract --compression GZIP --destination_format NEWLINE_DELIMITED_JSON k8s-gubernator:temp.triage gs://k8s-gubernator/triage_tests.json.gz
BigQuery error in extract operation: Error processing job
'k8s-gubernator:bqjob_r7ddb74232fc611d3_0000016c533a4156_1': Table
gs://k8s-gubernator/triage_tests.json.gz too large to be exported to a single
file. Specify a uri including a * to shard export. See 'Exporting data into one
or more files' in https://cloud.google.com/bigquery/docs/exporting-data.
triage failing entirely was fixed by https://github.com/kubernetes/test-infra/pull/13753
bailing out still happening eg: https://storage.googleapis.com/kubernetes-jenkins/logs/ci-test-infra-triage/1158788723303780358/build-log.txt
I0806 17:22:18.564] 286/3584, 137 failures, verify vendor-licenses
I0806 17:23:18.640] bailing early, taking too long!
I0806 17:23:57.304] 1417/3584, 38 failures, k8s.io/kubernetes/test/integration/scheduler TestVolumeProvision
I0806 17:25:06.662] bailing early, taking too long!
I0806 17:25:15.805] 1930/3584, 19 failures, k8s.io/kubernetes/test/integration/scheduler TestRescheduleProvisioning
I0806 17:26:21.639] bailing early, taking too long!
I0806 17:26:30.919] 2245/3584, 8 failures, k8s.io/kubernetes/test/integration/scheduler TestPVAffinityConflict
I0806 17:27:34.862] bailing early, taking too long!
Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale
.
Stale issues rot after an additional 30d of inactivity and eventually close.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta. /lifecycle stale
Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten
.
Rotten issues close after an additional 30d of inactivity.
If this issue is safe to close now please do so with /close
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta. /lifecycle rotten
Rotten issues close after 30d of inactivity.
Reopen the issue with /reopen
.
Mark the issue as fresh with /remove-lifecycle rotten
.
Send feedback to sig-testing, kubernetes/test-infra and/or fejta. /close
@fejta-bot: Closing this issue.
What happened: Not seeing as much of
TestVolumeProvision
listed in go.k8s.io/triage as we would expectA look at logs from a recent triage run reveals:
During the next phase
What you expected to happen: To see
TestVolumeProvision
listed in go.k8s.io/triageHow to reproduce it (as minimally and precisely as possible): Try running
triage/update-summaries.sh
(in its current state I would recommend commenting out the parts that write to buckets)Please provide links to example occurrences, if any: See above
Anything else we need to know?:
My guess is there is something common between those two failures, and it's probably large failure text. If so, we could address either by truncating failure text more aggressively, or by increasing the timeout before bailing
/help /area triage /priority important-longterm