Closed mrniranjan closed 1 day ago
/test e2e-upgrade
@mrniranjan: This pull request references Jira Issue OCPBUGS-41487, which is valid. The bug has been moved to the POST state.
Requesting review from QA contact: /cc @mrniranjan
The bug has been updated to refer to the pull request using the external bug tracker.
@openshift-ci-robot: GitHub didn't allow me to request PR reviews from the following users: mrniranjan.
Note that only openshift members and repo collaborators can review this PR, and authors cannot review their own PRs.
Requesting review from QA contact: /cc @mrniranjan
The bug has been updated to refer to the pull request using the external bug tracker.
In response to [this](https://github.com/openshift/cluster-node-tuning-operator/pull/1154): >- Changes primarily done to check nodepools instead of mcp for hypershift. >- replace dataplane test client instead of generic testclient.Client Instructions for interacting with me using PR comments are available [here](https://prow.ci.openshift.org/command-help?repo=openshift%2Fcluster-node-tuning-operator). If you have questions or suggestions related to my behavior, please file an issue against the [openshift-eng/jira-lifecycle-plugin](https://github.com/openshift-eng/jira-lifecycle-plugin/issues/new) repository.
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository.
@mrniranjan Please add the suite to the make target, because currently it's not running.
/retest-required
/lgtm
Thanks!
/approve
/approved
/retest-required
Remaining retests: 0 against base HEAD 6d2e1edef5b898f63617b1cdba589b3083b87331 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD 6d2e1edef5b898f63617b1cdba589b3083b87331 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD 6d2e1edef5b898f63617b1cdba589b3083b87331 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD 2e07e3ab309a29dbf4a147c76869291e8fb1e350 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD 2e07e3ab309a29dbf4a147c76869291e8fb1e350 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
Seems like actual failure
> Enter [BeforeEach] [rfe_id:49062][workloadHints] Telco friendly workload specific PerformanceProfile API - /go/src/github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/8_performance_workloadhints/workloadhints.go:61 @ 10/27/24 11:59:26.333
I1027 11:59:26.358328 20478 workloadhints.go:859] updated nodes from map[string]string{"node-role.kubernetes.io/worker-cnf":""}: []
I1027 11:59:26.358350 20478 workloadhints.go:861] updated nodes matching optional selector: []
[FAILED] cannot find RT enabled worker nodes
Expected
<[]v1.Node | len:0, cap:0>: nil
not to be empty
In [BeforeEach] at: /go/src/github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/8_performance_workloadhints/workloadhints.go:863 @ 10/27/24 11:59:2
/retest-required
Remaining retests: 0 against base HEAD a98b16a6c6d9e1f0fc575fba137d0ffd22178f0b and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD bc3ecaea5131f120ef8a282039c3ffd013cb8a76 and 1 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD bc3ecaea5131f120ef8a282039c3ffd013cb8a76 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD bc3ecaea5131f120ef8a282039c3ffd013cb8a76 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/retest-required
Remaining retests: 0 against base HEAD bc3ecaea5131f120ef8a282039c3ffd013cb8a76 and 2 for PR HEAD f16d3d727d007174db9786e85bdad86e9c28de5f in total
/hold
STEP: Waiting for TuneD to start on nodes - /go/src/github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/8_performance_workloadhints/workloadhints.go:123 @ 10/30/24 10:52:22.36
[FAILED] Unexpected error:
<*errors.errorString | 0xc000223300>:
failed to find a TuneD Pod for node ip-10-0-142-236.ec2.internal
{
s: "failed to find a TuneD Pod for node ip-10-0-142-236.ec2.internal",
}
occurred
In [It] at: /go/src/github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/8_performance_workloadhints/workloadhints.go:132 @ 10/30/24 10:57:22.476
@Tal-or why doesn't the tuned pod start on the HCP nodes ?, when i execute the tests locally i see tuned pod running and tests proceed
/test e2e-hypershift-pao
We reach the jira API quota so it failed on the CPU test again
Nov 6 23:14:39.818: [WARNING]: failed to retrieve status of Jira issue OCPBUGS-43280: failed to get jira status of OCPBUGS-43280: 429 429 Too Many Requests
[1mSTEP:[0m fetch Default cpu set from cpu manager state file before restart [38;5;243m- /go/src/github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/1_performance/cpu_management.go:312 @ 11/06/24 23:14:39.818[0m
Nov 6 23:14:39.821: [INFO]: daemonset "node-inspector-ns" "node-inspector" desired 3 scheduled 3 ready 3
Nov 6 23:14:39.825: [INFO]: found daemon pod node-inspector-m22vt for node ip-10-0-130-199.ec2.internal
cpuset = 0,2
Nov 6 23:14:39.882: [INFO]: pre kubelet restart default cpuset: 0,2
Nov 6 23:14:39.885: [INFO]: daemonset "node-inspector-ns" "node-inspector" desired 3 scheduled 3 ready 3
Nov 6 23:14:39.889: [INFO]: found daemon pod node-inspector-m22vt for node ip-10-0-130-199.ec2.internal
Nov 6 23:15:39.979: [INFO]: post kubele restart: waiting for node "ip-10-0-130-199.ec2.internal": to be ready
Nov 6 23:15:39.985: [INFO]: node "ip-10-0-130-199.ec2.internal" ready=true
Nov 6 23:15:39.985: [INFO]: post kubele restart: node "ip-10-0-130-199.ec2.internal": reported ready
Nov 6 23:15:39.985: [INFO]: post restart: entering cooldown time: 1m0s
Nov 6 23:16:39.985: [INFO]: post restart: finished cooldown time: 1m0s
[1mSTEP:[0m fetch Default cpuset from cpu manager state after restart [38;5;243m- /go/src/github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/1_performance/cpu_management.go:332 @ 11/06/24 23:16:39.985[0m
Nov 6 23:16:39.990: [INFO]: daemonset "node-inspector-ns" "node-inspector" desired 3 scheduled 3 ready 3
Nov 6 23:16:39.993: [INFO]: found daemon pod node-inspector-m22vt for node ip-10-0-130-199.ec2.internal
cpuset = 0-3
[38;5;9m[FAILED] Expected
<cpuset.CPUSet>: {
elems: {0: {}, 2: {}},
}
to equal
<cpuset.CPUSet>: {
elems: {1: {}, 2: {}, 3: {}, 0: {}},
/unhold
/test e2e-hypershift-pao
Reached timeout again. Too many tests are running on the same lane. On OCP we're having a separate lane for the workloadhints but it seems too much here (on HCP).
@mrniranjan can we classified the workloadhints tests and run only the most critical ones in order to cut tests execution time?
Reached timeout again. Too many tests are running on the same lane. On OCP we're having a separate lane for the workloadhints but it seems too much here (on HCP).
@mrniranjan can we classified the workloadhints tests and run only the most critical ones in order to cut tests execution time?
I have classified each workload hints test as tier3 and 1 test as tier-0, Modified the Makefile to skip all tests with workload-hints and tier-3 , so executing only 1 tier-0 tests.
/test okd-scos-e2e-aws-ovn
[APPROVALNOTIFIER] This PR is APPROVED
This pull-request has been approved by: jmencak, MarSik, mrniranjan, shajmakh
The full list of commands accepted by this bot can be found here.
The pull request process is described here
/retest-required
/retest-required
/retest-required
/unhold
/retest-required
/retest
/retest-required
/test okd-scos-e2e-aws-ovn
/retest-required
[FAILED] Unexpected error:
<*fmt.wrapError | 0xc0002a2b20>:
failed to run command [/bin/sh -c tuned-adm profile_info openshift-node-performance-performance 2>/dev/null | grep ^openshift-]: output ""; error ""; command terminated with exit code 1
{
msg: "failed to run command [/bin/sh -c tuned-adm profile_info openshift-node-performance-performance 2>/dev/null | grep ^openshift-]: output \"\"; error \"\"; command terminated with exit code 1",
err: <exec.CodeExitError>{
Err: <*errors.errorString | 0xc000437[220](https://prow.ci.openshift.org/view/gs/test-platform-results/pr-logs/pull/openshift_cluster-node-tuning-operator/1154/pull-ci-openshift-cluster-node-tuning-operator-master-e2e-hypershift-pao/1859177980660551680#1:build-log.txt%3A220)>{
s: "command terminated with exit code 1",
},
Code: 1,
},
}
occurred
Might be a flake, lets follow
@mrniranjan: all tests passed!
Full PR test history. Your PR dashboard.
@Tal-or can you have a look
/lgtm /label acknowledge-critical-fixes-only test code, not going to be part of OCP core payload
@mrniranjan: Jira Issue OCPBUGS-41487: All pull requests linked via external trackers have merged:
Jira Issue OCPBUGS-41487 has been moved to the MODIFIED state.