Closed Tal-or closed 4 months ago
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by: Tal-or Once this PR has been reviewed and has the lgtm label, please assign ffromani for approval. For more information see the Kubernetes Code Review Process.
The full list of commands accepted by this bot can be found here.
/test ci/prow/e2e-hypershift-pao
@Tal-or: The specified target(s) for /test
were not found.
The following commands are available to trigger required jobs:
/test e2e-aws-operator
/test e2e-aws-ovn
/test e2e-aws-ovn-techpreview
/test e2e-gcp-pao
/test e2e-gcp-pao-updating-profile
/test e2e-gcp-pao-workloadhints
/test e2e-hypershift
/test e2e-hypershift-pao
/test e2e-no-cluster
/test e2e-upgrade
/test images
/test unit
/test verify
/test vet
The following commands are available to trigger optional jobs:
/test e2e-telco5g-cnftests
/test lint
/test okd-scos-images
Use /test all
to run the following jobs that were automatically triggered:
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-aws-operator
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-aws-ovn
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-aws-ovn-techpreview
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-gcp-pao
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-gcp-pao-updating-profile
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-gcp-pao-workloadhints
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-hypershift
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-hypershift-pao
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-no-cluster
pull-ci-openshift-cluster-node-tuning-operator-master-e2e-upgrade
pull-ci-openshift-cluster-node-tuning-operator-master-images
pull-ci-openshift-cluster-node-tuning-operator-master-lint
pull-ci-openshift-cluster-node-tuning-operator-master-unit
pull-ci-openshift-cluster-node-tuning-operator-master-verify
pull-ci-openshift-cluster-node-tuning-operator-master-vet
/test e2e-hypershift-pao
/test e2e-hypershift-pao
/test e2e-hypershift-pao
/test e2e-hypershift-pao
Note to self:
For some reason which is not clear, the testpod
has two nodeSelectors
:
NodeSelector: {
kubernetes.io/hostname: ip-10-0-129-17.ec2.internal,
node-role.kubernetes.io/worker-cnf: ""
},
Now, the node (for another reason which I don't understand) does not contains the worker-cnf
label (although it gets labeled at the beginning of the test).
Checking carefully I see the node that was labeled at the beginning has gone during the test runtime, so the node the pod is scheduled into, does not contain the label anymore.
The coincidence of:
A. A node which was removed from a cluster.
B. An addition if undesired nodeSelector
for the testpod
is the cause for this bug.
All that "left" is to figure out what causes A
and B
/test e2e-hypershift-pao
/test e2e-gcp-pao
@Tal-or: The following test failed, say /retest
to rerun all failed tests or /retest-required
to rerun all mandatory failed tests:
Test name | Commit | Details | Required | Rerun command |
---|---|---|---|---|
ci/prow/e2e-gcp-pao | d0850b3df25afec98a117105f9bcbe2a21bf0122 | link | true | /test e2e-gcp-pao |
Full PR test history. Your PR dashboard.
Note to self: For some reason which is not clear, the
testpod
has twonodeSelectors
:NodeSelector: { kubernetes.io/hostname: ip-10-0-129-17.ec2.internal, node-role.kubernetes.io/worker-cnf: "" },
Now, the node (for another reason which I don't understand) does not contains the
worker-cnf
label (although it gets labeled at the beginning of the test). Checking carefully I see the node that was labeled at the beginning has gone during the test runtime, so the node the pod is scheduled into, does not contain the label anymore. The coincidence of: A. A node which was removed from a cluster. B. An addition if undesirednodeSelector
for the testpod is the cause for this bug. All that "left" is to figure out what causesA
andB
Seems like the worker-cnf has been appended by the API server. definitely, we didn't append this label to the pod in any phase of the pod construction.
Note to self: For some reason which is not clear, the
testpod
has twonodeSelectors
:NodeSelector: { kubernetes.io/hostname: ip-10-0-129-17.ec2.internal, node-role.kubernetes.io/worker-cnf: "" },
Now, the node (for another reason which I don't understand) does not contains the
worker-cnf
label (although it gets labeled at the beginning of the test). Checking carefully I see the node that was labeled at the beginning has gone during the test runtime, so the node the pod is scheduled into, does not contain the label anymore. The coincidence of: A. A node which was removed from a cluster. B. An addition if undesirednodeSelector
for the testpod is the cause for this bug. All that "left" is to figure out what causesA
andB
Seems like the worker-cnf has been appended by the API server. definitely, we didn't append this label to the pod in any phase of the pod construction.
Indeed, but atm we should first make sure that after every time that performance profile is being applied we'll restore the labels for the nodes. IOW set the worker-cnf label. This will fix the bug and we'll handle with the investigation of the appended node selector to the pod later.
/test e2e-hypershift-pao
PR needs rebase.
/hold DNM!