IBM / cloud-pak-deployer

Configuration-based installation of OpenShift and Cloud Pak for Data/Integration/Watson AIOps on various private and public cloud infrastructure providers. Deployment attempts to achieve the end-state defined in the configuration. If something fails along the way, you only need to restart the process to continue the deployment.
https://ibm.github.io/cloud-pak-deployer/
Apache License 2.0
136 stars 67 forks source link

when running cp-deploy.sh env apply with --accept-all-licenses, TASK[cp4d-ocp-tuned] failed #661

Closed imicer closed 5 months ago

imicer commented 6 months ago

Describe the bug when running cp-deploy.sh env apply with --accept-all-licenses, TASK[cp4d-ocp-tuned] failed as eblow:

TASK [cp4d-ocp-tuned : Get first compute node in the cluster] ******************
Monday 18 March 2024  15:31:15 +0000 (0:00:00.036)       0:01:52.292 **********
fatal: [localhost]: FAILED! => changed=true
  cmd: |-
    oc get no  -l node-role.kubernetes.io/worker,cluster.ocs.openshift.io/openshift-storage!=""  -o jsonpath='{.items[0].metadata.name}'
  delta: '0:00:00.093389'
  end: '2024-03-18 15:31:15.371319'
  msg: non-zero return code
  rc: 1
  start: '2024-03-18 15:31:15.277930'
  stderr: |-
    error: error executing jsonpath "{.items[0].metadata.name}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template:
            template was:
                    {.items[0].metadata.name}
            object given to jsonpath engine was:
                    map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}}
  stderr_lines: <omitted>
  stdout: ''
  stdout_lines: <omitted>

PLAY RECAP *********************************************************************
localhost                  : ok=515  changed=60   unreachable=0    failed=1    skipped=306  rescued=0    ignored=0

Checked node as below:

oc get no  -l node-role.kubernetes.io/worker,cluster.ocs.openshift.io/openshift-storage
NAME                                 STATUS   ROLES    AGE    VERSION
worker0.test04cp4d.cp.fyre.ibm.com   Ready    worker   141d   v1.25.16+6df2177
worker1.test04cp4d.cp.fyre.ibm.com   Ready    worker   141d   v1.25.16+6df2177
worker2.test04cp4d.cp.fyre.ibm.com   Ready    worker   141d   v1.25.16+6df2177
worker3.test04cp4d.cp.fyre.ibm.com   Ready    worker   141d   v1.25.16+6df2177
worker4.test04cp4d.cp.fyre.ibm.com   Ready    worker   141d   v1.25.16+6df2177
worker5.test04cp4d.cp.fyre.ibm.com   Ready    worker   141d   v1.25.16+6df2177

oc get no  -l node-role.kubernetes.io/worker,cluster.ocs.openshift.io/openshift-storage!=""
No resources found

oc get node --show-labels
NAME                                 STATUS   ROLES                  AGE    VERSION            LABELS
master0.test04cp4d.cp.fyre.ibm.com   Ready    control-plane,master   141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=master0.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos
master1.test04cp4d.cp.fyre.ibm.com   Ready    control-plane,master   141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=master1.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos
master2.test04cp4d.cp.fyre.ibm.com   Ready    control-plane,master   141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=master2.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos
worker0.test04cp4d.cp.fyre.ibm.com   Ready    worker                 141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,cluster.ocs.openshift.io/openshift-storage=,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker0.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos
worker1.test04cp4d.cp.fyre.ibm.com   Ready    worker                 141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,cluster.ocs.openshift.io/openshift-storage=,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker1.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos
worker2.test04cp4d.cp.fyre.ibm.com   Ready    worker                 141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,cluster.ocs.openshift.io/openshift-storage=,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker2.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos
worker3.test04cp4d.cp.fyre.ibm.com   Ready    worker                 141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,cluster.ocs.openshift.io/openshift-storage=,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker3.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos
worker4.test04cp4d.cp.fyre.ibm.com   Ready    worker                 141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,cluster.ocs.openshift.io/openshift-storage=,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker4.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos
worker5.test04cp4d.cp.fyre.ibm.com   Ready    worker                 141d   v1.25.16+6df2177   beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,cluster.ocs.openshift.io/openshift-storage=,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker5.test04cp4d.cp.fyre.ibm.com,kubernetes.io/os=linux,node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos

ODF 4.12 has been installed in OCP 4.12.

Could you please check how to resolve the issue? Thanks.

fketelaars commented 6 months ago

@imicer All of your compute nodes have label cluster.ocs.openshift.io/openshift-storage=. When installing ODF/FDF, only the nodes that run the storage cluster should have this label. Please remove the storage label from the nodes that must carry the Cloud Pak workloads.

fketelaars commented 5 months ago

Environment issue, nothing to fix.