Closed pavanats closed 4 years ago
From the error, it seems that the controller is unable to connect back to node01
. Could be an improper proxy setting. Is this the correct node01 IP 30.30.30.11
?
I tested the above command on my local setup and it is able to retrieve logs.
Can you try the same on your shell?
kubectl logs -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name --field-selector spec.nodeName=node01 | grep ovs-ovn)
Hi Amr, I am not using any proxy configuration. Some of the pods are shown as running. Here's the output of the cli cmd you tried: [root@controller ~]# kubectl logs -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name --field-selector spec.nodeName=node01 | grep ovs-ovn) Error from server: Get https://30.30.30.11:10250/containerLogs/kube-system/ovs-ovn-645b9/openvswitch: dial tcp 30.30.30.11:10250: connect: connection refused
Amr, can we have a quick online meeting. We have been stuck for sometime on just the deployment. Pavan
From: Amr Mokhtar notifications@github.com Sent: Monday, July 13, 2020 4:59 PM To: open-ness/openness-experience-kits openness-experience-kits@noreply.github.com Cc: Pavan Gupta pavan.gupta@atsgen.com; Author author@noreply.github.com Subject: Re: [open-ness/openness-experience-kits] Error in running Ansible playbook for the edge node (#33)
From the error, it seems that the controller is unable to connect back to node01. Could be an improper proxy setting. Is this the correct node01 IP 30.30.30.11 ? I tested the above command on my local setup and it is able to retrieve logs. Can you try the same on your shell?
kubectl logs -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name --field-selector spec.nodeName=node01 | grep ovs-ovn)
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHubhttps://github.com/open-ness/openness-experience-kits/issues/33#issuecomment-657503141, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APSLZC7NYRFUEXC5LXIEDT3R3LV2LANCNFSM4OXL3YNQ.
What is the status of the pods in the cluster?
kubectl get pods -A -o wide
Here's the output. Please note, I haven't been able to deploy the edge node yet without error.
[root@controller ~]# kubectl get pods -o wide -A
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
cdi cdi-operator-76b6694845-7f4jq 0/1 Terminating 0 2d
From: Amr Mokhtar notifications@github.com Sent: Monday, July 13, 2020 5:34 PM To: open-ness/openness-experience-kits openness-experience-kits@noreply.github.com Cc: Pavan Gupta pavan.gupta@atsgen.com; Author author@noreply.github.com Subject: Re: [open-ness/openness-experience-kits] Error in running Ansible playbook for the edge node (#33)
What is the status of the pods in the cluster?
kubectl get pods -A -o wide
— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHubhttps://github.com/open-ness/openness-experience-kits/issues/33#issuecomment-657520367, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APSLZC6K3DBAOXLY527D65TR3LZ33ANCNFSM4OXL3YNQ.
As mentioned earlier, we currently support OpenNESS in bare-metal only deployment. Given that the using VMs is not running smooth, I would suggest that you start by deploying on bare-metal to get familiar with the system, then start migrating to VM-based setup.
@pavanats - Are you ok to close this issue for now and open a new one when you have any issues with the bare-metal installation?
Hi Amr, I am not trying the setup on 2 baremetal servers. You can close this ticket. I will connect with you again if the baremetal deployment fails. Pavan
From: Amr Mokhtar notifications@github.com Sent: Wednesday, July 15, 2020 6:08 PM To: open-ness/openness-experience-kits openness-experience-kits@noreply.github.com Cc: Pavan Gupta pavan.gupta@atsgen.com; Mention mention@noreply.github.com Subject: Re: [open-ness/openness-experience-kits] Error in running Ansible playbook for the edge node (#33)
@pavanatshttps://github.com/pavanats - Are you ok to close this issue for now and open a new one when you have any issues with the bare-metal installation?
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHubhttps://github.com/open-ness/openness-experience-kits/issues/33#issuecomment-658742494, or unsubscribehttps://github.com/notifications/unsubscribe-auth/APSLZC5247XQIJ5RDHAWIITR3WPNBANCNFSM4OXL3YNQ.
Hi Amr, I have tried the edge node deployment on another new server and see the same error again:
"changed": false,
"cmd": "set -o pipefail && kubectl logs -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name --field-selector spec.nodeName=node01 | grep ovs-ovn)\n",
"delta": "0:00:00.256773",
"end": "2020-07-15 19:12:14.914509",
"rc": 1,
"start": "2020-07-15 19:12:14.657736"
}
STDERR:
Error from server: Get https://x.x.x.x:10250/containerLogs/kube-system/ovs-ovn-tgfq6/openvswitch: dial tcp x.x.x.x:10250: connect: connection refused
MSG:
non-zero return code ...ignoring
TASK [kubernetes/cni/kubeovn/worker : end the playbook] ***** task path: /root/openness-experience-kits/roles/kubernetes/cni/kubeovn/worker/tasks/main.yml:84 fatal: [node01]: FAILED! => { "changed": false }
MSG:
end the playbook: either ovs-ovn pod did not start or the socket was not created
This error isn't really related to using VM for setup. It can be seen on both hardware and VMs. Output for different pods is shows below:
[root@controller ~]# kubectl get pods -o wide -A
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
cdi cdi-apiserver-885758cc4-m4dhn 1/1 Terminating 0 3h54m 10.16.0.25 node01
Hi, I could install the controller on a VM, however, edge node installation on physical server has failed with the following error: TASK [kubernetes/cni/kubeovn/worker : try to get ovs-ovn execution logs] *** task path: /root/openness-experience-kits/roles/kubernetes/cni/kubeovn/worker/tasks/main.yml:75 fatal: [node01 -> 30.30.30.22]: FAILED! => { "changed": false, "cmd": "set -o pipefail && kubectl logs -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name --field-selector spec.nodeName=node01 | grep ovs-ovn)\n", "delta": "0:00:00.444066", "end": "2020-07-11 12:03:37.659635", "rc": 1, "start": "2020-07-11 12:03:37.215569" }
STDERR:
Error from server: Get https://30.30.30.11:10250/containerLogs/kube-system/ovs-ovn-645b9/openvswitch: dial tcp 30.30.30.11:10250: connect: connection refused
If there is a known solution or a workaround, please let us know.