Closed sbbroot closed 7 months ago
We had the same issue in an offline env, and fixed it by doing a kubeadm reset
on each of the masters, then running the playbook again.
The Kubernetes project currently lacks enough contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle stale
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues.
This bot triages un-triaged issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/remove-lifecycle rotten
/close
Please send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle rotten
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.
This bot triages issues according to the following rules:
lifecycle/stale
is appliedlifecycle/stale
was applied, lifecycle/rotten
is appliedlifecycle/rotten
was applied, the issue is closedYou can:
/reopen
/remove-lifecycle rotten
Please send feedback to sig-contributor-experience at kubernetes/community.
/close not-planned
@k8s-triage-robot: Closing this issue, marking it as "Not Planned".
@lord0gnome terrible advice. kubeadm reset
wiped my cluster.
I'm very sorry @clayrisser . I should have mentioned that we were installing a new cluster and so there was no consequence to doing a kubeadm reset. My understanding of what we're doing at the time was minimal and I was just trying to help others who may have been stuck in similar situations.
Environment:
Cloud provider or hardware configuration: Offline deployment.
OS (
printf "$(uname -srm)\n$(cat /etc/os-release)\n"
):ALMALINUX_MANTISBT_PROJECT="AlmaLinux-8" ALMALINUX_MANTISBT_PROJECT_VERSION="8.5"
ansible [core 2.12.5] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.8/dist-packages/ansible ansible collection location = /root/.ansible/collections:/usr/share/ansible/collections executable location = /usr/local/bin/ansible python version = 3.8.10 (default, Jun 22 2022, 20:18:18) [GCC 9.4.0] jinja version = 2.11.3 libyaml = True
Python 3.8.10
Calico
all: hosts: node1: ansible_host: 192.168.122.211 ip: 192.168.122.211 access_ip: 192.168.122.211 node2: ansible_host: 192.168.122.212 ip: 192.168.122.212 access_ip: 192.168.122.212 node3: ansible_host: 192.168.122.213 ip: 192.168.122.213 access_ip: 192.168.122.213 children: kube_control_plane: hosts: node1: node2: kube_node: hosts: node1: node2: node3: etcd: hosts: node1: node2: node3: k8s_cluster: children: kube_control_plane: kube_node: calico_rr: hosts: {}
ansible-playbook -i inventory/onprem/hosts.yaml --become --user=root --become-user=root cluster.yml
TASK [kubernetes/control-plane : kubeadm | Copy kubeadm patches from inventory files] *****
skipping: [node1] skipping: [node2]
TASK [kubernetes/control-plane : kubeadm | Initialize first master] ***
skipping: [node1] skipping: [node2]
TASK [kubernetes/control-plane : set kubeadm certificate key] *****
TASK [kubernetes/control-plane : Create hardcoded kubeadm token for joining nodes with 24h expiration (if defined)] ***
skipping: [node1] skipping: [node2]
TASK [kubernetes/control-plane : Create kubeadm token for joining nodes with 24h expiration (default)] ****
FAILED - RETRYING: [node1]: Create kubeadm token for joining nodes with 24h expiration (default) (5 retries left). FAILED - RETRYING: [node2 -> node1]: Create kubeadm token for joining nodes with 24h expiration (default) (5 retries left). FAILED - RETRYING: [node2 -> node1]: Create kubeadm token for joining nodes with 24h expiration (default) (4 retries left). FAILED - RETRYING: [node1]: Create kubeadm token for joining nodes with 24h expiration (default) (4 retries left). FAILED - RETRYING: [node2 -> node1]: Create kubeadm token for joining nodes with 24h expiration (default) (3 retries left). FAILED - RETRYING: [node1]: Create kubeadm token for joining nodes with 24h expiration (default) (3 retries left). FAILED - RETRYING: [node2 -> node1]: Create kubeadm token for joining nodes with 24h expiration (default) (2 retries left). FAILED - RETRYING: [node1]: Create kubeadm token for joining nodes with 24h expiration (default) (2 retries left). FAILED - RETRYING: [node1]: Create kubeadm token for joining nodes with 24h expiration (default) (1 retries left). FAILED - RETRYING: [node2 -> node1]: Create kubeadm token for joining nodes with 24h expiration (default) (1 retries left). fatal: [node2 -> node1(192.168.122.201)]: FAILED! => {"attempts": 5, "changed": false, "cmd": ["/usr/local/bin/kubeadm", "--kubeconfig", "/etc/kubernetes/admin.conf", "token", "create"], "delta": "0:01:15.090228 ", "end": "2023-03-16 14:36:35.958446", "msg": "non-zero return code", "rc": 1, "start": "2023-03-16 14:35:20.868218", "stderr": "timed out waiting for the condition\nTo see the stack trace of this error execute with --v=5 or higher", "stderr_lines": ["timed out waiting for the condition", "To see the stack trace of this error execute with --v=5 or higher"], "stdout": "", "stdout_lines": []} fatal: [node1]: FAILED! => {"attempts": 5, "changed": false, "cmd": ["/usr/local/bin/kubeadm", "--kubeconfig", "/etc/kubernetes/admin.conf", "token", "create"], "delta": "0:01:15.111611", "end": "2023-03-16 14:3 6:35.982683", "msg": "non-zero return code", "rc": 1, "start": "2023-03-16 14:35:20.871072", "stderr": "timed out waiting for the condition\nTo see the stack trace of this error execute with --v=5 or higher", "s tderr_lines": ["timed out waiting for the condition", "To see the stack trace of this error execute with --v=5 or higher"], "stdout": "", "stdout_lines": []}
NO MORE HOSTS LEFT ****
PLAY RECAP ****
localhost : ok=3 changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 node1 : ok=569 changed=16 unreachable=0 failed=1 skipped=754 rescued=0 ignored=2 node2 : ok=529 changed=14 unreachable=0 failed=1 skipped=647 rescued=0 ignored=2 node3 : ok=474 changed=12 unreachable=0 failed=0 skipped=577 rescued=0 ignored=1
root@node1:~# /usr/local/bin/kubeadm --kubeconfig /etc/kubernetes/admin.conf token create --v=8 I0316 15:00:22.062350 114713 token.go:119] [token] validating mixed arguments I0316 15:00:22.062534 114713 token.go:128] [token] getting Clientsets from kubeconfig file I0316 15:00:22.063512 114713 loader.go:374] Config loaded from file: /etc/kubernetes/admin.conf I0316 15:00:22.064669 114713 token.go:243] [token] loading configurations I0316 15:00:22.065165 114713 interface.go:432] Looking for default routes with IPv4 addresses I0316 15:00:22.065289 114713 interface.go:437] Default route transits interface "enp1s0" I0316 15:00:22.065514 114713 interface.go:209] Interface enp1s0 is up I0316 15:00:22.065782 114713 interface.go:257] Interface "enp1s0" has 2 addresses :[192.168.122.201/24 fe80::5054:ff:fe5f:e352/64]. I0316 15:00:22.065935 114713 interface.go:224] Checking addr 192.168.122.201/24. I0316 15:00:22.066041 114713 interface.go:231] IP found 192.168.122.201 I0316 15:00:22.066180 114713 interface.go:263] Found valid IPv4 address 192.168.122.201 for interface "enp1s0". I0316 15:00:22.066247 114713 interface.go:443] Found active IP 192.168.122.201 I0316 15:00:22.066421 114713 kubelet.go:196] the value of KubeletConfiguration.cgroupDriver is empty; setting it to "systemd" I0316 15:00:22.077310 114713 token.go:250] [token] creating token I0316 15:00:22.077882 114713 round_trippers.go:463] GET https://192.168.122.201:6443/api/v1/namespaces/kube-system/secrets/bootstrap-token-xy5i84?timeout=10s I0316 15:00:22.077991 114713 round_trippers.go:469] Request Headers: I0316 15:00:22.078175 114713 round_trippers.go:473] Accept: application/json, / I0316 15:00:22.078278 114713 round_trippers.go:473] User-Agent: kubeadm/v1.25.6 (linux/amd64) kubernetes/ff2c119 I0316 15:00:22.081208 114713 round_trippers.go:574] Response Status: in 2 milliseconds I0316 15:00:22.081383 114713 round_trippers.go:577] Response Headers: I0316 15:00:22.082563 114713 request.go:1172] Request Body: {"kind":"Secret","apiVersion":"v1","metadata":{"name":"bootstrap-token-xy5i84","namespace":"kube-system","creationTimestamp":null},"data":{"auth-extra-groups":"c3lzdGVtOmJvb3RzdHJhcHBlcnM6a3ViZWFkbTpkZWZhdWx0LW5vZGUtdG9rZW4=","expiration":"MjAyMy0wMy0xN1QxNTowMDoyMlo=","token-id":"eHk1aTg0","token-secret":"Mmh1Z210MDZlMjJuOWdqeg==","usage-bootstrap-authentication":"dHJ1ZQ==","usage-bootstrap-signing":"dHJ1ZQ=="},"type":"bootstrap.kubernetes.io/token"} ... timed out waiting for the condition
containerd.yml: