Open Nello-Angelo opened 5 months ago
v3.1.0-alpha.5 works
Encountered the same problem. The phenomenon is that the registry node and the master node are the same, so there will be problems with containerd, and you will find that the /etc/containerd directory is missing.
What is version of KubeKey has the issue?
3.1.1
What is your os environment?
debian 11.5
KubeKey config file
root@debian:~# kk create cluster -f config-sample.yaml
| | / / | | | | / / | |/ / | | __| |/ / | | | | | ' \ / \ \ / \ | | | | |\ \ || | |) | / |\ \ / || | _| _/_,|./ \_| _/_|__, | _/ | |/
22:53:42 MSK [GreetingsModule] Greetings 22:53:42 MSK message: [worker2] Greetings, KubeKey! 22:53:42 MSK message: [master] Greetings, KubeKey! 22:53:42 MSK message: [worker1] Greetings, KubeKey! 22:53:42 MSK success: [worker2] 22:53:42 MSK success: [master] 22:53:42 MSK success: [worker1] 22:53:42 MSK [NodePreCheckModule] A pre-check on nodes 22:53:43 MSK success: [worker1] 22:53:43 MSK success: [master] 22:53:43 MSK success: [worker2] 22:53:43 MSK [ConfirmModule] Display confirmation form +---------+------+------+---------+----------+-------+-------+---------+-----------+--------+--------+------------+------------+-------------+------------------+--------------+ | name | sudo | curl | openssl | ebtables | socat | ipset | ipvsadm | conntrack | chrony | docker | containerd | nfs client | ceph client | glusterfs client | time | +---------+------+------+---------+----------+-------+-------+---------+-----------+--------+--------+------------+------------+-------------+------------------+--------------+ | master | y | y | y | y | y | | | y | | | v1.7.13 | | | | MSK 22:53:43 | | worker1 | y | y | y | y | y | | | y | | | v1.7.13 | | | | MSK 22:53:43 | | worker2 | y | y | y | y | y | | | y | | | v1.7.13 | | | | MSK 22:53:43 | +---------+------+------+---------+----------+-------+-------+---------+-----------+--------+--------+------------+------------+-------------+------------------+--------------+
This is a simple check of your environment. Before installation, ensure that your machines meet all requirements specified at https://github.com/kubesphere/kubekey#requirements-and-recommendations
Continue this installation? [yes/no]: yes 22:53:45 MSK success: [LocalHost] 22:53:45 MSK [NodeBinariesModule] Download installation binaries 22:53:45 MSK message: [localhost] downloading amd64 kubeadm v1.28.0 ... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 48.3M 100 48.3M 0 0 7358k 0 0:00:06 0:00:06 --:--:-- 9034k 22:53:53 MSK message: [localhost] downloading amd64 kubelet v1.28.0 ... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 105M 100 105M 0 0 8688k 0 0:00:12 0:00:12 --:--:-- 10.5M 22:54:06 MSK message: [localhost] downloading amd64 kubectl v1.28.0 ... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 47.5M 100 47.5M 0 0 7416k 0 0:00:06 0:00:06 --:--:-- 8901k 22:54:13 MSK message: [localhost] downloading amd64 helm v3.14.3 ... 22:54:13 MSK message: [localhost] helm exists 22:54:13 MSK message: [localhost] downloading amd64 kubecni v1.2.0 ... 22:54:13 MSK message: [localhost] kubecni exists 22:54:13 MSK message: [localhost] downloading amd64 crictl v1.29.0 ... 22:54:14 MSK message: [localhost] crictl exists 22:54:14 MSK message: [localhost] downloading amd64 etcd v3.5.13 ... 22:54:14 MSK message: [localhost] etcd exists 22:54:14 MSK message: [localhost] downloading amd64 containerd 1.7.13 ... 22:54:14 MSK message: [localhost] containerd exists 22:54:14 MSK message: [localhost] downloading amd64 runc v1.1.12 ... 22:54:14 MSK message: [localhost] runc exists 22:54:14 MSK message: [localhost] downloading amd64 calicoctl v3.27.3 ... 22:54:15 MSK message: [localhost] calicoctl exists 22:54:15 MSK success: [LocalHost] 22:54:15 MSK [ConfigureOSModule] Get OS release 22:54:15 MSK success: [worker2] 22:54:15 MSK success: [master] 22:54:15 MSK success: [worker1] 22:54:15 MSK [ConfigureOSModule] Prepare to init OS 22:54:15 MSK success: [worker2] 22:54:15 MSK success: [worker1] 22:54:15 MSK success: [master] 22:54:15 MSK [ConfigureOSModule] Generate init os script 22:54:15 MSK success: [master] 22:54:15 MSK success: [worker1] 22:54:15 MSK success: [worker2] 22:54:15 MSK [ConfigureOSModule] Exec init os script 22:54:16 MSK stdout: [worker2] net.ipv4.conf.default.rp_filter = 1 net.ipv4.conf.all.rp_filter = 1 net.ipv4.ip_forward = 1 net.bridge.bridge-nf-call-arptables = 1 net.bridge.bridge-nf-call-ip6tables = 1 net.bridge.bridge-nf-call-iptables = 1 net.ipv4.ip_local_reserved_ports = 30000-32767 net.core.netdev_max_backlog = 65535 net.core.rmem_max = 33554432 net.core.wmem_max = 33554432 net.core.somaxconn = 32768 net.ipv4.tcp_max_syn_backlog = 1048576 net.ipv4.neigh.default.gc_thresh1 = 512 net.ipv4.neigh.default.gc_thresh2 = 2048 net.ipv4.neigh.default.gc_thresh3 = 4096 net.ipv4.tcp_retries2 = 15 net.ipv4.tcp_max_tw_buckets = 1048576 net.ipv4.tcp_max_orphans = 65535 net.ipv4.tcp_keepalive_time = 600 net.ipv4.tcp_keepalive_intvl = 30 net.ipv4.tcp_keepalive_probes = 10 net.ipv4.udp_rmem_min = 131072 net.ipv4.udp_wmem_min = 131072 net.ipv4.conf.all.arp_accept = 1 net.ipv4.conf.default.arp_accept = 1 net.ipv4.conf.all.arp_ignore = 1 net.ipv4.conf.default.arp_ignore = 1 vm.max_map_count = 262144 vm.swappiness = 0 vm.overcommit_memory = 0 fs.inotify.max_user_instances = 524288 fs.inotify.max_user_watches = 524288 fs.pipe-max-size = 4194304 fs.aio-max-nr = 262144 kernel.pid_max = 65535 kernel.watchdog_thresh = 5 kernel.hung_task_timeout_secs = 5 net.ipv6.conf.all.disable_ipv6 = 0 net.ipv6.conf.default.disable_ipv6 = 0 net.ipv6.conf.lo.disable_ipv6 = 0 net.ipv6.conf.all.forwarding = 1 22:54:16 MSK stdout: [worker1] net.ipv4.conf.default.rp_filter = 1 net.ipv4.conf.all.rp_filter = 1 net.ipv4.ip_forward = 1 net.bridge.bridge-nf-call-arptables = 1 net.bridge.bridge-nf-call-ip6tables = 1 net.bridge.bridge-nf-call-iptables = 1 net.ipv4.ip_local_reserved_ports = 30000-32767 net.core.netdev_max_backlog = 65535 net.core.rmem_max = 33554432 net.core.wmem_max = 33554432 net.core.somaxconn = 32768 net.ipv4.tcp_max_syn_backlog = 1048576 net.ipv4.neigh.default.gc_thresh1 = 512 net.ipv4.neigh.default.gc_thresh2 = 2048 net.ipv4.neigh.default.gc_thresh3 = 4096 net.ipv4.tcp_retries2 = 15 net.ipv4.tcp_max_tw_buckets = 1048576 net.ipv4.tcp_max_orphans = 65535 net.ipv4.tcp_keepalive_time = 600 net.ipv4.tcp_keepalive_intvl = 30 net.ipv4.tcp_keepalive_probes = 10 net.ipv4.udp_rmem_min = 131072 net.ipv4.udp_wmem_min = 131072 net.ipv4.conf.all.arp_accept = 1 net.ipv4.conf.default.arp_accept = 1 net.ipv4.conf.all.arp_ignore = 1 net.ipv4.conf.default.arp_ignore = 1 vm.max_map_count = 262144 vm.swappiness = 0 vm.overcommit_memory = 0 fs.inotify.max_user_instances = 524288 fs.inotify.max_user_watches = 524288 fs.pipe-max-size = 4194304 fs.aio-max-nr = 262144 kernel.pid_max = 65535 kernel.watchdog_thresh = 5 kernel.hung_task_timeout_secs = 5 net.ipv6.conf.all.disable_ipv6 = 0 net.ipv6.conf.default.disable_ipv6 = 0 net.ipv6.conf.lo.disable_ipv6 = 0 net.ipv6.conf.all.forwarding = 1 22:54:16 MSK stdout: [master] net.ipv4.conf.default.rp_filter = 1 net.ipv4.conf.all.rp_filter = 1 net.ipv4.ip_forward = 1 net.bridge.bridge-nf-call-arptables = 1 net.bridge.bridge-nf-call-ip6tables = 1 net.bridge.bridge-nf-call-iptables = 1 net.ipv4.ip_local_reserved_ports = 30000-32767 net.core.netdev_max_backlog = 65535 net.core.rmem_max = 33554432 net.core.wmem_max = 33554432 net.core.somaxconn = 32768 net.ipv4.tcp_max_syn_backlog = 1048576 net.ipv4.neigh.default.gc_thresh1 = 512 net.ipv4.neigh.default.gc_thresh2 = 2048 net.ipv4.neigh.default.gc_thresh3 = 4096 net.ipv4.tcp_retries2 = 15 net.ipv4.tcp_max_tw_buckets = 1048576 net.ipv4.tcp_max_orphans = 65535 net.ipv4.tcp_keepalive_time = 600 net.ipv4.tcp_keepalive_intvl = 30 net.ipv4.tcp_keepalive_probes = 10 net.ipv4.udp_rmem_min = 131072 net.ipv4.udp_wmem_min = 131072 net.ipv4.conf.all.arp_accept = 1 net.ipv4.conf.default.arp_accept = 1 net.ipv4.conf.all.arp_ignore = 1 net.ipv4.conf.default.arp_ignore = 1 vm.max_map_count = 262144 vm.swappiness = 0 vm.overcommit_memory = 0 fs.inotify.max_user_instances = 524288 fs.inotify.max_user_watches = 524288 fs.pipe-max-size = 4194304 fs.aio-max-nr = 262144 kernel.pid_max = 65535 kernel.watchdog_thresh = 5 kernel.hung_task_timeout_secs = 5 net.ipv6.conf.all.disable_ipv6 = 0 net.ipv6.conf.default.disable_ipv6 = 0 net.ipv6.conf.lo.disable_ipv6 = 0 net.ipv6.conf.all.forwarding = 1 22:54:16 MSK success: [worker2] 22:54:16 MSK success: [worker1] 22:54:16 MSK success: [master] 22:54:16 MSK [ConfigureOSModule] configure the ntp server for each node 22:54:16 MSK skipped: [worker2] 22:54:16 MSK skipped: [master] 22:54:16 MSK skipped: [worker1] 22:54:16 MSK [KubernetesStatusModule] Get kubernetes cluster status 22:54:16 MSK success: [master] 22:54:16 MSK [InstallContainerModule] Sync containerd binaries 22:54:16 MSK skipped: [master] 22:54:16 MSK skipped: [worker2] 22:54:16 MSK skipped: [worker1] 22:54:16 MSK [InstallContainerModule] Generate containerd service 22:54:16 MSK skipped: [master] 22:54:16 MSK skipped: [worker2] 22:54:16 MSK skipped: [worker1] 22:54:16 MSK [InstallContainerModule] Generate containerd config 22:54:16 MSK skipped: [master] 22:54:16 MSK skipped: [worker2] 22:54:16 MSK skipped: [worker1] 22:54:16 MSK [InstallContainerModule] Enable containerd 22:54:16 MSK skipped: [master] 22:54:16 MSK skipped: [worker2] 22:54:16 MSK skipped: [worker1] 22:54:16 MSK [InstallContainerModule] Sync crictl binaries 22:54:16 MSK skipped: [master] 22:54:16 MSK skipped: [worker2] 22:54:16 MSK skipped: [worker1] 22:54:16 MSK [InstallContainerModule] Generate crictl config 22:54:16 MSK success: [worker2] 22:54:16 MSK success: [master] 22:54:16 MSK success: [worker1] 22:54:16 MSK [PullModule] Start to pull images on all nodes 22:54:16 MSK message: [worker2] downloading image: kubesphere/pause:3.9 22:54:16 MSK message: [master] downloading image: kubesphere/etcd:v3.5.13 22:54:16 MSK message: [worker1] downloading image: kubesphere/pause:3.9 22:54:17 MSK message: [worker1] downloading image: kubesphere/kube-proxy:v1.28.0 22:54:17 MSK message: [master] downloading image: kubesphere/pause:3.9 22:54:17 MSK message: [worker2] downloading image: kubesphere/kube-proxy:v1.28.0 22:54:17 MSK message: [worker1] downloading image: coredns/coredns:1.9.3 22:54:17 MSK message: [worker2] downloading image: coredns/coredns:1.9.3 22:54:17 MSK message: [master] downloading image: kubesphere/kube-apiserver:v1.28.0 22:54:17 MSK message: [worker1] downloading image: kubesphere/k8s-dns-node-cache:1.22.20 22:54:17 MSK message: [worker2] downloading image: kubesphere/k8s-dns-node-cache:1.22.20 22:54:17 MSK message: [master] downloading image: kubesphere/kube-controller-manager:v1.28.0 22:54:17 MSK message: [worker1] downloading image: calico/kube-controllers:v3.27.3 22:54:17 MSK message: [worker2] downloading image: calico/kube-controllers:v3.27.3 22:54:17 MSK message: [master] downloading image: kubesphere/kube-scheduler:v1.28.0 22:54:17 MSK message: [worker1] downloading image: calico/cni:v3.27.3 22:54:17 MSK message: [worker2] downloading image: calico/cni:v3.27.3 22:54:17 MSK message: [master] downloading image: kubesphere/kube-proxy:v1.28.0 22:54:17 MSK message: [worker1] downloading image: calico/node:v3.27.3 22:54:17 MSK message: [worker2] downloading image: calico/node:v3.27.3 22:54:17 MSK message: [master] downloading image: coredns/coredns:1.9.3 22:54:17 MSK message: [worker1] downloading image: calico/pod2daemon-flexvol:v3.27.3 22:54:17 MSK message: [worker2] downloading image: calico/pod2daemon-flexvol:v3.27.3 22:54:17 MSK message: [master] downloading image: kubesphere/k8s-dns-node-cache:1.22.20 22:54:17 MSK message: [worker1] downloading image: library/haproxy:2.9.6-alpine 22:54:17 MSK message: [worker2] downloading image: library/haproxy:2.9.6-alpine 22:54:17 MSK message: [master] downloading image: calico/kube-controllers:v3.27.3 22:54:17 MSK message: [master] downloading image: calico/cni:v3.27.3 22:54:17 MSK message: [master] downloading image: calico/node:v3.27.3 22:54:17 MSK message: [master] downloading image: calico/pod2daemon-flexvol:v3.27.3 22:54:17 MSK success: [worker1] 22:54:17 MSK success: [worker2] 22:54:17 MSK success: [master] 22:54:17 MSK [InstallKubeBinariesModule] Synchronize kubernetes binaries 22:54:40 MSK success: [master] 22:54:40 MSK success: [worker2] 22:54:40 MSK success: [worker1] 22:54:40 MSK [InstallKubeBinariesModule] Change kubelet mode 22:54:40 MSK success: [worker1] 22:54:40 MSK success: [master] 22:54:40 MSK success: [worker2] 22:54:40 MSK [InstallKubeBinariesModule] Generate kubelet service 22:54:40 MSK success: [worker1] 22:54:40 MSK success: [master] 22:54:40 MSK success: [worker2] 22:54:40 MSK [InstallKubeBinariesModule] Enable kubelet service 22:54:41 MSK success: [worker2] 22:54:41 MSK success: [worker1] 22:54:41 MSK success: [master] 22:54:41 MSK [InstallKubeBinariesModule] Generate kubelet env 22:54:41 MSK success: [worker1] 22:54:41 MSK success: [master] 22:54:41 MSK success: [worker2] 22:54:41 MSK [InitKubernetesModule] Generate kubeadm config 22:54:41 MSK success: [master] 22:54:41 MSK [InitKubernetesModule] Generate audit policy 22:54:41 MSK skipped: [master] 22:54:41 MSK [InitKubernetesModule] Generate audit webhook 22:54:41 MSK skipped: [master] 22:54:41 MSK [InitKubernetesModule] Init cluster using kubeadm 22:59:44 MSK stdout: [master] W0423 22:54:41.834178 5820 utils.go:69] The recommended value for "clusterDNS" in "KubeletConfiguration" is: [10.233.0.10]; the provided value is: [169.254.25.10] [init] Using Kubernetes version: v1.28.0 [preflight] Running pre-flight checks [preflight] Pulling images required for setting up a Kubernetes cluster [preflight] This might take a minute or two, depending on the speed of your internet connection [preflight] You can also perform this action in beforehand using 'kubeadm config images pull' [WARNING ImagePull]: failed to pull image kubesphere/etcd:v3.5.13: output: E0423 22:55:39.163611 5916 remote_image.go:180] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"docker.io/kubesphere/etcd:v3.5.13\": failed to resolve reference \"docker.io/kubesphere/etcd:v3.5.13\": docker.io/kubesphere/etcd:v3.5.13: not found" image="kubesphere/etcd:v3.5.13" time="2024-04-23T22:55:39+03:00" level=fatal msg="pulling image: rpc error: code = NotFound desc = failed to pull and unpack image \"docker.io/kubesphere/etcd:v3.5.13\": failed to resolve reference \"docker.io/kubesphere/etcd:v3.5.13\": docker.io/kubesphere/etcd:v3.5.13: not found" , error: exit status 1 [certs] Using certificateDir folder "/etc/kubernetes/pki" [certs] Generating "ca" certificate and key [certs] Generating "apiserver" certificate and key [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local lb.kubesphere.local localhost master master.cluster.local worker1 worker1.cluster.local worker2 worker2.cluster.local] and IPs [10.233.0.1 192.168.163.130 127.0.0.1 192.168.163.131 192.168.163.132] [certs] Generating "apiserver-kubelet-client" certificate and key [certs] Generating "front-proxy-ca" certificate and key [certs] Generating "front-proxy-client" certificate and key [certs] Generating "etcd/ca" certificate and key [certs] Generating "etcd/server" certificate and key [certs] etcd/server serving cert is signed for DNS names [localhost master] and IPs [192.168.163.130 127.0.0.1 ::1] [certs] Generating "etcd/peer" certificate and key [certs] etcd/peer serving cert is signed for DNS names [localhost master] and IPs [192.168.163.130 127.0.0.1 ::1] [certs] Generating "etcd/healthcheck-client" certificate and key [certs] Generating "apiserver-etcd-client" certificate and key [certs] Generating "sa" key and public key [kubeconfig] Using kubeconfig folder "/etc/kubernetes" [kubeconfig] Writing "admin.conf" kubeconfig file [kubeconfig] Writing "kubelet.conf" kubeconfig file [kubeconfig] Writing "controller-manager.conf" kubeconfig file [kubeconfig] Writing "scheduler.conf" kubeconfig file [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests" [control-plane] Using manifest folder "/etc/kubernetes/manifests" [control-plane] Creating static Pod manifest for "kube-apiserver" [control-plane] Creating static Pod manifest for "kube-controller-manager" [control-plane] Creating static Pod manifest for "kube-scheduler" [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env" [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml" [kubelet-start] Starting the kubelet [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s [kubelet-check] Initial timeout of 40s passed.
Unfortunately, an error has occurred: timed out waiting for the condition
This error is likely caused by:
If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
Additionally, a control plane component may have crashed or exited when started by the container runtime. To troubleshoot, list all containers using your preferred container runtimes CLI. Here is one example how you may list all running Kubernetes containers by using crictl:
The reset process does not clean CNI configuration. To do so, you must remove /etc/cni/net.d
The reset process does not reset or clean up iptables rules or IPVS tables. If you wish to reset iptables, you must do so manually by using the "iptables" command.
If your cluster was setup to utilize IPVS, run ipvsadm --clear (or similar) to reset your system's IPVS tables.
The reset process does not clean your kubeconfig files and you must remove them manually. Please, check the contents of the $HOME/.kube/config file. 22:59:45 MSK message: [master] init kubernetes cluster failed: Failed to exec command: sudo -E /bin/bash -c "/usr/local/bin/kubeadm init --config=/etc/kubernetes/kubeadm-config.yaml --ignore-preflight-errors=FileExisting-crictl,ImagePull" W0423 22:54:41.834178 5820 utils.go:69] The recommended value for "clusterDNS" in "KubeletConfiguration" is: [10.233.0.10]; the provided value is: [169.254.25.10] [init] Using Kubernetes version: v1.28.0 [preflight] Running pre-flight checks [preflight] Pulling images required for setting up a Kubernetes cluster [preflight] This might take a minute or two, depending on the speed of your internet connection [preflight] You can also perform this action in beforehand using 'kubeadm config images pull' [WARNING ImagePull]: failed to pull image kubesphere/etcd:v3.5.13: output: E0423 22:55:39.163611 5916 remote_image.go:180] "PullImage from image service failed" err="rpc error: code = NotFound desc = failed to pull and unpack image \"docker.io/kubesphere/etcd:v3.5.13\": failed to resolve reference \"docker.io/kubesphere/etcd:v3.5.13\": docker.io/kubesphere/etcd:v3.5.13: not found" image="kubesphere/etcd:v3.5.13" time="2024-04-23T22:55:39+03:00" level=fatal msg="pulling image: rpc error: code = NotFound desc = failed to pull and unpack image \"docker.io/kubesphere/etcd:v3.5.13\": failed to resolve reference \"docker.io/kubesphere/etcd:v3.5.13\": docker.io/kubesphere/etcd:v3.5.13: not found" , error: exit status 1 [certs] Using certificateDir folder "/etc/kubernetes/pki" [certs] Generating "ca" certificate and key [certs] Generating "apiserver" certificate and key [certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local lb.kubesphere.local localhost master master.cluster.local worker1 worker1.cluster.local worker2 worker2.cluster.local] and IPs [10.233.0.1 192.168.163.130 127.0.0.1 192.168.163.131 192.168.163.132] [certs] Generating "apiserver-kubelet-client" certificate and key [certs] Generating "front-proxy-ca" certificate and key [certs] Generating "front-proxy-client" certificate and key [certs] Generating "etcd/ca" certificate and key [certs] Generating "etcd/server" certificate and key [certs] etcd/server serving cert is signed for DNS names [localhost master] and IPs [192.168.163.130 127.0.0.1 ::1] [certs] Generating "etcd/peer" certificate and key [certs] etcd/peer serving cert is signed for DNS names [localhost master] and IPs [192.168.163.130 127.0.0.1 ::1] [certs] Generating "etcd/healthcheck-client" certificate and key [certs] Generating "apiserver-etcd-client" certificate and key [certs] Generating "sa" key and public key [kubeconfig] Using kubeconfig folder "/etc/kubernetes" [kubeconfig] Writing "admin.conf" kubeconfig file [kubeconfig] Writing "kubelet.conf" kubeconfig file [kubeconfig] Writing "controller-manager.conf" kubeconfig file [kubeconfig] Writing "scheduler.conf" kubeconfig file [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests" [control-plane] Using manifest folder "/etc/kubernetes/manifests" [control-plane] Creating static Pod manifest for "kube-apiserver" [control-plane] Creating static Pod manifest for "kube-controller-manager" [control-plane] Creating static Pod manifest for "kube-scheduler" [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env" [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml" [kubelet-start] Starting the kubelet [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s [kubelet-check] Initial timeout of 40s passed.
Unfortunately, an error has occurred: timed out waiting for the condition
This error is likely caused by:
If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
Additionally, a control plane component may have crashed or exited when started by the container runtime. To troubleshoot, list all containers using your preferred container runtimes CLI. Here is one example how you may list all running Kubernetes containers by using crictl:
Relevant log output