Closed untraceablez closed 10 months ago
I notice your inventory IPs are on a different range than metal_lb_ip_range and apiserver_endpoint. Is it possible your issue is related to tis discussion?
https://github.com/techno-tim/k3s-ansible/discussions/158#discussioncomment-4082525
I notice your inventory IPs are on a different range than metal_lb_ip_range and apiserver_endpoint. Is it possible your issue is related to tis discussion?
I'll be honest, I thought that those were on addresses on the cluster network, didn't realize those needed to be modified to match my local LAN. I'll go ahead and update those IPs and re-run site.yml and report back.
My bad, hit close instead of just comment. 😅
@redmetablue was correct, I needed to update my metallb IP range and api-server IP to match my local LAN. Re-ran the site.yml and copied the kube config and confirmed connectivity to the cluster.
When running with a fresh copy of the repo, only modifications being k3s-token and the hosts.ini file, the playbook hangs at the k3s_agent: Enable and check K3S service task, and never completes.
Expected Behavior
The playbook should run all the way through and join the worker nodes to the cluster.
Current Behavior
After modifying the hosts.ini file to point to my 3 control VMs on Proxmox (non-LXC) and my 6 raspberry pi 4Bs, I generated an alphanumeric token (truly alphanumeric, 0 symbols of any kind.)
I then ran the following command:
ansible-playbook site.yml -i inventory/untraceablez/hosts.ini
and the playbook will run through most of the tasks until arriving at the k3s_agent: Enable and check K3s service] task. Here, it will hang in perpetuity until manually killed.Steps to Reproduce
ansible-playbook site.yml -i inventory/mycluster/hosts.ini
Context (variables)
Operating system:
Hardware:
3 VMs have the following:
6 Raspberry Pi 4Bs are all 4GB model except for 1 8GB model.
Variables Used
all.yml
Hosts
host.ini
Results of
journalctl -uxe k3s-node.service
Possible Solution
I have run the reset.yml as well and rebooted the pis and the VMs, and tried running the site.yml again, but to no avail, I get the same result each time. I suspect there's a bug somewhere that's causing the worker nodes to point to localhost:6444 instead of the control nodes on port 6444, at least that's the error being indicated by the
journalctl
output.