The easiest way to bootstrap a self-hosted High Availability Kubernetes cluster. A fully automated HA k3s etcd install with kube-vip, MetalLB, and more. Build. Destroy. Repeat.
I have 2 clusters
Production - 6 node (3 etcd, 3 worker)
Development - 3 node (3 etcd).
Production is cloud-init backed Ubuntu 22.04 machines with static ip's set in cloud init. When building the cluster the first time its online and runs without issues. When I get a power outage or reboot. all nodes come back online but cluster is not available.
On reboot of development my nodes get ip's
1 - 10.0.99.104/24
2 - 10.0.99.105/24 + 10.0.99.104/32
3 - 10.0.99.106/24
example ip a on the node which kills the cluster
ip a show eno1
2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000
link/ether 00:23:24:9a:99:c4 brd ff:ff:ff:ff:ff:ff
altname enp0s25
inet 10.0.99.105/24 brd 10.0.99.255 scope global eno1
valid_lft forever preferred_lft forever
inet 10.0.99.104/32 scope global eno1
valid_lft forever preferred_lft forever
inet6 fe80::223:24ff:fe9a:99c4/64 scope link
valid_lft forever preferred_lft forever
Expected Behavior
Cluster starts up without issues
Current Behavior
Cluster is unable to start, restart required on all non first node to trigger IP move.
Steps to Reproduce
Build a new k3s cluster based on the playbook
Define static IP address for each node (Cloud-init/Netplan)
Reboot cluster nodes
check ip a (second node gets first nodes IP with /32 subnet
I am wondering if this is a cloud-init | proxmox | k3s issue. I am only seeing this issue on nodes 1/2 of my clusters. It started happening about 2 months ago when I was using Debian 12, saw there was a cloud-init bug about ip's. Switched to Ubuntu 22.04 and seeing the same issues.
I have 2 clusters Production - 6 node (3 etcd, 3 worker) Development - 3 node (3 etcd).
Production is cloud-init backed Ubuntu 22.04 machines with static ip's set in cloud init. When building the cluster the first time its online and runs without issues. When I get a power outage or reboot. all nodes come back online but cluster is not available.
On reboot of development my nodes get ip's 1 - 10.0.99.104/24 2 - 10.0.99.105/24 + 10.0.99.104/32 3 - 10.0.99.106/24
example
ip a
on the node which kills the clusterExpected Behavior
Cluster starts up without issues
Current Behavior
Cluster is unable to start, restart required on all non first node to trigger IP move.
Steps to Reproduce
Context (variables)
Operating system: Ubuntu 22.04 | Debian 12 Hardware: Lenovo Tiny m900 (Production) | Lenovo Tiny M703 (Ubuntu 22.04 Server)
Variables Used
all.yml
Hosts
host.ini
Possible Solution
I am wondering if this is a cloud-init | proxmox | k3s issue. I am only seeing this issue on nodes 1/2 of my clusters. It started happening about 2 months ago when I was using Debian 12, saw there was a cloud-init bug about ip's. Switched to Ubuntu 22.04 and seeing the same issues.