Closed zjfroot closed 5 years ago
👋 Thanks for opening your first issue here! If you're reporting a 🐞 bug, please make sure you include steps to reproduce it.
Hi @zjfroot, are you intentionally putting your master nodes into a subnet outside your --cluster-cidr
IP address range? @palma21 to your knowledge is that a viable network configuration?
I think clusterSubnet range might be the problem (would using 10.11.0.0/16 work?), but this should be a possible config
CC @khenidak @juan-lee for thoughts
@jackfrancis When you say --cluster-cidr
, do you mean the one 10.11.0.0/19
in clusterSubnet
of kubernetesConfig
?
In our case, the master and worker nodes subnets are not next to each other, if we think cidr wise. How can we specify a clusterSubnet that covers both?
Also from the cluster definition doc, the clusterSubnet is:
The IP subnet used for allocating IP addresses for pod network interfaces.
I assumed it is only for the worker nodes, that is why the cidr range for clusterSubnet in my example is 10.11.0.0/19
, which is the worker nodes subnet.
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.
Is this a request for help?:
Yes
Is this an ISSUE or FEATURE REQUEST? (choose one):
ISSUE
What version of aks-engine?:
v0.32.3
Kubernetes version:
1.12
What happened: We were trying to deploy a k8s cluster with aks-engine to our existing vnet.
We have two subnets:
worker nodes subnet has a cidr of
10.11.0.0/19
master nodes subnet has a cidr of10.11.255.0/24
The vnet has a cidr
10.11.0.0/16
We have a following cluster config for aks-engine (VMSS, Multi Zone, Custom Vnet/Subnets):
We tried a few times, it always fail to deploy the master vmss. After checking the
/var/log/azure/cluster-provision.log
of the master nodes, it turns outetcd
couldn't start.It looks like
etcd
has following command line parameters:It seems like it gives a wrong --init-cluster parameter. It passes
https://10.11.0.4:2380
,https://10.11.0.35:2380
andhttps://10.11.0.66:2380
, but they are actually worker nodes IPs, not master nodes.If I understand correctly,
--init-cluster
parameter ofetcd
should contain the master nodes, not the worker nodes.What you expected to happen: We expect masters get correctly provisioned and etcd up and running.
How to reproduce it (as minimally and precisely as possible): Try to use a similar json config as the example above, with the same subnets ranges, deploy it with aks-engine.
Anything else we need to know: Maybe our config is not really correct? We might need to specify something extra, so that aks-engine can pick up the correct master IPs when starting
etcd
?