Azure / aksArc

# Welcome to the Azure Kubernetes Service enabled by Azure Arc (AKS Arc) repo This is where the AKS Arc team will track features and issues with AKS Arc. We will monitor this repo in order to engage with our community and discuss questions, customer scenarios, or feature requests. Checkout our projects tab to see the roadmap for AKS Arc!
MIT License
111 stars 45 forks source link

kubeconfig-clustergroup-management returned a non zero exit code 1 #87

Closed willam-gregoire closed 3 years ago

willam-gregoire commented 3 years ago

In the process of creating an AKS on Azure HCI (nested virtualization) with WAC.

I can´t create the cluster, I get this error

Install-AksHci - C:\Program Files\AksHci\kvactl.exe create --configfile C:\ClusterStorage\aksvol\AksHCI\yaml\appliance.yaml --outfile C:\ClusterStorage\aksvol\AksHCI\kubeconfig-clustergroup-management returned a non zero exit code 1 []Duration: 8 minutes 30 seconds

willam-gregoire commented 3 years ago

and by the way, these failes called byt Install-akshci like kvactl.exe is not existing in the folder.

madhanrm commented 3 years ago

Can we collect logs using get-akshcilogs

madhanrm commented 3 years ago

Also this file please: c:\ProgramData\Kva\kva.log

madhanrm commented 3 years ago

Can you check if your HCI hardware is registered with Azure.

mohaom commented 3 years ago

I'm getting the same issue, i think there is a problem with parsing missing data in appliance.yaml, the error is "Error: strconv.ParseUint: parsing "": invalid syntax", I'm guessing some value are not being filled during the deployment, which result in parsing empty string to an integer.

willam-gregoire commented 3 years ago

Can you check if your HCI hardware is registered with Azure.

I'm using nested virtualization in Azure. Stack HCI is on Hyper-V on an Azure VM.

willam-gregoire commented 3 years ago

Also this file please: c:\ProgramData\Kva\kva.log

These are the contents of the kva.log (which I could only find on my second HCI node:


2021-04-05T10:07:24-07:00 DEBUG Retrieved Config from File Location {"Config File Location": "C:\ClusterStorage\aksvol\AksHCI\yaml\appliance.yaml"} 2021-04-05T10:07:24-07:00 DEBUG Appliance client created with configuration: controlplaneendpoint: "" clustername: aks-management-cluster-1 kubernetesversion: v sshauthorizedkey: ssh-rsa AA (I have removed the key here) lowprivilegekubeconfig: false waitforbootstrapcompletion: true applianceagents: reference: "" chartvariables: [] connectagent: namespace: connectagent kubeconfigsecretname: kubeconfig privatekeysecretname: privatekey configmapname: connectAgentConfig location: westus onboardingagent: subscriptionid: "" tenantid: "" resourcegroup: "" location: "" environment: "" infrastructure: "" clusterapi: configurationendpoint: /var/lib/msk8s/deployment infrastructureprovider: azurestackhci bootstrapprovider: kubeadm controlplaneprovider: kubeadm coreprovider: cluster-api containerregistry: name: ecpacr.azurecr.io username: (I have removed username) password: (I have removed password) networking: controlplanecidr: 10.240.0.0/24 clustercidr: 10.200.0.0/16 podcidr: 10.244.0.0/16 proxy: http: "" https: "" noproxy: localhost,127.0.0.1,.svc,10.0.0.0/8,172.16.0.0/12,192.168.0.0/16 certfilename: "" deploymentmanifest: cloudoperatormanifestpath: C:\ClusterStorage\aksvol\AksHCI\cloud-operator.yaml cnitype: flannel cnimanifestpath: "" certificates: [] azurestackhciprovider: cloudagent: address: ca-0b1a427c-083f-40ce-aa24-ed81b5651ab1.azshci.local port: 55000 authenticationport: 65000 cloudconfigfile: C:\Users\labadmin.wssd\cloudconfig insecure: false appliancevm: imagename: "" vmsize: Standard_A4_v2 loadbalancer: imagename: "" vmsize: Standard_K8S_v1 type: stacked-kube-vip location: MocLocation group: clustergroup virtualnetwork: name: ComputeSwitch type: Transparent macpoolname: "" vlanid: 0 ipaddressprefix: 192.168.0.0/16 gateway: 192.168.0.1 dnsservers:

2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: RetrieveKubeconfig 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Ensure Prerequisites 2021-04-05T10:07:24-07:00 DEBUG Failed to find vippool at location MocLocation 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciling Group 2021-04-05T10:07:24-07:00 DEBUG Reconciling HCI Group {"Group": {"Name":"clustergroup","Location":"MocLocation"}} 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciling Keyvault 2021-04-05T10:07:24-07:00 DEBUG Reconciling HCI Keyvault {"Keyvault": {"Name":"kva","Group":"clustergroup"}} 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciling Virtual Network 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciling Load Balancer 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciling Load Balancer resource 2021-04-05T10:07:24-07:00 DEBUG Reconciling HCI LoadBalancer Configuration {"LoadBalancer": {"Name":"aks-management-cluster-1-load-balancer","Group":"clustergroup","BackendPoolName":"aks-management-cluster-1-backend-pool","VnetName":"ComputeSwitch","FrontendIP":"","FrontendPort":6443,"BackendPort":6443}} 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciling Control Plane Endpoint 2021-04-05T10:07:24-07:00 INFO azurestackhciProvider: Reconciled Control Plane Endpoint: 192.168.0.150

2021-04-05T10:07:34-07:00 DEBUG Retrieved Config from File Location {"Config File Location": "C:\ClusterStorage\aksvol\AksHCI\yaml\appliance.yaml"} 2021-04-05T10:07:34-07:00 DEBUG Appliance client created with configuration: controlplaneendpoint: "" clustername: aks-management-cluster-1 kubernetesversion: v sshauthorizedkey: ssh-rsa (I have removed the key) lowprivilegekubeconfig: false waitforbootstrapcompletion: true applianceagents: reference: "" chartvariables: [] connectagent: namespace: connectagent kubeconfigsecretname: kubeconfig privatekeysecretname: privatekey configmapname: connectAgentConfig location: westus onboardingagent: subscriptionid: "" tenantid: "" resourcegroup: "" location: "" environment: "" infrastructure: "" clusterapi: configurationendpoint: /var/lib/msk8s/deployment infrastructureprovider: azurestackhci bootstrapprovider: kubeadm controlplaneprovider: kubeadm coreprovider: cluster-api containerregistry: name: ecpacr.azurecr.io username: (I have removed user) password: (I have removed password) networking: controlplanecidr: 10.240.0.0/24 clustercidr: 10.200.0.0/16 podcidr: 10.244.0.0/16 proxy: http: "" https: "" noproxy: localhost,127.0.0.1,.svc,10.0.0.0/8,172.16.0.0/12,192.168.0.0/16 certfilename: "" deploymentmanifest: cloudoperatormanifestpath: C:\ClusterStorage\aksvol\AksHCI\cloud-operator.yaml cnitype: flannel cnimanifestpath: "" certificates: [] azurestackhciprovider: cloudagent: address: ca-0b1a427c-083f-40ce-aa24-ed81b5651ab1.azshci.local port: 55000 authenticationport: 65000 cloudconfigfile: C:\Users\labadmin.wssd\cloudconfig insecure: false appliancevm: imagename: "" vmsize: Standard_A4_v2 loadbalancer: imagename: "" vmsize: Standard_K8S_v1 type: stacked-kube-vip location: MocLocation group: clustergroup virtualnetwork: name: ComputeSwitch type: Transparent macpoolname: "" vlanid: 0 ipaddressprefix: 192.168.0.0/16 gateway: 192.168.0.1 dnsservers:

2021-04-05T10:07:34-07:00 INFO azurestackhciProvider: Delete 2021-04-05T10:07:34-07:00 INFO azurestackhciProvider: Deleting Identity 2021-04-05T10:07:34-07:00 DEBUG Deleting Identity {"Identity": {"Name":"CAPI","Group":"clustergroup"}} 2021-04-05T10:07:34-07:00 INFO azurestackhciProvider: Deleting Group 2021-04-05T10:07:34-07:00 DEBUG Deleting Group {"Group": {"Name":"clustergroup","Location":"MocLocation"}}

willam-gregoire commented 3 years ago

Can we collect logs using get-akshcilogs

hereby the logs, thanks in advance for helpoing

akshcilogs.zip

madhanrm commented 3 years ago

Can you please try the GA build?

mkostersitz commented 3 years ago

@willam-gregoire were you able to verify that the GA build resolves the issue? If not please re-open the issue