Closed bradbeam closed 6 years ago
@abrarshivani as we discussed offline, this breaking upgrade needs to be documented. I am assigning this issue to you as well, to help with proper documentation.
Are there any specifics you're able to share?
We are seeing this on RHEL 7.4 with Kube 1.9.0, I am confused by the current status of this bug, can we get an update? Thanks!
@bradbeam As you mentioned,
but on a fresh install of 1.9.6
Did you found this error on fresh install or upgrade? And if you have upgrade the cluster, did you upgrade kubernetes cluster in the following way?
1. Create k8s cluster of version v1.8.2.
2. Deploy an app with persistent volume (vSphere volume).
3. Update k8s master to v1.9.6.
4. Delete v1.8.2 node (or Drain v1.8.2 node)
5. Upgrade node to v1.9.6.
And detach is failing after step 4
The errors were seen on an upgrade. Not on a fresh install.
Steps 1/2/3/5 are correct. Step 4 was not done -- the node was not drained.
Detach is failing after step 5.
@bradbeam This is known issue and has been fixed in v1.10.2. Here are the issues:
@bradbeam Can we close this issue now?
Yeah, we haven't been able to do any more testing on it; will reopen if we find anything else. Thanks.
Is this a BUG REPORT or FEATURE REQUEST?: Bug Report
/kind bug
What happened: After upgrading our cluster from 1.8.2 to 1.9.6, persistent volumes via vsphere are no longer getting detached.
What you expected to happen: Persistent volumes should get detached properly
How to reproduce it (as minimally and precisely as possible): Deploy a cluster on 1.8.2 Deploy an app w/ persistent volume (
helm install grafana
orhelm install prometheus
are good candidates ) Upgrade cluster to 1.9.6 Delete pod / Cordon node if necessary to get pod to get rescheduled on a new node.Anything else we need to know?:
Environment:
Kubernetes version (use
kubectl version
): 1.8.2:1.9.6:
Cloud provider or hardware configuration: virtual machines w/ vsphere cloud provider
OS (e.g. from /etc/os-release):
Kernel (e.g.
uname -a
):Install tools: kubespray
Others:
After the cluster was upgraded, you could manually detach the volume with the following and it would properly get attached to the new node:
I wasn't able to fix this on the existing cluster, but after deploying 1.9.6 fresh I compared some of the details and one of the things that stood out was looking at the volume attachment info on each of the kubernetes nodes.
on 1.8.2,
- kubernetes.io/vsphere-volume/[vsanDatastore] kubevols/kubernetes-dynamic-pvc-365b444a-2e5a-11e8-aefb-005056a256f5.vmdk
but on a fresh install of 1.9.6,- kubernetes.io/vsphere-volume/[vsanDatastore] 8abe845a-146b-f38d-c08d-801844dfb6b6/kubernetes-dynamic-pvc-365b444a-2e5a-11e8-aefb-005056a256f5.vmdk
After the cluster was upgraded from 1.8.2 to 1.9.6, the
kubevols
path was still referenced in the volume above.Here are some of the logs ( note they are from two different occurances ) kubelet logs:
controller-manager logs: