Open todeb opened 5 days ago
Did you reboot without draining the node? The nvme kernel initiator will keep trying to connect for some time. If you really want to do this, then you need to do something like this on the host:
for dev in /sys/class/nvme/*/ctrl_loss_tmo; do echo 10 | sudo tee -a $dev; done
But I suggest you drain the node first before rebooting otherwise the filesystems may not unmount gracefully and potentially result in data loss.
do you mean kubectl-mayastor drain command or usual kubectl drain to evict pods? Im not draining node because im using single replica and have apps replicated by their own, so don't want that they will be scheduled on other nodes.
do you mean kubectl-mayastor drain command or usual kubectl drain to evict pods?
I mean the usual kubectl drain to evict the pods using the mayastor volumes to other nodes.
Im not draining node because im using single replica and have apps replicated by their own, so don't want that they will be scheduled on other nodes.
Ah I see... are you using the GracefulNodeShutdown
feature?
IIRC that should ensure the apps are stopped gracefully, but IIRC the volumeattachment
and connections would still remain? CC @Abhinandan-Purkait
Ah I see... are you using the
GracefulNodeShutdown
feature? IIRC that should ensure the apps are stopped gracefully, but IIRC thevolumeattachment
and connections would still remain? CC @Abhinandan-Purkait
Yes, that's true. IIRC we need to remove the attachments manually.
not using it.
Describe the bug after initiating reboot of k8s node with mayastor, OS in 8 mins was trying to reconnect nvme.
To Reproduce init reboot of OS
Expected behavior I dont know how these nvmes are handled but imo nvme should not lock system reboot, especially if they failing in connection
Screenshots
OS info (please complete the following information):
Additional context Add any other context about the problem here.