Open slejdops opened 4 years ago
Interested to find out if you've figured out anything. We're seeing the same thing at LSST in what I think is a similar circumstance: dask using Kubespawner in a JupyterHub+user-containers+K8s environment, when we try to do any operation on data that doesn't fit into memory (which is kind of the reason to be using dask).
Any ideas?
Dask workers lose connection with dask scheduler and do not exit when the user pod is terminated.