Currently the dask cluster is started within the main processing logic (wrapper).
I think it would be advantageous to decouple these steps for several reasons:
This might help down the road when we want to use other ways of spinning up a cluster than the coiled API
If the main prefect flow fails, this might be a more robust way of 'catching' the cluster and closing it after both success or failure see #10
This could also help in executing the logic manually (e.g. on one of the pangeo deployments/local machine). One might be able to spin up a cluster and then just provide the needed information to the execution step, without repeatedly opening/closing clusters.
I would need to find a way to pipe the cluster object (or just the address?) into the main processing step
Currently the dask cluster is started within the main processing logic (
wrapper
).I think it would be advantageous to decouple these steps for several reasons:
I would need to find a way to pipe the cluster object (or just the address?) into the main processing step