After the model training split and the train-test split, we will have dask data bags with data-frames. These DataFrames can also be used in DNN trainings, as in the example python script.
Currently when training for large number of gridcells, we parallel the workflow by submitting each grid cell as a separate SLURM job, as this example. We can improve this by using Dask SLURMCluster instead of directly submit via SLURM. In this way one can reserve a continuous time on the HPC computation resources and get less chance to be interrupted by a SLURM queue.
Goal: have 1) an example Jupyter NB and 2) and example python script to:
load example data as Xarray following this example
After the model training split and the train-test split, we will have dask data bags with data-frames. These DataFrames can also be used in DNN trainings, as in the example python script.
Currently when training for large number of gridcells, we parallel the workflow by submitting each grid cell as a separate SLURM job, as this example. We can improve this by using Dask SLURMCluster instead of directly submit via SLURM. In this way one can reserve a continuous time on the HPC computation resources and get less chance to be interrupted by a SLURM queue.
Goal: have 1) an example Jupyter NB and 2) and example python script to: