larsbuntemeyer / notebooks

MIT License
0 stars 0 forks source link

Notebooks for prototyping

nbviewer Binder

These notebooks contain some prototype code for working with large CMIP or CORDEX datasets. Most notebooks are not documented very well but are used only to try out some coding concepts. Most notebooks require data access to the DKRZ data pool using their intake catalogs. However, in principle you should also be able to run CMIP6 notebooks if you switch to the PANGEO datastore url.

Requirements

The most important packages that will make your work with heavy cmip datasets in ipython easier are:

You can use the environment.yaml file to easily install all dependencies, e.g.

conda env create -f environment.yaml

This should create a basic kernel called notebooks that you can use to run most of the notebooks in this repo. However, sometimes you might need to install additional packages if they are missing. Here is a documentation of how to install the kernel in the DKRZ Jupyterhub.

Server settings - DKRZ

Here are my recommended server settings for the JupyterLab at DKRZ.

Server Settings DKRZ

Note, that on the shared partition at DKRZ, you don't have internet access (actually, you only have internet access on the prepost partition). However, some packages might need internet access to cache data (e.g., regionmask regionmask.defined_regions, or xarray.tutorial). You can trigger those downloads on the login node. Once, the data is cached in you HOME directory, you can use it also on the shared partition. Also py-cordex needs some tables with domain definitions to download. You can trigger that e.g., using

import cordex as cx
cx.domains.table