Right now joblib does just fine for multi-core parallelization but for distributed computing might be good to check out Dask: https://docs.dask.org/en/latest/
We'll probably need this if we want to do >10,000,000 particles.
Will also need something like Dask for when the datasets can't fit in memory anymore. Easy for us to split datasets but might be nicer to switch to Dask?
Right now joblib does just fine for multi-core parallelization but for distributed computing might be good to check out Dask: https://docs.dask.org/en/latest/
We'll probably need this if we want to do >10,000,000 particles.