Closed hcorson-dosch-usgs closed 2 years ago
I'm currently using Jake's grouping approach for the GCM runs -- grouping the 18 runs for each lake into a single branch -- unmerged fork here.
For reference -- Jake's setup of group length, creating another column that is the target group based on the group length and using the grouped table for running another target.
See also how Jake set OpenMP environment variables in the pipeline here. With this approach, you still need to specify the # of workers in your tar_make_clustermq()
call. Jake said he usually sets that roughly equal to the p3_openmp_threads
target (~40-50). If submitting the targets command in a slurm script, he noted you could do the calculation in the slurm script instead.
When we scale up to the whole footprint, we should consider moving away from our current
targets
workflow where we have uniquetargets
branch for each model simulation (18 branches per lake) toward an approach where we are bundling multiple simulations into a single branch.Jordan notes:
I asked Jake "Zwart, Jacob A didn't you run into some limits with # of branches?" and he weighed in: