Closed jessjaco closed 1 year ago
Think I resolved this.
Solution was to just install gcc on the planetary computer image before installing dea_tools, so hdstats wouldn't error out (Sachin's solution above worked, but using the "official" dea_tools, which doesn't have hdstats).
My only complaint is this makes the docker image rather large. I'm considering maintaining the use of the vanilla planetary computer image when running through kbatch for step 2 above.
Here's the current workflow (putting here rather than readme as it may change):
Currently 1 is done locally using he dea-australia/coastlines based docker image. It can be run remotely, as long as the tidal model data is available. It hasn't yet been run in the notebook or kbatch environment.
2 is done using kbatch and the planetary computer docker image (as noted in the yaml) with the addition of dep_tools (which is installed in src/run_on_pc.sh).
3 is done using kbatch and the dea-australia/coastlines based docker image.
We will need to decide if we want to maintain these different dev environments or try to use a single one. Here are confounding issues:
The tidal models are big, so they make large docker images. In addition, output can not be written to blob storage directly, so kbatch is a no-go. (This could be solved by some manual-uploading perhaps). Nevertheless, we could still standardize on the docker image with the requirement that the tidal models are mounted in a directory.
I have not been able to install dea_tools in the MSPC environment / atop the docker image, due to dependencies on GCC. However, @sopac has told me:
It'd be nice to use the MSPC instance as much as possible, with as little additional installations as possible, so code may function in the notebook environment if possible.