Open mabruzzo opened 1 year ago
I'm guessing it's related to using non-cubical blocks (! (nx == ny == nz)); I didn't test that case, and HDF5 flips the axes around which could be related to the problem. (If blocks are cubical it should work fine, I've been using it at scale with PPM+Grackle+Gravity and restarting with different processor counts with no apparent issues.) I'll look into it.
Checkpoint relies on "order_morton"
, which in turn (currently) requires a cubical blocking since it coarsens to negative levels and assumes it bottoms out at a single block.
Great! Thanks for figuring that out!
I'm going to leave this issue open since this should probably be addressed before the version 2 release.
Last week, I started working on creating new versions of the checkpoint restart tests for the automated testing infrastructure, and I was focusing my attention on the new checkpoint-restart infrastructure.
It appears that there are a number of bugs related to Checkpoint-Restart with Unigrid Hydro Simulations. I have attached 2 simplified examples (1 of them uses PPM and the other uses VL+CT).
For each example, you need to untar the file,
cd
into the untarred directory, modify theCHARMRUN
andENZOE_BIN
paths at the top ofrun_example.sh
, and then you can callbash run_example.sh
.There are a few problems:
Method:check:include_ghosts
set to true, the restart fails with a similar error.Method:check:include_ghosts
set tofalse
. However, restart clearly messes up when it tries to read in field data. You can see this by invokingpython3 check_file.py
(included in each example). This shows that the active zone for the density field on one of the blocks is totally wrong after it gets read back in.As an aside: the example checkpoint-restart files,
input/Checkpoint/test_cosmo-check.in
andinput/Checkpoint/test_cosmo-restart.in
don't actually use a hydro-solver. So I'm not sure if this problem also occurs in that case...@jobordner am I doing anything obviously wrong?