Closed clarkpede closed 5 years ago
@clarkpede : Thank you for the detailed post, and your hunch is likely correct. Unfortunately, the existing periodic BC implementation has some limitations due to how closely it is coupled to the old mesh partitioning routines during setup.
Those original partitioning routines had become difficult to maintain or expand (very hard to make quick fixes), so they were rewritten from scratch in PR #513. Now, the periodic BC is also being rewritten cleanly (hopefully for the last time :) ). A prototype can be seen in feature_periodic that is already working for Euler problems, and the rest is in progress now. I am aware that several folks are in need of this, but know that a new version is coming.
Alright. Thanks for the update. I'll wait to close this issue until the periodic BCs are fixed.
Resolved by #652 - @clarkpede : if you still find any issues with the new implementation, do not hesitate to open a new issue.
I've recently run into a problem with periodic geometry when I run a RANS problem on 16 cores or more (256+ MPI tasks). While initializing the Jacobian structure for the turbulence model, I run into one of two errors, depending on the core count.
The first error results in the following error message:
The second error just leads to the solver hanging indefinitely at the
Initialize Jacobian structure (SA model)
step. I'm guessing that an MPI send/receive is left dangling.I have not seen these problems at lower core counts (2-4 cores with 2-32 MPI tasks).
The errors seem to be tied to the way the periodic send/receives are set up. If I change the periodic boundaries to far-field boundaries, the error vanishes.
I've also done a lot of work to weed out possible causes:
SU2_MSH
and thesu2perio
Fortran tool.I've got a minimal example that you can use to test this for yourself, in the attached files. It should be self-explanatory.
MPI_Failure_Example.tar.gz