Open jswhit opened 2 weeks ago
Will work on this, thanks for reporting @jswhit !
Alrighty, so the gdasstage_ic
job picks up your ICs "correctly" using m_prefix=20210323.210000
, which is based on model_start_date_current_cycle
minus 3hrs because DOIAU=YES
. The gdasfcst_seg0
job then initially sets model_start_date_current_cycle
to the same time (from a log from my reproduction of the issue):
1453 + forecast_predet.sh[94]: model_start_date_current_cycle=2021032321
...but later on it gets set to the cycle that's running because the experiment is cold-starting, which means IAU is off and the model start date would not be 3hrs earlier:
1907 + forecast_det.sh[27]: model_start_date_current_cycle=2021032400
That happens in forecast_det.sh
here: https://github.com/NOAA-EMC/global-workflow/blob/develop/ush/forecast_det.sh#L27
Based on the above, either:
Pretty sure option 2 is what is needed. Thoughts?
I believe 2 was how things worked before
FWIW, this fixes my particular case (cold start for atmosphere, warm starts for ocean/ice)
diff --git a/ush/forecast_postdet.sh b/ush/forecast_postdet.sh
index 8af90549..2adf1aa1 100755
--- a/ush/forecast_postdet.sh
+++ b/ush/forecast_postdet.sh
@@ -415,7 +415,8 @@ MOM6_postdet() {
restart_date="${RERUN_DATE}"
else # "${RERUN}" == "NO"
restart_dir="${COMIN_OCEAN_RESTART_PREV}"
- restart_date="${model_start_date_current_cycle}"
+ #restart_date="${model_start_date_current_cycle}"
+ restart_date="${current_cycle_begin}"
fi
# Copy MOM6 ICs
@@ -565,7 +566,8 @@ CICE_postdet() {
seconds=$(to_seconds "${restart_date:8:2}0000") # convert HHMMSS to seconds
cice_restart_file="${DATArestart}/CICE_RESTART/cice_model.res.${restart_date:0:4}-${restart_date:4:2}-${restart_date:6:2}-${seconds}.nc"
else # "${RERUN}" == "NO"
- restart_date="${model_start_date_current_cycle}"
+ #restart_date="${model_start_date_current_cycle}"
+ restart_date="${current_cycle_begin}"
cice_restart_file="${COMIN_ICE_RESTART_PREV}/${restart_date:0:8}.${restart_date:8:2}0000.cice_model.res.nc"
if [[ "${DO_JEDIOCNVAR:-NO}" == "YES" ]]; then
cice_restart_file="${COMIN_ICE_ANALYSIS}/${restart_date:0:8}.${restart_date:8:2}0000.cice_model_anl.res.nc"
Good to know, thanks @jswhit ! Didn't get a chance to look deep into this yesterday, will aim to today.
@jswhit I see now that the staging needed adjusting. When I tested it it worked but I see now that you had symlinks from the 20210323.210000.MOM.res*.nc
files to the correct 20210324.000000.MOM.res*.nc
files so it was a false success for me. I updated the staging yaml files to fix the issue in issue #2890 and it seems to have fixed things for the staging job in this case too.
I just ran the gdasstage_ic
and gdasfcst_seg0
job for your case and they worked. Would you mind copying the yaml from my clone on Hera (/scratch1/NCEPDEV/global/Kate.Friedman/git/develop_fork/parm/stage
) into your clone's parm/stage
folder and try the staging and fcst jobs for your case? Let me know if it works as anticipated. Thanks!
sorry for the late reply @KateFriedman-NOAA. When I copy your parm/stage directory, the staging job seems to run fine but I'm getting this error in gdasfcst_seg0.log. I think it's probably unrelated to this issue, but I don't seem to have a sorc/upp.fd/parm/gfs directory (which parm/post/gfs is symlinked to).
+ forecast_predet.sh[544]: /bin/cp -p /scratch2/BMC/gsienkf/whitaker/global-workflow-jswhit2/parm/post/gfs/postxconfig-NT-gfs-two.txt /scratch1/NCEPDEV/stmp2/Jeffrey.S.Whitaker/RUNDIRS/C96coupled3dvar_test/gdas.2021032400/gdasfcst.2021032400/fcst.998793/postxconfig-NT.txt
/bin/cp: cannot stat '/scratch2/BMC/gsienkf/whitaker/global-workflow-jswhit2/parm/post/gfs/postxconfig-NT-gfs-two.txt': No such file or directory
@jswhit There was an update to the system related to UPP and its parm txt files so you'll either want to make a fresh clone or do a submodule update command (and then link script) in your clone to remedy the issue.
Okay, got the submodules updated correctly. Now I get this error
forecast_postdet.sh[440]: /bin/cp -p /scratch2/BMC/gsienkf/whitaker/GWTESTS/COMROOT/C96coupled3dvar_test/gdas.20210323/18//model/ocean/restart/20210324.000000.MOM.res.nc /scratch1/NCEPDEV/stmp2/Jeffrey.S.Whitaker/RUNDIRS/C96coupled3dvar_test/gdas.2021032400/gdasfcst.2021032400/fcst.1355321/INPUT/MOM.res.nc
/bin/cp: cannot stat '/scratch2/BMC/gsienkf/whitaker/GWTESTS/COMROOT/C96coupled3dvar_test/gdas.20210323/18//model/ocean/restart/20210324.000000.MOM.res.nc': No such file or directory
+ forecast_postdet.sh[441]: echo 'FATAL ERROR: Unable to copy MOM6 IC, ABORT!'
FATAL ERROR: Unable to copy MOM6 IC, ABORT!
+ forecast_postdet.sh[441]: exit 1
in gdasfcst_seg0 (see /scratch2/BMC/gsienkf/whitaker/GWTESTS/COMROOT/C96coupled3dvar_test/logs/2021032400)
The file that was staged is 20210323.210000.MOM.res.nc
, not 20210324.000000.MOM.res.nc
What is wrong?
for a coupled 3dVar cycling experiment with cold starts for 2021032400
gdasstage_ic
stages oceans restarts with m_prefix = 20210323.180000, butgdasfcst_seg0
then looks for restarts with m_prefix = 20210324.0000. Here's the error fromgdasfcst_sego
:and the relevant output from
gdasstage_ic
:What should have happened?
gdasstage_ic stages ocean restarts with the same filenames expected by gdasfcst_seg0 for cold starts.
What machines are impacted?
Hera
Steps to reproduce
run a cycled 3DVar coupled atm/ocean experiment. I ran
from /scratch2/BMC/gsienkf/whitaker/global-workflow-jswhit, with the patch from issue #2864 applied so that
gdasstage_ic
does not fail.Additional information
N/A
Do you have a proposed solution?
no