Closed ktravis213 closed 1 year ago
Thanks for writing @ktravis213. You might need to update #SBATCH --mem=85000
, as you might be running out of memory for the job. You might need 95000 or 100000.
Also do you have the slurm*.out log? That might give you an idea of what the run failure was.
Thanks @yantosca. It seems like it has enough memory - here is my .out file. That's why I thought it might have something to do with the stack size?
Job Resource Usage Summary for 64787662
Job Ended : Mon Nov 7 09:15:22 EST 2022
Partition : compute
Head Node : borgu208
Charged to : s2199
Estimated SBUs : 14.91
Total Mem Requested / Allocated : 83.00G / 83.00G
Max Real Mem Used (on any node) : 58.43G
Max Pct Memory Used : 70.39%
Walltime Requested : 20:00:00
Walltime Used : 18:25:06
Pct Walltime Used : 92.09%
CPUs Requested / Allocated : 24 / 28
Total CPU-Time Allocated : 21-11:42:48 (Walltime Used * #CPUs)
CPU Utilization of __last time cmd only__:
User CPU : 00:00.012 of 18-10:02:00
System CPU : 00:00.008 of 18-10:02:00
Pct CPU Utilization : 0%
HI @ktravis213. Can you check your HEMCO_Config.rc.gmao_metfields
to see if it's using the global 0.25x0.3125 or pre-cropped "AS" files? I discovered this weekend that when selecting one of the supported met fields, we no longer seem to be adding in the nested domain token in the met field path. I wonder if this is causing your issues. At the very least it will drastically slow down your simulations if using the global high-resolution files.
Thanks @msulprizio. I am using the cropped "CH" files. Maybe I will have to ask my IT folks if there is no obvious solution.
This issue has been automatically marked as stale because it has not had recent activity. If there are no updates within 7 days it will be closed. You can add the "never stale" tag to prevent the Stale bot from closing this issue.
Closing due to inactivity
What institution are you from?
NASA LaRC
Description of the problem
I am running v14.4.0 at 0.25x0.3125 over Asia, using boundary conditions at 2x25. I have not been able to get past LINOZ, see the end of my log file below. The run hangs there until the simulation runs out of time.
Description of troubleshooting performed
I have increased KMP_STACKSIZE from 500m to 800m, set -c to 24. I am not sure what else to do. My nested runs for 13.4.0 ran fine with this run script below.
GEOS-Chem version
14.0.0
Description of modifications
None
Log files
Photolysis is activated -- rates computed by FAST-JX
Interpolating Linoz fields for apr
###############################################################################
Software versions