Open ktindiana opened 6 months ago
The looks like a kind of post-analysis study, driven by the regular per-model sphinx outputs. I am imagining a script that reads the output files and generates new, aggregate forecast jsons that can be run through sphinx in a second pass.
For the SEP Scoreboard, how can we look at the models as an ensemble to better inform how operators use them. Do the models together give us more reliable information than a model alone? For example:
Are there features, reporting, or a workflow we can add to SPHINX to evaluate this?