-
Notifications
You must be signed in to change notification settings - Fork 22
Refreshing the ensemble at the beginning of a season
Nicholas G Reich edited this page Oct 25, 2019
·
12 revisions
This page documents the steps for refreshing the ensemble at the beginning of a new season.
- Update the participation guidelines, as appropriate.
- By reaching out to teams, establish a working sense of what models will be submitted.
- Before making any changes to the repository, create a tagged version and/or branch marking files that represent the end of the last season. This makes it easy to switch back to files as they were at a certain point in time.
- For models that were used last season and will be used again use
./scripts/move-last-years-files.R
to migrate files from the./model-forecasts/real-time-component-models
to the./model-forecasts/component-models
directory. - Update the "ground truth" data in the
./scores/target-multivals.csv
file by running./scripts/calculate-targets.R
. - Recalculate all scores for component models in the
master
branch using./scripts/generate-scores.js
. See Code Documentation for more details. - Update
model-forecasts/component-models/complete-modelids.csv
andmodel-forecasts/component-models/model-id-map.csv
. Currently this is done usingscripts/generate-id-mappings.js
(As of Oct 2019, we are not sure if any piece of visualization infrastructure relies on the definition of a model being "complete" as used by these scripts. We don't think so.) - Establish a list of candidate ensemble models to test for the given season. Ideally, write this down on GitHub somewhere, so we have a time-stamp of the "data analysis plan", including the method we will use to choose an ensemble.
- Generate weights for all candidate ensemble models. For the 2019/2020 season, this requires using a combination of the
scripts/calculate-weights.R
and the makefile inscripts/static_adaptive_ensemble/
. (According to Tom, usingmake rebuild staticAdaptive
should re-estimate the weight file.) - Confirm that ensemble weights for all ensemble specifications are saved and stored in GitHub.
- Generate leave-one-season-out cross-validation forecast files for all ensemble models using this the set of ensemble models defined above. This will require minor updates to and running of the
scripts/make-cv-ensemble-forecast-files.R
file. - Re-generate scores (see scoring instructions above) so that the updated ensemble forecasts are now scored as well.
- Summarize scores and pick one ensemble to submit.
- Re-organize the
real-time-ensemble-models
folder to make sure that there is a folder for each model. Also, put last year's files into thereal-time-ensemble-models\archive-20xx-20xx
folder. - Copy the files from last year's
submissions
folder into thesubmissions-archive
folder. - Wait for the season to start and once it does, use the Creating ensemble submission guidelines to create weekly submissions.