This folder contains metadata files for the models submitting to the FluSight forecasting collaboration. The specification for these files has been adapted to be consistent with model metadata guidelines in the hubverse documentation.
Each model is required to have metadata in yaml format, e.g. see this metadata file.
These instructions provide detail about the data format as well as validation that you can do prior to a pull request with a metadata file.
This section describes each of the variables (keys) in the yaml document. Please order the variables in this order.
The name of your team that is less than 50 characters.
The name of your team that is less than 16 characters.
The name of your model that is less than 50 characters.
An abbreviated name for your model that is less than 16 alphanumeric characters.
A list of all individuals involved in the forecasting effort. A names, affiliations, and email address is required for each contributor. Individuals may also include an optional orcid identifiers. All email addresses provided will be added to an email distribution list for model contributors.
The syntax of this field should be
model_contributors: [
{
"name": "Modeler Name 1",
"affiliation": "Institution Name 1",
"email": "[email protected]",
"orcid": "1234-1234-1234-1234"
},
{
"name": "Modeler Name 2",
"affiliation": "Institution Name 2",
"email": "[email protected]",
"orcid": "1234-1234-1234-1234"
}
]
One of the accepted licenses.
We encourage teams to submit as a "cc-by-4.0" to allow the broadest possible uses including private vaccine production (which would be excluded by the "cc-by-nc-4.0" license).
A team-specified boolean indicator (true
or false
) for whether the model should be considered eligible for inclusion in a Hub ensemble and public visualization. In general, a team may specify up to two models as a designated_model for inclusion. Models which have a designated_model value of 'False' will still be included in internal forecasting hub evaluations. Please send an email to [email protected] if you would like to submit more than two methodologically distinct models for inclusion in the ensemble. This email should provide evidence of out of sample performance assessment and/or a statement of differences between the proposed forecast models to justify inclusion in the ensemble.
List or description of the data sources used to inform the model. Particularly those used beyond the target data of confirmed influenza hospital admissions.
A brief description of your forecasting methodology that is less than 200 characters.
A full description of the methods used by this model. Among other details, this should include whether spatial correlation is considered and how the model accounts for uncertainty. If the model is modified, this field can also be used to provide the date of the modification and a description of the change.
A boolean value (true
or false
) that indicates whether a model is an ensemble of any separate component models.
A boolean value (true
or false
) that indicates whether a model is an ensemble specifically of other models submited to the FluSight forecasting hub.
A url to a website that has additional data about your model. We encourage teams to submit the most user-friendly version of your model, e.g. a dashboard, or similar, that displays your model forecasts.
A brief description of the methodology used to generate sample trajectories, as well as the total number of samples generated if they exceed 100.
An identifier of the version of the model
A github (or similar) repository url containing code for the model.
One or more citations to manuscripts or preprints with additional model details. For example, "Gibson GC , Reich NG , Sheldon D. Real-time mechanistic bayesian forecasts of Covid-19 mortality. medRxiv. 2020. https://doi.org/10.1101/2020.12.22.20248736".
Any information about funding source(s) for the team or members of the team that would be natural to include on any resulting FluSight publications. For example, "National Institutes of General Medical Sciences (R01GM123456). The content is solely the responsibility of the authors and does not necessarily represent the official views of NIGMS."
Optionally, you may validate a model metadata file locally before submitting it to the hub in a pull request. Note that this is not required, since the validations will also run on the pull request. To run the validations locally, follow these steps:
- Create a fork of the
FluSight-forecast-hub
repository and then clone the fork to your computer. - Create a draft of the model metadata file for your model and place it in the
model-metadata
folder of this clone. - Install the hubValidations package for R by running the following command from within an R session:
remotes::install_github("Infectious-Disease-Modeling-Hubs/hubValidations")
- Validate your draft metadata file by running the following command in an R session:
hubValidations::validate_model_metadata(
hub_path="<path to your clone of the hub repository>",
file_path="<name of your metadata file>")
For example, if your working directory is the root of the hub repository, you can use a command similar to the following:
hubValidations::validate_model_metadata(hub_path=".", file_path="UMass-trends_ensemble.yml")
If all is well, you should see output similar to the following:
✔ model-metadata-schema.json: File exists at path hub-config/model-metadata-schema.json.
✔ UMass-trends_ensemble.yml: File exists at path model-metadata/UMass-trends_ensemble.yml.
✔ UMass-trends_ensemble.yml: Metadata file extension is "yml" or "yaml".
✔ UMass-trends_ensemble.yml: Metadata file directory name matches "model-metadata".
✔ UMass-trends_ensemble.yml: Metadata file contents are consistent with schema specifications.
✔ UMass-trends_ensemble.yml: Metadata file name matches the `model_id` specified within the metadata file.
If there are any errors, you will see a message describing the problem.