Who?

This guide is for vetted community members interested in helping to review an EMD submission. It explains how submissions reach you, what has already been checked, what needs your judgement, and how to record your decision.


Overview

EMD entries are submitted through online forms. Before a submission reaches you it passes through automated validation. Your role is to assess the scientific content that machines cannot judge — whether the description is accurate, the component classifications are sensible, and the entry is not a duplicate.


How submissions reach you

Step 1: Submission

A modelling group fills in a structured form. No programming or Git knowledge is needed.

flowchart LR
    A(["Modelling group\ncompletes form"]) --> B["Submission created\nwith structured fields"]

    style A fill:#f0f4f8,stroke:#90b4d6,color:#1e293b
    style B fill:#fafafa,stroke:#b0bec5,color:#1e293b

Step 2: Automated validation

The submission is checked against the EMD schema. If anything is missing or incorrectly formatted, the submitter is asked to correct it. Only valid submissions proceed further.

flowchart LR
    B["Submission"] --> C{"Automated\nvalidation"}
    C -->|Pass| D["Submission\nqueued for review"]
    C -->|Fail| E["Sent back to\nsubmitter with\nspecific errors"]

    style C fill:#f3edf5,stroke:#b39dba,color:#1e293b
    style D fill:#eef5ee,stroke:#8cb88c,color:#1e293b
    style E fill:#faf0f0,stroke:#d19a9a,color:#1e293b
    style B fill:#fafafa,stroke:#b0bec5,color:#1e293b

Step 3: Review assignment

A reviewer is selected based on scientific domain, independence from the submitting group, and current workload. The reviewer (you) is notified via the pull request.

flowchart LR
    D["Valid submission"] --> F{"Reviewer\nassigned"}
    F --> G(["You receive\nthe review request"])

    style D fill:#fafafa,stroke:#b0bec5,color:#1e293b
    style F fill:#f3edf5,stroke:#b39dba,color:#1e293b
    style G fill:#f0f4f8,stroke:#90b4d6,color:#1e293b

Step 4: Your decision

You review the flagged items and choose one of three actions: approve, request changes, or reject.

flowchart LR
    G(["You review\nthe submission"]) --> K{"Decision"}
    K -->|Approve| L["Entry registered\nin the EMD"]
    K -->|Request changes| N["Submitter\nupdates and\nresubmits to you"]
    K -->|Reject| M["Submission closed\nwith explanation"]

    style G fill:#f0f4f8,stroke:#90b4d6,color:#1e293b
    style K fill:#fdf5eb,stroke:#d4b07a,color:#1e293b
    style L fill:#eef5ee,stroke:#8cb88c,color:#1e293b
    style M fill:#faf0f0,stroke:#d19a9a,color:#1e293b
    style N fill:#fdf5eb,stroke:#d4b07a,color:#1e293b

What has already been checked

By the time a submission reaches you, the following have been verified automatically. You do not need to re-check these.

Check What was tested
Required fields All mandatory fields are present and correctly formatted
Unique identifier The proposed ID does not clash with any existing entry
Cross-references Any referenced grids, families, or components exist in the registry
Published references DOI links are valid and reachable
Controlled vocabularies Values such as grid types, calendar types, and arrangements match the accepted lists
Numeric plausibility Values such as layer counts, resolutions, and release years fall within expected ranges

What needs your judgement

The review report on the pull request highlights the items below. These require scientific expertise that automated checks cannot provide.

Description quality — Is the scientific description accurate and informative? For models, does it mention all component types (dynamically simulated, prescribed, and omitted)? Would a researcher unfamiliar with this model understand what it does from the description alone?

Naming conventions — Does the name follow the patterns used by other entries of the same type? For models, is the source_id consistent with CMIP naming expectations?

Potential duplicates — A similar entry may have been flagged. Is this genuinely a new entry, or should the submitter be updating an existing one?

Component completeness — For model submissions: do the dynamic, prescribed, and omitted component assignments together account for all the scientific domains expected for this type of model? Is anything missing or misclassified?

Scientific plausibility — Do the values make sense as a whole? For example: a 1 km resolution global ocean grid would be unusual; a land-only model that omits the land surface component would be contradictory; a release year before the model family was established would be inconsistent.


Recording your decision

Take one of the following actions directly on the pull request.

Approve — Use the GitHub review interface to approve. The entry is registered immediately and appears on the EMD site within minutes.

Request changes — Leave review comments explaining what needs to change. Be specific: name the field, state the problem, and suggest a correction where possible. The submitter is notified and may update the submission, which then returns to you.

Reject — Close the pull request with a written explanation. The submitter may open a new submission after addressing the concerns.


Practicalities

You were assigned because your expertise or interests matche the submitted content. You have 4 days to respond before the review is automatically reassigned. If you have a conflict of interest or are unable to review, reassign the pull request to another reviewer via the assignee field.

Once you approve and the entry is merged:

  • The entry appears on the EMD github within a few minutes.
  • Other submitters will then be able to reference it in their own submissions within the next few hours (for example, selecting a registered grid or model family from a dropdown).
  • For model submissions, the source_id will need to be linked in the dropdown in CMIP7 controlled vocabularies. This will automatically be accepted.