Skip to content

Conversation

@jfrost-mo
Copy link
Member

@jfrost-mo jfrost-mo commented Sep 17, 2025

Still a draft, but please do comment!

Contribution checklist

Aim to have all relevant checks ticked off before merging. See the developer's guide for more detail.

  • Documentation has been updated to reflect change.
  • New code has tests, and affected old tests have been updated.
  • All tests and CI checks pass.
  • Ensured the pull request title is descriptive.
  • Conda lock files have been updated if dependencies have changed.
  • Attributed any Generative AI, such as GitHub Copilot, used in this PR.
  • Marked the PR as ready to review.

@jfrost-mo jfrost-mo self-assigned this Sep 17, 2025
@jfrost-mo jfrost-mo added documentation Improvements or additions to documentation full_review Requires a technical, scientific, and portability review labels Sep 17, 2025
@github-actions
Copy link
Contributor

github-actions bot commented Sep 17, 2025

Coverage

@jfrost-mo jfrost-mo force-pushed the JOSS_paper branch 2 times, most recently from 36a59f6 to 050ab4f Compare September 22, 2025 12:33
@jfrost-mo
Copy link
Member Author

Current version: paper.pdf

@jfrost-mo jfrost-mo changed the title Initial version of JOSS paper Add JOSS paper Sep 22, 2025
Copy link
Contributor

@daflack daflack left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Overall looking good. I've made several comments for what are generally minor wording tweaks just to improve accessibility, but see what you think.


<!-- A summary describing the high-level functionality and purpose of the software for a diverse, non-specialist audience. -->

The _Convective- [and turbulence-] Scale Evaluation Toolkit_ (**CSET**) is a community-driven open source library, command line tool, and workflow designed to support the evaluation of weather and climate models at convective and turbulent scales.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is it better to put convective and turbulent scales as something more accessible like: km and 100 m scales or kilometre and hectometric scales?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

That sounds good, especially as we already mention the convective and turbulence scales in CSET's name. Hectometric is probably too jargony, so maybe

support the evaluation of weather and climate models at kilometre and hundred metre scales.

<!-- A summary describing the high-level functionality and purpose of the software for a diverse, non-specialist audience. -->

The _Convective- [and turbulence-] Scale Evaluation Toolkit_ (**CSET**) is a community-driven open source library, command line tool, and workflow designed to support the evaluation of weather and climate models at convective and turbulent scales.
Developed by the Met Office in collaboration with the [Momentum® Partnership][momentum_partnership] and broader research community, CSET provides a reproducible, modular, and extensible framework for model diagnostics and verification.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not clear what "broader research community" means here - are we talking industry, academic partners, something else?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Essentially anyone who is not in the Momentum Partnership. Currently this is just academics.

Evaluation is essential for the model development process in atmospheric sciences.
Typically, an evaluation includes both context and justification to demonstrate the benefit of model changes against other models or previous model versions.
The verification provides the context or baseline for understanding the model’s performance through comparison against observation.
The evaluation then demonstrates the benefit through comparison against theoretical expectations or previous or different version of the model and other models for similar application areas using diagnostics derived from model output to explain the context.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
The evaluation then demonstrates the benefit through comparison against theoretical expectations or previous or different version of the model and other models for similar application areas using diagnostics derived from model output to explain the context.
The evaluation then demonstrates the benefit through comparison against theoretical expectations or known characteristics of previous or different versions of the same model, and (or) other models for similar applications using diagnostics derived from model output to explain the context.

I'm getting a bit lost in this sentence. I've altered it a little, but see what you think, I'm still not fully happy with what I have suggested. So, I think the main message is that there needs to be some further iteration on this sentence.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'll have a crack at it.

The recipes and operators within CSET are well-documented, tested, and peer reviewed, increasing discoverability and giving confidence to users.
The documentation covers information on the applicability and interpretation of diagnostics, ensuring they are appropriately used.

CSET has been built with portability in mind.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should not be starting a sentence with an acronym.

CSET has been built with portability in mind.
It can run on a range of platforms, from laptops to supercomputers, and can be easily installed from conda-forge.
It is built on a modern software stack that is underpinned by Cylc (a workflow engine for complex computational tasks) [@cylc8], Python 3, and Iris (a Python library for meteorological data analysis) [@scitools_iris].
CSET is open source under the Apache-2.0 licence, and actively developed on GitHub, with extensive automatic unit and integration testing.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should not be starting a sentence with an acronym.


# Conclusion

CSET shows the benefits of open source evaluation software.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should not be starting a sentence with an acronym.


<!-- Acknowledgement of any financial support. -->

We acknowledge contributions and support from the Met Office and Momentum® Partnership for this project.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You should also be acknowledging funding e..g WCSSP projects that are contributing to people's time/have contributed towards peoples time and naming who is being supported by them.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do we have a good list? Off the top of my head:

  • WCSSP South East Asia
  • WCSSP South Afric
  • NGMS, ML-INT (Not sure as these are internal Met Office programs.)


# Contribution to the field

CSET addresses the need for an evaluation system that supports consistent and comparable evaluation.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please don't start a sentence with an acronym.

jfrost-mo and others added 4 commits September 24, 2025 09:56
Copy link
Member Author

@jfrost-mo jfrost-mo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you for the comments.

I'll go ahead and adjust some of the wording as you suggest, and I'll also have a go at deacronyming the sentence starts.

The justification side of CSET consists of a range of diagnostics derived from model output.
These derived diagnostics include process-based diagnostics for specific atmospheric phenomena and impact-based diagnostics that can be used to understand how model changes will affect customers.

Compared to alternative open source evaluation tools, such as ESMValTool [@esmvaltool], CSET is focused on weather-relevant time scales and evaluating models towards a goal of operational usage.
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nothing JOSS specific.

I chose that reference as ESMValTool has a CITATION.cff file in their repository, which is used to indicate the preferred citation.

There are plenty of other good ESMValTool papers1 though, so I'm not really sure which one to pick.

Footnotes

  1. Such as https://gmd.copernicus.org/articles/18/4009/2025/ and https://gmd.copernicus.org/articles/18/1169/2025/

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Any suggestions for nicer images for CSET?

title: "CSET: Toolkit for evaluation of weather and climate models"
date: 17 September 2025
bibliography: paper.bib
tags:
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Any additional keywords to suggest?

- Weather
- Climate
- Atmospheric Science
authors:
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

To start with I've ordered the authors by amount of code/commits, with a cut off for those that have changed $\lt 10$ lines of code (which is the same as those who have changed $\lt 100$ lines).

I'm not very experienced with this aspect of publication, so would welcome any advice.

Additionally, we need to get everyone listed here to sign off on the paper, and make sure we have their preferred name and affiliation.


<!-- A summary describing the high-level functionality and purpose of the software for a diverse, non-specialist audience. -->

The _Convective- [and turbulence-] Scale Evaluation Toolkit_ (**CSET**) is a community-driven open source library, command line tool, and workflow designed to support the evaluation of weather and climate models at convective and turbulent scales.
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

That sounds good, especially as we already mention the convective and turbulence scales in CSET's name. Hectometric is probably too jargony, so maybe

support the evaluation of weather and climate models at kilometre and hundred metre scales.

<!-- A summary describing the high-level functionality and purpose of the software for a diverse, non-specialist audience. -->

The _Convective- [and turbulence-] Scale Evaluation Toolkit_ (**CSET**) is a community-driven open source library, command line tool, and workflow designed to support the evaluation of weather and climate models at convective and turbulent scales.
Developed by the Met Office in collaboration with the [Momentum® Partnership][momentum_partnership] and broader research community, CSET provides a reproducible, modular, and extensible framework for model diagnostics and verification.
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Essentially anyone who is not in the Momentum Partnership. Currently this is just academics.

It gives users easy access to a wide selection of peer-reviewed diagnostics, including spatial plots, time series, vertical profiles, probability density functions, and aggregated analysis over multiple model simulations, replacing bespoke evaluation scripts.
To cater for the full evaluation process, CSET provides a range of verification diagnostics to compare against observations and derived diagnostics based on model output, allowing for both physical process-based and impact-based understanding.

<!-- TODO: Find a better image. -->
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Definitely. I wanted an image of the output, and mainly put it in here to figure out the layout, but this was just a random screenshot I had to hand. I would welcome suggestions for what would make an interesting screenshot.


<!-- Acknowledgement of any financial support. -->

We acknowledge contributions and support from the Met Office and Momentum® Partnership for this project.
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do we have a good list? Off the top of my head:

  • WCSSP South East Asia
  • WCSSP South Afric
  • NGMS, ML-INT (Not sure as these are internal Met Office programs.)


# Design

CSET is built using operators, recipes and a workflow:
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is there a way to start the sentence without the acronym?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If you feel there is no way to start it without the acronym it should be spelt out in full rather than using the acronym (even if the acronym has already been defined).

Evaluation is essential for the model development process in atmospheric sciences.
Typically, an evaluation includes both context and justification to demonstrate the benefit of model changes against other models or previous model versions.
The verification provides the context or baseline for understanding the model’s performance through comparison against observation.
The evaluation then demonstrates the benefit through comparison against theoretical expectations or previous or different version of the model and other models for similar application areas using diagnostics derived from model output to explain the context.
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'll have a crack at it.

- name: John M Edwards
orcid:
affiliation: 1
affiliations:
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should have "MetOffice@Reading" for Reading folk.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation full_review Requires a technical, scientific, and portability review

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants