Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add benchmark info of representative runs? #104

Open
GanZhang-GFD opened this issue Jul 24, 2024 · 2 comments
Open

Add benchmark info of representative runs? #104

GanZhang-GFD opened this issue Jul 24, 2024 · 2 comments

Comments

@GanZhang-GFD
Copy link

Would including the benchmark info of representative runs in the documentation be helpful?

These runs could include a 14-day ensemble simulation, a 1-year deterministic simulation, and so on. It would also be interesting to see the performance differences between GPU and TPU.

The information can help users check their local configuration and make informed decisions about their experiments (e.g., Google Cloud or local Nvidia machines).

@shoyer
Copy link
Collaborator

shoyer commented Jul 24, 2024

We have have training and inference runtimes for all our different models in an Extended Data table from the paper:
https://www.nature.com/articles/s41586-024-07744-y/tables/1

image

The inference numbers are all reported for a single core of a Google TPU v4 chip.

Is there something else you were thinking of?

@GanZhang-GFD
Copy link
Author

Thanks. I misread the table and thought the benchmark chip was T4 of Colab. The lower performance on a local implementation also confused me.

As a data point for the community, the inference time of a benchmark comparable task with JAX (cuda) + Nvidia L40S (A100-level) is approximately 40s. This is a preliminary number with a naive implementation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants