You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Would including the benchmark info of representative runs in the documentation be helpful?
These runs could include a 14-day ensemble simulation, a 1-year deterministic simulation, and so on. It would also be interesting to see the performance differences between GPU and TPU.
The information can help users check their local configuration and make informed decisions about their experiments (e.g., Google Cloud or local Nvidia machines).
The text was updated successfully, but these errors were encountered:
Thanks. I misread the table and thought the benchmark chip was T4 of Colab. The lower performance on a local implementation also confused me.
As a data point for the community, the inference time of a benchmark comparable task with JAX (cuda) + Nvidia L40S (A100-level) is approximately 40s. This is a preliminary number with a naive implementation.
Would including the benchmark info of representative runs in the documentation be helpful?
These runs could include a 14-day ensemble simulation, a 1-year deterministic simulation, and so on. It would also be interesting to see the performance differences between GPU and TPU.
The information can help users check their local configuration and make informed decisions about their experiments (e.g., Google Cloud or local Nvidia machines).
The text was updated successfully, but these errors were encountered: