Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cdsapi: ERA5 total precipitation accumulation #23

Open
Aquila96 opened this issue Jul 6, 2023 · 4 comments
Open

cdsapi: ERA5 total precipitation accumulation #23

Aquila96 opened this issue Jul 6, 2023 · 4 comments

Comments

@Aquila96
Copy link

Aquila96 commented Jul 6, 2023

In data_factory/download_era5.py, it downloads 6-hour total precipitation by specifying 6 hour intevals.

But after reviewing the data and reading the ERA5 cdsapi demo, It seems that the api only downloads 1-hour accumlated precipitation, for multi-hour accumulated precipitation, one have to download all the previous hours and perform accumulation afterwards.

@daniellucs2002
Copy link

Yeah, I agree with you. Also, in the GraphCast paper, the authors also mentioned that the precipitation data needs to be accumulated from the past six hours. So, have you tried to modify the code or figured out how to prepare the training data at least? (I plan to train the model from scratch if the provided weights file cannot be utilized)

@Aquila96
Copy link
Author

Yeah, I agree with you. Also, in the GraphCast paper, the authors also mentioned that the precipitation data needs to be accumulated from the past six hours. So, have you tried to modify the code or figured out how to prepare the training data at least? (I plan to train the model from scratch if the provided weights file cannot be utilized)

cdsapi is rather straight forward to use, you can refer to NVIDIA's download demo here and here. For precipitation specifically, I just fetch data from all 6 previous hours and add them together, similar to the approach in the ECMWF demo.

@daniellucs2002
Copy link

Thanks for answering! I remember you mentioned that the inconsistency in model architecture between huggingface checkpoint and this repository in another issue post. Have you addressed that problem and run the project on the GPU cluster?

@Aquila96
Copy link
Author

Thanks for answering! I remember you mentioned that the inconsistency in model architecture between huggingface checkpoint and this repository in another issue post. Have you addressed that problem and run the project on the GPU cluster?

No problem. Regarding the model weights, I wouldn't recommend to try to modify the source code to match the weights provided.
For FourCastNet, we have compared results both from this version and the official version from NVIDIA, and the results from this version is worse, though I couldn't remember by how large of a margin.
For GraphCast, deepmind have since released their implementation and weights in jax, and the results are mostly on par with their claimed state-of-the-art performance. You can also read the implementation in modulus for the pytorch.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants