-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
decide on attributes and metadata for downscaled data #179
Comments
I think there is a more essential low-hanging fruit version of this. As it is right now, we're just dumping blob data into azure storage. We need at least enough basic metadata to know what we're looking at, at the simplest level (source_id, variable_id, etc...), even if that information is not encoded in the data path. |
That makes sense. So first pass can be all of the basic identifiers - e.g. SSP, model, variable, frequency, etc - along with methods used (e.g. QDM/AIQPD, wet day frequency for precip). We can iron out the CF-conventions-compliance part of this as a next step. Sound good? In terms of what data actually needs metadata, I'm thinking the bias corrected and downscaled outputs - any intermediate outputs as well? Cleaned and rechunked CMIP6 input data to bias correction perhaps? Thinking we add a function in |
The immediate issue is that metadata needs to be preserved between workflow steps because it's useful to validate and write final-ish output. All of this "immediately essential" data (stuff we use for I/O) is present as soon as we download it from CMIP6-in-the-cloud. We just need to hold on to it. Right now, ClimateImpactLab/dodola#116 is likely the most obvious offender. |
For an example of what we currently have in place, here is a copy of the Dataset metadata for output: Our additions have a I need additional feedback to continue. Any other input? |
We need to iterate and decide on how we want to set up our attributes and metadata for downscaled zarr stores - this is just a placeholder to start the discussion.
The text was updated successfully, but these errors were encountered: