Skip to content

Time series look-up from a large Zarr file #7131

Answered by shoyer
justurbo asked this question in Q&A
Discussion options

You must be logged in to vote

Your data is chunked with Zarr into blocks of size (50, 721, 1440), which means that every request data for the entire world. This means every look up loads the full 42 GB of data, stored in several hundred files!

To enable efficient queries of your data, you will need to "rechunk" it so you can query data a single location with less waste, e.g., by chunking along latitude and longitude. You can do this with .chunk() and writing a new Zarr file in xarray, or with a tool like Rechunker.

Replies: 2 comments 1 reply

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@JustasUrbonas
Comment options

Answer selected by shoyer
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
4 participants