-
Notifications
You must be signed in to change notification settings - Fork 116
Open
Labels
enhancementNew feature or requestNew feature or request
Description
from the pangeo working meeting discussion with @mgrover1 @jmunroe @norlandrhagen
Here's an outline for an intermediate tutorial talking about dask chunking specifically for Xarray users
Motivation: why care about chunk size?
- demonstrate relation between chunk size and computation time / number of tasks with a simple example?
- maybe even memory usage
- https://tutorial.dask.org/02_array.html#Choosing-good-chunk-sizes
- https://docs.dask.org/en/stable/array-chunks.html
Keeping track
- monitoring chunk sizes and num tasks throughout the pipeline using the repr
- use some images
- while output blocks may be small (say after a big reduction), intermediate blocks need not be.
- So keep monitoring chunksizes (and tasks) throughout the pipeline.
Why is it important to choose appropriate chunks early in the pipeline?
- Demonstrate that rechunking is not cheap in most cases
Specify chunks when reading data
- Avoid
chunks="auto"
. - Specifying
chunks
during data readopen_dataset
open_mfdataset
- Analysis vs storage chunks:
vanderwb and scottyhqnorlandrhagen and mgrover1
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or request
Activity
scottyhq commentedon Jan 31, 2023
Agreed this would be great to document thoroughly. See also this relevant issue + discussion in rioxarray corteva/rioxarray#253
djhoese commentedon Feb 1, 2023
This would be huge! This comes up often in Satpy where users want to process satellite images on their local machine but they only have 8GB or 16GB of memory. If someone can make a good diagram showing chunks being processed by a worker thread/process and how changing the size of all chunks or number of workers contributes to the overall memory usage that would be such a help when explaining this to users.
dcherian commentedon Feb 2, 2023
forgot to cc @rybchuk!