File tree Expand file tree Collapse file tree 2 files changed +31
-0
lines changed Expand file tree Collapse file tree 2 files changed +31
-0
lines changed Original file line number Diff line number Diff line change 1+ # Frequently Asked Questions (FAQ)
2+
3+ ## Performance and Memory
4+
5+ ### Why does ` to_multiscales ` perform computation immediately with large datasets?
6+
7+ For both small and large datasets, ` to_multiscales ` returns a simple Python
8+ dataclass composed of basic Python datatypes and lazy dask arrays. The lazy dask
9+ arrays, as with all dask arrays, are a task graph that defines how to generate
10+ those arrays and do not exist in memory. This is the case for both regular sized
11+ datasets and very large datasets.
12+
13+ For very large datasets, though, data conditioning and task graph engineering is
14+ performed during construction to improve performance and avoid running out of
15+ system memory. This preprocessing step ensures that when you eventually compute
16+ the arrays, the operations are optimized and memory-efficient.
17+
18+ If you want to avoid this behavior entirely, you can pass ` cache=False ` to
19+ ` to_multiscales ` :
20+
21+ ``` python
22+ multiscales = to_multiscales(image, cache = False )
23+ ```
24+
25+ ** Warning:** Disabling caching may cause you to run out of memory when working
26+ with very large datasets!
27+
28+ The lazy evaluation approach allows ngff-zarr to handle extremely large datasets
29+ that wouldn't fit in memory, while still providing optimal performance through
30+ intelligent task graph optimization.
Original file line number Diff line number Diff line change 3737mcp.md
3838itk.md
3939methods.md
40+ faq.md
4041development.md
4142```
4243
You can’t perform that action at this time.
0 commit comments