How to apply a custom function on a dataset's chunks to reduce its size and save it on disk ? #7913
-
Hello, This question is a clone of something I asked on StackOverflow. I have a dataset (110 Gb) that is created by using In short: how can I write on disk a dataset that is too big to be stored in the RAM ? Why am I doing that ? I want to apply a computationally intensive method that goes iteratively through every 2D pixel of my dataset ( ############################################################### Here is a code reproducing what I am trying to do (also as a python script):
|
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 19 replies
-
Does any of the material in https://e-marshall.github.io/itslive/intro.html help? cc @e-marshall |
Beta Was this translation helpful? Give feedback.
-
Hey @vdevauxchupin ! I've struggled with this type of thing of thing with ITSLIVE before but don't think I've figured out much of a solution, and haven't figured out a way to do directly what you're asking. A few thoughts:
|
Beta Was this translation helpful? Give feedback.
-
It sounds more like you should do this:
|
Beta Was this translation helpful? Give feedback.
It sounds more like you should do this:
{'x': 100, 'y': 100. 'time': -1}
(so same as chunks on disk expect for opening all time chunks as contiguous).