Best practices for distributing memory? #16618
Unanswered
rkruegs123
asked this question in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have a memory-intensive computation that I want to scale up by using multiple GPUs. I have been experimenting with
xmap
,pmap
, andshmap
. My understanding is that these would typically be use to distribute batches in a training set to parallelize computation. Note that I am not training a neural network, and therefore the computation I am trying to parallelize is not analogous -- instead, the computation I am trying to vectorize (and distribute across GPUs) is in ascan
.I have found two things:
scan
/jitted function isshmap
. This is fine.So, does anybody know how to distribute memory across GPUs when the computation to be vectorized is within another function? A minimal example would be extremely helpful.
Thank you in advance.
Beta Was this translation helpful? Give feedback.
All reactions