Multi GPU with Vulkan out of memory issue. #5720
Replies: 1 comment
-
Solved in: #5848 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I'm trying to load a model on two GPUs with Vulkan, the model's size is Q6_K quant of
26.27 GiB (6.56 BPW)
My GPUs have 20 and 11 gigs of VRAM
Loading it with
-ts "20,11" -c 512
yields:A Q5_K_M quant at
22.65 GiB (5.66 BPW)
works perfectly fine until I increase the context to 4096.Which makes no sense to me.
Any idea why this happens?
Beta Was this translation helpful? Give feedback.
All reactions