Set Different GPU Memory Utilization for Each GPU #10201
MohmedMonsef
announced in
Q&A
Replies: 1 comment 1 reply
-
No, this is not possible. It would complicate things too much to support this. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have two GPUs with varying workloads, so one might have more memory than the other. Is there a way to configure vllm to use different amounts of memory for each GPU individually?
I am using
pipeline_parallel_size: 1
tensor_parallel_size: 2
as my two GPUs are located on one node.
And according to my understanding, the same
gpu-memory-utilization
setting is applied to all GPUs.Beta Was this translation helpful? Give feedback.
All reactions