Does vllm support deploy models on vuda? #9691
Gaohang0804
announced in
Q&A
Replies: 1 comment
-
遇到同样问题了,在k8s上 tencent.com/vcuda-core 设成100 可以部署 改成25 这样报错 。请问你解决了吗 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm trying to deploy internVL2 model with vllm on k8s,I have succeed on 1 A800 80G GPU,but because the low memory required of my model,many gpu memory is wasted.
I just wondering could deploy my model on virtualized cuda,and required lower memory on one k8s pod.
Beta Was this translation helpful? Give feedback.
All reactions