vLLM for SBERT #1575
david-koleckar
announced in
Q&A
vLLM for SBERT
#1575
Replies: 1 comment
-
Hi @david-koleckar, thanks for the proposal. For now, we are not planning to support SBERT model or other embedding/encoder models. vLLM is optimized for generative (auto-regressive) Transformers. Its core technologies like PagedAttention cannot be used for embedding/encoder models. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi there,
would it be possible to add sbert model (eg. paraphrase-multilingual-mpnet-base-v2) to vLLM as described in adding a new model?
Thank you, David
Beta Was this translation helpful? Give feedback.
All reactions