vLLM serve wrong output #1525
Closed
Martins6
started this conversation in
Weird model output
Replies: 2 comments 1 reply
-
This is likely the result of missing chat templating (docs here). In general I'm not totally sure why we still provide the outlines serve functionality. I usually refer people to vLLM instead, which supports Outlines as a constrained decoding backend. |
Beta Was this translation helpful? Give feedback.
0 replies
-
From what I've inspected, it seems that it was just a wrong line of code! I've opened this PR to fix this in case you guys want to maintain the vLLM serve @cpfiffer But nice, I'll also check out vLLM directly! Thank you very much! 🚀 |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
As I was testing the vLLM serve from pure Outlines perspective, I've noticed that the outputs have a weird format of the prompt that I've used + the result that I want.
To replicate first run on a Terminal:
python -m outlines.serve.serve --model="microsoft/Phi-3-mini-4k-instruct"
Then run this Python script that depends on requests and Pydantic. I'm using Python3.11 too.
The result that I get is:
See? I belive the expected result is just:
Beta Was this translation helpful? Give feedback.
All reactions