sample params difference between vLLM and Huggingface transformers #539
zhaoyang-star
announced in
Q&A
Replies: 2 comments
-
Did you figure it out? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Here are some merged pull requests focusing on the alignment of sampling parameters and results: #753, #1424, #1577, and #1885. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
We know evaluation is very important for a production env. I tested the code completion ability of starcoder-15b by HumanEval using vLLM and HF as two different backends.
I found it is hard to get the same pass@1 using vLLM and HF transformers. So how to get the same pass@1 as HF when I use vLLM?
There may be two reasons and the first may be the main part.
For the first one. I compared the main params as following.
I use the same sampling params in vLLM and HF.
Beta Was this translation helpful? Give feedback.
All reactions