sample params difference between vLLM and Huggingface transformers #539
Unanswered
zhaoyang-star
asked this question in
Q&A
Replies: 2 comments
-
Did you figure it out? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Here are some merged pull requests focusing on the alignment of sampling parameters and results: #753, #1424, #1577, and #1885. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
We know evaluation is very important for a production env. I tested the code completion ability of starcoder-15b by HumanEval using vLLM and HF as two different backends.
I found it is hard to get the same pass@1 using vLLM and HF transformers. So how to get the same pass@1 as HF when I use vLLM?
There may be two reasons and the first may be the main part.
For the first one. I compared the main params as following.
I use the same sampling params in vLLM and HF.
Beta Was this translation helpful? Give feedback.
All reactions