You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For the model kaitchup/Qwen2.5-72B-Instruct-AutoRoundGPTQ-8bit, the leaderboard_ifeval results differ significantly between the HF backend and the vLLM backend. Could you provide insights into the possible reasons or help debug the issue? Thanks in advance!
For the model
kaitchup/Qwen2.5-72B-Instruct-AutoRoundGPTQ-8bit
, theleaderboard_ifeval
results differ significantly between the HF backend and the vLLM backend. Could you provide insights into the possible reasons or help debug the issue? Thanks in advance!HF backend
hf (pretrained=./,parallelize=True,dtype=float16), gen_kwargs: (None), limit: 10.0, num_fewshot: None, batch_size: 16
vllm backend
vllm (pretrained=./,tensor_parallel_size=2,dtype=float16), gen_kwargs: (None), limit: 10.0, num_fewshot: None, batch_size: auto
limit 10
The text was updated successfully, but these errors were encountered: