Tool calling with llm.chat #12557
Closed
alexanderbrodko
announced in
Q&A
Replies: 1 comment
-
My bad. I do not need tokenize when I use
In fact, the model is Qwen2.5-Coder-Instruct-0.5B |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
According to this example
https://github.com/vllm-project/vllm/blob/27b78c73cad00f5c7bb3b2431f02dc680f7034bc/examples/offline_inference/chat_with_tools.py
I create some model:
Do inference:
Then I ask the model about temperature in San Francisco and it fails.
Any suggestions?
Beta Was this translation helpful? Give feedback.
All reactions