-
Notifications
You must be signed in to change notification settings - Fork 12.4k
Description
Prerequisites
- I am running the latest code. Mention the version if possible as well.
- I carefully followed the README.md.
- I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
- I reviewed the Discussions, and have a new and useful enhancement to share.
Feature Description
Currently, the llama-server can run the deepseek-r1-0528 unsolth model normally, but when adding the --jinja parameter, the tool calling functionality fails to meet expectations. Despite multiple attempts with the --chat-template-file parameter (e.g., --chat-template-file tool_chat_template_deepseekr1.jinja), it still doesn't work. Could this be an issue with the parsing of the chat-template-file
Motivation
llama-server -m /models/UD-IQ2_XXS/DeepSeek-R1-0528-UD-IQ2_XXS-00001-of-00005.gguf --n-gpu-layers 99 --host 0.0.0.0 --port 9996 --parallel 2 --ctx-size 16384 --cache-type-k q4_0 --api-key sk-123456789 --threads 24 -a DeepSeek-R1-0528-UD-IQ2_XXS --temp 0.6 --top_p 0.95 --min_p 0.01 --seed 3407 --jinja --chat-template-file ~/tool_chat_template_deepseekr1.jinja
Possible Implementation
No response