Replies: 3 comments
-
@JohnClaw Prompt template just says to use ### Problem ### Solution keywords, so something like this should work, you might need to change -ngl to lower number if you do not have enough VRAM on you gpu or remove it completely if you do not use GPU at all: |
Beta Was this translation helpful? Give feedback.
-
You can also run in conversation mode just add -cnv to that, remove -p "What is 500 * 10000?" -n 100 options |
Beta Was this translation helpful? Give feedback.
-
Here a full demo: https://github.com/foldl/chatllm.cpp/blob/master/docs/tool_calling.md#numinamath |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Link to weights: https://huggingface.co/AI-MO/NuminaMath-7B-TIR
Link to gguf: https://huggingface.co/bartowski/NuminaMath-7B-TIR-GGUF
What is the compatible chat template and llama-cli's command-line options for this great model?
Beta Was this translation helpful? Give feedback.
All reactions