Replies: 1 comment 3 replies
-
Hello, the server is printing the chat template at startup. You can set --verbose for verbose logs in addition with --log-format text. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
For example, if I start the server with this:
./server -m ../llama/llama-2-13b/ggml-model-f16.gguf
Then I hit it from python using this:
I can't tell if I need to change my user content's text to match the llama-2-13b model I'm running's prompt format from their paper, which is the following format:
<s>[INST] <<sys>> systemprompthere <</sys>> prompt [/INST]
Any help is appreciated!
P.S.- is there a way to require the server output json? I couldn't find a flag for using a json grammar with the server.
Beta Was this translation helpful? Give feedback.
All reactions