-
I compiled
then ran each model with for instance:
but when I try to interact with it using OpenAI Python API library the output doesn't stop:
how can I set the prompt format to:
|
Beta Was this translation helpful? Give feedback.
Replies: 5 comments
-
this is what I'm experiencing: 2024-10-22-21.38.00.368923547.mp4 |
Beta Was this translation helpful? Give feedback.
-
I don't know about server, but for
I spent an hour trying to implement it into the defined templates (wiki) but had to give up at the |
Beta Was this translation helpful? Give feedback.
-
EDIT: I was missing the newlines at the end of "<|end_of_text|>", but note that there should NOT be newlines after "<|end_of_role|>" according to the official template. I just added this after the exaone3 entry in llama_chat_apply_template_internal and worked for me, no other patch was required. You should be able to manually set chat template to "granite" or just let it detect from the presence of "<|start_of_role|>" from the built-in model chat template. This was tested with whatever version of llama.cpp that is required for the latest llama-cpp-python (0.3.1).
|
Beta Was this translation helpful? Give feedback.
-
I saw arch-btw has made a PR to fix this #10013 |
Beta Was this translation helpful? Give feedback.
-
@kth8 sorry for taking so long, I finished it depending on the review. |
Beta Was this translation helpful? Give feedback.
@kth8 sorry for taking so long, I finished it depending on the review.