Infill /Infill api not working on llama.cpp server #9861
Unanswered
deusmachinea
asked this question in
Q&A
Replies: 2 comments
-
You need to add The |
Beta Was this translation helpful? Give feedback.
0 replies
-
Thank you Sir.
the code_before snippet is
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I had downloaded the llama.cpp https://github.com/ggerganov/llama.cpp/releases/download/b3909/llama-b3909-bin-macos-arm64.zip unzipped it and running the llama-server like this
MODEL is basically Qwen2.5-Coder-7B-Instruct-Q4_K_L.gguf
/completions apis is working correctly but /infill request is returning errors .
This is how I am trying
code_before is just a code snippet that is only 100 tokens long.
The error I am getting is
Please let us know how to make /infill request working.
Beta Was this translation helpful? Give feedback.
All reactions