You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I wonder what can cause difference in Llama3-8b answers for the same promt when use curl to llamacpp http server endpoint vs when use chat ui requesting the same llamacpp server.
curl respond with messy output, looks like the model ignores rules in system_prompt parameter, while chatui case looks good to me.
My guess is that reason is chatPromptTemplate which I use as parameter in chatui config vs raw string as prompt in curl request. Can this cause differences. If so, how can I set chat template when use /completion endpoint
Thank you for any suggestions
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I wonder what can cause difference in Llama3-8b answers for the same promt when use
curl
to llamacpp http server endpoint vs when use chat ui requesting the same llamacpp server.curl respond with messy output, looks like the model ignores rules in
system_prompt
parameter, while chatui case looks good to me.My guess is that reason is
chatPromptTemplate
which I use as parameter in chatui config vs raw string asprompt
in curl request. Can this cause differences. If so, how can I set chat template when use/completion
endpointThank you for any suggestions
Beta Was this translation helpful? Give feedback.
All reactions