Chatbox AI app backend support #1047
magikRUKKOLA
started this conversation in
Show and tell
Replies: 1 comment 6 replies
-
Is there a way to fold the <think> produced by R1 series? My local deployment is too thinkable that its <think> is longer than its answer... |
Beta Was this translation helpful? Give feedback.
6 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
In case someone would want to use ktransformers (multiple ones, running at different machines) as a backend for apps like Chatbox AI that support Ollama API they could use the following.
[EDIT] (link to the latest version of nginx config updated)
chatboxai/chatbox#2221 (comment)
the picture below shows the usage of the llms from three different backends (ollama and 2 ktransformers).
Beta Was this translation helpful? Give feedback.
All reactions