How to pass vLLM inference server error messages to GUI / open-webui users? #10857
c-hoffmann
announced in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I run vLLM and open-webui, both installed via pip instead of docker.
My issue is that vLLM error messages don't get passed through to the open-webui user as expected. An easy way to show this is to have a conversation that exceeds the max_model_len / context length. Here's my vLLM output when this occurs:
And here's what open-webui shows:

How can I pass the issue to the GUI more appropriately? Perhaps in this case telling the user to start a new conversation?
Beta Was this translation helpful? Give feedback.
All reactions