Strange behaviour in VLLMOpenAI #29309
Replies: 1 comment 2 replies
-
Hello @to-sora! I'm here to assist you with any bugs, questions, or contributions you have. Let's tackle this together! I found a similar unsolved discussion that might be relevant to your issue: langchain-openai couldn't work while the openai package works fine at the same parameters [1]. Regarding your issue, it seems like the LangChain implementation is not behaving as expected compared to the direct OpenAI API usage. Here are a few things you might want to check or try:
If these steps do not resolve the issue, it might be worth exploring the differences in how LangChain and the OpenAI package handle requests and responses, especially in terms of API compatibility and parameter handling. To continue talking to Dosu, mention @dosu. Help Dosu learn! Give it feedback: Great Response | Irrelevant Answer | Incorrect Sources | Too Verbose | Hallucination | Other |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I am using vllm and want to use batch process.
The vllm is start by
As a compare testI run the code in vllm docs
And the result is reasonable with backed end log called
How ever, when i run
The langchain code
and return
with vllm log
as vllm docs in clear say https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html
Supported APIs
We currently support the following OpenAI APIs:
Completions API (/v1/completions)
Only applicable to text generation models (--task generate).
Note: suffix parameter is not supported.
Chat Completions API (/v1/chat/completions)
Only applicable to text generation models (--task generate) with a chat template.
Note: parallel_tool_calls and user parameters are ignored.
Embeddings API (/v1/embeddings)
Only applicable to embedding models (--task embed).
May I know is I making any mistake or is a bug.
FYI, the following generated result is also meaning less
Beta Was this translation helpful? Give feedback.
All reactions