Force disable thinking for gemini-2.5-flash model? #10450
Dima-Mediator
started this conversation in
General
Replies: 1 comment 2 replies
-
|
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Following #10121
Does anybody understand how to disable thinking for the gemini-2.5-flash-preview-04-17 model on LiteLLM side?
The client (Open WebUI) is unfortunately not currently capable of transmitting the Anthropic-style "thinking" parameter and only supports reasoning_effort which is not enough. And thinking is currently enabled by default for this model.
In the original PR, someone suggested using using YAML model-level litellm_params to force a thinking setting (with budget 0) but I have not figured out the exact syntax to make it work. Has someone succeeded?
Beta Was this translation helpful? Give feedback.
All reactions