Add support for prompt caching through litellm #2429
mboret
started this conversation in
Feature Requests
Replies: 1 comment 5 replies
-
Hi, did you try checking the prompt caching checkbox? We added that recently. |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I'm using Roo Code with litellm as an OpenAI-compatible endpoint. Prompt caching is enabled.
If I use an OpenAI model like o3-mini (Azure), prompt caching works, but it does not when I switch to Claude 3.7 sonnet (AWS bedrock).
When I inspect the request sent to litellm by roo code, no caching parameter is defined (which makes sense as OpenAI doesn't need extra parameters to do prompt caching), but Claude 3.7 sonnet needs one.
Example of request to litellm with prompt caching parameter:
litellm doc
It would be nice to be able to add this extra parameter to use prompt caching
Beta Was this translation helpful? Give feedback.
All reactions