v1.68.1.dev4
What's Changed
- Contributor PR - Return 404 when
delete_verification_tokens
(POST /key/delete
) fai… by @ishaan-jaff in #10605 - Fix otel - follow genai semantic conventions + support 'instructions' param for tts by @krrishdholakia in #10608
- make openai model O series conditional accept provider/model by @aholmberg in #10591
- add gemini-2.5-pro-preview-05-06 model prices and context window by @marty-sullivan in #10597
- Fix: Ollama integration KeyError when using JSON response format by @aravindkarnam in #10611
- [Feat] V2 Emails - Fixes for sending emails when creating keys + Resend API support by @ishaan-jaff in #10602
- [Feat] Add User invitation emails when inviting users to litellm by @ishaan-jaff in #10615
- [Fix] SCIM - Creating SCIM tokens on Admin UI by @ishaan-jaff in #10628
- Filter on logs table by @NANDINI-star in #10644
- [Feat] Bedrock Guardrails - Add support for PII Masking with bedrock guardrails by @ishaan-jaff in #10642
- [Feat] Add endpoints to manage email settings by @ishaan-jaff in #10646
- Contributor PR - MCP Server DB Schema (#10634) by @ishaan-jaff in #10641
- Ollama - fix custom price cost tracking + add 'max_completion_token' support by @krrishdholakia in #10636
- fix cerebras llama-3.1-70b model_prices_and_context_window, not llama3.1-70b by @xsg22 in #10648
- Fix cache miss for gemini models with response_format by @casparhsws in #10635
- Add user management functionality to Python client library & CLI by @msabramo in #10627
- [BETA] Support unified file id (managed files) for batches by @krrishdholakia in #10650
- Fix Slack alerting not working if using a DB by @hypermoose in #10370
- Add support for Nscale (EU-Sovereign) Provider by @tomukmatthews in #10638
- Add New Perplexity Models by @keyute in #10652
New Contributors
- @aholmberg made their first contribution in #10591
- @aravindkarnam made their first contribution in #10611
- @xsg22 made their first contribution in #10648
- @casparhsws made their first contribution in #10635
- @hypermoose made their first contribution in #10370
- @tomukmatthews made their first contribution in #10638
- @keyute made their first contribution in #10652
Full Changelog: v1.68.1-nightly...v1.68.1.dev4
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.68.1.dev4
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 190.0 | 233.10816080888745 | 6.241336822394705 | 0.0 | 1868 | 0 | 166.93079599997418 | 5406.457653000075 |
Aggregated | Passed ✅ | 190.0 | 233.10816080888745 | 6.241336822394705 | 0.0 | 1868 | 0 | 166.93079599997418 | 5406.457653000075 |