v1.69.0-nightly
·
16 commits
to beae5cfea97651bc6b63ffb78a663708f4ecf23d
since this release
What's Changed
- build: update model in test by @krrishdholakia in #10706
- fix: support for python 3.11- (re datetime UTC) (#10471) by @ishaan-jaff in #10701
- [FIX] Update token fields in schema.prisma to use BigInt for improved… by @husnain7766 in #10697
- [Refactor] Use pip package for enterprise/ folder by @ishaan-jaff in #10709
- [Feat] Add streaming support for using bedrock invoke models with /v1/messages by @ishaan-jaff in #10710
- Add
--version
flag tolitellm-proxy
CLI by @msabramo in #10704 - Add management client docs by @msabramo in #10703
- fix(caching_handler.py): fix embedding str caching result by @krrishdholakia in #10700
- Azure LLM: fix passing through of azure_ad_token_provider parameter by @claralp in #10694
- set correct context window length for all gemini 2.5 variants by @mollux in #10690
- Fix log table bugs (after filtering logic was added) by @NANDINI-star in #10712
- fix(router.py): write file to all deployments by @krrishdholakia in #10708
- Litellm Unified File ID output file id support by @krrishdholakia in #10713
- complete unified batch id support - replace model in jsonl to be deployment model name by @krrishdholakia in #10719
- [UI] Bug Fix - Allow Copying Request / Response on Logs Page by @ishaan-jaff in #10720
- [UI] QA Logs page - Fix bug where log did not remain in focus + text overflow on error logs by @ishaan-jaff in #10725
- Add target model name validation by @krrishdholakia in #10722
- [Bug fix] - allow using credentials for /moderations by @ishaan-jaff in #10723
- [DB] Add index for session_id on LiteLLM_SpendLogs by @ishaan-jaff in #10727
- [QA Bug fix] fix: ensure model info does not get overwritten when editing a model on UI by @ishaan-jaff in #10726
- Mutable default arguments on embeddings/completion headers parameters breaks watsonx by @terylt in #10728
New Contributors
- @husnain7766 made their first contribution in #10697
- @claralp made their first contribution in #10694
- @mollux made their first contribution in #10690
- @terylt made their first contribution in #10728
Full Changelog: v1.68.2-nightly...v1.69.0-nightly
Docker Run LiteLLM Proxy
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.69.0-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Passed ✅ | 250.0 | 292.69430995024163 | 6.184694862389694 | 0.0 | 1849 | 0 | 216.9113210000262 | 60025.948276999996 |
Aggregated | Passed ✅ | 250.0 | 292.69430995024163 | 6.184694862389694 | 0.0 | 1849 | 0 | 216.9113210000262 | 60025.948276999996 |