Skip to content

v1.68.0-nightly

Compare
Choose a tag to compare
@github-actions github-actions released this 04 May 06:30
· 3 commits to 914ab0080594299591d64fecefb7a23c3ac9931f since this release

What's Changed

  • [Contributor PR] Support Llama-api as an LLM provider (#10451) by @ishaan-jaff in #10538
  • UI - fix(model_management_endpoints.py): allow team admin to update model info + fix request logs - handle expanding other rows when existing row selected + fix(organization_endpoints.py): enable proxy admin with 'all-proxy-model' access to create new org with specific models by @krrishdholakia in #10539
  • [Bug Fix] UnicodeDecodeError: 'charmap' on Windows during litellm import by @ishaan-jaff in #10542
  • fix(converse_transformation.py): handle meta llama tool call response by @krrishdholakia in #10541

Full Changelog: v1.67.6.dev1...v1.68.0-nightly

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.68.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Docker Run LiteLLM Proxy

docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.68.0-nightly

Don't want to maintain your internal proxy? get in touch 🎉

Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat

Load Test LiteLLM Proxy Results

Name Status Median Response Time (ms) Average Response Time (ms) Requests/s Failures/s Request Count Failure Count Min Response Time (ms) Max Response Time (ms)
/chat/completions Passed ✅ 180.0 210.99923315604772 6.1894793990457675 0.0 1852 0 166.69672900002297 3755.0343799999837
Aggregated Passed ✅ 180.0 210.99923315604772 6.1894793990457675 0.0 1852 0 166.69672900002297 3755.0343799999837