Skip to content

Commit eb998ee

Browse files
authored
[v1.66.0-stable] Release notes (#9952)
* release notes * docs release notes * docs fix release notes * docs clean up * docs clean up * release notes * docs sso tag management
1 parent 25d4cf1 commit eb998ee

File tree

4 files changed

+179
-0
lines changed

4 files changed

+179
-0
lines changed
Loading
60.5 KB
Loading
Loading
Lines changed: 179 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,179 @@
1+
---
2+
title: v1.66.0-stable
3+
slug: v1.66.0-stable
4+
date: 2025-04-12T10:00:00
5+
authors:
6+
- name: Krrish Dholakia
7+
title: CEO, LiteLLM
8+
url: https://www.linkedin.com/in/krish-d/
9+
image_url: https://media.licdn.com/dms/image/v2/D4D03AQGrlsJ3aqpHmQ/profile-displayphoto-shrink_400_400/B4DZSAzgP7HYAg-/0/1737327772964?e=1749686400&v=beta&t=Hkl3U8Ps0VtvNxX0BNNq24b4dtX5wQaPFp6oiKCIHD8
10+
- name: Ishaan Jaffer
11+
title: CTO, LiteLLM
12+
url: https://www.linkedin.com/in/reffajnaahsi/
13+
image_url: https://pbs.twimg.com/profile_images/1613813310264340481/lz54oEiB_400x400.jpg
14+
15+
tags: []
16+
hide_table_of_contents: false
17+
---
18+
19+
import Image from '@theme/IdealImage';
20+
import Tabs from '@theme/Tabs';
21+
import TabItem from '@theme/TabItem';
22+
23+
## Deploy this version
24+
25+
<Tabs>
26+
<TabItem value="docker" label="Docker">
27+
28+
``` showLineNumbers title="docker run litellm"
29+
docker run
30+
-e STORE_MODEL_IN_DB=True
31+
-p 4000:4000
32+
ghcr.io/berriai/litellm:main-v1.66.0-stable
33+
```
34+
</TabItem>
35+
36+
<TabItem value="pip" label="Pip">
37+
38+
``` showLineNumbers title="pip install litellm"
39+
pip install litellm==1.66.0.post1
40+
```
41+
</TabItem>
42+
</Tabs>
43+
44+
v1.66.0-stable is live now, here are the key highlights of this release
45+
46+
## Key Highlights
47+
- **Microsoft SSO Auto-sync**: Auto-sync groups and group members from Azure Entra ID to LiteLLM
48+
- **Unified File IDs**: Use the same file id across LLM API providers.
49+
- **New Models**: `xAI grok-3` support, `realtime api` cost tracking and logging
50+
- **Security Fixes**: Fixed [CVE-2025-0330](https://www.cve.org/CVERecord?id=CVE-2025-0330) and [CVE-2024-6825](https://www.cve.org/CVERecord?id=CVE-2024-6825) vulnerabilities
51+
52+
Let's dive in.
53+
54+
## Microsoft SSO Auto-sync
55+
56+
<Image
57+
img={require('../../img/release_notes/sso_sync.png')}
58+
style={{width: '100%', display: 'block'}}
59+
/>
60+
<p style={{textAlign: 'left', color: '#666'}}>
61+
Auto-sync groups and members from Azure Entra ID to LiteLLM
62+
</p>
63+
64+
This release adds support for auto-syncing groups and members on Microsoft Entra ID with LiteLLM. This means that litellm proxy administrators can spend less time managing teams and members and LiteLLM handles the following:
65+
66+
- Auto-create Teams that existing on Microsoft Entra ID
67+
- Sync team members on Microsoft Entra ID with LiteLLM Teams
68+
69+
Get started with this [here](https://docs.litellm.ai/docs/tutorials/msft_sso)
70+
71+
## Unified File ID
72+
73+
74+
75+
## New Models / Updated Models
76+
77+
- xAI
78+
1. Added cost tracking for `xai/grok-3` models [PR](https://github.com/BerriAI/litellm/pull/9920)
79+
2. Added reasoning_effort support for `xai/grok-3-mini-beta` model family [PR](https://github.com/BerriAI/litellm/pull/9932)
80+
81+
- Hugging Face
82+
1. Hugging Face - Added inference providers support [PR](https://github.com/BerriAI/litellm/pull/9773)
83+
84+
- Azure
85+
1. Azure - Added azure/gpt-4o-realtime-audio cost tracking [PR](https://github.com/BerriAI/litellm/pull/9893)
86+
87+
- VertexAI
88+
1. VertexAI - Added enterpriseWebSearch tool support [PR](https://github.com/BerriAI/litellm/pull/9856)
89+
2. VertexAI - Moved to only passing in accepted keys by vertex ai response schema [PR](https://github.com/BerriAI/litellm/pull/8992)
90+
91+
- Google AI Studio
92+
1. Google AI Studio - Added cost tracking for `gemini-2.5-pro` [PR](https://github.com/BerriAI/litellm/pull/9837)
93+
2. Google AI Studio - Fixed pricing for 'gemini/gemini-2.5-pro-preview-03-25' [PR](https://github.com/BerriAI/litellm/pull/9896)
94+
3. Google AI Studio - Fixed handling file_data being passed in [PR](https://github.com/BerriAI/litellm/pull/9786)
95+
96+
- Azure
97+
1. Azure - Updated Azure Phi-4 pricing [PR](https://github.com/BerriAI/litellm/pull/9862)
98+
2. Azure - Added azure/gpt-4o-realtime-audio cost tracking [PR](https://github.com/BerriAI/litellm/pull/9893)
99+
100+
- Databricks
101+
1. Databricks - Removed reasoning_effort from parameters [PR](https://github.com/BerriAI/litellm/pull/9811)
102+
2. Fixed custom endpoint check for Databricks [PR](https://github.com/BerriAI/litellm/pull/9925)
103+
104+
- General
105+
1. Function Calling - Handle pydantic base model in message tool calls, handle tools = [], and support fake streaming on tool calls for meta.llama3-3-70b-instruct-v1:0 [PR](https://github.com/BerriAI/litellm/pull/9774)
106+
2. LiteLLM Proxy - Allow passing `thinking` param to litellm proxy via client sdk [PR](https://github.com/BerriAI/litellm/pull/9386)
107+
3. Reasoning - Added litellm.supports_reasoning() util to track if an llm supports reasoning [PR](https://github.com/BerriAI/litellm/pull/9923)
108+
4. Fixed correctly translating 'thinking' param for litellm [PR](https://github.com/BerriAI/litellm/pull/9904)
109+
110+
111+
## Spend Tracking Improvements
112+
113+
1. Realtime API Cost tracking with token usage metrics in spend logs [PR](https://github.com/BerriAI/litellm/pull/9795)
114+
2. Fixed Claude Haiku cache read pricing per token [PR](https://github.com/BerriAI/litellm/pull/9834)
115+
3. Added cost tracking for Claude responses with base_model [PR](https://github.com/BerriAI/litellm/pull/9897)
116+
4. Fixed Anthropic prompt caching cost calculation and trimmed logged message in db [PR](https://github.com/BerriAI/litellm/pull/9838)
117+
5. Added token tracking and log usage object in spend logs [PR](https://github.com/BerriAI/litellm/pull/9843)
118+
6. Handle custom pricing at deployment level [PR](https://github.com/BerriAI/litellm/pull/9855)
119+
120+
121+
## Management Endpoints / UI
122+
123+
1. Test Key Tab:
124+
1. Added rendering of Reasoning content, ttft, usage metrics on test key page [PR](https://github.com/BerriAI/litellm/pull/9931)
125+
126+
<Image
127+
img={require('../../img/release_notes/chat_metrics.png')}
128+
style={{width: '100%', display: 'block'}}
129+
/>
130+
<p style={{textAlign: 'left', color: '#666'}}>
131+
View input, output, reasoning tokens, ttft metrics.
132+
</p>
133+
2. Tag / Policy Management:
134+
1. Added Tag/Policy Management [PR](https://github.com/BerriAI/litellm/pull/9813)
135+
136+
<Image
137+
img={require('../../img/release_notes/tag_management.png')}
138+
style={{width: '100%', display: 'block'}}
139+
/>
140+
<p style={{textAlign: 'left', color: '#666'}}>
141+
Tag / Policy Management
142+
</p>
143+
3. Redesigned Login Screen:
144+
1. Polished login screen [PR](https://github.com/BerriAI/litellm/pull/9778)
145+
2. Microsoft SSO Auto-Sync:
146+
1. Added debug route to allow admins to debug SSO JWT fields [PR](https://github.com/BerriAI/litellm/pull/9835)
147+
2. Added ability to use MSFT Graph API to assign users to teams [PR](https://github.com/BerriAI/litellm/pull/9865)
148+
3. Connected LiteLLM to Azure Entra ID Enterprise Application [PR](https://github.com/BerriAI/litellm/pull/9872)
149+
4. Added ability for admins to set `default_team_params` for when litellm SSO creates default teams [PR](https://github.com/BerriAI/litellm/pull/9895)
150+
5. Fixed MSFT SSO to use correct field for user email [PR](https://github.com/BerriAI/litellm/pull/9886)
151+
6. Added UI support for setting Default Team setting when LiteLLM SSO auto creates teams [PR](https://github.com/BerriAI/litellm/pull/9918)
152+
5. UI Bug Fixes:
153+
1. Prevented team, key, org, model numerical values changing on scrolling [PR](https://github.com/BerriAI/litellm/pull/9776)
154+
2. Instantly reflect key and team updates in UI [PR](https://github.com/BerriAI/litellm/pull/9825)
155+
156+
## Logging / Guardrail Improvements
157+
158+
1. Prometheus:
159+
- Emit Key and Team Budget metrics on a cron job schedule [PR](https://github.com/BerriAI/litellm/pull/9528)
160+
161+
## Security Fixes
162+
163+
1. Fixed [CVE-2025-0330](https://www.cve.org/CVERecord?id=CVE-2025-0330) - Leakage of Langfuse API keys in team exception handling [PR](https://github.com/BerriAI/litellm/pull/9830)
164+
2. Fixed [CVE-2024-6825](https://www.cve.org/CVERecord?id=CVE-2024-6825) - Remote code execution in post call rules [PR](https://github.com/BerriAI/litellm/pull/9826)
165+
166+
## Helm
167+
168+
1. Added service annotations to litellm-helm chart [PR](https://github.com/BerriAI/litellm/pull/9840)
169+
2. Added extraEnvVars to the helm deployment [PR](https://github.com/BerriAI/litellm/pull/9292)
170+
171+
## Demo
172+
173+
Try this on the demo instance [today](https://docs.litellm.ai/docs/proxy/demo)
174+
175+
## Complete Git Diff
176+
177+
See the complete git diff since v1.65.4-stable, [here](https://github.com/BerriAI/litellm/releases/tag/v1.66.0-stable)
178+
179+

0 commit comments

Comments
 (0)