Releases: BerriAI/litellm
v1.61.11-nightly
What's Changed
- fix(team_endpoints.py): allow team member to view team info by @krrishdholakia in #8644
- build: build ui by @krrishdholakia in #8654
- (UI + Proxy) Cache Health Check Page - Cleanup/Improvements by @ishaan-jaff in #8665
- (Bug Fix Redis) - Fix running redis.mget operations with
None
Keys by @ishaan-jaff in #8666 - (Bug fix) prometheus - safely set latency metrics by @ishaan-jaff in #8669
- extract
<think>..</think>
block for amazon deepseek r1 and put inreasoning_content
by @krrishdholakia in #8664 - Add all
/key/generate
api params to UI + add metadata fields on team AND org add/update by @krrishdholakia in #8667
Full Changelog: v1.61.9-nightly...v1.61.11-nightly
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.11-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 120.0 | 146.33082595240526 | 6.457801208431416 | 6.457801208431416 | 1933 | 1933 | 97.35924100004922 | 4080.5825460000165 |
Aggregated | Failed ❌ | 120.0 | 146.33082595240526 | 6.457801208431416 | 6.457801208431416 | 1933 | 1933 | 97.35924100004922 | 4080.5825460000165 |
v1.61.9.dev1
What's Changed
- fix(team_endpoints.py): allow team member to view team info by @krrishdholakia in #8644
- build: build ui by @krrishdholakia in #8654
- (UI + Proxy) Cache Health Check Page - Cleanup/Improvements by @ishaan-jaff in #8665
Full Changelog: v1.61.9-nightly...v1.61.9.dev1
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.9.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 180.0 | 209.72659395983104 | 6.321588488030633 | 6.321588488030633 | 1892 | 1892 | 147.1097109999846 | 3268.0857999999944 |
Aggregated | Failed ❌ | 180.0 | 209.72659395983104 | 6.321588488030633 | 6.321588488030633 | 1892 | 1892 | 147.1097109999846 | 3268.0857999999944 |
v1.61.9-nightly
What's Changed
- Pass router tags in request headers -
x-litellm-tags
+ fix openai metadata param check by @krrishdholakia in #8609 - (Fix) Redis async context usage for Redis Cluster + 94% lower median latency when using Redis Cluster by @ishaan-jaff in #8622
- add openrouter/google/gemini-2.0-flash-001 by @HeMuling in #8619
- feat: add oss license check for related packages by @krrishdholakia in #8623
- fix(model_cost_map): fix json parse error on model cost map + add uni… by @krrishdholakia in #8629
- [Feature]: Redis Caching - Allow setting a namespace for redis cache by @ishaan-jaff in #8624
- Cleanup ui filter icon + pass timeout for Sagemaker messages API by @krrishdholakia in #8630
- Add Elroy to projects built with litellm by @elroy-bot in #8642
- Add OSS license check to ci/cd by @krrishdholakia in #8626
- Fix parallel request limiter on proxy by @krrishdholakia in #8639
- Cleanup user <-> team association on
/team/delete
+ Fix bedrock/deepseek_r1/ translation by @krrishdholakia in #8640 - (Polish/Fixes) - Fixes for Adding Team Specific Models by @ishaan-jaff in #8645
New Contributors
- @HeMuling made their first contribution in #8619
- @elroy-bot made their first contribution in #8642
Full Changelog: v1.61.8-nightly...v1.61.9-nightly
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.9-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 180.0 | 203.54644847482734 | 6.3054769799102575 | 6.3054769799102575 | 1887 | 1887 | 146.3379119999786 | 3805.3281139999626 |
Aggregated | Failed ❌ | 180.0 | 203.54644847482734 | 6.3054769799102575 | 6.3054769799102575 | 1887 | 1887 | 146.3379119999786 | 3805.3281139999626 |
v1.61.8-nightly
What's Changed
- (UI) Allow adding models for a Team (#8598) by @ishaan-jaff in #8601
- (UI) Refactor Add Models for Specific Teams by @ishaan-jaff in #8592
- (UI) Improvements to Add Team Model Flow by @ishaan-jaff in #8603
Full Changelog: v1.61.7...v1.61.8-nightly
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.8-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.8-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.8-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 120.0 | 129.19425708375965 | 6.54112229454407 | 6.54112229454407 | 1958 | 1958 | 94.39574200001744 | 2020.834275000027 |
Aggregated | Failed ❌ | 120.0 | 129.19425708375965 | 6.54112229454407 | 6.54112229454407 | 1958 | 1958 | 94.39574200001744 | 2020.834275000027 |
v1.61.7.dev1
What's Changed
- (UI) Allow adding models for a Team (#8598) by @ishaan-jaff in #8601
- (UI) Refactor Add Models for Specific Teams by @ishaan-jaff in #8592
- (UI) Improvements to Add Team Model Flow by @ishaan-jaff in #8603
Full Changelog: v1.61.7...v1.61.7.dev1
v1.61.7-nightly
What's Changed
- docs: update README.md API key and model example typos by @colesmcintosh in #8590
- Fix typo in main readme by @scosman in #8574
- (UI) Allow adding models for a Team by @ishaan-jaff in #8598
- feat(ui): alert when adding model without STORE_MODEL_IN_DB by @Aditya8840 in #8591
- Revert "(UI) Allow adding models for a Team" by @ishaan-jaff in #8600
- Litellm stable UI 02 17 2025 p1 by @krrishdholakia in #8599
New Contributors
- @colesmcintosh made their first contribution in #8590
- @scosman made their first contribution in #8574
- @Aditya8840 made their first contribution in #8591
Full Changelog: v1.61.6-nightly...v1.61.7-nightly
v1.61.7
What's Changed
- docs(perplexity.md): removing
return_citations
documentation by @miraclebakelaser in #8527 - (docs - cookbook) litellm proxy x langfuse by @ishaan-jaff in #8541
- UI Fixes and Improvements (02/14/2025) p1 by @krrishdholakia in #8546
- (Feat) - Add
/bedrock/meta.llama3-3-70b-instruct-v1:0
tool calling support + cost tracking + base llm unit test for tool calling by @ishaan-jaff in #8545 - fix(general_settings.tsx): filter out empty dictionaries post fallbac… by @krrishdholakia in #8550
- (perf) Fix memory leak on
/completions
route by @ishaan-jaff in #8551 - Org Flow Improvements by @krrishdholakia in #8549
- feat(openai/o_series_transformation.py): support native streaming for o1 by @krrishdholakia in #8552
- fix(team_endpoints.py): fix team info check to handle team keys by @krrishdholakia in #8529
- build: ui build update by @krrishdholakia in #8553
- Optimize Alpine Dockerfile by removing redundant apk commands by @PeterDaveHello in #5016
- fix(main.py): fix key leak error when unknown provider given by @krrishdholakia in #8556
- (Feat) - return
x-litellm-attempted-fallbacks
in responses from litellm proxy by @ishaan-jaff in #8558 - Add remaining org CRUD endpoints + support deleting orgs on UI by @krrishdholakia in #8561
- Enable update/delete org members on UI by @krrishdholakia in #8560
- (Bug Fix) - Add Regenerate Key on Virtual Keys Tab by @ishaan-jaff in #8567
- (Bug Fix + Better Observability) - BudgetResetJob: for reseting key, team, user budgets by @ishaan-jaff in #8562
- (Patch/bug fix) - UI, filter out litellm ui session tokens on Virtual Keys Page by @ishaan-jaff in #8568
- refactor(teams.tsx): refactor to display all teams, across all orgs by @krrishdholakia in #8565
- docs: update README.md API key and model example typos by @colesmcintosh in #8590
- Fix typo in main readme by @scosman in #8574
- (UI) Allow adding models for a Team by @ishaan-jaff in #8598
- feat(ui): alert when adding model without STORE_MODEL_IN_DB by @Aditya8840 in #8591
- Revert "(UI) Allow adding models for a Team" by @ishaan-jaff in #8600
- Litellm stable UI 02 17 2025 p1 by @krrishdholakia in #8599
New Contributors
- @PeterDaveHello made their first contribution in #5016
- @colesmcintosh made their first contribution in #8590
- @scosman made their first contribution in #8574
- @Aditya8840 made their first contribution in #8591
Full Changelog: v1.61.3...v1.61.7
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.7
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.7
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 180.0 | 206.98769618433857 | 6.145029010811349 | 6.145029010811349 | 1839 | 1839 | 146.21495699998377 | 3174.8161250000067 |
Aggregated | Failed ❌ | 180.0 | 206.98769618433857 | 6.145029010811349 | 6.145029010811349 | 1839 | 1839 | 146.21495699998377 | 3174.8161250000067 |
v1.61.6.dev1
What's Changed
- docs: update README.md API key and model example typos by @colesmcintosh in #8590
- Fix typo in main readme by @scosman in #8574
- (UI) Allow adding models for a Team by @ishaan-jaff in #8598
- feat(ui): alert when adding model without STORE_MODEL_IN_DB by @Aditya8840 in #8591
- Revert "(UI) Allow adding models for a Team" by @ishaan-jaff in #8600
- Litellm stable UI 02 17 2025 p1 by @krrishdholakia in #8599
- (UI) Allow adding models for a Team (#8598) by @ishaan-jaff in #8601
- (UI) Refactor Add Models for Specific Teams by @ishaan-jaff in #8592
- (UI) Improvements to Add Team Model Flow by @ishaan-jaff in #8603
New Contributors
- @colesmcintosh made their first contribution in #8590
- @scosman made their first contribution in #8574
- @Aditya8840 made their first contribution in #8591
Full Changelog: v1.61.6-nightly...v1.61.6.dev1
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.6.dev1
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 170.0 | 197.04136517618934 | 6.316924319787487 | 6.316924319787487 | 1890 | 1890 | 142.7094059999945 | 2646.323271999961 |
Aggregated | Failed ❌ | 170.0 | 197.04136517618934 | 6.316924319787487 | 6.316924319787487 | 1890 | 1890 | 142.7094059999945 | 2646.323271999961 |
v1.61.6-nightly
What's Changed
- refactor(teams.tsx): refactor to display all teams, across all orgs by @krrishdholakia in #8565
Full Changelog: v1.61.5-nightly...v1.61.6-nightly
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.6-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 170.0 | 197.37858561234376 | 6.172709160882249 | 6.172709160882249 | 1847 | 1847 | 139.8097940000298 | 3194.1706680000266 |
Aggregated | Failed ❌ | 170.0 | 197.37858561234376 | 6.172709160882249 | 6.172709160882249 | 1847 | 1847 | 139.8097940000298 | 3194.1706680000266 |
v1.61.5-nightly
What's Changed
- Optimize Alpine Dockerfile by removing redundant apk commands by @PeterDaveHello in #5016
- fix(main.py): fix key leak error when unknown provider given by @krrishdholakia in #8556
- (Feat) - return
x-litellm-attempted-fallbacks
in responses from litellm proxy by @ishaan-jaff in #8558 - Add remaining org CRUD endpoints + support deleting orgs on UI by @krrishdholakia in #8561
- Enable update/delete org members on UI by @krrishdholakia in #8560
- (Bug Fix) - Add Regenerate Key on Virtual Keys Tab by @ishaan-jaff in #8567
- (Bug Fix + Better Observability) - BudgetResetJob: for reseting key, team, user budgets by @ishaan-jaff in #8562
- (Patch/bug fix) - UI, filter out litellm ui session tokens on Virtual Keys Page by @ishaan-jaff in #8568
New Contributors
- @PeterDaveHello made their first contribution in #5016
Full Changelog: v1.61.3.dev1...v1.61.5-nightly
## Docker Run LiteLLM Proxy
```
docker run \
-e STORE_MODEL_IN_DB=True \
-p 4000:4000 \
ghcr.io/berriai/litellm:main-v1.61.5-nightly
Don't want to maintain your internal proxy? get in touch 🎉
Hosted Proxy Alpha: https://calendly.com/d/4mp-gd3-k5k/litellm-1-1-onboarding-chat
Load Test LiteLLM Proxy Results
Name | Status | Median Response Time (ms) | Average Response Time (ms) | Requests/s | Failures/s | Request Count | Failure Count | Min Response Time (ms) | Max Response Time (ms) |
---|---|---|---|---|---|---|---|---|---|
/chat/completions | Failed ❌ | 150.0 | 169.92952748954406 | 6.233287189548679 | 6.233287189548679 | 1865 | 1865 | 130.2254270000276 | 1515.568768999998 |
Aggregated | Failed ❌ | 150.0 | 169.92952748954406 | 6.233287189548679 | 6.233287189548679 | 1865 | 1865 | 130.2254270000276 | 1515.568768999998 |