Added LiteLLM to the stack
This commit is contained in:
@@ -0,0 +1,291 @@
|
||||
---
|
||||
title: "v1.74.15-stable"
|
||||
slug: "v1-74-15"
|
||||
date: 2025-08-02T10:00:00
|
||||
authors:
|
||||
- name: Krrish Dholakia
|
||||
title: CEO, LiteLLM
|
||||
url: https://www.linkedin.com/in/krish-d/
|
||||
image_url: https://pbs.twimg.com/profile_images/1298587542745358340/DZv3Oj-h_400x400.jpg
|
||||
- name: Ishaan Jaffer
|
||||
title: CTO, LiteLLM
|
||||
url: https://www.linkedin.com/in/reffajnaahsi/
|
||||
image_url: https://pbs.twimg.com/profile_images/1613813310264340481/lz54oEiB_400x400.jpg
|
||||
|
||||
hide_table_of_contents: false
|
||||
---
|
||||
|
||||
import Image from '@theme/IdealImage';
|
||||
import Tabs from '@theme/Tabs';
|
||||
import TabItem from '@theme/TabItem';
|
||||
|
||||
## Deploy this version
|
||||
|
||||
<Tabs>
|
||||
<TabItem value="docker" label="Docker">
|
||||
|
||||
``` showLineNumbers title="docker run litellm"
|
||||
docker run \
|
||||
-e STORE_MODEL_IN_DB=True \
|
||||
-p 4000:4000 \
|
||||
ghcr.io/berriai/litellm:v1.74.15-stable
|
||||
```
|
||||
</TabItem>
|
||||
|
||||
<TabItem value="pip" label="Pip">
|
||||
|
||||
``` showLineNumbers title="pip install litellm"
|
||||
pip install litellm==1.74.15.post2
|
||||
```
|
||||
|
||||
</TabItem>
|
||||
</Tabs>
|
||||
|
||||
---
|
||||
|
||||
## Key Highlights
|
||||
|
||||
- **User Agent Activity Tracking** - Track how much usage each coding tool gets.
|
||||
- **Prompt Management** - Use Git-Ops style prompt management with prompt templates.
|
||||
- **MCP Gateway: Guardrails** - Support for using Guardrails with MCP servers.
|
||||
- **Google AI Studio Imagen4** - Support for using Imagen4 models on Google AI Studio.
|
||||
|
||||
---
|
||||
|
||||
## User Agent Activity Tracking
|
||||
|
||||
<Image
|
||||
img={require('../../img/agent_1.png')}
|
||||
style={{width: '100%', display: 'block', margin: '2rem auto'}}
|
||||
/>
|
||||
|
||||
<br/>
|
||||
|
||||
This release brings support for tracking usage and costs for AI-powered coding tools like Claude Code, Roo Code, Gemini CLI through LiteLLM. You can now track LLM cost, total tokens used, and DAU/WAU/MAU for each coding tool.
|
||||
|
||||
This is great to central AI Platform teams looking to track how they are helping developer productivity.
|
||||
|
||||
[Read More](https://docs.litellm.ai/docs/tutorials/cost_tracking_coding)
|
||||
|
||||
---
|
||||
|
||||
## Prompt Management
|
||||
|
||||
<br/>
|
||||
|
||||
|
||||
|
||||
[Read More](../../docs/proxy/prompt_management)
|
||||
|
||||
---
|
||||
|
||||
## New Models / Updated Models
|
||||
|
||||
#### New Model Support
|
||||
|
||||
| Provider | Model | Context Window | Input ($/1M tokens) | Output ($/1M tokens) | Cost per Image |
|
||||
| ----------- | -------------------------------------- | -------------- | ------------------- | -------------------- | -------------- |
|
||||
| OpenRouter | `openrouter/x-ai/grok-4` | 256k | $3 | $15 | N/A |
|
||||
| Google AI Studio | `gemini/imagen-4.0-generate-preview-06-06` | N/A | N/A | N/A | $0.04 |
|
||||
| Google AI Studio | `gemini/imagen-4.0-ultra-generate-preview-06-06` | N/A | N/A | N/A | $0.06 |
|
||||
| Google AI Studio | `gemini/imagen-4.0-fast-generate-preview-06-06` | N/A | N/A | N/A | $0.02 |
|
||||
| Google AI Studio | `gemini/imagen-3.0-generate-002` | N/A | N/A | N/A | $0.04 |
|
||||
| Google AI Studio | `gemini/imagen-3.0-generate-001` | N/A | N/A | N/A | $0.04 |
|
||||
| Google AI Studio | `gemini/imagen-3.0-fast-generate-001` | N/A | N/A | N/A | $0.02 |
|
||||
|
||||
#### Features
|
||||
|
||||
- **[Google AI Studio](../../docs/providers/gemini)**
|
||||
- Added Google AI Studio Imagen4 model family support - [PR #13065](https://github.com/BerriAI/litellm/pull/13065), [Get Started](../../docs/providers/google_ai_studio/image_gen)
|
||||
- **[Azure OpenAI](../../docs/providers/azure/azure)**
|
||||
- Azure `api_version="preview"` support - [PR #13072](https://github.com/BerriAI/litellm/pull/13072), [Get Started](../../docs/providers/azure/azure#setting-api-version)
|
||||
- Password protected certificate files support - [PR #12995](https://github.com/BerriAI/litellm/pull/12995), [Get Started](../../docs/providers/azure/azure#authentication)
|
||||
- **[AWS Bedrock](../../docs/providers/bedrock)**
|
||||
- Cost tracking via Anthropic `/v1/messages` - [PR #13072](https://github.com/BerriAI/litellm/pull/13072)
|
||||
- Computer use support - [PR #13150](https://github.com/BerriAI/litellm/pull/13150)
|
||||
- **[OpenRouter](../../docs/providers/openrouter)**
|
||||
- Added Grok4 model support - [PR #13018](https://github.com/BerriAI/litellm/pull/13018)
|
||||
- **[Anthropic](../../docs/providers/anthropic)**
|
||||
- Auto Cache Control Injection - Improved cache_control_injection_points with negative index support - [PR #13187](https://github.com/BerriAI/litellm/pull/13187), [Get Started](../../docs/tutorials/prompt_caching)
|
||||
- Working mid-stream fallbacks with token usage tracking - [PR #13149](https://github.com/BerriAI/litellm/pull/13149), [PR #13170](https://github.com/BerriAI/litellm/pull/13170)
|
||||
- **[Perplexity](../../docs/providers/perplexity)**
|
||||
- Citation annotations support - [PR #13225](https://github.com/BerriAI/litellm/pull/13225)
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **[Gemini](../../docs/providers/gemini)**
|
||||
- Fix merge_reasoning_content_in_choices parameter issue - [PR #13066](https://github.com/BerriAI/litellm/pull/13066), [Get Started](../../docs/tutorials/openweb_ui#render-thinking-content-on-open-webui)
|
||||
- Added support for using `GOOGLE_API_KEY` environment variable for Google AI Studio - [PR #12507](https://github.com/BerriAI/litellm/pull/12507)
|
||||
- **[vLLM/OpenAI-like](../../docs/providers/vllm)**
|
||||
- Fix missing extra_headers support for embeddings - [PR #13198](https://github.com/BerriAI/litellm/pull/13198)
|
||||
|
||||
---
|
||||
|
||||
## LLM API Endpoints
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **[/generateContent](../../docs/generateContent)**
|
||||
- Support for query_params in generateContent routes for API Key setting - [PR #13100](https://github.com/BerriAI/litellm/pull/13100)
|
||||
- Ensure "x-goog-api-key" is used for auth to google ai studio when using /generateContent on LiteLLM - [PR #13098](https://github.com/BerriAI/litellm/pull/13098)
|
||||
- Ensure tool calling works as expected on generateContent - [PR #13189](https://github.com/BerriAI/litellm/pull/13189)
|
||||
- **[/vertex_ai (Passthrough)](../../docs/pass_through/vertex_ai)**
|
||||
- Ensure multimodal embedding responses are logged properly - [PR #13050](https://github.com/BerriAI/litellm/pull/13050)
|
||||
|
||||
---
|
||||
|
||||
## [MCP Gateway](../../docs/mcp)
|
||||
|
||||
#### Features
|
||||
|
||||
- **Health Check Improvements**
|
||||
- Add health check endpoints for MCP servers - [PR #13106](https://github.com/BerriAI/litellm/pull/13106)
|
||||
- **Guardrails Integration**
|
||||
- Add pre and during call hooks initialization - [PR #13067](https://github.com/BerriAI/litellm/pull/13067)
|
||||
- Move pre and during hooks to ProxyLogging - [PR #13109](https://github.com/BerriAI/litellm/pull/13109)
|
||||
- MCP pre and during guardrails implementation - [PR #13188](https://github.com/BerriAI/litellm/pull/13188)
|
||||
- **Protocol & Header Support**
|
||||
- Add protocol headers support - [PR #13062](https://github.com/BerriAI/litellm/pull/13062)
|
||||
- **URL & Namespacing**
|
||||
- Improve MCP server URL validation for internal/Kubernetes URLs - [PR #13099](https://github.com/BerriAI/litellm/pull/13099)
|
||||
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **UI**
|
||||
- Fix scrolling issue with MCP tools - [PR #13015](https://github.com/BerriAI/litellm/pull/13015)
|
||||
- Fix MCP client list failure - [PR #13114](https://github.com/BerriAI/litellm/pull/13114)
|
||||
|
||||
|
||||
[Read More](../../docs/mcp)
|
||||
|
||||
|
||||
---
|
||||
|
||||
## Management Endpoints / UI
|
||||
|
||||
#### Features
|
||||
|
||||
- **Usage Analytics**
|
||||
- New tab for user agent activity tracking - [PR #13146](https://github.com/BerriAI/litellm/pull/13146)
|
||||
- Daily usage per user analytics - [PR #13147](https://github.com/BerriAI/litellm/pull/13147)
|
||||
- Default usage chart date range set to last 7 days - [PR #12917](https://github.com/BerriAI/litellm/pull/12917)
|
||||
- New advanced date range picker component - [PR #13141](https://github.com/BerriAI/litellm/pull/13141), [PR #13221](https://github.com/BerriAI/litellm/pull/13221)
|
||||
- Show loader on usage cost charts after date selection - [PR #13113](https://github.com/BerriAI/litellm/pull/13113)
|
||||
- **Models**
|
||||
- Added Voyage, Jinai, Deepinfra and VolcEngine providers on UI - [PR #13131](https://github.com/BerriAI/litellm/pull/13131)
|
||||
- Added Sagemaker on UI - [PR #13117](https://github.com/BerriAI/litellm/pull/13117)
|
||||
- Preserve model order in `/v1/models` and `/model_group/info` endpoints - [PR #13178](https://github.com/BerriAI/litellm/pull/13178)
|
||||
|
||||
- **Key Management**
|
||||
- Properly parse JSON options for key generation in UI - [PR #12989](https://github.com/BerriAI/litellm/pull/12989)
|
||||
- **Authentication**
|
||||
- **JWT Fields**
|
||||
- Add dot notation support for all JWT fields - [PR #13013](https://github.com/BerriAI/litellm/pull/13013)
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **Permissions**
|
||||
- Fix object permission for organizations - [PR #13142](https://github.com/BerriAI/litellm/pull/13142)
|
||||
- Fix list team v2 security check - [PR #13094](https://github.com/BerriAI/litellm/pull/13094)
|
||||
- **Models**
|
||||
- Fix model reload on model update - [PR #13216](https://github.com/BerriAI/litellm/pull/13216)
|
||||
- **Router Settings**
|
||||
- Fix displaying models for fallbacks in UI - [PR #13191](https://github.com/BerriAI/litellm/pull/13191)
|
||||
- Fix wildcard model name handling with custom values - [PR #13116](https://github.com/BerriAI/litellm/pull/13116)
|
||||
- Fix fallback delete functionality - [PR #12606](https://github.com/BerriAI/litellm/pull/12606)
|
||||
|
||||
---
|
||||
|
||||
## Logging / Guardrail Integrations
|
||||
|
||||
#### Features
|
||||
|
||||
- **[MLFlow](../../docs/proxy/logging#mlflow)**
|
||||
- Allow adding tags for MLFlow logging requests - [PR #13108](https://github.com/BerriAI/litellm/pull/13108)
|
||||
- **[Langfuse OTEL](../../docs/proxy/logging#langfuse)**
|
||||
- Add comprehensive metadata support to Langfuse OpenTelemetry integration - [PR #12956](https://github.com/BerriAI/litellm/pull/12956)
|
||||
- **[Datadog LLM Observability](../../docs/proxy/logging#datadog)**
|
||||
- Allow redacting message/response content for specific logging integrations - [PR #13158](https://github.com/BerriAI/litellm/pull/13158)
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **API Key Logging**
|
||||
- Fix API Key being logged inappropriately - [PR #12978](https://github.com/BerriAI/litellm/pull/12978)
|
||||
- **MCP Spend Tracking**
|
||||
- Set default value for MCP namespace tool name in spend table - [PR #12894](https://github.com/BerriAI/litellm/pull/12894)
|
||||
|
||||
---
|
||||
|
||||
## Performance / Loadbalancing / Reliability improvements
|
||||
|
||||
#### Features
|
||||
|
||||
- **Background Health Checks**
|
||||
- Allow disabling background health checks for specific deployments - [PR #13186](https://github.com/BerriAI/litellm/pull/13186)
|
||||
- **Database Connection Management**
|
||||
- Ensure stale Prisma clients disconnect DB connections properly - [PR #13140](https://github.com/BerriAI/litellm/pull/13140)
|
||||
- **Jitter Improvements**
|
||||
- Fix jitter calculation (should be added not multiplied) - [PR #12901](https://github.com/BerriAI/litellm/pull/12901)
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **Anthropic Streaming**
|
||||
- Always use choice index=0 for Anthropic streaming responses - [PR #12666](https://github.com/BerriAI/litellm/pull/12666)
|
||||
- **Custom Auth**
|
||||
- Bubble up custom exceptions properly - [PR #13093](https://github.com/BerriAI/litellm/pull/13093)
|
||||
- **OTEL with Managed Files**
|
||||
- Fix using managed files with OTEL integration - [PR #13171](https://github.com/BerriAI/litellm/pull/13171)
|
||||
|
||||
---
|
||||
|
||||
## General Proxy Improvements
|
||||
|
||||
#### Features
|
||||
|
||||
- **Database Migration**
|
||||
- Move to use_prisma_migrate by default - [PR #13117](https://github.com/BerriAI/litellm/pull/13117)
|
||||
- Resolve team-only models on auth checks - [PR #13117](https://github.com/BerriAI/litellm/pull/13117)
|
||||
- **Infrastructure**
|
||||
- Loosened MCP Python version restrictions - [PR #13102](https://github.com/BerriAI/litellm/pull/13102)
|
||||
- Migrate build_and_test to CI/CD Postgres DB - [PR #13166](https://github.com/BerriAI/litellm/pull/13166)
|
||||
- **Helm Charts**
|
||||
- Allow Helm hooks for migration jobs - [PR #13174](https://github.com/BerriAI/litellm/pull/13174)
|
||||
- Fix Helm migration job schema updates - [PR #12809](https://github.com/BerriAI/litellm/pull/12809)
|
||||
|
||||
#### Bugs
|
||||
|
||||
- **Docker**
|
||||
- Remove obsolete `version` attribute in docker-compose - [PR #13172](https://github.com/BerriAI/litellm/pull/13172)
|
||||
- Add openssl in runtime stage for non-root Dockerfile - [PR #13168](https://github.com/BerriAI/litellm/pull/13168)
|
||||
- **Database Configuration**
|
||||
- Fix DB config through environment variables - [PR #13111](https://github.com/BerriAI/litellm/pull/13111)
|
||||
- **Logging**
|
||||
- Suppress httpx logging - [PR #13217](https://github.com/BerriAI/litellm/pull/13217)
|
||||
- **Token Counting**
|
||||
- Ignore unsupported keys like prefix in token counter - [PR #11954](https://github.com/BerriAI/litellm/pull/11954)
|
||||
---
|
||||
|
||||
## New Contributors
|
||||
* @5731la made their first contribution in https://github.com/BerriAI/litellm/pull/12989
|
||||
* @restato made their first contribution in https://github.com/BerriAI/litellm/pull/12980
|
||||
* @strickvl made their first contribution in https://github.com/BerriAI/litellm/pull/12956
|
||||
* @Ne0-1 made their first contribution in https://github.com/BerriAI/litellm/pull/12995
|
||||
* @maxrabin made their first contribution in https://github.com/BerriAI/litellm/pull/13079
|
||||
* @lvuna made their first contribution in https://github.com/BerriAI/litellm/pull/12894
|
||||
* @Maximgitman made their first contribution in https://github.com/BerriAI/litellm/pull/12666
|
||||
* @pathikrit made their first contribution in https://github.com/BerriAI/litellm/pull/12901
|
||||
* @huetterma made their first contribution in https://github.com/BerriAI/litellm/pull/12809
|
||||
* @betterthanbreakfast made their first contribution in https://github.com/BerriAI/litellm/pull/13029
|
||||
* @phosae made their first contribution in https://github.com/BerriAI/litellm/pull/12606
|
||||
* @sahusiddharth made their first contribution in https://github.com/BerriAI/litellm/pull/12507
|
||||
* @Amit-kr26 made their first contribution in https://github.com/BerriAI/litellm/pull/11954
|
||||
* @kowyo made their first contribution in https://github.com/BerriAI/litellm/pull/13172
|
||||
* @AnandKhinvasara made their first contribution in https://github.com/BerriAI/litellm/pull/13187
|
||||
* @unique-jakub made their first contribution in https://github.com/BerriAI/litellm/pull/13174
|
||||
* @tyumentsev4 made their first contribution in https://github.com/BerriAI/litellm/pull/13134
|
||||
* @aayush-malviya-acquia made their first contribution in https://github.com/BerriAI/litellm/pull/12978
|
||||
* @kankute-sameer made their first contribution in https://github.com/BerriAI/litellm/pull/13225
|
||||
* @AlexanderYastrebov made their first contribution in https://github.com/BerriAI/litellm/pull/13178
|
||||
|
||||
## **[Full Changelog](https://github.com/BerriAI/litellm/compare/v1.74.9-stable...v1.74.15.rc)**
|
Reference in New Issue
Block a user