v1.70.1-stable - Gemini Realtime API Support
Deploy this version​
- Docker
- Pip
docker run
-e STORE_MODEL_IN_DB=True
-p 4000:4000
ghcr.io/berriai/litellm:main-v1.70.1-stable
pip install litellm==1.70.1
Key Highlights​
LiteLLM v1.70.1-stable is live now. Here are the key highlights of this release:
- Gemini Realtime API: You can now call Gemini's Live API via the OpenAI /v1/realtime API
- Spend Logs Retention Period: Enable deleting spend logs older than a certain period.
- PII Masking 2.0: Easily configure masking or blocking specific PII/PHI entities on the UI
Gemini Realtime API​
This release brings support for calling Gemini's realtime models (e.g. gemini-2.0-flash-live) via OpenAI's /v1/realtime API. This is great for developers as it lets them easily switch from OpenAI to Gemini by just changing the model name.
Key Highlights:
- Support for text + audio input/output
- Support for setting session configurations (modality, instructions, activity detection) in the OpenAI format
- Support for logging + usage tracking for realtime sessions
This is currently supported via Google AI Studio. We plan to release VertexAI support over the coming week.
Spend Logs Retention Period​
This release enables deleting LiteLLM Spend Logs older than a certain period. Since we now enable storing the raw request/response in the logs, deleting old logs ensures the database remains performant in production.
PII Masking 2.0​
This release brings improvements to our Presidio PII Integration. As a Proxy Admin, you now have the ability to:
- Mask or block specific entities (e.g., block medical licenses while masking other entities like emails).
- Monitor guardrails in production. LiteLLM Logs will now show you the guardrail run, the entities it detected, and its confidence score for each entity.
New Models / Updated Models​
- Gemini (VertexAI + Google AI Studio)
- Google AI Studio
- /realtime- Gemini Multimodal Live API support
- Audio input/output support, optional param mapping, accurate usage calculation - PR
 
 
- VertexAI
- /chat/completion- Fix llama streaming error - where model response was nested in returned streaming chunk - PR
 
 
- Ollama
- /chat/completion- structure responses fix - PR
 
 
- Bedrock
- Nvidia NIM
- /chat/completion- Add tools, tool_choice, parallel_tool_calls support - PR
 
 
- Novita AI
- New Provider added for /chat/completionroutes - PR
 
- New Provider added for 
- Azure
- /image/generation- Fix azure dall e 3 call with custom model name - PR
 
 
- Cohere
- /embeddings- Migrate embedding to use /v2/embed- adds support for output_dimensions param - PR
 
- Migrate embedding to use 
 
- Anthropic
- /chat/completion- Web search tool support - native + openai format - Get Started
 
 
- VLLM
- /embeddings- Support embedding input as list of integers
 
 
- OpenAI
- /chat/completion- Fix - b64 file data input handling - Get Started
- Add ‘supports_pdf_input’ to all vision models - PR
 
 
LLM API Endpoints​
- Responses API
- Fix delete API support - PR
 
- Rerank API
- /v2/reranknow registered as ‘llm_api_route’ - enabling non-admins to call it - PR
 
Spend Tracking Improvements​
- /chat/completion,- /messages
- /audio/transcription
- /embeddings- Azure AI - Add cohere embed v4 pricing - PR
 
Management Endpoints / UI​
- Models
- Ollama - adds api base param to UI
 
- Logs
- Add team id, key alias, key hash filter on logs - https://github.com/BerriAI/litellm/pull/10831
- Guardrail tracing now in Logs UI - https://github.com/BerriAI/litellm/pull/10893
 
- Teams
- Patch for updating team info when team in org and members not in org - https://github.com/BerriAI/litellm/pull/10835
 
- Guardrails
- Add Bedrock, Presidio, Lakers guardrails on UI - https://github.com/BerriAI/litellm/pull/10874
- See guardrail info page - https://github.com/BerriAI/litellm/pull/10904
- Allow editing guardrails on UI - https://github.com/BerriAI/litellm/pull/10907
 
- Test Key
- select guardrails to test on UI
 
Logging / Alerting Integrations​
- StandardLoggingPayload
- Log any x-headers in requester metadata - Get Started
- Guardrail tracing now in standard logging payload - Get Started
 
- Log any 
- Generic API Logger
- Support passing application/json header
 
- Arize Phoenix
- PagerDuty
- Pagerduty is now a free feature - PR
 
- Alerting
- Sending slack alerts on virtual key/user/team updates is now free - PR
 
Guardrails​
- Guardrails
- New /apply_guardrailendpoint for directly testing a guardrail - PR
 
- New 
- Lakera
- /v2endpoints support - PR
 
- Presidio
- Aim Security
- Support for anonymization in AIM Guardrails - PR
 
Performance / Loadbalancing / Reliability improvements​
- Allow overriding all constants using a .env variable - PR
- Maximum retention period for spend logs
General Proxy Improvements​
- Authentication
- Handle Bearer $LITELLM_API_KEY in x-litellm-api-key custom header PR
 
- New Enterprise pip package - litellm-enterprise- fixes issue whereenterprisefolder was not found when using pip package
- Proxy CLI
- Add models importcommand - PR
 
- Add 
- OpenWebUI
- Configure LiteLLM to Parse User Headers from Open Web UI
 
- LiteLLM Proxy w/ LiteLLM SDK
- Option to force/always use the litellm proxy when calling via LiteLLM SDK
 
New Contributors​
- @imdigitalashish made their first contribution in PR #10617
- @LouisShark made their first contribution in PR #10688
- @OscarSavNS made their first contribution in PR #10764
- @arizedatngo made their first contribution in PR #10654
- @jugaldb made their first contribution in PR #10805
- @daikeren made their first contribution in PR #10781
- @naliotopier made their first contribution in PR #10077
- @damienpontifex made their first contribution in PR #10813
- @Dima-Mediator made their first contribution in PR #10789
- @igtm made their first contribution in PR #10814
- @shibaboy made their first contribution in PR #10752
- @camfarineau made their first contribution in PR #10629
- @ajac-zero made their first contribution in PR #10439
- @damgem made their first contribution in PR #9802
- @hxdror made their first contribution in PR #10757
- @wwwillchen made their first contribution in PR #10894
Demo Instance​
Here's a Demo Instance to test changes:
- Instance: https://demo.litellm.ai/
- Login Credentials:
- Username: admin
- Password: sk-1234
 
