These are the changes since v1.63.2-stable
.
This release is primarily focused on:
- [Beta] Responses API Support
- Snowflake Cortex Support
- UI - Credential Management, re-use credentials when adding new models
- UI - Test Connection to LLM Provider before adding a model
This release will be live on 03/16/2025
Demo Instance​
Here's a Demo Instance to test changes:
- Instance: https://demo.litellm.ai/
- Login Credentials:
- Username: admin
- Password: sk-1234
New / Updated Models in Model Cost Map​
- Add support for Amazon Nova Canvas model PR
- Add pricing for Jamba new models PR
- Add pricing for Amazon EU models PR
- Add Bedrock Deepseek R1 model pricing PR
- Update Gemini pricing: Gemma 3, Flash 2 thinking update, LearnLM PR
- Mark Cohere Embedding 3 models as Multimodal PR
- Add Azure Data Zone pricing PR
- LiteLLM Tracks cost for
azure/eu
andazure/us
models
- LiteLLM Tracks cost for
LLM Translation​
- New Endpoints
- [Beta] POST
/responses
API. Get Started
- New LLM Providers
- Snowflake Cortex Get Started
- New models
- Support OpenRouter
reasoning_content
on streaming Get Started - Support Bedrock converse cache token tracking Get Started
- Bug Fixes
- Return
code
,param
andtype
on OpenAI bad request error More information on litellm exceptions - Fix Bedrock converse chunk parsing to only return empty dict on tool use PR
- Fix Azure Function Calling Bug & Update Default API Version to
2025-02-01-preview
PR - Fix Perplexity incorrect streaming response PR
- Fix Triton streaming completions bug PR
- Fix: String
data:
stripped from entire content in streamed Gemini responses PR - Fix: Support bytes.IO when handling audio files for transcription PR
- Fix: "system" role has become unacceptable in Ollama PR
- Handle HTTP 201 status code in Vertex AI response PR
Spend Tracking Improvements​
- Cost Tracking for Responses API Get Started
- Fix Azure Whisper cost tracking PR
UI​
Re-Use Credentials on UI​
You can now onboard LLM provider credentials on LiteLLM UI. Once these credentials are added you can re-use them when adding new models
Test Connections before adding models​
Before adding a model you can test the connection to the LLM provider to verify you have setup your API Base + API Key correctly

General UI Improvements​
- Add Models Page
- Allow adding Cerebras, Sambanova, Perplexity, Fireworks, Openrouter, TogetherAI Models, Text-Completion OpenAI on Admin UI
- Allow adding EU OpenAI models
- Fix: Instantly show edit + deletes to models
- Keys Page
- Fix: Instantly show newly created keys on Admin UI (don't require refresh)
- Fix: Allow clicking into Top Keys when showing users Top API Key
- Fix: Allow Filter Keys by Team Alias, Key Alias and Org
- UI Improvements: Show 100 Keys Per Page, Use full height, increase width of key alias
- Users Page
- Fix: Show correct count of internal user keys on Users Page
- Fix: Metadata not updating in Team UI
- Logs Page
- UI Improvements: Keep expanded log in focus on LiteLLM UI
- UI Improvements: Minor improvements to logs page
- Fix: Allow internal user to query their own logs
Security​
- Support for Rotating Master Keys Getting Started
- Fix: Internal User Viewer Permissions, don't allow
internal_user_viewer
role to seeTest Key Page
orCreate Key Button
Role based access controls - Emit audit logs on All user + model Create/Update/Delete endpoints Get Started
- JWT
- Support multiple JWT OIDC providers Get Started
- Fix JWT access with Groups not working when team is assigned All Proxy Models access
- Using K/V pairs in 1 AWS Secret Get Started
Logging Integrations​
- Prometheus: Track Azure LLM API latency metric Get Started here
- Allow switching off storing Error Logs in DB Get Started here
- Added tags, user_feedback and model_options to additional_keys which can be sent to Athina Get Started here
Performance / Reliability improvements​
- Fix Redis cluster mode for routers PR
- Delegate router Azure client init logic to Azure provider PR
- Fix Azure AI services URL PR
- Support extra_headers on Bedrock PR
General Improvements​
- UI API Playground for testing LiteLLM translation PR
- Fix: Correctly use
PROXY_LOGOUT_URL
when set PR - Bing Search Pass Through endpoint PR
- OpenWebUI Integration - display
thinking
tokens
- Guide on getting started with LiteLLM x OpenWebUI. Get Started
- Display
thinking
tokens on OpenWebUI (Bedrock, Anthropic, Deepseek) Get Started
