Releases: BerriAI/litellm
Releases · BerriAI/litellm
v1.79.1.rc.2
Full Changelog: v1.79.1-nightly...v1.79.1.rc.2
v1.79.1.dev5
What's Changed
- (feat) Milvus - search vector store support + (fix) Passthrough Endpoints - support multi-part form data on passthrough by @krrishdholakia in #16035
- (feat) Azure AI Vector Stores - support "virtual" indexes + create vector store on passthrough API by @krrishdholakia in #16160
- Add
custom_llm_providersupport for video endpoints (non-generation) by @Sameerlite in #16121 - fix: resolve memory accumulation caused by Pydantic 2.11+ deprecation warnings by @AlexsanderHamir in #16110
- add: minimum resource requirement for production by @AlexsanderHamir in #16146
- Remove minimum validation for cache control injection index by @steve-gore-snapdocs in #16149
- Fix empty assistant message handling in AWS Bedrock Converse API to prevent 400 Bad Request errors by @Sameerlite in #15850
- Add LLM provider response headers to Responses API by @Sameerlite in #16091
- Fix azure doesn't accept extra body param by @Sameerlite in #16116
- Add E2E Container API Support by @Sameerlite in #16136
- Add new model pricing for nemotron-nano-9b-v2 by @timbmg in #16120
- docs - v1.79.1 by @ishaan-jaff in #16163
- UI - Fix regression where Guardrail Entity Could not be selected and entity was not displayed by @ishaan-jaff in #16165
- Fix Anthropic token counting for VertexAI by @steve-gore-snapdocs in #16171
- [Feature] UI - Delete Team Member with friction by @yuneng-jiang in #16167
- [Feature] UI - Add Model Existing Credentials Improvement by @yuneng-jiang in #16166
- [Feature] UI - Guardrail Info Page Show PII Config by @yuneng-jiang in #16164
- strip base64 in s3 by @deepanshululla in #16157
- fix(ui_sso.py): support dot notation on ui sso by @krrishdholakia in #16135
- Support responses API streaming in langfuse otel by @eycjur in #16153
- fix: return the diarized transcript when it's required in the request by @gvioss in #16133
- build: Squashed commit of the following: by @krrishdholakia in #16176
- feat: change guardrail_information to list type by @uc4w6c in #16127
- Milvus - Passthrough API support - adds create + read vector store support via passthrough API's by @krrishdholakia in #16170
- Fix index field not populated in streaming mode with n>1 and tool calls by @Sameerlite in #15962
- Update perplexity cost tracking by @Sameerlite in #15743
- Fix image_config.aspect_ratio not working for gemini-2.5-flash-image by @Sameerlite in #15999
New Contributors
- @steve-gore-snapdocs made their first contribution in #16149
- @timbmg made their first contribution in #16120
Full Changelog: v1.79.1-nightly...v1.79.1.dev5
v1.79.1.dev3
What's Changed
- (feat) Milvus - search vector store support + (fix) Passthrough Endpoints - support multi-part form data on passthrough by @krrishdholakia in #16035
- (feat) Azure AI Vector Stores - support "virtual" indexes + create vector store on passthrough API by @krrishdholakia in #16160
- Add
custom_llm_providersupport for video endpoints (non-generation) by @Sameerlite in #16121 - fix: resolve memory accumulation caused by Pydantic 2.11+ deprecation warnings by @AlexsanderHamir in #16110
- add: minimum resource requirement for production by @AlexsanderHamir in #16146
- Remove minimum validation for cache control injection index by @steve-gore-snapdocs in #16149
- Fix empty assistant message handling in AWS Bedrock Converse API to prevent 400 Bad Request errors by @Sameerlite in #15850
- Add LLM provider response headers to Responses API by @Sameerlite in #16091
- Fix azure doesn't accept extra body param by @Sameerlite in #16116
- Add E2E Container API Support by @Sameerlite in #16136
- Add new model pricing for nemotron-nano-9b-v2 by @timbmg in #16120
- docs - v1.79.1 by @ishaan-jaff in #16163
- UI - Fix regression where Guardrail Entity Could not be selected and entity was not displayed by @ishaan-jaff in #16165
- Fix Anthropic token counting for VertexAI by @steve-gore-snapdocs in #16171
- [Feature] UI - Delete Team Member with friction by @yuneng-jiang in #16167
- [Feature] UI - Add Model Existing Credentials Improvement by @yuneng-jiang in #16166
- [Feature] UI - Guardrail Info Page Show PII Config by @yuneng-jiang in #16164
- strip base64 in s3 by @deepanshululla in #16157
- fix(ui_sso.py): support dot notation on ui sso by @krrishdholakia in #16135
- Support responses API streaming in langfuse otel by @eycjur in #16153
- fix: return the diarized transcript when it's required in the request by @gvioss in #16133
- build: Squashed commit of the following: by @krrishdholakia in #16176
- feat: change guardrail_information to list type by @uc4w6c in #16127
- Milvus - Passthrough API support - adds create + read vector store support via passthrough API's by @krrishdholakia in #16170
- Fix index field not populated in streaming mode with n>1 and tool calls by @Sameerlite in #15962
- Update perplexity cost tracking by @Sameerlite in #15743
- Fix image_config.aspect_ratio not working for gemini-2.5-flash-image by @Sameerlite in #15999
New Contributors
- @steve-gore-snapdocs made their first contribution in #16149
- @timbmg made their first contribution in #16120
Full Changelog: v1.79.1-nightly...v1.79.1.dev3
v1.79.1.dev2
What's Changed
- (feat) Milvus - search vector store support + (fix) Passthrough Endpoints - support multi-part form data on passthrough by @krrishdholakia in #16035
- (feat) Azure AI Vector Stores - support "virtual" indexes + create vector store on passthrough API by @krrishdholakia in #16160
- Add
custom_llm_providersupport for video endpoints (non-generation) by @Sameerlite in #16121 - fix: resolve memory accumulation caused by Pydantic 2.11+ deprecation warnings by @AlexsanderHamir in #16110
- add: minimum resource requirement for production by @AlexsanderHamir in #16146
- Remove minimum validation for cache control injection index by @steve-gore-snapdocs in #16149
- Fix empty assistant message handling in AWS Bedrock Converse API to prevent 400 Bad Request errors by @Sameerlite in #15850
- Add LLM provider response headers to Responses API by @Sameerlite in #16091
- Fix azure doesn't accept extra body param by @Sameerlite in #16116
- Add E2E Container API Support by @Sameerlite in #16136
- Add new model pricing for nemotron-nano-9b-v2 by @timbmg in #16120
- docs - v1.79.1 by @ishaan-jaff in #16163
- UI - Fix regression where Guardrail Entity Could not be selected and entity was not displayed by @ishaan-jaff in #16165
- Fix Anthropic token counting for VertexAI by @steve-gore-snapdocs in #16171
- [Feature] UI - Delete Team Member with friction by @yuneng-jiang in #16167
- [Feature] UI - Add Model Existing Credentials Improvement by @yuneng-jiang in #16166
- [Feature] UI - Guardrail Info Page Show PII Config by @yuneng-jiang in #16164
- strip base64 in s3 by @deepanshululla in #16157
- fix(ui_sso.py): support dot notation on ui sso by @krrishdholakia in #16135
- Support responses API streaming in langfuse otel by @eycjur in #16153
- fix: return the diarized transcript when it's required in the request by @gvioss in #16133
- build: Squashed commit of the following: by @krrishdholakia in #16176
- feat: change guardrail_information to list type by @uc4w6c in #16127
- Milvus - Passthrough API support - adds create + read vector store support via passthrough API's by @krrishdholakia in #16170
- Fix index field not populated in streaming mode with n>1 and tool calls by @Sameerlite in #15962
- Update perplexity cost tracking by @Sameerlite in #15743
- Fix image_config.aspect_ratio not working for gemini-2.5-flash-image by @Sameerlite in #15999
New Contributors
- @steve-gore-snapdocs made their first contribution in #16149
- @timbmg made their first contribution in #16120
Full Changelog: v1.79.1-nightly...v1.79.1.dev2
v1.79.1.rc.1
What's Changed
- 1-79-0 docs by @ishaan-jaff in #15936
- feat(lasso): Upgrade to Lasso API v3 and fix ULID generation by @oroxenberg in #15941
- Enable OpenTelemetry context propagation by external tracers by @eycjur in #15940
- Fix documentation for videos by @Sameerlite in #15937
- Fix duplicate trace in langfuse_otel by @eycjur in #15931
- [Feat] add support for dynamic client registration by @uc4w6c in #15921
- Update IBM Guardrails to correctly use SSL Verify argument by @RobGeada in #15975
- feat: support during_call for model armor guardrails by @bjornjee in #15970
- docs(openrouter): add base_url config with environment variables by @shanto12 in #15946
- [Buf fix] - Azure OpenAI, fix ContextWindowExceededError is not mapped from Azure openai errors by @ishaan-jaff in #15981
- [Fix] DD logging - ensure key's metadata + guardrail is logged on DD by @ishaan-jaff in #15980
- [Feat] OTEL - Ensure error information is logged on OTEL by @ishaan-jaff in #15978
- [Fix] Minor fix proxy - ensure User API key and team id and user id missing from custom callback is not misfiring by @ishaan-jaff in #15982
- [Fix] Azure OpenAI - Add handling for
v1under azure api versions by @ishaan-jaff in #15984 - Fix: Respect
LiteLLM-Disable-Message-Redactionheader for Responses API by @Sameerlite in #15966 - [Feat] UI - Changed API Base from Select to Input in New LLM Credentials by @yuneng-jiang in #15987
- [Bug Fix] Remove limit from admin UI numerical input fix by @yuneng-jiang in #15991
- [Feature] UI - Key Already Exist Error Notification by @yuneng-jiang in #15993
- [Fix] - Responses API - add /openai routes for responses API. (Azure OpenAI SDK Compatibility) by @ishaan-jaff in #15988
- Add deprecation dates for models by @dima-hx430 in #15976
- docs(guardrails/ibm_guardrails): add additional detail to
ibm_guardrails.mdby @m-misiura in #15971 - Perf speed up pytest by @uc4w6c in #15951
- fix: Preserve Bedrock inference profile IDs in health checks by @ylgibby in #15947
- Fix: Support tool usage messages with Langfuse OTEL integration by @eycjur in #15932
- Add Haiku 4.5 pricing for open router by @Somtom in #15909
- fix(opik): enhance requester metadata retrieval from API key auth by @Thomas-Mildner in #15897
- [feat]: graceful degradation for pillar service when using litellm by @afogel in #15857
- Add GitlabPromptCache and enable subfolder access by @deepanshululla in #15712
- Add OpenAI client usage documentation for videos and fix navigation visibility by @Sameerlite in #15996
- [Feature] Config Models should not be editable by @yuneng-jiang in #16020
- [Fix] Guardrails - Ensure Key Guardrails are applied by @ishaan-jaff in #16025
- [UI] Feature - Add Apply Guardrail Testing Playground by @ishaan-jaff in #16030
- [Fix] SQS Logger - Add Base64 handling by @ishaan-jaff in #16028
- Fix mutation of original request for gemini request by @Sameerlite in #16002
- Fix: Redact reasoning summaries in ResponsesAPI output when message logging is disabled by @Sameerlite in #15965
- fix: Support text.format parameter in Responses API for providers without native ResponsesAPIConfig by @rodolfo-nobrega in #16023
- Remove unnecessary model variable assignment by @Mte90 in #16008
- Add license metadata to health/readiness endpoint. by @bernata in #15997
- chore(deps): bump hono from 4.9.7 to 4.10.3 in /litellm-js/spend-logs by @dependabot[bot] in #15915
- docs: improve Grayswan guardrail documentation by @TeddyAmkie in #15875
- fix(apscheduler): prevent memory leaks from jitter and frequent job intervals by @jatorre in #15846
- Python entry-point for CustomLLM subclasses by @AlbertDeFusco in #15881
- Allow using ARNs when generation images via Bedrock by @komarovd95 in #15789
- Added fallback logic for detecting file content-type when S3 returns generic by @langpingxue in #15635
- fix: prevent httpx DeprecationWarning memory leak in AsyncHTTPHandler by @AlexsanderHamir in #16024
- [Feat] Add FAL AI Image Generations on LiteLLM by @ishaan-jaff in #16067
- Feat: Mistral API - add codestral-embed-2505 by @ishaan-jaff in #16071
- Fix spend tracking for OCR/aOCR requests (log
pages_processed+ recognizeOCRResponse) by @OrionCodeDev in #16070 - [Feat] Add support for Batch API Rate limiting - PR1 adds support for input based rate limits by @ishaan-jaff in #16075
- [Feature] UI - Validation for Proxy Base URL in SSO Settings by @yuneng-jiang in #16082
- [Feature] UI - Test Key UI Embeddings by @yuneng-jiang in #16065
- [Feature] UI - Add Key Type Select in Key Settings by @yuneng-jiang in #16034
- feat(guardrails): Add per-request profile overrides to PANW Prisma AIRS by @jroberts2600 in #16069
- docs: use custom-llm-provider header in examples by @tlecomte in #16055
- Improve(mcp): respect X-Forwarded- headers in OAuth endpoints by @talalryz in #16036
- Add OpenAI-compatible annotations support for Cohere v2 citations by @Sameerlite in #16038
- Opik user auth key metadata Documentation by @Thomas-Mildner in #16004
- [Fix] Circle CI UI E2E Test Fix by @yuneng-jiang in #16100
- Fix: Moderations endpoint now respects
api_baseconfiguration parameter by @Sameerlite in #16087 - [Infra] Litellm Backend SSO Changes by @yuneng-jiang in #16029
- Litellm docs readme fixes by @ishaan-jaff in #16107
- [Docs] readme fixes add supported providers by @ishaan-jaff in #16109
- Add gemini/gemini-embedding-001 pricing entry for Google GenAI API by @sbinnee in #16078
- Add supports vision field, to qwen-vl models, in model_prices_and_context_window.json by @JetoPistola in #16106
- feat: handle Deepgram detected language when available by @gvioss in #16093
- Fix vercel_ai_gateway entry for glm-4.6 by @pale-aura in #16084
- Add OCI Signer Authentication. Closes #16048, Closes #15654 by @gotsysdba in #16064
- Changes to fix frequency_penalty and presence_penalty issue for gemini-2.5-pro model by @tanvithakur94 in #16041
- Add Add per model group header forwarding for Bedrock Invoke API by @Sameerlite in #16042
- fix: Handle multiple rate limit types per descriptor and prevent IndexError by @orolega in #16039
- prompt_caching.md: Fix wrong prompt_tokens definition by @li-boxuan in #16044
- Add Xai websearch cost by @Sameerlite in #16001
- Don't add "accounts/fireworks/models" prefix for Fireworks Provider by @1stprinciple in #15938
- fix: make debug log debug by @raghav-stripe in #16137
- [Feat] UI + Management - Allow setting all routing strategies, tag filtering on UI by @ishaan-jaff in #16139
- [Feature] UI - Add tags in model creation by @yuneng-jiang in #16138
- [Feature] UI - Add Tags to update model by @yuneng-jiang in #16140
- [Feature] UI - Config Guardrails should not be editable and guardrail info fix by @yuneng-jiang in #16142
- [Feat] UI - Allow setting cache settings on UI by @ishaan-jaff in #16143
- build(deps): bump starlette from 0.47.2 to 0.49.1 by @dependabot[bot] in https://github.com/BerriAI/lite...
v1.79.1-nightly
What's Changed
- 1-79-0 docs by @ishaan-jaff in #15936
- feat(lasso): Upgrade to Lasso API v3 and fix ULID generation by @oroxenberg in #15941
- Enable OpenTelemetry context propagation by external tracers by @eycjur in #15940
- Fix documentation for videos by @Sameerlite in #15937
- Fix duplicate trace in langfuse_otel by @eycjur in #15931
- [Feat] add support for dynamic client registration by @uc4w6c in #15921
- Update IBM Guardrails to correctly use SSL Verify argument by @RobGeada in #15975
- feat: support during_call for model armor guardrails by @bjornjee in #15970
- docs(openrouter): add base_url config with environment variables by @shanto12 in #15946
- [Buf fix] - Azure OpenAI, fix ContextWindowExceededError is not mapped from Azure openai errors by @ishaan-jaff in #15981
- [Fix] DD logging - ensure key's metadata + guardrail is logged on DD by @ishaan-jaff in #15980
- [Feat] OTEL - Ensure error information is logged on OTEL by @ishaan-jaff in #15978
- [Fix] Minor fix proxy - ensure User API key and team id and user id missing from custom callback is not misfiring by @ishaan-jaff in #15982
- [Fix] Azure OpenAI - Add handling for
v1under azure api versions by @ishaan-jaff in #15984 - Fix: Respect
LiteLLM-Disable-Message-Redactionheader for Responses API by @Sameerlite in #15966 - [Feat] UI - Changed API Base from Select to Input in New LLM Credentials by @yuneng-jiang in #15987
- [Bug Fix] Remove limit from admin UI numerical input fix by @yuneng-jiang in #15991
- [Feature] UI - Key Already Exist Error Notification by @yuneng-jiang in #15993
- [Fix] - Responses API - add /openai routes for responses API. (Azure OpenAI SDK Compatibility) by @ishaan-jaff in #15988
- Add deprecation dates for models by @dima-hx430 in #15976
- docs(guardrails/ibm_guardrails): add additional detail to
ibm_guardrails.mdby @m-misiura in #15971 - Perf speed up pytest by @uc4w6c in #15951
- fix: Preserve Bedrock inference profile IDs in health checks by @ylgibby in #15947
- Fix: Support tool usage messages with Langfuse OTEL integration by @eycjur in #15932
- Add Haiku 4.5 pricing for open router by @Somtom in #15909
- fix(opik): enhance requester metadata retrieval from API key auth by @Thomas-Mildner in #15897
- [feat]: graceful degradation for pillar service when using litellm by @afogel in #15857
- Add GitlabPromptCache and enable subfolder access by @deepanshululla in #15712
- Add OpenAI client usage documentation for videos and fix navigation visibility by @Sameerlite in #15996
- [Feature] Config Models should not be editable by @yuneng-jiang in #16020
- [Fix] Guardrails - Ensure Key Guardrails are applied by @ishaan-jaff in #16025
- [UI] Feature - Add Apply Guardrail Testing Playground by @ishaan-jaff in #16030
- [Fix] SQS Logger - Add Base64 handling by @ishaan-jaff in #16028
- Fix mutation of original request for gemini request by @Sameerlite in #16002
- Fix: Redact reasoning summaries in ResponsesAPI output when message logging is disabled by @Sameerlite in #15965
- fix: Support text.format parameter in Responses API for providers without native ResponsesAPIConfig by @rodolfo-nobrega in #16023
- Remove unnecessary model variable assignment by @Mte90 in #16008
- Add license metadata to health/readiness endpoint. by @bernata in #15997
- chore(deps): bump hono from 4.9.7 to 4.10.3 in /litellm-js/spend-logs by @dependabot[bot] in #15915
- docs: improve Grayswan guardrail documentation by @TeddyAmkie in #15875
- fix(apscheduler): prevent memory leaks from jitter and frequent job intervals by @jatorre in #15846
- Python entry-point for CustomLLM subclasses by @AlbertDeFusco in #15881
- Allow using ARNs when generation images via Bedrock by @komarovd95 in #15789
- Added fallback logic for detecting file content-type when S3 returns generic by @langpingxue in #15635
- fix: prevent httpx DeprecationWarning memory leak in AsyncHTTPHandler by @AlexsanderHamir in #16024
- [Feat] Add FAL AI Image Generations on LiteLLM by @ishaan-jaff in #16067
- Feat: Mistral API - add codestral-embed-2505 by @ishaan-jaff in #16071
- Fix spend tracking for OCR/aOCR requests (log
pages_processed+ recognizeOCRResponse) by @OrionCodeDev in #16070 - [Feat] Add support for Batch API Rate limiting - PR1 adds support for input based rate limits by @ishaan-jaff in #16075
- [Feature] UI - Validation for Proxy Base URL in SSO Settings by @yuneng-jiang in #16082
- [Feature] UI - Test Key UI Embeddings by @yuneng-jiang in #16065
- [Feature] UI - Add Key Type Select in Key Settings by @yuneng-jiang in #16034
- feat(guardrails): Add per-request profile overrides to PANW Prisma AIRS by @jroberts2600 in #16069
- docs: use custom-llm-provider header in examples by @tlecomte in #16055
- Improve(mcp): respect X-Forwarded- headers in OAuth endpoints by @talalryz in #16036
- Add OpenAI-compatible annotations support for Cohere v2 citations by @Sameerlite in #16038
- Opik user auth key metadata Documentation by @Thomas-Mildner in #16004
- [Fix] Circle CI UI E2E Test Fix by @yuneng-jiang in #16100
- Fix: Moderations endpoint now respects
api_baseconfiguration parameter by @Sameerlite in #16087 - [Infra] Litellm Backend SSO Changes by @yuneng-jiang in #16029
- Litellm docs readme fixes by @ishaan-jaff in #16107
- [Docs] readme fixes add supported providers by @ishaan-jaff in #16109
- Add gemini/gemini-embedding-001 pricing entry for Google GenAI API by @sbinnee in #16078
- Add supports vision field, to qwen-vl models, in model_prices_and_context_window.json by @JetoPistola in #16106
- feat: handle Deepgram detected language when available by @gvioss in #16093
- Fix vercel_ai_gateway entry for glm-4.6 by @pale-aura in #16084
- Add OCI Signer Authentication. Closes #16048, Closes #15654 by @gotsysdba in #16064
- Changes to fix frequency_penalty and presence_penalty issue for gemini-2.5-pro model by @tanvithakur94 in #16041
- Add Add per model group header forwarding for Bedrock Invoke API by @Sameerlite in #16042
- fix: Handle multiple rate limit types per descriptor and prevent IndexError by @orolega in #16039
- prompt_caching.md: Fix wrong prompt_tokens definition by @li-boxuan in #16044
- Add Xai websearch cost by @Sameerlite in #16001
- Don't add "accounts/fireworks/models" prefix for Fireworks Provider by @1stprinciple in #15938
- fix: make debug log debug by @raghav-stripe in #16137
- [Feat] UI + Management - Allow setting all routing strategies, tag filtering on UI by @ishaan-jaff in #16139
- [Feature] UI - Add tags in model creation by @yuneng-jiang in #16138
- [Feature] UI - Add Tags to update model by @yuneng-jiang in #16140
- [Feature] UI - Config Guardrails should not be editable and guardrail info fix by @yuneng-jiang in #16142
- [Feat] UI - Allow setting cache settings on UI by @ishaan-jaff in #16143
- build(deps): bump starlette from 0.47.2 to 0.49.1 by @dependabot[bot] in https://github.com/BerriAI/lite...
v1.79.0-stable
Full Changelog: v1.79.0-nightly...v1.79.0-stable
v1.78.5.rc.4
Full Changelog: v1.78.5-nightly...v1.78.5.rc.4
v1.79.1.dev1
What's Changed
- 1-79-0 docs by @ishaan-jaff in #15936
- feat(lasso): Upgrade to Lasso API v3 and fix ULID generation by @oroxenberg in #15941
- Enable OpenTelemetry context propagation by external tracers by @eycjur in #15940
- Fix documentation for videos by @Sameerlite in #15937
- Fix duplicate trace in langfuse_otel by @eycjur in #15931
- [Feat] add support for dynamic client registration by @uc4w6c in #15921
- Update IBM Guardrails to correctly use SSL Verify argument by @RobGeada in #15975
- feat: support during_call for model armor guardrails by @bjornjee in #15970
- docs(openrouter): add base_url config with environment variables by @shanto12 in #15946
- [Buf fix] - Azure OpenAI, fix ContextWindowExceededError is not mapped from Azure openai errors by @ishaan-jaff in #15981
- [Fix] DD logging - ensure key's metadata + guardrail is logged on DD by @ishaan-jaff in #15980
- [Feat] OTEL - Ensure error information is logged on OTEL by @ishaan-jaff in #15978
- [Fix] Minor fix proxy - ensure User API key and team id and user id missing from custom callback is not misfiring by @ishaan-jaff in #15982
- [Fix] Azure OpenAI - Add handling for
v1under azure api versions by @ishaan-jaff in #15984 - Fix: Respect
LiteLLM-Disable-Message-Redactionheader for Responses API by @Sameerlite in #15966 - [Feat] UI - Changed API Base from Select to Input in New LLM Credentials by @yuneng-jiang in #15987
- [Bug Fix] Remove limit from admin UI numerical input fix by @yuneng-jiang in #15991
- [Feature] UI - Key Already Exist Error Notification by @yuneng-jiang in #15993
- [Fix] - Responses API - add /openai routes for responses API. (Azure OpenAI SDK Compatibility) by @ishaan-jaff in #15988
- Add deprecation dates for models by @dima-hx430 in #15976
- docs(guardrails/ibm_guardrails): add additional detail to
ibm_guardrails.mdby @m-misiura in #15971 - Perf speed up pytest by @uc4w6c in #15951
- fix: Preserve Bedrock inference profile IDs in health checks by @ylgibby in #15947
- Fix: Support tool usage messages with Langfuse OTEL integration by @eycjur in #15932
- Add Haiku 4.5 pricing for open router by @Somtom in #15909
- fix(opik): enhance requester metadata retrieval from API key auth by @Thomas-Mildner in #15897
- [feat]: graceful degradation for pillar service when using litellm by @afogel in #15857
- Add GitlabPromptCache and enable subfolder access by @deepanshululla in #15712
- Add OpenAI client usage documentation for videos and fix navigation visibility by @Sameerlite in #15996
- [Feature] Config Models should not be editable by @yuneng-jiang in #16020
- [Fix] Guardrails - Ensure Key Guardrails are applied by @ishaan-jaff in #16025
- [UI] Feature - Add Apply Guardrail Testing Playground by @ishaan-jaff in #16030
- [Fix] SQS Logger - Add Base64 handling by @ishaan-jaff in #16028
- Fix mutation of original request for gemini request by @Sameerlite in #16002
- Fix: Redact reasoning summaries in ResponsesAPI output when message logging is disabled by @Sameerlite in #15965
- fix: Support text.format parameter in Responses API for providers without native ResponsesAPIConfig by @rodolfo-nobrega in #16023
- Remove unnecessary model variable assignment by @Mte90 in #16008
- Add license metadata to health/readiness endpoint. by @bernata in #15997
- chore(deps): bump hono from 4.9.7 to 4.10.3 in /litellm-js/spend-logs by @dependabot[bot] in #15915
- docs: improve Grayswan guardrail documentation by @TeddyAmkie in #15875
- fix(apscheduler): prevent memory leaks from jitter and frequent job intervals by @jatorre in #15846
- Python entry-point for CustomLLM subclasses by @AlbertDeFusco in #15881
- Allow using ARNs when generation images via Bedrock by @komarovd95 in #15789
- Added fallback logic for detecting file content-type when S3 returns generic by @langpingxue in #15635
- fix: prevent httpx DeprecationWarning memory leak in AsyncHTTPHandler by @AlexsanderHamir in #16024
- [Feat] Add FAL AI Image Generations on LiteLLM by @ishaan-jaff in #16067
- Feat: Mistral API - add codestral-embed-2505 by @ishaan-jaff in #16071
- Fix spend tracking for OCR/aOCR requests (log
pages_processed+ recognizeOCRResponse) by @OrionCodeDev in #16070 - [Feat] Add support for Batch API Rate limiting - PR1 adds support for input based rate limits by @ishaan-jaff in #16075
- [Feature] UI - Validation for Proxy Base URL in SSO Settings by @yuneng-jiang in #16082
- [Feature] UI - Test Key UI Embeddings by @yuneng-jiang in #16065
- [Feature] UI - Add Key Type Select in Key Settings by @yuneng-jiang in #16034
- feat(guardrails): Add per-request profile overrides to PANW Prisma AIRS by @jroberts2600 in #16069
- docs: use custom-llm-provider header in examples by @tlecomte in #16055
- Improve(mcp): respect X-Forwarded- headers in OAuth endpoints by @talalryz in #16036
- Add OpenAI-compatible annotations support for Cohere v2 citations by @Sameerlite in #16038
- Opik user auth key metadata Documentation by @Thomas-Mildner in #16004
- [Fix] Circle CI UI E2E Test Fix by @yuneng-jiang in #16100
- Fix: Moderations endpoint now respects
api_baseconfiguration parameter by @Sameerlite in #16087 - [Infra] Litellm Backend SSO Changes by @yuneng-jiang in #16029
New Contributors
- @RobGeada made their first contribution in #15975
- @shanto12 made their first contribution in #15946
- @dima-hx430 made their first contribution in #15976
- @m-misiura made their first contribution in #15971
- @ylgibby made their first contribution in #15947
- @Somtom made their first contribution in #15909
- @rodolfo-nobrega made their first contribution in #16023
- @bernata made their first contribution in #15997
- @AlbertDeFusco made their first contribution in #15881
- @komarovd95 made their first contribution in #15789
- @langpingxue made their first contribution in #15635
- @OrionCodeDev made their first contribution in #16070
Full Changelog: v1.79.0-nightly...v1.79.1.dev1
v1.77.7.dev3
What's Changed
- 1-79-0 docs by @ishaan-jaff in #15936
- feat(lasso): Upgrade to Lasso API v3 and fix ULID generation by @oroxenberg in #15941
- Enable OpenTelemetry context propagation by external tracers by @eycjur in #15940
- Fix documentation for videos by @Sameerlite in #15937
- Fix duplicate trace in langfuse_otel by @eycjur in #15931
- [Feat] add support for dynamic client registration by @uc4w6c in #15921
- Update IBM Guardrails to correctly use SSL Verify argument by @RobGeada in #15975
- feat: support during_call for model armor guardrails by @bjornjee in #15970
- docs(openrouter): add base_url config with environment variables by @shanto12 in #15946
- [Buf fix] - Azure OpenAI, fix ContextWindowExceededError is not mapped from Azure openai errors by @ishaan-jaff in #15981
- [Fix] DD logging - ensure key's metadata + guardrail is logged on DD by @ishaan-jaff in #15980
- [Feat] OTEL - Ensure error information is logged on OTEL by @ishaan-jaff in #15978
- [Fix] Minor fix proxy - ensure User API key and team id and user id missing from custom callback is not misfiring by @ishaan-jaff in #15982
- [Fix] Azure OpenAI - Add handling for
v1under azure api versions by @ishaan-jaff in #15984 - Fix: Respect
LiteLLM-Disable-Message-Redactionheader for Responses API by @Sameerlite in #15966 - [Feat] UI - Changed API Base from Select to Input in New LLM Credentials by @yuneng-jiang in #15987
- [Bug Fix] Remove limit from admin UI numerical input fix by @yuneng-jiang in #15991
- [Feature] UI - Key Already Exist Error Notification by @yuneng-jiang in #15993
- [Fix] - Responses API - add /openai routes for responses API. (Azure OpenAI SDK Compatibility) by @ishaan-jaff in #15988
- Add deprecation dates for models by @dima-hx430 in #15976
- docs(guardrails/ibm_guardrails): add additional detail to
ibm_guardrails.mdby @m-misiura in #15971 - Perf speed up pytest by @uc4w6c in #15951
- fix: Preserve Bedrock inference profile IDs in health checks by @ylgibby in #15947
- Fix: Support tool usage messages with Langfuse OTEL integration by @eycjur in #15932
- Add Haiku 4.5 pricing for open router by @Somtom in #15909
- fix(opik): enhance requester metadata retrieval from API key auth by @Thomas-Mildner in #15897
- [feat]: graceful degradation for pillar service when using litellm by @afogel in #15857
- Add GitlabPromptCache and enable subfolder access by @deepanshululla in #15712
- Add OpenAI client usage documentation for videos and fix navigation visibility by @Sameerlite in #15996
- [Feature] Config Models should not be editable by @yuneng-jiang in #16020
- [Fix] Guardrails - Ensure Key Guardrails are applied by @ishaan-jaff in #16025
- [UI] Feature - Add Apply Guardrail Testing Playground by @ishaan-jaff in #16030
- [Fix] SQS Logger - Add Base64 handling by @ishaan-jaff in #16028
- Fix mutation of original request for gemini request by @Sameerlite in #16002
- Fix: Redact reasoning summaries in ResponsesAPI output when message logging is disabled by @Sameerlite in #15965
- fix: Support text.format parameter in Responses API for providers without native ResponsesAPIConfig by @rodolfo-nobrega in #16023
- Remove unnecessary model variable assignment by @Mte90 in #16008
- Add license metadata to health/readiness endpoint. by @bernata in #15997
- chore(deps): bump hono from 4.9.7 to 4.10.3 in /litellm-js/spend-logs by @dependabot[bot] in #15915
- docs: improve Grayswan guardrail documentation by @TeddyAmkie in #15875
- fix(apscheduler): prevent memory leaks from jitter and frequent job intervals by @jatorre in #15846
- Python entry-point for CustomLLM subclasses by @AlbertDeFusco in #15881
- Allow using ARNs when generation images via Bedrock by @komarovd95 in #15789
- Added fallback logic for detecting file content-type when S3 returns generic by @langpingxue in #15635
- fix: prevent httpx DeprecationWarning memory leak in AsyncHTTPHandler by @AlexsanderHamir in #16024
- [Feat] Add FAL AI Image Generations on LiteLLM by @ishaan-jaff in #16067
- Feat: Mistral API - add codestral-embed-2505 by @ishaan-jaff in #16071
New Contributors
- @RobGeada made their first contribution in #15975
- @shanto12 made their first contribution in #15946
- @dima-hx430 made their first contribution in #15976
- @m-misiura made their first contribution in #15971
- @ylgibby made their first contribution in #15947
- @Somtom made their first contribution in #15909
- @rodolfo-nobrega made their first contribution in #16023
- @bernata made their first contribution in #15997
- @AlbertDeFusco made their first contribution in #15881
- @komarovd95 made their first contribution in #15789
- @langpingxue made their first contribution in #15635
Full Changelog: v1.79.0-nightly...v1.77.7.dev3