Open
Conversation
refactor: consolidate route auth for UI and API tokens
Align with aget_agent_card and the DEFAULT_A2A_AGENT_TIMEOUT env var so A2A message/send uses the same default as agent card fetch instead of a hardcoded 60s HTTP read timeout. Also correct aget_agent_card docstring for the timeout parameter. Made-with: Cursor
Filter all explicitly-passed keys from remaining_kwargs before spreading into async_responses_websocket(). The router now injects custom_llm_provider into kwargs (via #25334), which collides with the explicit custom_llm_provider= argument.
…AGENT_TIMEOUT Made-with: Cursor
- Removed unused imports and streamlined type hints in `litellm/utils.py` and `litellm/files/main.py`. - Moved `FileContentStreamingResult` to a new `litellm/files/types.py` for better organization. - Updated `FileContentStreamingResponse` in `litellm/files/streaming.py` to include asynchronous close methods and improved logging capabilities. - Enhanced tests to ensure proper closure of streaming iterators in `tests/test_litellm/llms/openai/test_openai_file_content_streaming.py` and `tests/test_litellm/proxy/openai_files_endpoint/test_files_endpoint.py`.
…s streaming usage Made-with: Cursor
merge main
…03_01 beta header enum
…ader in /messages path
Add advisor-tool-2026-03-01 to anthropic_beta_headers_config.json so the beta headers manager forwards it to Anthropic (was being silently dropped). Mark as null for all non-native providers.
[Fix] Responses WebSocket Duplicate Keyword Argument Error
…-prompt-double-count fix(bedrock): avoid double-counting cache tokens in Anthropic Messages streaming usage
…eld_handling [Fix] Align v1 guardrail and agent list responses with v2 field handling
…l absent Prevents Anthropic 400 invalid_request_error on follow-up turns where the caller has removed the advisor tool but message history still contains server_tool_use(advisor) + advisor_tool_result blocks.
bump: version 1.83.5 → 1.83.6
[Fix] Flush Tremor Tooltip timers in user_edit_view tests
[Infra] Merge Dev Branch with Main
…v1_83_7rc1 [Docs] Add release notes for v1.83.3-stable and v1.83.7.rc.1
…t-role fix: default invite user modal global role to least-privilege
…seline The previous v1.83.3 changelog was generated against v1.83.0-nightly and missed ~3 weeks of work. This regenerates it against the previous stable release and restructures the LLM API Endpoints section to group by API type (Responses, Batch, Count Tokens, Video Generation, Pass-Through, etc.) matching the convention used in v1.82.3, v1.82.0, and v1.81.14. Adds ~25 previously uncited PRs, cross-section duplications for cross-cutting changes, and a verified first-time-contributors list.
[Docs] Regenerate v1.83.3-stable release notes from previous stable
[Refactor] Remove Chat UI link from Swagger docs message
…Qwen3.5-9B Mixtral-8x7B-Instruct-v0.1 is no longer on Together AI's serverless tier and now requires a dedicated endpoint, causing multiple tests to fail in CI: - test_together_ai.py::TestTogetherAI::test_empty_tools - test_completion.py::test_completion_together_ai_stream - test_completion.py::test_customprompt_together_ai - test_completion.py::test_completion_custom_provider_model_name - test_text_completion.py::test_async_text_completion_together_ai Qwen/Qwen3.5-9B is currently serverless on Together AI and supports function calling, satisfying BaseLLMChatTest capability requirements.
…odel [Fix] Test - Together AI: replace deprecated Mixtral with serverless Qwen3.5-9B
fallbacks image
Adds a GHA that fails PRs to main unless the head branch is 'litellm_internal_staging' or 'litellm_hotfix_*'. Also fails merge_group events since merge queue is not in use.
Bedrock GPT-OSS occasionally emits truncated toolUse.input deltas
(e.g. accumulated args of '{"":"'), which causes
test_function_calling_with_tool_response to hard-fail on json.loads.
Other overrides in TestBedrockGPTOSS already handle similar
model-side flakiness; apply retries=6 delay=5 scoped to this subclass
so other providers keep strict behavior.
GPT-OSS on Bedrock intermittently emits truncated toolUse.input deltas
(e.g. accumulated args of '{"":"'), causing
test_function_calling_with_tool_response to hard-fail on json.loads.
The model flakiness is not a litellm regression: the same base test
passes for Anthropic in the same CI run, and the streaming delta path
at invoke_handler.py has not changed recently.
Follow the existing override pattern in TestBedrockGPTOSS
(test_prompt_caching, test_completion_cost, test_tool_call_no_arguments)
and stub the test to pass. The underlying bedrock converse streaming
tool-call path is already covered by Claude/Nova/Llama Converse suites
in test_bedrock_completion.py and test_bedrock_llama.py, so removing
the live GPT-OSS check loses no unique litellm-side signal.
Complements the stubbed-out live integration test by verifying the outgoing Bedrock Converse request body for GPT-OSS is well-formed when the caller supplies a tool schema with OpenAI-style metadata ($id, $schema, additionalProperties, strict): - correct converse URL for bedrock/converse/openai.gpt-oss-20b-1:0 - toolConfig.tools[0].toolSpec has the expected name/description - inputSchema.json keeps type/properties/required and strips fields Bedrock does not accept
…Call [Test] Replace flaky bedrock gpt-oss tool-call live test with request-body mock
…l_fallbacks docs update
…age-path fix: remove non-existent litellm_mcps_tests_coverage from coverage combine
…eout fix(ci): increase test-server-root-path timeout to 30m
bump: version 1.83.7 → 1.83.8
[Infra] Guard main to only accept PRs from staging and hotfix branches
Litellm day 0 opus 4.7 support
Litellm hotfix opus 4.7
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
See Commits and Changes for more details.
Created by
pull[bot] (v2.0.0-alpha.4)
Can you help keep this open source service alive? 💖 Please sponsor : )