Skip to content

[pull] main from BerriAI:main#17

Open
pull[bot] wants to merge 4919 commits into0xnxxh:mainfrom
BerriAI:main
Open

[pull] main from BerriAI:main#17
pull[bot] wants to merge 4919 commits into0xnxxh:mainfrom
BerriAI:main

Conversation

@pull
Copy link
Copy Markdown

@pull pull bot commented Feb 13, 2026

See Commits and Changes for more details.


Created by pull[bot] (v2.0.0-alpha.4)

Can you help keep this open source service alive? 💖 Please sponsor : )

@pull pull bot locked and limited conversation to collaborators Feb 13, 2026
@pull pull bot added the ⤵️ pull label Feb 13, 2026
ryan-crabbe-berri and others added 28 commits April 10, 2026 08:55
refactor: consolidate route auth for UI and API tokens
Align with aget_agent_card and the DEFAULT_A2A_AGENT_TIMEOUT env var so
A2A message/send uses the same default as agent card fetch instead of a
hardcoded 60s HTTP read timeout.

Also correct aget_agent_card docstring for the timeout parameter.

Made-with: Cursor
Filter all explicitly-passed keys from remaining_kwargs before
spreading into async_responses_websocket(). The router now injects
custom_llm_provider into kwargs (via #25334), which collides with
the explicit custom_llm_provider= argument.
- Removed unused imports and streamlined type hints in `litellm/utils.py` and `litellm/files/main.py`.
- Moved `FileContentStreamingResult` to a new `litellm/files/types.py` for better organization.
- Updated `FileContentStreamingResponse` in `litellm/files/streaming.py` to include asynchronous close methods and improved logging capabilities.
- Enhanced tests to ensure proper closure of streaming iterators in `tests/test_litellm/llms/openai/test_openai_file_content_streaming.py` and `tests/test_litellm/proxy/openai_files_endpoint/test_files_endpoint.py`.
Add advisor-tool-2026-03-01 to anthropic_beta_headers_config.json so the
beta headers manager forwards it to Anthropic (was being silently dropped).
Mark as null for all non-native providers.
[Fix] Responses WebSocket Duplicate Keyword Argument Error
…-prompt-double-count

fix(bedrock): avoid double-counting cache tokens in Anthropic Messages streaming usage
…eld_handling

[Fix] Align v1 guardrail and agent list responses with v2 field handling
…l absent

Prevents Anthropic 400 invalid_request_error on follow-up turns where
the caller has removed the advisor tool but message history still contains
server_tool_use(advisor) + advisor_tool_result blocks.
bump: version 1.83.5 → 1.83.6
[Fix] Flush Tremor Tooltip timers in user_edit_view tests
yuneng-berri and others added 30 commits April 14, 2026 16:32
…v1_83_7rc1

[Docs] Add release notes for v1.83.3-stable and v1.83.7.rc.1
…t-role

fix: default invite user modal global role to least-privilege
…seline

The previous v1.83.3 changelog was generated against v1.83.0-nightly and
missed ~3 weeks of work. This regenerates it against the previous stable
release and restructures the LLM API Endpoints section to group by API
type (Responses, Batch, Count Tokens, Video Generation, Pass-Through,
etc.) matching the convention used in v1.82.3, v1.82.0, and v1.81.14.
Adds ~25 previously uncited PRs, cross-section duplications for
cross-cutting changes, and a verified first-time-contributors list.
[Docs] Regenerate v1.83.3-stable release notes from previous stable
[Refactor] Remove Chat UI link from Swagger docs message
…Qwen3.5-9B

Mixtral-8x7B-Instruct-v0.1 is no longer on Together AI's serverless tier
and now requires a dedicated endpoint, causing multiple tests to fail in CI:

  - test_together_ai.py::TestTogetherAI::test_empty_tools
  - test_completion.py::test_completion_together_ai_stream
  - test_completion.py::test_customprompt_together_ai
  - test_completion.py::test_completion_custom_provider_model_name
  - test_text_completion.py::test_async_text_completion_together_ai

Qwen/Qwen3.5-9B is currently serverless on Together AI and supports
function calling, satisfying BaseLLMChatTest capability requirements.
…odel

[Fix] Test - Together AI: replace deprecated Mixtral with serverless Qwen3.5-9B
Adds a GHA that fails PRs to main unless the head branch is
'litellm_internal_staging' or 'litellm_hotfix_*'. Also fails merge_group
events since merge queue is not in use.
Bedrock GPT-OSS occasionally emits truncated toolUse.input deltas
(e.g. accumulated args of '{"":"'), which causes
test_function_calling_with_tool_response to hard-fail on json.loads.
Other overrides in TestBedrockGPTOSS already handle similar
model-side flakiness; apply retries=6 delay=5 scoped to this subclass
so other providers keep strict behavior.
GPT-OSS on Bedrock intermittently emits truncated toolUse.input deltas
(e.g. accumulated args of '{"":"'), causing
test_function_calling_with_tool_response to hard-fail on json.loads.
The model flakiness is not a litellm regression: the same base test
passes for Anthropic in the same CI run, and the streaming delta path
at invoke_handler.py has not changed recently.

Follow the existing override pattern in TestBedrockGPTOSS
(test_prompt_caching, test_completion_cost, test_tool_call_no_arguments)
and stub the test to pass. The underlying bedrock converse streaming
tool-call path is already covered by Claude/Nova/Llama Converse suites
in test_bedrock_completion.py and test_bedrock_llama.py, so removing
the live GPT-OSS check loses no unique litellm-side signal.
Complements the stubbed-out live integration test by verifying the
outgoing Bedrock Converse request body for GPT-OSS is well-formed when
the caller supplies a tool schema with OpenAI-style metadata
($id, $schema, additionalProperties, strict):
- correct converse URL for bedrock/converse/openai.gpt-oss-20b-1:0
- toolConfig.tools[0].toolSpec has the expected name/description
- inputSchema.json keeps type/properties/required and strips fields
  Bedrock does not accept
…Call

[Test] Replace flaky bedrock gpt-oss tool-call live test with request-body mock
…age-path

fix: remove non-existent litellm_mcps_tests_coverage from coverage combine
…eout

fix(ci): increase test-server-root-path timeout to 30m
bump: version 1.83.7 → 1.83.8
[Infra] Guard main to only accept PRs from staging and hotfix branches
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Projects

None yet

Development

Successfully merging this pull request may close these issues.