Skip to content

Conversation

rootfs
Copy link
Collaborator

@rootfs rootfs commented Sep 15, 2025

What type of PR is this?

When testing different reasoning models, the reason bench needs to adjust max token to avoid partial responses that result in wrong answers to many datasets include MMLU and GPQA

What this PR does / why we need it:

Which issue(s) this PR fixes:

Fixes #

Release Notes: Yes/No

Copy link

netlify bot commented Sep 15, 2025

Deploy Preview for vllm-semantic-router ready!

Name Link
🔨 Latest commit 2bc24ff
🔍 Latest deploy log https://app.netlify.com/projects/vllm-semantic-router/deploys/68c81e57a4e263000851174b
😎 Deploy Preview https://deploy-preview-137--vllm-semantic-router.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

Copy link

👥 vLLM Semantic Team Notification

The following members have been identified for the changed files in this PR and have been automatically assigned:

📁 bench

Owners: @yuezhu1, @Xunzhuo
Files changed:

  • bench/vllm_semantic_router_bench/router_reason_bench_multi_dataset.py

vLLM

🎉 Thanks for your contributions!

This comment was automatically generated based on the OWNER files in the repository.

@rootfs rootfs marked this pull request as draft September 15, 2025 23:22
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants