Skip to content

[WIP] Updated DSv4 vllm B300 MTP#1271

Open
wzhao18 wants to merge 3 commits intomainfrom
nv/dsv4-b300-mtp-new
Open

[WIP] Updated DSv4 vllm B300 MTP#1271
wzhao18 wants to merge 3 commits intomainfrom
nv/dsv4-b300-mtp-new

Conversation

@wzhao18
Copy link
Copy Markdown
Collaborator

@wzhao18 wzhao18 commented May 4, 2026

No description provided.

@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented May 4, 2026

Thanks for the contribution! For vLLM & SGLang, please ensure that your recipes is similar to the official vLLM recipes and/or the SGLang cookbook

If it is not, please create a PR first before we can merge your PR into the master branch. Let's ensure that the documentation is first class such that the entire ML community can benefit from your hard work! Thank you

PR authors are responsible for ensuring that after merging, all GitHub Action jobs fully pass. A lot of the time, failures are just flakes and simply re-running the failed jobs will fix it. If re-running failed jobs is attempted, PR authors are responsible for ensuring it passes. See GitHub's docs on re-running failed jobs: https://docs.github.com/en/actions/how-tos/manage-workflow-runs/re-run-workflows-and-jobs#re-running-failed-jobs-in-a-workflow

As a rule of thumb, generally, PR authors should request a review & get a PR approval from the respective companies' CODEOWNERS before requesting a review from core maintainers.

If additional help is needed, PR authors can reach out to core maintainers over Slack.

@wzhao18 wzhao18 marked this pull request as ready for review May 4, 2026 02:17
@wzhao18 wzhao18 requested a review from a team May 4, 2026 02:17
@claude
Copy link
Copy Markdown
Contributor

claude Bot commented May 4, 2026

Claude finished @wzhao18's task in 1m 28s —— View job


Review of PR #1271

  • Gather context and read changed files
  • Validate master config changes
  • Validate perf-changelog.yaml
  • Validate benchmark script

Summary

LGTM - no blocking issues found.

What this PR does: Upgrades the vLLM image from v0.20.0-cu130 to v0.20.1-cu130 for the B300 MTP config and adds new data-parallel attention + expert parallelism search space entries. The benchmark script is updated to conditionally set --moe-backend deep_gemm_mega_moe and MAX_NUM_BATCHED_TOKENS=2048 when DP_ATTENTION=true.

Checks passed:

  • Image vllm/vllm-openai:v0.20.1-cu130 is publicly accessible on Docker Hub
  • Model prefix dsv4 is valid
  • perf-changelog.yaml entry is correctly appended at the end of the file
  • Expert parallelism is correctly conditional on EP_SIZE (pre-existing)
  • --dsv4 flag implies --use-chat-template (verified in benchmark_lib.sh)
  • Server launch command is properly formatted with line continuations

@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented May 4, 2026

@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented May 4, 2026

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

Status: No status

Development

Successfully merging this pull request may close these issues.

1 participant