Adjust MiniMax MI355X block size for TP8 EP8#1228
Adjust MiniMax MI355X block size for TP8 EP8#1228jiacao-amd wants to merge 3 commits intoSemiAnalysisAI:mainfrom
Conversation
c01e0b6 to
d66409b
Compare
|
@jiacao-amd Kicking off a sweep. Run: https://github.com/SemiAnalysisAI/InferenceX/actions/runs/25121853688 |
|
/sweep test-config --config-files .github/configs/amd-master.yaml --config-keys minimaxm2.5-fp8-mi355x-vllm |
|
@jiacao-amd Kicking off a sweep. Run: https://github.com/SemiAnalysisAI/InferenceX/actions/runs/25330781481 |
9c6bfd2 to
37962d4
Compare
|
/sweep test-config --config-files .github/configs/amd-master.yaml --config-keys minimaxm2.5-fp8-mi355x-vllm |
|
@jiacao-amd Kicking off a sweep. Run: https://github.com/SemiAnalysisAI/InferenceX/actions/runs/25336346089 |
|
Superseded by #1276. The replacement PR uses the same MiniMax MI355X vLLM scheduling change, but the branch is pushed directly to SemiAnalysisAI/InferenceX instead of the fork so CI/automation should avoid the fork-PR permission issues. |
Summary
Testing