-
Notifications
You must be signed in to change notification settings - Fork 1.7k
[TRTLLM-5331] large-scale EP: perf - Replace allgather with AllToAllPrepare #5570
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Signed-off-by: Fred Wei <[email protected]>
/bot run |
PR_Github #10235 [ run ] triggered by Bot |
PR_Github #10235 [ run ] completed with state |
juney-nvidia
approved these changes
Jun 30, 2025
ameynaik-hub
pushed a commit
to ameynaik-hub/TensorRT-LLM
that referenced
this pull request
Jun 30, 2025
Signed-off-by: Fred Wei <[email protected]>
Shunkangz
pushed a commit
to Shunkangz/TensorRT-LLM
that referenced
this pull request
Jul 2, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 9, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 10, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 10, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 10, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 10, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 11, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 11, 2025
Signed-off-by: Fred Wei <[email protected]>
dominicshanshan
pushed a commit
to dominicshanshan/TensorRT-LLM
that referenced
this pull request
Jul 11, 2025
Signed-off-by: Fred Wei <[email protected]>
nvzhihanj
pushed a commit
to nvzhihanj/TensorRT-LLM
that referenced
this pull request
Jul 17, 2025
Signed-off-by: Fred Wei <[email protected]>
nvzhihanj
added a commit
that referenced
this pull request
Jul 22, 2025
) Signed-off-by: Fred Wei <[email protected]> Co-authored-by: WeiHaocheng <[email protected]>
3 tasks
wenscarl
pushed a commit
to flashinfer-ai/flashinfer
that referenced
this pull request
Aug 27, 2025
<!-- .github/pull_request_template.md --> ## 📌 Description This PR adds `mnnvl_moe_alltoallv_prepare_without_allgather` from [TensorRT-LLM](NVIDIA/TensorRT-LLM#5570). This is a more efficient way to prepare alltoallv info. ## 🔍 Related Issues <!-- Link any related issues here --> ## 🚀 Pull Request Checklist Thank you for contributing to FlashInfer! Before we review your pull request, please make sure the following items are complete. ### ✅ Pre-commit Checks - [x] I have installed `pre-commit` by running `pip install pre-commit` (or used your preferred method). - [x] I have installed the hooks with `pre-commit install`. - [x] I have run the hooks manually with `pre-commit run --all-files` and fixed any reported issues. > If you are unsure about how to set up `pre-commit`, see [the pre-commit documentation](https://pre-commit.com/). ## 🧪 Tests ``` tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[0-2-16-20-8-512] PASSED [ 69%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[0-2-16-16-3-300] PASSED [ 73%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[0-4-20-24-8-4000] PASSED [ 78%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[0-8-96-96-8-1000] PASSED [ 82%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[3-8-128-128-8-1000] PASSED [ 86%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[3-8-128-144-8-1] PASSED [ 91%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[0-4-72-80-4-2256] PASSED [ 95%] tests/test_trtllm_alltoall.py::test_moe_alltoall_prepare[0-4-72-80-6-3333] PASSED [100%] ``` ## Reviewer Notes <!-- Optional: anything you'd like reviewers to focus on, concerns, etc. -->
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
PR title
Please write the PR title by following template:
[JIRA ticket link/nvbug link/github issue link][fix/feat/doc/infra/...] <summary of this PR>
For example, assume I have a PR hope to support a new feature about cache manager of Jira TRTLLM-1000 ticket, it would be like
[TRTLLM-1000][feat] Support a new feature about cache manager
Description
Please explain the issue and the solution in short.
Test Coverage
GitHub Bot Help
/bot [-h] ['run', 'kill', 'skip', 'reuse-pipeline'] ...
Provide a user friendly way for developers to interact with a Jenkins server.
Run
/bot [-h|--help]
to print this help message.See details below for each supported subcommand.
run [--disable-fail-fast --skip-test --stage-list "A10-1, xxx" --gpu-type "A30, H100_PCIe" --add-multi-gpu-test --only-multi-gpu-test --disable-multi-gpu-test --post-merge --extra-stage "H100_PCIe-[Post-Merge]-1, xxx"]
Launch build/test pipelines. All previously running jobs will be killed.
--disable-fail-fast
(OPTIONAL) : Disable fail fast on build/tests/infra failures.--skip-test
(OPTIONAL) : Skip all test stages, but still run build stages, package stages and sanity check stages. Note: Does NOT update GitHub check status.--stage-list "A10-1, xxx"
(OPTIONAL) : Only run the specified test stages. Examples: "A10-1, xxx". Note: Does NOT update GitHub check status.--gpu-type "A30, H100_PCIe"
(OPTIONAL) : Only run the test stages on the specified GPU types. Examples: "A30, H100_PCIe". Note: Does NOT update GitHub check status.--only-multi-gpu-test
(OPTIONAL) : Only run the multi-GPU tests. Note: Does NOT update GitHub check status.--disable-multi-gpu-test
(OPTIONAL) : Disable the multi-GPU tests. Note: Does NOT update GitHub check status.--add-multi-gpu-test
(OPTIONAL) : Force run the multi-GPU tests. Will also run L0 pre-merge pipeline.--post-merge
(OPTIONAL) : Run the L0 post-merge pipeline instead of the ordinary L0 pre-merge pipeline.--extra-stage "H100_PCIe-[Post-Merge]-1, xxx"
(OPTIONAL) : Run the ordinary L0 pre-merge pipeline and specified test stages. Examples: --extra-stage "H100_PCIe-[Post-Merge]-1, xxx".For guidance on mapping tests to stage names, see
docs/source/reference/ci-overview.md
.kill
kill
Kill all running builds associated with pull request.
skip
skip --comment COMMENT
Skip testing for latest commit on pull request.
--comment "Reason for skipping build/test"
is required. IMPORTANT NOTE: This is dangerous since lack of user care and validation can cause top of tree to break.reuse-pipeline
reuse-pipeline
Reuse a previous pipeline to validate current commit. This action will also kill all currently running builds associated with the pull request. IMPORTANT NOTE: This is dangerous since lack of user care and validation can cause top of tree to break.