-
Notifications
You must be signed in to change notification settings - Fork 368
[mxfp8 moe training] integrate mxfp8 dim0 triton kernel #3129
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[mxfp8 moe training] integrate mxfp8 dim0 triton kernel #3129
Conversation
51b9be2 to
168d4b7
Compare
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/3129
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: This comment was automatically generated by Dr. CI and updates every 15 minutes. |
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/3129
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: ❌ 1 New FailureAs of commit 168d4b7 with merge base cd21d0e ( NEW FAILURE - The following job has failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
c62b0f0 to
ff0f0c7
Compare
| key=["n_rows", "n_cols", "INNER_BLOCK_SIZE"], | ||
| ) | ||
| @triton.jit | ||
| def to_mxfp8_dim0_kernel( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
isn't this added in the previous PR
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This needs a rebase, will get back to it shortly (the original devgpu with the stack-pr stack is being re-imaged, so I'm having to manage the remote branches individually from a differnet devgpu now).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think i'm just going to close this and reopen a fresh PR with these changes to get around some weird git history/branch issues
cbf6277 to
527317f
Compare
Stacked PRs:
[mxfp8 moe training] integrate mxfp8 dim0 triton kernel
Test plan
pytest test/prototype/moe_training/test_scaled_grouped_mm.py -k dq -sBenchmarks