Skip to content

Conversation

danielvegamyhre
Copy link
Contributor

@danielvegamyhre danielvegamyhre commented Aug 28, 2025

Summary

  • Add conditional logic to benchmarks and tests to only run if the SM arch supports the underlying GEMMs that are going to run (when applicable), to address float8 rowwise scaled grouped mm doesn't support B200 #2904 (torch._scaled_grouped_mm not supported on B200 yet)
  • Make all MoE training tests have identical comprehensive parameterization where applicable, for robustness + consistency. All tests now have cases for all combinations of:
    • recipes: mxfp8, fp8 rowwise
    • execution: eager, compiled
    • target fqns: routed experts only, routed experts + shared experts
    • distributed tests cover FDSP, TP, FSDP+TP. EP tests coming next.
  • Add TFLOPs and speedup calculations to grouped gemms bench script
  • In benchmark_scaled_grouped_mm_dq.py split into (1) fwd time, and (2) e2e fwd+bwd time. Backward 2d-2d GEMM is not ready yet so I want to measure + optimize forward.
  • Add more tests to test_everything.sh and ensure everything passes.

Test plan

  • ./test/prototype/moe_training/test_everything.sh

@danielvegamyhre danielvegamyhre added the topic: not user facing Use this tag if you don't want this PR to show up in release notes label Aug 28, 2025
Copy link

pytorch-bot bot commented Aug 28, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/2905

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

⏳ No Failures, 7 Pending

As of commit 7eb6bfd with merge base 083d0c3 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Aug 28, 2025
@danielvegamyhre danielvegamyhre force-pushed the danielvegamyhre/stack/64 branch from 843448d to 02e246a Compare August 29, 2025 17:16
Copy link
Contributor

@vkuzo vkuzo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

stamping prototype code

@danielvegamyhre danielvegamyhre force-pushed the danielvegamyhre/stack/64 branch from 02e246a to e64d4b5 Compare August 29, 2025 18:53
@danielvegamyhre danielvegamyhre changed the base branch from danielvegamyhre/stack/64 to main August 29, 2025 20:02
@danielvegamyhre danielvegamyhre merged commit 568c193 into main Aug 29, 2025
18 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. topic: not user facing Use this tag if you don't want this PR to show up in release notes
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants