Skip to content

Conversation

@Aya-ZIbra
Copy link
Contributor

Summary: This diff adds support for 64 head dimension in the Blackwell Decode attention algorithm. The code changes include a dispatch macro for head dimension and a test case for the new head dimension. The test case is skipped for known numerical precision issues with FP8 and head_dim=64 in GQA mode.

Differential Revision: D86774487

Summary: This diff adds support for 64 head dimension in the Blackwell Decode attention algorithm. The code changes include a dispatch macro for head dimension and a test case for the new head dimension. The test case is skipped for known numerical precision issues with FP8 and head_dim=64 in GQA mode.

Differential Revision: D86774487
@netlify
Copy link

netlify bot commented Nov 11, 2025

Deploy Preview for pytorch-fbgemm-docs ready!

Name Link
🔨 Latest commit abadbd0
🔍 Latest deploy log https://app.netlify.com/projects/pytorch-fbgemm-docs/deploys/69136bbd4ae7f900084b2e5c
😎 Deploy Preview https://deploy-preview-5114--pytorch-fbgemm-docs.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

@meta-codesync
Copy link
Contributor

meta-codesync bot commented Nov 11, 2025

@Aya-ZIbra has exported this pull request. If you are a Meta employee, you can view the originating Diff in D86774487.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant