forked from pytorch/FBGEMM
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Block-wise FP8 matmul (pytorch#2780)
Summary: Pull Request resolved: pytorch#2780 Introduce a CUTLASS-based matmul for block-scaled fp8 tensors. This is based on the regular ("slow" accum) fp8 matmul in CUTLASS, with its fp8 accumulator class changed to do a fused multiply-and-add instead of a regular add into the global accumulator. This required changes throughout the stack, which is why I ended up copying sizeable chunks of CUTLASS into this diff. Reviewed By: ipiszy, jiawenliu64 Differential Revision: D57965065 fbshipit-source-id: 0b92b2ac1b3c687f23e820ea05255149dac686dc
- Loading branch information
1 parent
24140d5
commit 5a5b0e6
Showing
4 changed files
with
1,882 additions
and
30 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.