-
Notifications
You must be signed in to change notification settings - Fork 501
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Make some fbgemm fp8 triton ops pt2 friendly #3188
Conversation
This pull request was exported from Phabricator. Differential Revision: D63560103 |
✅ Deploy Preview for pytorch-fbgemm-docs ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
9544784
to
4f38802
Compare
This pull request was exported from Phabricator. Differential Revision: D63560103 |
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Reviewed By: jwfromm Differential Revision: D63560103
This pull request was exported from Phabricator. Differential Revision: D63560103 |
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Reviewed By: jwfromm Differential Revision: D63560103
4f38802
to
c6b13bc
Compare
This pull request was exported from Phabricator. Differential Revision: D63560103 |
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Differential Revision: D63560103
c6b13bc
to
259dd02
Compare
This pull request was exported from Phabricator. Differential Revision: D63560103 |
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Differential Revision: D63560103
259dd02
to
0385aa4
Compare
This pull request was exported from Phabricator. Differential Revision: D63560103 |
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Reviewed By: henrylhtsang Differential Revision: D63560103
0385aa4
to
63a4721
Compare
This pull request was exported from Phabricator. Differential Revision: D63560103 |
63a4721
to
48569d0
Compare
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Reviewed By: henrylhtsang Differential Revision: D63560103
Summary: X-link: facebookresearch/FBGEMM#283 Pull Request resolved: pytorch#3188 Make some fbgemm fp8 triton ops pt2 friendly.. # What this diff tries to do * stop using TensorWrapper and tl.reinterpret * Remove the use of triton_heuristics for _kernel_matmul_fp8_row # What this diff won't help: * triton_herustics use cases of EVEN_K. One option is to just merge that into the autotuning configs # need to do in the future: * Update other ops, like quantize_fp8_row. * Update documentation. Feels pretty outdated, and some still reference to TensorWrapper. Reviewed By: henrylhtsang Differential Revision: D63560103
This pull request was exported from Phabricator. Differential Revision: D63560103 |
48569d0
to
be457d5
Compare
This pull request has been merged in d27acbd. |
Summary:
Make some fbgemm fp8 triton ops pt2 friendly..
What this diff tries to do
What this diff won't help:
need to do in the future:
Differential Revision: D63560103