pytorch / FBGEMM

FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/
Other
1.2k stars 498 forks source link

FP8 cutlass lowering enablement #2946

Open samanamp opened 3 months ago

samanamp commented 3 months ago

Summary: Currently when lowering with FP8 linear modules, we get AssertionError:mutable f8f8bf16_rowwise.default is not supported with cpp_wrapper. Adding this change, bypasses the error.

Differential Revision: D60916286

netlify[bot] commented 3 months ago

Deploy Preview for pytorch-fbgemm-docs ready!

Name Link
Latest commit 18a3d866ca712136d960810495d2fd46e42b91d0
Latest deploy log https://app.netlify.com/sites/pytorch-fbgemm-docs/deploys/66b3c621e1282f000822644a
Deploy Preview https://deploy-preview-2946--pytorch-fbgemm-docs.netlify.app
Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

facebook-github-bot commented 3 months ago

This pull request was exported from Phabricator. Differential Revision: D60916286

facebook-github-bot commented 3 months ago

This pull request was exported from Phabricator. Differential Revision: D60916286