pytorch / executorch

On-device AI across mobile, embedded and edge for PyTorch
https://pytorch.org/executorch/
Other
1.39k stars 228 forks source link

Add fp16 qb4w scalar kernels #3626

Closed GregoryComer closed 4 weeks ago

GregoryComer commented 1 month ago

Summary: Add qb4w scalar 1x2, 1x4, 1x8, 2x2, 2x4, 2x8, and 4x4 kernels to XNNPACK. Add test coverage in ExecuTorch op-level linear test coverage for 4-bit blockwise weights / fp16.

Differential Revision: D57335871

pytorch-bot[bot] commented 1 month ago

:link: Helpful Links

:test_tube: See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/3626

Note: Links to docs will display an error until the docs builds have been completed.

:white_check_mark: No Failures

As of commit 055cde2fabd57e021b0fc2b229f15129d25fa75d with merge base 47a29a13f4f2574b6787be03249633d5531b795a (image): :green_heart: Looks good so far! There are no failures yet. :green_heart:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

facebook-github-bot commented 1 month ago

This pull request was exported from Phabricator. Differential Revision: D57335871

facebook-github-bot commented 1 month ago

This pull request was exported from Phabricator. Differential Revision: D57335871

facebook-github-bot commented 1 month ago

This pull request was exported from Phabricator. Differential Revision: D57335871

digantdesai commented 4 weeks ago

Thanks!

facebook-github-bot commented 4 weeks ago

This pull request was exported from Phabricator. Differential Revision: D57335871

facebook-github-bot commented 4 weeks ago

This pull request has been merged in pytorch/executorch@4b7c6db5d284d7bcca9b5a6db45758517fa1031d.