Open LeiWang1999 opened 2 months ago
@LeiWang1999 Would you be able to share a minimal example for using FP8 W8A8? I would love to test this!
I'm recently working on quantizing an FP8 W/A language model, this project has been a great help to me, thanks :)
Hello @mgoin, it's a pleasant surprise to discover this project. Thank you for your contributions to BitBLAS. We have recently added support for FP8 Matmul, hoping it will help this project.