Open chutaklee opened 2 years ago
assert query.dtype == torch.half, "sparse attention only supports training in fp16 currently, please file a github issue if you need fp32 support"
Any update on fp32 support for spase attention? Currently I'm happy with training sparse model on fp16 but it seems to me fp32 is more versatile.
I can work on this but I'm not familiar with triton and gpu programming. Could you point out some hints at least?
Bumping this! I would really appreciate fp32 support!
Any update on fp32 support for spase attention? Currently I'm happy with training sparse model on fp16 but it seems to me fp32 is more versatile.
I can work on this but I'm not familiar with triton and gpu programming. Could you point out some hints at least?