jundaf2 / INT8-Flash-Attention-FMHA-Quantization

151 stars 16 forks source link