flashinfer-ai / flashinfer

FlashInfer: Kernel Library for LLM Serving
https://flashinfer.ai
Apache License 2.0
1.21k stars 111 forks source link

perf: optimize warp layout for prefill operator #304

Closed yzh119 closed 3 months ago

yzh119 commented 3 months ago

185 is out of sync with main branch, this PR rebase #185 on main.