issues
search
flashinfer-ai
/
flashinfer
FlashInfer: Kernel Library for LLM Serving
https://flashinfer.ai
Apache License 2.0
768
stars
64
forks
source link
Revert "feat: support cuda graph for batched multi-query(prefill/append) attention"
#276
Closed
yzh119
closed
1 month ago
yzh119
commented
1 month ago
Reverts flashinfer-ai/flashinfer#275
Reverts flashinfer-ai/flashinfer#275