flashinfer-ai / flashinfer

FlashInfer: Kernel Library for LLM Serving
https://flashinfer.ai
Apache License 2.0
760 stars 64 forks source link

bugfix: fix cascade test #315

Closed yzh119 closed 2 weeks ago

yzh119 commented 2 weeks ago

Cascade inference test was failed for a while, this PR fixes it.

Also fixes some of formats issues of previous PR #310.