Closed acon96 closed 2 months ago
Fix startup error when using llama.cpp backend and add flash attention to llama.cpp backend
Fix startup error when using llama.cpp backend and add flash attention to llama.cpp backend