Closed jeffbolznv closed 2 months ago
Ping for review.
I'm curious, is it better to do these sort of fixes in llama.cpp or in ggml?
Sorry for the delay. Pinging @0cc4m as they are the main maintainer of the vulkan backend. If we don't get a response in a day or two, will try to review myself - please ping again if necessary.
I'm curious, is it better to do these sort of fixes in llama.cpp or in ggml?
llama.cpp
has more eyes in general, so chances for review are better. Although some of the Vulkan PRs are also stale over there (https://github.com/ggerganov/llama.cpp/pull/9407).
GGML_VULKAN_RUN_TESTS currently has build errors like:
and runtime assertions like:
Fix these, and also add teraflops calculation to the matmul tests.