Open riverzhou opened 9 months ago
@riverzhou the last llama.cpp has changed llama_backend_init to have bool argument, the version 0.0.7 is updated to match it
@riverzhou the last llama.cpp has changed llama_backend_init to have bool argument, the version 0.0.7 is updated to match it
They removed numa argument for llama_backend_init at Feb 16. In my test, b2277 do not have this argument.
commit f486f6e1e5e9d01603d9325ab3e05f1edb362a95
Author: bmwl <brian.marshall@tolko.com>
Date: Fri Feb 16 01:31:07 2024 -0800
ggml : add numa options (#5377)
diff --git a/llama.h b/llama.h
index 4a26bd61..f4ec6ea6 100644
--- a/llama.h
+++ b/llama.h
@@ -312,7 +312,10 @@ extern "C" {
// Initialize the llama + ggml backend
// If numa is true, use NUMA optimizations
// Call once at the start of the program
- LLAMA_API void llama_backend_init(bool numa);
+ LLAMA_API void llama_backend_init(void);
+
+ //optional:
+ LLAMA_API void llama_numa_init(enum ggml_numa_strategy numa);
// Call once at the end of the program - currently only used for MPI
@riverzhou the last llama.cpp has changed llama_backend_init to have bool argument, the version 0.0.7 is updated to match it
They removed numa argument for llama_backend_init at Feb 16. In my test, b2277 do not have this argument.
commit f486f6e1e5e9d01603d9325ab3e05f1edb362a95 Author: bmwl <brian.marshall@tolko.com> Date: Fri Feb 16 01:31:07 2024 -0800 ggml : add numa options (#5377) diff --git a/llama.h b/llama.h index 4a26bd61..f4ec6ea6 100644 --- a/llama.h +++ b/llama.h @@ -312,7 +312,10 @@ extern "C" { // Initialize the llama + ggml backend // If numa is true, use NUMA optimizations // Call once at the start of the program - LLAMA_API void llama_backend_init(bool numa); + LLAMA_API void llama_backend_init(void); + + //optional: + LLAMA_API void llama_numa_init(enum ggml_numa_strategy numa); // Call once at the end of the program - currently only used for MPI
I checked your source code. Both in 0.0.6 and 0.0.7, they have numa argument. So they can not work on upstream llama.cpp after Feb 16.
void llama_backend_init(
bool numa,
) {
return _llama_backend_init(
numa,
);
}
that weird, I will double check
@riverzhou you are correct, turns out my git pull
did not update llama.cpp
code, I had to hard reset, please try the last update
@riverzhou you are correct, turns out my
git pull
did not updatellama.cpp
code, I had to hard reset, please try the last update
Great! Thanks!
Version: llama_cpp_dart 0.0.6 llama.cpp tag: b2277
logcat: